This talk explores building a complete self-hosted LLM stack in Rust: Paddler, a distributed load balancer for serving LLMs at scale, and Poet, a static site generator that consumes those LLMs for AI-powered content features.
This session we will delve into the sometimes murky world of procedural macros - showing some of the great tooling available for understanding the code generated, such as cargo expand, and the key building blocks we will need for writing our own.
I'll share a few tricks to help you write cleaner, more powerful declarative macros. You'll also get a sneak peek at the nightly features to see what's coming next macro_rules! world.
This talk explores what it means to write scientific software that lives up to the standards we expect of science itself.