Front Matter
Front Matter

Who Should Read This Book

Now that you know what the book covers, the natural next question is: is it for you? If you have ever wondered how ChatGPT works under the hood, wanted to build an AI product but were not sure where to start, or needed to make informed decisions about LLM adoption at your company, the answer is yes. This book serves anyone who wants to build real systems with LLMs, regardless of whether your background is in software engineering, machine learning, product management, or research. Here is how different readers typically engage with the material.

Who This Book Is For

If you are an engineer focused on shipping, three recurring features deserve special mention. Library Shortcut boxes appear after each from-scratch implementation and show the equivalent one-liner in Hugging Face, LangChain, or another popular framework, so you can move quickly once you understand the concept. Warning boxes flag subtle production pitfalls (silent tokenizer mismatches, GPU memory cliffs, non-deterministic outputs) that can cost hours of debugging. And the appendices on Docker containerization, inference serving, and distributed ML provide step-by-step guides for the infrastructure side of deploying LLM systems.

What Background Is Assumed

RequiredNot Required (covered in the book)
Python proficiency (functions, classes, standard library) Prior machine learning experience (Chapter 0 covers this)
Basic linear algebra: vectors, matrices, dot products NLP background (Chapters 1 through 3 build from scratch)
Familiarity with APIs and JSON GPU programming or CUDA knowledge
Comfort with the command line and Git Deep learning framework experience (Chapter 0.3 teaches PyTorch)
Novelty: Coverage of 2024 through 2026 Developments

This book covers techniques and protocols that most textbooks have not yet addressed: the Model Context Protocol (MCP) for standardized tool integration, the AG-UI protocol for agent-user streaming interaction, reasoning models like OpenAI's o1/o3 and DeepSeek-R1 (with a dedicated chapter on reasoning and test-time compute in Chapter 08), automated prompt optimization with DSPy (in Section 11.3), vision-language-action models for robotics (in Section 28.7), and the latest advances in mixture-of-experts architectures. Each of these is treated not as a curiosity but as a production-relevant technology with code examples.

Personalized Pathways

Not sure where to start? The book offers 20 personalized reader pathways tailored to specific roles (ML Engineer, Product Builder, Startup CTO, Researcher, and more). Each pathway recommends a reading order and highlights the chapters most relevant to your goals. There are also 4 university course syllabi and 5 specialty tracks for instructors who want to use this book in academic settings.

What Comes Next

Ready to start? The next section explains the conventions used throughout every chapter. Proceed to FM.3: How to Use This Book to learn about callout types, code conventions, and labs, or jump to Problem-Solution Key if you want to look up a specific task.