No Starch Press · Available Spring/Summer 2026

Post-Training:
A Practical Guide for
AI Engineers and Developers

Capable by default. Reliable by design.

If you’re a practitioner who has watched a promising AI demo fail to survive contact with production — where prompting hits its ceiling, retrieval isn’t enough, and the model still can’t be trusted with your domain — post-training is what you’ve been missing. This is a practical guide to turning foundation models into production-ready systems: reshaping behavior, aligning to your values, and deploying with confidence. Each technique is taught concept-first, then implementation-through-code, so you understand not just what to run, but what you’re actually changing inside the model.

§1

What you’ll learn

Fine-tune models on curated datasets using supervised fine-tuning, LoRA, and QLoRA without destroying the base model’s general capabilities

Apply reinforcement learning from verifiable rewards (RLVR) and modern preference optimisation methods including DPO, ORPO, and beyond, to shape model behaviour

Evaluate models rigorously: design benchmarks, detect regression, and measure quality claims that survive scrutiny

Adapt models to specialised domains — from clinical language to legal text — turning general capability into a defensible competitive advantage

Train agentic models that take sequences of actions reliably, not just models that talk about taking actions

Quantise and compress fine-tuned models for deployment without sacrificing the gains you trained for

§2

Why this book

Written with the enterprise practitioner in mind.

The literature on post-training is focused either on small educational use cases that do not consider enterprise realities, or presuppose the workflow of foundation labs. There’s nothing for the crucial middle: enterprise practitioners with real compute budgets who need to customise, align and deploy AI at scale. This book fills that gap.

Trade-offs, not best practices.

The book treats post-training decisions as trade-offs rather than best practices, helping practitioners match techniques to constraints. It provides decision frameworks, clearly documenting trade-offs and benefits.

From principles to practice.

Combines technical depth with strategic context. Includes companion Jupyter notebooks covering practical implementation. Shows how to embed proprietary knowledge, organisational values and domain expertise into foundation models.

§3

About the author

Chris von Csefalvay is a Principal at HCLTech’s AI Practice, where he leads post-training research and clinical intelligence. He has held senior data science leadership roles across major enterprises, published extensively on distributed computing for ML, and designed language models for applications ranging from pharmacovigilance to social dynamics. He is also the author of Computational Modeling of Infectious Disease (Elsevier, 2023), a monograph on computational epidemiology. He holds degrees from the University of Oxford and Cardiff University and is a Fellow of the Royal Society for Public Health and Senior Member of IEEE.

chrisvoncsefalvay.com
§4

What’s inside

Part I: The Foundation

  • Chapter 1: Post-Training Essentials: What It Is and Why It Matters
  • Chapter 2: Prerequisites for Success: Before You Fine-Tune

Part II: The Tools

  • Chapter 3: Supervised Fine-Tuning: The Foundation Technique
  • Chapter 4: Reinforcement Learning: Better Each Time
  • Chapter 5: Preference Optimization: Modern Alternatives to PPO
  • Chapter 6: Evaluation Strategies: Measuring Model Quality

Part III: The Craft

  • Chapter 7: Efficiency Techniques: Quantization and Compression
  • Chapter 8: Domain Adaptation: Make It Yours
  • Chapter 9: Agentic Models: Deeds, Not Words
  • Chapter 10: Reasoning Capabilities: Training for Complex Thought

Part IV: The Frontier

  • Chapter 11: Synthetic Training: Self-Play and Generated Data
  • Chapter 12: Multimodal Systems: Post-Training Beyond Text
  • Chapter 13: Future Directions: What Comes Next
§5

Get notified when the book launches

Be the first to know when the book is available for early access. No spam — just a single email when it’s ready.

Post-training is where models stop being impressive and start being useful.