LLMs in Production Audiobook By Christopher Brousseau, Matt Sharp cover art

LLMs in Production

Engineering AI Applications

Preview

Get this deal Try for $0.00
Offer ends April 30, 2025 at 11:59PM PT.
Prime logo Prime members: New to Audible? Get 2 free audiobooks during trial.
Pick 1 audiobook a month from our unmatched collection.
Listen all you want to thousands of included audiobooks, Originals, and podcasts.
Access exclusive sales and deals.
Premium Plus auto-renews for $14.95/mo after 3 months. Cancel anytime.
Pick 1 audiobook a month from our unmatched collection.
Listen all you want to thousands of included audiobooks, Originals, and podcasts.
Access exclusive sales and deals.
Premium Plus auto-renews for $14.95/mo after 30 days. Cancel anytime.

LLMs in Production

By: Christopher Brousseau, Matt Sharp
Narrated by: Christopher Kendrick
Get this deal Try for $0.00

$14.95/mo. after 3 months. Offer ends April 30, 2025 11:59PM PT. Cancel anytime.

$14.95/month after 30 days. Cancel anytime.

Buy for $24.95

Buy for $24.95

Confirm purchase
Pay using card ending in
By confirming your purchase, you agree to Audible's Conditions of Use and Amazon's Privacy Notice. Taxes where applicable.
Cancel

About this listen

Unlock the potential of Generative AI with this Large Language Model production-ready playbook for seamless deployment, optimization, and scaling. This hands-on guide takes you beyond theory, offering expert strategies for integrating LLMs into real-world applications using retrieval-augmented generation (RAG), vector databases, PEFT, LoRA, and scalable inference architectures. Whether you're an ML engineer, data scientist, or MLOps practitioner, you’ll gain the technical know-how to operationalize LLMs efficiently, reduce compute costs, and ensure rock-solid reliability in production.

What You’ll Learn:

  • Master LLM Fundamentals – Understand tokenization, transformer architectures, and the evolution linguistics to the creation of foundation models.
  • RAG & Vector Databases – Augment model capabilities with real-time retrieval and memory-optimized embeddings.
  • Training vs Fine-tuning – Learn how to train your own model as well as cutting edge techniques like Distillation, RLHF, PEFT, LoRA, and QLoRA for cost-effective adaptation.
  • Prompt Engineering – Discover the quickly evolving world of prompt engineering and go beyond simple prompt and pray methods and learn how to implement structured outputs, complex workflows, and LLM agents.
  • Scaling & Cost Optimization – Deploy LLMs into your favorite cloud of choice, on commodity hardware, Kubernetes clusters, and edge devices.
  • Securing AI Workflows – Implement guardrails for hallucination mitigation, adversarial testing, and compliance monitoring.
  • MLOps for LLMs – Learn all about LLMOps, automate model lifecycle management, retraining pipelines, and continuous evaluation.

Hands-on Projects Include:

• Training a custom LLM from scratch – Build and optimize an industry-specific model.

• AI-Powered VSCode Extension – Use LLMs to enhance developer productivity with intelligent code completion.

• Deploying on Edge Devices – Run a lightweight LLM on a Raspberry Pi or Jetson Nano for real-world AI applications.

©2024 Manning Publications (P)2025 Manning Publications
Computer Science Programming & Software Development Architecture Software Development
adbl_web_global_use_to_activate_webcro768_stickypopup

What listeners say about LLMs in Production

Average customer ratings

Reviews - Please select the tabs below to change the source of reviews.