Overview

1 What makes conversational AI work?

This chapter sets the stage for building conversational AI that users actually want to use. It defines conversational AI and the main solution types—question answering, process-oriented workflows, and routing—and explains why so many assistants fail: they misread intent, impose unnecessary complexity, or trigger immediate opt-outs. The chapter introduces a simple, universal flow for successful systems: understand what the user wants, gather only the information needed, and deliver the result quickly and ethically. It emphasizes user-centered design, the interplay between intent models, dialogue, and APIs, and the importance of designing for the channel and context to reduce friction and personalize help.

The chapter then introduces generative AI as a complementary tool to classic techniques. Large language models can bolster intent understanding, simplify and improve copy, power retrieval-augmented answers, and accelerate builder workflows such as data augmentation and dialogue drafting. Because LLMs can be biased or hallucinate, the chapter stresses practical guardrails: choosing appropriate models and training data, adding contextual prompts, pre- and post-filtering for unsafe content, and keeping humans in the loop when risk is high. It also highlights the need to experiment with models and parameters for each task, optimizing for both performance and safety.

Finally, the chapter advocates a disciplined, continuous improvement cycle: measure, identify a problem tied to business outcomes, implement targeted fixes, deploy, and repeat. Small, incremental changes are preferred over large, risky overhauls because they deliver value sooner, are easier to diagnose, and create more learning opportunities. Success depends on improving the full chain—engagement, understanding, and fulfillment—and communicating progress in business terms. Teams should tie technical work to metrics like containment, average handle time, time to resolution, and customer satisfaction, ensuring stakeholders see clear, compounding value from ongoing enhancements.

A painful chat experience with a process-oriented bot that puts cognitive burden on the user. The AI has not provided any value in three conversational turns.
A delightful experience that uses context and reasonable assumptions to complete the user's goal quickly. The context could be loaded from a log-in process (chat) or from caller phone number (voice).
Flow diagram for conversational AI. In many use cases “additional information” includes user profile data.
Conversational AI logical architecture annotated with password reset example.
It takes a dream team with diverse skills to build an enterprise-ready conversational AI.
Adding context in the prompt is an important way to guide a large language model.
Impact of changing one LLM parameter (repetition penalty).
Cumulative success in a process is dependent on success in each of the individual steps. Visually it looks like a funnel that narrows after each step.
A continuous improvement lifecycle for conversational AI.
Large changes — like retraining all intents — take a long time and have less predictable outcomes.
Many small changes — like retraining one intent at a time — has a smaller “blast zone” for each change, bringing quicker value and more learning.
Area over the dotted line is additional business value over “big bang” change. Working code in production delivers value!

Summary

  • Conversational AI must be built with the user experience in mind. Good conversational AI helps users complete their tasks quickly. Bad conversational AI frustrates users.
  • There are thousands of generative AI models. Large language models are a subtype of generative AI models good at generating text.
  • LLMs can perform many tasks with impressive performance but also have significant risks including hallucination. It takes thoughtful guidance and guardrails to use LLMs effectively and responsibly.
  • LLM technology can supplement conversational AI. LLMs can respond to users directly and also assist you in building your conversational AI.
  • Continuous improvement is possible and necessary for effective conversational AI.
  • Iterative improvement delivers higher business value with lower risk.

FAQ

What are the main types of conversational AI?Three common types: 1) Question-answering (FAQ bots that reply directly), 2) Process-oriented or transactional assistants (guide users through multi-step tasks and often call APIs), and 3) Routing agents (triage and hand off to the right bot or human). Many real systems combine all three.
What are the top failure modes to watch for?The big three are: 1) The bot doesn’t understand user intent, 2) The flow puts too much complexity on the user, and 3) Users immediately opt out. Fixes include improving intent training with representative data, simplifying and personalizing dialogue with context, and writing concise, engaging copy.
How does a conversational AI work at a high level?It follows three steps: 1) Figure out what the user wants (NLU/intent recognition), 2) Gather needed information (dialogue management, state, and API orchestration), and 3) Fulfill the request (execute a transaction, answer, or route to a human).
What skills and components are needed to build an effective assistant?You need a cross‑functional team (design, data science, development, product/compliance). Core components include an intent classifier, well-defined APIs, and a conversation flow that collects the right info for fulfillment while respecting channel constraints and security.
When should I use generative AI versus classic techniques?Use classic intent/flow orchestration for reliable, procedural tasks; add generative AI to enhance understanding, generate answers from your content (RAG), summarize, and improve copy or training data. They work best together: generative AI augments, not replaces, classic approaches.
How do I reduce hallucinations and keep generative AI safe?Apply layered guardrails: pick suitable models and training data, pre-filter inputs for hate/abuse/profanity, provide clear contextual prompts (and retrieved documents), post-filter outputs, and, for higher risk, keep a human in the loop. Monitor before, during, and after deployment.
What is retrieval‑augmented generation (RAG) and why use it?RAG retrieves relevant documents from your trusted sources and uses them as context for an LLM to generate answers. It grounds responses in your content, improving relevance and reducing hallucinations for question‑answering use cases.
How do I choose the right LLM and parameters?Match the model to the task (generation, classification, extraction, summarization, RAG) and experiment across prompts and parameters (e.g., repetition penalty, temperature). Don’t generalize from a single test; evaluate cost, latency, and quality on representative inputs.
What does a continuous improvement cycle look like?Measure a baseline, identify a problem tied to a business metric, implement a targeted change, deploy, and repeat. Favor many small, reversible changes over big‑bang rewrites—they deliver value sooner, reduce risk, and create more learning opportunities.
Which business metrics should I use to show value to stakeholders?Link technical work to outcomes like containment, average handle time, human touches (for routing), time to resolution, NPS, and compliance. Communicate in business terms (impact on cost and satisfaction), not just technical metrics like intent F1 scores.

pro $24.99 per month

  • access to all Manning books, MEAPs, liveVideos, liveProjects, and audiobooks!
  • choose one free eBook per month to keep
  • exclusive 50% discount on all purchases
  • renews monthly, pause or cancel renewal anytime

lite $19.99 per month

  • access to all Manning books, including MEAPs!

team

5, 10 or 20 seats+ for your team - learn more


choose your plan

team

monthly
annual
$49.99
$399.99
only $33.33 per month
  • five seats for your team
  • access to all Manning books, MEAPs, liveVideos, liveProjects, and audiobooks!
  • choose another free product every time you renew
  • choose twelve free products per year
  • exclusive 50% discount on all purchases
  • renews monthly, pause or cancel renewal anytime
  • renews annually, pause or cancel renewal anytime
  • Effective Conversational AI ebook for free
choose your plan

team

monthly
annual
$49.99
$399.99
only $33.33 per month
  • five seats for your team
  • access to all Manning books, MEAPs, liveVideos, liveProjects, and audiobooks!
  • choose another free product every time you renew
  • choose twelve free products per year
  • exclusive 50% discount on all purchases
  • renews monthly, pause or cancel renewal anytime
  • renews annually, pause or cancel renewal anytime
  • Effective Conversational AI ebook for free
choose your plan

team

monthly
annual
$49.99
$399.99
only $33.33 per month
  • five seats for your team
  • access to all Manning books, MEAPs, liveVideos, liveProjects, and audiobooks!
  • choose another free product every time you renew
  • choose twelve free products per year
  • exclusive 50% discount on all purchases
  • renews monthly, pause or cancel renewal anytime
  • renews annually, pause or cancel renewal anytime
  • Effective Conversational AI ebook for free