Skip to main content
Intents, utterances, and entities are the three foundational building blocks of conversational language understanding systems. Understanding how they work together helps you design accurate intent classification and reliable entity extraction pipelines for chatbots, virtual assistants, and other conversational applications.
  • Keywords: intents, utterances, entities, conversational AI, natural language understanding (NLU), entity extraction, intent classification, slot filling.

What is an utterance?

An utterance is the raw piece of natural language input a user speaks or types to your application. It can be a question, command, or statement. Examples of utterances:
  • “What’s the weather like tomorrow?”
  • “Turn on the bedroom light.”
  • “Set the heater to 25 degrees Celsius.”

What is an intent?

An intent represents the user’s goal or purpose behind an utterance—the action the user wants the system to perform. Correct intent classification tells your system which flow, API call, or response to trigger. Example intents:
  • GetWeather — user requests weather information
  • TurnOnDevice — user issues a device activation command
  • AdjustDevice — user requests a device setting change
Mapping utterances to intents is the primary routing mechanism in an NLU system.

What is an entity?

Entities are structured, named data extracted from an utterance that provide context for an intent. Entities (also called slots) make responses precise and actionable. Examples:
  • Utterance: “What time is it in Paris?” → Intent: GetTime, Entity: Location = Paris
  • Utterance: “Will it rain tomorrow?” → Intent: GetWeather, Entity: Time = tomorrow
  • Utterance: “Set the heater to 25 degrees Celsius.” → Intent: AdjustDevice, Entities: Device = heater, Value = 25°C
A slide titled "Utterances, Intents, and Entities in Language Models" showing three labeled panels: 01 Utterances ("what users say in natural language"), 02 Intents ("the purpose or action behind an utterance"), and 03 Entities ("extract details from user input, making AI responses more precise").

Quick reference table: utterances → intents → entities

UtteranceLikely IntentExtracted Entity Example
”What’s the weather like tomorrow?”GetWeatherTime = tomorrow
”What time is it in Paris?”GetTimeLocation = Paris
”Turn on the bedroom light.”TurnOnDeviceDevice = bedroom light
”Set the heater to 25 degrees Celsius.”AdjustDeviceDevice = heater, Value = 25°C

Pre-built entity types and why they matter

Many platforms (including Azure Language services) provide pre-built entity extractors that recognize common data types without custom training. These speed up development and improve accuracy for standard patterns.
Pre-built entity extractors accelerate development by automatically detecting common data formats—numbers, dates, emails, phone numbers, and URLs—so you can focus training on domain-specific entities.
Common pre-built entity types and examples:
Entity TypeUse CaseExample
QuantitiesPercentages, counts, measures”Increase brightness to 50%” → Quantity = 50%
Date & TimeAbsolute and relative expressions”Remind me tomorrow at 7 p.m.” → DateTime = tomorrow at 7 p.m.
Email AddressesContact extraction”Contact me at user@domain.com” → Email = user@domain.com
Phone NumbersLocal and international formats”Call +1 234 567 8900” → PhoneNumber = +1 234 567 8900
URLsWeb addresses in text”Check out https://example.com” → URL = https://example.com
A presentation slide titled "Prebuilt Entity Components" containing three boxed panels. They list "Quantities" (numerical values), "Date and Time" (specific and relative times), and "Email Addresses," each with a simple icon.

Tips for designing intents and entities

  • Keep intents focused and action-oriented (e.g., GetWeather, BookFlight, AdjustDevice).
  • Design entities as the minimal pieces of context needed to fulfill the intent (e.g., Location, Time, Device, Value).
  • Use pre-built entities where applicable and reserve custom entities for domain-specific concepts.
  • Provide diverse utterance examples during training to cover synonyms, colloquialisms, and different phrasing.
Ambiguous utterances can lead to incorrect routing or extraction. Add disambiguation prompts in your dialog flow (e.g., “Do you mean Paris, France or Paris, Texas?”) and validate critical entities before taking irreversible actions.

Where to learn more

By combining clearly defined intents, well-scoped entities, and representative utterances—while leveraging pre-built extractors where suitable—you can significantly improve both the accuracy and usability of your conversational applications.

Watch Video