Edit

Share via


Migrate from retired intent recognition

Intent recognition in Azure AI Speech was retired on September 30, 2025. Applications can no longer use intent recognition via Azure AI Speech. However, you can still perform intent recognition using Azure AI Language Service or Azure OpenAI.

This change doesn't affect other Azure AI Speech capabilities such as speech to text (including no change to speaker diarization), text to speech, and speech translation.

Azure AI Speech previously exposed the IntentRecognizer object family in the Speech SDK. These APIs depended on a Language Understanding Intelligent Service (LUIS) application or simple pattern matching constructs. With the retirement:

  • IntentRecognizer, pattern matching intents/entities, and related parameters are no longer available.
  • Existing applications must remove direct Speech SDK intent logic and adopt a two-step approach (speech to text, then intent classification) or a single prompt-based approach.

Choose an alternative

Requirement Recommended service Why
Structured intent and entity extraction with labeled training data Azure AI Language Service Conversational Language Understanding (CLU) Purpose-built for multi-intent classification and entity extraction; supports versions, testing, and analytics.
Few-shot or zero-shot dynamic intent determination Azure OpenAI Use GPT models with example prompts; rapidly adapt without schema changes.
Combine transcription with generative reasoning (summaries + intents) Azure OpenAI + Speech Transcribe audio then enrich with GPT outputs for complex reasoning.
Multilingual speech input flowed into consistent intent schema Speech (STT) + CLU Speech handles transcription; CLU handles normalization and classification.

Migration steps

  1. Replace any Speech SDK IntentRecognizer usage with SpeechRecognizer or ConversationTranscriber to obtain text.
  2. For structured intent/entity needs, create a CLU project and deploy a model. Send transcribed utterances to the CLU prediction API.
  3. For flexible or rapid scenarios, craft a prompt for an Azure OpenAI model including representative user utterances and expected JSON intent output.
  4. Remove dependencies on LanguageUnderstandingModel and any LUIS application IDs or endpoints from configuration.
  5. Eliminate pattern matching code referencing PatternMatchingIntent or PatternMatchingEntity types.
  6. Validate accuracy by comparing historic IntentRecognizer outputs to CLU classification results or OpenAI completions, adjusting training data or prompts as needed.
  7. Update monitoring: shift any existing intent latency/accuracy dashboards to new sources (CLU evaluation logs or OpenAI prompt result tracking).

Sample architecture

  1. Speech to text transcribes audio into text with real-time or batch mode.
  2. Text is sent to CLU or Azure OpenAI depending on your intent strategy.
  3. Response is normalized into a common JSON shape (for example: { "intent": "BookFlight", "entities": { "Destination": "Seattle" } }).
  4. Business logic routes the normalized output to downstream services (booking, knowledge base, workflow engine).

Result format considerations

Aspect CLU Azure OpenAI
Schema stability High (defined intents/entities) Flexible (prompt-defined)
Versioning Built-in model versions Manual prompt versioning
Training effort Requires labeled dataset Few-shot examples in prompt
Edge cases Requires more labeled data Add examples or instructions
Latency Prediction API call Completion API call (similar)

Frequently asked questions

Do I need to re-label data? If you used LUIS, you need to export and reimport data into CLU, then retrain. Mapping is often direct (intents, entities). Pattern matching intents might require manual conversion to examples.

Can I combine CLU and Azure OpenAI? Yes. Use CLU for deterministic classification and OpenAI for summarization or fallback classification when confidence is low.

Is speaker diarization affected? No. Diarization features continue; you just process each speaker segment through CLU or OpenAI after transcription.