Note
Access to this page requires authorization. You can try signing in or changing directories.
Access to this page requires authorization. You can try changing directories.
Intent recognition in Azure AI Speech was retired on September 30, 2025. Applications can no longer use intent recognition via Azure AI Speech. However, you can still perform intent recognition using Azure AI Language Service or Azure OpenAI.
This change doesn't affect other Azure AI Speech capabilities such as speech to text (including no change to speaker diarization), text to speech, and speech translation.
Azure AI Speech previously exposed the IntentRecognizer object family in the Speech SDK. These APIs depended on a Language Understanding Intelligent Service (LUIS) application or simple pattern matching constructs. With the retirement:
IntentRecognizer, pattern matching intents/entities, and related parameters are no longer available.- Existing applications must remove direct Speech SDK intent logic and adopt a two-step approach (speech to text, then intent classification) or a single prompt-based approach.
Choose an alternative
| Requirement | Recommended service | Why |
|---|---|---|
| Structured intent and entity extraction with labeled training data | Azure AI Language Service Conversational Language Understanding (CLU) | Purpose-built for multi-intent classification and entity extraction; supports versions, testing, and analytics. |
| Few-shot or zero-shot dynamic intent determination | Azure OpenAI | Use GPT models with example prompts; rapidly adapt without schema changes. |
| Combine transcription with generative reasoning (summaries + intents) | Azure OpenAI + Speech | Transcribe audio then enrich with GPT outputs for complex reasoning. |
| Multilingual speech input flowed into consistent intent schema | Speech (STT) + CLU | Speech handles transcription; CLU handles normalization and classification. |
Migration steps
- Replace any Speech SDK
IntentRecognizerusage withSpeechRecognizerorConversationTranscriberto obtain text. - For structured intent/entity needs, create a CLU project and deploy a model. Send transcribed utterances to the CLU prediction API.
- For flexible or rapid scenarios, craft a prompt for an Azure OpenAI model including representative user utterances and expected JSON intent output.
- Remove dependencies on
LanguageUnderstandingModeland any LUIS application IDs or endpoints from configuration. - Eliminate pattern matching code referencing
PatternMatchingIntentorPatternMatchingEntitytypes. - Validate accuracy by comparing historic
IntentRecognizeroutputs to CLU classification results or OpenAI completions, adjusting training data or prompts as needed. - Update monitoring: shift any existing intent latency/accuracy dashboards to new sources (CLU evaluation logs or OpenAI prompt result tracking).
Sample architecture
- Speech to text transcribes audio into text with real-time or batch mode.
- Text is sent to CLU or Azure OpenAI depending on your intent strategy.
- Response is normalized into a common JSON shape (for example:
{ "intent": "BookFlight", "entities": { "Destination": "Seattle" } }). - Business logic routes the normalized output to downstream services (booking, knowledge base, workflow engine).
Result format considerations
| Aspect | CLU | Azure OpenAI |
|---|---|---|
| Schema stability | High (defined intents/entities) | Flexible (prompt-defined) |
| Versioning | Built-in model versions | Manual prompt versioning |
| Training effort | Requires labeled dataset | Few-shot examples in prompt |
| Edge cases | Requires more labeled data | Add examples or instructions |
| Latency | Prediction API call | Completion API call (similar) |
Frequently asked questions
Do I need to re-label data? If you used LUIS, you need to export and reimport data into CLU, then retrain. Mapping is often direct (intents, entities). Pattern matching intents might require manual conversion to examples.
Can I combine CLU and Azure OpenAI? Yes. Use CLU for deterministic classification and OpenAI for summarization or fallback classification when confidence is low.
Is speaker diarization affected? No. Diarization features continue; you just process each speaker segment through CLU or OpenAI after transcription.