An Azure service that integrates speech processing into apps and services.
Significant Regression in Code-Switching (EN-ZH) recognition for zh-HK after April 2026 Engine Update
Problem Description
We are reporting a critical regression in the Cantonese (zh-HK) Speech-to-Text service that started occurring after the March 31, 2026 API retirement and the subsequent rollout of the latest engine (MAI-Transcribe-1).
The Issue: Common English terms (e.g., "Wifi", "Meeting", "MTR") mixed within Cantonese sentences are being systematically filtered out or misrecognized as Cantonese homophones, even when using the latest Speech SDK (v1.49.1) and high Phrase List weights (2.0).
Technical Details
Region: East Asia (Hong Kong) / Southeast Asia
Locale: zh-HK (Cantonese Traditional)
SDK Version: JavaScript SDK 1.49.1 (Updated from 1.36.0)
Deployment Type: Base Model (and previously Custom Model)
Expected Behavior: "邊度有 Wifi 機?" (Where is the Wifi machine?)
Actual Behavior: "邊度有機?" (The word "Wifi" is completely omitted by the AI Post-Processor).
Evidence & Logs (Telemetry)
The following log shows that the Lexical (raw) output already fails to capture the English token, suggesting the issue is within the new decoder's language-purity constraint:
Result ID: [請貼上你 log 裡的 Result ID] Lexical (raw): 邊 度 有 機 ITN: 邊 度 有 機 Display: 邊度有機? Note: The speaker clearly said "Wifi 機", but "Wifi" was treated as disfluency and removed.
Steps we have already taken to troubleshoot:
SDK Upgrade: Upgraded to 1.49.1 to ensure compatibility with new service properties.
Phrase List: Added English terms to PhraseListGrammar with phraseListWeight set to 2.0 (via setServiceProperty).
Post-Processing: Toggled PostProcessingOption between "None" and "TrueText", but the English words remain missing.
Custom Model Removal: The issue persists even on the Base Model, indicating a global regression in the zh-HK engine's ability to handle code-switching.
Requested Action
Please investigate if the MAI-Transcribe-1 model for zh-HK has an over-aggressive language filter for English tokens.
Clarify if the property phraseListWeight is being correctly honored by the new engine for zh-HK.
- Provide a way to disable the mandatory "Semantic Segmentation" or "Disfluency Removal" that seems to be filtering out short English terms in Cantonese contexts.
Problem Description
We are reporting a critical regression in the Cantonese (zh-HK) Speech-to-Text service that started occurring after the March 31, 2026 API retirement and the subsequent rollout of the latest engine (MAI-Transcribe-1). The Issue: Common English terms (e.g., "Wifi", "Meeting", "MTR") mixed within Cantonese sentences are being systematically filtered out or misrecognized as Cantonese homophones, even when using the latest Speech SDK (v1.49.1) and high Phrase List weights (2.0).Technical Details
- Region: East Asia (Hong Kong) / Southeast Asia
- Locale: zh-HK (Cantonese Traditional)
- SDK Version: JavaScript SDK 1.49.1 (Updated from 1.36.0)
- Deployment Type: Base Model (and previously Custom Model)
- Expected Behavior: "邊度有 Wifi 機?" (Where is the Wifi machine?)
- Actual Behavior: "邊度有機?" (The word "Wifi" is completely omitted by the AI Post-Processor).
Evidence & Logs (Telemetry)
The following log shows that the Lexical (raw) output already fails to capture the English token, suggesting the issue is within the new decoder's language-purity constraint:Lexical (raw): 邊 度 有 機 ITN: 邊 度 有 機 Display: 邊度有機? Note: The speaker clearly said "Wifi 機", but "Wifi" was treated as disfluency and removed.
Steps we have already taken to troubleshoot:
- SDK Upgrade: Upgraded to 1.49.1 to ensure compatibility with new service properties.
- Phrase List: Added English terms to
PhraseListGrammarwithphraseListWeightset to 2.0 (viasetServiceProperty). - Post-Processing: Toggled
PostProcessingOptionbetween "None" and "TrueText", but the English words remain missing. - Custom Model Removal: The issue persists even on the Base Model, indicating a global regression in the
zh-HKengine's ability to handle code-switching.
Requested Action
- Please investigate if the MAI-Transcribe-1 model for
zh-HKhas an over-aggressive language filter for English tokens. - Clarify if the property
phraseListWeightis being correctly honored by the new engine forzh-HK. - Provide a way to disable the mandatory "Semantic Segmentation" or "Disfluency Removal" that seems to be filtering out short English terms in Cantonese contexts.