Share via

Significant Regression in Code-Switching (EN-ZH) recognition for zh-HK after April 2026 Engine Update

Peter Ng 0 Reputation points
2026-04-22T11:53:03.27+00:00

Problem Description

We are reporting a critical regression in the Cantonese (zh-HK) Speech-to-Text service that started occurring after the March 31, 2026 API retirement and the subsequent rollout of the latest engine (MAI-Transcribe-1).

The Issue: Common English terms (e.g., "Wifi", "Meeting", "MTR") mixed within Cantonese sentences are being systematically filtered out or misrecognized as Cantonese homophones, even when using the latest Speech SDK (v1.49.1) and high Phrase List weights (2.0).

Technical Details

Region: East Asia (Hong Kong) / Southeast Asia

Locale: zh-HK (Cantonese Traditional)

SDK Version: JavaScript SDK 1.49.1 (Updated from 1.36.0)

Deployment Type: Base Model (and previously Custom Model)

Expected Behavior: "邊度有 Wifi 機?" (Where is the Wifi machine?)

Actual Behavior: "邊度有機?" (The word "Wifi" is completely omitted by the AI Post-Processor).

Evidence & Logs (Telemetry)

The following log shows that the Lexical (raw) output already fails to capture the English token, suggesting the issue is within the new decoder's language-purity constraint:

Result ID: [請貼上你 log 裡的 Result ID] Lexical (raw): 邊 度 有 機 ITN: 邊 度 有 機 Display: 邊度有機? Note: The speaker clearly said "Wifi 機", but "Wifi" was treated as disfluency and removed.

Steps we have already taken to troubleshoot:

SDK Upgrade: Upgraded to 1.49.1 to ensure compatibility with new service properties.

Phrase List: Added English terms to PhraseListGrammar with phraseListWeight set to 2.0 (via setServiceProperty).

Post-Processing: Toggled PostProcessingOption between "None" and "TrueText", but the English words remain missing.

Custom Model Removal: The issue persists even on the Base Model, indicating a global regression in the zh-HK engine's ability to handle code-switching.

Requested Action

Please investigate if the MAI-Transcribe-1 model for zh-HK has an over-aggressive language filter for English tokens.

Clarify if the property phraseListWeight is being correctly honored by the new engine for zh-HK.

  1. Provide a way to disable the mandatory "Semantic Segmentation" or "Disfluency Removal" that seems to be filtering out short English terms in Cantonese contexts.

    Problem Description

    We are reporting a critical regression in the Cantonese (zh-HK) Speech-to-Text service that started occurring after the March 31, 2026 API retirement and the subsequent rollout of the latest engine (MAI-Transcribe-1). The Issue: Common English terms (e.g., "Wifi", "Meeting", "MTR") mixed within Cantonese sentences are being systematically filtered out or misrecognized as Cantonese homophones, even when using the latest Speech SDK (v1.49.1) and high Phrase List weights (2.0).

    Technical Details

    • Region: East Asia (Hong Kong) / Southeast Asia
    • Locale: zh-HK (Cantonese Traditional)
    • SDK Version: JavaScript SDK 1.49.1 (Updated from 1.36.0)
    • Deployment Type: Base Model (and previously Custom Model)
    • Expected Behavior: "邊度有 Wifi 機?" (Where is the Wifi machine?)
    • Actual Behavior: "邊度有機?" (The word "Wifi" is completely omitted by the AI Post-Processor).

    Evidence & Logs (Telemetry)

    The following log shows that the Lexical (raw) output already fails to capture the English token, suggesting the issue is within the new decoder's language-purity constraint:

    Lexical (raw): 邊 度 有 機 ITN: 邊 度 有 機 Display: 邊度有機? Note: The speaker clearly said "Wifi 機", but "Wifi" was treated as disfluency and removed.

    Steps we have already taken to troubleshoot:

    1. SDK Upgrade: Upgraded to 1.49.1 to ensure compatibility with new service properties.
    2. Phrase List: Added English terms to PhraseListGrammar with phraseListWeight set to 2.0 (via setServiceProperty).
    3. Post-Processing: Toggled PostProcessingOption between "None" and "TrueText", but the English words remain missing.
    4. Custom Model Removal: The issue persists even on the Base Model, indicating a global regression in the zh-HK engine's ability to handle code-switching.

    Requested Action

    1. Please investigate if the MAI-Transcribe-1 model for zh-HK has an over-aggressive language filter for English tokens.
    2. Clarify if the property phraseListWeight is being correctly honored by the new engine for zh-HK.
    3. Provide a way to disable the mandatory "Semantic Segmentation" or "Disfluency Removal" that seems to be filtering out short English terms in Cantonese contexts.
Azure Speech in Foundry Tools

Your answer

Answers can be marked as 'Accepted' by the question author and 'Recommended' by moderators, which helps users know the answer solved the author's problem.