Result Details
Unifying Global and Near-Context Biasing in a Single Trie Pass
Villatoro-Tello Esau
Zuluaga Juan Pablo
Kumar Shashi
Burdisso Sergio
Rangappa Pradeep
Carofilis Andres
Madikeri Srikanth
Motlíček Petr, doc. Ing., Ph.D., DCGM (FIT)
Pandia Karthik
Hacioglu Kadri
Stolcke Andreas
Despite the success of end-to-end automatic speech recognition (ASR) models, challenges persist in recognizing rare, out-of-vocabulary wordsincluding named entities (NE)-and in adapting to new domains using only text data. This work presents a practical approach to address these challenges through an unexplored combination of an NE bias list and a word-level n-gram language model (LM). This solution balances simplicity and effectiveness, improving entities' recognition while maintaining or even enhancing overall ASR performance. We efficiently integrate this enriched biasing method into a transducer-based ASR system, enabling context adaptation with almost no computational overhead. We present our results on three datasets spanning four languages and compare them to state-of-the-art biasing strategies We demonstrate that the proposed combination of keyword biasing and n-gram LM improves entity recognition by up to 32% relative and reduces overall WER by up to a 12% relative.
Contextualisation and adaptation of ASR, real-time ASR, Aho-Corasick algorithm, Transformer-Transducer
@inproceedings{BUT201441,
author="{} and {} and {} and {} and {} and {} and {} and {} and Petr {Motlíček} and {} and {} and {}",
title="Unifying Global and Near-Context Biasing in a Single Trie Pass",
booktitle="Lecture Notes in Artificial Intelligence",
year="2026",
journal="Lecture Notes in Computer Science",
volume="16029",
pages="170--181",
publisher="Springer Nature",
address="CHAM",
doi="10.1007/978-3-032-02548-7\{_}15",
isbn="978-3-032-02547-0",
url="https://www.fit.vut.cz/research/group/speech/public/publi/2025/Iuliia_TSD2025_2025_co-author_Motlicek.pdf"
}