Multilingual & Cross-Cultural AI

Retrieval-augmented methods for rare vocabulary and cross-lingual transfer.

Overview

Multilingual AI systems often fail on rare words, domain-specific terminology, and low-resource languages. My research develops retrieval-augmented methods that enable models to “look things up” during inference.


Current Projects

Retrieval-Augmented Speech Translation

Status: Published at EMNLP 2024 Paper
Key insight: Cross-modal retrieval (speech-to-speech) outperforms text-based retrieval for rare word translation.

What we built:

  • Knowledge-based retrieval-augmented end-to-end speech translation
  • Cross-modal retriever: speech-to-speech, speech-to-text, text-to-text
  • Zero-shot terminology translation for open-domain recognition

Results:

  • 17.6% improvement in rare word accuracy with gold examples
  • 8.5% improvement with retrieved examples
  • Speech-to-speech retrieval shows higher robustness to unseen speakers

Why it matters: Rare words (names, technical terms, neologisms) are where translation systems fail most catastrophically. Retrieval provides a path to reliable terminology handling without massive retraining.


Research Questions I’m Exploring

  1. Cross-lingual retrieval — Can we retrieve relevant examples across language pairs?
  2. Domain adaptation — How do retrieval-augmented methods transfer to new domains?
  3. Efficient indexing — How do we scale retrieval to millions of examples?