Yung-Sung Chuang

MIT EECS PhD Student @ CSAIL
Office: 32-G436

Email: yungsung [AT] mit.edu

Yung-Sung Chuang (莊永松) How to pronounce?

I am a final-year PhD student working with Jim Glass at MIT CSAIL. My research focuses on large language models: hallucinations, factuality, and retrieval-augmented generation. In addition, I worked on pre-training MetaCLIP 2, a multilingual vision-language model pre-trained on worldwide web-scale data, during my internship at Meta FAIR.

My research has introduced several approaches for improving LLM factuality. DoLa enhances factuality through layer-wise knowledge contrasting during decoding. Lookback Lens detects and mitigates hallucinations by analyzing attention patterns under RAG settings. Most recently, SelfCite enables LLMs to generate accurate citations without external supervision. I also used to work on retrieval methods, developing DiffCSE for better sentence embeddings and Query Reranking for more accurate passage retrieval.

Before MIT, I conducted research in speech processing and NLP with Hung-Yi Lee, Yun-Nung Chen, and Lin-shan Lee at National Taiwan University, where I obtained my B.S. degree in Electrical Engineering in 2020.

Recent News

Talks

  • Reducing Hallucinations in LLMs via Decoding, Detection, and Citation

    MIT CSAIL: EI Seminar, April 24, 2025
    A comprehensive overview of my research on reducing hallucinations in large language models through three key approaches: improved decoding strategies (DoLa), detection methods (Lookback Lens), and citation generation (SelfCite).

For my other talks at conferences, please see https://www.youtube.com/@yung-sung


Selected Publications

For a full list of papers, see my Google Scholar.

Multi-modal Large-scale Pre-training

  • MetaCLIP 2 teaser

    MetaCLIP 2: A Worldwide Scaling Recipe

    arXiv preprint arXiv:2507.22062, 2025.

Factuality & Transparency

  • Lookback Lens teaser

    Lookback Lens: Detecting and Mitigating Contextual Hallucinations in Large Language Models Using Only Attention Maps

  • DoLa teaser

    DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models

  • AT2 teaser

    Learning to Attribute with Attention

    arXiv preprint arXiv:2504.13752, 2025.

Retrieval-based Methods

  • SAIL teaser

    SAIL: Search-Augmented Instruction Learning

Speech Processing

  • DUAL teaser

    DUAL: Discrete Spoken Unit Adaptive Learning for Textless Spoken Question Answering

    In Interspeech, 2022.
  • Semi-Supervised Spoken Language Understanding teaser

    Semi-Supervised Spoken Language Understanding via Self-Supervised Speech and Language Model Pretraining

  • SUPERB teaser

    SUPERB: Speech processing Universal PERformance Benchmark

    In Interspeech, 2021.
  • SpeechBERT teaser

    SpeechBERT: An Audio-and-text Jointly Learned Language Model for End-to-end Spoken Question Answering

    Yung-Sung Chuang, Chi-Liang Liu, Hung-Yi Lee, Lin-shan Lee

Honors

Services

Reviewer

NeurIPS 2021, 2022, 2023, 2024, 2025 ICLR 2022, 2023, 2024, 2025 ICML 2022, 2023, 2024, 2025 ACL ARR 2023, 2024, 2025 EMNLP 2022, 2023 ACL 2023 AAAI 2023 ICASSP 2022, 2023 TASL 2023, 2024, 2025
This website is built from the source code of Nelson F. Liu's awesome website (nelsonliu.me).

BibTeX Citation