Yung-Sung Chuang
Yung-Sung Chuang

MIT EECS PhD Student @ CSAIL
Office: 32-G436

Yung-Sung Chuang (莊永松) How to pronounce?

Hi! I'm a fourth-year PhD student in Electrical Engineering and Computer Science at Massachusetts Institute of Technology, where I work with Jim Glass at CSAIL.

My research primarily focuses on natural language processing and large language models (LLMs), with a particular interest in improving their factuality and reliability. In DoLa, we proposed a decoding strategy that enhances LLM factuality by contrasting the knowledge across different transformer layers. In Lookback Lens, we introduced a method that detects and mitigates contextual hallucinations in LLMs using attention maps. Most recently, in SelfCite, we developed a self-supervised framework that enables LLMs to generate fine-grained, sentence-level citations by leveraging context ablation as a reward signal.

I also explore retrieval-based approaches to strengthen LLM by grounding answers in real documents. For instance, in Expand, Rerank, and Retrieve, we proposed query reranking to achieve more accurate retrieval results for open-domain QA. In DiffCSE, we built a contrastive learning method based on the differences between similar sentences to further boost the quality of sentence embeddings.

I was fortunate to intern at FAIR Meta, Microsoft, and MIT-IBM Watson AI Lab. Before joining MIT, I was an undergraduate student in Electrical Engineering at National Taiwan University, where I worked with Hung-Yi Lee, Yun-Nung (Vivian) Chen, and Lin-shan Lee. Here is my Curriculum Vitae.

Email: yungsung [AT] mit.edu

Links: [CV] [Twitter] [Github] [Google Scholar] [DBLP] [Blog] [Linkedin] [Instagram]

Recent News


Publications

For a full list of papers, see my Google Scholar.

2025

  • SelfCite: Self-Supervised Alignment for Context Attribution in Large Language Models
    Yung-Sung Chuang, Benjamin Cohen-Wang, Shannon Zejiang Shen, Zhaofeng Wu, Hu Xu, Xi Victoria Lin, James Glass, Shang-Wen Li, Wen-tau Yih
    ArXiv preprint, 2025.
    [bib] [abstract] [pdf] [code]

2024

  • Lookback Lens: Detecting and Mitigating Contextual Hallucinations in Large Language Models Using Only Attention Maps
    Yung-Sung Chuang Linlu Qiu, Cheng-Yu Hsieh, Ranjay Krishna, Yoon Kim, James Glass
    In The Conference on Empirical Methods in Natural Language Processing, 2024.
    [bib] [abstract] [pdf] [code]
  • DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models
    Yung-Sung Chuang, Yujia Xie, Hongyin Luo, Yoon Kim, James Glass, Pengcheng He
    In The Twelfth International Conference on Learning Representations (ICLR), 2024.
    [bib] [abstract] [pdf] [code]
  • Curiosity-driven Red-teaming for Large Language Models
    Zhang-Wei Hong, Idan Shenfeld, Tsun-Hsuan Wang, Yung-Sung Chuang, Aldo Pareja, James Glass, Akash Srivastava, Pulkit Agrawal
    In The Twelfth International Conference on Learning Representations (ICLR), 2024.
    [bib] [abstract] [pdf] [code]
  • Joint Dual Learning With Mutual Information Maximization for Natural Language Understanding and Generation in Dialogues
    Shang-Yu Su, Yung-Sung Chung, Yun-Nung Chen
    In IEEE/ACM Transactions on Audio, Speech, and Language Processing (Volume: 32), pages 2445 - 2452, March 2024.
    [bib] [abstract] [pdf]

2023

  • On Robustness-Accuracy Characterization of Large Language Models using Synthetic Datasets
    Ching-Yun Ko, Pin-Yu Chen, Payel Das, Yung-Sung Chuang, Luca Daniel
    In Workshop on Efficient Systems for Foundation Models@ ICML2023 , July 2023.
    [bib] [abstract] [pdf]
  • Revealing the Blind Spot of Sentence Encoder Evaluation by HEROS
    Cheng-Han Chiang, Yung-Sung Chuang, James Glass, Hung-yi Lee
    In Proceedings of the 8th Workshop on Representation Learning for NLP (RepL4NLP 2023) , 2023.
    [bib] [abstract] [pdf] [data]
  • Expand, Rerank, and Retrieve: Query Reranking for Open-Domain Question Answering
    Yung-Sung Chuang, Wei Fang, Shang-Wen Li, Wen-tau Yih, James Glass
    In The 61st Annual Meeting of the Association for Computational Linguistics: Findings, 2023.
    [bib] [abstract] [pdf] [code]
  • SAIL: Search-Augmented Instruction Learning
    Hongyin Luo, Yung-Sung Chuang, Yuan Gong, Tianhua Zhang, Yoon Kim, Xixin Wu, Danny Fox, Helen Meng, James Glass
    In Findings of the Association for Computational Linguistics: EMNLP, 2023.
    [bib] [abstract] [pdf] [code] [demo] [web]
  • Interpretable Unified Language Checking
    Hongyin Luo, Tianhua Zhang, Yung-Sung Chuang, Wei Fang, Luc Gaitskell, Thomas Hartvigsen, Xixin Wu, Danny Fox, Helen Meng, James Glass
    In arXiv preprint, 2023.
    [bib] [abstract] [pdf] [code]
  • C2KD: Cross-Lingual Cross-Modal Knowledge Distillation for Multilingual Text-Video Retrieval
    Andrew Rouditchenko, Yung-Sung Chuang, Nina Shvetsova, Samuel Thomas, Rogerio Feris, Brian Kingsbury, Leonid Karlinsky, David Harwath, Hilde Kuehne, James Glass
    In IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2023.
    [bib] [abstract] [pdf] [code & dataset]
  • Visual Language Pretrained Multiple Instance Zero-Shot Transfer for Histopathology Images
    Ming Y Lu, Bowen Chen, Andrew Zhang, Drew FK Williamson, Richard J Chen, Tong Ding, Long Phi Le, Yung-Sung Chuang, Faisal Mahmood
    In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023.
    [bib] [abstract] [pdf] [supplementary material]

2022

  • DUAL: Discrete Spoken Unit Adaptive Learning for Textless Spoken Question Answering
    Guan-Ting Lin, Yung-Sung Chuang, Ho-Lam Chung, Shu-wen Yang, Hsuan-Jui Chen, Shuyan Dong, Shang-Wen Li, Abdelrahman Mohamed, Hung-yi Lee, Lin-shan Lee.
    In Interspeech, 2022.
    [bib] [abstract] [pdf] [code]

2021

2020

  • Lifelong Language Knowledge Distillation
    Yung-Sung Chuang, Shang-Yu Su, Yun-Nung Chen.
    In The Conference on Empirical Methods in Natural Language Processing (EMNLP), 2020.
    [bib] [abstract] [pdf] [code] [video]

  • Dual Inference for Improving Language Understanding and Generation
    Yung-Sung Chuang*, Shang-Yu Su*, Yun-Nung Chen.
    In The Conference on Empirical Methods in Natural Language Processing: Findings (EMNLP), 2020.
    [bib] [abstract] [pdf] [code]

  • SpeechBERT: An Audio-and-text Jointly Learned Language Model for End-to-end Spoken Question Answering
    Yung-Sung Chuang, Chi-Liang Liu, Hung-Yi Lee, Lin-shan Lee.
    In Interspeech, 2020. (Interspeech 2020 Travel Grant)
    [bib] [abstract] [pdf] [video]

2019


Honors

  • Dean’s list (4 times), Electrical Engineering Dept. at NTU, Spring ’18, Spring ’19, Fall ’19, Spring ’20
  • Irving T. Ho Memorial Scholarship (2 times), EECS at NTU, Fall ’18, Fall ’19
  • Travel Grant, INTERSPEECH 2020 conference, Sep. 2020
  • Appier Best Application Award, 2020 NTU CSIE Undergrad Special Research Exhibition, Jun. 2020
  • 2nd Place & Appier 1st Award, 2019 NTU CSIE Undergrad Special Research Exhibition, Jun. 2019
  • 2nd Place, 2019 NTUEE Undergraduate Innovation Award, Jun. 2019
  • 1st Place, 2018 H. Spectrum Demo Day (out of 21 teams), Jul. 2018
  • 1st Place, NCTS Health Hackathon 2018 (out of 18 teams), Jun. 2018
  • Top 8 Finalist, Microsoft Imagine Cup Taiwan National Final 2018, Apr. 2018
  • Best Tech Award & Microsoft Enterprise Award, MakeNTU 2018 (out of 50 teams), Mar. 2018
  • 1st place of Dept. of Transportation, HackNTU 2017 (out of 100+ teams), Jul. 2017

Services

  • Reviewer:
    NeurIPS 2021, 2022, 2023, 2024
    ICLR 2022, 2023, 2024, 2025
    ICML 2022, 2023, 2024, 2025
    ACL ARR 2023, 2024, 2025
    EMNLP 2022, 2023
    ACL 2023
    AAAI 2023
    ICASSP 2022, 2023
    TASL 2023, 2024, 2025