Timezone: »
In recent years, pre-trained large language models (LLMs) have demonstrated remarkable efficiency in achieving an inference-time few-shot learning capability known as in-context learning. However, existing literature has highlighted the sensitivity of this capability to the selection of few-shot demonstrations. Current understandings of the underlying mechanisms by which this capability arises from regular language model pretraining objectives remain disconnected from the real-world LLMs. This study aims to examine the in-context learning phenomenon through a Bayesian lens, viewing real-world LLMs as latent variable models. On this premise, we propose an algorithm to select optimal demonstrations from a set of annotated data with a small LM, and then directly generalize the selected demonstrations to larger LMs. We demonstrate significant improvement over baselines, averaged over eight GPT models on eight real-world text classification datasets. We also demonstrate the real-world usefulness of our algorithm on GSM8K, a math word problem dataset. Our empirical findings support our hypothesis that LLMs implicitly infer a latent variable containing task information.
Author Information
Xinyi Wang (University of California, Santa Barbara)
Wanrong Zhu (University of California, Santa Barbara)
Michael Saxon (UC Santa Barbara)
Mark Steyvers (UC Irvine)
William Yang Wang (University of California, Santa Barbara)
William Wang is the Co-Director of UC Santa Barbara's Natural Language Processing group and Center for Responsible Machine Learning. He is the Duncan and Suzanne Mellichamp Chair in Artificial Intelligence and Designs, and an Associate Professor in the Department of Computer Science at the University of California, Santa Barbara. He received his PhD from School of Computer Science, Carnegie Mellon University. He has broad interests in Artificial Intelligence, including statistical relational learning, information extraction, computational social science, dialog & generation, and vision. He has published more than 100 papers at leading NLP/AI/ML conferences and journals, and received best paper awards (or nominations) at ASRU 2013, CIKM 2013, EMNLP 2015, and CVPR 2019, a DARPA Young Faculty Award (Class of 2018), an IEEE AI's 10 to Watch Award (Class of 2020), an NSF CAREER Award (2021), two Google Faculty Research Awards (2018, 2019), three IBM Faculty Awards (2017-2019), two Facebook Research Awards (2018, 2019), an Amazon AWS Machine Learning Research Award, a JP Morgan Chase Faculty Research Award, an Adobe Research Award in 2018, and the Richard King Mellon Presidential Fellowship in 2011. He frequently serves as an Area Chair or Senior Area Chair for NAACL, ACL, EMNLP, and AAAI. He is an elected member of IEEE Speech and Language Processing Technical Committee (2021-2023) and a member of ACM Future of Computing Academy. In addition to research, William enjoys writing scientific articles that impact the broader online community. His work and opinions appear at major tech media outlets such as Wired, VICE, Scientific American, Fortune, Fast Company, NASDAQ, The Next Web, Law.com, and Mental Floss.
More from the Same Authors
-
2021 : VALUE: A Multi-Task Benchmark for Video-and-Language Understanding Evaluation »
Linjie Li · Jie Lei · Zhe Gan · Licheng Yu · Yen-Chun Chen · Rohit Pillai · Yu Cheng · Luowei Zhou · Xin Wang · William Yang Wang · Tamara L Berg · Mohit Bansal · Jingjing Liu · Lijuan Wang · Zicheng Liu -
2021 : A Dataset for Answering Time-Sensitive Questions »
Wenhu Chen · Xinyi Wang · William Yang Wang -
2022 : LAD: Language Augmented Diffusion for Reinforcement Learning »
Edwin Zhang · Yujie Lu · William Yang Wang · Amy Zhang -
2022 : Offline Reinforcement Learning with Closed-Form Policy Improvement Operators »
Jiachen Li · Edwin Zhang · Ming Yin · Qinxun Bai · Yu-Xiang Wang · William Yang Wang -
2022 : Off-policy Reinforcement Learning with Optimistic Exploration and Distribution Correction »
Jiachen Li · Shuo Cheng · Zhenyu Liao · Huayan Wang · William Yang Wang · Qinxun Bai -
2023 : ToolDec: Syntax Error-Free and Generalizable Tool Use for LLMs via Finite-State Decoding »
Hongqiao Chen · Kexun Zhang · Lei Li · William Yang Wang -
2023 : Improving Few-Shot Generalization by Exploring and Exploiting Auxiliary Data »
Alon Albalak · Colin Raffel · William Yang Wang -
2023 : Efficient Online Data Mixing For Language Model Pre-Training »
Alon Albalak · Liangming Pan · Colin Raffel · William Yang Wang -
2023 : Efficient Online Data Mixing For Language Model Pre-Training »
Alon Albalak · Liangming Pan · Colin Raffel · William Yang Wang -
2023 : Efficient Online Data Mixing For Language Model Pre-Training »
Alon Albalak · Liang-Ming Pan · Colin Raffel · William Yang Wang -
2023 : VELMA: Verbalization Embodiment of LLM Agents for Vision and Language Navigation in Street View »
Raphael Schumann · Wanrong Zhu · Weixi Feng · Tsu-Jui Fu · Stefan Riezler · William Yang Wang -
2023 : VELMA: Verbalization Embodiment of LLM Agents for Vision and Language Navigation in Street View »
Raphael Schumann · Wanrong Zhu · Weixi Feng · Tsu-Jui Fu · Stefan Riezler · William Yang Wang -
2023 Poster: Flexible Attention-Based Multi-Policy Fusion for Efficient Deep Reinforcement Learning »
Zih-Yun Chiu · Yi-Lin Tuan · William Yang Wang · Michael Yip -
2023 Poster: LayoutGPT: Compositional Visual Planning and Generation with Large Language Models »
Weixi Feng · Wanrong Zhu · Tsu-Jui Fu · Varun Jampani · Arjun Akula · Xuehai He · S Basu · Xin Eric Wang · William Yang Wang -
2023 Poster: LLMScore: Unveiling the Power of Large Language Models in Text-to-Image Synthesis Evaluation »
Yujie Lu · Xianjun Yang · Xiujun Li · Xin Eric Wang · William Yang Wang -
2023 Poster: Improving Few-Shot Generalization by Exploring and Exploiting Auxiliary Data »
Alon Albalak · Colin Raffel · William Yang Wang -
2023 Poster: ALGO: Synthesizing Algorithmic Programs with Generated Oracle Verifiers »
Kexun Zhang · Danqing Wang · Jingtao Xia · William Yang Wang · Lei Li -
2023 Poster: Multimodal C4: An Open, Billion-scale Corpus of Images Interleaved with Text »
Wanrong Zhu · Jack Hessel · Anas Awadalla · Samir Yitzhak Gadre · Jesse Dodge · Alex Fang · Youngjae Yu · Ludwig Schmidt · William Yang Wang · Yejin Choi -
2023 Poster: VisIT-Bench: A Dynamic Benchmark for Evaluating Instruction-Following Vision-and-Language Models »
Yonatan Bitton · Hritik Bansal · Jack Hessel · Rulin Shao · Wanrong Zhu · Anas Awadalla · Josh Gardner · Rohan Taori · Ludwig Schmidt -
2021 Poster: Local Explanation of Dialogue Response Generation »
Yi-Lin Tuan · Connor Pryor · Wenhu Chen · Lise Getoor · William Yang Wang -
2021 Poster: Combining Human Predictions with Model Probabilities via Confusion Matrices and Calibration »
Gavin Kerrigan · Padhraic Smyth · Mark Steyvers -
2021 Poster: Counterfactual Maximum Likelihood Estimation for Training Deep Networks »
Xinyi Wang · Wenhu Chen · Michael Saxon · William Yang Wang -
2020 Poster: Can I Trust My Fairness Metric? Assessing Fairness with Unlabeled Data and Bayesian Inference »
Disi Ji · Padhraic Smyth · Mark Steyvers -
2013 Poster: Scoring Workers in Crowdsourcing: How Many Control Questions are Enough? »
Qiang Liu · Alexander Ihler · Mark Steyvers -
2013 Spotlight: Scoring Workers in Crowdsourcing: How Many Control Questions are Enough? »
Qiang Liu · Alexander Ihler · Mark Steyvers -
2010 Spotlight: Learning concept graphs from text with stick-breaking priors »
America Chambers · Padhraic Smyth · Mark Steyvers -
2010 Poster: Learning concept graphs from text with stick-breaking priors »
America Chambers · Padhraic Smyth · Mark Steyvers -
2009 Poster: The Wisdom of Crowds in the Recollection of Order Information »
Mark Steyvers · Michael D Lee · Brent Miller · Pernille Hemmer