Poster
|
Tue 8:45
|
Towards Efficient Pre-Trained Language Model via Feature Correlation Distillation
Kun Huang · Xin Guo · Meng Wang
|
|
Workshop
|
|
Pre-trained Language Models Do Not Help Auto-regressive Text-to-Image Generation
Yuhui Zhang · Brandon McKinzie · Zhe Gan · Vaishaal Shankar · Alexander Toshev
|
|
Workshop
|
|
Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning
Mengzhou Xia · Tianyu Gao · Zhiyuan Zeng · Danqi Chen
|
|
Workshop
|
|
Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning
Mengzhou Xia · Tianyu Gao · Zhiyuan Zeng · Danqi Chen
|
|
Poster
|
Wed 15:00
|
Connecting Pre-trained Language Model and Downstream Task via Properties of Representation
Chenwei Wu · Holden Lee · Rong Ge
|
|
Workshop
|
|
FDAPT: Federated Domain-adaptive Pre-training for Language Models
Lekang Jiang · Filip Svoboda · Nicholas Lane
|
|
Workshop
|
|
Efficient Online Data Mixing For Language Model Pre-Training
Alon Albalak · Liangming Pan · Colin Raffel · William Yang Wang
|
|
Workshop
|
|
Efficient Online Data Mixing For Language Model Pre-Training
Alon Albalak · Liangming Pan · Colin Raffel · William Yang Wang
|
|
Workshop
|
|
Learning to Act without Actions
Dominik Schmidt · Minqi Jiang
|
|
Poster
|
Tue 15:15
|
Bootstrapping Vision-Language Learning with Decoupled Language Pre-training
Yiren Jian · Chongyang Gao · Soroush Vosoughi
|
|
Workshop
|
|
Exploring Data Augmentations on Self-/Semi-/Fully- Supervised Pre-trained Models
Shentong Mo · Zhun Sun · Chao Li
|
|
Poster
|
Thu 8:45
|
How does GPT-2 compute greater-than?: Interpreting mathematical abilities in a pre-trained language model
Michael Hanna · Ollie Liu · Alexandre Variengien
|
|