Workshop
|
|
Transformer-based Causal Language Models from a Meta-Learning Perspective
Xinbo Wu · Lav Varshney
|
|
Workshop
|
|
Transformer-Based Large Language Models Are Not General Learners: A Universal Circuit Perspective
Yang Chen · Yitao Liang · Zhouchen Lin
|
|
Poster
|
Wed 8:45
|
Blockwise Parallel Transformers for Large Context Models
Hao Liu · Pieter Abbeel
|
|
Workshop
|
|
Trained Transformers Learn Linear Models In-Context
Ruiqi Zhang · Spencer Frei · Peter Bartlett
|
|
Workshop
|
|
Trained Transformers Learn Linear Models In-Context
Ruiqi Zhang · Spencer Frei · Peter Bartlett
|
|
Workshop
|
|
Can Transformers In-Context Learn Task Mixtures?
Nilesh Tripuraneni · Lyric Doshi · Steve Yadlowsky
|
|
Workshop
|
|
Can Transformer Models Generalize Via In-Context Learning Beyond Pretraining Data?
Steve Yadlowsky · Lyric Doshi · Nilesh Tripuraneni
|
|
Workshop
|
|
DDxT: Deep Generative Transformer Models for Differential Diagnosis
Mohammad Mahmudul Alam · Edward Raff · Tim Oates · Cynthia Matuszek
|
|
Poster
|
Wed 8:45
|
MEGABYTE: Predicting Million-byte Sequences with Multiscale Transformers
LILI YU · Daniel Simig · Colin Flaherty · Armen Aghajanyan · Luke Zettlemoyer · Mike Lewis
|
|
Workshop
|
|
Optimizing protein language models with Sentence Transformers
Istvan Redl
|
|
Poster
|
Thu 8:45
|
TabMT: Generating tabular data with masked transformers
Manbir Gulati · Paul Roysdon
|
|
Workshop
|
|
Ring Attention with Blockwise Transformers for Near-Infinite Context
Hao Liu · Matei A Zaharia · Pieter Abbeel
|
|