Skip to yearly menu bar Skip to main content


Search All 2023 Events
 

60 Results

<<   <   Page 1 of 5   >   >>
Workshop
Transformer-based Causal Language Models from a Meta-Learning Perspective
Xinbo Wu · Lav Varshney
Workshop
Transformer-Based Large Language Models Are Not General Learners: A Universal Circuit Perspective
Yang Chen · Yitao Liang · Zhouchen Lin
Poster
Wed 8:45 Blockwise Parallel Transformers for Large Context Models
Hao Liu · Pieter Abbeel
Workshop
Trained Transformers Learn Linear Models In-Context
Ruiqi Zhang · Spencer Frei · Peter Bartlett
Workshop
Trained Transformers Learn Linear Models In-Context
Ruiqi Zhang · Spencer Frei · Peter Bartlett
Workshop
Can Transformers In-Context Learn Task Mixtures?
Nilesh Tripuraneni · Lyric Doshi · Steve Yadlowsky
Workshop
Can Transformer Models Generalize Via In-Context Learning Beyond Pretraining Data?
Steve Yadlowsky · Lyric Doshi · Nilesh Tripuraneni
Workshop
DDxT: Deep Generative Transformer Models for Differential Diagnosis
Mohammad Mahmudul Alam · Edward Raff · Tim Oates · Cynthia Matuszek
Poster
Wed 8:45 MEGABYTE: Predicting Million-byte Sequences with Multiscale Transformers
LILI YU · Daniel Simig · Colin Flaherty · Armen Aghajanyan · Luke Zettlemoyer · Mike Lewis
Workshop
Optimizing protein language models with Sentence Transformers
Istvan Redl
Poster
Thu 8:45 TabMT: Generating tabular data with masked transformers
Manbir Gulati · Paul Roysdon
Workshop
Ring Attention with Blockwise Transformers for Near-Infinite Context
Hao Liu · Matei A Zaharia · Pieter Abbeel