Workshop
|
|
Multi-Layer Transformers Gradient Can be Approximated in Almost Linear Time
Yingyu Liang · Zhizhou Sha · Zhenmei Shi · Zhao Song · Yufa Zhou
|
|
Poster
|
Thu 16:30
|
Transforming Vision Transformer: Towards Efficient Multi-Task Asynchronous Learner
Hanwen Zhong · Jiaxin Chen · Yutong Zhang · Di Huang · Yunhong Wang
|
|
Poster
|
Thu 11:00
|
Ada-MSHyper: Adaptive Multi-Scale Hypergraph Transformer for Time Series Forecasting
Zongjiang Shang · Ling Chen · Binqing Wu · Dongliang Cui
|
|
Poster
|
Fri 16:30
|
A Global Depth-Range-Free Multi-View Stereo Transformer Network with Pose Embedding
Yitong Dong · Yijin Li · Zhaoyang Huang · Weikang Bian · Jingbo Liu · Hujun Bao · Zhaopeng Cui · Hongsheng Li · Guofeng Zhang
|
|
Workshop
|
|
Shared Recurrent Memory Improves Multi-agent Pathfinding
Alsu Sagirova · Yury Kuratov · Mikhail Burtsev
|
|
Workshop
|
|
BAKU: An Efficient Transformer for Multi-Task Policy Learning
Siddhant Haldar · Zhuoran Peng · Lerrel Pinto
|
|
Poster
|
Wed 11:00
|
Remix-DiT: Mixing Diffusion Transformers for Multi-Expert Denoising
Gongfan Fang · Xinyin Ma · Xinchao Wang
|
|
Poster
|
Fri 16:30
|
How Transformers Utilize Multi-Head Attention in In-Context Learning? A Case Study on Sparse Linear Regression
Xingwu Chen · Lei Zhao · Difan Zou
|
|
Poster
|
Fri 16:30
|
Federated Transformer: Multi-Party Vertical Federated Learning on Practical Fuzzily Linked Data
Zhaomin Wu · Junyi Hou · Yiqun Diao · Bingsheng He
|
|
Poster
|
Fri 11:00
|
Kraken: Inherently Parallel Transformers For Efficient Multi-Device Inference
Rohan Baskar Prabhakar · Hengrui Zhang · David Wentzlaff
|
|
Workshop
|
|
On Layer-wise Representation Similarity: Application for Multi-Exit Models with a Single Classifier
Jiachen Jiang · Jinxin Zhou · Zhihui Zhu
|
|
Poster
|
Thu 11:00
|
BAKU: An Efficient Transformer for Multi-Task Policy Learning
Siddhant Haldar · Zhuoran Peng · Lerrel Pinto
|
|