Skip to yearly menu bar Skip to main content


Poster

Understanding the Expressive Power and Mechanisms of Transformer for Sequence Modeling

Mingze Wang · Weinan E


Abstract:

We conduct a systematic study of the approximation properties of Transformer for sequence modeling with long, sparse and complicated memory. We investigate the mechanisms through which different components of Transformer, such as the dot-product self-attention, positional encoding and feed-forward layer, affect its expressive power, and we study their combined effects through establishing explicit approximation rates.Our study reveals the roles of critical parameters in the Transformer, such as the number of layers and the number of attention heads.These theoretical insights are validated experimentally and offer natural suggestions for alternative architectures.

Live content is unavailable. Log in and register to view live content