Problem
Research questions and friction points this paper is trying to address.
In-context universality of transformers
Comparison with MLPs universality
Exploring transformer's success factors
Innovation
Methods, ideas, or system contributions that make the work stand out.
MLPs with trainable activation functions
universal in-context learning
challenges transformer superiority