Attention free transformer
Abstract:
Attention-free transformers are disclosed. Various implementations of attention-free transformers include a gating and pooling operation that allows the attention-free transformers to provide comparable or better results to those of a standard attention-based transformer, with improved efficiency and reduced computational complexity with respect to space and time.
Public/Granted literature
Information query
Patent Agency Ranking
0/0