The Universal Transformer is a generalization of the Transformer architecture. Universal Transformers combine the parallelizability and global receptive field of feed-forward sequence models like the Transformer with the recurrent inductive bias of RNNs. They also utilise a dynamic per-position halting mechanism.
Source: Universal TransformersPaper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
Sentence | 4 | 11.76% |
Language Modelling | 3 | 8.82% |
Text Generation | 2 | 5.88% |
Reinforcement Learning (RL) | 1 | 2.94% |
Semantic Similarity | 1 | 2.94% |
Semantic Textual Similarity | 1 | 2.94% |
Video Inpainting | 1 | 2.94% |
Automatic Speech Recognition (ASR) | 1 | 2.94% |
Speech Recognition | 1 | 2.94% |