Showing 1–20 of 23 results
/ Date/ Name
Jun 26, 2020Object-Centric Learning with Slot AttentionJun 18, 2021How to train your ViT? Data, Augmentation, and Regularization in Vision TransformersJun 12, 2017Attention Is All You NeedOct 20, 2020Towards End-to-End In-Image Neural Machine TranslationSep 12, 2018Music TransformerJun 4, 2019KERMIT: Generative Insertion-Based Modeling for SequencesJun 6, 2019Scaling Autoregressive Video ModelsJun 16, 2017One Model To Learn Them AllMar 6, 2018Self-Attention with Relative Position RepresentationsNov 7, 2018Blockwise Parallel Decoding for Deep Autoregressive ModelsOct 29, 2019An Empirical Study of Generation Order for Machine TranslationOct 22, 2020An Image is Worth 16x16 Words: Transformers for Image Recognition at ScaleFeb 15, 2018Image TransformerApr 15, 2017Neural Paraphrase Identification of Questions with Noisy PretrainingMay 4, 2021MLP-Mixer: An all-MLP Architecture for VisionNov 6, 2016Hierarchical Question Answering for Long DocumentsMar 9, 2018Fast Decoding in Sequence Models using Discrete Latent VariablesJul 10, 2018Universal TransformersFeb 8, 2019Insertion Transformer: Flexible Sequence Generation via Insertion OperationsNov 25, 2021Scene Representation Transformer: Geometry-Free Novel View Synthesis Through Set-Latent Scene Representations