https://ruder.io/recent-advances-lm-fine-tuning/ Recent Advances in Language Model Fine-tuningThis article provides an overview of recent methods to fine-tune large pre-trained language models.Sebastian Ruder24 Feb 2021 “While pre-training is compute-intensive, fine-tuning can be
What I Read: Transformer Networks to Answer Questions About Images
https://medium.com/dataseries/microsoft-uses-transformer-networks-to-answer-questions-about-images-with-minimum-training-f978c018bb72 Microsoft Uses Transformer Networks to Answer Questions About Images With Minimum TrainingUnified VLP can understand concepts about scenic images by using pretrained models.Jesus RodriguezJan 12 “Can we build deep
What I Read: HuggingFace Transformers
https://medium.com/georgian-impact-blog/how-to-incorporate-tabular-data-with-huggingface-transformers-b70ac45fcfb4 How to Incorporate Tabular Data with HuggingFace TransformersGeorgianOct 23 “At Georgian, we find ourselves working with supporting tabular feature information as well as unstructured text data. We found that
What I Read: Revisiting Sutton’s Bitter Lesson for AI
https://blog.exxactcorp.com/compute-goes-brrr-revisiting-suttons-bitter-lesson-artificial-intelligence/ Deep LearningCompute Goes Brrr: Revisiting Sutton’s Bitter Lesson for Artificial IntelligenceMarketing, October 27, 2020 “The main driver of AI progress, according to Sutton, is the increasing availability of compute
What I Read: Attention with Performers
https://ai.googleblog.com/2020/10/rethinking-attention-with-performers.html Rethinking Attention with PerformersFriday, October 23, 2020Posted by Krzysztof Choromanski and Lucy Colwell, Research Scientists, Google Research “To resolve these issues, we introduce the Performer, a Transformer architecture with