Author: Andrew Fairless

What I Read: Ensemble, knowledge distillation, and self-distillation

https://www.microsoft.com/en-us/research/blog/three-mysteries-in-deep-learning-ensemble-knowledge-distillation-and-self-distillation/ Three mysteries in deep learning: Ensemble, knowledge distillation, and self-distillationPublished January 19, 2021By Zeyuan Allen-Zhu , Senior Researcher Yuanzhi Li , Assistant Professor, Carnegie Mellon University “…besides this smallContinue readingWhat I Read: Ensemble, knowledge distillation, and self-distillation

What I Read: Transformer Networks to Answer Questions About Images

https://medium.com/dataseries/microsoft-uses-transformer-networks-to-answer-questions-about-images-with-minimum-training-f978c018bb72 Microsoft Uses Transformer Networks to Answer Questions About Images With Minimum TrainingUnified VLP can understand concepts about scenic images by using pretrained models.Jesus RodriguezJan 12 “Can we build deepContinue readingWhat I Read: Transformer Networks to Answer Questions About Images