https://newsletter.maartengrootendorst.com/p/a-visual-guide-to-mamba-and-state A Visual Guide to Mamba and State Space ModelsMaarten GrootendorstFeb 19, 2024 “To further improve LLMs, new architectures are developed that might even outperform the Transformer architecture. One of
What I Read: Kalman Filter
https://www.youtube.com/watch?v=-DiZGpAh7T4 Kalman Filter – VISUALLY EXPLAINED!Kapil Sachdeva “This tutorial explains the Kalman Filter from Bayesian Probabilistic View and as a special case of Bayesian Filtering.”
What I Read: Mamba Explained
https://thegradient.pub/mamba-explained Mamba ExplainedKola Ayonrinde27.Mar.2024 “Mamba promises similar performance (and crucially similar scaling laws) as the Transformer whilst being feasible at long sequence lengths (say 1 million tokens).”