Do we need Attention? A Mamba Primer
Youtube : Download Convert to MP3
Description :
A Technical Primer on Mamba and Friends. With Yair Schiff (yair-schiff.github.io/) Slides: github.com/srush/mamba-primer/blob/main/mamba.pdf Main focus: Mamba: Linear-Time Sequence Modeling with Selective State Spaces arxiv.org/abs/2312.00752 from Albert Gu and Tri Dao. Simplified State Space Layers for Sequence Modeling arxiv.org/abs/2208.04933 ...
Related Videos :
Mamba: Linear-Time Sequence Modeling with Selective State Spaces (Paper Explained) By: Yannic Kilcher |
Do we need Attention? - Linear RNNs and State Space Models (SSMs) for NLP By: Sasha Rush 🤗 |
Visualizing transformers and attention | Talk for TNG Big Tech Day '24 By: Grant Sanderson |
Andrew Ng Explores The Rise Of AI Agents And Agentic Reasoning | BUILD 2024 Keynote By: Snowflake Inc. |
12,419 Days Of Strandbeest Evolution By: Veritasium |
Long-Context LLM Extension By: Sasha Rush 🤗 |