phd @ mila/udem, prev. @ uwaterloo
averyryoo.github.io 🇨🇦🇰🇷
Link: arxiv.org/abs/2506.05320
A big thank you to my co-authors: @nandahkrishna.bsky.social*, @ximengmao.bsky.social*, @mehdiazabou.bsky.social, Eva Dyer, @mattperich.bsky.social, and @glajoie.bsky.social!
🧵7/7
Link: arxiv.org/abs/2506.05320
A big thank you to my co-authors: @nandahkrishna.bsky.social*, @ximengmao.bsky.social*, @mehdiazabou.bsky.social, Eva Dyer, @mattperich.bsky.social, and @glajoie.bsky.social!
🧵7/7
🧵6/7
🧵6/7
Excitingly, we find that POSSM pretrained solely on monkey reaching data achieves SOTA performance when decoding imagined handwriting in human subjects! This shows the potential of leveraging NHP data to bootstrap human BCI decoding in low-data clinical settings.
🧵5/7
Excitingly, we find that POSSM pretrained solely on monkey reaching data achieves SOTA performance when decoding imagined handwriting in human subjects! This shows the potential of leveraging NHP data to bootstrap human BCI decoding in low-data clinical settings.
🧵5/7
✅ High R² across the board
✅ 9× faster inference than Transformers
✅ <5ms latency per prediction
🧵4/7
✅ High R² across the board
✅ 9× faster inference than Transformers
✅ <5ms latency per prediction
🧵4/7
Using POYO-style tokenization, we encode spikes in 50ms windows and stream them to a recurrent model (e.g., Mamba, GRU) for fast, frequent predictions over time.
🧵3/7
Using POYO-style tokenization, we encode spikes in 50ms windows and stream them to a recurrent model (e.g., Mamba, GRU) for fast, frequent predictions over time.
🧵3/7
😔 RNNs offer efficient, causal inference, but rely on rigid, binned input formats - limiting generalization to new neurons or sessions.
😔 Transformers enable generalization via tokenization, but have high computational costs due to the attention mechanism.
🧵2/7
😔 RNNs offer efficient, causal inference, but rely on rigid, binned input formats - limiting generalization to new neurons or sessions.
😔 Transformers enable generalization via tokenization, but have high computational costs due to the attention mechanism.
🧵2/7