Ilyass Moummad
@ilyassmoummad.bsky.social
370 followers 330 following 55 posts
Postdoctoral Researcher @ Inria Montpellier (IROKO, Pl@ntNet) SSL for plant images Interested in Computer Vision, Natural Language Processing, Machine Listening, and Biodiversity Monitoring Website: ilyassmoummad.github.io
Posts Media Videos Starter Packs
Pinned
ilyassmoummad.bsky.social
Hi everyone! I'm Ilyass, a PhD student at IMT Atlantique, France, working on representation learning and few-shot learning, currently focusing on species recognition (audio/vision) until the end of 2024. I love sharing cool ideas and papers in Deep Learning—let's learn together!
ilyassmoummad.bsky.social
I heard that the Linux client is buggy, I use it on the browser and it's working ok.
ilyassmoummad.bsky.social
for the curious, the code, slides and the article are on Github: github.com/BastienPasde...
ilyassmoummad.bsky.social
love it haha wish I were there to hear Prostitute Disfigurement in an amphitheater
ilyassmoummad.bsky.social
A website to visually browse and explore the ImageNet-1k dataset (there are other supported datasets: IN-12M, WikiMedia, ETH Images, Pixabay, Fashion) navigu.net#imagenet
(Maybe this is already known, but I was happy to discover it this morning)
NAVIGU: a powerful image collection explorer.
NAVIGU lets you dive into the ocean of images. Drag the image sphere or double-click on an image you like to browse large collections.
navigu.net
ilyassmoummad.bsky.social
Im interested in the quantum and footnotesize, how much params should they have 😂
ilyassmoummad.bsky.social
Learning Deep Representations of Data Distributions
Sam Buchanan · Druv Pai · Peng Wang · Yi Ma

ma-lab-berkeley.github.io/deep-represe...

The best Deep Learning book is out, I've been waiting for its release for more than a year. Let's learn how to build intelligent systems via compression.
Learning Deep Representations of Data Distributions
Landing page for the book Learning Deep Representations of Data Distributions.
ma-lab-berkeley.github.io
ilyassmoummad.bsky.social
It feels like we can now fit more noise with more model capacity 🤔 (Figure 6), maybe we need newer architectures and/or newer training losses.
Reposted by Ilyass Moummad
abursuc.bsky.social
1/ Can open-data models beat DINOv2? Today we release Franca, a fully open-sourced vision foundation model. Franca with ViT-G backbone matches (and often beats) proprietary models like SigLIPv2, CLIP, DINOv2 on various benchmarks setting a new standard for open-source research.
ilyassmoummad.bsky.social
👋 I worked on bioacoustics during my PhD, but I post mostly about AI
ilyassmoummad.bsky.social
the best discovery I've had in recent years, I'm addicted to it now as well 😁
ilyassmoummad.bsky.social
Thank you for making this accessible to everyone! I've read some sections, it is very instructive.
Reposted by Ilyass Moummad
phillipisola.bsky.social
Our computer vision textbook is now available for free online here:
visionbook.mit.edu

We are working on adding some interactive components like search and (beta) integration with LLMs.

Hope this is useful and feel free to submit Github issues to help us improve the text!
Foundations of Computer Vision
The print version was published by
visionbook.mit.edu
Reposted by Ilyass Moummad
klara-cz.bsky.social
⚠️❗Open PhD and Postdoc positions in Prague with Lukas Neumann! ❗⚠️

We rank #5 in computer vision in Europe and Lukas is a great supervisor, so this is a great opportunity!

If you are interested, contact him, he will also be at CVPR with his group :)
Reposted by Ilyass Moummad
klara-cz.bsky.social
We will be presenting the 🍄 FungiTastic 🍄, a multimodal, highly challenging dataset and benchmark covering many ML problems at @fgvcworkshop.bsky.social CVPR-W on Wednesday!

⏱️ 16:15
📍104 E, Level 1
📸 www.kaggle.com/datasets/pic...
📃 arxiv.org/abs/2408.13632

@cvprconference.bsky.social
ilyassmoummad.bsky.social
One of the best conferences that I have been to, happy to have met old friends and having made new ones, hopefully future collaborations as well. Many thanks for organizing this 🙏
Reposted by Ilyass Moummad
davnords.bsky.social
Want stronger Vision Transformers? Use octic-equivariant layers (arxiv.org/abs/2505.15441).

TLDR; We extend @bokmangeorg.bsky.social's reflection-equivariant ViTs to the (octic) group of 90-degree rotations and reflections and... it just works... (DINOv2+DeiT)

Code: github.com/davnords/octic-vits
Reposted by Ilyass Moummad
rbalestr.bsky.social
Want to use SOTA Self Supervised Learning (SSL) methods on noisy data? We provide a novel training curriculum that significantly improves test performance on clean and noisy samples! The approach is fully SSL and works on any method (DINOv2, MoCo, ...)
arxiv.org/abs/2505.12191
Ditch the Denoiser: Emergence of Noise Robustness in Self-Supervised Learning from Data Curriculum
Self-Supervised Learning (SSL) has become a powerful solution to extract rich representations from unlabeled data. Yet, SSL research is mostly focused on clean, curated and high-quality datasets. As a...
arxiv.org
ilyassmoummad.bsky.social
Super présentation, bravo et félicitations ! 👏 👏