Guillaume Astruc
@gastruc.bsky.social
350 followers 79 following 10 posts
2nd Year PhD Student from Imagine-ENPC/IGN/CNES Working on Self-supervised Cross-modal Geospatial Learning. Personal WebPage: https://gastruc.github.io/
Posts Media Videos Starter Packs
Pinned
gastruc.bsky.social
🤔 What if embedding multimodal EO data was as easy as using a ResNet on images?
Introducing AnySat: one model for any resolution (0.2m–250m), scale (0.3–2600 hectares), and modalities (choose from 11 sensors & time series)!
Try it with just a few lines of code:
gastruc.bsky.social
Super interesting to see pure SSL outperforms text alignement on a super competitive but text-aligned suited task 🤯
nicolasdufour.bsky.social
🚀 DinoV3 just became the new go-to backbone for geoloc!
It outperforms CLIP-like models (SigLip2, finetuned StreetCLIP)… and that’s shocking 🤯
Why? CLIP models have an innate advantage — they literally learn place names + images. DinoV3 doesn’t.
gastruc.bsky.social
🛰️ At #CVPR2025 presenting "AnySat: An Earth Observation Model for Any Resolutions, Scales, and Modalities" - Saturday afternoon, Poster 355!
If you're here and want to discuss geolocation or geospatial foundation models, let's connect!
Reposted by Guillaume Astruc
elliotvincent.bsky.social
I will be presenting our work on the detection of archaeological looting with satellite image time series at CVPR 2025 EarthVision workshop tomorrow!

Honored and grateful that this paper received the best student paper award!
Reposted by Guillaume Astruc
ryanboustany.bsky.social
📢 New preprint!
“When majority rules, minority loses: bias amplification of gradient descent”

We often blame biased data but training also amplifies biases. Our paper explores how ML algorithms favor stereotypes at the expense of minority groups.

➡️ arxiv.org/abs/2505.13122

(1/3)
When majority rules, minority loses: bias amplification of gradient descent
Despite growing empirical evidence of bias amplification in machine learning, its theoretical foundations remain poorly understood. We develop a formal framework for majority-minority learning tasks, ...
arxiv.org
gastruc.bsky.social
We've added new experiments demonstrating robust generalization capabilities! Notably, AnySat shows strong performance on HLS Burn Scars - a sensor never seen during pretraining! 🔥🛰️
Check it out:
📄 Paper: arxiv.org/abs/2412.14123
🌐 Project: gastruc.github.io/anysat
Reposted by Guillaume Astruc
imagineenpc.bsky.social
Looking forward to #CVPR2025! We will present the following papers:
Reposted by Guillaume Astruc
antoine-guedon.bsky.social
💻We've released the code for our #CVPR2025 paper MAtCha!

🍵MAtCha reconstructs sharp, accurate and scalable meshes of both foreground AND background from just a few unposed images (eg 3 to 10 images)...

...While also working with dense-view datasets (hundreds of images)!
Reposted by Guillaume Astruc
davidpicard.bsky.social
🔥🔥🔥 CV Folks, I have some news! We're organizing a 1-day meeting in center Paris on June 6th before CVPR called CVPR@Paris (similar as NeurIPS@Paris) 🥐🍾🥖🍷

Registration is open (it's free) with priority given to authors of accepted papers: cvprinparis.github.io/CVPR2025InPa...

Big 🧵👇 with details!
Reposted by Guillaume Astruc
imagineenpc.bsky.social
Starter pack including some of the lab members: go.bsky.app/QK8j87w
Reposted by Guillaume Astruc
thibautloiseau.bsky.social
🧩 Excited to share our paper "RUBIK: A Structured Benchmark for Image Matching across Geometric Challenges" (arxiv.org/abs/2502.19955) accepted to #CVPR2025! We created a benchmark that systematically evaluates image matching methods across well-defined geometric difficulty levels. 🔍
Reposted by Guillaume Astruc
nicolasdufour.bsky.social
Weights for CAD are finally available. It's one of the smallest diffusion models on the market, achieving performance close to SD and Pixart, featuring a Perceiver-like architecture.
We leverage our coherence aware training to improve the textual understanding
davidpicard.bsky.social
🚨 Just a quick note that following requests, we trained a 512px version of our Coherence-Aware Diffusion model (CVPR'24) and updated the paper on arxiv: arxiv.org/abs/2405.20324

It has a package and pretrained models!

🖥️ nicolas-dufour.github.io/cad.html
🤖 github.com/nicolas-dufo...
gastruc.bsky.social
🚀 Even better: AnySat supports linear probing for semantic segmentation!
That means you can fine-tune just a few thousand parameters and achieve SOTA results on challenging tasks—all with minimal effort.
gastruc.bsky.social
AnySat achieves SOTA performance on 6 tasks across 10 datasets:
🌱 Land cover mapping
🌾 Crop type segmentation
🌳 Tree species classification
🌊 Flood detection
🌍 Change detection
gastruc.bsky.social
We trained AnySat on 5 multimodal datasets simultaneously:
📡 11 distinct sensors
📏 Resolutions: 0.2m–500m
🔁 Revisit: single date to weekly
🏞️ Scales: 0.3–150 hectares

The pretrained model can adapt to truly diverse data, and probably yours too!
gastruc.bsky.social
🔍Thanks to our modified JEPA training scheme and scale-adaptive spatial encoders, AnySat trains on datasets with diverse scales, resolutions, and modalities!
🧠 75% of its parameters are shared across all inputs, enabling unmatched flexibility.
gastruc.bsky.social
🤔 What if embedding multimodal EO data was as easy as using a ResNet on images?
Introducing AnySat: one model for any resolution (0.2m–250m), scale (0.3–2600 hectares), and modalities (choose from 11 sensors & time series)!
Try it with just a few lines of code:
Reposted by Guillaume Astruc
arxiv-cs-cv.bsky.social
Guillaume Astruc, Nicolas Gonthier, Clement Mallet, Loic Landrieu
AnySat: An Earth Observation Model for Any Resolutions, Scales, and Modalities
https://arxiv.org/abs/2412.14123
Reposted by Guillaume Astruc
antoine-guedon.bsky.social
⚠️Reconstructing sharp 3D meshes from a few unposed images is a hard and ambiguous problem.

☑️With MAtCha, we leverage a pretrained depth model to recover sharp meshes from sparse views including both foreground and background, within mins!🧵

🌐Webpage: anttwo.github.io/matcha/
Reposted by Guillaume Astruc
nicolasdufour.bsky.social
🌍 Guessing where an image was taken is a hard, and often ambiguous problem. Introducing diffusion-based geolocation—we predict global locations by refining random guesses into trajectories across the Earth's surface!

🗺️ Paper, code, and demo: nicolas-dufour.github.io/plonk
gastruc.bsky.social
Hi, I am a PhD student from @imagineenpc.bsky.social. Could you also add us both please?