Kashyap Chitta
@kashyap7x.bsky.social
2.2K followers 600 following 61 posts
kashyap7x.github.io Postdoc at NVIDIA. Previously at the University of Tübingen and CMU. Robot Learning, Autonomous Driving.
Posts Media Videos Starter Packs
Pinned
kashyap7x.bsky.social
🚗 Pseudo-simulation combines the efficiency of open-loop and robustness of closed-loop evaluation. It uses real data + 3D Gaussian Splatting synthetic views to assess error recovery, achieving strong correlation with closed-loop simulations while requiring 6x less compute. arxiv.org/abs/2506.04218
Reposted by Kashyap Chitta
eugenevinitsky.bsky.social
We're finally out of stealth: percepta.ai
We're a research / engineering team working together in industries like health and logistics to ship ML tools that drastically improve productivity. If you're interested in ML and RL work that matters, come join us 😀
Percepta | A General Catalyst Transformation Company
Transforming critical institutions using applied AI. Let's harness the frontier.
percepta.ai
Reposted by Kashyap Chitta
eugenevinitsky.bsky.social
We've hired some *fantastic* researchers but our startup is still looking for 2-3 more people with skills in ML/RL/LLMs. If you'd like to work on some transformative applied problems, hit me up. We'll be launching publicly soon too...
Reposted by Kashyap Chitta
jbhuang0604.bsky.social
How AI Taught Itself to See

Self-supervised learning is fascinating! How can AI learn from images only without labels?

In this video, we’ll build the method from first principles and uncover the key ideas behind CLIP, MAE, SimCLR, and DINO (v1–v3).

Video link: youtu.be/oGTasd3cliM
How AI Taught Itself to See [DINOv3]
YouTube video by Jia-Bin Huang
youtu.be
kashyap7x.bsky.social
Putting the final touches on your submission? Just a few days left to enter the @iccv.bsky.social NAVSIM Challenge! Deadline: September 20. $8K in prizes and several travel grants are on the line!
kashyap7x.bsky.social
Announcing the @iccv.bsky.social NAVSIM Challenge! What's new? We're testing not only on real recordings, but also perturbed futures generated from the real ones via pseudo-simulation! $8K in prizes + several $1.5k travel grants. Submit by September 20! opendrivelab.com/challenge2025/ 🧵👇
Reposted by Kashyap Chitta
eugenevinitsky.bsky.social
Found this marvelous little course full of readings to trace the evolution of computer science and its canonical ideas
graphics.stanford.edu/courses/cs20...
CS208: Canon of Computer Science, Spring 2010
graphics.stanford.edu
Reposted by Kashyap Chitta
ellisinsttue.bsky.social
Our Principal Investigators Antonio Orvieto, Celestine Mendler-Dünner, Maximilian Dax, Rediet Abebe, @shiweiliu.bsky.social, T. Konstantin Rusch, and @wielandbrendel.bsky.social are looking for motivated students interested in internships.
Apply here: docs.google.com/forms/d/e/1F...
Reposted by Kashyap Chitta
kashyap7x.bsky.social
Announcing the @iccv.bsky.social NAVSIM Challenge! What's new? We're testing not only on real recordings, but also perturbed futures generated from the real ones via pseudo-simulation! $8K in prizes + several $1.5k travel grants. Submit by September 20! opendrivelab.com/challenge2025/ 🧵👇
kashyap7x.bsky.social
Announcing the @iccv.bsky.social NAVSIM Challenge! What's new? We're testing not only on real recordings, but also perturbed futures generated from the real ones via pseudo-simulation! $8K in prizes + several $1.5k travel grants. Submit by September 20! opendrivelab.com/challenge2025/ 🧵👇
Reposted by Kashyap Chitta
ellis.eu
ELLIS @ellis.eu · Aug 29
🎓 Interested in a #PhD in machine learning or #AI? The ELLIS PhD Program connects top students with leading researchers across Europe. The application portal opens on Oct 1st. Curious? Join our info session on the same day. Get all the info 👉
ELLIS PhD Program: Call for Applications 2025
The ELLIS mission is to create a diverse European network that promotes research excellence and advances breakthroughs in AI, as well as a pan-European PhD program to educate the next generation of AI researchers. ELLIS also aims to boost economic growt...
ellis.eu
Reposted by Kashyap Chitta
mlcv-at-ista.bsky.social
Let's push for the obvious solution: Dear @neuripsconf.bsky.social ! Allow authors to present accepted papers at EurIPS instead of NeurIPS rather than just additionally. Likely, at least 500 papers would move to Copenhagen, problem solved.
Reposted by Kashyap Chitta
haoyuhe.bsky.social
🚀 Introducing our new paper, MDPO: Overcoming the Training-Inference Divide of Masked Diffusion Language Models.

📄 Paper: www.scholar-inbox.com/papers/He202...
arxiv.org/pdf/2508.13148
💻 Code: github.com/autonomousvi...
🌐 Project Page: cli212.github.io/MDPO/
Reposted by Kashyap Chitta
maxseitzer.bsky.social
Introducing DINOv3 🦕🦕🦕

A SotA-enabling vision foundation model, trained with pure self-supervised learning (SSL) at scale.
High quality dense features, combining unprecedented semantic and geometric scene understanding.

Three reasons why this matters👇
Reposted by Kashyap Chitta
bernhard-jaeger.bsky.social
Our paper "CaRL: Learning Scalable Planning Policies with Simple Rewards" has been accepted to the Conference on Robot Learning (CoRL 2025).
See you in Seoul at the end of September.

Code & Paper:
github.com/autonomousvi...
GitHub - autonomousvision/CaRL: [CoRL 2025] CaRL: Learning Scalable Planning Policies with Simple Rewards
[CoRL 2025] CaRL: Learning Scalable Planning Policies with Simple Rewards - autonomousvision/CaRL
github.com
Reposted by Kashyap Chitta
rowantmc.bsky.social
At TRI’s newest division: Automated Driving Advanced Development, we’re building a clean-slate, end-to-end autonomy stack. We're hiring, with open roles in learning, infra, and validation: www.tri.global/careers#open...
Reposted by Kashyap Chitta
abursuc.bsky.social
1/ Can open-data models beat DINOv2? Today we release Franca, a fully open-sourced vision foundation model. Franca with ViT-G backbone matches (and often beats) proprietary models like SigLIPv2, CLIP, DINOv2 on various benchmarks setting a new standard for open-source research.
Reposted by Kashyap Chitta
natolambert.bsky.social
Adding a nice way to visualize the PPO objective to the rlhf book.
Core for policy-gradient is L is proportional to R*A (R=policy ratio, A = advantage).
PPO makes good actions more likely, up to a point.
PPO makes bad actions less likely, up to a point.
kashyap7x.bsky.social
What if the cherry was actually cake? (source: arxiv.org/abs/2506.08007)
Reposted by Kashyap Chitta
ingmarweber.de
🚨Job Alert
W2 (TT W3) Professorship in Computer Science "AI for People & Society"
@saarland-informatics-campus.de/@uni-saarland.de is looking to appoint an outstanding individual in the field of AI for people and society who has made significant contributions in one or more of the following areas:
Reposted by Kashyap Chitta
euripsconf.bsky.social
EurIPS is coming! 📣 Mark your calendar for Dec. 2-7, 2025 in Copenhagen 📅

EurIPS is a community-organized conference where you can present accepted NeurIPS 2025 papers, endorsed by @neuripsconf.bsky.social and @nordicair.bsky.social and is co-developed by @ellis.eu

eurips.cc
Reposted by Kashyap Chitta
jbhuang0604.bsky.social
In an era of billion-parameter models everywhere, it's incredibly refreshing to see how a fundamental question can be formulated and solved with simple, beautiful math.

- How should we orient a solar panel ☀️🔋? -

Zero AI! If you enjoy math, you'll love this!

Video: www.youtube.com/watch?v=ZKzL...
Reposted by Kashyap Chitta
ellis.eu
ELLIS @ellis.eu · Jul 16
📢 Present your NeurIPS paper in Europe!

Join EurIPS 2025 + ELLIS UnConference in Copenhagen for in-person talks, posters, workshops and more. Registration opens soon; save the date:

📅 Dec 2–7, 2025
📍 Copenhagen 🇩🇰
🔗eurips.cc

#EurIPS
@euripsconf.bsky.social
Reposted by Kashyap Chitta
chriswolfvision.bsky.social
I really like this paper on relative positional encodings using projective geometry for multi-view transformers, by Li et al. (Berkeley/Nvidia/HKU).

It is elegant: in special situations, it defaults to known baselines like GTA (if identity intrinsics) and RoPE (same cam).

arxiv.org/abs/2507.10496
Reposted by Kashyap Chitta
bernhard-jaeger.bsky.social
We have released the code for our work, CaRL: Learning Scalable Planning Policies with Simple Rewards.

The repository contains the first public code base for training RL agents with the CARLA leaderboard 2.0 and nuPlan.

github.com/autonomousvi...
GitHub - autonomousvision/CaRL: [ArXiv 2025] CaRL: Learning Scalable Planning Policies with Simple Rewards
[ArXiv 2025] CaRL: Learning Scalable Planning Policies with Simple Rewards - autonomousvision/CaRL
github.com