Siddarth Venkatraman @ NeurIPS 2024
@hyperpotatoneo.bsky.social
330 followers 450 following 31 posts
PhD student at Mila | Diffusion models and reinforcement learning 🧐 | hyperpotatoneo.github.io
Posts Media Videos Starter Packs
hyperpotatoneo.bsky.social
Honestly, it feels like as an AI researcher it might actually be worth it to throw your dignity aside and pay Elon for Twitter blue to advertise your papers. Getting papers famous is literally just a social media clout game now.
hyperpotatoneo.bsky.social
See the second part of my post - yes, they are likely using explicit search to improve performance at test time. But the focus should be on the search through reasoning chains itself, which the model has been trained to do with RL. Even for the explicit search, you require the RL value functions.
Reposted by Siddarth Venkatraman @ NeurIPS 2024
bennokrojer.bsky.social
Few fields reward quick pivoting as much as AI, or vice versa punish the very thing a phd is usually meant to be: stick with one research direction for 5 years no matter what, go really deep, becoming a niche expert

for your research to be relevant in AI, you might wanna pivot every 1-2 years
hyperpotatoneo.bsky.social
I think the intersection of builders and researchers is higher in machine learning, compared to other disciplines.
hyperpotatoneo.bsky.social
You could still wrap this with explicit search techniques like MCTS if you have value functions for partial sequences (which would also be a product of the RL training). This could further improve performance, similar to fast vs slow policy in AlphaZero.
hyperpotatoneo.bsky.social
Saying o3 is just a “more principled search technique” is quite reductive. The o series of models don’t require “explicit search” strategies in the form of tree search, wrapped in loops etc. Instead, RL is used to train the model to “learn to search” using long CoT chains.
markriedl.bsky.social
Six months ago someone put a for-loop around GPT-4o and got 50% on the ARC-AGI test set and 72% on a held-out training set redwoodresearch.substack.com/p/getting-50... Just sample 8000 times with beam search.

o3 is probably a more principled search technique...
Getting 50% (SoTA) on ARC-AGI with GPT-4o
You can just draw more samples
redwoodresearch.substack.com
hyperpotatoneo.bsky.social
You’re correct, there’s plenty of simulated environments we can’t solve yet. But do you consider having 1 million parallel instances of an environment sped up 100x solving it with PPO with low wall clock time a desirable solution?
hyperpotatoneo.bsky.social
This isn’t a general solution to RL. The point is to make learning algorithms sample efficient. If the environment you are doing RL on is the real world, you can’t make the “environment go fast”.

With “infinite samples”, you can random sample policies till you stumble on one with high reward.
Reposted by Siddarth Venkatraman @ NeurIPS 2024
Reposted by Siddarth Venkatraman @ NeurIPS 2024
rl-conference.bsky.social
If you're at NeurIPS, RLC is hosting an RL event from 8 till late at The Pearl on Dec. 11th. Join us, meet all the RL researchers, and spread the word!
hyperpotatoneo.bsky.social
Even his current claim that o1 is “better than most humans in most tasks” is pretty wild imo. What are “most tasks” here even? Obviously not any physical tasks because there is no embodiment. Can o1 actually completely replace a human in any job? Can it manage a project from start to finish?
hyperpotatoneo.bsky.social
It also doesn’t help when OpenAI staff post about how o1 is already AGI (yes this happened today).

Unfortunately the dialogue is directed by those on either end of the spectrum (AI is useless vs AGI is already here) without much room for nuance.
hyperpotatoneo.bsky.social
www.newsweek.com/united-healt...

I have anecdotal evidence from a friend who works at a client company for a popular insurance firm. They are using shitty “AI models” which are basically just CatBoost to mass process claims. They know the models are shit, but that’s also the point. Truly sickening.
A year before CEO shooting, lawsuit alleged UHC used AI to deny coverage
The lawsuit accuses UnitedHealthcare of using artificial intelligence to deny coverage to elderly patients.
www.newsweek.com
hyperpotatoneo.bsky.social
It is reductive to blame it all on a single CEO, but I find it hard to believe how you are “shocked” by this public reaction. UHC has the highest claim denial rate among insurance providers, resulting in untold medical bankruptcies and preventable deaths. I’m shocked this doesn’t happen more often.
hyperpotatoneo.bsky.social
Subtlety and nuance go out the window when strong political feelings are thrown in the mix. I understand why AI researchers can get defensive/angry due to toxic comments, but we should still try to understand the origin of people’s anger. Imo, right wing AI silicon valley billionaires are the root.
hyperpotatoneo.bsky.social
I think the recent conflict between AI researchers and the anti-AI clique hints at the latter. This broad left leaning user base could fracture again as differences in opinions between the farther left and moderate factions get amplified.
hyperpotatoneo.bsky.social
This app is an interesting social experiment. Assuming Bluesky doesn’t just fizzle out, will hostile social relations as in Twitter resurface here too? If hostilities do return, will it be because conservatives come to this app, or will it be new political tensions within left leaning communities?
hyperpotatoneo.bsky.social
Another thing; let’s reflect if they actually have a point. When I deeply reflect upon it, I am not even personally convinced that in the grand scheme of things AI is going to be a net good for humanity. So, maybe the distaste is warranted and we’re the ones in the bubble?
hyperpotatoneo.bsky.social
As AI researchers, we shouldn’t demonize people outside our space who have a passionate distaste for AI. You have to understand that most of the pro-AI sentiment people see online comes from absolutely vile “AI-bros”, especially on twitter. We just need to distinguish ourselves as academics.
hyperpotatoneo.bsky.social
Yeah, it will definitely not be “true OT” at end, but it works to get surprisingly smooth ODE paths which can be easily numerically integrated. You can train a CIFAR 10 flow model which can generate high quality images with 5-10 Euler steps.
hyperpotatoneo.bsky.social
Sure, that argument works from a utilitarian perspective.

From monkey brain casual user point of view, it looks ugly and outdated. And I think this is what should be focused on.
hyperpotatoneo.bsky.social
Anyone has thoughts about which generative models are also the best for representation learning features for downstream tasks?

My guess is GANs are a dark horse and the latents carry important abstract features. But we haven’t explored this much since they are hard to train.
hyperpotatoneo.bsky.social
You can just have a verification system like the system in pre-Elon twitter, where blue check marks are verified accounts.