Levi Lelis
@programsynthesis.bsky.social
310 followers 480 following 81 posts
Associate Professor - University of Alberta Canada CIFAR AI Chair with Amii Machine Learning and Program Synthesis he/him; ele/dele 🇨🇦 🇧🇷 https://www.cs.ualberta.ca/~santanad
Posts Media Videos Starter Packs
Pinned
programsynthesis.bsky.social
I recently spoke at IPAM's Naturalistic Approaches to Artificial Intelligence Workshop, sharing some of the programmatic perspectives we're exploring in reinforcement learning research.

youtu.be/UNpg05yxc3o?...
Levi Lelis - Learning Libraries of Programmatic Policies - IPAM at UCLA
YouTube video by Institute for Pure & Applied Mathematics (IPAM)
youtu.be
Reposted by Levi Lelis
matthewguz.bsky.social
Excited to announce that our work on Reinforcement Learning for Arachnophobia treatment has been accepted at ACM Transactions on Interactive Intelligent Systems! We found that an RL agent could more effectively adapt VR spiders to achieve specified anxiety levels in users compared to current SOTA.
A graph showing that a rules-based approach consistently underperformed at achieving desired anxiety levels measured in normalized SCL compared to an RL approach. A brownish red virtual spider a medium distance away A close by black fuzzy spider
Reposted by Levi Lelis
eugenevinitsky.bsky.social
Was talking to a student who wasn't sure about why one would get a PhD. So I wrote up a list of reasons!
www.eugenevinitsky.com/posts/reason...
Eugene Vinitsky
www.eugenevinitsky.com
Reposted by Levi Lelis
programsynthesis.bsky.social
Previous work has shown that programmatic policies—computer programs written in a domain-specific language—generalize to out-of-distribution problems more easily than neural policies.

Is this really the case? 🧵
programsynthesis.bsky.social
1. Is the representation expressive enough to find solutions that generalize?
2. Can our search procedure find a policy that generalizes?
programsynthesis.bsky.social
So, when should we use neural vs. programmatic policies for OOD generalization?

Rather than treating programmatic policies as the default, we should ask:
programsynthesis.bsky.social
As an illustrative example, we changed the grid-world task so that a solution policy must use a queue or stack to solve a navigation task. FunSearch found a Python program that provably generalizes. As one would expect, neural nets couldn’t solve the problem.
programsynthesis.bsky.social
Are neural and programmatic policies similar in terms OOD generalization? We don't think so. We think that benchmark problems used in previous work actually undervalue what programmatic representations can do.
programsynthesis.bsky.social
Programmatic policies appeared to generalize better in previous work because they never learned to go fast in the easy training tracks. Neural nets optimized speed well, which made it difficult to generalize to tracks with sharp curves.
programsynthesis.bsky.social
In a car-racing task, we adjusted the reward to encourage cautious driving. Neural nets generalized just as well as programmatic policies.
programsynthesis.bsky.social
We had to perform simple changes to the neural policies' training pipeline to attain similar OOD generalization to that exhibited by programmatic ones.

In a grid-world problem, we used the same sparse observation space as used with the programmatic policies augmented with the agent's last action.
programsynthesis.bsky.social
In a preprint, led by my Master's student Amirhossein Rajabpour, we revisit some of these OOD generalization claims and show that neural policies generalize just as well as programmatic ones on benchmark problems used in previous work.

Preprint: arxiv.org/abs/2506.14162
arXiv.org e-Print archive
arxiv.org
programsynthesis.bsky.social
Previous work has shown that programmatic policies—computer programs written in a domain-specific language—generalize to out-of-distribution problems more easily than neural policies.

Is this really the case? 🧵
Reposted by Levi Lelis
sharky6000.bsky.social
If like me your Discover feed has been even worse lately and you are here for ML/AI news and discussion, check out these two feeds:

- Paper Skygest
- ML Feed: Trending

Links below 👇
Reposted by Levi Lelis
martinklissarov.bsky.social
As AI agents face increasingly long and complex tasks, decomposing them into subtasks becomes increasingly appealing.

But how do we discover such temporal structure?

Hierarchical RL provides a natural formalism-yet many questions remain open.

Here's our overview of the field🧵
Reposted by Levi Lelis
markgongloff.bsky.social
As hot as this summer is, it’s also one of the coolest we’ll ever enjoy again.

Just how much hotter and deadlier summers will get is still up to us. Right now we’re working hard to make them worse

🎁 link to my @opinion.bloomberg.com column:

www.bloomberg.com/opinion/arti...
The Heat Dome Wants a Word With Climate-Change Deniers
The temperatures gripping the US this week were made up to five times more likely by the fact that the atmosphere is simply hotter.
www.bloomberg.com
Reposted by Levi Lelis
eugenevinitsky.bsky.social
Hiring a postdoc to scale up and deploy RL-based planning onto some self-driving cars! We'll be building on arxiv.org/abs/2502.03349 and learn what the limits and challenges of RL planning are. Shoot me a message if interested and help spread the word please!

Full posting to come in a bit.
Robust Autonomy Emerges from Self-Play
Self-play has powered breakthroughs in two-player and multi-player games. Here we show that self-play is a surprisingly effective strategy in another domain. We show that robust and naturalistic drivi...
arxiv.org
programsynthesis.bsky.social
In addition to Sat's pointers, I would also take a look at the following recent paper by @swarat.bsky.social:

www.cs.utexas.edu/~swarat/pubs...

Also, the following paper covers most of the recent works on neuro-guided bottom-up synthesis algorithms:

webdocs.cs.ualberta.ca/~santanad/pa...
www.cs.utexas.edu
Reposted by Levi Lelis
matthewguz.bsky.social
We’re extending the AIIDE deadline! Partially due to author requests, partially due to a significant increase in submissions meaning I need to increase the PC!
programsynthesis.bsky.social
I wanted to thank the folks who reviewed our paper. Your feedback helped us improve our work, especially by asking us to include experiments on more difficult instances and the TSP. Thank you!
programsynthesis.bsky.social
Still, many important problems with real-world applications, such as the TSP and program synthesis, share some of the properties we assume in this work.
programsynthesis.bsky.social
The work has a few limitations. The policy learning scheme was evaluated only on needle-in-the-haystack deterministic problems. Also, since we are using tree search algorithms, we assume the agent has access to an efficient forward model.
programsynthesis.bsky.social
In other cases, where clustering seems unable to find relevant structure, the subgoal-based policies do not seem to harm the search, as in Sokoban problems.
programsynthesis.bsky.social
The empirical results are strong when clustering effectively detects the problem's underlying structure.