Kale-ab Tessera
@kale-ab.bsky.social
2.3K followers 240 following 50 posts
ML PhD Student @ Uni. of Edinburgh, working on Multi-Agent Problems. | Organiser @deeplearningindaba.bsky.social‬ @rl-agents-rg.bsky.social‬ | 🇪🇹🇿🇦 kaleabtessera.com
Posts Media Videos Starter Packs
Reposted by Kale-ab Tessera
rl-agents-rg.bsky.social
📢 RL reading group Thursday @ 16:00 BST 📢

Speaker: Alex Lewandowski

Title: The World Is Bigger: A Computationally-Embedded Perspective on the Big World Hypothesis 🌍

Details: edinburgh-rl.github.io/reading-group
UoE RL Reading Group
University of Edinburgh Reinforcement Learning Reading Group
edinburgh-rl.github.io
kale-ab.bsky.social
Refreshing to see posts like this compared to "we have 15 papers accepted at X" 🙌
Reposted by Kale-ab Tessera
eugenevinitsky.bsky.social
None of our impactful papers have had an easy path through traditional venues.
Most cited paper? Rejected four times.
Most impactful paper? Poster at a conference.
But none of it matters because arxiv makes everything work
kale-ab.bsky.social
Great first couple of days at DLI @deeplearningindaba.bsky.social in Kigali 🇷🇼, some highlights include amazing talks talks by @verenarieser.bsky.social and Max Welling, great pracs and tuts, and of course the opening party ( before the rain 😢) 🎉 #DLI2025
Reposted by Kale-ab Tessera
deeplearningindaba.bsky.social
We’re excited to unveil the first #DLI2025 lineup of tutorials and practicals:

✨ Machine Learning Foundations
✨ Generative Models & LLMs for African languages

All tutorial content will also be available online after the Indaba. Don’t miss out, subscribe here 👉 lnkd.in/eCgXRqsV
kale-ab.bsky.social
🇨🇦 Heading to @rl-conference.bsky.social next week to present HyperMARL (@cocomarl-workshop.bsky.social) and Remember Markov (Finding The Frame Workshop).

If you are around, hmu, happy to chat about Multi-Agent Systems (MARL, agentic systems), open-endedness, environments, or anything related! 🎉
Remembering Markov poster.
Reposted by Kale-ab Tessera
deeplearningindaba.bsky.social
We are thrilled to announce our next keynote speaker
@wellingmax.bsky.social, Professor at the University of Amsterdam, Visiting Professor at Caltech and CTO & Co-Founder of CuspAI.
Catch his talk “How AI could transform the sciences” on August 18 at 4:30 PM GMT+2.
#DLI2025
Reposted by Kale-ab Tessera
rl-agents-rg.bsky.social
RL reading group TODAY @ 15:00 BST 🔥

Speaker: Cam Allen (Postdoc, UC Berkeley)

Title: The Agent Must Choose the Problem Model

Details: edinburgh-rl.github.io/reading-group
UoE RL Reading Group
University of Edinburgh Reinforcement Learning Reading Group
edinburgh-rl.github.io
kale-ab.bsky.social
Always nice to see when simpler methods + good evaluations > more complicated ones. 👌
kale-ab.bsky.social
Reading group is back for those interested in RL/MARL/agents/open-endedness and alike... First session today at 3pm BST, @mattieml.bsky.social is presenting the Simplifying TD learning/PQN paper. 🎉 Meeting link: bit.ly/4lfdaGR Sign up: bit.ly/40xNQDR
rl-agents-rg.bsky.social
We are super excited to kick things off again with Mattie Fellows (Postdoc @ FLAIR in Oxford) today 15:00 BST!

Paper: Simplifying Deep Temporal Difference Learning

Check out our website for full info edinburgh-rl.github.io/reading-grou...
Reposted by Kale-ab Tessera
rl-agents-rg.bsky.social
Hello world! This is the RL & Agents Reading Group

We organise regular meetings to discuss recent papers in Reinforcement Learning (RL), Multi-Agent RL and related areas (open-ended learning, LLM agents, robotics, etc).

Meetings take place online and are open to everyone 😊
kale-ab.bsky.social
This has happened to me too many times 🤦‍♂️ Also doesn't help that Jax and PyTorch use different default initialisations for dense layers.
kale-ab.bsky.social
Well done & well deserved!! 🎉🎉 It has been awesome to see this project evolve from the early days.
kale-ab.bsky.social
The Edinburgh one will be back and running soon. We are just updating the website and other things. There is this form for people interested - forms.gle/DAbkpN9b4cUt...
Edinburgh RL Reading Group
Please add your details so that you can remain on the mailing list for the RL Reading Group.
forms.gle
kale-ab.bsky.social
Forgot to also add ⚡ quickstart link for people who like to experiment on notebooks: github.com/KaleabTesser...
github.com
kale-ab.bsky.social
Thanks for checking it out! 👍 Good point, there might be an interesting link between MoEs and hypernets. We used hypernets since they're simpler (no need to pick or combine experts), and maximally expressive (gen weights directly).

Lol yes, will had a .gitignore, missed it when copying things over.
kale-ab.bsky.social
🎯 TL;DR: HyperMARL is a versatile approach for adaptive MARL -- no changes to the RL objective, preset diversity, or seq. updates needed. See paper & code below!

Work with Arrasy Rahman, Amos Storkey & Stefano Albrecht.

📜: arxiv.org/abs/2412.04233
👩‍💻: github.com/KaleabTessera/HyperMARL
HyperMARL: Adaptive Hypernetworks for Multi-Agent RL
Adaptability to specialised or homogeneous behaviours is critical in cooperative multi-agent reinforcement learning (MARL). Parameter sharing (PS) techniques, common for efficient adaptation, often li...
arxiv.org
kale-ab.bsky.social
⚠️ Limitations (+opportunity): HyperMARL uses vanilla hypernets, which can inc. param. count esp. MLP hypernets. In RL/MARL this matters less (actor-critic nets are small), and params grow ~const with #agents, so scaling remains strong. Future work could explore chunked hypernets.
kale-ab.bsky.social
🔎 We also do ablations and see the importance of the decoupling and the simple initialisation scheme we follow.
kale-ab.bsky.social
📊 We validate HyperMARL across various diverse envs (18 settings; up to 20 agents) and find that it achieves competitive mean episode returns compared to NoPS, FuPS, and modern diversity-focused methods -- without using diversity losses, preset diversity levels or seq. updates.
kale-ab.bsky.social
💡To address the coupling problem, we propose 𝐇𝐲𝐩𝐞𝐫𝐌𝐀𝐑𝐋: a method that explicitly 𝐝𝐞𝐜𝐨𝐮𝐩𝐥𝐞𝐬 obs- and agent-conditioned gradients with hypernetworks. This means obs grad noise is avg. per agent (Zᵢ) before applying agent-cond. grads (Jᵢ) -- unlike FuPS, which entangles both.
kale-ab.bsky.social
🔬 We isolate FuPS’s failure in matrix games: shared policies struggle when agents need to act differently. Inter-agent gradient interference is at play -- especially when obs and agent IDs are 𝐜𝐨𝐮𝐩𝐥𝐞𝐝. Surprisingly, using only IDs (no obs) performed better and reduced interference.
Specialisation matrix game. Performance and gradient interference plots.
kale-ab.bsky.social
❓Existing methods add a diversity loss, use sequential updates or require knowing the optimal task diversity level beforehand. These can be hard to tune or inefficient. We ask: can shared policies adapt without any of the above?
kale-ab.bsky.social
⚖️ 𝐖𝐡𝐚𝐭’𝐬 𝐭𝐡𝐞 𝐢𝐬𝐬𝐮𝐞? In MARL, optimal performance requires representing the right behaviours. Separate networks per agent (NoPS) enable agent specialisation but is costly & sample-inefficient; shared networks (FuPS) are efficient but lack agent diversity/specialisation.