Bálint Gyevnár
@gbalint.bsky.social
110 followers 210 following 37 posts
Postdoc at CMU • Safeguarding scientific integrity in the age of AI scientists • PhD at University of Edinburgh • gbalint.me • 🇭🇺🏴󠁧󠁢󠁳󠁣󠁴󠁿
Posts Media Videos Starter Packs
Pinned
gbalint.bsky.social
Preprint alert 🎉 Introducing the Agentic eXplanations via Interrogative Simulations (AXIS) algo.

AXIS integrates multi-agent simulators with LLMs by having the LLMs interrogate the simulator with counterfactual queries over multiple rounds for explaining agent behaviour.

arxiv.org/pdf/2505.17801
Flowchart of the AXIS algorithm with 5 parts. The top-left has the memory, the centre-left has the user query, the centre-bottom has the final explanation, the centre has the LLM, and the right has the multi-agent simulator. Screenshot of the arXiv paper
gbalint.bsky.social
Ironically, the “AI scientist”-written paper that was accepted to an ICLR workshop was about LSTMs and the paper didn’t cite Schmidhuber. AI scientists should have a Schmidhuber agent that verifies these important details in the future.
Reposted by Bálint Gyevnár
gracekind.net
Never ask a man his age, a woman her salary, or GPT-5 whether a seahorse emoji exists
Reposted by Bálint Gyevnár
kashhill.bsky.social
Adam Raine, 16, died from suicide in April after months on ChatGPT discussing plans to end his life. His parents have filed the first known case against OpenAI for wrongful death.

Overwhelming at times to work on this story, but here it is. My latest on AI chatbots: www.nytimes.com/2025/08/26/t...
A Teen Was Suicidal. ChatGPT Was the Friend He Confided In.
www.nytimes.com
gbalint.bsky.social
"I have wait a long time for this moment, my little red friend." - Emperor Palpatine, probably
teamcherry.bsky.social
Hollow Knight: Silksong will be available September 4 on all platforms and day one on Xbox Game Pass!

Watch the release trailer: youtu.be/6XGeJwsUP9c
Title: Hollow Knight: Silksong - Release Trailer
YouTube video by Team Cherry
youtu.be
gbalint.bsky.social
Having been playing The Art of Fauna while procrastinating, and it is the most relaxing game with beautiful illustrations of animals. I cannot recommend it enough: theartof.app/fauna/
The Art of Fauna: Cozy Puzzles
Discover the wonders of nature with this cozy puzzle game. Download Now!
theartof.app
gbalint.bsky.social
For a week, we visited @iyadrahwan.bsky.social's Center for Humans and Machines in Berlin, where we have met an incredible array of interdisciplinary researchers.

Special thanks to @neeleengelmann.bsky.social for hosting us and @alice-ross.bsky.social for organising the trip from the start!
gbalint.bsky.social
Have been neglecting Bluesky recently, so I am happy to share a big update 🎉
I will join @cmu.edu as a postdoc in September working with the incomparable @atoosakz.bsky.social and Nihar Shah on understanding risks from LLM co-scientists. If you are in Pittsburgh, I would love to connect!
gbalint.bsky.social
The Edinburgh RL & Agents reading group is back in action after a hiatus. Previous speakers come for all across the world, including DeepMind, CMU, Oxford, NUS, etc. Sign up for some great discussions about cutting-edge RL and agents research.
rl-agents-rg.bsky.social
Hello world! This is the RL & Agents Reading Group

We organise regular meetings to discuss recent papers in Reinforcement Learning (RL), Multi-Agent RL and related areas (open-ended learning, LLM agents, robotics, etc).

Meetings take place online and are open to everyone 😊
gbalint.bsky.social
The absolute state of peer review...
neuralnoise.com
"in 2025 we will have flying cars" 😂😂😂
Reposted by Bálint Gyevnár
peterjelinek.bsky.social
Orban: Budapest Pride ist verboten.

Budapest Pride: Liebe kannst du nicht verbieten♥️🏳️‍🌈

Angeblich mehr als 500.000 Menschen vor Ort.
gbalint.bsky.social
The @braiduk.bsky.social gathering did an amazing job with presenting artists and researchers who address real-world questions around AI by actually engaging with people and learning from them. After hearing two weeks of technical talks at CHAI and RLDM, this was a most welcome break of pace.
Shannon Vallor and Fabio Tollon on stage presenting their landscape study of responsible AI
gbalint.bsky.social
I had the most amazing time at RLDM learning a lot about RL and agent foundations, catching up with and meeting new friends.

Two things that really stood out to me are:
- Agency is Frame Dependent by from Dave Abel
- Rethinking Foundations of Continual RL by Michael Bowling

#RLDM2025
Members of the Edinburgh RL group in front of the RLDM poster
gbalint.bsky.social
I am heading to RLDM in Dublin this week to present our work on objective evaluation metrics for explainable RL. Hit me up there or send me a DM to connect if you are around.
gbalint.bsky.social
New preprint 🥳

Explainable RL is suffering from a reproducibility crisis and a lack of standards. To start fixing this, we curate and compare a range of objective human behavioural metrics to build reproducible, comparable, and epistemically grounded research.

arxiv.org/abs/2501.19256
Objective Metrics for Human-Subjects Evaluation in Explainable Reinforcement Learning
Explanation is a fundamentally human process. Understanding the goal and audience of the explanation is vital, yet existing work on explainable reinforcement learning (XRL) routinely does not consult ...
arxiv.org
gbalint.bsky.social
Preprint alert 🎉 Introducing the Agentic eXplanations via Interrogative Simulations (AXIS) algo.

AXIS integrates multi-agent simulators with LLMs by having the LLMs interrogate the simulator with counterfactual queries over multiple rounds for explaining agent behaviour.

arxiv.org/pdf/2505.17801
Flowchart of the AXIS algorithm with 5 parts. The top-left has the memory, the centre-left has the user query, the centre-bottom has the final explanation, the centre has the LLM, and the right has the multi-agent simulator. Screenshot of the arXiv paper
Reposted by Bálint Gyevnár
atoosakz.bsky.social
What can policy makers learn from “AI safety for everyone” (Read here: www.nature.com/articles/s42... ; joint work with @gbalint.bsky.social )? I wrote about some policy lessons for Tech Policy Press.
Reposted by Bálint Gyevnár
profwinikoff.bsky.social
Delighted to share our most recent paper: "A Scoresheet for Explainable AI" (with John & Sebastian).

It will be presented at @aamasconf.bsky.social later this month.

🎬 Short YouTube summary (5 minutes): www.youtube.com/watch?v=GCpf...

📝 Link to the paper on arXiv: arxiv.org/abs/2502.098...
A Scoresheet for Explainable AI
YouTube video by Michael Winikoff
www.youtube.com
gbalint.bsky.social
90-year-old Masako Wakamiya at the final keynote of #CHI2025 shared a cautiously optimistic vision of the future of AI and humanity, especially for the elderly, as we enter the age of 100-year-long lives. Her speech and work is truly inspiring.
View of an auditorium stage with slides showing in the centre and Wakamiya-san and a sign-language interpreter showing on the sides.
gbalint.bsky.social
Oh yeah. In English, because apparently it sounds sophisticated, or at least that is what I have heard on the internet... So it must be true
White body towel in plastic packaging with black text
gbalint.bsky.social
One of the things I find most unique about Japan are the unnecessary and questionable motivational quotes on just about anything.

"Humans can only put out what has been put into them."

says my pre-packed body towel in the fanciest of fonts.
Inspiring stuff
gbalint.bsky.social
The #CHI2025 plushie is looking too cute:
A white maneki-neko plushie with a CHI2025 scarf looking extra cute
gbalint.bsky.social
#CHI2025 has a badge tag for rejected author. 🥲I couldn't resist getting one for future use.
White translucent badge tag on a wooden table that says rejected author
gbalint.bsky.social
Our key takeaways are:
1. Designing causality for explanations from first principles is essential to fully understand what explanations to give to people about autonomous agents;
2. People prefer goal-oriented explanations for AVs, so focusing on those first might be beneficial.

🧵 7/7