#SaferAI
⚡ Efficient Models: Tiny AI matches big ones via compression.
🤖 Expert Reasoning: New AIs solve multi-step tasks like pros.
🔍 Safer AI: Interpretability and robustness improve transparency.
#AI2025 #EfficientAI #ExpertReasoning #SaferAI
View in Timelines
December 15, 2025 at 3:01 PM
AI is a mirror. If shortcuts trump values, the reflection will not flatter. Ask not 'how fast,' but 'how wisely' can we adopt. buff.ly/Hmjr4xj #DigitalTrust #SAFERAI #ExecLeadership
buff.ly
October 15, 2025 at 5:55 PM
We ignore one-size-fits-all bots. Style-aligned, RAG-grounded assistants meet us where we are - earning trust and shaping safer decisions in the moment. buff.ly/FeY4jgK #AIethics #Personalisation #SAFERAI
buff.ly
October 3, 2025 at 10:01 PM
29/31 2028: Through careful iteration, Safer-2 becomes both aligned and transparent. The massive compute advantage helps US stay ahead of China despite slower progress.

Cyberwar balance tips toward US as capability gaps grow.
#SaferAI
August 2, 2025 at 9:29 AM
Meta Overtakes Google DeepMind: Despite Progress in AI Risk Management, Industry Remains Far Below Adequate Standards, New SaferAI Ratings Reveal drive.google.com/file/d/1Q1pQ...
SaferAI Release.pdf
drive.google.com
July 24, 2025 at 5:54 AM
"ベンチマーク「ELEPHANT」"
開発:スタンフォード大学、カーネギーメロン大学、オックスフォード大学研究チーム
>大規模言語モデル(LLM)が人間よりも一貫して高い頻度でシカファンシーを示す…明らかに

非営利団体セーファーAI(SaferAI)ヘンリー・パパダトス常務理事
"現在、AIモデルは世界中の数百万人に対して猛スピードで導入され…ユーザーに対する説得力…ユーザーに関する情報を保持する能力が向上"
"これらすべてが「大惨事のすべての要素を備えた危険な状況」を生み出している"

"「安全性を確保するには時間がかかりますが、そのための時間が十分に費やされているとは思えません」"
June 2, 2025 at 3:06 AM
[15/25] 🚨 DAS xAI-PROBLEM:

“Concerning”: Elon Musks xAI hat:

• Nur einen ENTWURF (Feb 2025)
• VERPASSTE eigene Frist für finales Framework
• "Sehr schwache" Risikomanagement-Praktiken (SaferAI-Studie)
• Chatbot "Grok" mit problematischem Verhalten, gerade zuletzt
May 23, 2025 at 12:07 PM
A group representing a number of child safety organisations, calling itself thge "SaferAI for Children Coaliiton", has called on the new federal government to prioritise the use of AI for child sexual abuse, urging action on technological solutions and educating young people
May 23, 2025 at 12:28 AM
"[W]e are on a trajectory to build machines that are smarter and smarter. And one day, it's very plausible that they will be smarter than us, and then they will have their own agency, their own goals, which may not be aligned with ours. What happens to us then?" #AI #SaferAI
When I realized how dangerous the current agency-driven AI trajectory could be for future generations, I knew I had to do all I could to make AI safer. I recently shared this personal experience, and outlined the scientific solution I envision @TEDTalks⤵️
www.ted.com/talks/yoshua...
The catastrophic risks of AI — and a safer path
Yoshua Bengio — the world's most-cited computer scientist and a "godfather" of artificial intelligence — is deadly concerned about the current trajectory of the technology. As AI models race toward fu...
www.ted.com
May 22, 2025 at 7:15 AM
➡️ Chloé Touzet, Policy Lead chez SaferAI
May 14, 2025 at 9:16 AM
Link to paper: arxiv.org/abs/2504.118...
Joint work with: @ggarbacea.bsky.social Alexis Bellot, Jonathan Richens, Henry Papadatos, Simeon Campos, and Rohin Shah from Google DeepMind, University of Chicago, and SaferAI
Evaluating the Goal-Directedness of Large Language Models
To what extent do LLMs use their capabilities towards their given goal? We take this as a measure of their goal-directedness. We evaluate goal-directedness on tasks that require information gathering,...
arxiv.org
April 17, 2025 at 9:52 AM
Yes! Great Idea!⚡️
I’ve been thinking of this, too, as well as:
#BoycottGOOGLE
Because they are participating in the #COUP
- and also recently dissolved their #SaferAI-agreements
😎🇺🇸😎🌍😎 > Google too!
February 27, 2025 at 8:41 PM
Yes! Great Idea!⚡️
I’ve been thinking of this, too, as well as:
#BoycottGOOGLE
Because they are participating in the #COUP
- and also recently dissolved their #SaferAI-agreements
😎🇺🇸😎🌍😎 > Google too!
So, we are dumping ALL Amazon products as of this week. Permanently. Finding other sources for what we need that are not dumping DEI. THIS ECONOMIC BOYCOTT SHOULD BE HAPPENING EVERY FRIDAY, NOT JUST THIS FRIDAY.
February 27, 2025 at 8:41 PM
The best contribution of SaferAI to the conversation has always been to point out that we already have lots of good risk management frameworks. No need to reinvent something for AI: arxiv.org/abs/2502.06656
A Frontier AI Risk Management Framework: Bridging the Gap Between Current AI Practices and Established Risk Management
The recent development of powerful AI systems has highlighted the need for robust risk management frameworks in the AI industry. Although companies have begun to implement safety frameworks, current a...
arxiv.org
February 21, 2025 at 7:13 PM
🗣️ “Quand on se penche sur l'industrie aérospatiale, ou l'aviation, ou le nucléaire, on retrouve les mêmes éléments clés de gestion du risque, expose Chloé Touzet du think tank français SaferAI. On n'a pas de raison de penser qu'elles ne s'appliquent pas au domaine de l'IA aussi.”
February 12, 2025 at 10:15 AM
Building a Smarter, Safer Future: Why the Right Human-in-the-Loop Is Critical for Effective AI
#HiTL #SmaterAI #SaferAI #EffectiveAI #LLM #GenAI
tinyurl.com/4ekv7ahx
Right Human-in-the-Loop
Building a Smarter, Safer Future: Why the Right Human-in-the-Loop Is Critical for Effective AI
tinyurl.com
January 21, 2025 at 5:39 AM
Prevent AI's dark side:
Lock it with rules,
Share the responsibility,
Manage risks with care. #AIethics #SaferAI
December 14, 2024 at 12:48 PM
We face a series of challenges with AI, and not least among them is protecting children. As a part of the SaferAI for Children Coalition, I recommend the discussion paper we have authored over the last few months and was published today #ai #childprotection #onlinesafety icmec.org.au/saferai-for-...
SaferAI for Children Coalition Discussion Paper | ICMEC Australia
Explore ICMEC Australia's SaferAI for Children Coalition Discussion Paper, addressing AI's role in child protection.
icmec.org.au
November 26, 2024 at 2:30 AM