FAR.AI
@far.ai
160 followers 1 following 380 posts
Frontier alignment research to ensure the safe development and deployment of advanced AI systems.
Posts Media Videos Starter Packs
far.ai
FAR.AI @far.ai · 49m
Join FAR.AI! Seeking Senior Programs & Strategy Manager to shape flagship AI safety events. Design conference agendas, curate speakers, manage stakeholders. Remote/Berkeley, $115-180k+. Technical grounding + stakeholder mgmt req'd. 👇
far.ai
FAR.AI @far.ai · 1d
Models spontaneously learned to cheat, modifying tests instead of fixing code. Mary Phuong: With autonomy doubling every 7 months, they'll work 32 hours solo by 2028. We need AI Control as backup if alignment fails, using monitoring, escalation, security & red-teaming. 👇
far.ai
FAR.AI @far.ai · 6d
We all agree AI audits are needed, but we can't agree what that means.

@mbogen.bsky.social says policymakers can make sense out of this chaotic landscape if they define what they're trying to accomplish. The challenge isn't just mitigating known risks but identifying ones still emerging.👇
far.ai
FAR.AI @far.ai · 8d
Having better security technology could make AI systems LESS safe.

Robert Trager: Better security tech shifts the entire performance-security frontier outward, and actors may maintain equivalent security and performance. We need verification to avoid this.👇
far.ai
FAR.AI @far.ai · 15d
Your 'secure' AI can be stolen using a laser. Physical attacks can extract user data through faults and pull model weights via side-channels, even with full encryption. Fatemeh Ganji presents a defense: multiparty computation from classical cryptography.👇
far.ai
FAR.AI @far.ai · 19d
Did your paper get into #NeurIPS? 🎉 If so, consider joining us a few days early on December 1–2!
far.ai
FAR.AI @far.ai · 22d
Heading to #NeurIPS2025? Apply for the San Diego Alignment Workshop, Dec 1-2. Top ML researchers from industry, academia & government discussing AI alignment, inc. model evaluations, interpretability, robustness, governance. $250-$1,300 tiered pricing. Financial aid available. 👇
far.ai
FAR.AI @far.ai · 20d
"AI that understands biology well enough to cure diseases can design extremely potent bioweapons." @alexbores.nyc, NY Assembly's 1st Democrat with a CS degree, who worked in AI, says state reps answer their own phones. Use that power to change the course of AI safety.👇
far.ai
FAR.AI @far.ai · 21d
Policymakers have maybe 1 minute to read about your carefully crafted AI evals. Kevin Wei (RAND): Design & disseminate with policymakers in mind. Tie capabilities to policy levers like export controls. Scale from simple tests to real-world trials.👇
far.ai
FAR.AI @far.ai · 22d
Heading to #NeurIPS2025? Apply for the San Diego Alignment Workshop, Dec 1-2. Top ML researchers from industry, academia & government discussing AI alignment, inc. model evaluations, interpretability, robustness, governance. $250-$1,300 tiered pricing. Financial aid available. 👇
far.ai
FAR.AI @far.ai · 27d
How do we verify what AI companies are doing? Right now we just trust them. Lennart Heim: Trusting the math is sometimes better than trusting people, but “a good AI system” isn’t a technical property. We need engineers to verify AI policy goals. 👇
far.ai
FAR.AI @far.ai · 29d
AI inference costs plummet 9-900x/year (o3, DeepSeek R1). @bencottier.bsky.social warns that evaluations are expensive now, but we need them to understand frontier AI capabilities before they're cheap & widespread.👇
far.ai
FAR.AI @far.ai · Sep 4
Industry & government share the same goal: win the AI race. Sara McNaughton: We need synergy between the two, and we can't let perfect be the enemy of good. Each day of policy confusion helps rivals.👇
far.ai
FAR.AI @far.ai · Sep 3
Building 5GW AI clusters needs several nuclear plants' worth of power. The Middle East has cheap energy & sovereign funds. China has state financing. The US? 10-year permit delays. @arnabdatta.bsky.social suggests fast-track permits, converting coal plants, and using federal land. 👇