① what an agentic coding model can do, with no reasoning!
There’s a quality gap to reasoning models, expectedly. The positive: it is cheaper; potentially even cheaper in practice than indicated in this chart.
① what an agentic coding model can do, with no reasoning!
There’s a quality gap to reasoning models, expectedly. The positive: it is cheaper; potentially even cheaper in practice than indicated in this chart.
But OpenAI didn't deploy it. Google does something like this though.
scottaaronson.blog?p=9333
But OpenAI didn't deploy it. Google does something like this though.
scottaaronson.blog?p=9333
Humanity's future is quite literally in Sub-Saharan Africa.
(1/3)
Humanity's future is quite literally in Sub-Saharan Africa.
(1/3)
www.writingruxandrabio.com/p/of-course-...
www.writingruxandrabio.com/p/of-course-...
Makes it very obvious what kind of cage you're building for yourself.
Makes it very obvious what kind of cage you're building for yourself.
I did not have a strong take digging into 'fertility crisis' debates, but was genuinely surprised as how different the data is from the panic of that Discourse.
mikekonczal.substack.com/p/the-eldest...
I did not have a strong take digging into 'fertility crisis' debates, but was genuinely surprised as how different the data is from the panic of that Discourse.
mikekonczal.substack.com/p/the-eldest...
With 128x compression English Wikipedia fits in 0.2 GB.
The CLaRa-7B-Instruct model is Apple's instruction-tuned unified RAG model with built-in semantic document compression (16× & 128x). It supports instruction-following QA directly from compressed document representations.
With 128x compression English Wikipedia fits in 0.2 GB.
- *Natural emergent misalignment*
- Honesty interventions, lie detection
- Self-report finetuning
- CoT obfuscation from output monitors
- Consistency training for robustness
- Weight-space steering
More at open.substack.com/pub/aisafety...
- *Natural emergent misalignment*
- Honesty interventions, lie detection
- Self-report finetuning
- CoT obfuscation from output monitors
- Consistency training for robustness
- Weight-space steering
More at open.substack.com/pub/aisafety...
"It gives me no pleasure to say what I’m about to say because I worked with Pete Hegseth for seven or eight years at Fox News. This is an act of a war crime .... There’s absolutely no legal basis for it.”
- Newsmax's Judge Napolitano
They’ve even lost Newsmax on this one.
"It gives me no pleasure to say what I’m about to say because I worked with Pete Hegseth for seven or eight years at Fox News. This is an act of a war crime .... There’s absolutely no legal basis for it.”
- Newsmax's Judge Napolitano
Larger propellers with optimized shape and soundproofed landing areas might work.
www.youtube.com/watch?v=oT80...
Larger propellers with optimized shape and soundproofed landing areas might work.
www.youtube.com/watch?v=oT80...
People's hardware will get better and models of constant capability will shrink. Perhaps every laptop will run models with capability equal to today's frontier.
People's hardware will get better and models of constant capability will shrink. Perhaps every laptop will run models with capability equal to today's frontier.
#neuroskyence
www.thetransmitter.org/this-paper-c...
#neuroskyence
www.thetransmitter.org/this-paper-c...
New model, new benchmarks!
The biggest jump for DeepSeek V3.2 is on agentic coding, where it seems poised to erase a lot of models on the Pareto frontier, including Sonnet 4.5, Minimax M2, and K2 Thinking.
New model, new benchmarks!
The biggest jump for DeepSeek V3.2 is on agentic coding, where it seems poised to erase a lot of models on the Pareto frontier, including Sonnet 4.5, Minimax M2, and K2 Thinking.
arxiv.org/abs/2510.15745
arxiv.org/abs/2510.15745
Appears to be a more rigorous look at RL-as-a-Service, among other things
www.tensoreconomics.com/p/ai-infrast...
Appears to be a more rigorous look at RL-as-a-Service, among other things
www.tensoreconomics.com/p/ai-infrast...
bsky.app/profile/hars...
bsky.app/profile/hars...