My next blog post will be my yearly prediction post.
mkannen.tech/ai-and-the-n...
My next blog post will be my yearly prediction post.
mkannen.tech/ai-and-the-n...
Especially now that you can use their models for much cheaper elsewhere. Look at this contrast. Both OpenAI models. One is a new reasoning model at over 2000t/s for pennies and the other one is their mini model for twice the price.
Especially now that you can use their models for much cheaper elsewhere. Look at this contrast. Both OpenAI models. One is a new reasoning model at over 2000t/s for pennies and the other one is their mini model for twice the price.
Someone training a pure MLP model with xxB parameters?
arxiv.org/abs/2506.14761
Someone training a pure MLP model with xxB parameters?
arxiv.org/abs/2506.14761
arxiv.org/abs/2506.09804
arxiv.org/abs/2506.09804
Let me explain why this is a terrible way to compare models:
Let me explain why this is a terrible way to compare models:
deepmind.google/discover/blo...
deepmind.google/discover/blo...
All support 1 Million context
Big focus on improved coding capabilities.
4.1 Pricing:
Input/Output per million 2/8
Mini: 0.4/1.6
Nano: 0.1/0.4
openai.com/index/gpt-4-1/
All support 1 Million context
Big focus on improved coding capabilities.
4.1 Pricing:
Input/Output per million 2/8
Mini: 0.4/1.6
Nano: 0.1/0.4
openai.com/index/gpt-4-1/
The first one was the most interesting benchmark to follow.
The first one was the most interesting benchmark to follow.