Tim Kellogg
banner
timkellogg.me
Tim Kellogg
@timkellogg.me
AI Architect | North Carolina | AI/ML, IoT, science

WARNING: I talk about kids sometimes
Pinned
Social Media “Nutrition Label” for me for the last several days (thanks nano banana!)
Reposted by Tim Kellogg
At the risk of starting the flame war to end all flame wars...

Modern LLMs (GPT-5.1, Claude 4.5, Gemini 3) produce excellent code and can be a significant productivity boost to software engineers who take the time to learn how to effectively apply them - especially if used with coding agent tools
November 27, 2025 at 7:55 PM
i hereby am requesting that @ai2.bsky.social make a 1T base model and have PrimeIntellect posttrain it
November 27, 2025 at 7:41 PM
they take GLM-4.5-Air (small model) and post train it to out-perform the already near-SOTA 3x-larger GLM-4.5

but the model is an advertisement for their infrastructure (that you should use!), and so peek at that too! You should be able to replicate this for your domain
November 27, 2025 at 4:28 PM
DeepSeek-Math-V2: self-verification

Fascinating paper that explores how to RL but focused on process over outcome

It’s sort of similar to a GAN, but with loops for each the generator & verifier as well as an outer loop

github.com/deepseek-ai/...
November 27, 2025 at 2:19 PM
Reposted by Tim Kellogg
thank u for choosing linkedin. here are a bunch of comments made by people you do not care about on posts from people you somehow care even less about. also your inbox is a personalised advertising billboard. also your notifications are randomised every morning based on what you will hate the most.
November 27, 2025 at 12:37 AM
in the past, leftward-leaning folk were tied together by one uniting force: progress

that was Obama’s schtick. Social, tech, economic, any kind of progress will do

now it feels like the left and right are fighting over which kind of *regress* is better

seems like someone will probably win
November 27, 2025 at 12:16 AM
inb4 the left goes anti-solar along the path to being anti-AI
Whatever happens at the model level, one thing is clear: hyperscalers are building enough new infrastructure to put city-scale amounts of power and compute behind AI.
November 26, 2025 at 11:49 PM
dropping truths into the gc
November 26, 2025 at 11:03 PM
for the 12 days of openai, i’m really hoping they release both GPT-5o and o4 + o5-mini. That would really round things out
November 26, 2025 at 4:48 PM
towards the end, Ilya has a part where he makes the case that the genome (thus evolution) doesn't dictate intelligence

i have a hunch that that's why he's taking a lot of crap from some parts of the tech bro crowd that's started leaning into eugenics. Might have nothing to do with his AI views
November 26, 2025 at 12:56 PM
this is insane. how small can we go?

the even cooler part is this all independent research
The threshold for consistent English/query understanding is now 3M parameters.
November 26, 2025 at 11:58 AM
new summary of the Ilya podcast just landed
If he was in a band, he would be saying that this next album is going to suck but it’s “getting back to our roots” kind of suck.
November 26, 2025 at 11:53 AM
thinking more about Ilya’s strategy..

that’s why he’s on a podcast, to shape minds. he can’t just release a shitty model and be called a saint. he needs to control the narrative and provide context for what he’s done

if this doesn’t land, he’s likely screwed (ngl i don’t think it landed)
yeah, his idea is really like releasing a newborn baby and saying, “that’s it, work is done”

but it’s not done, it’s still got to learn

in our current approaches, it’s hard to conceive of that, because we’re bombarded by hype and marketing. i can’t imagine releasing an incapable model..
November 26, 2025 at 2:52 AM
Summary — He's got a divergent view of AGI

We're all pursuing a single behemoth that is *already* smarter than all humans when it's launched

He's pursuing an entity that is *capable of* being smarter

i.e. he's all in on continual learning
November 26, 2025 at 2:02 AM
alright, Opus 4.5 got me through a gnarly debugging session (not one shot) that Gemini 3 couldn't figure out

Opus 4.5 basically does not do doom loops, period. It's legit, I'm impressed.
November 25, 2025 at 10:41 PM
oh shit, NVIDIA’s in trouble
November 25, 2025 at 9:45 PM
we've reached AGI
November 25, 2025 at 9:33 PM
too many people seem to be convinced that LLM vendors set prices on a cost plus basis

no, the advantage of closed weights is you can explore prices completely detached from cost. You’re free to set prices based purely on what people will pay, the value they get from it
November 25, 2025 at 3:29 PM
codex just taught me about jina.ai reader

an API you can easily use via curl that takes a URL and converts it to LLM-friendly text. Free to use, afaict

github.com/jina-ai/reader
GitHub - jina-ai/reader: Convert any URL to an LLM-friendly input with a simple prefix https://r.jina.ai/
Convert any URL to an LLM-friendly input with a simple prefix https://r.jina.ai/ - jina-ai/reader
github.com
November 25, 2025 at 2:20 PM
calling out bs on my own posts because microsoft can’t be trusted to not commit chart crimes
community note: using cost on the y axis makes it appear like cheaper models are more capable on pass@3
November 25, 2025 at 2:10 PM
one side conversation i had multiple times at AIE was that maybe mono repos are good

if all of your dependencies are sitting on disk, the agent doesn’t need to rely on documentation

even wo monorepos, it’s a good idea to clone tricky dependencies locally
November 25, 2025 at 2:04 PM
Fara 7B: A cheap & capable open weights computer use agent (CuA)

they got within a few points of o3’s performance using only 4k training data points (yes, synthetic)

www.microsoft.com/en-us/resear...
November 25, 2025 at 1:54 PM
Exa 2.1: both fast and accurate search (that’s not Google)

available both as an MCP server & web UI

exa.ai/blog/exa-api...
November 25, 2025 at 1:42 PM
this is the value of these new scaled models

GPT-5-Pro could probably do it too, but you’d pay like $30 for one shot

Gemini 3 & Opus 4.5 can still run fast & cheap bc they’re extremely sparse MoE, but solve very tricky problems

we truly need scale along both axes
Opus 4.5 solved a very tricky, complex problem in one session for me (VS Code Agent mode) that Sonnet 4.5 had been giving up on all day yesterday (I'm quite relentless).
November 25, 2025 at 12:48 PM