Tim Kellogg
banner
timkellogg.me
Tim Kellogg
@timkellogg.me
AI Architect | North Carolina | AI/ML, IoT, science

WARNING: I talk about kids sometimes
Pinned
Social Media “Nutrition Label” for me for the last several days (thanks nano banana!)
you can ABSOLUTELY lose money, and that can be a good thing in the right situations, like if you’re capturing market

it’s possible o3 didn’t actually make money after the price cut, but they were just angling to capture the market for gpt-5 which would make money after
November 25, 2025 at 3:51 PM
oh yeah, they work for pinning versions, but ime submodules are basically the worst way of managing dependencies and i’ve had my share of scars from them
November 25, 2025 at 3:30 PM
too many people seem to be convinced that LLM vendors set prices on a cost plus basis

no, the advantage of closed weights is you can explore prices completely detached from cost. You’re free to set prices based purely on what people will pay, the value they get from it
November 25, 2025 at 3:29 PM
you don’t even need submodules. i just clone into a .clones dir and add it to .gitignore, works fine

Also, most agents will seek out dependency code in venv, node_modules, etc.
November 25, 2025 at 3:19 PM
are you saying that based on their methods reported in the model card?
November 25, 2025 at 2:22 PM
codex just taught me about jina.ai reader

an API you can easily use via curl that takes a URL and converts it to LLM-friendly text. Free to use, afaict

github.com/jina-ai/reader
GitHub - jina-ai/reader: Convert any URL to an LLM-friendly input with a simple prefix https://r.jina.ai/
Convert any URL to an LLM-friendly input with a simple prefix https://r.jina.ai/ - jina-ai/reader
github.com
November 25, 2025 at 2:20 PM
calling out bs on my own posts because microsoft can’t be trusted to not commit chart crimes
community note: using cost on the y axis makes it appear like cheaper models are more capable on pass@3
November 25, 2025 at 2:10 PM
one side conversation i had multiple times at AIE was that maybe mono repos are good

if all of your dependencies are sitting on disk, the agent doesn’t need to rely on documentation

even wo monorepos, it’s a good idea to clone tricky dependencies locally
November 25, 2025 at 2:04 PM
community note: using cost on the y axis makes it appear like cheaper models are more capable on pass@3
November 25, 2025 at 1:59 PM
i only use it via MCP and i find it a bit shaky. the results are decent but it’s historically been slow. i’m hoping this changes things
November 25, 2025 at 1:56 PM
Fara 7B: A cheap & capable open weights computer use agent (CuA)

they got within a few points of o3’s performance using only 4k training data points (yes, synthetic)

www.microsoft.com/en-us/resear...
November 25, 2025 at 1:54 PM
Exa 2.1: both fast and accurate search (that’s not Google)

available both as an MCP server & web UI

exa.ai/blog/exa-api...
November 25, 2025 at 1:42 PM
well, i meant that as a joke. stuff in the pic aren’t exactly positive traits
November 25, 2025 at 1:22 PM
this is the value of these new scaled models

GPT-5-Pro could probably do it too, but you’d pay like $30 for one shot

Gemini 3 & Opus 4.5 can still run fast & cheap bc they’re extremely sparse MoE, but solve very tricky problems

we truly need scale along both axes
Opus 4.5 solved a very tricky, complex problem in one session for me (VS Code Agent mode) that Sonnet 4.5 had been giving up on all day yesterday (I'm quite relentless).
November 25, 2025 at 12:48 PM
that’s it, i’m calling it, software engineering is over

AI can do everything an engineer can do
Yeah it’ll do that now I’ve head
November 25, 2025 at 12:23 PM
lol what is this?
November 25, 2025 at 12:13 PM
Reposted by Tim Kellogg
Opus 4.5 solved a very tricky, complex problem in one session for me (VS Code Agent mode) that Sonnet 4.5 had been giving up on all day yesterday (I'm quite relentless).
November 25, 2025 at 7:32 AM
**i** haven’t tried doing that, but sonnet 4.5 and gpt-5.2 are both aware of their context length
November 25, 2025 at 12:12 PM
nano banana infographic => Veo 3 pan around character => create Sora character
November 25, 2025 at 1:27 AM
merging the kiddo’s trio of passions for space facts, hamsters, and K-pop demon hunter
November 25, 2025 at 12:44 AM
i built this for myself a few months ago. it worked well, except that i only launched them in subagents (to preserve the prefix cache). this would probably work a lot better

no such thing as too many tools!
A tool for searching for relevant tools to keep context clean?

Was thinking about this last night as I approached sleep and glad to find this morning that one of the thought leaders rolled out this capability

www.anthropic.com/engineering/...
Introducing advanced tool use on the Claude Developer Platform
Claude can now discover, learn, and execute tools dynamically to enable agents that take action in the real world. Here’s how.
www.anthropic.com
November 25, 2025 at 12:39 AM
o3 came out in april
November 25, 2025 at 12:29 AM
are you saying you won’t listen to the episode?
November 24, 2025 at 11:47 PM
for those unaware — Ilya was a core researcher at OpenAI (founder??). You can probably credit the success of LLMs mostly to him

he quit and started SSI saying that he won’t say a single thing until they reach ASI. got many billions in funding

and he hasn’t. it’s been completely silent, until now..
November 24, 2025 at 11:34 PM
are TPUs actually that much faster? i just figured they’re in the same general range
November 24, 2025 at 10:47 PM