Kyle Baxter
banner
kbaxter.bsky.social
Kyle Baxter
@kbaxter.bsky.social
Cause it feels like I've been, I've been here before. The lite brite is now black and white. Exploring LLMs and HCI

http://TightWind.net/
This post seems to run somewhat counter to Anthropic's message that Claude Code (or a harness like it) together with skills is a path toward a generally-applicable agent for many tasks (something I believe is correct). I don't know if I hav a thought behind that, but I find that interesting.
December 3, 2025 at 1:38 AM
Is plan + incremental work completing with "commitable" units + version control for work product generalizable? The post identifies this as an area of exploration. It feels like this can be stated in a more abstract manner that is widely applicable.
December 3, 2025 at 1:35 AM
Some thoughts.

1. I'm not sure this requires a separate "initialization agent," vs instructions for how to plan, and instructions for how to perform development.

2. This feels like a universal problem for a harness/orchestrator working on a task of meaningful complexity for any domain.
December 3, 2025 at 1:35 AM
To solve this, they encouraged the agent to do incremental work, commit it, and update the requirement status. This leaves the project in a clean state where a model with a fresh window can pick up a new requirement and the codebase and continue on.
December 3, 2025 at 1:35 AM
The problem they saw is the model wanted to one-shot tasks, leading to poor and incomplete work. They created a first step of drafting a detailed set of requirements to work against; this led to window overflow which compaction didn't solve.
December 3, 2025 at 1:35 AM
What’s especially unfortunate is we need to keep pulling more of the industry here, but there just isn’t a critical mass (or even near one) that would justify most folks only being here. Contributes to Twitter’s perpetual motion machine.
November 26, 2025 at 8:10 AM
Thanks! This is the same realization I’ve had over the last 1-2 months too, starting with realizing that agents (with the right combo of prompt and tools, like memory and filesystem) are generalizable (i.e. “deep agents,” but bad term). And then agent skills was like a bolt of lightning
November 22, 2025 at 11:23 PM
Those spirits can be given identities which are persistent. Under the hood, there is no “agent,” just an orchestrator, but you can operate with the fiction of them being a tangible persistent thing
November 22, 2025 at 10:59 PM
(Thinking out loud), where I’m ending up is an agent ecosystem (say, within an enterprise) benefits from an “orchestrator” primitive, whereas with 1:1 spirit:body pairing wasn’t req’d. Orchestrator executes a spirit, whatever complexity it has (ranging in analogy from single agent to multi-agent).
November 22, 2025 at 10:59 PM
In that sense, the “body” is infinitely copyable, and is more an actuator in an env for the spirit. But we still need the spirit+body to reason about what are the actors within an environment (I think). But it feels weird given that separation of spirit and body to recreate “agent” for that purpose
November 22, 2025 at 10:39 PM
imo this distinction is relatively new and really important. “Agent” before was usually a singular combo of what you refer to as the spirit and body; now the spirit is separately definable and loadable. I am still struggling tbh with what is the primary unit to reason about, spirit or body
November 22, 2025 at 10:37 PM
They hit on this in a post a few weeks ago: www.anthropic.com/engineering/... Skills really does feel like a durable phase change on top of model+tools
Code execution with MCP: building more efficient AI agents
Learn how code execution with the Model Context Protocol enables agents to handle more tools while using fewer tokens, reducing context overhead by up to 98.7%.
www.anthropic.com
November 22, 2025 at 3:06 PM
Yeah, that’s what’s captured my thoughts for the last few weeks. I’ve been experimenting with having Claude write skills for various things using primary source documents and it’s done a quite good job. A model doing that to formalize a task path seems very doable, and yeah, it is continual learning
November 22, 2025 at 2:59 PM
Any new takeaways from this talk re skills? I’ve been pretty fascinated with their concept of “tool” being a script the model retrieves (rather than a function call) and runs in a sandbox, with the model able to also formalize paths it takes to solve tasks as skills in its sandbox
November 22, 2025 at 5:10 AM
That makes sense. Knowledgebase/sources might be relevant to some fiction writing as well (to ground it, or even to create a knowledgebase for the fictional world to maintain consistency beyond context window horizon). Maybe. This would be a fun thing to play with with Claude
November 16, 2025 at 7:37 PM
Kind of curious how much overlap there ends up being in process for writing of different types. Fiction, nonfiction, technical analysis, memos, etc. There’s probably a good agent skills angle to generalizing it
November 16, 2025 at 7:30 PM
The library means we can add stuff over time to it and control what goes in our meal plans, while the meal plan list also creates a history of what we’ve cooked, so each week’s meal plan has variety. Quite useful, if boring.
November 16, 2025 at 6:05 PM
For example, I use Reminders for a Grocery list. I have a weekly meal plan list, and is the basis for updating the grocery list. I also created a Recipe Library list. Using these, each week, and I can have Claude create a meal plan using the library and update the grocery list.
November 16, 2025 at 6:05 PM