Local-first/Reactive Programming ⁙ LLM system evals ⁙ Startup lessons ⁙ Game design quips.
Longform: https://interjectedfuture.com
Podcast: https://www.youtube.com/@techniumpod
I can then ask the LLM to extract a prompt from my code by asking me questions about it. Then I use that prompt to apply it to other places in a refactor.
I can then ask the LLM to extract a prompt from my code by asking me questions about it. Then I use that prompt to apply it to other places in a refactor.
www.moltbook.com/m/blessthei...
www.moltbook.com/m/blessthei...
An even weirder experiment is to let them loose on a DAO. Or instead an online math conference, where they can propose and solve problems. It'd be like SETI@home.
An even weirder experiment is to let them loose on a DAO. Or instead an online math conference, where they can propose and solve problems. It'd be like SETI@home.
I use Grok to find the consensus view on a topic on Twitter.
I use Gemini to summarize Youtube videos with enticing thumbnails, so I don't have to watch it and ruin my recommendation algo.
I use Grok to find the consensus view on a topic on Twitter.
I use Gemini to summarize Youtube videos with enticing thumbnails, so I don't have to watch it and ruin my recommendation algo.
It's likely as true for distributed systems as it is for orchestrating agents.
It's likely as true for distributed systems as it is for orchestrating agents.
When we figure out which humans outperform others + why, we have a sense for what the critical skills are.
When we figure out which humans outperform others + why, we have a sense for what the critical skills are.
At the moment, that documentation is only useful in the moment for guiding the agent.
At the moment, that documentation is only useful in the moment for guiding the agent.
Even better is if I get to drive a tank or ride a bike to work to see if I can beat the time.
Even better is if I get to drive a tank or ride a bike to work to see if I can beat the time.
I think how Geordi LaForge uses "Computer" to do engineering is more akin to how we'll do it in the near future.
Rotating the Space: On LLMs as a Medium for Thought
sbgeoaiphd.github.io/rotating_the...
I think how Geordi LaForge uses "Computer" to do engineering is more akin to how we'll do it in the near future.
When you're expert and sure, they don't want to be second-guessed. It's a diminishment of status to have to confirm.
x.com/usgraphics/...
When you're expert and sure, they don't want to be second-guessed. It's a diminishment of status to have to confirm.
x.com/usgraphics/...
Their estimates of how long things take are an order of magnitude off. But even if estimates were in the ballpark, would LLMs be any better at doing estimation if they got feedback? Would their estimates also be spikey?
Their estimates of how long things take are an order of magnitude off. But even if estimates were in the ballpark, would LLMs be any better at doing estimation if they got feedback? Would their estimates also be spikey?
Now, I wonder if Terrance Tao had such of experience in Mathematics, or if his story was going to be, "I wanted to play football, but I found I was terrible at it"
x.com/getjonwithi...
Now, I wonder if Terrance Tao had such of experience in Mathematics, or if his story was going to be, "I wanted to play football, but I found I was terrible at it"
x.com/getjonwithi...
bsky.app/profile/sne...
bsky.app/profile/sne...
in these demos "thinking" is disabled, which makes the model return tokens very quickly (all videos are realtime), and I find these rapid responses pretty good for the use-cases I'm experimenting with, like:
executing simple diagrams ...
in these demos "thinking" is disabled, which makes the model return tokens very quickly (all videos are realtime), and I find these rapid responses pretty good for the use-cases I'm experimenting with, like:
executing simple diagrams ...
Meanwhile, vibe coders are convinced it's like compilers, and lots of people aren't looking at the code.
Meanwhile, vibe coders are convinced it's like compilers, and lots of people aren't looking at the code.