Tech Grandpa
techgrandpa.bsky.social
Tech Grandpa
@techgrandpa.bsky.social
Reposted by Tech Grandpa
Moonshot AI has released the updated Kimi K2-0905

- Enhanced coding capabilities, esp. front-end & tool-calling
- Context length extended to 256k tokens
- Improved integration with various agent scaffolds
September 5, 2025 at 7:14 AM
Does anyone know if this is the new DeepSeek model? No model card, no benchmarks so far, but unsloth already made a quant 😊 huggingface.co/collections/...
DeepSeek-V3.1 - a deepseek-ai Collection
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
huggingface.co
August 20, 2025 at 7:31 PM
More #gpt-5 + #cursor impressions...
I tried to fix a bug while running 80-90% context size and it basically circled around the same ideas, no matter what I told it for an hour back and forth. As the context got too big, I started a new chat with the "auto" mode (by accident) and it one-shotted it.
August 11, 2025 at 10:24 AM
#GPT-5 Take away:
- Coding is ok on medium size code bases
- Tool calling seems great, but I haven't tested it enough to tell
- I miss the variety of models before
- the price is a huge win for #openai - people don't yet understand the impact
- the router is annoying, but something others will adapt
August 10, 2025 at 10:26 PM
Well, the gpt-5 launch was ... interesting. I spent the last days testing it out in cursor and I am not impressed unfortunately. The thinking was taking many loops and working on profanities. The UI design were also not winning prices - not bad, don't get me wrong, just not great.
August 10, 2025 at 10:17 PM
Reposted by Tech Grandpa
🤣🤣🤣

Source: x.com/avichal/stat...
January 27, 2025 at 8:47 PM
Great comparison of local LLMs and their performance on consumer grade cards (24GB RAM limit):
www.reddit.com/r/LocalLLaMA... #AI #LLM #homelabai #localaiagent
From the LocalLLaMA community on Reddit: I benchmarked (almost) every model that can fit in 24GB VRAM (Qwens, R1 distils, Mistrals, even Llama 70b gguf)
Explore this post and more from the LocalLLaMA community
www.reddit.com
January 24, 2025 at 2:58 PM
I tested the new DeepSeek.r1 14b, which performs pretty well and is a good middle ground between speed, VRAM consumption and quality. It fails the “strawberry” test, but I can live with that 😉. If you can afford it, go with the 70b model though.

ollama.com/library/deep...
deepseek-r1:14b
DeepSeek's first generation reasoning models with comparable performance to OpenAI-o1.
ollama.com
January 22, 2025 at 8:42 PM
Reposted by Tech Grandpa
A short guide to run DeepSeek R1 (all 671B of it) on a home cluster of Macs with mlx.distributed.

gist.github.com/awni/ec071fd...
January 22, 2025 at 7:15 AM
I just posted a video about #NVIDIA and its project DIGITS. I not only think that it is a marvel of engineering to put so much #ai power into such a tiny package, but I also think that this is the beginning of a revolution: Edge AI computing for the masses. Check it out:
youtu.be/NEi9oJbwZC4
Will Project DIGITS Revolutionize The Way We Use LLMs Forever?
YouTube video by The Nitty-Gritty
youtu.be
January 21, 2025 at 11:35 PM
In case you have been living under a rock, deepseek has released its new r1 thinking model rivaling OpenAI’s o1 family, but being open source and MIT licensed! Ollama and HF already provide quants and distilled versions! Great times! github.com/deepseek-ai/...
GitHub - deepseek-ai/DeepSeek-R1
Contribute to deepseek-ai/DeepSeek-R1 development by creating an account on GitHub.
github.com
January 21, 2025 at 11:12 PM