Andrew Gross
gross.systems
Andrew Gross
@gross.systems
Engineer at YipitData.
NYC Area

https://github.com/andrewgross/
https://gross.systems

I was told I had to add AI Engineer to my profile for the bots to find me.

Views my own, not my employer etc etc.
CC behaves differently with Shift-Enter in the terminal vs as a Cursor extension. Its maddening.
January 13, 2026 at 6:48 PM
If if you let it iterate on local hardware before crafting Kaggle notebooks for submission, I bet a lot of the trick would be to get it to be able to run and monitor long async jobs, and make decisions about when to optimize training speed.
January 13, 2026 at 3:57 AM
I want to figure out a harness so that I can run Claude Code against a Kaggle competition. I feel like it could pretty reliably score silver in many competitions, assuming it was given access to the forums and had a decent harness
January 13, 2026 at 3:57 AM
New requirement for monospace fonts: When coding agents draw diagrams the arrow lines are aligned.
January 9, 2026 at 3:13 AM
I used to think that typing speed beyond a certainly level had diminishing returns for programming. That said, it seems like there is a lot of value in being able to quickly convey prompts via typing, as voice transcription lacks some capabilities for things like linking files or code.
January 8, 2026 at 6:00 PM
I remember a while back there was a lot of discourse around codebases that are "greppable" vs those that are expected to be navigated with an IDE. Seems like greppable ended up being the correct choice for coding agents (at leat for now).
January 7, 2026 at 9:50 PM
Trying out a new thing: "This code dictated but not read" to absolve myself of all responsibility for LLM generated code, just like executives in the 60s would place all the blame for their bad ideas on their secretary.
December 12, 2025 at 3:57 AM
Release PySpark Toolkit 0.5.0:

* support for S3 Presigned URLs in Pure Pyspark
* `fdtf` - A generic UDTF decorator that can wrap an arbitrary row level python function and append its results as columns to your dataframe. Cuz UDTFs are annoying to use.
December 2, 2025 at 9:14 PM
One thing I noticed with a lot of AI blog spam is that they never seem to properly link things. I just went over an entire article about a tool I wanted to check out, and not once did it actually link to the tool.
November 28, 2025 at 12:22 PM
Biggest work distractions.

1. Arguing with healthcare providers about billing
2. Meetings

Distant 3rd: IDK IT Issues or something
November 19, 2025 at 3:25 PM
Anyways, would need to reproduce it to understand more about it, but seems to try actually bridge the gap from math to real implementation via their algorithm.
November 1, 2025 at 2:58 PM
As for temperature, their work seems to focus on looking at the state of the token distribution for the last token. I think they arent looking to reverse each step of generation but instead take the state when the last token is generated and show you can get back to the prompt from there.
November 1, 2025 at 2:55 PM
Worth browsing the paper. They start with the mathematical approach ignoring some of the realities and showing that they are reversible. But they also build out an algorithm for being able to produce the input prompt based on a set of hidden states for a model.
November 1, 2025 at 2:53 PM
I wonder how long until we see all these tools that are meant to stop overly-aggressive AI data crawlers start poisoning their data www.anthropic.com/research/sma....
A small number of samples can poison LLMs of any size
Anthropic research on data-poisoning attacks in large language models
www.anthropic.com
October 26, 2025 at 6:05 PM
Every new benchmark or tool I see screams that the real limiting factor for making effective systems with LLMs/ML is context + evals. Model "intelligence" is rarely the deciding factor now.
September 26, 2025 at 1:54 PM
Astounding to me that OpenAI has had their new billing dashboard for this long without a good way to tie an API key to usage. API keys get human names, but billing refers to them by `key_XXXXXXXXX`, with no mapping between them. Have to use the legacy dashboard platform.openai.com/account/usag...
OpenAI Platform
Explore developer resources, tutorials, API docs, and dynamic examples to get the most out of OpenAI's platform.
platform.openai.com
September 22, 2025 at 3:38 PM
"May you create a successful open source project" - Ancient Developer Curse
September 14, 2025 at 7:06 PM
Context is also very poorly used in a couple ways. I feel like most tool calls etc need to be done in a sub agent with a concise summary returned to the main context (somehow not dropping important details). The info density of english tokens is bad and I can see using something else.
September 12, 2025 at 11:48 AM
Context retrieval is one of the worst areas right now. Since context fills up so fast, its hard to convey instructions effectively since most things fall off after 10-20k tokens. Not great to need to re import your claude md before every prompt.
September 12, 2025 at 11:46 AM
1mm context seems like a lot, and is in many ways, but it fills up extremely fast between code, prompts, instructions (claude md etc), tool descriptions, tool call results etc. We really need way more / a different system a la long term/short term memory. 100mm+ context to start
September 12, 2025 at 11:42 AM
Its becoming pretty apparenty from using agentic systems and tools that there are a few big blockers making them more effective

1. Context is way too small
2. Retrieval from that context sucks
3. Density of context is terrible
September 12, 2025 at 11:41 AM
I've tried using an lsp server with refactoring/renaming tools. It was great when Claude used it but the hardest part was getting Claude to actually use the MCP reliably. Probably a prompting issue. Could be fun to block the ability to use the Write() toolcall using Hooks and try to force it.
September 12, 2025 at 12:16 AM
It was a little annoying that Claude Code didn't have a way to limit the context so it was easy to use other models without manually running compact. I ended up hacking on the JS blob after reviewing the unminified code to find what I needed. I did feel only having 128K context vs the 200k/1mm
September 11, 2025 at 12:35 AM