Seems way more likely than an LLM is regurgitating close-to-verbatim content from its training set.
Seems way more likely than an LLM is regurgitating close-to-verbatim content from its training set.
Number one "complaint" I keep hearing about gridfinity - once you start you really want to do All The Things, and the print times are bruuuuuttal. Someone should sell injection molded snap-fit grids.
Number one "complaint" I keep hearing about gridfinity - once you start you really want to do All The Things, and the print times are bruuuuuttal. Someone should sell injection molded snap-fit grids.
Claude Code also uses this trick for it's built in tools (e.g. WebFetch)
lots more detail in the full post: blog.thepete.net/blog/2025/12...
Claude Code also uses this trick for it's built in tools (e.g. WebFetch)
lots more detail in the full post: blog.thepete.net/blog/2025/12...
- different system prompt
- different tools
- (hidden!) context management tricks to steer the model’s behavior
- task management tools to nudge the LLM into better planning
- sub-agents, which can have a *huge* impact on context management
🧵...
- different system prompt
- different tools
- (hidden!) context management tricks to steer the model’s behavior
- task management tools to nudge the LLM into better planning
- sub-agents, which can have a *huge* impact on context management
🧵...
🧵...
🧵...
- your agent can burn through it fast
- your LLM's performance starts taking a hit way before that "limit" is reached
further reading:
- blog.thepete.net/blog/2025/10...
- research.trychroma.com/context-rot
- github.com/adobe-resear...
- arxiv.org/abs/2404.06654
- your agent can burn through it fast
- your LLM's performance starts taking a hit way before that "limit" is reached
further reading:
- blog.thepete.net/blog/2025/10...
- research.trychroma.com/context-rot
- github.com/adobe-resear...
- arxiv.org/abs/2404.06654
Particularly when coupled with the increased overall volume of PRs that AI-assisted coding is producing... 😬
Particularly when coupled with the increased overall volume of PRs that AI-assisted coding is producing... 😬
No surprise: The LLM does a *great* impression of the huge pile of human code review examples in it's training set!
No surprise: The LLM does a *great* impression of the huge pile of human code review examples in it's training set!
Think of it as a super-powerful static analysis step in your CD pipeline.
Think of it as a super-powerful static analysis step in your CD pipeline.
- the LLM won't have the full context of the codebase, and
- an LLM only has mediocre software design skills (at best)
you're not going to get any sort of full critique of your proposed solution from an AI-powered tool.
- the LLM won't have the full context of the codebase, and
- an LLM only has mediocre software design skills (at best)
you're not going to get any sort of full critique of your proposed solution from an AI-powered tool.
Or if you can spot the backdoor being added: github.com/trailofbits/...
More on the lethal trifecta: simonwillison.net/2025/Jun/16/...
Or if you can spot the backdoor being added: github.com/trailofbits/...
More on the lethal trifecta: simonwillison.net/2025/Jun/16/...
point C is exactly the same guidance that I was giving 11 years ago(!!!): www.thoughtworks.com/insights/blo...
I was ahead of the curve I guess :D
point C is exactly the same guidance that I was giving 11 years ago(!!!): www.thoughtworks.com/insights/blo...
I was ahead of the curve I guess :D