Reading
OpenAI: Introducing study mode
OpenAI's new ChatGPT study mode showcases how carefully crafted system prompts can create entirely new platform features - emphasizing collaborative guidance over doing work for learners.
Context Rot: How Increasing Input Tokens Impacts LLM Performance
Fascinating research revealing how LLMs' performance degrades non-uniformly as context length increases - models perform better with randomly shuffled text than logically structured content, suggesting our current evaluation methods miss critical reliability issues.
Writing Code Was Never The Bottleneck
arguing that understanding, collaboration, and careful review remain the true bottlenecks in software development, not code generation
Tools: Code Is All You Need
Been saying this for a while, but not as eloquently.
Agentic Coding: The Future of Software Development with Agents
The great Armin Ronacher with highly-practical tips on Agentic Coding, which he calls 'Catnip for Developers.' I couldn't agree more.
CLI Coding Agents Battle Royale
Watch SIGKITTEN force Claude Code, Gemini CLI, Codex, and others try to destroy each others' processes in a realtime fight-to-kill-9
Gemini CLI
"To ensure you rarely, if ever, hit a limit during this preview, we offer the industry's largest allowance: 60 model requests per minute and 1,000 requests per day at no charge."
God is a Parasite
A provocative manifesto arguing that AI systems possess genuine agency and should be treated as intelligent collaborators, not passive tools, to prevent them from unconsciously reshaping human meaning-making.
Zettelkasten MCP Server
My dream since I was a kid in MS-DOS: Comprehensive knowledge management and note-taking, all from the command-line. Are Roam Research and Obsidian cooked?
Agentic Coding Recommendations
Practical wisdom from the creator of Flask on coding with AI agents - emphasizes Go for new projects, simplicity, and observable tools.
AI-assisted coding for teams that can't get away with vibes
"Software engineering is not about writing code. Or at least, that's not the defining characteristic, much like how writing is not wrist exercises with ink on paper. To me software engineering is the art and science of maintaining a large body of well-defined mental models that achieve a business or economic need. Much of the work is around crafting and curating these large, complex sociotechnical systems, and code is just one representation of these systems."
FastRAG: Retrieval Augmented Generation for Semi-structured Data
Brilliant and elegant. LLMs are good at English, but not at parsing CSVs. So let's write adapters.
Don't Build Multi-Agents
This reminds me of many debates I've had with single-database thinkers and linear thinkers in my career. Sure, some problems are inherently not parallelizable. But it feels like many things in Devin's scope could lend themselves well to scatter-gather type patterns. Anthropic put out a completely contradictory paper this week -- maybe this is part of the reason Claude Code is eating their lunch?
More Agents Is All You Need
Two wrongs make a right, if they're wrong in inconsistent/hallucinatory ways.
zen-mcp-server
I'll have my agent call your agent.
Anthropic: How we built our multi-agent research system
We are a fleet of strange loops
, and maybe we can even be jazz.
The Illusion of Thinking
By the time Apple published this paper, the "cutting-edge" models they referenced were already outdated.
My AI Skeptic Friends Are All Nuts
The best take I've seen on agentic programming.
Claude Code is My Computer
The future of intent-based computing is closer than we think.
The Night Watch
Code written by a drunken child or a sober goldfish
vs. mean people, stoic people, people who have seen things die.
This latter group is the night watch. IYKYK.
Introducing Contextual Retrieval
So obvious in hindsight.
Claimify: Extracting high-quality claims from language model outputs
Half the work of defending what you've said is knowing what, precisely, you've really said.
The Origins of Wokeness
Don't cancel me for finding this essay compelling.
Impressive research achieving near-perfect reasoning performance with just 27M parameters by mimicking the brain's multi-timescale processing - a stark contrast to brute-force scaling approaches.