The State of AI Coding
A cross-industry study on recent trends in AI software development.
Q1 2026 update.
Navigate the Report
Engineering Team Velocity
Measuring productivity gains across development workflows.
Median PR size increased 93% from March 2025 to March 2026, rising from 57 to 110 lines changed per PR.
Captured from Greptile internal data engineering team velocity
Lines of code per developer grew from 4,450 to 14,148 as AI coding tools act as a force multiplier.
Captured from Greptile internal data engineering team velocity
Medium teams (6-15 devs) increased output from 7,005 to 19,715 lines per developer.
Captured from Greptile internal data engineering team velocity
Median lines changed per file grew from 18 to 25 as PRs become denser.
Captured from Greptile internal data engineering team velocity
AI Tool Adoption
Tracking the rise of AI-powered development tools.
mem0 holds 58% market share. Zep dropped 9pp as smaller players gained ground.
PyPI + npm monthly downloads, Mar 2026
Weaviate extended its lead to 33% (+8pp). The remaining 7 converged between 5-11%.
PyPI + npm monthly downloads, Mar 2026
CLAUDE.md present in 75% of orgs. AGENTS.md overtook Cursor Rules, down 18pp.
Repos using all three formats dropped from 17% to 6% as teams standardize on fewer formats
Anthropic SDK reached 124M downloads in March 2026. OpenAI Agents grew 5x in Q1 2026 to 21M.
PyPI + npm monthly downloads, Apr 2025 – Mar 2026
LiteLLM overtook LangSmith in Q1 2026, hitting 98M monthly downloads.
PyPI + npm monthly downloads, Jun 2025 – Mar 2026
LangSmith is bundled with LangChain installs
Model Growth Trends
How AI models have scaled and evolved.
OpenAI hit 233M in Mar 2026. Anthropic surged to 83M. Google trails at 17M.
PyPI monthly downloads, Jan 2022 – Mar 2026
OpenAI-to-Anthropic ratio dropped from 3.7:1 (Dec 2025) to 2.8:1 (Mar 2026).
PyPI monthly downloads ratio, Jul 2023 – Mar 2026
Research & Content
Surfacing recent research that shaped how 2026 tools handle compression, context, multimodality, and long-horizon agents, so teams can interpret and apply it to their own systems.
Foundational Model Advances
TurboQuant: Online Vector Quantization with Near-optimal Distortion Rate
TurboQuant is an online, data-oblivious quantization method for KV-cache compression and vector search that targets both mean-squared error and inner-product distortion.
Recursive Language Models
RLMs treat the prompt as external state, letting the model inspect, partition, and recursively call itself over snippets instead of feeding everything through one context window.
Titans: Learning to Memorize at Test Time
Titans is a family of long-context architectures that pairs limited-window attention with a neural long-term memory that keeps learning at test time.
Kimi K2.5: Visual Agentic Intelligence
Kimi K2.5 is an open-source multimodal agentic model that jointly optimizes text and vision, with a parallel-agent execution framework layered on top.
Does RAG Really Perform Bad for Long Context?
RetroLM introduces KV-level retrieval for long-context tasks, treating the KV cache as the retrieval surface instead of raw text.
Rethinking Mixture-of-Agents
Self-MoA examines whether diverse model ensembles are actually necessary for strong Mixture-of-Agents performance.
Application-Layer Innovations
Chroma Context-1: Training a Self-Editing Search Agent
Context-1 is a 20B agentic search model derived from gpt-oss-20B that is designed to act as a retrieval subagent rather than answer the question directly.
Composer 2 Technical Report
Composer 2 is a domain-specialized model for agentic software engineering, built to handle long-horizon coding tasks while staying efficient enough for interactive use.
GEPA: Reflective Prompt Evolution Can Outperform RL
GEPA (Genetic-Pareto) is a reflective prompt-evolution method that optimizes instructions using execution traces instead of updating model weights.
SFR-DeepResearch: Single-Agent RL for Deep Web Research
SFR-DeepResearch (SFR-DR) is a reinforcement-learning framework for training a single web-research agent that decides when to search, browse, or execute code.
MEM1: Constant-Memory Long-Horizon Agents
MEM1 is an RL framework that trains LLM agents to operate over long multi-turn tasks while keeping memory usage nearly constant.
Search-R1: Training LLMs to Reason and Search with RL
Search-R1 trains models to interleave step-by-step reasoning with live search-engine queries.