Back to articles

Deep Signal Quarterly – Q1 2023

14th April 2023
By Steve Kilpatrick
Founder & Director
ML Systems and Infrastructure

474 Tracked Repos  |  52,249 Commits  |  5,717 Contributors

THE SIGNAL


The LLM gold rush arrived in open source this quarter, but it landed where most hiring teams aren’t looking. The unforgettable number: model optimization attracted 60% more contributors while churn spiked 461%. The people building the techniques that will define cost-per-token economics for the next two years are experimenting broadly and committing to nothing. Mean engagement sits at 3.0 weeks, the lowest of any category we track. If you want quantisation engineers, the sourcing window is open now. It won’t stay open.

Underneath the hype, a structural rotation is underway. Every category feeding inference surged while training frameworks, the largest category by raw volume, managed just 6%. The two biggest ML frameworks quietly became compiler projects: PyTorch’s commits concentrated on graph lowering and compilation passes, TensorFlow’s heaviest work on compiler backend consolidation. Our Q4 2022 baseline flagged this rotation. That call landed correct, and faster than we projected.

This edition tracks 5,717 contributors across 474 tracked repositories and 52,249 commits.

Q-over-Q Snapshot


Inference-adjacent categories surged while training-side work held steady; the table captures an industry rotating from “build the model” to “run it cheaply.”

Category Commits Contributors Active Repos Commits QoQ Contribs QoQ
GPU Kernels & Performance 3,104 438 24 +4% +50%
ML Compilers & Graph Optimization 7,216 622 21 +29% +28%
Distributed Training & Parallelism 1,502 293 16 +4% +30%
Inference Runtimes & Engines 2,285 319 18 +24% -2%
LLM Serving & Inference 1,911 303 16 +22% +59%
Training Frameworks & Model Architecture 16,346 2,155 45 +6% +8%
ML Platform & Orchestration 6,920 790 26 +4% -2%
Edge & On-Device ML 3,845 348 22 +64% +51%
Model Optimization & Compression 1,333 202 11 +51% +60%
Hardware-Software Co-Design 5,552 594 18 +12% +22%
ML Debugging & Tooling 1,268 135 12 +2% +4%
Agent Framework 967 289 3 +176% +445%
Top Projects by Contributor Count

Whatโ€™s Moving


๐Ÿงช Training Frameworks & Model Architecture

The largest category tells a story of transformation, not growth. PyTorchโ€™s engineering has migrated from training loops to dynamic shape handling and compilation passes. TensorFlow followed a parallel path through compiler backend consolidation. Both of the two largest ML frameworks are, in practice, becoming compiler projects. Downstream, Hugging Faceโ€™s libraries absorbed the model-architecture experimentation that previously happened inside core framework repos, with an unusually collaborative contributor culture for projects of their age.

The 2,155 contributors here are not a monolithic pool. Roughly a third now do work indistinguishable from compiler engineering; another third handle model integration; the remainder maintain training infrastructure. A job description that says โ€œPyTorch experience requiredโ€ without specifying which layer will attract the wrong candidates.

๐Ÿง  ML Compilers & Graph Optimization

New contributors are arriving and staying. Churn dropped 61% even as contributor count grew 28%. Test infrastructure across the categoryโ€™s top projects has matured to match production standards: this is no longer the exploratory compiler phase of 2022. The engineering is increasingly team-based, with deep coordination patterns across major projects confirming that productive compiler work requires institutional context. Hiring a lone compiler engineer and expecting standalone output will not work in these codebases.

A hardware vendorโ€™s new compiler project appeared this quarter, concentrated almost entirely in core runtime code with intensive benchmark coverage. The combination of new entrants and deepening engagement makes this the most structurally healthy talent pool in ML systems.

๐Ÿš€ LLM Serving & Inference

Production serving is forming a talent pool, but the depth isnโ€™t there yet. Mean engagement sits at 2.4 weeks: the lowest of any category. Engineers are touching serving code; very few are building the runtime internals that production systems need. The engineering splits between lightweight model-adapter integration and dense, low-level scheduling and memory management. The latter group is tiny.

This is early-stage talent formation. The projects hardening for real deployment are still small. By Q2, watch whether any serving project can retain contributors beyond a single quarter; that will determine whether a durable talent pool forms or the category stays shallow.

๐Ÿค– Agent Framework

The agent hype produced volume, not depth. The engineering across this category is overwhelmingly integration glue and documentation: connector modules, example notebooks, retrieval wrappers. Very little touches core orchestration internals. LangChain pulled 181 net-new contributors in a single quarter, but the work theyโ€™re doing wonโ€™t produce the systems engineers hiring teams actually need.

The people building integrations are application developers, not systems engineers. If you need someone who can architect reliable tool-use pipelines with state management and observability, the OSS contributor pool wonโ€™t help. Youโ€™re hiring from adjacent categories and retraining.

๐ŸŒ Edge & On-Device ML

On-device inference went from niche to strategic in a single quarter. The nature of the work shifted from mobile vision pipelines to LLM-on-device deployment. Hardware-optimised operator libraries posted hundreds of commits focused on operator coverage and runtime reliability, with engineering concentrated almost entirely in core code rather than integration layers. The contributor profile for โ€œedge ML engineerโ€ is bifurcating: one track optimises traditional vision workloads for mobile silicon, the other builds inference stacks for parameter-dense models on consumer hardware. These are different people with different skills, recruited under the same job title.

Quiet Corners


GPU Kernels saw a 50% contributor jump on just 4% more commits: an onboarding wave that hasnโ€™t yet translated to sustained output, with engineering focus on low-precision arithmetic and attention mechanisms. Hardware Co-Design grew steadily; Intelโ€™s LLVM fork and compute runtime absorbed most new arrivals, and contributors here show the deepest sustained engagement of any hardware-adjacent category. Model Optimization attracted experimenters, not builders: most contributors touched quantisation repos once and moved on.

Distributed Training added 30% more contributors on a flat commit base, pointing to onboarding rather than output. Inference Runtimes consolidated quietly, with the existing base deepening around hardware backend integration at the highest mean engagement in the report. ML Platform held steady with the longest contributor tenure of any category; this talent is locked up. ML Debugging barely moved.

Where Talent Is Moving


The strongest overlap sits between ML compilers and training frameworks: 191 engineers contributed to both. Nearly a third of the compiler categoryโ€™s contributor base also touches training framework repos. For hiring, โ€œcompiler engineerโ€ and โ€œframework engineerโ€ are converging into a single talent pool at the senior level. Sourcing them separately wastes effort.

A new overlap emerged between LLM serving and edge (72 shared contributors), driven by quantisation-adjacent projects. These contributors understand both cache scheduling and memory-mapped weight loading. Sixty-six engineers bridge GPU kernels and hardware co-design, concentrated around vendor-specific silicon stacks. The contributor-level migration data here tells a more granular story; one weโ€™re making available to a small number of hiring teams directly.

Talent Migration: Contributor Overlap Between Categories

What This Means If Youโ€™re Hiring


Compiler engineers are the scarcest and most strategically valuable profile in ML infrastructure right now. The 622 contributors in the compiler category have declining churn and team-based coordination patterns that make individual poaching difficult. ML compiler roles benchmark at $250K to $450K+ total comp, with a 30-50% specialist premium over generalist ML engineering. Even at those numbers, the addressable pool doing active compiler work globally is under 700.

Quantisation and model efficiency engineers present a different challenge: high availability, low depth. Over 200 contributors touched optimization repos this quarter, but most are experimenting. The handful doing kernel-level quantisation work are embedded in specialised projects. Sourcing from this group requires distinguishing between someone who fine-tuned an adapter once and someone who implemented a custom quantisation kernel.

Edge inference engineers who understand both traditional on-device optimization and LLM-specific techniques are the emerging unicorn profile. The overlap data shows 72 people working across both serving and edge categories. Thatโ€™s the total pool demonstrating both skill sets in public code. If youโ€™re building on-device LLM products, start sourcing now; by Q3, every foundation model company will be competing for these same people.

If any of these signals match what youโ€™re seeing in your own pipeline, the conversation is worth having sooner rather than later.

Predictions


  • Q2 2023: LLM serving contributor counts will overtake distributed training for the first time as production serving challenges absorb engineering attention from training scale-out.
  • Q3 2023: Model optimization churn will normalise sharply as the experimentation wave passes. Mean engagement will climb from 3.0 toward 6.0 weeks as the tourist phase ends. The contributors who remain will be the ones worth hiring.
  • Year-end 2023: At least one major framework will formally split contributor onboarding into compiler-track and application-track pathways, acknowledging the bifurcation visible in the commit data.

Six months from now, the cost of ignoring this rotation will show up in your pipeline. The talent is moving. Your job descriptions should be moving with it.


This report is powered by D33P S1GNL: a proprietary contributor intelligence engine. For access to the full contributor-level dataset or to discuss ML Systems hiring, contact [email protected]

Share Article

Get our latest articles and insight straight to your inbox

Hiring Machine Learning Talent?

We engageย exceptional humans for companies powered by AI

Upload your CV
One of our consultants will contact you
to find out more about you