Your AI Coding Assistant is Incomplete: Why DevX Needs a New Class of Data
- Mark Rose
- Sep 25
- 5 min read
Updated: Sep 26

AI-powered coding assistants have become an indispensable part of the modern developer's toolkit. In just a few years, tools like GitHub Copilot have fundamentally changed the mechanics of writing code, with surveys showing that over 90% of developers are already using them to boost productivity. They are remarkably good at completing lines of code, generating boilerplate, and even writing unit tests. But they are also fundamentally incomplete.
The current generation of AI assistants has been trained primarily on a diet of public code repositories. This has made them experts in syntax and patterns—the "what" of software development. However, it leaves them blind to the most critical, and ofte
n most painful, part of a developer's day: the "why." Why is this API so confusing? Why does this build process keep failing? Why can't I find the right information in the documentation?
These are not coding problems; they are experience problems. Despite the rise of AI assistance, developer frustration remains rampant, with high rates of burnout and job dissatisfaction.
These are not coding problems; they are experience problems. Despite the rise of AI assistance, developer frustration remains rampant, with high rates of burnout and job dissatisfaction. The truth is, the biggest drags on developer productivity aren't about typing code faster. They are about the friction in the surrounding ecosystem: the confusing toolchains, the opaque error messages, and the soul-crushing hunt for context. To solve these problems, AI needs a new kind of fuel—one that goes beyond code and captures the nuanced, qualitative reality of the developer experience (DevX).
The Rise of On-Device AI and the DevX Data Gap
The next frontier of AI-powered developer tools won't be found in ever-larger, cloud-based Large Language Models (LLMs). Instead, the innovation is happening at the edge, with Small Language Models (SLMs). These are highly efficient, domain-specific models designed to run directly on a developer's local machine.
The advantages of on-device SLMs for developers are immense:
Privacy and Security: Proprietary code and internal workflow data never leave the local machine, a critical requirement for enterprise security.
Low Latency: On-device processing provides instantaneous responses, keeping developers in a state of flow without network delays.
Offline Capability: The tools work anywhere, without requiring a constant internet connection.
However, these powerful SLMs are only as good as the data they are trained on. Fine-tuning a generic model on yet another code repository won't solve the core DevX challenges. To build a truly intelligent assistant, we need to train it on data that reflects the actual lived experience of developers.
This unique archive doesn't just contain code; it contains the context, the confusion, the "aha!" moments, and the subtle signals of frustration that define the real-world developer journey.
This is the DevX data gap. For nearly two decades, we at ConcreteUX have been systematically capturing this exact data, amassing thousands of hours of deep, qualitative research observing developers as they interact with SDKs, APIs, documentation, and complex toolchains. This unique archive doesn't just contain code; it contains the context, the confusion, the "aha!" moments, and the subtle signals of frustration that define the real-world developer journey. This is the data needed to build the next generation of AI developer tools.
Three Use Cases for AI That Truly Understands Developers
By fine-tuning specialized SLMs on this rich, qualitative DevX data, we can create a new class of AI assistants that go beyond code generation to actively improve the entire development workflow.
1. The On-Device, Context-Aware Workflow Assistant
The problem isn't always the code; it's getting to the code. Developers lose countless hours wrestling with environment setups, obscure command-line tools, and complex build processes. An SLM fine-tuned on observational data of these struggles becomes an expert platform engineer on your shoulder.
How It Works: Running locally within an IDE, this assistant understands the developer's workflow, not just their code. When a developer encounters a common environmental error, the SLM—trained on countless examples of others solving the same problem—can provide just-in-time guidance on the right command to run or the correct configuration file to edit.
Value Statement: This transforms the AI assistant from a code completer into a workflow accelerator. It dramatically reduces the "time-to-code" by solving the frustrating setup and tooling issues that block progress, allowing developers to stay focused on creative problem-solving.
2. The Proactive, Intelligent Documentation Engine
Poor documentation is a universal source of developer pain. An SLM trained on how developers actually use (and fail to use) documentation can turn a static library of information into a dynamic, intelligent partner.
How It Works: This on-device SLM is trained on a product's full documentation corpus and then fine-tuned with qualitative data showing where developers get confused, what they search for, and which explanations are ineffective. It can then power an intelligent search within the IDE that understands natural language queries and anticipates developer needs, proactively surfacing the right tutorial or API example at the moment it's needed.
Value Statement: This eliminates the costly context-switching of hunting for answers. It accelerates the onboarding of new developers and creates a data-driven feedback loop for technical writers, leading to a continuous improvement cycle for documentation quality and a massive boost in developer productivity.
3. The On-Device DevX Friction Detector
The most subtle and corrosive aspects of poor DevX are the small, repeated moments of friction that lead to cognitive load and eventual burnout. A multimodal AI model trained on the full spectrum of DevX research data—video, audio, and interaction patterns—can learn to identify these moments in real-time.
How It Works: A highly efficient SLM, running privately on-device, can detect the tell-tale signals of developer frustration: hesitations, erratic mouse movements, repeated compiler errors, and even audible sighs. It can then provide real-time, private feedback to the developer (e.g., suggesting a break) or generate anonymized "friction scores" for internal tools.
Value Statement: This provides a data-driven "health check" for the entire developer ecosystem. For the first time, organizations can get objective, quantifiable metrics on the tools and processes that cause the most friction, enabling targeted investments that improve developer retention, code quality, and create a more sustainable engineering culture.
The Future of Developer AI is Human-Centric
The next evolution of AI for developers will not be measured in the number of parameters or the volume of code it can generate. It will be measured by its ability to reduce friction, lower cognitive load, and create a more productive and satisfying developer experience. This requires a fundamental shift in how we train our models—moving from a code-first to a human-first approach.
Building these next-generation tools requires a unique and irreplaceable asset: a deep, longitudinal archive of human developer behavior. At ConcreteUX, we've spent nearly 20 years building that asset. We are now poised to fuel the intelligence supply chain with the real-world data essential for creating AI systems that don't just write code, but truly understand developers.

