


The UK government has completed its biggest test of AI coding tools, and the results show a clear pattern emerging across the tech industry. While developers are saving significant time using artificial intelligence assistants, they still don't fully trust what these tools produce.
The three-month trial run by the Government Digital Service tested AI coding assistants across 50+ government departments.
Over 2,500 licenses were distributed to civil servants, with 1,100 people actively using the tools. The findings mirror a global trend: developers want the speed boost AI provides, but they're spending extra time checking and fixing the output.
Results show government developers saved an average of 56 minutes per working day - equal to 28 full working days per year.
However, they only accepted 15.8% of the AI's code suggestions and just 39% actually used AI-generated code in their final work. This trust gap reflects worldwide surveys showing 84% of developers use AI tools, but 46% don't trust their accuracy.
This article will cover the trial's key findings, examine the global trust gap phenomenon, analyze implementation challenges, and explore what these results mean for organizations considering AI coding tools.
The government trial ran from November 2024 to February 2025, covering developers across dozens of departments. The time savings broke down into clear categories that show where AI helps most.
Code creation and analysis delivered the biggest benefit, saving 24 minutes daily per developer. This included 10 minutes saved when coding in familiar programming languages. Code review tasks saved another 21 minutes per day, while learning new skills saved 10 minutes.
Survey data shows 67% of users spent less time searching for information or examples, 65% completed tasks faster, and 56% solved problems more efficiently. These numbers suggest AI works best as a research assistant and starting point for code, rather than a complete solution.
The daily 56-minute saving adds up quickly. Over a standard work year, this equals 28 full working days per developer. For a government employing thousands of technical staff, this represents substantial productivity gains.
However, the raw time savings tell only part of the story. The way developers actually used the AI tools reveals a more complex relationship between human programmers and artificial intelligence.
Despite the time savings, government developers showed significant caution about AI-generated code. The average acceptance rate for GitHub Copilot suggestions was just 15.8%, meaning developers rejected more than 8 out of 10 suggestions from the AI.
This low acceptance rate aligns with global industry trends. Recent surveys show developers worldwide are becoming more skeptical of AI output quality. The 2025 Stack Overflow survey found 46% of developers don't trust AI accuracy, up from 31% the previous year.
The government trial data supports this cautious approach. While developers found the tools useful, they clearly see AI as needing human oversight rather than producing ready-to-use code. Only 39% of trial participants reported actually committing AI-suggested code to their projects.
This pattern reflects what industry experts call the "almost right" problem. AI often produces code that looks correct but contains subtle errors or doesn't fit the specific needs of a project. Developers have learned to treat AI suggestions as rough drafts requiring careful review.
The trust issue doesn't mean the tools are useless. Instead, it shows developers are learning to use AI effectively by combining its speed with human judgment. This approach may actually produce better results than blindly accepting AI output.
Interestingly, the trust gap doesn't translate to user dissatisfaction. The average satisfaction rating was 6.6 out of 10, and 58% of respondents said they wouldn't want to go back to working without AI assistance.
This seemingly contradictory result makes sense when you consider how developers actually use these tools. They're not looking for AI to replace their skills but to handle routine tasks and speed up research. Even if they reject most AI suggestions, the time saved on the accepted ones adds up.
72% of users agreed the tool offered good value for their organization, despite their cautious approach to trusting the output. This suggests developers see AI as a useful assistant rather than an autonomous programmer.
The satisfaction scores also varied based on how people used the tools. Developers who treated AI as a collaborative partner rather than a replacement seemed more satisfied with the results. This pattern appears in trials and studies worldwide.
User feedback highlighted specific benefits beyond just coding speed. Many reported enjoying their work more when routine tasks were automated. Others found AI helpful for learning new programming languages or exploring unfamiliar code bases.
The government trial revealed clear patterns about where AI coding assistants provide the biggest benefits. Understanding these patterns helps explain both the time savings and the trust issues.
AI performed best on routine, repetitive tasks. Writing boilerplate code, creating test cases that follow existing patterns, and generating documentation all saw significant time savings. These tasks have clear patterns that AI can follow reliably.
Learning and research activities also benefited strongly. Developers used AI to understand new programming languages, explore unfamiliar code bases, and get quick answers to technical questions. The tools served as an always-available reference that could explain code or suggest solutions.
Code review support proved valuable, with AI helping identify potential issues or suggesting improvements. However, developers consistently double-checked AI recommendations rather than accepting them automatically.
AI struggled more with complex, project-specific tasks. Custom business logic, security-sensitive code, and integration with existing systems required more human oversight. Developers learned to use AI for initial drafts but spent considerable time adapting the output to their specific needs.
The pattern shows AI works best as a smart autocomplete and research assistant rather than an independent programmer. This understanding helps explain why acceptance rates remain low even as satisfaction scores stay positive.
The government trial faced several practical challenges that other organizations should consider when adopting AI coding tools. These issues affected both adoption rates and user experience throughout the trial.
Marketing the trial required substantial effort, and success depended heavily on support from managers within each department. Some organizations were faster to adopt than others, creating inconsistent experiences across government departments.
Timing also created problems. The trial's middle month coincided with the December holiday period, reducing availability of both users and support staff. This timing issue affected the consistency of data collection and user engagement.
The temporary nature of the trial may have limited adoption. Since users knew their access would end after three months, some hesitated to fully integrate the tools into their daily workflow. This cautious approach likely affected both usage patterns and satisfaction scores.
License distribution proved challenging, with internal priorities delaying full rollout until the end of the first month in some departments. This meant some users had less time to learn and adapt to the tools than originally planned.
Technical integration varied across departments, with some organizations finding it easier to deploy the tools than others. Differences in existing development environments and security requirements created additional complexity.
The UK government findings fit within a broader global pattern of AI adoption in software development. Studies worldwide show similar results: meaningful productivity gains combined with persistent trust issues.
Recent research from Google found 21% productivity improvements in enterprise settings. Multi-company studies report average gains of 26%. These numbers align closely with the UK government's findings, suggesting consistent patterns across different organizations and contexts.
However, global studies also confirm the trust problem. Research consistently shows developers benefit from AI tools while remaining skeptical of their output. This cautious approach appears to be a feature rather than a bug - it helps maintain code quality while capturing productivity benefits.
The "almost right" problem appears in studies worldwide. Developers report that AI often produces code that looks correct but contains subtle errors or doesn't match project requirements. This issue explains why acceptance rates remain low even when satisfaction scores are positive.
Industry experts increasingly view the current state as healthy skepticism rather than a problem to solve. Developers who maintain critical evaluation of AI output are more likely to catch errors and maintain quality standards.
The global pattern suggests organizations should expect moderate but meaningful productivity gains rather than revolutionary changes. The most successful AI adoption strategies focus on augmenting human capabilities rather than replacing human judgment.
Government departments face unique security requirements that affected how developers approached AI coding tools during the trial. These considerations influenced both adoption patterns and trust levels throughout the study.
Security-sensitive code required extra scrutiny when AI tools were involved. Developers reported being more cautious about accepting AI suggestions for code that handled sensitive data or critical government functions. This appropriate caution likely contributed to lower acceptance rates.
Code quality standards in government work also affected AI usage patterns. The need for maintainable, well-documented code meant developers spent time reviewing and improving AI-generated output rather than using it directly.
Integration with existing government systems created additional complexity. AI tools often suggested modern coding patterns that didn't match older government systems, requiring developers to adapt the suggestions significantly.
Compliance requirements added another layer of consideration. Government code often needs to meet specific standards and regulations that AI tools don't automatically understand. This gap required human review and modification of AI suggestions.
Despite these challenges, many developers found ways to use AI tools while meeting security and quality requirements. The key was treating AI as a starting point rather than a final solution, which aligns with the overall pattern of supervised AI usage.
The trial's findings provide concrete data for evaluating the cost-effectiveness of AI coding tools in government settings. With clear time savings documented, organizations can make informed decisions about potential investments.
The 56 minutes of daily time savings per developer translates to roughly 12% of a standard 8-hour workday. Over a year, this represents significant value, especially when multiplied across large development teams.
72% of users agreed the tool offered good value for their organization, suggesting the benefits outweigh the costs from a user perspective. However, this calculation must include the overhead of training, deployment, and ongoing support.
The relatively low code acceptance rates don't necessarily reduce value. Even if developers only use 15.8% of AI suggestions directly, the time saved on research, learning, and initial drafts can justify the investment.
Organizations considering AI coding tools should also factor in indirect benefits. Improved job satisfaction, faster onboarding of new developers, and reduced time spent on routine tasks may provide value beyond direct productivity measures.
The government's approach of conducting a trial before full deployment provides a model for other organizations. This evidence-based approach helps ensure investments deliver expected returns while identifying potential issues before large-scale rollout.
The trial results have important implications for how government and other large organizations approach AI coding tool adoption. The findings suggest specific strategies for maximizing benefits while managing risks.
Training and support emerge as critical factors for successful adoption. Organizations should invest in helping developers learn effective AI collaboration techniques rather than simply providing tool access.
Process adaptation may be necessary to capture full benefits. Traditional code review processes may need adjustment to handle the increased volume of code that AI tools can generate.
Measuring success requires multiple metrics beyond simple productivity numbers. User satisfaction, code quality, and actual usage patterns provide a more complete picture than time savings alone.
The trust gap suggests organizations should embrace supervised AI usage rather than seeking autonomous solutions. Training developers to effectively evaluate and adapt AI output may be more valuable than trying to improve AI reliability.
Long-term adoption will likely require sustained access and organizational commitment. The trial's temporary nature may have limited full integration, suggesting permanent deployment would yield different results.
The UK government trial provides valuable lessons for other organizations considering AI coding assistant adoption. These insights can help guide implementation strategies and set appropriate expectations.
Start with realistic expectations focused on moderate productivity gains rather than revolutionary changes. The 20-30% improvements documented in various studies represent meaningful but incremental benefits.
Invest in change management and user training. Technical deployment alone isn't sufficient - developers need support to learn effective AI collaboration techniques.
Measure multiple dimensions of success. Time savings, user satisfaction, and actual usage patterns all provide important insights into AI tool effectiveness.
Plan for process adaptations. Traditional development workflows may need adjustment to accommodate AI-generated code and the review overhead it creates.
Consider security and compliance requirements from the start. Government and enterprise environments often have constraints that affect how AI tools can be used effectively.
Build internal expertise and support capabilities. Organizations need people who understand both the capabilities and limitations of AI coding tools to support effective adoption.
The UK government's AI coding assistant trial demonstrates the current state of artificial intelligence in software development: useful but requiring careful human oversight.
The 56 minutes of daily time savings represents meaningful productivity gains, yet developers' cautious approach - accepting only 15.8% of AI suggestions - shows they've learned to balance speed with quality.
This trust gap, where 84% use AI but 46% distrust its accuracy, reflects healthy skepticism rather than tool failure.
For organizations considering AI adoption, success lies in treating these tools as sophisticated assistants that augment human capabilities rather than replace human judgment.
The future points toward human-AI collaboration, with developers maintaining control while leveraging AI for routine tasks and research support.
Q1: How much time can developers actually save using AI coding assistants?
The UK government trial found developers saved 56 minutes daily, equaling 28 full working days annually. Studies worldwide typically show 20-30% productivity improvements on specific coding tasks.
Q2: Why do developers use AI tools if they don't trust the output?
Developers treat AI as a collaborative assistant, not an autonomous programmer. Even rejecting 80% of suggestions, the accepted ones save enough time to make tools valuable.
Q3: Are AI coding assistants worth the cost for organizations?
72% of UK trial users believed the tools offered good value. Organizations should expect moderate productivity gains rather than revolutionary changes, factoring in training and support costs.
Techdots has helped 15+ founders transform their visions into market-ready AI products. Each started exactly where you are now - with an idea and the courage to act on it.
Techdots: Where Founder Vision Meets AI Reality
Book Meeting