Transforming Developer Platforms with AI

Published by

on

AI is rapidly transforming developer platforms from static utilities into intelligent systems.

Internal developer platforms, such as GitLab, Jira, Jenkins, Nexus, Backstage, Vault, Grafana and Confluence have always operated behind the scenes.

Considered plumbing, these tools are viewed as necessary, but often unsexy. AI is about to change that, with the under-the-radar transformation of these systems in a way that will fundamentally shift software delivery in the large enterprise.

Here’s how…

AI: The New Infrastructure Platform

It’s rare to find a large enterprise with no developer tools in place. For the last decade or more, platforms have provided the key services to application developers internally, e.g. GitLab, Jira, Confluence, CI/CD pipelines, and API documentation sites.

They’ve always served as an infrastructure layer or as a utility to the application teams, but have never been a focus for innovative use cases.

That’s now beginning to change. AI has not yet created a new Uber or Amazon in the software delivery space, but it is slowly and quietly transforming the invisible plumbing of big company software development.

In place of a flood of new features with the usual fanfare, there’s a wave of integrations that are getting smarter over time. The result is a sea change in how teams build and ship software in large enterprises.

Developers spend up to 40% of their time on tasks unrelated to core coding, often fighting CI/CD, infrastructure, or security tools.
— Source: Stripe Developer Productivity Report


Automation: Before and after AI

AI use cases in the software delivery context almost always build on a bedrock of automation. For decades, DevOps teams have automated anything they could to minimise manual toil, ranging from cron jobs on Jenkins, YAML-based build scripts, static code checks, and more.

The problem was that most of these automations were brittle and rule-based, rapidly falling out of sync with the code and business processes they supported. They also, by their nature, required human intervention to get them into compliance and keep them there.

The difference now is that automation can be infused with intelligence, which can both recommend actions and optimise workflows.

In a CI/CD troubleshooting context, this means AIs that look at hundreds of thousands of lines of build logs and narrow down the actual root cause of a failure, not just the symptoms. Problem Records that took human hours of review can be summarised in seconds by an LLM.

Rather than relying on developer knowledge (SMEs), dev teams are also experimenting with AI change risk prediction tools that automatically apply context-aware analysis to code changes (commits), suggesting recommended actions to development teams.

Companies that adopted AI-based coding assistants saw a 55% faster coding task completion rate for junior developers.
— Source: McKinsey AI Developer Productivity Study, 2023

These AI tools utilise machine learning models trained on historical data, including previous commits, test outcomes, incident logs, deployment metrics, and even code ownership patterns, to assess the likelihood that a given change will cause production issues.


Developer experience: An area of increasing focus

In a large enterprise, finding the right documentation when dealing with a large platform (knowledge spread across multiple Confluence pages) is a day-to-day issue.

There are instances where multiple platforms are used to maintain the knowledge base, ranging from Confluence to SharePoint to Backstage. Application teams (users) often get confused about where to look for the right documentation about a platform.

With the rise of enterprise-trained AI copilots, developers can now interact with complex internal systems using simple, conversational prompts. Asking an enterprise-trained natural language interface, on the other hand, is a trivial proposition and is quickly being enabled by the rise of developer-facing copilots.

Instead of spending 20 minutes hunting down the right Helm override or Terraform module syntax, they can simply ask:

“Can you show me how to add a timeout to the staging job in this CI pipeline?”
Or:
“Why did the last deployment to service-X fail in production?”

Copilots can answer such questions contextually from the right source of truth, be it Git logs, Vault configuration, pipeline history, and more.

In short, this intelligence is being layered on top of the developer experience in ways that make the entire platform operate as a resourceful teammate.


Safety and compliance: Areas for even more improvement

AI is not just about enhancing efficiency; it’s also about improving safety, security, and compliance. These have always been important but often cumbersome aspects of developer work.

Historically, enforcing good practices has been a painful afterthought, with hard blocks or manual audits applied after code changes. But AI is beginning to enable more intelligent, pre-emptive forms of this work, providing a new level of reassurance for developers.

For Example:

An AI can detect a hardcoded JWT token and alert us with a message.

# login_service.py

def get_jwt_token():
jwt_token = "eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9..." # hardcoded JWT token
return jwt_token

Security Alert: This appears to be a hardcoded JWT token.
Recommendation: Move it to HashiCorp Vault (or any secrets manager) and retrieve it via a secure environment variable or runtime secrets API.

Traditionally, enforcing enterprise development standards for security, compliance, or infrastructure hygiene has relied on static linters, manual code reviews, and pre-commit hooks.

These tools often apply blanket rules without understanding the context or intent behind a developer’s code.

As a result, developers are either flooded with false positives or forced to memorise complex policies, slowing down productivity.

But with the introduction of AI copilots like GitHub Copilot Enterprise and GitLab Duo, we’re seeing the rise of intelligent, policy-aware development environments that can do far more than autocomplete code.


Some concerns and caveats

This type of work is not without risks. Poorly tuned AI agents, especially in low-latency contexts, can be quite noisy, which actually reduces clarity rather than improves it.

For example, an AI tool that constantly suggests irrelevant code completions, misinterprets infrastructure-as-code patterns, or repeatedly flags safe secrets as risky can overwhelm developers with false positives and low-quality recommendations.

Instead of boosting confidence, it introduces decision fatigue. Teams may start to ignore warnings altogether, eroding trust in the system; a dangerous outcome in security-sensitive environments.

For Example:

Snippet from .gitlab-ci.yml file.

Stages:
- test
- deploy

test_job:
stage: test
Script:
- pytest

It may hallucinate and give the “Missing required artefacts section for job output” error. While not all jobs require artefacts, this job only runs tests and doesn’t need to pass outputs.

Besides this, Enterprises have a huge amount of intellectual property in their logs, tickets, internal codebases and other artefacts, which they cannot compromise by training AI agents poorly.

There is also a question of scale: once models are in production, fine-tuning them on newer data sets is a challenging prospect in traditional LLMs.

So, balancing the usefulness with the responsibility of these tools is essential, and enterprises must design systems with the same thoughtfulness as the rest of the platform.


The future in sight

92% of developers using GitHub Copilot report that it helps them focus on more meaningful work, and 88% feel more productive.
– Source: GitHub Copilot Research 2023

Agentic AI is an inflexion point for internal developer platforms.

The platform is being reinvented as proactive, context-aware and sense-making systems that can interpret intent, take initiative and adapt dynamically in real time.

We are on the edge of developer platforms that no longer just react to inputs. Platforms that autonomously orchestrate the end-to-end software delivery workflow, generating pipelines from high-level objectives, summarising incidents with full-stack awareness and applying security and compliance controls that learn from team practices.

This is not science fiction; agentic AI is already emerging in tools like GitHub Copilot Enterprise, GitLab Duo and in a new wave of internal copilots trained on company proprietary models.

In a world of agentic systems, internal platforms are no longer simply enablers; they’re becoming copilots in their own right, helping to shape how software is built, how teams collaborate and how businesses ship value at scale.

The infrastructure is no longer just code. It’s starting to think.

Leave a Reply

Discover more from Products, People, and Processes: My Perspective

Subscribe now to keep reading and get access to the full archive.

Continue reading