If you ask most engineering leaders about AI in their development process, they’ll mention GitHub Copilot or Cursor and tell you it’s making their developers 20-30% faster at writing code. That’s real, but it’s also the least interesting part of the story.

Writing code was never the bottleneck. Understanding requirements, designing systems, reviewing code, writing tests, debugging production issues, and keeping documentation current — that’s where engineering time actually goes. And AI is starting to be genuinely useful in several of those areas.

Where AI Creates Real Value Today

Code review and security scanning. Human code reviewers get fatigued. They focus on logic and miss security vulnerabilities, especially in pull requests that touch authentication, authorization, or data handling. AI-powered code review tools (like GitHub’s Copilot for security scanning, Snyk Code, or Semgrep) catch classes of vulnerabilities that human reviewers systematically miss: SQL injection patterns, insecure deserialization, hardcoded secrets, and OWASP Top 10 issues.

At a recent client engagement, we implemented AI-assisted security scanning in the CI pipeline and caught 14 medium-to-high severity vulnerabilities in the first month — vulnerabilities that had passed human code review. That’s not replacing the human reviewer. It’s giving them a second set of eyes that never gets tired and never forgets the OWASP checklist.

Test generation. Writing tests is the thing developers skip when they’re under pressure. AI test generation tools can look at a function signature, its implementation, and its call sites, and generate meaningful test cases covering happy path, edge cases, and error conditions. The tests aren’t perfect — you’ll edit maybe 40% of them — but they dramatically lower the activation energy for achieving good test coverage.

I’ve been recommending clients use AI test generation specifically for inherited codebases where no tests exist. Getting from 0% to 60% coverage on a legacy codebase used to take months. With AI-assisted generation and human review, it takes weeks.

Incident response. When production goes down at 3am, the on-call engineer is reading through thousands of log lines trying to find the needle in the haystack. AI-powered log analysis (through tools built into Datadog, New Relic, or standalone solutions) can correlate logs across services, identify anomalous patterns, and suggest root causes in minutes rather than hours. This doesn’t replace the engineer’s judgment — it gives them a starting point instead of a blank screen.

Documentation. The perennial problem: code changes, documentation doesn’t. AI tools that generate documentation from code changes, update API docs when endpoints change, and flag documentation that’s drifted from the implementation are solving a problem that pure discipline never could. Not because engineers are undisciplined — because there are always more important things to do than update the README.

Where AI Is Overhyped (For Now)

Fully autonomous coding agents. The promise of “describe what you want and AI builds it” is compelling and partially real for simple, well-defined tasks. For anything involving architectural judgment, understanding existing system constraints, or navigating ambiguous requirements, current AI coding agents produce code that looks right but makes subtle, expensive mistakes. Use them for bounded tasks with clear specifications and human review. Don’t hand them open-ended features.

AI-generated architecture decisions. “Should we use microservices?” is not a question AI can answer well because the answer depends on your team’s capabilities, your deployment infrastructure, your growth trajectory, and a dozen other context-specific factors that AI models don’t have access to. Architecture is judgment, not pattern matching.

Replacing senior engineers. The companies I’ve seen try to use AI as a substitute for senior engineering talent — rather than as a force multiplier for existing talent — consistently produce systems that are harder to maintain, have more subtle bugs, and accumulate technical debt faster. AI makes good engineers more productive. It doesn’t make junior engineers senior.

The Governance Question

Here’s the part most companies skip: when AI writes or modifies code, who’s responsible for it? Your AI-generated code has the same security, compliance, and quality requirements as human-written code. But the review patterns are different.

I recommend treating AI-generated code like code from a new junior developer: it always gets reviewed, it always gets tested, and it never goes to production without a human explicitly approving it. Build this into your CI/CD pipeline as a policy, not a suggestion.

Companies in regulated industries need to go further. If you’re subject to SOX compliance, FDA software validation, or financial regulations that require documented code provenance, you need a clear policy on AI-generated code that your auditors have reviewed and accepted.

The Practical Starting Point

If you’re not using AI in your development workflow yet, start here: add AI-assisted code review to your pull request process, use AI test generation for your lowest-coverage modules, and give your on-call team access to AI-powered log analysis. These three applications have the clearest ROI, the lowest risk, and the most immediate impact on team productivity and system reliability.


Related: AI Strategy for Non-Technical CEOs | Engineering Metrics That Actually Matter | Security and Compliance Without a CISO