Managing the use of generative AI across the software development life cycle may require mixing and matching tools, metrics and methods
Developers adopt GenAI (generative AI) because it helps to them to create. Code fasterbut the tooling can expose organisations to unaccounted risks, especially if it is used unauthorised or if best practices are ignored.
With GenAI, we’ve seen both amazing and stunningly stupid results from the same development team. This tells us we need to work on the process and tools side. “My personal opinion is that AI can be a game changer. Dunning Kruger effect is incarnate.”
Dunning Kruger effect is a cognitive bias. The less knowledge or skill you have on a topic, the more likely it is that you will overestimate your competence in the area.
Some members of a team with average skill may review a piece code and think that it looks good. Colwell explains that true experts may look at a piece of code and see security vulnerabilities, bad packages, and other issues.
Dealing with this and avoiding the risk of multiple types of errors you haven’t seen previously can be difficult. There are tools available, but you must first have a robust security policy and practices that are enforced. GenAI is able to create code more quickly, so organisations may find it difficult to test code at a rate that keeps up with code production.
A 2025 Tricentis conducted a survey of 2,700 leaders, mainly executives, managers, IT professionals, and found that 63% admitted to shipping untested code and 90% said they trusted GenAI for software release decisions.
Defenses against superficially impressing results
Team members who are less knowledgeable can be asking Genes is a set of tools that can be used to create an app for any given task. The code that is generated can look impressive on the surface, but it may contain serious issues.
Software development, whether AI-based or not, should be secure in its process. If you commit code it must pass security checks, validation checks, dynamic scanning and more. Colwell confirms that you cannot eliminate all mistakes in code. “Code with stupidity in it”he says.
An example would be if an application or website that is age restricted allows users under 18 to access it, but the code does not deny them at certain entry points. Or if they accidentally click the wrong button or give an incorrect response. These simple errors are common and must be checked every time. All code, no matter how it is created, must pass the review and validation process. All code must be reviewed and validated.
It is important to document what the teams are doing. AI-powered testing tools, network monitoring, and backlog management can help detect code issues and prioritise the changes according to risk.
McKinsey’s study indicates that using surveys, data and backlog management can reduce customer-reported defects in software by 20-30%. App discovery software that detects AI usage, and data loss prevention tools (DLP) to pinpoint inappropriate information sharing are also valuable.
Tools for code coverage analysis can track which parts of the code are executed during functional testing. They may identify code bits that are not executed during the test, indicating AI-generated code which is unneeded or incorrect. A relevant tool can also identify redundant or irrelevant requirements. AIs can add strange things into code to make a test pass or satisfy a specific situation.
But Colwell notes that, above all, protecting organisations from risks introduced into code through unauthorised or improper GenAI usage means investing in education and training. Organisations should take note. You can trust their policies, documentation, and practices if you know that developer teams have been well-trained in best practice, and the risks they face if something goes wrong. Colwell says, “Teach the limitations of their tooling to your engineers and those using GenAI.” “A lot people will think that GenAI is more or less a magical thinking box. But what you have is a problem-solving tool with a short-term memory, a tendency for it to answer quickly rather than find concrete information, and no access to the external environment. Ankur Anand is the CIO of Nash Squared which owns Harvey Nash IT Recruitment. He says that the AI skills shortage in tech is the largest for 15 years.
AI includes understanding how to leverage platforms and CRM. It also includes learning about prompts and the “responsibility that comes with it”including auditing results before use. GenAI skillsets are in high demand for developers, project managers, and product managers, as well as data lineage, data governance, and data quality.
Below, Nash Squared May Report found an increased demand for Python programmers with knowledge of large language models (LLM), for example. GenAI is now a “nice-to-have” in many job descriptions. Anand says that this puts more pressure on the tech team, as it is not just about one area.
Facundo Guliani, a solutions engineer at CMS provider Storyblok, agrees that GenAI code can be a great starting point, if it is supervised by someone who knows what the machine is doing and what happens in the background. It is important to maintain control of the development process, especially when there are multiple teams involved.
Additional approaches for quality AI coding
Giuliani Does””https://www.techtarget.com/searchenterprisedesktop/opinion/End-users-can-code-with-AI-but-IT-must-be-wary””> Code generated from AI models trained on public internet datasets are often based upon datasets that do not represent a reliable source of truth. Some of the solutions that might indicate that something is wrong or missing in the code or that copy-and paste has been used excessively or that it is simply not correct, include long-way arounds or tangential ones.
Do you find more bugs than expected, or do things happen too fast or slowly? Attention to productivity There are a number of metrics that can be used, including DevOps Research and Assessment, Space/wellbeing Activity, Performance Communication, Efficiency/flow, SPACE metrics, talent capability scores, and contribution analysis.
It is necessary to have formal AI governance and AI model-risk management (MRM). There are also standards and frameworks that help assess AI risks.
The International Standards Organisation (ISO), standard 42,001, is about managing AI responsibly. And the US National Institute of Standards and Technology, NIST, is also involved. AI Risk Management Framework(and Playbook are being developed.
Giuliani said that any patterns that are out of the ordinary should be examined for bad coding practices. “You may see overly complex solutions to simple problems. It’s the same with code written by people who have no experience. A mentor or companion should help them improve their knowledge. “A person must be responsible for the code prior to production,” he says.
To check code adequately, you can use a variety of techniques or processes. For example, you could have a colleague or supervisor perform manual code checks prior to submission to production environments. Giuliani says that this is true regardless of the way code was written.
Jody Baily, chief product and tech officer (CPTO), at developer community Stack Overflow is in general agreement: “You have to make sure that the code is still of high quality.” You need oversight and review. Many people use prompts to write code, but they are still evaluating the prompts.”
The challenge for developers was never about how quickly you type or write code. Bailey says it’s more about having the right ideas, and thinking logically and efficiently about problems. He acknowledges that using AI could be a way to validate AI. An example of this would be to use Anthropic versus Gemini. Different modelsare different in their strengths and weaknesses.
On the various leaderboards this can change month to month. He says that some are more code-focused and others more general.
While you may not be able to completely eliminate shadow IT, more generalized shadow IT is a good alternative. Monitoring is a tool that can help, especially with web interactions and endpoints management. If people try something new and it works, then the organization may decide to continue with that.
I can’t but think about sports, when someone takes a shot, and the coach says ‘No, No, Not like that! ‘, then the goal is scored and it’s “Yay!” [instead],” says Bailey.
Depending on the circumstances and needs, the approach chosen will vary. However, code must be subject to oversight and quality control whether GenAI is being used or not. Bailey says that the alternative is to have a very restricted environment, where there is a risk of losing agility and innovation.
By John Moore.
What has generative AI done to cybersecurity? John Moore
By: Melinda Marks
How has generative AI affected cybersecurity?

