AI in DevSecOps: Report Highlights Security and Privacy Dilemma
GitLab Inc., the most comprehensive AI-powered DevSecOps platform, today announced its 9th Global DevSecOps Report: The State of AI in Software Development. GitLab surveyed more than 1,000 global senior technology executives, developers, and security and operations professionals on their successes, challenges, and priorities for AI adoption.
“The transformational opportunity with AI goes way beyond creating code,” said David DeSanto, chief product officer, GitLab. “According to the GitLab Global DevSecOps Report, only 25% of developers’ time is spent on code generation, but the data shows AI can boost productivity and collaboration in nearly 60% of developers’ day-to-day work. To realise AI’s full potential, it needs to be embedded across the software development lifecycle, allowing everyone involved in delivering secure software, not just developers, to benefit from the efficiency boost. GitLab’s AI-powered DevSecOps platform delivers a privacy-first, single application to help teams deliver secure software faster.”
Key takeaways of the report are:
Data Privacy, Intellectual Property, and Security are Top Concerns
Although organisations are enthusiastic about implementing AI, data privacy and intellectual property are key priorities when adopting new tools.
- 95% of senior technology executives said they prioritise privacy and protection of intellectual property when selecting an AI tool.
- 32% of respondents were “very” or “extremely” concerned about introducing AI into the software development lifecycle; of those:
- 39% cited they are concerned that AI-generated code may introduce security vulnerabilities and 48% said they are concerned that AI-generated code may not be subject to the same copyright protection as human-generated code.
Increased Developer Productivity May Widen the Existing Divide Between Developers and Security Professionals
Security professionals worry that AI-generated code could result in more security vulnerabilities—making more work for security professionals.
- Only 7% of developers’ time is spent identifying and mitigating security vulnerabilities and 11% is spent on testing code.
- 48% of developers were significantly more likely to identify faster cycle times as a benefit of AI, compared to 38% of security professionals.
- 51% of all respondents are already seeing productivity as a key benefit of AI implementation.
Insufficient Training and Resources Causing Widening AI Skills Gap
While respondents remain optimistic about their company’s use of AI, the data indicates a discrepancy between organisations’ and practitioners’ satisfaction with AI training resources. Despite 75% of respondents saying their organisation provides training and resources for using AI, a roughly equal proportion also said they are finding resources on their own, suggesting that the available resources and training may be insufficient.
- 81% cited they require training to successfully use AI in their daily work.
- 65% who use, or are planning to use, AI for software development said their organisation hired or will hire new talent to manage AI implementation.
- When asked what types of resources are being used to build AI skills, the top responses were:
- 49% utilise books, articles, and online videos, 49% watch educational courses, 47% practice with open-source projects, and 47% learn from peers and mentors.
“Enterprises are seeking out platforms that allow them to harness the power of AI while addressing potential privacy and security risks,” said Alexander Johnston, Research Analyst in the Data, AI & Analytics channel at 451 Research, a part of S&P Global Market Intelligence. “There is industry demand for privacy-first, sustainably adopted AI.”