
C-suite divisions slow GenAI adoption due to security worries
A new report from NTT DATA has highlighted a misalignment among senior executives regarding the adoption and security implications of generative artificial intelligence (GenAI) in organisations globally.
NTT DATA's report, "The AI Security Balancing Act: From Risk to Innovation," is based on survey responses from more than 2,300 senior GenAI decision makers, including over 1,500 C-level executives across 34 countries. The findings underscore a gap between the optimism of CEOs and the caution of Chief Information Security Officers (CISOs) concerning GenAI deployment.
C-Suite perspectives
The report indicates that 99% of C-Suite executives are planning to increase their GenAI investments over the next two years, with 67% of CEOs preparing for significant financial commitments. In comparison, 95% of Chief Information Officers (CIOs) and Chief Technology Officers (CTOs) report that GenAI is already influencing, or will soon drive, greater spending on cybersecurity initiatives. Improved security was named among the top three benefits realised from GenAI adoption in the past year.
Despite these high expectations, a considerable number of CISOs express reservations. Nearly half (45%) of CISOs surveyed shared negative sentiments about GenAI rollouts, identifying security gaps and the challenge of modernising legacy infrastructure as primary barriers.
The report also finds differences in the perception of policy clarity. More than half of CISOs (54%) stated that internal GenAI policies are unclear, compared with just 20% of CEOs. This suggests a disconnect between business leaders' strategic vision and concerns raised by operational security managers. "As organisations accelerate GenAI adoption, cybersecurity must be embedded from the outset to reinforce resilience. While CEOs champion innovation, ensuring seamless collaboration between cybersecurity and business strategy is critical to mitigating emerging risks," said Sheetal Mehta, Senior Vice President and Global Head of Cybersecurity at NTT DATA, Inc. "A secure and scalable approach to GenAI requires proactive alignment, modern infrastructure and trusted co-innovation to protect enterprises from emerging threats while unlocking AI's full potential."
Operational and skills challenges
The study highlights that, while 97% of CISOs consider themselves GenAI decision makers, 69% acknowledge their teams currently lack the necessary skills to work effectively with GenAI technologies. Only 38% of CISOs said their organisation's GenAI and cyber security strategies are aligned, compared with 51% of CEOs.
Another area of concern identified is the absence of clearly defined policies for GenAI use within organisations. According to the survey, 72% of respondents had yet to implement a formal GenAI usage policy, and just 24% of CISOs strongly agreed their company has an adequate framework for balancing the risks and rewards of GenAI adoption.
Infrastructure and technology barriers
Legacy technology also poses a significant challenge to GenAI integration. The research found that 88% of security leaders believe outdated infrastructure is negatively affecting both business agility and GenAI readiness. Upgrading systems such as Internet of Things (IoT), 5G, and edge computing was identified as crucial for future progress.
To address these obstacles, 64% of CISOs reported prioritising collaboration with strategic IT partners and co-innovation, rather than relying on proprietary AI solutions. When choosing GenAI technology partners, security leaders ranked end-to-end service integration as their most important selection criterion. "Collaboration is highly valued by line-of-business leaders in their relationships with CISOs. However, disconnects remain, with gaps between the organisation's desired risk posture and its current cybersecurity capabilities," said Craig Robinson, Research Vice President, Security Services at IDC. "While the use of GenAI clearly provides benefits to the enterprise, CISOs and Global Risk and Compliance leaders struggle to communicate the need for proper governance and guardrails, making alignment with business leaders essential for implementation."
Survey methodology
The report's data derives from a global survey of 2,300 senior GenAI decision makers. Of these respondents, 68% were C-suite executives, with the remainder comprising vice presidents, heads of department, directors, and senior managers. The research, conducted by Jigsaw Research, aimed to capture perspectives on both the opportunities and risks associated with GenAI across different regions and sectors.
The report points to the need for structured governance, clarity in strategic direction, and investment in modern infrastructure to ensure successful and secure GenAI deployments in organisations.

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


Otago Daily Times
2 days ago
- Otago Daily Times
‘Nanogirl' informs South on AI's use
Even though "Nanogirl", Dr Michelle Dickinson, has worked with world leading tech giants, she prefers to inspire the next generation. About 60 Great South guests were glued to their Kelvin Hotel seats on Thursday evening as the United Kingdom-born New Zealand nanotechnologist shared her knowledge and AI's future impact. Business needed to stay informed about technology so it could future-proof, she said. The days were gone where the traditional five year business plan would be enough to futureproof due to the breakneck speed technology has been advancing. Owners also needed to understand the importance of maintaining a customer-centric business or risk becoming quickly irrelevant. "I care about that we have empty stores." The number of legacy institutions closing was evidence of its model not moving with the customer. "Not being customer-centric is the biggest threat to business." Schools were another sector which needed to adapt to the changing world as it predominantly catered to produce an "average" student. "Nobody wants their kids to be average." Were AI technology to be implemented it could be used to develop personalised learning models while removing the stress-inducing and labour-intensive tasks from teachers' workload. "Now you can be the best teacher you can be and stay in the field you love. "I don't want our teachers to be burnt out, I want them to be excited to be teaching." In 30 seconds, new technology could now produce individualised 12-week teaching plans aligned to the curriculum, in both Ma¯ori and English she said. Agriculture was another sector to benefit from the developing technology. Better crop yields and cost savings could now be achieved through localised soil and crop tracking information which pinpointed what fertiliser needs or moisture levels were required in specific sections of a paddock. While AI was a problem-solving tool which provided outcomes on the information available to it, to work well, it still needed the creative ideas to come from humans, she said. "People are the fundamentals of the future . . . and human side of why we do things should be at the forefront. "We, as humans, make some pretty cool decisions that aren't always based on logic." Personal and commercial security had also become imperative now there was the ability to produce realistic "deep-fake" productions with videos and audio was about to hit us. She urged families and organisations to have "safe words" that would not be present in deep fake recordings and allow family members or staff to identify fake from genuine cries for help. "This is the stuff we need to be talking about with our kids right now." Great South chief executive Chami Abeysinghe said Dr Dickinson's presentation raised some "thought-provoking" questions for Southland's business leaders. She believed there needed to be discussions about how Southland could position itself to be at the forefront of tech-driven innovation. "I think some of the points that she really raised was a good indication that we probably need to get a bit quicker at adopting and adapting. "By the time we get around to thinking about it, it has already changed again." AI was able to process information and data in a fraction of the time humans did, but the technology did not come without risks and it was critical businesses protected their operations. "If we are going to use it, we need to be able to know that it's secure." Information on ChatGPT entered the public realm that everyone could have access to and business policies had not kept up. "You absolutely have to have a [AI security] policy."


Techday NZ
2 days ago
- Techday NZ
Nearly half of developers say over 50% of code is AI-generated
Cloudsmith's latest report shows that nearly half of all developers using AI in their workflows now have codebases that are at least 50% AI-generated. The 2025 Artifact Management Report from Cloudsmith surveyed 307 software professionals in the US and UK, all working with AI as part of their development, DevOps, or CI/CD processes. Among these respondents, 42% reported that at least half of their current codebase is now produced by AI tools. Despite the large-scale adoption of AI-driven coding, oversight remains inconsistent. Only 67% of developers who use AI review the generated code before every deployment. This means nearly one-third of those working with AI-assisted code are deploying software without always performing a human review, even as new security risks linked to AI-generated code are emerging. Security concerns The report points to a gap between the rapid pace of AI integration in software workflows and the implementation of safety checks and controls. Attacks such as 'slopsquatting'—where malicious actors exploit hallucinated or non-existent dependencies suggested by AI code assistants—highlight the risks when AI-generated code is left unchecked. Cloudsmith's data shows that while 59% of developers say they apply extra scrutiny to AI-generated packages, far fewer have more systematic approaches in place for risk mitigation. Only 34% use tools that enforce policies specific to AI-generated artifacts, and 17% acknowledge they have no controls in place at all for managing AI-written code or dependencies. "Software development teams are shipping faster, with more AI-generated code and AI agent-led updates," said Glenn Weinstein, CEO at Cloudsmith. "AI tools have had a huge impact on developer productivity, which is great. That said, with potentially less human scrutiny on generated code, it's more important that leaders ensure the right automated controls are in place for the software supply chain." Developer perceptions The research reveals a range of attitudes towards AI-generated code among developers. While 59% are cautious and take extra steps to verify the integrity of code created by AI, 20% said they trust AI-generated code "completely." This suggests a marked difference in risk appetite and perception within developer teams, even as the majority acknowledge the need for vigilance. Across the sample, 86% of developers reported an increase in the use of AI-influenced packages or software dependencies in the past year, and 40% described this increase as "significant." Nonetheless, only 29% of those surveyed felt "very confident" in their ability to detect potential vulnerabilities in open-source libraries, from which AI tools frequently pull suggestions. "Controlling the software supply chain is the first step towards securing it," added Weinstein. "Automated checks and use of curated artifact repositories can help developers spot issues early in the development lifecycle." Tooling and controls The report highlights that adoption of automated tools specifically designed for AI-generated code remains limited, despite the stated importance of security among software development teams. While AI technologies accelerate the pace of software delivery and updating, adoption of stricter controls and policy enforcement is not keeping up with the new risks posed by machine-generated code. The findings indicate a potential lag in upgrading security processes or artifact management solutions to match the growing use of AI in coding. Developers from a range of industries—including technology, finance, healthcare, and manufacturing—participated in the survey, with roles spanning development, DevOps management, engineering, and security leadership in enterprises with more than 500 employees. The full Cloudsmith 2025 Artifact Management Report also explores other key issues, including how teams decide which open-source packages to trust, the expanding presence of AI in build pipelines, and the persistent challenges in prioritising tooling upgrades for security benefits.


Techday NZ
2 days ago
- Techday NZ
Hitachi Vantara named leader for AI storage in GigaOm Radar 2025
Hitachi Vantara has been recognised as a Leader and Fast Mover in the 2025 GigaOm Radar for High-Performance Storage Optimised for AI Workloads. The newly released GigaOm Radar report evaluates advanced storage platforms designed specifically for artificial intelligence (AI) workloads. This is the first time the report has assessed high-performance storage optimised for AI, and Hitachi Vantara's positioning underscores the company's capabilities in supporting enterprise-scale data requirements for AI and machine learning (ML). Report evaluation The GigaOm Radar for High-Performance Storage Optimised for AI Workloads reviews market solutions based on their key capabilities, foundational criteria, feature completeness, and readiness for future demands. Leaders in this assessment are noted for their maturity and strategic alignment with evolving enterprise AI needs, as well as their ability to execute at scale. Hitachi Vantara's placement as a Leader and Fast Mover was attributed to its Hitachi iQ portfolio, which provides AI-ready infrastructure solutions. The report highlighted several strengths, including quality of service (QoS) and workload isolation, GPU-direct storage integration and AI-optimised data layout and management. The report noted, "Hitachi iQ offers world-class QoS and workload isolation capabilities." By combining file system-level policies with flexible cluster architecture, Hitachi iQ enables effective resource allocation, supporting consistent and reliable performance for high-priority AI and ML workloads, even in multi-tenant or shared environments. The platform's GPU-direct storage integration was identified as another area of strong performance. Optimised drivers within Hitachi iQ support efficient data transfer between storage and GPU memory, enhancing outcomes for AI and ML workflows. Additionally, Hitachi iQ's data management strategy utilises intelligent placement across storage tiers, such as TLC/QLC NVMe flash storage for high performance and object storage for additional capacity. Real-time monitoring and user-defined policies are built in to balance performance and cost efficiency for varying data patterns. Industry commentary "What really stood out about Hitachi Vantara's offerings is the quality of service and the ability to isolate workloads," said Whit Walters, Field CTO and Analyst at GigaOm. "They've delivered a well-integrated, scalable platform in Hitachi iQ, backed by enterprise-proven storage. This combination gives organisations a powerful and flexible foundation for operationalising AI at scale, especially as large-scale AI and GenAI workloads will require the ability to manage data and performance as demands continue to grow." The GigaOm report also referenced ongoing collaborations including a partnership with NVIDIA on the NVIDIA Data Platform for Hitachi iQ, a partnership that adds to the capabilities of the platform. The roadmap for Hitachi iQ includes new hardware and AI solutions, among them the Hitachi iQ M Series, announced earlier in the year. Integration with the Virtual Storage Platform One (VSP One) enables further intelligent tiering between NVMe flash and object storage, providing additional flexibility and performance optimisation. Octavian Tanase, Chief Product Officer at Hitachi Vantara, commented on the recognition. "AI isn't just pushing the boundaries of what infrastructure needs to do; it's completely redrawing them," he said. "Our goal with Hitachi iQ is to give customers a high-performance foundation that removes complexity, accelerates outcomes, and adapts to whatever their AI journey requires next. By integrating Hitachi iQ with our VSP One platform, we're enabling a flexible, intelligent storage strategy that's ready for what's now and what's next." Ongoing awards This rating follows recent industry recognition for Hitachi Vantara. In May 2025, the company was awarded the Sustainable Technology Award at the Global Tech & AI Awards for its efforts in sustainable data infrastructure with the VSP One Block solution. Earlier in the year, GigaOm also recognised Hitachi Vantara as a Leader and Outperformer in the GigaOm Radar for Primary Storage, related to its VSP One hybrid cloud data platform. The GigaOm Radar for High-Performance Storage Optimised for AI Workloads aims to shed light on platforms capable of addressing the increased performance, scalability, and operational requirements integral to enterprise AI and machine learning deployments.