logo
Oracle & NVIDIA expand OCI partnership with 160 AI tools

Oracle & NVIDIA expand OCI partnership with 160 AI tools

Techday NZ12-06-2025

Oracle and NVIDIA have expanded their partnership to enable customers to access more than 160 AI tools and agents while leveraging the necessary computing resources for AI development and deployment.
The collaboration brings NVIDIA AI Enterprise, a cloud-native software platform, natively to the Oracle Cloud Infrastructure (OCI) Console. Oracle customers can now use this platform across OCI's distributed cloud, including public regions, Government Clouds, and sovereign cloud solutions.
Platform access and capabilities
By integrating NVIDIA AI Enterprise directly through the OCI Console rather than a marketplace, Oracle allows customers to utilise their existing Universal Credits, streamlining transactions and support. This approach is designed to speed up deployment and help customers meet security, regulatory, and compliance requirements in enterprise AI processes.
Customers can now access over 160 AI tools focused on training and inference, including NVIDIA NIM microservices. These services aim to simplify the deployment of generative AI models and support a broad set of application-building and data management needs across various deployment scenarios. "Oracle has become the platform of choice for AI training and inferencing, and our work with NVIDIA boosts our ability to support customers running some of the world's most demanding AI workloads," said Karan Batta, Senior Vice President, Oracle Cloud Infrastructure. "Combining NVIDIA's full-stack AI computing platform with OCI's performance, security, and deployment flexibility enables us to deliver AI capabilities at scale to help advance AI efforts globally."
The partnership includes making NVIDIA GB200 NVL72 systems available on the OCI Supercluster, supporting up to 131,072 NVIDIA Blackwell GPUs. The new architecture provides a liquid-cooled infrastructure that targets large-scale AI training and inference requirements. Governments and enterprises can take advantage of the so-called AI factories, using platforms like NVIDIA's GB200 NVL72 for agentic AI tasks reliant on advanced reasoning models and efficiency enhancements.
Developer access to advanced resources
Oracle has become one of the first major cloud providers to integrate with NVIDIA DGX Cloud Lepton, which links developers to a global marketplace of GPU compute. This integration offers developers access to OCI's high-performance GPU clusters for a range of needs, including AI training, inference, digital twin implementations, and parallel HPC applications.
Ian Buck, Vice President of Hyperscale and HPC at NVIDIA, said: "Developers need the latest AI infrastructure and software to rapidly build and launch innovative solutions. With OCI and NVIDIA, they get the performance and tools to bring ideas to life, wherever their work happens."
With this integration, developers are also able to select compute resources in precise regions to help achieve both strategic and sovereign AI aims and satisfy long-term and on-demand requirements.
Customer projects using joint capabilities
Enterprises in Europe and internationally are making use of the enhanced partnership between Oracle and NVIDIA. For example, Almawave, based in Italy, utilises OCI AI infrastructure and NVIDIA Hopper GPUs to run generative AI model training and inference for its Velvet family, which supports Italian alongside other European languages and is being deployed within Almawave's AIWave platform. "Our commitment is to accelerate innovation by building a high-performing, transparent, and fully integrated Italian foundational AI in a European context—and we are only just getting started," said Valeria Sandei, Chief Executive Officer, Almawave. "Oracle and NVIDIA are valued partners for us in this effort, given our common vision around AI and the powerful infrastructure capabilities they bring to the development and operation of Velvet."
Danish health technology company Cerebriu is using OCI along with NVIDIA Hopper GPUs to build an AI-driven tool for clinical brain MRI analysis. Cerebriu's deep learning models, trained on thousands of multi-modal MRI images, aim to reduce the time required to interpret scans, potentially benefiting the clinical diagnosis of time-sensitive neurological conditions. "AI plays an increasingly critical role in how we design and differentiate our products," said Marko Bauer, Machine Learning Researcher, Cerebriu. "OCI and NVIDIA offer AI capabilities that are critical to helping us advance our product strategy, giving us the computing resources we need to discover and develop new AI use cases quickly, cost-effectively, and at scale. Finding the optimal way of training our models has been a key focus for us. While we've experimented with other cloud platforms for AI training, OCI and NVIDIA have provided us the best cloud infrastructure availability and price performance."
By expanding the Oracle-NVIDIA partnership, customers are now able to choose from a wide variety of AI tools and infrastructure options within OCI, supporting both research and production environments for AI solution development.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

Mirantis unveils architecture to speed & secure AI deployment
Mirantis unveils architecture to speed & secure AI deployment

Techday NZ

time2 days ago

  • Techday NZ

Mirantis unveils architecture to speed & secure AI deployment

Mirantis has released a comprehensive reference architecture to support IT infrastructure for AI workloads, aiming to assist enterprises in deploying AI systems quickly and securely. The Mirantis AI Factory Reference Architecture is based on the company's k0rdent AI platform and designed to offer a composable, scalable, and secure environment for artificial intelligence and machine learning (ML) workloads. According to Mirantis, the solution provides criteria for building, operating, and optimising AI and ML infrastructure at scale, and can be operational within days of hardware installation. The architecture leverages templated and declarative approaches provided by k0rdent AI, which Mirantis claims enables rapid provisioning of required resources. This, the company states, leads to accelerated prototyping, model iteration, and deployment—thereby shortening the overall AI development cycle. The platform features curated integrations, accessible via the k0rdent Catalog, for various AI and ML tools, observability frameworks, continuous integration and delivery, and security, all while adhering to open standards. Mirantis is positioning the reference architecture as a response to rising demand for specialised compute resources, such as GPUs and CPUs, crucial for the execution of complex AI models. "We've built and shared the reference architecture to help enterprises and service providers efficiently deploy and manage large-scale multi-tenant sovereign infrastructure solutions for AI and ML workloads," said Shaun O'Meara, chief technology officer, Mirantis. "This is in response to the significant increase in the need for specialized resources (GPU and CPU) to run AI models while providing a good user experience for developers and data scientists who don't want to learn infrastructure." The architecture addresses several high-performance computing challenges, including Remote Direct Memory Access (RDMA) networking, GPU allocation and slicing, advanced scheduling, performance tuning, and Kubernetes scaling. Additionally, it supports integration with multiple AI platform services, such as Gcore Everywhere Inference and the NVIDIA AI Enterprise software ecosystem. In contrast to typical cloud-native workloads, which are optimised for scale-out and multi-core environments, AI tasks often require the aggregation of multiple GPU servers into a single high-performance computing instance. This shift demands RDMA and ultra-high-performance networking, areas which the Mirantis reference architecture is designed to accommodate. The reference architecture uses Kubernetes and is adaptable to various AI workload types, including training, fine-tuning, and inference, across a range of environments. These include dedicated or shared servers, virtualised settings using KubeVirt or OpenStack, public cloud, hybrid or multi-cloud configurations, and edge locations. The solution addresses the specific needs of AI workloads, such as high-performance storage and high-speed networking technologies, including Ethernet, Infiniband, NVLink, NVSwitch, and CXL, to manage the movement of large data sets inherent to AI applications. Mirantis has identified and aimed to resolve several challenges in AI infrastructure, such as: Time-intensive fine-tuning and configuration compared to traditional compute systems; Support for hard multi-tenancy to ensure security, isolation, resource allocation, and contention management; Maintaining data sovereignty for data-driven AI and ML workloads, particularly where models contain proprietary information; Ensuring compliance with varied regional and regulatory standards; Managing distributed, large-scale infrastructure, which is common in edge deployments; Effective resource sharing, particularly of high-demand compute components such as GPUs; Enabling accessibility for users such as data scientists and developers who may not have specific IT infrastructure expertise. The composable nature of the Mirantis AI Factory Reference Architecture allows users to assemble infrastructure using reusable templates across compute, storage, GPU, and networking components, which can then be tailored to specific AI use cases. The architecture includes support for a variety of hardware accelerators, including products from NVIDIA, AMD, and Intel. Mirantis reports that its AI Factory Reference Architecture has been developed with the goal of supporting the unique operational requirements of enterprises seeking scalable, sovereign AI infrastructures, especially where control over data and regulatory compliance are paramount. The framework is intended as a guideline to streamline the deployment and ongoing management of these environments, offering modularity and integration with open standard tools and platforms.

Hitachi Vantara named leader for AI storage in GigaOm Radar 2025
Hitachi Vantara named leader for AI storage in GigaOm Radar 2025

Techday NZ

time3 days ago

  • Techday NZ

Hitachi Vantara named leader for AI storage in GigaOm Radar 2025

Hitachi Vantara has been recognised as a Leader and Fast Mover in the 2025 GigaOm Radar for High-Performance Storage Optimised for AI Workloads. The newly released GigaOm Radar report evaluates advanced storage platforms designed specifically for artificial intelligence (AI) workloads. This is the first time the report has assessed high-performance storage optimised for AI, and Hitachi Vantara's positioning underscores the company's capabilities in supporting enterprise-scale data requirements for AI and machine learning (ML). Report evaluation The GigaOm Radar for High-Performance Storage Optimised for AI Workloads reviews market solutions based on their key capabilities, foundational criteria, feature completeness, and readiness for future demands. Leaders in this assessment are noted for their maturity and strategic alignment with evolving enterprise AI needs, as well as their ability to execute at scale. Hitachi Vantara's placement as a Leader and Fast Mover was attributed to its Hitachi iQ portfolio, which provides AI-ready infrastructure solutions. The report highlighted several strengths, including quality of service (QoS) and workload isolation, GPU-direct storage integration and AI-optimised data layout and management. The report noted, "Hitachi iQ offers world-class QoS and workload isolation capabilities." By combining file system-level policies with flexible cluster architecture, Hitachi iQ enables effective resource allocation, supporting consistent and reliable performance for high-priority AI and ML workloads, even in multi-tenant or shared environments. The platform's GPU-direct storage integration was identified as another area of strong performance. Optimised drivers within Hitachi iQ support efficient data transfer between storage and GPU memory, enhancing outcomes for AI and ML workflows. Additionally, Hitachi iQ's data management strategy utilises intelligent placement across storage tiers, such as TLC/QLC NVMe flash storage for high performance and object storage for additional capacity. Real-time monitoring and user-defined policies are built in to balance performance and cost efficiency for varying data patterns. Industry commentary "What really stood out about Hitachi Vantara's offerings is the quality of service and the ability to isolate workloads," said Whit Walters, Field CTO and Analyst at GigaOm. "They've delivered a well-integrated, scalable platform in Hitachi iQ, backed by enterprise-proven storage. This combination gives organisations a powerful and flexible foundation for operationalising AI at scale, especially as large-scale AI and GenAI workloads will require the ability to manage data and performance as demands continue to grow." The GigaOm report also referenced ongoing collaborations including a partnership with NVIDIA on the NVIDIA Data Platform for Hitachi iQ, a partnership that adds to the capabilities of the platform. The roadmap for Hitachi iQ includes new hardware and AI solutions, among them the Hitachi iQ M Series, announced earlier in the year. Integration with the Virtual Storage Platform One (VSP One) enables further intelligent tiering between NVMe flash and object storage, providing additional flexibility and performance optimisation. Octavian Tanase, Chief Product Officer at Hitachi Vantara, commented on the recognition. "AI isn't just pushing the boundaries of what infrastructure needs to do; it's completely redrawing them," he said. "Our goal with Hitachi iQ is to give customers a high-performance foundation that removes complexity, accelerates outcomes, and adapts to whatever their AI journey requires next. By integrating Hitachi iQ with our VSP One platform, we're enabling a flexible, intelligent storage strategy that's ready for what's now and what's next." Ongoing awards This rating follows recent industry recognition for Hitachi Vantara. In May 2025, the company was awarded the Sustainable Technology Award at the Global Tech & AI Awards for its efforts in sustainable data infrastructure with the VSP One Block solution. Earlier in the year, GigaOm also recognised Hitachi Vantara as a Leader and Outperformer in the GigaOm Radar for Primary Storage, related to its VSP One hybrid cloud data platform. The GigaOm Radar for High-Performance Storage Optimised for AI Workloads aims to shed light on platforms capable of addressing the increased performance, scalability, and operational requirements integral to enterprise AI and machine learning deployments.

Oracle launches air-gapped cloud for security & sovereignty
Oracle launches air-gapped cloud for security & sovereignty

Techday NZ

time3 days ago

  • Techday NZ

Oracle launches air-gapped cloud for security & sovereignty

Oracle has unveiled Compute Cloud@Customer Isolated, a fully air-gapped cloud service designed for use by governments and regulated sectors needing stringent control over their data. This offering is positioned to address the requirements of agencies and industries with significant mission-critical security and compliance obligations, including sectors such as defence, intelligence, telecommunications, and healthcare. Compute Cloud@Customer Isolated has been launched as a deployable option that can be separated from the internet, a configuration commonly required to support sensitive and classified workloads. Security and sovereignty The new service is engineered to mitigate the risk of external threats, with a focus on data sovereignty tailored for entities with strict regulatory and security needs. According to Oracle, this is an area of particular importance in Australia and New Zealand, where national security and data handling regulations are strong priorities for government and critical infrastructure operators. Matt Leonard, Vice President, Edge Cloud Product Management at Oracle, said: "Protecting national security has always been a top priority for Oracle. Oracle Compute Cloud@Customer Isolated is designed to help regulated industries benefit from the cloud and AI, while also providing the flexibility to deploy anywhere." Compute Cloud@Customer Isolated provides organisations with access to the same compute, storage, and networking capabilities available from Oracle's existing Compute Cloud@Customer offering, but adds the option for an isolated deployment. This is intended to enhance organisations' abilities to accelerate artificial intelligence development and improve efficiency, whilst maintaining comprehensive data sovereignty and control over the underlying infrastructure. Deployment flexibility The system can be installed as a single rack and scaled up according to need, granting government and regulated sector customers the ability to quickly deploy Oracle Cloud Infrastructure (OCI) Compute services on their own premises—in isolated environments designed for locations with unique sovereignty demands. This capability is expected to make it viable for defence users and other sensitive operations to begin or expand cloud migration in regions where mission requirements dictate the need for air-gapped technology. Oracle presents multiple migration paths, including a Fast-Start option, which can deliver the Compute Cloud@Customer Isolated service on premises within six to eight weeks. This pathway allows customers to later develop a full Oracle Cloud Isolated Region as their requirements evolve. Alternatively, customers can take a phased approach, progressing from edge computing solutions, through isolated cloud deployments, and up towards a fully disconnected Oracle Cloud Isolated Region. This Isolated Region is entirely separate from public clouds and the wider internet, designed to provide control and security specifically demanded by defence and other regulated clients. Partner comments Andy Laidler, Chief Digital Officer for Fujitsu Defence and National Security, addressed the importance of robust and portable connectivity for organisations whose operational backbone now depends on digital services: "As digital services become the operational backbone—from logistics to command and control—forces need connectivity that moves with the mission. Fujitsu's Secure Cloud solution incorporates the full range of Oracle deployment options to ensure we deliver capability where it is needed most. Oracle Compute Cloud@Customer Isolated provides a secure, isolated environment and access to data anywhere, anytime, and under all conditions to help propel mission success." The Compute Cloud@Customer Isolated service will be made available globally later in the year, expanding Oracle's suite of cloud infrastructure offerings to clients operating under the world's most stringent data control laws. Follow us on: Share on:

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store