Latest news with #NVIDIAAIEnterprise


Techday NZ
2 days ago
- Business
- Techday NZ
Mirantis unveils architecture to speed & secure AI deployment
Mirantis has released a comprehensive reference architecture to support IT infrastructure for AI workloads, aiming to assist enterprises in deploying AI systems quickly and securely. The Mirantis AI Factory Reference Architecture is based on the company's k0rdent AI platform and designed to offer a composable, scalable, and secure environment for artificial intelligence and machine learning (ML) workloads. According to Mirantis, the solution provides criteria for building, operating, and optimising AI and ML infrastructure at scale, and can be operational within days of hardware installation. The architecture leverages templated and declarative approaches provided by k0rdent AI, which Mirantis claims enables rapid provisioning of required resources. This, the company states, leads to accelerated prototyping, model iteration, and deployment—thereby shortening the overall AI development cycle. The platform features curated integrations, accessible via the k0rdent Catalog, for various AI and ML tools, observability frameworks, continuous integration and delivery, and security, all while adhering to open standards. Mirantis is positioning the reference architecture as a response to rising demand for specialised compute resources, such as GPUs and CPUs, crucial for the execution of complex AI models. "We've built and shared the reference architecture to help enterprises and service providers efficiently deploy and manage large-scale multi-tenant sovereign infrastructure solutions for AI and ML workloads," said Shaun O'Meara, chief technology officer, Mirantis. "This is in response to the significant increase in the need for specialized resources (GPU and CPU) to run AI models while providing a good user experience for developers and data scientists who don't want to learn infrastructure." The architecture addresses several high-performance computing challenges, including Remote Direct Memory Access (RDMA) networking, GPU allocation and slicing, advanced scheduling, performance tuning, and Kubernetes scaling. Additionally, it supports integration with multiple AI platform services, such as Gcore Everywhere Inference and the NVIDIA AI Enterprise software ecosystem. In contrast to typical cloud-native workloads, which are optimised for scale-out and multi-core environments, AI tasks often require the aggregation of multiple GPU servers into a single high-performance computing instance. This shift demands RDMA and ultra-high-performance networking, areas which the Mirantis reference architecture is designed to accommodate. The reference architecture uses Kubernetes and is adaptable to various AI workload types, including training, fine-tuning, and inference, across a range of environments. These include dedicated or shared servers, virtualised settings using KubeVirt or OpenStack, public cloud, hybrid or multi-cloud configurations, and edge locations. The solution addresses the specific needs of AI workloads, such as high-performance storage and high-speed networking technologies, including Ethernet, Infiniband, NVLink, NVSwitch, and CXL, to manage the movement of large data sets inherent to AI applications. Mirantis has identified and aimed to resolve several challenges in AI infrastructure, such as: Time-intensive fine-tuning and configuration compared to traditional compute systems; Support for hard multi-tenancy to ensure security, isolation, resource allocation, and contention management; Maintaining data sovereignty for data-driven AI and ML workloads, particularly where models contain proprietary information; Ensuring compliance with varied regional and regulatory standards; Managing distributed, large-scale infrastructure, which is common in edge deployments; Effective resource sharing, particularly of high-demand compute components such as GPUs; Enabling accessibility for users such as data scientists and developers who may not have specific IT infrastructure expertise. The composable nature of the Mirantis AI Factory Reference Architecture allows users to assemble infrastructure using reusable templates across compute, storage, GPU, and networking components, which can then be tailored to specific AI use cases. The architecture includes support for a variety of hardware accelerators, including products from NVIDIA, AMD, and Intel. Mirantis reports that its AI Factory Reference Architecture has been developed with the goal of supporting the unique operational requirements of enterprises seeking scalable, sovereign AI infrastructures, especially where control over data and regulatory compliance are paramount. The framework is intended as a guideline to streamline the deployment and ongoing management of these environments, offering modularity and integration with open standard tools and platforms.


Web Release
3 days ago
- Business
- Web Release
VAST Data Powers Smarter, Evolving AI Agents with NVIDIA Data Flywheel
VAST Data, the AI Operating System company, announced today that it is delivering a complete data and compute platform that enables AI intelligence to continuously evolve. The VAST AI OS, combined with NVIDIA AI Enterprise, which includes NeMo microservices that power a data flywheel for continuous model improvement, creates a unified environment where AI pipelines can constantly learn, adapt, and improve. This reference workflow provides a self-optimizing foundation for scalable AI, laying the groundwork for billions of intelligent agents to fine-tune and evolve from their data and experiences. This solution provides enterprises with a converged software platform for data management, database services, and AI compute orchestration. Additionally, VAST AI OS AgentEngine uniquely shares feedback by providing the critical capability to map the intricate web of agent-data interactions through production logs. This granular traceability allows the flywheel to dissect these multi-step interactions, accurately identifying which specific elements require adjustment to enhance outcomes, accelerating model performance and accuracy at scale. 'AI-powered businesses need thinking machines designed for a future where billions of AI agents learn from their own experiences, fine-tune in real time, and create new possibilities through collaboration,' said Jeff Denworth, Co-Founder of VAST Data. 'By unifying NVIDIA's AI software and hardware technologies within the core of the VAST AI Operating System, we are giving customers the foundation to operationalize continuous improvements in AI intelligence at scale, with the security, governance, and service delivery tools required to manage these intelligent agents and the data they rely upon.' As AI moves from isolated projects to always-on infrastructure, businesses need systems that evolve in real time with every data point while addressing new security and governance challenges around fine-tuned models, agent interactions, and decentralized data pipelines. This collaboration makes it possible to run continuous, automated AI pipelines – from ingestion to inference to retraining – all managed within the VAST AI Operating System. Among the first to embrace this strategy is CACEIS, one of Europe's largest asset servicing firms. In collaboration with VAST and NVIDIA, CACEIS is exploring a real-time AI platform concept designed to securely capture, transcribe, and analyze 100% of client meetings. The vision is for the system to instantly generate meeting minutes, surface actionable insights, and deliver anonymized trend data — all seamlessly integrated into their sovereign CRM. With an end-to-end security model at its foundation, the platform is being designed to safeguard client privacy and data integrity while continuously evolving through every interaction. 'AI will be a game-changer, highlighting trends in current needs by analysing meeting reports so we can better serve clients,' said Arnaud Misset, Chief Digital Officer, CACEIS. Using VAST's AgentEngine that leverages the NVIDIA AI-Q Blueprint, CACEIS is developing a platform proof of concept that would enable AI agents to assist relationship managers in real time and help uncover new business opportunities. Built by NVIDIA NeMo microservices and the NVIDIA data flywheel blueprint, the envisioned CACEIS AI factory would continuously capture data and insights from every customer interaction. These feedback loops are intended to drive ongoing model refinement and training, allowing the system to improve and adapt with each meeting. As AI agents learn from one another and from human counterparts, this concept sets the stage for new ideas, collective intelligence, and enterprise-wide knowledge sharing to take shape. This capability was showcased during NVIDIA's Kari Briski's presentation at GTC Europe in Paris. Watch the demo replay here. This collaboration signals a broader shift in enterprise AI, from one-size-fits-all models to dynamic ecosystems of intelligent agents that continuously fine-tune, collaborate, and generate new ideas from their own data and interactions. Managing these agents at scale requires fine-grained security, governance, and access controls to ensure they operate safely and within defined boundaries. It also demands scalable, dynamic infrastructure capable of handling the varied and unpredictable demands of AI agents as they interact with diverse data tools and services. The VAST AI Operating System provides this real-time data infrastructure, along with compute orchestration, QoS tools that enforce fairness as different agents run within the environment, and a security framework to refine intelligence and operationalize AI innovation in a secure, scalable, and adaptive system. 'Data flywheels leverage each interaction with an AI agent to continuously improve system intelligence and value,' said Kari Briski, Vice President of Generative AI Software at NVIDIA. 'CACEIS is an exemplary pioneer with a vision of building an agentic AI data flywheel with NVIDIA and VAST to supercharge productivity for financial services in Europe.' Additional Resources: ? VAST + NVIDIA ? DEMO: AI Agents Unlocked: CACEIS Redefines Client Conversations With VAST Data and NVIDIA ? NVIDIA BLOG: Chat with Your Enterprise Data: Open-Source AI-Q NVIDIA Blueprint Puts Institutional Knowledge at Your Fingertips ? NVIDIA BLOG: Sovereign AI Agents Think Local, Act Global With NVIDIA AI Factories ? NVIDIA BLOG: Build Efficient AI Agents Through Model Distillation With NVIDIA's Data Flywheel Blueprint
Yahoo
4 days ago
- Business
- Yahoo
Nebius Group (NBIS) and Saturn Cloud Partner to Deliver Turnkey AI/ML Infrastructure Solution Built on NVIDIA Hopper GPUs
Nebius Group N.V. (NASDAQ:NBIS) is one of the . On June 11, Nebius Group N.V. (NASDAQ:NBIS) announced a partnership with Saturn Cloud to deliver a turnkey AI/ML infrastructure solution that is built on NVIDIA Hopper GPUs, with support for the NVIDIA AI Enterprise software stack. Saturn Cloud is the MLOps platform for AI/ML engineers. The collaboration would enable AI engineers to access an enterprise-grade AI/ML infrastructure-in-a-box with on-demand access to an enterprise-ready MLOps platform and NVIDIA Hopper GPUs. A computer screen showcasing Artificial Intelligence and Machine Learning algorithms at work. The solution would combine Saturn Cloud's engineer-loved MLOps platform with the flexibility and power of Nebius's AI cloud, allowing anyone to instantly sign up and run jobs and deployments, use Jupyter notebooks or other IDEs, and manage cloud resources on NVIDIA Hopper GPUs through Nebius Group's (NASDAQ:NBIS) infrastructure. The solution is a compelling option for all use case types because of its notably lower cost compared to traditional cloud service providers. Nebius Group N.V. (NASDAQ:NBIS) takes the ninth spot on our list of the top hot large-cap stocks to invest in. Nebius Group N.V. (NASDAQ:NBIS) is a technology company that provides services and infrastructure to AI builders across the globe. The company's offerings include Nebius AI, which is an AI-centric cloud platform that offers full-stack infrastructure. This includes cloud services, developer tools, and large-scale GPU clusters, cloud services, and developer tools. While we acknowledge the potential of NBIS as an investment, we believe certain AI stocks offer greater upside potential and carry less downside risk. If you're looking for an extremely undervalued AI stock that also stands to benefit significantly from Trump-era tariffs and the onshoring trend, see our free report on the best short-term AI stock. READ NEXT: The Best and Worst Dow Stocks for the Next 12 Months and 10 Unstoppable Stocks That Could Double Your Money. Disclosure: None.


TECHx
5 days ago
- Business
- TECHx
VAST Data Unveils AI OS to Power Evolving AI Agents
Home » Tech Value Chain » Global Brands » VAST Data Unveils AI OS to Power Evolving AI Agents VAST Data, the AI Operating System company, has announced a complete data and compute platform that enables continuous AI intelligence evolution. The VAST AI OS, integrated with NVIDIA AI Enterprise and NeMo microservices, powers a self-optimizing AI flywheel. This creates a unified environment where AI pipelines can learn, adapt, and improve at scale. The platform delivers: A converged software solution for data management, database services, and AI compute orchestration. AgentEngine, a tool offering feedback by mapping complex agent-data interactions through production logs. This approach allows enterprises to identify and adjust key performance elements, boosting model accuracy and learning speed. Jeff Denworth, Co-Founder of VAST Data, said the solution supports businesses building 'thinking machines' that fine-tune in real time and collaborate to generate new outcomes. By integrating NVIDIA's software and hardware within the VAST AI OS, the company enables customers to manage AI agents and data with security, governance, and service tools. As AI moves into real-time, always-on infrastructure, enterprises require systems that evolve with each data point while addressing security and governance for decentralized pipelines and agent interactions. This collaboration enables automated AI workflows from ingestion to inference and retraining within the VAST AI Operating System. CACEIS, one of Europe's largest asset servicing firms, is among the first adopters. In partnership with VAST and NVIDIA, it is exploring a real-time AI platform that: Captures and analyzes 100% of client meetings. Instantly generates meeting minutes and insights. Integrates anonymized data into a secure CRM. The platform features an end-to-end security model to protect privacy and ensure data integrity. CACEIS is also using VAST's AgentEngine with NVIDIA AI-Q Blueprint to develop AI agents that support relationship managers and reveal business opportunities. These agents will use NeMo microservices to refine models through continuous feedback and shared intelligence. Kari Briski, Vice President of Generative AI Software at NVIDIA, showcased this capability at GTC Europe in Paris. The demo highlighted the role of data flywheels in enhancing system intelligence. Briski stated that CACEIS is a pioneer in building agentic AI systems with NVIDIA and VAST to boost productivity in financial services. VAST Data's solution represents a shift from static models to dynamic AI ecosystems. These systems require scalable infrastructure and strict governance to support evolving intelligent agents across enterprise environments.
Yahoo
12-06-2025
- Business
- Yahoo
Nebius Expands AI Cloud in Europe With NVIDIA GB200 and Blackwell
Nebius Group N.V. NBIS has announced the general availability of NVIDIA GB200 Grace Blackwell Superchip capacity for its customers in Europe, marking a significant step in its mission to build out global, full-stack artificial intelligence (AI) infrastructure. The move underscores Nebius' commitment to accelerating AI innovation across Europe and beyond by integrating NVIDIA's powerful computing technology into its AI Cloud platform. In addition to launching GB200 support, Nebius has rolled out several new integrations with NVIDIA's AI ecosystem. These include the expansion of Nebius AI Cloud with NVIDIA AI Enterprise software, early access to NVIDIA Blackwell architecture through NVIDIA DGX Cloud Lepton, and first-in-Europe availability of NVIDIA Hopper GPUs via the Brev platform. Nebius will also introduce instances with the RTX PRO 6000 Blackwell Server Edition later this year. Nebius Group N.V. price-consensus-chart | Nebius Group N.V. Quote Nebius has already put the NVIDIA GB200 platform to work, collaborating with research group LMArena and NVIDIA to bring LMArena's Prompt-to-Leaderboard (P2L) system into production. Trained in just four days using NVIDIA DGX Cloud infrastructure on Nebius, the P2L system dynamically routes real-time user queries to the most accurate and cost-effective AI models, optimizing performance through live feedback. The inclusion of NVIDIA AI Enterprise within Nebius AI Cloud further enhances its offering for enterprise customers and national AI programs. The platform provides tools such as NVIDIA NIM microservices and NVIDIA NeMo, enabling faster development and deployment of AI applications. Nebius' infrastructure is already supporting groundbreaking work. Prima Mente, a biotech firm researching neurological diseases, is using Nebius' AI Cloud to train large-scale epigenetic models. As the first European-headquartered Reference Platform NVIDIA Cloud Partner (NCP), Nebius meets EU regulatory standards, making it ideal for sovereign AI initiatives. One such project is Milestone Systems' Project Hafnia, a collaborative effort between NVIDIA and Nebius to develop a compliant, anonymized video data platform for training vision-language AI models to support smart city solutions. All data is processed and stored within the EU. The company reported first-quarter 2025 results, wherein revenues surged 385% year over year, driven by strong growth in its core AI business. Annualized run-rate revenue, or ARR, saw a 700% surge, highlighting a structurally expanding revenue base. ARR for April was $310 million, which provides a strong start for the second quarter. Nebius is carrying strong momentum into the second quarter of 2025 and remains confident in achieving its full-year ARR guidance of $750 million to $1 billion. For 2025, the company also reaffirmed its overall revenue guidance of $500 million to $700 million. In the first quarter, Nebius significantly upgraded its AI cloud infrastructure through improvements to its Slurm-based cluster. These enhancements included automatic recovery for failed nodes and proactive system health checks designed to identify issues before they impact jobs. This directly lowers downtime and boosts capacity availability. According to the company, these changes led to an estimated 5% improvement in the availability of nodes for commercial use. Nebius is making substantial investments in improving its object storage capabilities, and the upgraded storage system ensures that big data sets can be easily accessed and saved quickly during model training, directly lowering time-to-result for end users. NBIS successfully graduated multiple platform services like MLflow and JupyterLab Notebook from beta to general availability. Nebius expanded integrations with external AI platforms like Metaflow, D Stack and SkyPilot, enabling customers to migrate tools with nominal friction. Nebius currently carries a Zacks Rank #3 (Hold). Shares of the company have surged 64.4% in the past six months compared with the Zacks Internet - Software and Services industry's growth of 25.2%. Image Source: Zacks Investment Research Some better-ranked stocks from the broader technology space are Blackbaud, Inc. BLKB, Criteo S.A. CRTO and Red Violet, Inc. RDVT. BLKB sports a Zacks Rank #1 (Strong Buy), while CRTO and RDVT carry a Zacks Rank #2 (Buy). You can see the complete list of today's Zacks #1 Rank stocks here. Blackbaud's earnings beat the Zacks Consensus Estimate in two of the trailing four quarters while meeting in one and missing in the other, with the average surprise being 1.2%. In the last reported quarter, BLKB delivered an earnings surprise of 6.67%. Its shares have lost 17.4% in the past year. Criteo's earnings beat the Zacks Consensus Estimate in each of the trailing four quarters, with the average surprise being 32.03%. In the last reported quarter, CRTO delivered an earnings surprise of 46.67%. Its shares have decreased 29.4% in the past year. Red Violet's earnings beat the Zacks Consensus Estimate in two of the trailing four quarters while in line in one and missing in one, with the average surprise being 21.63%. In the last reported quarter, RDVT delivered an earnings surprise of 43.48%. Its shares have surged 106.1% in the past year. Want the latest recommendations from Zacks Investment Research? Today, you can download 7 Best Stocks for the Next 30 Days. Click to get this free report Blackbaud, Inc. (BLKB) : Free Stock Analysis Report Criteo S.A. (CRTO) : Free Stock Analysis Report Red Violet, Inc. (RDVT) : Free Stock Analysis Report Nebius Group N.V. (NBIS) : Free Stock Analysis Report This article originally published on Zacks Investment Research ( Zacks Investment Research Error in retrieving data Sign in to access your portfolio Error in retrieving data Error in retrieving data Error in retrieving data Error in retrieving data