Latest news with #InstinctMI355X


Techday NZ
5 days ago
- Business
- Techday NZ
Vultr launches early access to AMD Instinct MI355X GPU for AI
Vultr has announced the availability of the AMD Instinct MI355X GPU as part of its cloud infrastructure services. As one of the first cloud providers to integrate the new AMD Instinct MI355X GPU, Vultr is now taking pre-orders for early access, with global availability scheduled for the third quarter of the year. The GPU forms part of AMD's latest focus on high-capacity computational demands, catering to artificial intelligence (AI) workloads as well as enterprise-scale applications. Product features The AMD Instinct MI355X GPU is based on AMD's 4th Generation CDNA architecture. According to Vultr, this GPU features 288 GB of HBM3E memory, delivers up to 8 TB/s of memory bandwidth, and supports expanded datatypes such as FP6 and FP4. These improvements are designed to address complex tasks ranging from AI training and inference to scientific simulations within high-performance computing (HPC) environments. For customers operating within higher-density data environments, the Instinct MI355X supports direct liquid cooling (DLC). This enhancement offers increased thermal efficiency, which is intended to unlock greater computing performance per rack and facilitate advanced, scalable cooling strategies. The GPU is also supported by the latest version of AMD's ROCm software, which further optimises tasks related to AI inference, training, and compatibility with various frameworks. This results in improved throughput and reduced latency for critical operations. AMD and Vultr partnership Vultr's portfolio already includes other AMD offerings, such as the AMD EPYC 9004 Series and EPYC 7003 Series central processing units (CPUs), as well as previous GPU models like the Instinct MI325X and MI300X. Customers using the MI355X in combination with AMD EPYC 4005 Series CPUs will benefit from a fully supported computing stack across both processing and acceleration functions, streamlining high-powered workloads from end to end. Negin Oliver, Corporate Vice President of Business Development, Data Centre GPU Business at AMD, stated: "AMD is the trusted AI solutions provider of choice, enabling customers to tackle the most ambitious AI initiatives, from building large-scale AI cloud deployments to accelerating AI-powered scientific discovery. AMD Instinct MI350 series GPUs paired with AMD ROCm software provide the performance, flexibility, and security needed to deliver tailored AI solutions that meet the diverse demands of the modern AI landscape." The collaboration builds on Vultr's efforts to support a range of AMD solutions tailored for enterprise, HPC, and AI sectors, reinforcing the company's capacity to cater to evolving customer workloads. Cloud market implications J.J. Kardwell, Chief Executive Officer of Vultr, highlighted the alignment of the new GPU with market requirements. Kardwell commented: "AMD MI355X GPUs are designed to meet the diverse and complex demands of today's AI workloads, delivering exceptional value and flexibility. As AI development continues to accelerate, the scalability, security, and efficiency these GPUs deliver are more essential than ever. We are proud to be among the first cloud providers worldwide to offer AMD MI355X GPUs, empowering our customers with next-generation AI infrastructure." AMD is recognised as a member of the Vultr Cloud Alliance, which supports a collaborative ecosystem of technology providers focused on offering integrated cloud computing solutions. The introduction of the MI355X GPU follows a period of upgrades across AMD's GPU lineup, including a greater emphasis on catering to both inferencing and enterprise-scale workloads. Vultr's offering is aimed at organisations seeking advanced compute resources for AI-driven applications and scientific tasks requiring significant computational capacity. Vultr's global network reportedly serves hundreds of thousands of customers across 185 countries, supplying services in cloud compute, GPU, bare metal infrastructure and cloud storage. The addition of AMD's latest GPU to its infrastructure underlines Vultr's commitment to providing a variety of options for businesses and developers pursuing AI and HPC advancements.


Hindustan Times
16-06-2025
- Business
- Hindustan Times
AMD Advancing AI event: 6 key details you need to know
At AMD's Advancing AI event in San Jose, California, United States, the brand unveiled a range of hardware as well as software-centric announcements, which at large serves as the vision for its open AI ecosystem. Here are the key announcements: AMD Instinct MI350 series GPUs announced: AMD announced the Instinct MI350X and Instinct MI355X GPUs and platforms, which the company claims allow for a four times generation-on-generation AI compute boost for better AI solutions across various industries. End-to-end open standards rack-scale infrastructure: At the keynote, AMD also showed its open standards rack-scale AI infrastructure, which is already rolling out with AMD Instinct MI350 series accelerators, 5th Gen AMD EPYC processors, and AMD Pensando Pollara NICs in hyperscaler deployments, including Oracle's Cloud Infrastructure. The company says the broad availability for the same is going to be the second half of 2025. AMD unveils its next-generation AI rack called Helios: AMD stated that this is going to be built on the next-gen AMD Instinct MI400 series GPUs. When you compare this to the former generation, it can allow for 10 times more performance when running inference on mixture-of-experts models. AMD announced the broad availability of its Developer Cloud: The company says this is purpose-built for fast, high-performance AI development and allows for developers to have access to a fully managed cloud environment with tools, allowing for fast development. The company says the combination of ROCm 7 and the AMD Developer Cloud enables the company to expand access to next-generation compute. It is already collaborating with AI leaders like Hugging Face, OpenAI, and Grok. AMD's latest version of its AI software stack, ROCm 7: The company says that this is to serve the growing demands of generative AI and the demand for more compute. The company says that this is going to enhance the experience for developers and features improved support for industry-standard frameworks and more. AMD also revealed its partner ecosystem: Seven out of ten of the largest AI model builders, including the likes of Meta, OpenAI, Microsoft, and xAI, are now on board with AMD for training their AI models, the company said. For instance, it was detailed how the Instinct MI300X is used for Llama 3 and for Llama 4 inference. Sam Altman, OpenAI CEO, also discussed hardware, software, and more. Other companies like Microsoft and Red Hat also joined them. MOBILE FINDER: iPhone 16 LATEST Price, Specs And More
Yahoo
13-06-2025
- Business
- Yahoo
AMD (AMD) Unveils 1,400W MI355X AI GPU to Challenge Nvidia's Blackwell
AMD (AMD, Financials) officially launched its Instinct MI355X GPU accelerator Wednesday, showcasing a massive leap in compute power and energy demands as it competes with Nvidia's Blackwell Ultra B300. Warning! GuruFocus has detected 3 Warning Signs with AMD. The MI355X is part of AMD's new CDNA 4 architecture and is optimized for AI inference. With support for FP4, FP6, FP8, and FP16 precision, the MI355X delivers up to 20.1 PFLOPS in FP4/FP6 workloads and 10.1 PFLOPS in FP8, slightly ahead of Nvidia's B300 at 15 FP4 PFLOPS. To support this performance, the MI355X consumes 1,400W peak, nearly doubling the 750W required by its predecessor, the MI300X. AMD expects some users may still air-cool the chip, but liquid cooling is the standard. The GPU includes 288 GB of HBM3E memory with bandwidth reaching 8 TB/s. A scaled 8-way configuration brings system-level performance to 161 PFLOPS (FP4) and 80.5 PFLOPS. While raw compute marks a win on paper, AMD still trails Nvidia in deployment scale and software ecosystem. Pegatron is reportedly preparing a 128-way MI350X system, but Nvidia remains dominant in large-scale AI training clusters. AMD's Chief Technology Officer Mark Papermaster said zettascale supercomputing by 2035 will require processors consuming up to 2,000W each. He projected that future AI systems may need nuclear-scale powerup to 500 MW per machine. This article first appeared on GuruFocus. Error in retrieving data Sign in to access your portfolio Error in retrieving data Error in retrieving data Error in retrieving data Error in retrieving data

Yahoo
13-06-2025
- Business
- Yahoo
AMD gains on Nvidia? Lisa Su reveals new chips in heated AI inference race
-- Advanced Micro Devices Inc (NASDAQ:AMD) made an aggressive bid for dominance in AI inference at its Advancing AI event Thursday, unveiling new chips that directly challenge NVIDIA Corporation's (NASDAQ:NVDA) supremacy in the data center GPU market. AMD claims its latest Instinct MI355X accelerators surpass Nvidia's most advanced Blackwell GPUs in inference performance while offering a significant cost advantage, a critical selling point as hyperscalers look to scale generative AI services affordably. The MI355X, which has just begun volume shipments, delivers a 35-fold generational leap in inference performance and, according to AMD, up to 40% more tokens-per-dollar compared to Nvidia's flagship chips. That performance boost, coupled with lower power consumption, is designed to help AMD undercut Nvidia's offerings in total cost of ownership at a time when major AI customers are re-evaluating procurement strategies. 'What has really changed is the demand for inference has grown significantly,' AMD CEO Lisa Su said at the event in San Jose. 'It says that we have really strong hardware, which we always knew, but it also shows that the open software frameworks have made tremendous progress.' AMD's argument hinges not just on silicon performance, but on architecture and economics. By pairing its GPUs with its own CPUs and networking chips inside open 'rack-scale' systems, branded Helios, AMD is building full-stack solutions to rival Nvidia's proprietary end-to-end ecosystem. These systems, launching next year with the MI400 series, were designed to enable hyperscale inference clusters while reducing energy and infrastructure costs. Su highlighted how companies like OpenAI, Meta Platforms Inc (NASDAQ:META), and Microsoft Corporation (NASDAQ:MSFT) are now running inference workloads on AMD chips, with OpenAI CEO Sam Altman confirming a close partnership on infrastructure innovation. 'It's gonna be an amazing thing,' Altman said during the event. 'When you first started telling me about the specs, I was like, there's no way, that just sounds totally crazy.' Oracle Corporation (NYSE:ORCL) Cloud Infrastructure intends to offer massive clusters of AMD chips, with plans to deploy up to 131,072 MI355X GPUs, positioning AMD as a scalable alternative to Nvidia's tightly integrated, and often more expensive, solutions. AMD officials emphasized the cost benefits, asserting that customers could achieve double-digit percent savings on power and capital expenditures when compared with Nvidia's GPUs. Despite the positive news, AMD shares were down roughly 2% ahead of market close. Wall Street remains cautious, but AMD's moves suggest it is committed to challenging Nvidia's leadership not only with performance parity, but also with a differentiated value and systems strategy. While Nvidia still commands more than 90% of the data center AI chip market, AMD's targeted push into inference, where workloads demand high efficiency and lower costs, marks a strategic front in the battle for AI dominance. With generative AI models driving a surge in inference demand across enterprises, AMD is betting that performance per dollar will matter more than ever. Related articles AMD gains on Nvidia? Lisa Su reveals new chips in heated AI inference race Wall Street weighs in on Boeing after digesting deadly Air India crash SharpLink Gaming stock tumbles after filing to offer common stock, warrants