News Posts matching #NVIDIA

Return to Keyword Browsing

Alphacool Releases New ES 1-Slot GPU Water Cooler

Alphacool International GmbH from Braunschweig is a pioneer in PC water cooling technology. With one of the industry's most comprehensive product portfolios and over 20 years of experience, Alphacool is now expanding its Enterprise Solutions series with the new ES 1-Slot GPU water cooler for the NVIDIA H200 141 GB - a cooling solution specifically designed for professional use.

The cooler impresses with its compact 1-slot design, making it ideal for use in racks and cases with limited space. The space-saving rear-facing port layout simplifies integration into existing water cooling loops - even under demanding installation conditions. The cooling block is made from high-quality, chrome-plated copper, which is significantly more durable than conventional nickel plating. This combination offers reliable protection against corrosion, scratches, and thermal stress - ideal for continuous 24/7 operation in professional environments.

ASUS Republic of Gamers Unveils ROG Astral GeForce RTX 5080 Dhahab CORE OC Edition

ASUS Republic of Gamers (ROG) today announced the ROG Astral GeForce RTX 5080 Dhahab CORE OC Edition graphics card, built to take style and performance to new frontiers. With the latest NVIDIA GPU architecture, cutting-edge thermal design and a premium aesthetic, the ROG Astral GeForce RTX 5080 Dhahab CORE OC is built for gamers who want a PC that plays well and looks incredible doing it.

The gold standard of GeForce RTX 5080 performance
The ROG Astral GeForce RTX 5080 Dhahab CORE OC Edition graphics card stands ready to let users reap the benefits of the new Blackwell architecture at the heart of the NVIDIA GeForce RTX 50 Series. This delivers fourth-generation ray tracing cores for incredible performance. Users also get NVIDIA DLSS 4 Super Resolution, Multi-Frame Generation and Ray Reconstruction, which help games run smoothly with graphics cranked up.

NVIDIA RTX 50 Series GPUs at MSRP in the Most Unexpected Place: US Navy

The US Navy Exchange (NEX) store has become a surprising platform for acquiring NVIDIA's RTX 50-series graphics cards at their manufacturer's suggested retail prices. A Reddit user, known as Accomplished-Feed123, shared that by combining various store promotions and credit card rewards, they managed to purchase a GeForce RTX 5090 Founders Edition for just $1,900, which is significantly below the typical retail price. Savvy shoppers have long discovered open‑box electronics and gaming hardware bargains there. On this occasion, the Reddit user noticed several "largeish brown boxes" hidden behind a locked display that usually houses Apple products.

Those boxes contained multiple RTX 5070 and RTX 5080 cards, along with a single RTX 5090, all priced at their suggested MSRPs of $550, $999, and $1,999, respectively. A quick online search of the part numbers confirmed that the top‑end card was indeed the Founders Edition model. After applying applicable discounts and card rewards, Accomplished‑Feed123 walked away, paying only $1,900 out of pocket. Access to NEX is restricted to active and retired military members and their families, operating under the motto "You Serve, You Save." However, many consumers may know someone eligible for these benefits. Other branches of the US armed forces maintain similar exchange stores, though GPU availability and pricing may differ by location.

AAEON Unveils World's First 8L Dual-GPU AI Inference Server, the MAXER-5100

Leading provider of advanced AI solutions AAEON has released a new addition to its AI Inference Server product line, the MAXER-5100 - the world's first 8L AI inference server equipped with two integrated GPUs. The MAXER-5100's default model comes with the 24 core, 32 thread 14th Generation Intel Core i9 processor 14900K and two onboard NVIDIA RTX 2000 Ada GPUs. Meanwhile, a barebone SKU is available, accommodating up to 65 W CPUs from across the 12th, 13th, and 14th Generation Intel Core processor lines as well as PCIe slots for other compatible GPUs, per project need.

Given the processing power and AI performance the system offers, the MAXER-5100 is primarily positioned as a central server for the management of multiple edge devices, particularly with its Certificate Authority (CA) support granting additional security for smart infrastructure, healthcare, and advanced manufacturing applications. Moreover, the MAXER-5100's use of a zero-trust secure tunnel and onboard TPM 2.0 allows for encrypted data transmission between the server and multiple edge devices, as well implement over-the-air updates and remote diagnostics.

Borderlands 4 PC Minimum Specs Emerge As Collector's Edition Without Game Key Dampens Spirits

Gearbox and 2K today opened Borderlands 4 pre-orders, with the rather unexpected $69.99 base game price resulting in a simultaneous sigh of relief following the gaming industry's recent move towards $80 AAA games. Following the pre-order launch, Borderlands 4's Steam page revealed the game's minimum system requirements as a bit of a mixed bag. According to the Steam page, the new looter shooter will only run on an AMD Radeon RX 5700 XT or NVIDIA GeForce RTX 2070, but this seems to largely be related to VRAM, since the minimum requirements also mentions that 8 GB is the minimum VRAM amount. An SSD and an eight-core CPU—listed as the AMD Ryzen 2700X or Intel Core i7-9700—will also be mandatory for Borderlands 4. Even the recommended specifications for Borderlands 4 are rather moderate, with the NVIDIA GeForce RTX 3080 and Radeon RX 6800 XT making the cut.

In addition to the reasonable PC specs, news also broke about an ECHO-4 Collector's Edition, a curious edition to the line-up of special edition bundles for Borderlands 4 that seems to be exclusive to GameStop for now. The Echo-4 Collector's Edition bundle is a host of physical Borderlands 4 memorabilia, including a cloth map of Kairos, an eight-inch Echo-4 figure with LED lights, a Vault symbol necklace, a card with five SHiFT codes for Borderlands 4 golden keys, and four double-sided lithography art prints—all for $149.99. The one glaring omission from the Echo-4 bundle is an actual game key for Borderlands 4. That is clearly labelled as "sold separately" on the GameStop page.

Shadow Launches Neo: The Next Generation Cloud Gaming PC

SHADOW, the Global leader in high-performance cloud computing, is proud to announce the launch of Neo, a brand-new cloud gaming PC offering designed to deliver next-level RTX experiences for gamers, creators, and professionals alike. Neo will officially roll out in Europe and North America starting June 16, 2025.

Building on the success of the company's previous offers, Neo replaces its widely adopted "Boost" tier and delivers major performance leaps—up to 150% more in gaming and 200% more in pro software performance. All existing Boost users are being upgraded to Neo at no additional cost while new users rates will start at $37.99 per month.

Biostar Launches NVIDIA-based Edge AI Developer Kit, AI-NONXS

BIOSTAR, a leading manufacturer of edge computing solutions, motherboards, graphics cards, and storage devices, is excited to introduce the AI-NONXS Developer Kit, a high-performance edge AI development platform engineered to accelerate edge AI deployment for system integrators and AI software developers. Designed to support NVIDIA Jetson Orin NX and Orin Nano series modules, the AI-NONXS Developer Kit empowers users to efficiently build and deploy AI-powered solutions across a broad range of real-world applications. From smart manufacture, smart retail, and automated warehouses to smart cities, transportation, and smart agriculture, the AI-NONXS provides the computing power, flexibility, and reliability required for advanced edge systems.

The AI-NONXS Developer Kit comes with a comprehensive set of industrial-grade interfaces, including 4 USB 3.2 Gen 2 Type-A ports, 1 Micro-USB port, dual Gigabit Ethernet (1 x 1GbE and 1 x 2.5GbE with optional Power over Ethernet), and RS232/422/485, CANBUS, TPM (opt.), audio Line-Out/MIC-In and Speaker out. It operates within a wide temperature range of -20°C to 60°C and supports a flexible power input range of 12 V to 20 V, ensuring reliable performance in harsh environments. For visual input and expansion capabilities, the kit includes HDMI 2.1 (for Orin NX) or HDMI 1.4 (for Orin Nano), dual MIPI CSI connectors, and three M.2 slots (Key-M, Key-E, and Key-B) for storage, wireless communication, and modular upgrades.

Next‑Gen HBM4 to HBM8: Toward Multi‑Terabyte Memory on 15,000 W Accelerators

In a joint briefing this week, KAIST's Memory Systems Laboratory and TERA's Interconnection and Packaging group presented a forward-looking roadmap for High Bandwidth Memory (HBM) standards and the accelerator platforms that will employ them. Shared via Wccftech and VideoCardz, the outline covers five successive generations, from HBM4 to HBM8, each promising substantial gains in capacity, bandwidth, and packaging sophistication. First up is HBM4, targeted for a 2026 rollout in AI GPUs and data center accelerators. It will deliver approximately 2 TB/s per stack at an 8 Gbps pin rate over a 2,048-bit interface. Die stacks will reach 12 to 16 layers, yielding 36-48 GB per package with a 75 W power envelope. NVIDIA's upcoming Rubin series and AMD's Instinct MI500 cards are slated to employ HBM4, with Rubin Ultra doubling the number of memory stacks from eight to sixteen and AMD targeting up to 432 GB per device.

Looking to 2029, HBM5 maintains an 8 Gbps speed but doubles the I/O lanes to 4,096 bits, boosting throughput to 4 TB/s per stack. Power rises to 100 W and capacity scales to 80 GB using 16‑high stacks of 40 Gb dies. NVIDIA's tentative Feynman accelerator is expected to be the first HBM5 adopter, packing 400-500 GB of memory into a multi-die package and drawing more than 4,400 W of total power. By 2032, HBM6 will double pin speeds to 16 Gbps and increase bandwidth to 8 TB/s over 4,096 lanes. Stack heights can grow to 20 layers, supporting up to 120 GB per stack at 120 W. Immersion cooling and bumpless copper-copper bonding will become the norm. The roadmap then predicts HBM7 in 2035, which includes 24 Gbps speeds, 8,192-bit interfaces, 24 TB/s throughput, and up to 192 GB per stack at 160 W. NVIDIA is preparing a 15,360 W accelerator to accommodate this monstrous memory.

Robust AI Demand Drives 6% QoQ Growth in Revenue for Top 10 Global IC Design Companies in 1Q25

TrendForce's latest investigations reveal that 1Q25 revenue for the global IC design industry reached US$77.4 billion, marking a 6% QoQ increase and setting a new record high. This growth was fueled by early stocking ahead of new U.S. tariffs on electronics and the ongoing construction of AI data centers around the world, which sustained strong chip demand despite the traditional off-season.

NVIDIA remained the top-ranking IC design company, with Q1 revenue surging to $42.3 billion—up 12% QoQ and 72% YoY—thanks to increasing shipments of its new Blackwell platform. Although its H20 chip is constrained by updated U.S. export controls and is expected to incur losses in Q2, the higher-margin Blackwell is poised to replace the Hopper platform gradually, cushioning the financial impact.

Premiere Pro June 2025 Update Adds NVIDIA Blackwell 4:2:2 Video GPU Acceleration Support

The June 2025 Update of Adobe Premiere Pro, the industry standard video editing application, added support for GPU-accelerated 4:2:2 color format video encoding and decoding on NVIDIA "Blackwell" GPUs. It implements Video Codec SDK 13.0, which NVIDIA released in February. This adds 4:2:2 color format support for H.264 and HEVC. The 4:2:2 color format provides twice the amount of color detail as 4:2:0, but with a nominal increase in stream bitrate or file-size. Adobe says that the added color detail of 4:2:2 is ideal for "precise color grading, cleaner chroma keying, and crisper text rendering."

NVIDIA GeForce NOW Gets 40% Off Performance Day Passes, More Games

Level up GeForce NOW experiences this summer with 40% off Performance Day Passes. Enjoy 24 hours of premium cloud gaming with RTX ON, delivering low latency and shorter wait times. The hot deal comes just in time for the cloud's highly anticipated launch of Dune: Awakening - a multiplayer survival game on a massive scale set on the unforgiving sands of Arrakis. It's perfect to pair with the nine games available this week, including the Frosthaven demo announced at Steam Next Fest.

Try Before You Buy
Level up to the cloud, no commitment required. For a limited time, grab a Performance Day Pass at a price that's less than an ice cream sundae and experience premium GeForce NOW gaming for 24 hours. With RTX ON, enjoy shorter wait times and lower latency for supported games, all powered by the cloud. Dive into popular games with upgraded visuals and smoother gameplay over free users, whether exploring vast open worlds or battling in fast-paced arenas. Take the experience even further by applying the value of the Day Pass toward a six-month Performance membership during the limited-time summer sale. It's the perfect way to try out premium cloud gaming before jumping into a longer-term membership.

NVIDIA TensorRT Boosts Stable Diffusion 3.5 Performance on NVIDIA GeForce RTX and RTX PRO GPUs

Generative AI has reshaped how people create, imagine and interact with digital content. As AI models continue to grow in capability and complexity, they require more VRAM, or video random access memory. The base Stable Diffusion 3.5 Large model, for example, uses over 18 GB of VRAM - limiting the number of systems that can run it well. By applying quantization to the model, noncritical layers can be removed or run with lower precision. NVIDIA GeForce RTX 40 Series and the Ada Lovelace generation of NVIDIA RTX PRO GPUs support FP8 quantization to help run these quantized models, and the latest-generation NVIDIA Blackwell GPUs also add support for FP4.

NVIDIA collaborated with Stability AI to quantize its latest model, Stable Diffusion (SD) 3.5 Large, to FP8 - reducing VRAM consumption by 40%. Further optimizations to SD3.5 Large and Medium with the NVIDIA TensorRT software development kit (SDK) double performance. In addition, TensorRT has been reimagined for RTX AI PCs, combining its industry-leading performance with just-in-time (JIT), on-device engine building and an 8x smaller package size for seamless AI deployment to more than 100 million RTX AI PCs. TensorRT for RTX is now available as a standalone SDK for developers.

TSMC Prepares "CoPoS": Next-Gen 310 × 310 mm Packages

As demand for ever-growing AI compute power continues to rise and manufacturing advanced nodes becomes more difficult, packaging is undergoing its golden era of development. Today's advanced accelerators often rely on TSMC's CoWoS modules, which are built on wafer cuts measuring no more than 120 × 150 mm in size. In response to the need for more space, TSMC has unveiled plans for CoPoS, or "Chips on Panel on Substrate," which could expand substrate dimensions to 310 × 310 mm and beyond. By shifting from round wafers to rectangular panels, CoPoS offers more than five times the usable area. This extra surface makes it possible to integrate additional high-bandwidth memory stacks, multiple I/O chiplets and compute dies in a single package. It also brings panel-level packaging (PLP) to the fore. Unlike wafer-level packaging (WLP), PLP assembles components on large, rectangular panels, delivering higher throughput and lower cost per unit. Systems with PLP will be actually viable for production runs and allow faster iterations over WLP.

TSMC will establish a CoPoS pilot line in 2026 at its Visionchip subsidiary. In 2027, the pilot facility will focus on refining the process, to meet partner requirements by the end of the year. Mass production is projected to begin between the end of 2028 and early 2029 at TSMC's Chiayi AP7 campus. That site, chosen for its modern infrastructure and ample space, is also slated to host production of multi-chip modules and System-on-Wafer technologies. NVIDIA is expected to be the launch partner for CoPoS. The company plans to leverage the larger panel area to accommodate up to 12 HBM4 chips alongside several GPU chiplets, offering significant performance gains for AI workloads. At the same time, AMD and Broadcom will continue using TSMC's CoWoS-L and CoWoS-R variants for their high-end products. Beyond simply increasing size, CoPoS and PLP may work in tandem with other emerging advances, such as glass substrates and silicon photonics. If development proceeds as planned, the first CoPoS-enabled devices could reach the market by late 2029.

MAINGEAR Unleashes ULTIMA 18 - The Ultimate 18" 4K Gaming Laptop

MAINGEAR, the leader in premium-quality, high-performance gaming PCs, today announced its most powerful laptop to date, the 18-inch ULTIMA 18. Developed in collaboration with CLEVO, ULTIMA 18 redefines what a gaming laptop can be by offering desktop-level specs, like a 4K@200 Hz G-SYNC display, Intel Core Ultra 9 275HX processor, and up to an NVIDIA GeForce RTX 5090 mobile GPU, all inside a sleek chassis outfitted with metal lid and palm rest.

Designed for elite gamers and creators who demand top-tier performance without compromise, ULTIMA 18 is MAINGEAR's first laptop to support modern dual-channel DDR5 memory, PCIe Gen 5 SSDs, dual Thunderbolt 5 ports, and Wi-Fi 7. Whether plugged in or on the move, this system delivers unprecedented power, quiet efficiency, and immersive visuals for the most demanding workloads and graphics-rich game titles.

AMD Instinct MI355X Draws up to 1,400 Watts in OAM Form Factor

Tomorrow evening, AMD will host its "Advancing AI" livestream to introduce the Instinct MI350 series, a new line of GPU accelerators designed for large-scale AI training and inference. First shown in prototype form at ISC 2025 in Hamburg just a day ago, each MI350 card features 288 GB of HBM3E memory, delivering up to 8 TB/s of sustained bandwidth. Customers can choose between the single-card MI350X and the higher-clocked MI355X or opt for a full eight-GPU platform that aggregates to over 2.3 TB of memory. Both chips are built on the CDNA 4 architecture, which now supports four different precision formats: FP16, FP8, FP6, and FP4. The addition of FP6 and FP4 is designed to boost throughput in modern AI workloads, where models of tomorrow with tens of trillions of parameters are trained on FP6 and FP4.

In half-precision tests, the MI350X achieves 4.6 PetaFLOPS on its own and 36.8 PetaFLOPS in eight-GPU platform form, while the MI355X surpasses those numbers, reaching 5.03 PetaFLOPS and just over 40 PetaFLOPS. AMD is also aiming to improve energy efficiency by a factor of thirty compared with its previous generation. The MI350X card runs within a 1,000 Watt power envelope and relies on air cooling, whereas the MI355X steps up to 1,400 Watts and is intended for direct-liquid cooling setups. That 400 Watt increase puts it right at NVIDIA's upcoming GB300 "Grace Blackwell Ultra" superchip, which is also a 1,400 W design. With memory capacity, raw computing, and power efficiency all pushed to new heights, the question remains whether real-world benchmarks will match these ambitious specifications. AMD now only lacks platform scaling beyond eight GPUs, which the Instinct MI400 series will address.

NVIDIA NVL72 GB200 Systems Accelerate the Journey to Useful Quantum Computing

The integration of quantum processors into tomorrow's supercomputers promises to dramatically expand the problems that can be addressed with compute—revolutionizing industries including drug and materials development.

In addition to being part of the vision for tomorrow's hybrid quantum-classical supercomputers, accelerated computing is dramatically advancing the work quantum researchers and developers are already doing to achieve that vision. And in today's development of tomorrow's quantum technology, NVIDIA GB200 NVL72 systems and their fifth-generation multinode NVIDIA NVLink interconnect capabilities have emerged as the leading architecture.

Europe Builds AI Infrastructure With NVIDIA to Fuel Region's Next Industrial Transformation

NVIDIA today announced it is working with European nations, and technology and industry leaders, to build NVIDIA Blackwell AI infrastructure that will strengthen digital sovereignty, support economic growth and position the continent as a leader in the AI industrial revolution. France, Italy, Spain and the U.K. are among the nations building domestic AI infrastructure with an ecosystem of technology and cloud providers, including Domyn, Mistral AI, Nebius and Nscale, and telecommunications providers, including Orange, Swisscom, Telefónica and Telenor.

These deployments will deliver more than 3,000 exaflops of NVIDIA Blackwell compute resources for sovereign AI, enabling European enterprises, startups and public sector organizations to securely develop, train and deploy agentic and physical AI applications. NVIDIA is establishing and expanding AI technology centers in Germany, Sweden, Italy, Spain, the U.K. and Finland. These centers build on NVIDIA's history of collaborating with academic institutions and industry through the NVIDIA AI Technology Center program and NVIDIA Deep Learning Institute to develop the AI workforce and scientific discovery throughout the regions.

NVIDIA Partners With Europe Model Builders and Cloud Providers to Accelerate Region's Leap Into AI

NVIDIA GTC Paris at VivaTech -- NVIDIA today announced that it is teaming with model builders and cloud providers across Europe and the Middle East to optimize sovereign large language models (LLMs), providing a springboard to accelerate enterprise AI adoption for the region's industries.

Model builders and AI consortiums Barcelona Supercomputing Center (BSC), Bielik.AI, Dicta, H Company, Domyn, LightOn, the National Academic Infrastructure for Supercomputing in Sweden (NAISS) together with KBLab at the National Library of Sweden, the Slovak Republic, the Technology Innovation Institute (TII), the University College of London, the University of Ljubljana and UTTER are teaming with NVIDIA to optimize their models with NVIDIA Nemotron techniques to maximize cost efficiency and accuracy for enterprise AI workloads, including agentic AI.

Pegatron Unveils AI-Optimized Server Innovations at GTC Paris 2025

PEGATRON, a globally recognized Design, Manufacturing, and Service (DMS) provider, is showcasing its latest AI server solutions at GTC Paris 2025. Built on NVIDIA Blackwell architecture, PEGATRON's cutting-edge systems are tailored for AI training, reasoning, and enterprise-scale deployment.

NVIDIA GB300 NVL72
At the forefront is the RA4802-72N2, built on the NVIDIA GB300 NVL72 rack system, featuring 72 NVIDIA Blackwell Ultra GPUs and 36 NVIDIA Grace CPUs. Designed for AI factories, it boosts output by up to 50X. PEGATRON's in-house developed Coolant Distribution Unit (CDU) delivers 310 kW of cooling capacity with redundant hot-swappable pumps, ensuring performance and reliability for mission-critical workloads.

NVIDIA DLSS 4 With Multi Frame Generation Coming To Another 8 Games, Including Stellar Blade and Dune: Awakening

More than 700 games and applications feature RTX technologies, and each week new games integrating NVIDIA DLSS, NVIDIA Reflex, and advanced ray-traced effects are released or announced, delivering the definitive PC experience for GeForce RTX players. DLSS 4 with Multi Frame Generation's adoption continues at rapid pace, and is the fastest-adopted NVIDIA gaming technology in our history.

This week, we can reveal another 8 games equipped with the frame rate multiplying technology, including Dragonkin: The Banished, Dune: Awakening, Marvel's Spider-Man 2, Mecha BREAK, MindsEye, REMATCH, and Stellar Blade. And Indiana Jones and the Great Circle: The Order of Giants DLC is launching September 4th, featuring full ray tracing and DLSS 4 with Multi Frame Generation and DLSS Ray Reconstruction. Each title also features DLSS Frame Generation, boosting frame rates for GeForce RTX 40 Series gamers, and DLSS Super Resolution, accelerating performance on every GeForce RTX GPU.

Lenovo Announces the All-New Workstations Solutions and Updates to the ThinkStation Desktop Portfolio

Lenovo, today at NXT BLD, announced its new portfolio of Workstation Solutions, a series of purpose-built, expertly-curated industry solutions that meet and exceed the rigorous performance and workflow requirements of engineers, designers, architects, data scientists, researchers, and creators so these power users can work smarter, faster, and more cost-effectively. Lenovo also unveiled the latest editions of its newest ThinkStation P2 and P3 desktop workstations designed to maximize performance and value.

Lenovo Workstations Solutions—Your Workflow, Perfected
Businesses need more than just powerful hardware—they need complete workflow solutions tailored to real-world industry challenges. Developed by Lenovo engineering experts through research and customer engagement to understand workflow bottlenecks and pain points, Lenovo Workstation Solutions deliver easily deployable blueprints—scalable and secure reference architectures powered by the state-of-the-art Lenovo Workstations—featuring superior hardware, software and services.

MSI Powers AI's Next Leap for Enterprises at ISC 2025

MSI, a global leader in high-performance server solutions, is showcasing its enterprise-grade, high-performance server platforms at ISC 2025, taking place June 10-12 at booth #E12. Built on standardized and modular architectures, MSI's AI servers are designed to power next-generation AI and accelerated computing workloads, enabling enterprises to rapidly advance their AI innovations.

"As AI workloads continue to grow and evolve toward inference-driven applications, we're seeing a significant shift in how enterprises approach AI deployment," said Danny Hsu, General Manager of Enterprise Platform Solutions at MSI. "With modular and standards-based architectures, enterprise data centers can now adopt AI technologies more quickly and cost-effectively than ever before. This marks a new era where AI is not only powerful but also increasingly accessible to businesses of all sizes.

ASUS Announces Key Milestone with Nebius and Showcases NVIDIA GB300 NVL72 System at GTC Paris 2025

ASUS today joined GTC Paris at VivaTech 2025 as a Gold Sponsor, highlighting its latest portfolio of AI infrastructure solutions and reinforcing its commitment to advancing the AI Factory vision with a full range of NVIDIA Blackwell Ultra solutions, delivering breakthrough performance from large-scale datacenter to personal desktop.

ASUS is also excited to announce a transformative partnership milestone in its partnership with Nebius. Together, the two companies are enabling a new era of AI innovation built on NVIDIA's advanced platforms. Building on the success of the NVIDIA GB200 NVL72 platform deployment, ASUS and Nebius are now moving forward with strategic collaborations featuring the next-generation NVIDIA GB300 NVL72 platform. This ongoing initiative underscores ASUS's role as a key enabler in AI infrastructure, committed to delivering scalable, high-performance solutions that help enterprises accelerate AI adoption and innovation.

Micron Ships HBM4 Samples: 12-Hi 36 GB Modules with 2 TB/s Bandwidth

Micron has achieved a significant advancement of the HBM4 architecture, which will stack 12 DRAM dies (12-Hi) to provide 36 GB of capacity per package. According to company representatives, initial engineering samples are scheduled to ship to key partners in the coming weeks, paving the way for full production in early 2026. The HBM4 design relies on Micron's established 1β ("one-beta") process node for DRAM tiles, in production since 2022, while it prepares to introduce EUV-enabled 1γ ("one-gamma") later this year for DDR5. By increasing the interface width from 1,024 to 2,048 bits per stack, each HBM4 chip can achieve a sustained memory bandwidth of 2 TB/s, representing a 20% efficiency improvement over the existing HBM3E standard.

NVIDIA and AMD are expected to be early adopters of Micron's HBM4. NVIDIA plans to integrate these memory modules into its upcoming Rubin-Vera AI accelerators in the second half of 2026. AMD is anticipated to incorporate HBM4 into its next-generation Instinct MI400 series, with further information to be revealed at the company's Advancing AI 2025 conference. The increased capacity and bandwidth of HBM4 will address growing demands in generative AI, high-performance computing, and other data-intensive applications. Larger stack heights and expanded interface widths enable more efficient data movement, a critical factor in multi-chip configurations and memory-coherent interconnects. As Micron begins mass production of HBM4, major obstacles to overcome will be thermal performance and real-world benchmarks, which will determine how effectively this new memory standard can support the most demanding AI workloads.
Micron HBM4 Memory

NVIDIA and HPE Join Forces to Construct Advanced Supercomputer in Germany

NVIDIA and Hewlett Packard Enterprise announced Tuesday at a supercomputing conference in Hamburg their partnership with Germany's Leibniz Supercomputing Centre to build a new supercomputer called Blue Lion which will deliver approximately 30 times more computing power than the current SuperMUC-NG system. The Blue Lion supercomputer will run on NVIDIA's upcoming Vera Rubin architecture. This setup combines the Rubin GPU with NVIDIA's first custom CPU Vera. The integrated system aims to unite simulation, data processing, and AI into one high-bandwidth low-latency platform. Optimized to support scientific research it boasts shared memory coherent compute abilities, and in-network acceleration.

HPE will build the system using its next-gen Cray technology by including NVIDIA GPUs along with cutting-edge storage and interconnect systems. The Blue Lion supercomputer will use HPE's 100% fanless direct liquid-cooling setup. This design circulates warm water through pipes for efficient cooling while the generated system's heat output will be reused to warm buildings nearby. The Blue Lion project comes after NVIDIA said Lawrence Berkeley National Lab in the US will also set up a Vera Rubin-powered system called Doudna next year. Scientists will have access to the Blue Lion supercomputer beginning in early 2027. The Blue Lion supercomputer, based in Germany will be utilized by researchers working on climate, physics, and machine learning. In contrast, Doudna, the U.S. Department of Energy's next supercomputer, will get its data from telescopes, genome sequencers, and fusion experiments.
Return to Keyword Browsing
Jun 19th, 2025 04:34 CDT change timezone

New Forum Posts

Popular Reviews

TPU on YouTube

Controversial News Posts