News Posts matching #GPU

Return to Keyword Browsing

AMD Releases Software Adrenalin 24.4.1 WHQL GPU Drivers

AMD has released the latest version of Adrenalin Edition graphics drivers, version 24.4.1 WHQL. It includes support for the upcoming Manor Lords game, as well as add performance improvements for HELLDIVERS 2 game, and adds AMD HYPR-Tune support to Nightingale and SKULL AND BONES games. New drivers also expand Vulkan API extensions support with VK_KHR_shader_maximal_reconvergence and VK_KHR_dynamic_rendering_local_read, as well as include support and optimizations for Topaz Gigapixel AI application, versions 7.1.0 and 7.1.1, with new "Recovery" and "Low Resolution" AI upscaling features.

New AMD Software Adrenalin Edition 24.4.1 WHQL drivers come with several fixes, including performance improvements for HELLDIVERS 2, fix for intermittent application crash in Lords of the Fallen on Radeon RX 6000 series graphics cards, various artifact issues in SnowRunner and Horizon Forbidden West Complete Edition on Radeon RX 6800 and Radeon RX 6000 series graphics cards, fix for intermittent application crash or driver timeout in Overwatch 2 when Radeon Boost is enabled on Radeon RX 6000 and above series graphics cards, intermittent crash while changing Anti-Aliasing settings in Enshrouded on Radeon 7000 series graphics cards, and various application freeze or crash issues with the SteamVR while using Quest Link on Meta Quest 2 or when screen sharing with Microsoft Teams.

DOWNLOAD: AMD Software Adrenalin 24.4.1 WHQL

Aetina Accelerates Embedded AI with High-performance, Small Form-factor Aetina IA380E-QUFL Graphics Card

Aetina, a leading Edge AI solution provider, announced the launch of the Aetina IA380E-QUFL at Embedded World 2024 in Nuremberg, Germany. This groundbreaking product is a small form factor PCIe graphics card powered by the high-performance Intel Arc A380E GPU.

Unmatched Power in a Compact Design
The Aetina IA380E-QUFL delivers workstation-level performance packed into a low-profile, single-slot form factor. This innovative solution consumes only 50 W, making it ideal for space and power-constrained edge computing environments. Embedded system manufacturers and integrators can leverage the power of 4.096 TFLOPs peak FP32 performance delivered by the Intel Arc A380E GPU.

Unreal Engine 5.4 is Now Available With Improvements to Nanite, AI and Machine Learning, TSR, and More

Unreal Engine 5.4 is here, and it's packed with new features and improvements to performance, visual fidelity, and productivity that will benefit game developers and creators across industries. With this release, we're delivering the toolsets we've been using internally to build and ship Fortnite Chapter 5, Rocket Racing, Fortnite Festival, and LEGO Fortnite. Here are some of the highlights.

Animation
Character rigging and animation authoring
This release sees substantial updates to Unreal Engine's built-in animation toolset, enabling you to quickly, easily, and enjoyably rig characters and author animation directly in engine, without the frustrating and time-consuming need to round trip to external applications. With an Experimental new Modular Control Rig feature, you can build animation rigs from understandable modular parts instead of complex granular graphs, while Automatic Retargeting makes it easier to get great results when reusing bipedal character animations. There are also extensions to the Skeletal Editor and a suite of new deformer functions to make the Deformer Graph more accessible.

AMD's RDNA 4 GPUs Could Stick with 18 Gbps GDDR6 Memory

Today, we have the latest round of leaks that suggest that AMD's upcoming RDNA 4 graphics cards, codenamed the "RX 8000-series," might continue to rely on GDDR6 memory modules. According to Kepler on X, the next-generation GPUs from AMD are expected to feature 18 Gbps GDDR6 memory, marking the fourth consecutive RDNA architecture to employ this memory standard. While GDDR6 may not offer the same bandwidth capabilities as the newer GDDR7 standard, this decision does not necessarily imply that RDNA 4 GPUs will be slow performers. AMD's choice to stick with GDDR6 is likely driven by factors such as meeting specific memory bandwidth requirements and cost optimization for PCB designs. However, if the rumor of 18 Gbps GDDR6 memory proves accurate, it would represent a slight step back from the 18-20 Gbps GDDR6 memory used in AMD's current RDNA 3 offerings, such as the RX 7900 XT and RX 7900 XTX GPUs.

AMD's first generation RDNA used GDDR6 with 12-14 Gbps speeds, RDNA 2 came with GDDR6 at 14-18 Gbps, and the current RDNA 3 used 18-20 Gbps GDDR6. Without an increment in memory generation, speeds should stay the same at 18 Gbps. However, it is crucial to remember that leaks should be treated with skepticism, as AMD's final memory choices for RDNA 4 could change before the official launch. The decision to use GDDR6 versus GDDR7 could have significant implications in the upcoming battle between AMD, NVIDIA, and Intel's next-generation GPU architectures. If AMD indeed opts for GDDR6 while NVIDIA pivots to GDDR7 for its "Blackwell" GPUs, it could create a disparity in memory bandwidth performance between the competing products. All three major GPU manufacturers—AMD, NVIDIA, and Intel with its "Battlemage" architecture—are expected to unveil their next-generation offerings in the fall of this year. As we approach these highly anticipated releases, more concrete details on specifications and performance capabilities will emerge, providing a clearer picture of the competitive landscape.

China Circumvents US Restrictions, Still Acquiring NVIDIA GPUs

A recent Reuters investigation has uncovered evidence suggesting Chinese universities and research institutes may have circumvented US sanctions on high-performance NVIDIA GPUs by purchasing servers containing the restricted chips. The sanctions tightened on November 17, 2023, prohibit the export of advanced NVIDIA GPUs like the consumer GeForce RTX 4090 to China. Despite these restrictions, Reuters found that at least ten China-based organizations acquired servers equipped with the sanctioned NVIDIA GPUs between November 20, 2023, and February 28, 2024. These servers were purchased from major vendors such as Dell, Gigabyte, and Supermicro, raising concerns about potential sanctions evasion. When contacted by Reuters, the companies provided varying responses.

Dell stated that it had not observed any instances of servers with restricted chips being shipped to China and expressed willingness to terminate relationships with resellers found to be violating export control regulations. Gigabyte, on the other hand, stated that it adheres to Taiwanese laws and international regulations. Notably, the sale and purchase of the sanctioned GPUs are not illegal in China. This raises the possibility that the restricted NVIDIA chips may have already been present in the country before the sanctions took effect on November 17, 2023. The findings highlight the challenges in enforcing export controls on advanced technologies, particularly in the realm of high-performance computing hardware. As tensions between the US and China continue to rise, the potential for further tightening of export restrictions on cutting-edge technologies remains a possibility.

ZOTAC to Show Scalable GPU Platforms and Industrial Solutions at Hannover Messe 2024

ZOTAC Technology is announcing a new lineup of enterprise and healthcare-oriented mini PCs designed for specific applications and scalable deployment, as well as a whole new class of external GPU acceleration platforms for Thunderbolt 3-compatible PCs. Aside from the all-new additions, ZOTAC is also refreshing its best-selling performance mini PCs with the newest generations of Intel Core Processors and NVIDIA RTX-enabled GPUs. ZOTAC will debut these rugged, innovative solutions and showcase other AI-ready compute solutions during Hannover Messe 2024, reaffirming ZOTAC's commitment to embrace the AI-driven future.

ZOTAC ZBOX Healthcare Series: Medical AI Solution
With the all-new ZOTAC Healthcare Series, ZOTAC is offering the reputed superior quality and performance of ZOTAC ZBOX Mini PCs to the realm of Healthcare. The ZBOX H39R5000W and ZBOX H37R3500W are equipped with 13th Generation Intel Core i9 or i7 laptop processors, as well as professional-grade NVIDIA RTX Ada Generation laptop GPUs. These mini PCs are ready to power medical imaging, algorithms, and more, with some of the latest and greatest hardware currently available.

Long-Time Linux Nouveau Driver Chief Ben Skeggs Joins NVIDIA

Ben Skeggs, a lead maintainer of the open-source NVIDIA GPU driver for Linux kernel called Nouveau, has joined NVIDIA. Working as an open-source contributor for the Nouveau driver for more than a decade, Ben Skeggs has achieved the remarkable feat of working to support the NVIDIA GPU hardware on open-source drivers. Before joining NVIDIA, Ben Skeggs worked at Red Hat up until September 18th of 2023. At that date, he posted that he was resigning from Red Hat and stepping back from the Nouveau open-source driver development. However, this news today comes as a bit of an interesting development, as Ben Skeggs is going to NVIDIA, which has been reluctant in the past to support open-source drivers.

Now, he is able to continue working on the driver directly from NVIDIA. He posted a set of 156 patches to the driver, affecting tens of thousands of lines of code. And he signed it all off from the official NVIDIA work address. This signals a potential turn in NVIDIA's approach to open-source software development, where the company might pay more attention to the movement and potentially hire more developers to support these projects. Back in 2012, NVIDIA had a different stance on open-source development, infamously provoking the creator of the Linux kernel, Linus Torvalds, to issue some snide remarks to the company. Hopefully, better days are ahead for the OSS world of driver development and collaboration with tech giants.

Intel Builds World's Largest Neuromorphic System to Enable More Sustainable AI

Today, Intel announced that it has built the world's largest neuromorphic system. Code-named Hala Point, this large-scale neuromorphic system, initially deployed at Sandia National Laboratories, utilizes Intel's Loihi 2 processor, aims at supporting research for future brain-inspired artificial intelligence (AI), and tackles challenges related to the efficiency and sustainability of today's AI. Hala Point advances Intel's first-generation large-scale research system, Pohoiki Springs, with architectural improvements to achieve over 10 times more neuron capacity and up to 12 times higher performance.

"The computing cost of today's AI models is rising at unsustainable rates. The industry needs fundamentally new approaches capable of scaling. For that reason, we developed Hala Point, which combines deep learning efficiency with novel brain-inspired learning and optimization capabilities. We hope that research with Hala Point will advance the efficiency and adaptability of large-scale AI technology." -Mike Davies, director of the Neuromorphic Computing Lab at Intel Labs

Minisforum V3 High-Performance AMD AI 3-in-1 Tablet Starts at $1199 Pre-Sale

Minisforum has unveiled a game-changing device that blurs the lines between tablets and laptops: the Minisforum V3. Today, the V3 laptop has hit the Minisforum store. This innovative 3-in-1 tablet is powered by the high-performance AMD Ryzen 7 8840U processor, offering a unique blend of portability and computing power. Unlike its traditional Mini PC design, Minisforum has adopted the popular form factor of Microsoft Surface and Lenovo Yoga tablet PCs with the V3. This versatile device can be a handheld tablet, a laptop with an included magnetic attachable keyboard, or a solo kickstand. At the heart of the Minisforum V3 lies the 8-core, 16-thread Ryzen 7 8840U processor, capable of delivering exceptional performance for demanding tasks. The tablet features a stunning 14-inch 2560 x 1600 IPS screen with a 165 Hz refresh rate and 100% DCI-P3 color gamut coverage, making it an ideal choice for creative professionals and content creators.

The V3's standout feature is its advanced cooling system, which allows the Ryzen 7 8840U and onboard Radeon 780M iGPU to operate at a stable 28 watts. This ensures smooth and efficient performance even under heavy workloads, making it a reliable device for all your tasks. The tablet's screen boasts a remarkable 500 nits of brightness, and its high color gamut coverage makes it perfect for professionals who require accurate color representation. Minisforum has priced the V3 competitively at $1199 at the pre-sale offering, making it an attractive option for those seeking a powerful and versatile device that can adapt to various scenarios. This primary option includes 32 GB of RAM and 1 TB SSD for storage. For early birds, Minisforum offers a V Pen, tempered glass screen protector, and laptop sleeve as a gift. Here is the link to the Minisforum V3 store.

Demand for NVIDIA's Blackwell Platform Expected to Boost TSMC's CoWoS Total Capacity by Over 150% in 2024

NVIDIA's next-gen Blackwell platform, which includes B-series GPUs and integrates NVIDIA's own Grace Arm CPU in models such as the GB200, represents a significant development. TrendForce points out that the GB200 and its predecessor, the GH200, both feature a combined CPU+GPU solution, primarily equipped with the NVIDIA Grace CPU and H200 GPU. However, the GH200 accounted for only approximately 5% of NVIDIA's high-end GPU shipments. The supply chain has high expectations for the GB200, with projections suggesting that its shipments could exceed millions of units by 2025, potentially making up nearly 40 to 50% of NVIDIA's high-end GPU market.

Although NVIDIA plans to launch products such as the GB200 and B100 in the second half of this year, upstream wafer packaging will need to adopt more complex and high-precision CoWoS-L technology, making the validation and testing process time-consuming. Additionally, more time will be required to optimize the B-series for AI server systems in aspects such as network communication and cooling performance. It is anticipated that the GB200 and B100 products will not see significant production volumes until 4Q24 or 1Q25.

Sony PlayStation 5 Pro Specifications Confirmed, Console Arrives Before Holidays

Thanks for the detailed information obtained by The Verge, today we confirm previously leaked details as Sony gears up to unveil the highly anticipated PlayStation 5 Pro, codenamed "Trinity." According to insider reports, Sony is urging developers to optimize their games for the PS5 Pro, with a primary focus on enhancing ray tracing capabilities. The console is expected to feature an RDNA 3 GPU with 30 WGP running BVH8, capable of 33.5 TeraFLOPS of FP32 single-precision computing power, and a slightly quicker CPU running at 3.85 GHz, enabling it to render games with ray tracing enabled or achieve higher resolutions and frame rates in select titles. Sony anticipates GPU rendering on the PS5 Pro to be approximately 45 percent faster than the standard PlayStation 5. The PS5 Pro GPU will be larger and utilize faster system memory to bolster ray tracing performance, boasting up to three times the speed of the regular PS5.

Additionally, the console will employ a more powerful ray tracing architecture, backed by PlayStation Spectral Super Resolution (PSSR), allowing developers to leverage graphics features like ray tracing more extensively. To support this endeavor, Sony is providing developers with test kits, and all games submitted for certification from August onward must be compatible with the PS5 Pro. Insider Gaming, the first to report the full PS5 Pro specs, suggests a potential release during the 2024 holiday period. The PS5 Pro will also feature modifications for developers regarding system memory, with Sony increasing the memory bandwidth from 448 GB/s to 576 GB/s, enhancing efficiency for an even more immersive gaming experience. To do AI processing, there is an custom AI accelerator capable of 300 8-bit INT8 TOPS and 67 16-bit FP16 TeraFLOPS, in addition to ACV audio codec running up to 35% faster.

ADLINK Reveals New Graphics Card with Intel Arc A380E GPU at Embedded World 2024

The industrial grade A380E graphics card features an exceptional cost/performance ratio, high reliability and low power consumption (50 W). As with all ADLINK industrial products, it delivers on longevity with availability guaranteed for a minimum of five years. In addition, the A380E graphics card is slim and compact with a single slot design, measuring only 69 mm x 156 mm.

Flexible application
Although the core market is likely to be commercial gaming, the A380E graphics card is also suited to industrial Edge AI applications such as Industrial IoT and retail analytics. Video wall graphics and media processing and delivery are examples of the many other potential uses.

ASUS IoT Announces PE8000G

ASUS IoT, the global AIoT solution provider, today announced PE8000G at Embedded World 2024, a powerful edge AI computer that supports multiple GPU cards for high performance—and is expertly engineered to handle rugged conditions with resistance to extreme temperatures, vibration and variable voltage. PE8000G is powered by formidable Intel Core processors (13th and 12th gen) and the Intel R680E chipset to deliver high-octane processing power and efficiency.

With its advanced architecture, PE8000G excels at running multiple neural network modules simultaneously in real-time—and represents a significant leap forward in edge AI computing. With its robust design, exceptional performance and wide range of features, PE8000G series is poised to revolutionize AI-driven applications across multiple industries, elevating edge AI computing to new heights and enabling organizations to tackle mission-critical tasks with confidence and to achieve unprecedented levels of productivity and innovation.

Intel Arc Battlemage Could Arrive Before Black Friday, Right in Time for Holidays

According to the latest report from ComputerBase, Intel had a strong presence at the recently concluded Embedded World 2024 conference. The company officially showcased its Arc series of GPUs for the embedded market, based on the existing Alchemist chips rebranded as the "E series." However, industry whispers hint at a more significant development—the impending launch of Intel's second-generation Arc Xe² GPUs, codenamed "Battlemage," potentially before the lucrative Black Friday shopping season. While Alchemist serves as Intel's current offering for embedded applications, many companies in attendance expressed keen interest in Battlemage, the successor to Alchemist. These firms often cover a broad spectrum, from servers and desktops to notebooks and embedded systems, necessitating a hardware platform that caters to this diverse range of applications.

Officially, Intel had previously stated that Battlemage would "hopefully" arrive before CES 2025, implying a 2024 launch. However, rumors from the trade show floor suggest a more ambitious target—a release before Black Friday, which falls on November 29th this year. This timeline aligns with Intel's historical launch patterns, as the original Arc A380 and notebook GPUs debuted in early October 2022, albeit with a staggered and limited rollout. Intel's struggles with the Alchemist launch serve as a learning experience for the company. Early promises and performance claims for the first-generation Arc GPUs failed to materialize, leading to a stuttering market introduction. This time, Intel has adopted a more reserved approach, avoiding premature and grandiose proclamations about Battlemage's capabilities.

Intel Unleashes Enterprise AI with Gaudi 3, AI Open Systems Strategy and New Customer Wins

At the Intel Vision 2024 customer and partner conference, Intel introduced the Intel Gaudi 3 accelerator to bring performance, openness and choice to enterprise generative AI (GenAI), and unveiled a suite of new open scalable systems, next-gen products and strategic collaborations to accelerate GenAI adoption. With only 10% of enterprises successfully moving GenAI projects into production last year, Intel's latest offerings address the challenges businesses face in scaling AI initiatives.

"Innovation is advancing at an unprecedented pace, all enabled by silicon - and every company is quickly becoming an AI company," said Intel CEO Pat Gelsinger. "Intel is bringing AI everywhere across the enterprise, from the PC to the data center to the edge. Our latest Gaudi, Xeon and Core Ultra platforms are delivering a cohesive set of flexible solutions tailored to meet the changing needs of our customers and partners and capitalize on the immense opportunities ahead."

Acer Launches New Nitro 14 and Nitro 16 Gaming Laptops Powered by AMD Ryzen 8040 Series Processors

Acer today announced the new Nitro 14 and Nitro 16 gaming laptops, powered by AMD Ryzen 8040 Series processors with Ryzen AI[1]. With up to NVIDIA GeForce RTX 4060[2] Laptop GPUs supported by DLSS 3.5 technology, both are backed by NVIDIA's RTX AI platform, providing an array of capabilities in over 500 games and applications, enhanced by AI. Gamers are immersed in their 14- and 16-inch NVIDIA G-SYNC compatible panels with up to WQXGA (2560x1600) resolution.

Whether in call or streaming in-game, Acer PurifiedVoice 2.0 harnesses the power of AI to block out external noises, while Acer PurifiedView keeps users always front and center of all the action. Microsoft Copilot in Windows (with a dedicated Copilot key) helps accelerate everyday tasks on these AI laptops, and with one month of Xbox Game Pass Ultimate included with every device, players will enjoy hundreds of high-quality PC games. To seamlessly take command of device performance and customizations, one click of the NitroSense key directs users to the control center and the library of available AI-related functions through the new Experience Zone.

U.S. Updates Advanced Semiconductor Ban, Actual Impact on the Industry Will Be Insignificant

On March 29th, the United States announced another round of updates to its export controls, targeting advanced computing, supercomputers, semiconductor end-uses, and semiconductor manufacturing products. These new regulations, which took effect on April 4th, are designed to prevent certain countries and businesses from circumventing U.S. restrictions to access sensitive chip technologies and equipment. Despite these tighter controls, TrendForce believes the practical impact on the industry will be minimal.

The latest updates aim to refine the language and parameters of previous regulations, tightening the criteria for exports to Macau and D:5 countries (China, North Korea, Russia, Iran, etc.). They require a detailed examination of all technology products' Total Processing Performance (TPP) and Performance Density (PD). If a product exceeds certain computing power thresholds, it must undergo a case-by-case review. Nevertheless, a new provision, Advanced Computing Authorized (ACA), allows for specific exports and re-exports among selected countries, including the transshipment of particular products between Macau and D:5 countries.

Imagination's new Catapult CPU is Driving RISC-V Device Adoption

Imagination Technologies today unveils the next product in the Catapult CPU IP range, the Imagination APXM-6200 CPU: a RISC-V application processor with compelling performance density, seamless security and the artificial intelligence capabilities needed to support the compute and intuitive user experience needs for next generation consumer and industrial devices.

"The number of RISC-V based devices is skyrocketing with over 16Bn units forecast by 2030, and the consumer market is behind much of this growth" says Rich Wawrzyniak, Principal Analyst at SHD Group. "One fifth of all consumer devices will have a RISC-V based CPU by the end of this decade. Imagination is set to be a force in RISC-V with a strategy that prioritises quality and ease of adoption. Products like APXM-6200 are exactly what will help RISC-V achieve the promised success."

AIO Workstation Combines 128-Core Arm Processor and Four NVIDIA GPUs Totaling 28,416 CUDA Cores

All-in-one computers are often traditionally seen as lower-powered alternatives to traditional desktop workstations. However, a new offering from Alafia AI, a startup focused on medical imaging appliances, aims to shatter that perception. The company's upcoming Alafia Aivas SuperWorkstation packs serious hardware muscle, demonstrating that all-in-one systems can match the performance of their more modular counterparts. At the heart of the Aivas SuperWorkstation lies a 128-core Ampere Altra processor, running at 3.0 GHz clock speed. This CPU is complemented by not one but three NVIDIA L4 GPUs for compute, and a single NVIDIA RTX 4000 Ada GPU for video output, delivering a combined 28,416 CUDA cores for accelerated parallel computing tasks. The system doesn't skimp on other components, either. It features a 4K touch display with up to 360 nits of brightness, an extensive 2 TB of DDR4 RAM, and storage options up to an 8 TB solid-state drive. This combination of cutting-edge CPU, GPU, memory, and storage is squarely aimed at the demands of medical imaging and AI development workloads.

The all-in-one form factor packs this incredible hardware into a sleek, purposefully designed clinical research appliance. While initially targeting software developers, Alafia AI hopes that institutions that can optimize their applications for the Arm architecture can eventually deploy the Aivas SuperWorkstation for production medical imaging workloads. The company is aiming for application integration in Q3 2024 and full ecosystem device integration by Q4 2024. With this powerful new offering, Alafia AI is challenging long-held assumptions about the performance limitations of all-in-one systems. The Aivas SuperWorkstation demonstrates that the right hardware choices can transform these compact form factors into true powerhouse workstations. Especially with a combined total output of three NVIDIA L4 compute units, alongside RTX 4000 Ada graphics card, the AIO is more powerful than some of the high-end desktop workstations.

X-Silicon Startup Wants to Combine RISC-V CPU, GPU, and NPU in a Single Processor

While we are all used to having a system with a CPU, GPU, and, recently, NPU—X-Silicon Inc. (XSi), a startup founded by former Silicon Valley veterans—has unveiled an interesting RISC-V processor that can simultaneously handle CPU, GPU, and NPU workloads in a chip. This innovative chip architecture, which will be open-source, aims to provide a flexible and efficient solution for a wide range of applications, including artificial intelligence, virtual reality, automotive systems, and IoT devices. The new microprocessor combines a RISC-V CPU core with vector capabilities and GPU acceleration into a single chip, creating a versatile all-in-one processor. By integrating the functionality of a CPU and GPU into a single core, X-Silicon's design offers several advantages over traditional architectures. The chip utilizes the open-source RISC-V instruction set architecture (ISA) for both CPU and GPU operations, running a single instruction stream. This approach promises lower memory footprint execution and improved efficiency, as there is no need to copy data between separate CPU and GPU memory spaces.

Called the C-GPU architecture, X-Silicon uses RISC-V Vector Core, which has 16 32-bit FPUs and a Scaler ALU for processing regular integers as well as floating point instructions. A unified instruction decoder feeds the cores, which are connected to a thread scheduler, texture unit, rasterizer, clipping engine, neural engine, and pixel processors. All is fed into a frame buffer, which feeds the video engine for video output. The setup of the cores allows the users to program each core individually for HPC, AI, video, or graphics workloads. Without software, there is no usable chip, which prompts X-Silicon to work on OpenGL ES, Vulkan, Mesa, and OpenCL APIs. Additionally, the company plans to release a hardware abstraction layer (HAL) for direct chip programming. According to Jon Peddie Research (JPR), the industry has been seeking an open-standard GPU that is flexible and scalable enough to support various markets. X-Silicon's CPU/GPU hybrid chip aims to address this need by providing manufacturers with a single, open-chip design that can handle any desired workload. The XSi gave no timeline, but it has plans to distribute the IP to OEMs and hyperscalers, so the first silicon is still away.

Apple M3 Ultra Chip Could be a Monolithic Design Without UltraFusion Interconnect

As we witness Apple's generational updates of the M series of chips, the highly anticipated SKU of the 3rd generation of Apple M series yet-to-be-announced top-of-the-line M3 Ultra chip is growing speculations from industry insiders. The latest round of reports suggests that the M3 Ultra might step away from its predecessor's design, potentially adopting a monolithic architecture without the UltraFusion interconnect technology. In the past, Apple has relied on a dual-chip design for its Ultra variants, using the UltraFusion interconnect to combine two M series Max chips. For example, the second generation M Ultra chip, M2 Ultra, boasts 134 billion transistors across two 510 mm² chips. However, die-shots of the M3 Max have sparked discussions about the absence of dedicated chip space for the UltraFusion interconnect.

While the absence of visible interconnect space on early die-shots is not conclusive evidence, as seen with the M1 Max not having visible UltraFusion interconnect and still being a part of M1 Ultra with UltraFusion, industry has led the speculation that the M3 Ultra may indeed feature a monolithic design. Considering that the M3 Max has 92 billion transistors and is estimated to have a die size between 600 and 700 mm², going Ultra with these chips may be pushing the manufacturing limit. Considering the maximum die size limit of 848 mm² for the TSMC N3B process used by Apple, there may not be sufficient space for a dual-chip M3 Ultra design. The potential shift to a monolithic design for the M3 Ultra raises questions about how Apple will scale the chip's performance without the UltraFusion interconnect. Competing solutions, such as NVIDIA's Blackwell GPU, use a high-bandwidth C2C interface to connect two 104 billion transistor chips, achieving a bandwidth of 10 TB/s. In comparison, the M2 Ultra's UltraFusion interconnect provided a bandwidth of 2.5 TB/s.

US Government Wants Nuclear Plants to Offload AI Data Center Expansion

The expansion of AI technology affects not only the production and demand for graphics cards but also the electricity grid that powers them. Data centers hosting thousands of GPUs are becoming more common, and the industry has been building new facilities for GPU-enhanced servers to serve the need for more AI. However, these powerful GPUs often consume over 500 Watts per single card, and NVIDIA's latest Blackwell B200 GPU has a TGP of 1000 Watts or a single kilowatt. These kilowatt GPUs will be present in data centers with 10s of thousands of cards, resulting in multi-megawatt facilities. To combat the load on the national electricity grid, US President Joe Biden's administration has been discussing with big tech to re-evaluate their power sources, possibly using smaller nuclear plants. According to an Axios interview with Energy Secretary Jennifer Granholm, she has noted that "AI itself isn't a problem because AI could help to solve the problem." However, the problem is the load-bearing of the national electricity grid, which can't sustain the rapid expansion of the AI data centers.

The Department of Energy (DOE) has been reportedly talking with firms, most notably hyperscalers like Microsoft, Google, and Amazon, to start considering nuclear fusion and fission power plants to satisfy the need for AI expansion. We have already discussed the plan by Microsoft to embed a nuclear reactor near its data center facility and help manage the load of thousands of GPUs running AI training/inference. However, this time, it is not just Microsoft. Other tech giants are reportedly thinking about nuclear as well. They all need to offload their AI expansion from the US national power grid and develop a nuclear solution. Nuclear power is a mere 20% of the US power sourcing, and DOE is currently financing a Holtec Palisades 800-MW electric nuclear generating station with $1.52 billion in funds for restoration and resumption of service. Microsoft is investing in a Small Modular Reactors (SMRs) microreactor energy strategy, which could be an example for other big tech companies to follow.

Intel Arc "Battlemage" Xe2-HPG BMG-10 & BMG-21 GPUs Discovered in Shipping Manifest

Speculated lower-end Intel second generation Arc GPUs popped up via SiSoftware Sandra database entries around mid-March—evaluation samples are likely in the hands of trusted hardware partners. Yesterday, momomo_us happened upon another interesting shipping manifest, following a series of AMD-related leaks. The latest list reveals five "Battlemage" products—three utilizing the BMG-21 GPU, and the remaining two being based on the BMG-10 design. These identifiers have appeared in older leaks, although the latter has been viewed in place sight—chez Intel Malaysia's Failure Analysis Lab.

Previous leaks suggest that these second generation Arc models (Xe2) reside within a "High-Performance Graphics" (HPG) discrete GPU family—the Xe2-HPG BMG-10 range is likely targeting an "enthusiast" market segment, while the Xe2-HPG BMG-21 tier is rumored to offer mid-tier performance. Intel staffers have expressed confidence about a possible late 2024 launch window. Back in January, Tom "TAP" Petersen revealed that the Arc hardware team had already moved onto third-gen "Celestial" GPU endeavors: "I'd say about 30% of our engineers are working on Battlemage, mostly on the software side because our hardware team is on the next thing." The first-gen deck has not been cleared fully it seems—the Alchemist family could be joined by two new variants in the near future.

Unannounced AMD Instinct MI388X Accelerator Pops Up in SEC Filing

AMD's Instinct family has welcomed a new addition—the MI388X AI accelerator—as discovered in a lengthy regulatory 10K filing (submitted to the SEC). The document reveals that the unannounced SKU—along with the MI250, MI300X and MI300A integrated circuits—cannot be sold to Chinese customers due to updated US trade regulations (new requirements were issued around October 2023). Versal VC2802 and VE2802 FPGA products are also mentioned in the same section. Earlier this month, AMD's Chinese market-specific Instinct MI309 package was deemed to be too powerful for purpose by the US Department of Commerce.

AMD has not published anything about the Instinct MI388X's official specification, and technical details have not emerged via leaks. The "X" tag likely implies that it has been designed for AI and HPC applications, akin to the recently launched MI300X accelerator. The designation of a higher model number could (naturally) point to a potentially more potent spec sheet, although Tom's Hardware posits that MI388X is a semi-custom spinoff of an existing model.

ASUS ROG Strix GeForce RTX 4090 D Tweaked to Match RTX 4090 FE Performance

NVIDIA's GeForce RTX 4090 D GPU was launched late last year in China—this weakened variant (of the standard RTX 4090) was designed with US trade regulations in mind. Chinese media outlets have toyed around with various custom models for several months—January 2024 evaluations indicated a 5% performance disadvantage when lined up against unrestricted models. The GeForce RTX 4090 D GPU is a potent beast despite a reduced core count and restricted TDP limit, but Chinese enthusiasts have continued to struggle with the implementation of worthwhile overclocks. HKEPC—a Hong Kong-situated PC hardware review outlet—has bucked that trend.

The mega-sized flagship ZOTAC RTX 4090 D PGF model has the technical credentials to break beyond the expected overclock increase of "2 to 5%," courtesy of a powerful 28-phase power PCB design and 530 W max. TGP limit. The Expreview team pulled a paltry 3.7% extra bit of performance from ZOTAC China's behemoth. In contrast, HKEPC wrangled out some bigger numbers with a sampled ASUS ROG STRIX RTX 4090 GAMING OC graphics card—matching unrestricted variants: "it turns out that NVIDIA only does not allow AIC manufacturers to preset overclocking, but it does not restrict users from overclocking by themselves. After a high degree of overclocking adjustment, the ROG Strix RTX 4090 D actually has a way to achieve the performance level of the RTX 4090 FE."
Return to Keyword Browsing
Apr 29th, 2024 13:37 EDT change timezone

New Forum Posts

Popular Reviews

Controversial News Posts