News Posts matching #NVIDIA

Return to Keyword Browsing

AAEON BOXER-8653AI & BOXER-8623AI Expand Vertical Market Potential in a More Compact Form

Leading provider of embedded PC solutions, AAEON, is delighted to announce the official launch of two new additions to its rich line of embedded AI systems, the BOXER-8653AI and BOXER-8623AI, which are powered by the NVIDIA Jetson Orin NX and Jetson Orin Nano, respectively. Measuring just 180 mm x 136 mm x 75 mm, both systems are compact and easily wall-mounted for discreet deployment, which AAEON indicate make them ideal for use in both indoor and outdoor settings such as factories and parking lots. Adding to this is the systems' environmental resilience, with the BOXER-8653AI sporting a wide -15°C to 60°C temperature tolerance and the BOXER-8623AI able to operate between -15°C and 65°C, with both supporting a 12 V ~ 24 V power input range via a 2-pin terminal block.

The BOXER-8653AI benefits from the NVIDIA Jetson Orin NX module, offering up to 70 TOPS of AI inference performance for applications that require extremely fast analysis of vast quantities of data. Meanwhile, the BOXER-8623AI utilizes the more efficient, yet still powerful NVIDIA Jetson Orin Nano module, capable of up to 40 TOPS. Both systems consequently make use of the 1024-core NVIDIA Ampere architecture GPU with 32 Tensor Cores.

ServiceNow, Hugging Face & NVIDIA Release StarCoder2 - a New Open-Access LLM Family

ServiceNow, Hugging Face, and NVIDIA today announced the release of StarCoder2, a family of open-access large language models for code generation that sets new standards for performance, transparency, and cost-effectiveness. StarCoder2 was developed in partnership with the BigCode Community, managed by ServiceNow, the leading digital workflow company making the world work better for everyone, and Hugging Face, the most-used open-source platform, where the machine learning community collaborates on models, datasets, and applications. Trained on 619 programming languages, StarCoder2 can be further trained and embedded in enterprise applications to perform specialized tasks such as application source code generation, workflow generation, text summarization, and more. Developers can use its code completion, advanced code summarization, code snippets retrieval, and other capabilities to accelerate innovation and improve productivity.

StarCoder2 offers three model sizes: a 3-billion-parameter model trained by ServiceNow; a 7-billion-parameter model trained by Hugging Face; and a 15-billion-parameter model built by NVIDIA with NVIDIA NeMo and trained on NVIDIA accelerated infrastructure. The smaller variants provide powerful performance while saving on compute costs, as fewer parameters require less computing during inference. In fact, the new 3-billion-parameter model matches the performance of the original StarCoder 15-billion-parameter model. "StarCoder2 stands as a testament to the combined power of open scientific collaboration and responsible AI practices with an ethical data supply chain," emphasized Harm de Vries, lead of ServiceNow's StarCoder2 development team and co-lead of BigCode. "The state-of-the-art open-access model improves on prior generative AI performance to increase developer productivity and provides developers equal access to the benefits of code generation AI, which in turn enables organizations of any size to more easily meet their full business potential."

NVIDIA Grace Hopper Systems Gather at GTC

The spirit of software pioneer Grace Hopper will live on at NVIDIA GTC. Accelerated systems using powerful processors - named in honor of the pioneer of software programming - will be on display at the global AI conference running March 18-21, ready to take computing to the next level. System makers will show more than 500 servers in multiple configurations across 18 racks, all packing NVIDIA GH200 Grace Hopper Superchips. They'll form the largest display at NVIDIA's booth in the San Jose Convention Center, filling the MGX Pavilion.

MGX Speeds Time to Market
NVIDIA MGX is a blueprint for building accelerated servers with any combination of GPUs, CPUs and data processing units (DPUs) for a wide range of AI, high performance computing and NVIDIA Omniverse applications. It's a modular reference architecture for use across multiple product generations and workloads. GTC attendees can get an up-close look at MGX models tailored for enterprise, cloud and telco-edge uses, such as generative AI inference, recommenders and data analytics. The pavilion will showcase accelerated systems packing single and dual GH200 Superchips in 1U and 2U chassis, linked via NVIDIA BlueField-3 DPUs and NVIDIA Quantum-2 400 Gb/s InfiniBand networks over LinkX cables and transceivers. The systems support industry standards for 19- and 21-inch rack enclosures, and many provide E1.S bays for nonvolatile storage.

NVIDIA Accused of Acting as "GPU Cartel" and Controlling Supply

World's most important fuel of the AI frenzy, NVIDIA, is facing accusations of acting as a "GPU cartel" and controlling supply in the data center market, according to statements made by executives at rival chipmaker Groq and former AMD executive Scott Herkelman. In an interview with the Wall Street Journal, Groq CEO Jonathan Ross alleged that some of NVIDIA's data center customers are afraid to even meet with rival AI chipmakers out of fear that NVIDIA will retaliate by delaying shipments of already ordered GPUs. This is despite NVIDIA's claims that it is trying to allocate supply fairly during global shortages. "This happens more than you expect, NVIDIA does this with DC customers, OEMs, AIBs, press, and resellers. They learned from GPP to not put it into writing. They just don't ship after a customer has ordered. They are the GPU cartel, and they control all supply," said former Senior Vice President and General Manager at AMD Radeon, Scott Herkelman, in response to the accusations on X/Twitter.

Video Adverts Coming to GeForce NOW Free-tier

NVIDIA is reported to have started distributing an email newsletter regarding "video sponsorship messages" for its GeForce NOW free-tier. Team Green's entry-level/no-payment-required cloud gaming service grants access to time limited sessions (1-hour) on their least powerful server hardware (AKA Basic Rigs). Free-tier members are required to queue up for online server availability, while paying customers get "Priority" access—starting at $9.99 per month. NVIDIA's latest newsletter outlines an upcoming deployment of video adverts (running up to two minutes in length) for queued free-tier customers. The Verge must have heard early rumors of said "modification"—yesterday's late night report opened with: "GeForce Now is about to be very slightly less of a deal—on Wednesday February 28th, users will start seeing ads."

Sean Hollister, a senior editor at The Verge, established contact with an NVIDIA representative and managed to get a comment on the free-tier adjustment. Company spokesperson Stephanie Ngo confirmed that: "free users will start to see up to two minutes of ads while waiting in queue to start a gaming session." Team Green proposed that advertising will "help pay for the free-tier of service." They also anticipate reduced average wait times "for free users over time," due to the implementation of sponsored video content. Naturally, the "Priority" and "Ultimate" GeForce NOW membership tiers are not affected.

KFA2 Intros GeForce RTX 4070 SUPER EX Gamer in Pink and White Trims

KFA2, the brand owned by Galax for specific markets in Europe, announced its GeForce RTX 4070 SUPER EX Gamer graphics card in two new trims, Pink, and White. The regular RTX 4070 SUPER EX Gamer the company launched in January comes in black. The Pink trim involves a bubblegum pink shade taking over the cooler shroud and metal backplate. Ditto with the white trim. The PCB underneath both cards remains black, but since the cooling solution dwarfs it, it most remains out of sight when installed.

Both cards feature frosted acrylic fan impellers, with addressable RGB LEDs located in the fan hubs. Besides the illuminated fans, the GeForce RTX logo on top of the card comes with RGB LED illumination. Both cards come with a mobile app-based control for the lighting. This app talks to a service installed in your PC over the Internet, which interacts with the graphics card. Much like the regular black RTX 4070 SUPER EX Gamer, the two cards stick with NVIDIA-reference clock speeds of 2475 MHz boost for the GPU, and 21 Gbps (GDDR6X-effective) for the memory. KFA2 is pricing both the Pink and White variants at 679€ including taxes, a 20€ premium over the regular black EX Gamer card.

NVIDIA GH200 72-core Grace CPU Benched Against AMD Threadripper 7000 Series

GPTshop.ai is building prototypes of their "ultimate high-end desktop supercomputer," running the NVIDIA GH200 "Grace" CPU for AI and HPC workloads. Michael Larabel—founder and principal author of Phoronix—was first allowed to "remote access" a GPTshop.ai GH200 576 GB workstation converted model in early February—for the purpose of benchmarking it against systems based on AMD EPYC Zen 4 and Intel Xeon Emerald Rapids processors. Larabel noted: "it was a very interesting battle" that demonstrated the capabilities of 72 Arm Neoverse-V2 cores (in Grace). With this GPTshop.ai GH200 system actually being in workstation form, I also ran some additional benchmarks looking at the CPU capabilities of the GH200 compared to AMD Ryzen Threadripper 7000 series workstations."

Larabel had on-site access to two different Threadripper systems—a Hewlett-Packard (HP) Z6 G5 A workstation and a System76 Thelio Major semi-custom build. No comparable Intel "Xeon W hardware" was within reach, so the Team Green desktop supercomputer was only pitched against AMD HEDT processors. The HP review sample was configured with an AMD Ryzen Threadripper PRO 7995WX 96-core / 192-thread Zen 4 processor, 8 x 16 GB DDR5-5200 memory, and NVIDIA RTX A4000 GPU. Larabel said that it was an "all around nice high-end AMD workstation." The System76 Thelio Major was specced with an AMD Ryzen Threadripper 7980X processor "as the top-end non-PRO SKU." It is a 64-core / 128-thread part, working alongside 4 x 32 GB DDR5-4800 memory and a Radeon PRO W7900 graphics card.

Dell Announces New Laptops and Mobile Workstations with Focus on AI

Dell Technologies will introduce the broadest portfolio of commercial AI laptops and mobile workstations designed to bring organizations and employee productivity into the AI era. "The next generation of PCs is emerging at a pivotal time - with upcoming refresh cycles and new capabilities on the PC creating the perfect storm," said Patrick Moorhead, founder and CEO of Moor Insights & Strategy. "Dell's commercial AI PCs and workstations, coupled with its ecosystem of peripherals, software and services, offer an AI continuum designed to enhance the user experience today and set organizations up for success in the future."

"Every company that wants to remain competitive will have to implement AI in some way, and AI PCs will be central to that," said Sam Burd, president, Client Solutions Group, Dell Technologies. "From running complex AI workloads on workstations to using day-to-day AI-powered applications on laptops, the AI PC will be an important investment that pays dividends in productivity and paves the way to a smarter, more efficient future. Dell's advantage starts with offering more AI PCs across the commercial portfolio from day one, giving customers the ability to start future proofing for AI today."

AMD Radeon RX 7900 GRE To Launch Globally on February 27

AMD's Radeon RX 7900 GRE, or Golden Rabbit Edition, which was previously available only to the Chinese market, will launch globally on February 27. According to the leaked slides, the Radeon RX 7900 GRE will launch at $549, and AMD is comparing it to the NVIDIA GeForce RTX 4070 non-SUPER graphics card. In case you missed it, the AMD Radeon RX 7900 GRE is based on the Navi 31 XL GPU with 80 Compute Units (CUs), which leaves it with 5120 Stream Processors, and comes with 16 GB of 18 Gbps GDDR6 memory on a 256-bit memory interface, which adds up to a maximum bandwidth of 576 GB/s. The Radeon RX 7900 GRE should fit nicely between the Radeon RX 7900 XT and the Radeon RX 7800 XT.

According to the leaked slides, AMD is comparing the Radeon RX 7900 GRE against the NVIDIA GeForce RTX 4070 non-SUPER, which now shares the same price after the recent $50 price cut. According to AMD's own slides, the Radeon RX 7900 GRE should provide around 14 percent more performance per buck on average, and is between 1 and 32 percent faster, at least in games tested by AMD.

NVIDIA Announces RTX 500 and 1000 Professional Ada Generation Laptop GPUs

With generative AI and hybrid work environments becoming the new standard, nearly every professional, whether a content creator, researcher or engineer, needs a powerful, AI-accelerated laptop to help users tackle their industry's toughest challenges - even on the go. The new NVIDIA RTX 500 and 1000 Ada Generation Laptop GPUs will be available in new, highly portable mobile workstations, expanding the NVIDIA Ada Lovelace architecture-based lineup, which includes the RTX 2000, 3000, 3500, 4000 and 5000 Ada Generation Laptop GPUs.

AI is rapidly being adopted to drive efficiencies across professional design and content creation workflows and everyday productivity applications, underscoring the importance of having powerful local AI acceleration and sufficient processing power in systems. The next generation of mobile workstations with Ada Generation GPUs, including the RTX 500 and 1000 GPUs, will include both a neural processing unit (NPU), a component of the CPU, and an NVIDIA RTX GPU, which includes Tensor Cores for AI processing. The NPU helps offload light AI tasks, while the GPU provides up to an additional 682 TOPS of AI performance for more demanding day-to-day AI workflows.

ASRock Rack Offers World's Smallest NVIDIA Grace Hopper Superchip Server, Other Innovations at MWC 2024

ASRock Rack is a subsidiary of ASRock that deals with servers, workstations, and other data-center hardware, and comes with the enormous brand trust not just of ASRock, but also the firm hand of parent company and OEM giant Pegatron. At the 2024 Mobile World Congress, ASRock Rack introduced several new server innovations relevant to the AI Edge, and 5G cellular carrier industries. A star attraction here is the new ASRock Rack MECAI-GH200, claimed to be the world's smallest server powered by the NVIDIA GH200 Grace Hopper Superchip.

The ASRock Rack MECAI-GH200 executes one of NVIDIA's original design goals behind the Grace Hopper Superchip—AI deployments in an edge environment. The GH200 module combines an NVIDIA Grace CPU with a Hopper AI GPU, and a performance-optimized NVLink interconnect between them. The CPU features 72 Arm Neoverse V2 cores, and 480 GB of LPDDR5X memory; while the Hopper GPU has 132 SM with 528 Tensor cores, and 96 GB of HBM3 memory across a 6144-bit memory interface. Given that the newer HBM3e version of the GH200 won't come out before Q2-2024, this has to be the version with HBM3. What makes the MECAI-GH200 the world's smallest server with the GH200 has to be its compact 2U form-factor—competing solutions tend to be 3U or larger.

Supermicro Accelerates Performance of 5G and Telco Cloud Workloads with New and Expanded Portfolio of Infrastructure Solutions

Supermicro, Inc. (NASDAQ: SMCI), a Total IT Solution Provider for AI, Cloud, Storage, and 5G/Edge, delivers an expanded portfolio of purpose-built infrastructure solutions to accelerate performance and increase efficiency in 5G and telecom workloads. With one of the industry's most diverse offerings, Supermicro enables customers to expand public and private 5G infrastructures with improved performance per watt and support for new and innovative AI applications. As a long-term advocate of open networking platforms and a member of the O-RAN Alliance, Supermicro's portfolio incorporates systems featuring 5th Gen Intel Xeon processors, AMD EPYC 8004 Series processors, and the NVIDIA Grace Hopper Superchip.

"Supermicro is expanding our broad portfolio of sustainable and state-of-the-art servers to address the demanding requirements of 5G and telco markets and Edge AI," said Charles Liang, president and CEO of Supermicro. "Our products are not just about technology, they are about delivering tangible customer benefits. We quickly bring data center AI capabilities to the network's edge using our Building Block architecture. Our products enable operators to offer new capabilities to their customers with improved performance and lower energy consumption. Our edge servers contain up to 2 TB of high-speed DDR5 memory, 6 PCIe slots, and a range of networking options. These systems are designed for increased power efficiency and performance-per-watt, enabling operators to create high-performance, customized solutions for their unique requirements. This reassures our customers that they are investing in reliable and efficient solutions."

LG Announces US Pricing and Availability of 2024 Gram Pro Notebooks

LG Electronics USA (LG) today announced pricing and availability of its 2024 premium lineup of laptops - the LG gram Pro and LG gram Pro 2-in-1. The 16- and 17-inch LG gram Pro models retail for $2399 (16Z90SP-A.ADB9U1) and $2499 (17Z90SP-E.ADB9U1) respectively while the CES 2024 Innovation Award-winning 16-inch gram Pro 2-in-1 (16T90SP-K.ADB9U1) retails for $2099. For a limited time, customers shopping on LG.com can pre-order the 16- and 17-inch LG gram Pro and 16-inch LG gram Pro 2-in-1.

Throughout the duration of the pre-order period from February 21, 2024, to March 10, 2024, customers will be able to purchase the 32 GB-RAM/2 TB-SSD LG gram Pro laptop for the price of a 16 GB-RAM/1 TB-SSD model of the same screen size. They'll also receive an LG gram +view IPS portable monitor (16MR70.ASDU) and expedited shipping at no additional cost. All standard terms of purchase apply.

SmartCow introduces Uranus Plus AI Fanless Embedded System Powered by NVIDIA Jetson Orin

SmartCow, an AI engineering company specializing in building complex hardware and software solutions for artificial intelligence at the edge, announces the launch of their latest product, Uranus Plus, an AI fanless embedded system powered by the latest NVIDIA Jetson Orin NX and Jetson Orin Nano system-on-modules. With its thermally efficient design and compact form factor, Uranus Plus is suitable for various smart applications. Uranus Plus comes with options for 5G, 4G, and Wi-Fi connectivity and includes a 256Gb NVMe SSD, enabling the simultaneous operation of multiple neural networks and processing of high-resolution images, enhancing a groundbreaking benchmark in AI-driven capabilities at the edge with support up to 100 TOPS of AI compute.

Uranus Plus supercharges vision AI application development at the edge with NVIDIA Metropolis Microservices for Jetson through app stack modernization. Uranus Plus developers now get access to the latest generative AI capabilities through simple API calls, along with a far faster path to development and cloud-native deployment of vision AI applications at the far edge.

NVIDIA Prepared to Offer Custom Chip Designs to AI Clients

NVIDIA is reported to be setting up an AI-focused semi-custom chip design business unit, according to inside sources known to Reuters—it is believed that Team Green leadership is adapting to demands leveraged by key data-center customers. Many companies are seeking cheaper alternatives, or have devised their own designs (budget/war chest permitting)—NVIDIA's current range of AI GPUs are simply off-the-shelf solutions. OpenAI has generated the most industry noise—their alleged early 2024 fund-raising pursuits have attracted plenty of speculative/kind-of-serious interest from notable semiconductor personalities.

Team Green is seemingly reacting to emerging market trends—Jensen Huang (CEO, president and co-founder) has hinted that NVIDIA custom chip designing services are on the cusp. Stephen Nellis—a Reuters reporter specializing in tech industry developments—has highlighted select NVIDIA boss quotes from an incoming interview piece: "We're always open to do that. Usually, the customization, after some discussion, could fall into system reconfigurations or recompositions of systems." The Team Green chief teased that his engineering team is prepared to take on the challenge meeting exact requests: "But if it's not possible to do that, we're more than happy to do a custom chip. And the benefit to the customer, as you can imagine, is really quite terrific. It allows them to extend our architecture with their know-how and their proprietary information." The rumored NVIDIA semi-custom chip design business unit could be introduced in an official capacity at next month's GTC 2024 Conference.

NVIDIA Expects Upcoming Blackwell GPU Generation to be Capacity-Constrained

NVIDIA is anticipating supply issues for its upcoming Blackwell GPUs, which are expected to significantly improve artificial intelligence compute performance. "We expect our next-generation products to be supply constrained as demand far exceeds supply," said Colette Kress, NVIDIA's chief financial officer, during a recent earnings call. This prediction of scarcity comes just days after an analyst noted much shorter lead times for NVIDIA's current flagship Hopper-based H100 GPUs tailored to AI and high-performance computing. The eagerly anticipated Blackwell architecture and B100 GPUs built on it promise major leaps in capability—likely spurring NVIDIA's existing customers to place pre-orders already. With skyrocketing demand in the red-hot AI compute market, NVIDIA appears poised to capitalize on the insatiable appetite for ever-greater processing power.

However, the scarcity of NVIDIA's products may present an excellent opportunity for significant rivals like AMD and Intel. If both companies can offer a product that could beat NVIDIA's current H100 and provide a suitable software stack, customers would be willing to jump to their offerings and not wait many months for the anticipated high lead times. Intel is preparing the next-generation Gaudi 3 and working on the Falcon Shores accelerator for AI and HPC. AMD is shipping its Instinct MI300 accelerator, a highly competitive product, while already working on the MI400 generation. It remains to be seen if AI companies will begin the adoption of non-NVIDIA hardware or if they will remain a loyal customer and agree to the higher lead times of the new Blackwell generation. However, capacity constrain should only be a problem at launch, where the availability should improve from quarter to quarter. As TSMC improves CoWoS packaging capacity and 3 nm production, NVIDIA's allocation of the 3 nm wafers will likely improve over time as the company moves its priority from H100 to B100.

NVIDIA App Doesn't Need a Login, Unlike GeForce Experience

We found out that the new NVIDIA App doesn't need an NVIDIA Account login, and yet gives you nearly all of its functionality. NVIDIA today rolled out the GeForce 551.61 WHQL drivers, and with it, the new NVIDIA App, as we detailed in the driver's news report. NVIDIA App is the company's latest take on a Control Panel application that combines the functionality of the over 20-year-old NVIDIA Control Panel Win32 application, and the modern GeForce Experience app (GFE). The former focuses on settings related to the display head, with one or more settings for the GPU, but has no hardware monitoring or performance overlay features. GFE is more of a concentric outer layer focused on the games installed in your PC, to which you can figure out and apply optimal settings. The new NVIDIA App essentially combines the functionalities of the two, but it has an ace up its sleeve—you don't need an NVIDIA Account to use it.

One of the biggest drawbacks of GeForce Experience is that it mandates you to create an NVIDIA Account, and keeps you logged into this account to use its functionality. Not everyone wants an app that does this; and so some gamers would want to skip installation of GFE altogether during the GeForce driver installation. NVIDIA App takes a refreshingly different approach. It is currently a public beta, isn't part of the driver package, isn't found on Microsoft Store, but is being distributed as a standalone app with its own installer.

NVIDIA GeForce 551.61 WHQL with RTX HDR Released, NVIDIA Finally Has a Modern Control Panel

NVIDIA today released the latest version of GeForce drivers. While the actual driver package of the new GeForce 551.61 is WHQL, it also includes a beta version of the NVIDIA App. This is NVIDIA's take on a modern control panel application that combines functionality of the classic NVIDIA Control Panel that hasn't quite changed in 20+ years; and the GeForce Experience application that not everyone likes to have installed. The new NVIDIA App has a lot in common with the AMD Radeon Software application, in that you can configure your displays, monitor and tune your GPU, as well as manage and optimize your installed games. The app also helps gather and present software update options spanning both the main drivers and application profiles for DLSS and other technologies. Most importantly, the NVIDIA App provides a comprehensive performance overlay suite. We will be doing an article detailing the NVIDIA App before the weekend.

NVIDIA is also debuting RTX HDR, a technology that adds HDR capability to games that otherwise lack it. It does so by utilizing Tensor cores and an AI model that attempts to add HDR to SDR content. The driver adds optimization and optimal settings for Granblue Fantasy: Relink, Nightingale, Pacific Drive, and Skull and Bones. A stability issue with Tom Clancy's Rainbow Six Siege running using the Vulkan API has been fixed. Among the general issues fixed concern incorrect tone mapping for RTX Video HDR; poor SDR video playback quality in Chromium-based web-browsers with Windows HDR setting enabled; filter settings not carrying over with NVIDIA Freestyle; and a couple of bugs with Adobe Substance 3D.

DOWNLOAD: NVIDIA GeForce 551.61 WHQL

NVIDIA GeForce NOW Now Offers 1,800 Games to Stream

Top-tier games from publishing partners Bandai Namco Entertainment and Inflexion Games are joining GeForce NOW this week as the cloud streaming service's fourth-anniversary celebrations continue. Eleven new titles join the over 1,800 supported games in the GeForce NOW library, including Nightingale from Inflexion Games and Bandai Namco Entertainment's Tales of Arise, Katamari Damacy REROLL and Klonoa Phantasy Reverie Series.

"Happy fourth anniversary, GeForce NOW!" cheered Jarrett Lee, head of publishing at Inflexion Games. "The platform's ease of access and seamless performance comprise a winning combination, and we're excited to see how cloud gaming evolves."

NVIDIA GeForce RTX 4070 and 4060 Ti Founders Edition Get Price Drop in China

Both the NVIDIA GeForce RTX 4070 and the RTX 4060 Ti Founders Edition graphics cards have seen a price drop in China, pushing them below the original MSRP. Targeting the mainstream segment, these two graphics cards held at MSRP for quite some time, but now, some retailers are listing these Founders Edition at a reduced price.

The NVIDIA GeForce RTX 4070 Founders Edition is currently listed for 4599 RMB at JD.com, which is 200 RMB lower than its official MSRP. The same goes for the GeForce RTX 4060 Ti Founders Edition, which is currently listed at 2999 RMB. Some custom versions from NVIDIA AIC partners can be found with an even lower price tag, and it could be a result of AMD's push to sell the Radeon RX 6750 GRE in China and slash the sales of the new Radeon RX 7600 series. It is yet to be seen if these price cuts will be coming to the rest of the world, as we will certainly keep a close eye on US and European retailers/e-tailers.

Modders Pull Off 16GB GeForce RTX 2080 Upgrade, Modded Card Posts 8% Performance Boost

Brazilian tech enthusiast Paulo Gomes, in association with Jefferson Silva, and Ygor Mota, successfully modded an EVGA GeForce RTX 2080 "Turing" graphics card to 16 GB. This was done by replacing each of its 8 Gbit GDDR6 memory chips with ones that have double the density, at 16 Gbit. Over the GPU's 256-bit wide memory bus, eight of these chips add up to 16 GB. The memory speed was unchanged at 14 Gbps reference, as were the GPU clocks.

The process of modding involves de-soldering each of the eight 8 Gbit chips, clearing out the memory pads of any shorted pins, using a GDDR6 stencil to place replacement solder balls, and then soldering the new 16 Gbit chips onto the pad under heat. Besides replacing the memory chips, a series of SMD jumpers need to be adjusted near the BIOS ROM chip, which lets the GPU correctly recognize the 16 GB memory size. The TU104 silicon by default supports higher density memory, as NVIDIA uses this chip on some of its professional graphics cards with 16 GB memory, such as the Quadro RTX 5000.

NVIDIA Announces Q4 and Fiscal 2024 Results, Clocks 126% YoY Revenue Growth, Gaming Just 1/6th of Data Center Revenues

NVIDIA (NASDAQ: NVDA) today reported revenue for the fourth quarter ended January 28, 2024, of $22.1 billion, up 22% from the previous quarter and up 265% from a year ago. For the quarter, GAAP earnings per diluted share was $4.93, up 33% from the previous quarter and up 765% from a year ago. Non-GAAP earnings per diluted share was $5.16, up 28% from the previous quarter and up 486% from a year ago.

For fiscal 2024, revenue was up 126% to $60.9 billion. GAAP earnings per diluted share was $11.93, up 586% from a year ago. Non-GAAP earnings per diluted share was $12.96, up 288% from a year ago. "Accelerated computing and generative AI have hit the tipping point. Demand is surging worldwide across companies, industries and nations," said Jensen Huang, founder and CEO of NVIDIA.

Arm Launches Next-Generation Neoverse CSS V3 and N3 Designs for Cloud, HPC, and AI Acceleration

Last year, Arm introduced its Neoverse Compute Subsystem (CSS) for the N2 and V2 series of data center processors, providing a reference platform for the development of efficient Arm-based chips. Major cloud service providers like AWS with Graviton 4 and Trainuium 2, Microsoft with Cobalt 100 and Maia 100, and even NVIDIA with Grace CPU and Bluefield DPUs are already utilizing custom Arm server CPU and accelerator designs based on the CSS foundation in their data centers. The CSS allows hyperscalers to optimize Arm processor designs specifically for their workloads, focusing on efficiency rather than outright performance. Today, Arm has unveiled the next generation CSS N3 and V3 for even greater efficiency and AI inferencing capabilities. The N3 design provides up to 32 high-efficiency cores per die with improved branch prediction and larger caches to boost AI performance by 196%, while the V3 design scales up to 64 cores and is 50% faster overall than previous generations.

Both the N3 and V3 leverage advanced features like DDR5, PCIe 5.0, CXL 3.0, and chiplet architecture, continuing Arm's push to make chiplets the standard for data center and cloud architectures. The chiplet approach enables customers to connect their own accelerators and other chiplets to the Arm cores via UCIe interfaces, reducing costs and time-to-market. Looking ahead, Arm has a clear roadmap for its Neoverse platform. The upcoming CSS V4 "Adonis" and N4 "Dionysus" designs will build on the improvements in the N3 and V3, advancing Arm's goal of greater efficiency and performance using optimized chiplet architectures. As more major data center operators introduce custom Arm-based designs, the Neoverse CSS aims to provide a flexible, efficient foundation to power the next generation of cloud computing.

Google's Gemma Optimized to Run on NVIDIA GPUs, Gemma Coming to Chat with RTX

NVIDIA, in collaboration with Google, today launched optimizations across all NVIDIA AI platforms for Gemma—Google's state-of-the-art new lightweight 2 billion- and 7 billion-parameter open language models that can be run anywhere, reducing costs and speeding innovative work for domain-specific use cases.

Teams from the companies worked closely together to accelerate the performance of Gemma—built from the same research and technology used to create the Gemini models—with NVIDIA TensorRT-LLM, an open-source library for optimizing large language model inference, when running on NVIDIA GPUs in the data center, in the cloud and on PCs with NVIDIA RTX GPUs. This allows developers to target the installed base of over 100 million NVIDIA RTX GPUs available in high-performance AI PCs globally.

Jensen Huang to Unveil Latest AI Breakthroughs at GTC 2024 Conference

NVIDIA today announced it will host its flagship GTC 2024 conference at the San Jose Convention Center from March 18-21. More than 300,000 people are expected to register to attend in person or virtually. NVIDIA founder and CEO Jensen Huang will deliver the keynote from the SAP Center on Monday, March 18, at 1 p.m. Pacific time. It will be livestreamed and available on demand. Registration is not required to view the keynote online. Since Huang first highlighted machine learning in his 2014 GTC keynote, NVIDIA has been at the forefront of the AI revolution. The company's platforms have played a crucial role in enabling AI across numerous domains including large language models, biology, cybersecurity, data center and cloud computing, conversational AI, networking, physics, robotics, and quantum, scientific and edge computing.

The event's 900 sessions and over 300 exhibitors will showcase how organizations are deploying NVIDIA platforms to achieve remarkable breakthroughs across industries, including aerospace, agriculture, automotive and transportation, cloud services, financial services, healthcare and life sciences, manufacturing, retail and telecommunications. "Generative AI has moved to center stage as governments, industries and organizations everywhere look to harness its transformative capabilities," Huang said. "GTC has become the world's most important AI conference because the entire ecosystem is there to share knowledge and advance the state of the art. Come join us."
Return to Keyword Browsing
Jun 3rd, 2024 03:21 EDT change timezone

New Forum Posts

Popular Reviews

Controversial News Posts