Skip to main content

The future of fast PC graphics? Connecting directly to SSDs

Performance boosts are expected with each new generation of the best graphics cards, but it seems that Nvidia and IBM have their sights set on greater changes.

The companies teamed up to work on Big accelerator Memory (BaM), a technology that involves connecting graphics cards directly to superfast SSDs. This could result in larger GPU memory capacity and faster bandwidth while limiting the involvement of the CPU.

A chart breaks down Nvidia and IBM's BaM technology.
Image source: Arxiv Image used with permission by copyright holder

This type of technology has already been thought of, and worked on, in the past. Microsoft’s DirectStorage application programming interface (API) works in a somewhat similar way, improving data transfers between the GPU and the SSD. However, this relies on external software, only applies to games, and only works on Windows. Nvidia and IBM researchers are working together on a solution that removes the need for a proprietary API while still connecting GPUs to SSDs.

Recommended Videos

The method, amusingly referred to as BaM, was described in a paper written by the team that designed it. Connecting a GPU directly to an SSD would provide a performance boost that could prove to be viable, especially for resource-heavy tasks such as machine learning. As such, it would mostly be used in professional high-performance computing (HPC) scenarios.

Get your weekly teardown of the tech behind PC gaming
Check your inbox!

The technology that is currently available for processing such heavy workloads requires the graphics card to rely on large amounts of special-purpose memory, such as HBM2, or to be provided with efficient access to SSD storage. Considering that datasets are only growing in size, it’s important to optimize the connection between the GPU and storage in order to allow for efficient data transfers. This is where BaM comes in.

“BaM mitigates the I/O traffic amplification by enabling the GPU threads to read or write small amounts of data on-demand, as determined by the compute,” said the researchers in their paper, first cited by The Register. “The goal of BaM is to extend GPU memory capacity and enhance the effective storage access bandwidth while providing high-level abstractions for the GPU threads to easily make on-demand, fine-grain access to massive data structures in the extended memory hierarchy.”

An Nvidia GPU core sits on a table.
Niels Broekhuijsen / Digital Trends

For many people who don’t work directly with this subject, the details may seem complicated, but the gist of it is that Nvidia wants to rely less on the processor and connect directly to the source of the data. This would both make the process more efficient and free up the CPU, making the graphics card much more self-sufficient. The researchers claim that this design would be able to compete with DRAM-based solutions while remaining cheaper to implement.

Although Nvidia and IBM are undoubtedly breaking new ground with their BaM technology, AMD worked in this area first: In 2016, it unveiled the Radeon Pro SSG, a workstation GPU with integrated M.2 SSDs. However, the Radeon Pro SSG was intended to be strictly a graphics solution, and Nvidia is taking it a few steps further, aiming to deal with complex and heavy compute workloads.

The team working on BaM plans to release the details of their software and hardware optimization as open source, allowing others to build on their findings. There is no mention as to when, if ever, BaM might find itself implemented in future Nvidia products.

Monica J. White
Monica is a computing writer at Digital Trends, focusing on PC hardware. Since joining the team in 2021, Monica has written…
Nvidia’s RTX 5080 may be better than the RTX 5090 in one small way
The PNY RTX 4080 XLR8 installed in a PC.

The launch of Nvidia's next-gen best graphics cards is right around the corner, and we're getting new leaks about the specs almost every day. Today, Benchlife reveals that the RTX 5080 may be the only RTX 50-series GPU to receive 30Gbps memory modules from the get-go. This would give the RTX 5080 a slight advantage, but there's also some conflicting information about the memory configuration for this GPU.

All of Nvidia's next-gen graphics cards are said to use new GDDR7 memory, and yesterday's Zotac leak confirmed that the RTX 5090 will sport 32GB of GDDR7 VRAM. That's a massive upgrade over the previous generation, but the RTX 5080 won't enjoy the same improvements -- the GPU is said to retain both the 16GB memory and the 256-bit bus we've already seen in the RTX 4080 (and its Super version).

Read more
Nvidia may not budge on its VRAM choices
Logo on the RTX 4060 Ti graphics card.

According to new leaks about the RTX 50-series, Nvidia may still keep its most popular GPU starved for VRAM. Wccftech claims that the RTX 5060 will retain an 8GB memory configuration combined with a 128-bit bus. Does this mean that the RTX 5060 won't find its footing among some of the best graphics cards? Not necessarily.

The publication cites its own sources as it reveals some of the specs for Nvidia's more affordable GPUs, ranging from the RTX 5070 Ti to the RTX 5060. And while there are some changes, it does seem that, for the most part, Nvidia is satisfied with its approach to video memory -- which games like Indiana Jones and the Great Circle are constantly putting to the test. Newer AAA games will only push for higher memory capacities, which we may not find in Nvidia's most affordable GPU, but the rest of the stack is looking a little better. Let's go over the specs.

Read more
As a PC gamer, 2024 just made me sad
Fingers holding an Intel 285K.

Fine, I'll say it: 2024 wasn't a great year for PC gaming hardware. I'd even go as far as to call it pretty lame. There were plenty of great PC games to enjoy, but when it comes to hardware, it felt like one big letdown.

A lot of my most anticipated launches ended up getting delayed, and most of the upgrades we got were a bit of a wet blanket. Here are all the various things that proved to be a disappointment in 2024, both to me and to many other PC gamers, but why I'm feeling hopeful for 2025.
The least impressive generation of CPUs in a while

Read more