Skip to main content

When diagnosis time means life or death, NVIDIA’s advanced AI can save lives

We may buy new smartphones and laptops every year or two, but when it comes expensive medical computers, that’s not an option. There are more than three million medical equipment installed in hospitals today, and more than 100,000 new instruments added each year — that’s according Nvidia CEO Jensen Huang said at the company’s GPU Technology Conference (GTC). At this rate, it would take more than 30 years to replace all the old hospital equipment. So how do we advance medical care without adding more cost?

Nvidia’s technique is to leverage the cloud to provide a “virtual upgrade” to existing medical equipment. Dubbed Project Clara, the medical cloud is described by Huang as a “medical imaging supercomputer that is a data center virtualized, remoted, and is a multi-modality, multi-user supercomputer.”

Recommended Videos

It’s an end-to-end solution leveraging the power of Nvidia’s GPU and its cloud infrastructure, allowing medical practitioners the capability to upload data, analyze and interpret data. Herer at GTC this year, Nvidia is showing off how it uses deep learning to make inferences to detect diseases and pathologies at an earlier state, which could save lives.

Get your weekly teardown of the tech behind PC gaming
Check your inbox!

Early Detection to Save Lives

Nvidia DGX 2
Nvidia CEO Jensen Huang boasts of the GPU’s computing might at GTC 2018. Image used with permission by copyright holder

Early detection is extremely important in the case of sepsis, a disease that claims more lives each year than the combined mortality rates of breast and prostate cancers. It’s also an area of interest to Johns Hopkins University researchers, and Associate Professor of Computer Science Suchi Saria has worked on models to train AI to make early detection a reality. And similar to Huang, Saria’s AI-trained model examines existing patient sensor data without the need for hospitals to purchase more costly equipment to improve early detection and make medical diagnoses.

Deep learning has completely turbocharged modern AI.

This is particularly important because early signs of sepsis are hard to detect, and the condition is often misdiagnosed or ignored until it is too late to treat, Saria said. In fact, for each hour that treatment is delayed — likely as a result of incorrect diagnosis — the mortality rate for sepsis jumps seven to eight percent. Fortunately, though, sepsis is treatable, but only if it’s detected early, Saria noted, highlighting that the condition is the eleventh leading cause of death.

In a case study on sepsis, a female patient was admitted to Johns Hopkins Medical Center for what was believed to be medical pneumonia, Saria said. Doctors administered the usual course of antibiotics, and they weren’t too concerned. Her condition worsens, and on the seventh day, the patient demonstrated visible symptoms of septic shock and was subsequently transferred to the intensive care unit. Once she was in the ICU, her kidneys and lungs began to fail, and the patient passed away on day 22.

“The challenge is that sepsis is extremely hard to diagnose,” Saria said. A study conducted by Harvard University revealed that medical experts weren’t able to agree on early diagnosis of sepsis when presented with symptoms of the condition. However, late stages of sepsis, including septic shock, are easier to identify, but at that point, the mortality rate dramatically jumps.

To make early sepsis detection possible, Saria and her team created a framework called TREWS (pronounced “trues”), which stands for Targeted, Real-Time Early Warning System. TREWS is a class of machine learning that utilizes deep learning to identify symptoms and makes medical diagnosis.

“Deep learning has completely turbocharged modern AI,” Huang exclaimed. “This incredible algorithm that can automatically detect important features out of data, and from this algorithm it can construct hierarchically knowledge representations. And if you feed it more data, it will become more and more robust.”

The earlier the detection, the better.

Making comparisons to a modern smartphone, the Nvidia medical cloud essentially allows hospitals and medical providers to upload existing data and information collected from patients, create models and leverage the power of artificial intelligence. The result is that diseases can be detected earlier, pathologies can be modeled and easier to understand, and scans become richer with more details and information.

The system “leverages high dimensional, noisy health system data to build things that are very patient specific,” explained Saria. “This brings precision into precision healthcare.” Essentially, TREWS takes a look at all the data and “moves from a reactive to a proactive prevention system.”

The challenge with deep learning, Huang explained, is that “it needs a ton of data and a ton of computers.”

In the case of identifying sepsis, Saria relies on historical data from past patients, utilizing a sequential learning framework, and the end result of TREWS is to have the AI system detect sepsis as early as possible and alert doctors, or ask doctors to perform more tests to determine if the patient does in fact have sepsis.

In the example of the patient succumbing to sepsis, AI would have been able to detect sepsis 12 hours before doctors did. But even beyond sepsis, AI could be used to detect all sorts of other medical conditions such as hypertension and heart disease, Saria said.

Inferring More Information

To demonstrate some of the advancements that Nvidia’s hardware and software provides researchers with, AI was used to make inferences to show how a patient’s left ventricle would look in 3D and display data such as the heart’s ejection fraction.

In another example, Philips, makers of ultrasound machines, was able to take a black-and-white ultrasound of a fetus in 2D and reconstruct it into a three-dimensional image complete to give it a more life-like appearance. Additionally, using the GPU’s ray tracing capabilities, the ultrasound scan was visualized as if there was virtual light inside the uterus, complete with some subsurface scattering for tones of flesh.

Outside of Project Clara, Nvidia is also building the hardware needed to make some of these complex processes happen quickly. Medical imaging requires even more powerful hardware, and Huang claims that “the world wants a gigantic GPU.”

To facilitate early disease detection, more comprehensive scans and deep machine learning and artificial intelligence uses, Nvidia introduced the DGX-2 supercomputer at the GPU Technology Conference.

The company states that the DGX-2 is up to ten times faster than the DGX-1, which was introduced a mere six months prior, and the system is capable of replacing 300 dual-core CPU servers valued at $3 million. Essentially, the DGX-2 is an eighth the cost and occupies a sixtieth the space of existing equipment, while only consuming only an eighteenth the power.

“Our strategy at Nvidia is to advance GPU computing for deep learning in AI at the speed of light,” Huang said.

Chuong Nguyen
Silicon Valley-based technology reporter and Giants baseball fan who splits his time between Northern California and Southern…
Windows PCs now works with the Quest 3, and I tried it out for myself
i tried windows new mixed reality link with my quest 3 alan truly sits in front of a pc and adjusts virtual screen while wear

Microsoft and Meta teamed up on a new feature that lets me use my Windows PC while wearing a Quest 3 or 3S, and it’s super easy to connect and use. I simply glance at my computer and tap a floating button to use Windows in VR on large displays only I can see.

Meta’s new Quest 3 and 3S are among the best VR headsets for standalone gaming and media consumption. When I want more performance or need to run one of the best Windows apps that aren’t yet available in VR, I can connect to a much more powerful Windows PC.
Setting up Mixed Reality Link
Scanning Microsoft's Mixed Reality Link QR code with a Meta Quest 3 Photo by Tracey Truly / Digital Trends

Read more
How to transfer your books from Goodreads to StoryGraph
Front page of a book on Onyx BOOX Go 10.3 tablet.

Goodreads has been the only game in town for Android and iOS book-tracking for a long time now, and like most monopolies, it has grown old and fat. Acquired by Amazon in 2013, avid book readers have had lots to complain about in recent years, with the service languishing unloved, with no serious updates and an aging interface. It's been due some serious competition for a long time, and lo and behold, some has arrived. StoryGraph is a book-tracking app that offers everything you'll find on Goodreads but with an algorithm that lets you know about what you might love, and adds features any bibliophile will know are essential — like a Did Not Finish list.

Read more
I played Black Myth: Wukong on the new MSI handheld to prove it was possible
Black Myth: Wukong running on the MSI Claw 8 AI+.

I scoffed when MSI put the Claw 8 AI+ in my hands with Black Myth: Wukong selected. I'd spent 80 hours in the game on my full desktop packing an RTX 4090, and I knew just how demanding the game was. It's a pipedream for a handheld gaming PC.

I pressed Continue and loaded up at the Pool of Shattered Jade rest point -- the ideal spot to farm; if you know, you know -- and proceeded to run up to the cocoons spotted around the area, unleash my spirit ability, and run back. Sitting in a dimly-lit New York City bar, I continued the loop a few more times. I'd done plenty of farming in the game before.

Read more