Nvidia wants to revolutionize the software microservices world — and unsurprisingly, it’s starting with AI deployment

New NIM tools look to simplify AI deployment

When you purchase through links on our site, we may earn an affiliate commission.Here’s how it works.

In a bid to help businesses of all sizes embrace the new AI-driven world,Nvidiahas taken the wraps off a new approach for software and microservice access that it says could change everything.

The company’s Nvidia Inference Microservices, or NIM, offerings will look to replace the myriad of code and services currently needed to create or run software.

Instead, a NIM will collate together a collection of containerized models and their dependencies in a single package, which can then be distributed and deployed where needed.

NIM-ble

NIM-ble

In his keynote speech at the recentNvidia GTC 2024event, company CEO Jensen Huang said that the new approach signals a shift change for businesses everywhere.

“It is unlikely that you’ll write it from scratch or write a whole bunch of Python code or anything like that,” Huang said. “It is very likely that you assemble a team of AI.”

“This is how we’re going to write software in the future.”

Huang noted thatAI toolsand LLMs will likely be a common sight in NIM deployments as companies across the world look to embrace the latest technologies. He gave one example of how Nvidia itself is using one such NIM to create an internalchatbotdesigned to solve common problems encountered when building chips, helping improve knowledge and capabilities across the board.

Are you a pro? Subscribe to our newsletter

Are you a pro? Subscribe to our newsletter

Sign up to the TechRadar Pro newsletter to get all the top news, opinion, features and guidance your business needs to succeed!

Nvidia adds that NIMs built for portability and control, and can be deployed across not only cloud, but also on-premise data centers and even local workstations, including its RTX workstations and PCs as well as its DGX and DGX Cloud services.

Developers can access AI models through APIs that adhere to the industry standards for each domain, simplifying application development, and NIM will be available as part of Nvidia AI Enterprise, the company’s new platform and hub for AI services, offering a one-stop shop for businesses to understand and access new tools, with NIM AI use cases spreading across LLMs, VLMS, drug discovery, medical imaging and more.

More from TechRadar Pro

Mike Moore is Deputy Editor at TechRadar Pro. He has worked as a B2B and B2C tech journalist for nearly a decade, including at one of the UK’s leading national newspapers and fellow Future title ITProPortal, and when he’s not keeping track of all the latest enterprise and workplace trends, can most likely be found watching, following or taking part in some kind of sport.

This new malware utilizes a rare programming language to evade traditional detection methods

Google puts Nvidia on high alert as it showcases Trillium, its rival AI chip, while promising to bring H200 Tensor Core GPUs within days

Arcane season 2 confirms the hit series isn’t just one of the best Netflix shows ever made – it’s an animated legend that’ll stand the test of time