THE LATEST NEWS
Nvidia Boosts LLM Inference with Open-Source Library

SANTA CLARA, CALIF. — Nvidia has doubled large language model (LLM) inference performance on its H100, A100 and L4 GPUs with a new open-source software library called TensorRT-LLM.

As evidenced by benchmark results that improve round after round for the same hardware, software is often as important as the hardware when it comes to squeezing the best possible performance out of specialized AI chips.

“A huge part of what we do is a combination of hardware and software, and today Nvidia has more software engineers than hardware engineers,” Ian Buck, VP and general manager of Nvidia’s hyperscale and HPC computing business, told EE Times. “This is part of a decision going back to the original CUDA and the motivation around delivering not just a chip with an instruction set, but a complete stack to meet developers where they are.

“This offers an opportunity to innovate at all the levels: change the hardware architecture, change the instruction set, change the compilers, change the drivers, change the tools, the libraries, everything, so we can move the whole platform forward,” he said. “That’s played itself out multiple times in the last 20 years of doing accelerated computing, and it’s true for AI inference too.”

 ----Form EE Times

Back
Imec Makes Solid-State Nanopores Using EUV Lithography
For more than two decades, solid-state nanopores have hovered on the edge of commercial relevance. Researchers have repeatedly demonstrate...
More info
What is Groq-Nvidia Deal Really About?
When news broke late on Christmas Eve that Nvidia had all-but acquired Groq, two things were immediately clear. First, the announcemen...
More info
China EUV Breakthrough and the Rise of the ‘Silicon Curtain’
Inside a secure facility overseen by the Central Science and Technology Commission, Chinese engineers have activated an Extreme Ultr...
More info
0.1287s