女同毛片免费网站_蜜桃视频免费观看视频_亚洲黄色免费观看_狠狠欧美 - 日本一区二区三区在线观看

THE LATEST NEWS
Nvidia Boosts LLM Inference with Open-Source Library

SANTA CLARA, CALIF. — Nvidia has doubled large language model (LLM) inference performance on its H100, A100 and L4 GPUs with a new open-source software library called TensorRT-LLM.

As evidenced by benchmark results that improve round after round for the same hardware, software is often as important as the hardware when it comes to squeezing the best possible performance out of specialized AI chips.

“A huge part of what we do is a combination of hardware and software, and today Nvidia has more software engineers than hardware engineers,” Ian Buck, VP and general manager of Nvidia’s hyperscale and HPC computing business, told EE Times. “This is part of a decision going back to the original CUDA and the motivation around delivering not just a chip with an instruction set, but a complete stack to meet developers where they are.

“This offers an opportunity to innovate at all the levels: change the hardware architecture, change the instruction set, change the compilers, change the drivers, change the tools, the libraries, everything, so we can move the whole platform forward,” he said. “That’s played itself out multiple times in the last 20 years of doing accelerated computing, and it’s true for AI inference too.”

 ----Form EE Times

Back
STMicro Advances PiezoMEMS Development in Singapore
STMicroelectronics, in partnership with Singapore’s A*STAR Institute of Microelectronics (IME), the A*STAR Institute of...
More info
ZeroRISC Gets $10 Million Funding, Says Open-Source Silicon Security ‘Inevitable’
There is often skepticism around the concept of open-source silicon, especially when it comes to security, according to Dominic Rizzo, C...
More info
EnCharge Picks The PC For Its First Analog AI Chip
Analog AI accelerator startup EnCharge AI announced its first product, the 200-TOPS (INT8) EN100 AI accelerator designed f...
More info