NVIDIA Announces TensorRT 6; Breaks 10 Millisecond Barrier for BERT-Large

Today, NVIDIA released TensorRT 6 which includes new capabilities that dramatically accelerate conversational AI applications, speech recognition, 3D image segmentation for medical applications, as well as image-based applications in industrial automation. TensorRT is a high-performance deep learning inference optimizer and runtime that delivers low latency, high-throughput inference for AI applications. With today’s release, TensorRT continues

Continue reading NVIDIA Announces TensorRT 6; Breaks 10 Millisecond Barrier for BERT-Large