The TensorRT inference server provides an easy-to-use, production-ready inference solution, enabling simple GPU-acceleration for inferencing in the cloud.