Platform
Platform
Resources
Resources
Pricing
Pricing
Docs
Docs
Log in
Get started
Pankaj Gupta
Co-Founder
Model performance
FP8: Efficient model inference with 8-bit floating point numbers
Pankaj Gupta
1 other
Model performance
40% faster Stable Diffusion XL inference with NVIDIA TensorRT
Pankaj Gupta
2 others
Model performance
Unlocking the full power of NVIDIA H100 GPUs for ML inference with TensorRT
Pankaj Gupta
1 other
Model performance
Faster Mixtral inference with TensorRT-LLM and quantization
Pankaj Gupta
2 others
Infrastructure
Technical deep dive: Truss live reload
Pankaj Gupta
1
2
Explore Baseten today
Start deploying
Talk to an engineer