All Inference optimization Listings
Loading
...
Loading
...
Loading
...
Loading
...
Loading
...
Loading
...
...
...
...
...
...
...
...
...
...
...
...
...
Vector (and Scalar) Quantization, in Pytorch
[CVPR 2023] DepGraph: Towards Any Structural Pruning
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
Finetune Llama 3.3, Mistral, Phi, Qwen 2.5 & Gemma LLMs 2-5x faster with 80% less memory
[ICML 2023] SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
Listings per page