ONNX Runtime can accelerate training and inferencing popular Hugging Face NLP models.
- General export and inference: Hugging Face Transformers
- Accelerate GPT2 model on CPU
- Accelerate BERT model on CPU
- Accelerate BERT model on GPU
- Blog post: Faster and smaller quantized NLP with Hugging Face and ONNX Runtime
- Blog post: Accelerate your NLP pipelines using Hugging Face Transformers and ONNX Runtime