Stars
Pruna is a model optimization framework built for developers, enabling you to deliver faster, more efficient models with minimal overhead.
Concrete ML: Privacy Preserving ML framework using Fully Homomorphic Encryption (FHE), built on top of Concrete, with bindings to traditional ML frameworks.
π€ PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
Track emissions from Compute and recommend ways to reduce their impact on the environment.
Accelerated inference of π€ models using FuriosaAI NPU chips.
π€ Optimum Intel: Accelerate inference with Intel optimization tools
Easy, fast and very cheap training and inference on AWS Trainium and Inferentia chips.
The official Python client for the Hugging Face Hub.
π A compiled checklist of 300+ tips for protecting digital security and privacy in 2024
π Accelerate inference and training of π€ Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization tools
Efficient, scalable and enterprise-grade CPU/GPU inference server for π€ Hugging Face transformer models π
Easy and lightning fast training of π€ Transformers on Habana Gaudi processor (HPU)
π€ Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
Blazing fast training of π€ Transformers on Graphcore IPUs