Stars
A simple, performant and scalable Jax LLM!
JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs welcome).
lightweight, standalone C++ inference engine for Google's Gemma models.
Meaningful control of data in distributed systems.
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
Flax is a neural network library for JAX that is designed for flexibility.
Unsupervised text tokenizer for Neural Network-based text generation.
Official codebase used to develop Vision Transformer, SigLIP, MLP-Mixer, LiT and more.
A high-throughput and memory-efficient inference and serving engine for LLMs
An Open Source Machine Learning Framework for Everyone
Libraries for executing federated programs and computations.
A machine learning compiler for GPUs, CPUs, and ML accelerators