-
Meta
- Portland
- jwfromm.com
Stars
Composable Kernel: Performance Portable Programming Model for Machine Learning Tensor Operators
Universal LLM Deployment Engine with ML Compilation
High-performance In-browser LLM Inference Engine
Bringing stable diffusion models to web browsers. Everything runs inside the browser with no server support.
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
Highly optimized inference engine for Binarized Neural Networks
🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization tools
cmu-catalyst / collage
Forked from apache/tvmSystem for automated integration of deep learning backends.
Transform ONNX model to PyTorch representation
Open deep learning compiler stack for cpu, gpu and specialized accelerators