Highlights
- Pro
Stars
Fully open data curation for reasoning models
Code to reproduce the experiments in the paper Training on the Test Task Confounds Evaluation and Emergence.
Modeling, training, eval, and inference code for OLMo
PyTorch building blocks for the OLMo ecosystem
Meta Lingua: a lean, efficient, and easy-to-hack codebase to research LLMs.
Improving Alignment and Robustness with Circuit Breakers
Machine Learning Engineering Open Book
Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.
Minimalistic large language model 3D-parallelism training
Friends don't let friends make certain types of data visualization - What are they and why are they bad.
A concise but complete full-attention transformer with a set of promising experimental features from various papers
The hub for EleutherAI's work on interpretability and learning dynamics
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
Cramming the training of a (BERT-type) language model into limited compute.
Algorithmically create or extend categorical colour palettes
An autoregressive character-level language model for making more things
Figure sizes, font sizes, fonts, and more configurations at minimal overhead. Fix your journal papers, conference proceedings, and other scientific publications.
Code to run the TILT transfer learning experiments
NAACL 2022: MCSE: Multimodal Contrastive Learning of Sentence Embeddings
A library for parameter-efficient and composable transfer learning for NLP with sparse fine-tunings.
An original implementation of "Noisy Channel Language Model Prompting for Few-Shot Text Classification"
This code accompanies the paper "Bayesian Framework for Information-Theoretic Probing" published in EMNLP 2021.
Train Dense Passage Retriever (DPR) with a single GPU
Run Effective Large Batch Contrastive Learning Beyond GPU/TPU Memory Constraint