Stars
[Medical_NLP ➟ Awesome-AI4Med] medical-related LLMs, Multimodal systems, Datasets, Benchmarks, and more.
A Next-Generation Training Engine Built for Ultra-Large MoE Models
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
🔥🔥 LLaVA++: Extending LLaVA with Phi-3 and LLaMA-3 (LLaVA LLaMA-3, LLaVA Phi-3)
verl: Volcano Engine Reinforcement Learning for LLMs
An Easy-to-use, Scalable and High-performance Agentic RL Framework based on Ray (PPO & DAPO & REINFORCE++ & TIS & vLLM & Ray & Async RL)
A one stop repository for generative AI research updates, interview resources, notebooks and much more!
Curated list of data science interview questions and answers
Data processing for and with foundation models! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷
A benchmark for few-shot evaluation of foundation models for electronic health records (EHRs)
✨✨Latest Advances on Multimodal Large Language Models
Multimodal Question Answering in the Medical Domain: A summary of Existing Datasets and Systems
MIMIC Code Repository: Code shared by the research community for the MIMIC family of databases
Tools for curating biomedical training data for large-scale language modeling
Efficient Retrieval Augmentation and Generation Framework
Toolkit for fine-tuning, ablating and unit-testing open-source LLMs.
MedType: Improving Medical Entity Linking with Semantic Type Prediction
[Arxiv-2024] CheXagent: Towards a Foundation Model for Chest X-Ray Interpretation
Meditron is a suite of open-source medical Large Language Models (LLMs).
MedAlign is a clinician-generated dataset for instruction following with electronic medical records.
A quick guide (especially) for trending instruction finetuning datasets
This repository contains code for extending the Stanford Alpaca synthetic instruction tuning to existing instruction-tuned models such as Flan-T5.
Approaching Clinical NER as a MRC problem
Papers and online resources related to machine learning fairness
a library for named entity recognition developed by UF HOBI NLP lab featuring SOTA algorithms
A simple and working implementation of Electra, the fastest way to pretrain language models from scratch, in Pytorch
BertViz: Visualize Attention in Transformer Models