-
@Seoul National University
- Seoul, Korea
Highlights
- Pro
Stars
ArkVale: Efficient Generative LLM Inference with Recallable Key-Value Eviction (NIPS'24)
Efficient Infinite Context Transformers with Infini-attention Pytorch Implementation + QwenMoE Implementation + Training Script + 1M context keypass retrieval
[NeurIPS'24 Spotlight, ICLR'25, ICML'25] To speed up Long-context LLMs' inference, approximate and dynamic sparse calculate the attention, which reduces inference latency by up to 10x for pre-filli…
Awesome-LLM-KV-Cache: A curated list of 📙Awesome LLM KV Cache Papers with Codes.
📰 Must-read papers and blogs on LLM based Long Context Modeling 🔥
Collection of leaked system prompts
This repository collects debiasing methods for recommendation
Seamless operability between C++11 and Python
Fast hierarchical clustering routines for R and Python.
Performance of various open source GBM implementations
A library to train, evaluate, interpret, and productionize decision forest models such as Random Forest and Gradient Boosted Decision Trees.
A simple and efficient tool to parallelize Pandas operations on all available CPUs
Code and models for ICML 2024 paper, NExT-GPT: Any-to-Any Multimodal Large Language Model
Official implementation of the paper "Acoustic Music Understanding Model with Large-Scale Self-supervised Training".
This package aims at simplifying the download of the AudioCaps dataset.
LP-MusicCaps: LLM-Based Pseudo Music Captioning [ISMIR23]
A feature-rich command-line audio/video downloader
Command-line program to download videos from YouTube.com and other video sites
Emu Series: Generative Multimodal Models from BAAI
Audiocraft is a library for audio processing and generation with deep learning. It features the state-of-the-art EnCodec audio compressor / tokenizer, along with MusicGen, a simple and controllable…
An open source implementation of CLIP.
A configurable, tunable, and reproducible library for CTR prediction https://fuxictr.github.io
[NeurIPS 2023 Datasets and Benchmarks Track] LAMM: Multi-Modal Large Language Models and Applications as AI Agents
🦦 Otter, a multi-modal model based on OpenFlamingo (open-sourced version of DeepMind's Flamingo), trained on MIMIC-IT and showcasing improved instruction-following and in-context learning ability.
✨✨Latest Advances on Multimodal Large Language Models