-
Xihua University
- Tokyo
Highlights
- Pro
Starred repositories
Learning to Relate to Previous Turns in Conversational Search. A codebase for KDD 2023 accepted paper.
Codebase for Merging Language Models (ICML 2024)
本项目旨在分享大模型相关技术原理以及实战经验(大模型工程化、大模型应用落地)
MNBVC(Massive Never-ending BT Vast Chinese corpus)超大规模中文语料集。对标chatGPT训练的40T数据。MNBVC数据集不但包括主流文化,也包括各个小众文化甚至火星文的数据。MNBVC数据集包括新闻、作文、小说、书籍、杂志、论文、台词、帖子、wiki、古诗、歌词、商品介绍、笑话、糗事、聊天记录等一切形式的纯文本中文数据。
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
Official code for the NAACL 2022 paper "Fuse It More Deeply! A Variational Transformer with Layer-Wise Latent Variable Inference for Text Generation"
程序员延寿指南 | A programmer's guide to live longer
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
[ICLR'23] DiffuSeq: Sequence to Sequence Text Generation with Diffusion Models
Code for ACL'2021 paper WARP 🌀 Word-level Adversarial ReProgramming. Outperforming `GPT-3` on SuperGLUE Few-Shot text classification. https://aclanthology.org/2021.acl-long.381/