-
Shanghai Jiao Tong University
- Shanghai, China
-
03:06
(UTC +08:00) - gszfwsb.github.io
- @ShaoboWang6
Highlights
- Pro
Lists (11)
Sort Name ascending (A-Z)
Starred repositories
Agent framework and applications built upon Qwen>=3.0, featuring Function Calling, MCP, Code Interpreter, RAG, Chrome extension, etc.
Tools for merging pretrained large language models.
The official implementation for [NeurIPS2025 Oral] Gated Attention for Large Language Models: Non-linearity, Sparsity, and Attention-Sink-Free
Enjoy the magic of Diffusion models!
Search Self-Play: Pushing the Frontier of Agent Capability without Supervision
Scaling Preference Data Curation via Human-AI Synergy
Code repository for Group-MATES Group-Level Data Selection for Efficient Pretraining
gszfwsb / Socratic-Zero
Forked from Frostlinx/Socratic-ZeroSocratic-Zero is a fully autonomous framework that generates high-quality training data for mathematical reasoning
This is the official implementation for Human-Agent Collaborative Paper-to-Page Crafting for Under $0.1.
ERGO (Efficient Reasoning & Guided Observation) is a large vision–language model trained with reinforcement learning on efficiency objectives.
Official implementation of paper: SFT Memorizes, RL Generalizes: A Comparative Study of Foundation Model Post-training
2026 AI/ML internship & new graduate job list updated daily
The best repository showing why transformers might not be the answer for time series forecasting and showcasing the best SOTA non transformer models.
Towards a Unified View of Large Language Model Post-Training
[Survey] A Comprehensive Survey of Self-Evolving AI Agents: A New Paradigm Bridging Foundation Models and Lifelong Agentic Systems
Official PyTorch Implementation of "Diffusion Transformers with Representation Autoencoders"
An Efficient and User-Friendly Scaling Library for Reinforcement Learning with Large Language Models
Search-R1: An Efficient, Scalable RL Training Framework for Reasoning & Search Engine Calling interleaved LLM based on veRL
A version of verl to support diverse tool use
Pretraining and inference code for a large-scale depth-recurrent language model