Stars
Graph-R1: Towards Agentic GraphRAG Framework via End-to-end Reinforcement Learning
Qwen Code is a coding agent that lives in the digital world.
🦜🔗 The platform for reliable agents.
[EMNLP2025] "LightRAG: Simple and Fast Retrieval-Augmented Generation"
拼好RAG:手搓并融合了GraphRAG、LightRAG、Neo4j-llm-graph-builder进行知识图谱构建以及搜索;整合DeepSearch技术实现私域RAG的推理;自制针对GraphRAG的评估框架| Integrate GraphRAG, LightRAG, and Neo4j-llm-graph-builder for knowledge graph construct…
A modular graph-based Retrieval-Augmented Generation (RAG) system
Neo4j graph construction from unstructured data using LLMs
Docs2KG: A Human-LLM Collaborative Approach to Unified Knowledge Graph Construction from Heterogeneous Documents
🎉 A magical vue admin https://panjiachen.github.io/vue-element-admin
black20009 / VCIN
Forked from LivXue/VCINAuthors's code for "Variational Causal Inference Network for Explanatory Visual Question Answering" and "Integrating Neural-Symbolic Reasoning with Variational Causal Inference Network for Explanat…
An Industrial-Level Controllable and Efficient Zero-Shot Text-To-Speech System
Open-source patent documentation: 'A Computing Task Allocation System and Method for Distributed Inference in Large Language Models'.
An open source multi-modal trip planner
GPT4V-level open-source multi-modal model based on Llama3-8B
mPLUG-Owl: The Powerful Multi-modal Large Language Model Family
Superfast AI decision making and intelligent processing of multi-modal data.
A comprehensive list of papers using large language/multi-modal models for Robotics/RL, including papers, codes, and related websites
OmniGen: Unified Image Generation. https://arxiv.org/pdf/2409.11340
🤯 LobeHub - an open-source, modern design AI Agent Workspace. Supports multiple AI providers, Knowledge Base (file upload / RAG ), one click install MCP Marketplace and Artifacts / Thinking. One-cl…
MiniCPM-V 4.5: A GPT-4o Level MLLM for Single Image, Multi Image and High-FPS Video Understanding on Your Phone
🦉 OWL: Optimized Workforce Learning for General Multi-Agent Assistance in Real-World Task Automation
[TPAMI 2024 & CVPR 2023] PyTorch code for DGM4: Detecting and Grounding Multi-Modal Media Manipulation and beyond
Transformer: PyTorch Implementation of "Attention Is All You Need"
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
Plotting heatmaps with the self-attention of the [CLS] tokens in the last layer.
Deep Modular Co-Attention Networks for Visual Question Answering