-
๐ซ How to reach me: [email protected]
-
๐ Main interest: AI, Computer Vision, Machine Learning ...
-
๐ฏ More about me: https://xingruiwang.github.io/
CS PhD @ Johns Hopkins University
Multimodal & Spatial reasoning
-
Johns Hopkins University
- Baltimore, MD
-
19:46
(UTC -05:00) - https://xingruiwang.github.io/
- @XingruiWang
Highlights
- Pro
Pinned Loading
-
Spatial457
Spatial457 Public[CVPR'25] A vision question answering (VQA) benchmark for 6D spatial reasoning.
-
open-compass/VLMEvalKit
open-compass/VLMEvalKit PublicOpen-source evaluation toolkit of large multi-modality models (LMMs), support 220+ LMMs, 80+ benchmarks
-
DynSuperCLEVR
DynSuperCLEVR PublicA video question answering dataset that focuses on the dynamics properties of objects (velocity, acceleration) and their collisions within 4D scenes.
Python 18
-
3D-Aware-VQA
3D-Aware-VQA PublicOfficial Code for the NeurIPS'23 paper "3D-Aware Visual Question Answering about Parts, Poses and Occlusions"
Jupyter Notebook 19
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.