About Me

Zehan Wang (王泽寒) is a PhD student in the College of Computer Science at Zhejiang University, supervised by Prof. Zhou Zhao. I have published 10+ first-author papers at the top international AI conferences such as NeurIPS/ICLR/ICML/CVPR. My current research interests broadly span two areas:

  1. Unified Multimodal Perception (2D pixels, 3D geometry, language, audio)

  2. Spatial Intelligence in Visual Content (From understanding to generation)

My research is supported by the Fundamental Research Project for Young Ph.D. students from NSFC (国家自然科学基金博士青年基金).

I am actively seeking collaboration opportunities in both academia and industry — feel free to reach out via email!

🔥 News

  • 2025.02: 2 papers accepted by CVPR 2025! (SpatialCLIP)
  • 2025.01: 6 papers are accepted by ICLR 2025! (OmniBind)
  • 2024.12: We release Orient Anything, the foundation model to estimate object orientation in images.
  • 2024.10: 6 papers are accepted by NeurIPS 2024! (Chat-Scene and Ex-MCR)
  • 2024.05: 2 papers are accepted by ICML 2024! (FreeBind)
  • 2023.10: 1 paper is accepted by EMNLP 2023! (3DRP-Net)
  • 2023.09: 1 paper is accepted by NeurIPS 2023! (C-MCR)
  • 2023.06: 2 papers are accepted by ICCV 2023! (WS-3DVG)
  • 2023.05: 1 paper are accepted by ACL 2023!

📝 Representative Publications

Unified Multimodal Perception

  • Unified Representations: C-MCR (NeurIPS 2023), Ex-MCR (NeurIPS 2024), FreeBind (ICML 2024), OmniBind (ICLR 2025)

Spatial Intelligence in Visual Content

  • Point Cloud Understanding: Chat-3D (NAACL 2023) / Chat-Scene (NeurIPS 2024) for 3D MLLM, 3DRP-Net (EMNLP 2023) / WS-3DVG (ICCV 2023) for 3D visual grounding.

  • Spatial-aware Image Understanding: Orient Anything (Arxiv 2025), SpatialCLIP (CVPR 2025)

  • Spatial-aware Image Generation: 6DoF-Gen (Working on), GenSpace (Working on)

Arxiv 2025
sym
CVPR 2025
sym
NeurIPS 2024
sym
ICLR 2025
sym
ICML 2024
sym
NeurIPS 2023
sym
  • Connecting Multi-modal Contrastive Representations Zehan Wang, Yang Zhao, Xize Cheng, Haifeng Huang, Jiageng Liu, Li Tang, Linjun Li, Yongqi Wang, Aoxiong Yin, Ziang Zhang, Zhou Zhao NeurIPS 2023
  • Learning multimodal contrastive representations without requiring paired data.

Full Publication List

2025

2024

2023

📖 Educations

  • 2022.09 - Present, Ph.D Student, Zhejiang University, Hangzhou.

  • 2018.09 - 2022.06, Undergraduate, Zhejiang Univeristy, Hangzhou.

🎖 Honors and Awards

  • 2025: Fundamental Research Project for Young Ph.D. students from NSFC (国家自然科学基金博士青年基金)
  • 2022: Excellent Graduate, Zhejiang Province
  • 2020: National Scholarships
  • 2019: National Scholarships
  • 2019, 2020, 2021: Zhejiang University First-class Scholarships