πŸ€” About-me

I earned my Master’s degree in Artificial Intelligence from Tsinghua University , where I conducted research under the supervision of Prof. Haoqian Wang and collaborated closely with Prof. Yebin Liu on 3D computer vision. Prior to this, I completed my B.Eng. in Measurement and Control Technology & Instruments at Southeast University . During my graduate studies, I also had the privilege of visiting Harvard University as a research intern, working with Prof. Hanspeter Pfister on computational imaging projects.

I am currently an Researcher at ByteDance , focusing on cutting-edge challenges in generative AI and embodied intelligence. My work bridges 3D vision with real-world applications, particularly in dynamic scene understanding and human-AI interaction.

Research Directions

Core Expertise: 3D computer vision (NeRF, 3D Gaussian Splatting, multi-view reconstruction).

Emerging Focus: Embodied AI-driven video generation, robot-scene interaction, and physics-aware simulation.

Technical Vision: Building scalable frameworks that connect 3D reconstruction, generative models (video/3D assets), and embodied agents for industrial applications.

Open Opportunities

I am actively recruiting research interns to collaborate on:

πŸ“Œ 3D Content Creation: 3D Reconstruction, Video Generation, 3D Generation

πŸ“Œ 3D Scene Perception: 3D Foundation Model

πŸ“Œ Embodied AI: LLM/Vision-Language models for robot interaction, simulation environments

If you are seeking any form of academic cooperation, please feel free to email me at qinminghan1999@gmail.com.

πŸ”₯ News

  • 2025.02: Β πŸŽ‰πŸŽ‰ 2 paper accepted to CVPR 2025 !!!
  • 2024.09: Β πŸŽ‰πŸŽ‰ 1 paper accepted to NeurIPS 2024 !!!
  • 2024.07: Β πŸŽ‰πŸŽ‰ 1 paper accepted to ACM MM 2024 !!!
  • 2024.02: Β πŸŽ‰πŸŽ‰ 2 paper accepted to ECCV 2024 !!!
  • 2024.02: Β πŸŽ‰πŸŽ‰ LangSplat has been selected as CVPR 2024 Highlight !!!
  • 2024.02: Β πŸŽ‰πŸŽ‰ 1 paper accepted to CVPR 2024 !!!
  • 2023.11: Β πŸŽ‰πŸŽ‰ 1 paper accepted to AAAI 2024 !!!

πŸ“ Selected Publications

CVPR 2024 Highlight
sym

LangSplat: 3D Language Gaussian Splatting

Minghan Qin*, Wanhua Li*†, Jiawei Zhou*, Haoqian Wang†, Hanspeter Pfister

Website

minghanqin%2FLangSplat | Trendshift

  • We introduces LangSplat, which constructs a 3D language field that enables precise and efficient open-vocabulary querying within 3D spaces.
CVPR 2025
sym

4D LangSplat: 4D Language Gaussian Splatting via Multimodal Large Language Models

Wanhua Li, Renping Zhou, Jiawei Zhou, Yingwei Song, Johannes Herter, Minghan Qin, Gao Huang, Hanspeter Pfister

Website

  • We present 4D LangSplat, an approach to constructing a dynamic 4D language field in evolving scenes, leveraging Multimodal Large Language Models.
CVPR 2025
sym

HRAvatar: High-Quality and Relightable Gaussian Head Avatar

Dongbin Zhang, Yunfei Liu, Lijian Lin, Ye Zhu, Kangjie Chen, Minghan Qin, Yu Li†, Haoqian Wang†

Website

  • With monocular video input, HRAvatar reconstructs a high-quality, animatable 3D head avatar that enables realistic relighting effects and simple material editing.
NeurIPS 2024
sym

HDR-GS: Efficient High Dynamic Range Novel View Synthesis at 1000x Speed via Gaussian Splatting

Yuanhao Cai , Zihao Xiao, Yixun Liang, Minghan Qin, Yulun Zhang, Xiaokang Yang, Yaoyao Liu, Alan Yuille

Website

  • The first 3D Gaussian splatting-based method for high dynamic range imaging
ACM MM 2024
sym

Animatable 3d gaussian: Fast and high-quality reconstruction of multiple human avatars

Yang Liu, Xiang Huang, Minghan Qin, Qinwei Lin, Haoqian Wang (* indicates equal contribution)

Website

  • We propose Animatable 3D Gaussian, a novel neural representation for fast and high-fidelity reconstruction of multiple animatable human avatars, which can animate and render the model at interactive rate.
ECCV 2024
sym

Gaussian in the Wild: 3D Gaussian Splatting for Unconstrained Image Collections

Dongbin Zhang*, Chuming Wang*, Weitao Wang, Peihao Li, Minghan Qin, Haoqian Wang†

Website

  • We utilize 3D Gaussian Splatting with introduced separated intrinsic and dynamic appearance to reconstruct scenes from uncontrolled images, achieving high-quality results and a 1000 Γ— rendering speed increase.
ECCV 2024
sym

Category-level Object Detection, Pose Estimation and Reconstruction from Stereo Images

Chuanrui Zhang*, Yonggen Ling*†, Minglei Lu, Minghan Qin, Haoqian Wang†

Website Datasets

  • We present CODERS, a one-stage approach for Category-level Object Detection, pose Estimation and Reconstruction from Stereo images.
AAAI 2024
sym

High-Fidelity 3D Head Avatars Reconstruction through Spatially-Varying Expression Conditioned Neural Radiance Field

Minghan Qin*, Yifan Liu*, Yuelang Xu, Xiaochen Zhao, Yebin Liu†, Haoqian Wang†

Website

  • We introduce a novel Spatially-Varying Expression (SVE) conditioning, encompassing both spatial positional features and global expression information.

πŸŽ– Honors and Awards

πŸ’» Research Experience

  • 2023.09 - 2024.4, Harvard University - VCG Lab - Computer Vision Group. I spent a good time with Prof. Hanspeter Pfister.

πŸ’ Academic Service

Reviewers of: CVPR, ECCV, ICCV, NeurIPS, ACM MM, AAAI