Visual Artificial Intelligence
Visual Artificial Intelligence
Visual Artificial Intelligence
Visual Artificial Intelligence

About Us


Panoptic Captioning: Seeking an equivalency bridge
between image and text for comprehensive understanding.
Fin3R: Fine-tuning feed-forward 3D reconstruction
models via monocular knowledge distillation.
Wukong's 72 Transformations: High-fidelity
3D morphing via flow models.
Inpaint4Drag: Repurposing inpainting models for
drag-based image editing via bidirectional warping.
GAMEBot: Transparent assessment of LLM
reasoning capabilities in games.

Visual AI Lab (VAIL) is a research group directed by Prof. Kai Han, working on computer vision, machine learning, and artificial intelligence, at School of Computing and Data Science, The University of Hong Kong. The overarching goal of our research is to achieve principled and comprehensive visual understanding, close the intelligence gap between machines and humans, and build reliable AI systems for open-world use. Our current research focuses on open-world learning, 3D vision, generative AI, foundation models and their relevant fields. Our goal is to achieve principled and comprehensive visual understanding, close the intelligence gap between machines and humans, and build reliable AI systems for open-world use.

🚩 Openings:
(1) PhD students: We are always looking for strong students to work on exciting research problems (☞ fellowships and scholarships).
(2) Postdocs: Positions in computer vision and deep learning are available.
Please drop Prof. Kai Han an email with your resume if you are interested in working with us.

News and Updates


Nov 2025:

One paper on semantic correspondence is accepted to TPAMI.
Sept 2025:

Seven papers (Panoptic Captioning, 3DRS, Fin3R, Wukong, VaMP, SEAL, GSPN-2) are accepted to NeurIPS 2025.
Aug 2025:

Prof. Han will serve as an Area Chair for CVPR 2026, an Area Chair for ICLR 2026.
June 2025:

Two papers (Inpaint4Drag, GRAB) are accepted to ICCV 2025.
June 2025:

Prof. Han will serve as an Area Chair for AAAI 2026.
May 2025:

Two papers (GAMEBot, PruneVid) are accepted to ACL 2025.
Mar 2025:

Splat4D is accepted to SIGGRAPH 2025.
Feb 2025:

Six papers (ICE, HypCD, Mr. DETR, v-CLR, GSPN, PASS) are accepted to CVPR 2025.
Jan 2025:

Five papers (HiLo, DebGCD, BiGR, Needle Threading, AvatarGO) are accepted to ICLR 2025.
Sept 2024:

SciFIBench is accepted to NeurIPS 2024.
Sept 2024:

Prof. Han will serve as an Area Chair for ICLR 2025 and an Area Chair for CVPR 2025
Aug 2024:

One paper on dissecting OOD and OSR is accepted to IJCV.
July 2024:

Three papers (RegionDrag, PromptCCD, and ConceptExpress) are accepted to ECCV 2024.
March 2024:

Three papers (IBD-SLAM, DreamAvatar, and SD4Match) are accepted to CVPR 2024.
Feb 2024:

CiPR is accepted to TMLR 2024.
Jan 2024:

Two papers (on generalized category discovery/open-vocabulary action recognition) are accepted to ICLR 2024.
Oct 2023:

Prof. Han will serve as an Area Chair for ECCV 2024.
Sept 2023:

One paper on text-guided 3D head avatar generation and editing is accepted to NeurIPS 2023.
Aug 2023:

One paper on visual correspondence is accepted to TPAMI.
July 2023:

Two papers (on generalized category discovery/open-vocabulary semantic segmentation) are accepted to ICCV 2023.
July 2023:

Prof. Han will serve as an Area Chair for CVPR 2024.
Mar 2023:

We are organizing OOD-CV workshop @ ICCV 2023. Welcome participants from all over!
Feb 2023:

Two papers (on compositional zero-shot learning/3D human digitization) are accepted to CVPR 2023.
Jul 2022:

One paper on novel category discovery without forgetting is accepted to ECCV 2022.
Jun 2022:

Best Paper Runner-Up Award at CVPR 2022 Workshop on Continual Learning in Computer Vision.
Mar 2022:

Three papers (about generalized category discovery/3D human reconstruction/instance segmentation) are accepted to CVPR 2022.
Jan 2022:

One paper about open-set recognition is accepted to ICLR 2022.
Sep 2021:

One paper about novel category discovery is accepted to NeurIPS 2021.