Our research focuses on multimodal intelligence and perception systems
Research on integrating vision, language, and other modalities for intelligent perception systems.
Developing Vision-Language-Action (VLA) model for autonomous robot.
Advanced research in semantic segmentation, object detection, and scene understanding.
Applying AI and deep learning to medical imaging, diagnosis, and healthcare applications. Delving into genomics and protein design with AI.
Publishing cutting-edge research at top-tier conferences including CVPR, ICCV, ECCV, and NeurIPS.
Working with international collaborators and industry partners on innovative AI solutions.
We welcome a new MS student (Sangjin Lee) and new undergraduate students to MIP Lab.
A paper from MIP Lab is accepted to CVPR 2026. <Delta velocity rectified flow for text-to-image editing, Gaspard Beaudouin, Minghan Li, Jaeyeon Kim, Sung-Hoon Yoon*, Mengyu Wang*>
MIP Lab has been selected as a beta service participant for the 「Advanced GPU Utilization Support Program」.
Multimodal Intelligence and Perception (MIP) Lab has launched at DGIST EECS.