Hi, I'm Youngmin Kim. I'm a researcher at Yonsei University, MIRLAB (Multimodal Intelligance Research Lab) advised by Youngjae Yu. I received my bachelor's degree in Economics and Computer Science, and I am currently pursuing an integrated MS/Ph.D program in Artificial Intelligence.
My research question is "How can we enable AI systems to deeply understand and extract meaningful information from videos, and how can this understanding be effectively connected to human perception and communication?". I'm deeply interested in how AI can comprehend the complex and rich information embedded in videos, and how this understanding can facilitate natural interactions between humans and AI. Ultimately, I see AI as a tool designed to improve people’s lives, and I believe human-AI interaction should focus on making AI more accessible and effective for users. With this perspective, I'm particularly interested in the understanding and generation of nonverbal expressions, as well as the recognition and production of sign language.
MICCAI 2025
TLDR; We introduce ScalpVision, an AI system for comprehensive scalp disease and alopecia diagnosis that uses innovative hair segmentation and DiffuseIT-M, a generative model for dataset augmentation, to improve severity assessment and prediction accuracy.
Sensors (IF: 3.847)
TLDR; We introduce the effective preprocessing pipeline for sign language translation without glosses, combining skeleton-based motion features, keypoint normalization, and stochastic frame selection to enhance model performance.
IEEE BigData 2022
TLDR; We introduce a two-stage vehicle class and orientation detection model using synthetic-to-real image translation and meta-table fusion to improve real-world prediction accuracy.
Korea Information Processing Society (KIPS)
TLDR; (Korean) We introduce a real-time tram collision prediction system that combines fast object detection with YOLOv5 and a modified local dense optical flow to estimate object speed and predict collision time and probability using a single camera image.
Korea Computer Congress 2021 (🥇Best Paper Award)
TLDR; (Korean) We introduce a real-time pedestrian collision prediction system that uses YOLOv5 for fast object detection and a Local Dense Optical Flow method to quickly estimate pedestrian direction and speed, enabling accurate prediction of collision time and location.
Korea Information Processing Society (KIPS) Special Session
TLDR; (Korean) We survey recent advances in optical flow estimation, comparing traditional and deep learning-based methods, and highlight their applications in autonomous driving, medical imaging, and surveillance systems.