Interactive Multimodal Machine Learning lab.
Career History
- 2006 : B.S. Computer Science and Engineering / Mathematics, POSTECH
- 2009 : M.S. Computer Science, University of Southern California
- 2015 : Ph.D. Computer Science, Toyota Technological Institute at Chicago
- 2008 ~ 2009 : Graduate Programmer, Information Science Institute
- 2014 : Lab Associate(Intern), Disney Research, Pittsburgh
- 2009 ~ 2015 : Research Assistant, Toyota Technological Institute at Chicago
- 2016 ~ 2017 : Post-doc, California Institute of Technology
- 2017 ~ 2019 : Lead Research Scientist, Oben, Inc.
- 2020 ~ 2021 : Applied Scientist, Amazon Alexa AI
- 2021 ~ Present : Assistant Professor, UNIST
Intro
Our lab aims to help understanding and implement human intelligence for most common communication media: vision, natural language, and speech. Since they are connected and correlated to each other, we work on developing effective and efficient machine learning models for multi-modalities. In the Interactive Multimodal Machine Learning Lab lab, we are interested in Machine Learning and applications to Computer Vision and Language Processing. Specifically, we work on Multimodal Learning, Generative Models, and Deep Learning and our research topics include (but not limited to) Multimodal LLM (Vision-Language Model), Embodied AI/Vision Language Action Model, Text-to-image generation, Multi-modal conversational models, Video understanding and question answering, and Explainable AI. Additionally, we also work on AI for Science.
Research Field
Multimodal Learning, Generative AI, Interactive AI
