Hi, I'm Chan Hee (Luke) Song.
I am a CS PhD student at The Ohio State University advised by Yu Su.
My research focuses on multi-modality, particularly embodied AI and multimodal foundation models.
During my undergraduate at Notre Dame, I was part of the ND NLP.
I have interned at Nvidia Research and Adobe Research.
Jun 2024
BioCLIP won the best student paper award at CVPR 2024! Honored to be part of the team.
Feb 2024
I will be interning at Nvidia Learning and Perception Research Group this summer. Catch me in Seattle!
Jul 2023
LLM-Planner, a paper on using large language models for vison-and-language navigation accepted to ICCV 2023.
Mar 2023
Our SalsaBot work for Amazon Alexa Prize Challenge has been accepted to the Embodied AI Workshop at CVPR 2023!
Mar 2023
I will be interning at Adobe Research this summer. Catch me in San Jose!
See full list in Publications.
RoboSpatial: Teaching Spatial Understanding to 2D and 3D Vision-Language Models for Robotics
preprint
BioCLIP: A Vision Foundation Model for the Tree of Life
Dual-View Visual Contextualization for Web Navigation
CVPR 2024 Paper
LLM-Planner: Few-Shot Grounded Planning for Embodied Agents with Large Language Models
One Step at a Time: Long-Horizon Vision-and-Language Navigation with Milestones
CVPR 2022 Paper
Feel free to contact me if you are interested in my research or want to discuss anything :)