Hello, I am a Ph.D student at Machine Learning and Intelligence Lab (MLILAB) in KAIST, advised by Prof. Eunho Yang.

My research focuses on enhancing the efficiency of foundation models, particularly auto-regressive generative models. I aim to improve inference-time efficiency by optimizing memory usage and reducing latency, leveraging techniques such as speculative decoding and knowledge distillation, among others.

Publications

  • LANTERN: Accelerating Visual Autoregressive Models with Relaxed Speculative Decoding [paper]
    Doohyuk Jang*, Sihwan Park*, June Yong Yang, Yeonsung Jung, Jihun Yun, Souvik Kundu, Sungyub Kim, Eunho Yang
    ICLR 2025

  • PromptKD: Distilling Student-Friendly Knowledge for Generative Language Models via Prompt Tuning [paper]
    Gyeongman Kim, Doohyuk Jang, Eunho Yang
    Findings of EMNLP 2024

  • SeamsTalk: Seamless Talking Face Generation via Flow-Guided Inpainting [paper]
    Yeongho Jeong, Gyeongman Kim, Doohyuk Jang, Jaeryong Hwang, Eunho Yang
    IEEE Access 2024

  • Med-PerSAM: One-Shot Visual Prompt Tuning for Personalized Segment Anything Model in Medical Domain [paper]
    Hangyul Yoon, Doohyuk Jang, Jungeun Kim, Eunho Yang
    Under Review

Education

  • Ph.D. in Graduate School of AI, Korea Advanced Institute of Science and Technology (KAIST), Sep. 2024 - Present

  • M.S. in Graduate School of AI, Korea Advanced Institute of Science and Technology (KAIST), Mar. 2023 - Aug. 2024

  • B.S. in Electrical Engineering, Computer Science, Korea Advanced Institute of Science and Technology (KAIST), Mar. 2018 - Feb. 2023

Work Experiences

  • Intern, Synopsys Korea, Gyeonggi, South Korea, Mar 2021 - Aug 2021

Projects

  • Developing a conversational language model for virtual doctors, AITRICS, Apr. 2024 - May. 2024

Acamdeic Services

  • Workshop Reviewer
    • SCOPE@ICLR 2025

Teaching Experience

  • Teaching Assistant, Machine Learning for AI (AI501), KAIST