Jaewoo Ahn

jaewoo.ahn AT vision.snu.ac.kr

prof_outside.jpg

Hi, I’m a Ph.D. candidate in the Department of Computer Science and Engineering at Seoul National University where I am advised by Prof. Gunhee Kim as a member of Vision & Learning Lab. In addition, I am currently a Research Scientist Intern at KRAFTON AI.

I’m interested in Natural Language Processing (NLP) and Multimodal Artificial Intelligence (Multimodal AI), with the long‑term goal of building human-like embodied conversational agents capable of both talking and acting naturally across diverse (e.g., chat, game, embodied) environments to assist humans as cognitive partners.

My current research interests include, but are not limited to:

  • Conversational agents for social interaction/simulation
  • Robust multimodal (vision-language) perception and reasoning
  • Systematic evaluation of cognitive behaviors in language models and agents
  • Post-training LLM/VLM agents for sequential decision making (e.g., multi-turn RL)

You can refer to my Research Statement: Towards Coherent Embodied Conversational Agent, if interested in.

News

Oct 09, 2025 I will be a keynote speaker at the Wordplay Workshop @ EMNLP 2025!
Sep 27, 2025 Our FlashAdventure & Orak papers were accepted (as Spotlight & Outstanding, respectively) and will be presented at the Wordplay Workshop @ EMNLP 2025!
Aug 20, 2025 Our FlashAdventure paper got accepted to EMNLP 2025!
Jul 24, 2025 Our ChartCap paper got accepted to ICCV 2025 as a Highlight Poster!
Jun 05, 2025 Our Orak benchmark for video game agents is released!

Publications (* equal contribution)

  1. FlashAdventure: A Benchmark for GUI Agents Solving Full Story Arcs in Diverse Adventure Games
    In EMNLP, 2025
  2. Orak: A Foundational Benchmark for Training and Evaluating LLM Agents on Diverse Video Games
    In Wordplay @ EMNLP, 2025 Outstanding
  3. ChartCap: Mitigating Hallucination of Dense Chart Captioning
    Junyoung Lim, Jaewoo Ahn, and Gunhee Kim
    In ICCV, 2025 Highlight
  4. Can LLMs Deceive CLIP? Benchmarking Adversarial Compositionality of Pre-trained Multimodal Representation via Text Updates
    Jaewoo Ahn*Heeseung Yun*Dayoon Ko, and Gunhee Kim
    In ACL, 2025
  5. Is a Peeled Apple Still Red? Evaluating LLMs’ Ability for Conceptual Combination with Property Type
    In NAACL, 2025 Oral
  6. TimeChara: Evaluating Point-in-Time Character Hallucination of Role-Playing Large Language Models
    In ACL Findings, 2024
  7. Who Wrote this Code? Watermarking for Code Generation
    In ACL, 2024
  8. mRedditSum: A Multimodal Abstractive Summarization Dataset of Reddit Threads with Images
    In EMNLP, 2023
  9. MPCHAT: Towards Multimodal Persona-Grounded Conversation
    Jaewoo Ahn, Yeda SongSangdoo Yun, and Gunhee Kim
    In ACL, 2023
  10. Sequential Latent Knowledge Selection for Knowledge-Grounded Dialogue
    Byeongchang Kim, Jaewoo Ahn, and Gunhee Kim
    In ICLR, 2020 Spotlight