Seonghyeon Ye

Publications


2026

  • DreamZero: World Action Models are Zero-shot Policies
    Seonghyeon Ye†, Yunhao Ge*, Kaiyuan Zheng*, Shenyuan Gao*, Sihyun Yu*, George Kurian*, Suneel Indupuru*, You Liang Tan*, Chuning Zhu, Jiannan Xiang, Ayaan Malik, Kyungmin Lee, William Liang, Nadun Ranawaka, Jiasheng Gu, Yinzhen Xu, Guanzhi Wang, Fengyuan Hu, Avnish Narayan, Johan Bjorck, Jing Wang, Gwanghyun Kim, Dantong Niu, Ruijie Zheng, Yuqi Xie, Jimmy Wu, Qi Wang, Ryan Julian, Danfei Xu, Yilun Du, Yevgen Chebotar, Scott Reed, Jan Kautz, Yuke Zhu†, Linxi "Jim" Fan†, Joel Jang†
    [paper] [code][website]
  • DreamDojo: A Generalist Robot World Model from Large-Scale Human Videos
    Shenyuan Gao†, William Liang†, Kaiyuan Zheng*, Ayaan Malik*, Seonghyeon Ye, Sihyun Yu, Wei-Cheng Tseng, Yuzhu Dong, Kaichun Mo, Chen-Hsuan Lin, Qianli Ma, Seungjun Nah, Loic Magne, Jiannan Xiang, Yuqi Xie, Ruijie Zheng, Dantong Niu, You Liang Tan, K.R. Zentner, George Kurian, Suneel Indupuru, Pooya Jannaty, Jinwei Gu, Jun Zhang, Jitendra Malik, Pieter Abbeel, Ming-Yu Liu, Yuke Zhu†, Joel Jang†, Linxi "Jim" Fan†
    [paper] [code] [website]

2025

  • DreamGen: Unlocking Generalization in Robot Learning through Neural Trajectories
    Joel Jang*, Seonghyeon Ye*, Zongyu Lin*, Jiannan Xiang*, Johan Bjorck, Yu Fang, Fengyuan Hu, Spencer Huang, Kaushil Kundalia, Yen-Chen Lin, Loic Magne, Ajay Mandlekar, Avnish Narayan, You Liang Tan, Guanzhi Wang, Jing Wang, Qi Wang, Yinzhen Xu, Xiaohui Zeng, Kaiyuan Zheng, Ruijie Zheng, Ming-Yu Liu, Luke Zettlemoyer, Dieter Fox, Jan Kautz, Scott Reed*, Yuke Zhu*, Linxi "Jim" Fan*
    CoRL 2025
    [paper] [website]
  • Comsos-Predict2.5: World Simulation with Video Foundation Models for Physical AI
    NVIDIA
    Tech Report
    [paper] [code]
  • FLARE: Robot learning with implicit world modeling
    Ruijie Zheng*, Jing Wang*, Scott Reed*, Johan Bjorck, Yu Fang, Fengyuan Hu, Joel Jang, Kaushil Kundalia, Zongyu Lin, Loic Magne, Avnish Narayan, You Liang Tan, Guanzhi Wang, Qi Wang, Jiannan Xiang, Yinzhen Xu, Seonghyeon Ye, Jan Kautz, Furong Huang, Yuke Zhu, Linxi Fan
    CoRL 2025
    [paper] [website]
  • GR00T N1: An Open Foundation Model for Generalist Humanoid Robots
    NVIDIA GEAR
    Tech Report
    [paper] [code] [website]
  • Magma: A Foundation Model for Multimodal AI Agents
    Jianwei Yang, Reuben Tan, Qianhui Wu, Ruijie Zheng, Baolin Peng, Yongyuan Liang, Yu Gu, Mu Cai, Seonghyeon Ye, Joel Jang, Yuquan Deng, Lars Liden, Jianfeng Gao
    CVPR 2025
    [paper] [code] [website]
  • Latent Action Pretraining from Videos
    Seonghyeon Ye*, Joel Jang*, Byeongguk Jeon, Sejune Joo, Jianwei Yang, Baolin Peng, Ajay Mandlekar, Reuben Tan, Yu-Wei Chao, Bill Yuchen Lin, Lars Liden, Kimin Lee*, Jianfeng Gao*, Luke Zettlemoyer*, Dieter Fox*, Minjoon Seo*
    ICLR 2025
    LangRob Workshop @ CoRL 2024 Best Paper
    [paper] [code] [website]
  • The biggen bench: A principled benchmark for fine-grained evaluation of language models with language models
    Seungone Kim, Juyoung Suk, Ji Yong Cho, Shayne Longpre, Chaeeun Kim, Dongkeun Yoon, Guijin Son, Yejin Cho, Sheikh Shafayat, Jinheon Baek, Sue Hyun Park, Hyeonbin Hwang, Jinkyung Jo, Hyowon Cho, Haebin Shin, Seongyun Lee, Hanseok Oh, Noah Lee, Namgyu Ho, Se June Joo, Miyoung Ko, Yoonjoo Lee, Hyungjoo Chae, Jamin Shin, Joel Jang, Seonghyeon Ye, Bill Yuchen Lin, Sean Welleck, Graham Neubig, Moontae Lee, Kyungjae Lee, Minjoon Seo
    NAACL 2025 Best Paper
    [paper] [code]
  • Bridging the Data Provenance Gap Across Text, Speech and Video
    Shayne Longpre, Nikhil Singh, Manuel Cherep, Kushagra Tiwary, Joanna Materzynska, William Brannon, Robert Mahari, Manan Dey, Mohammed Hamdy, Nayan Saxena, Ahmad Mustafa Anis, Emad A Alghamdi, Vu Minh Chien, Naana Obeng-Marnu, Da Yin, Kun Qian, Yizhi Li, Minnie Liang, An Dinh, Shrestha Mohanty, Deividas Mataciunas, Tobin South, Jianguo Zhang, Ariel N Lee, Campbell S Lund, Christopher Klamm, Damien Sileo, Diganta Misra, Enrico Shippole, Kevin Klyman, Lester JV Miranda, Niklas Muennighoff, Seonghyeon Ye, Seungone Kim, Vipul Gupta, Vivek Sharma, Xuhui Zhou, Caiming Xiong, Luis Villa, Stella Biderman, Alex Pentland, Sara Hooker, Jad Kabbara
    ICLR 2025
    [paper]

2024

  • How Do Large Language Models Acquire Factual Knowledge During Pretraining?
    Hoyeon Chang, Jinho Park, Seonghyeon Ye, Sohee Yang, Youngkyung Seo, Du-Seong Chang, Minjoon Seo
    NeurIPS 2024
    [paper] [code]
  • Consent in Crisis: The Rapid Decline of the AI Data Commons
    Shayne Longpre, Robert Mahari, Ariel Lee, Campbell Lund, Hamidah Oderinwale, William Brannon, Nayan Saxena, Naana Obeng-Marnu, Tobin South, Cole Hunter, Kevin Klyman, Christopher Klamm, Hailey Schoelkopf, Nikhil Singh, Manuel Cherep, Ahmad Anis, An Dinh, Caroline Chitongo, Da Yin, Damien Sileo, Deividas Mataciunas, Diganta Misra, Emad Alghamdi, Enrico Shippole, Jianguo Zhang, Joanna Materzynska, Kun Qian, Kush Tiwary, Lester Miranda, Manan Dey, Minnie Liang, Mohammed Hamdy, Niklas Muennighoff, Seonghyeon Ye, Seungone Kim, Shrestha Mohanty, Vipul Gupta, Vivek Sharma, Vu Minh Chien, Xuhui Zhou, Yizhi Li, Caiming Xiong, Luis Villa, Stella Biderman, Hanlin Li, Daphne Ippolito, Sara Hooker, Jad Kabbara, Sandy Pentland
    NeurIPS 2024
    [paper]
  • Instruction Matters, a Simple yet Effective Task Selection Approach in Instruction Tuning for Specific Tasks
    Changho Lee, Janghoon Han, Seonghyeon Ye, Stanley Jungkyu Choi, Honglak Lee, Kyunghoon Bae
    EMNLP 2024
    [paper] [code]
  • Self-Explore: Enhancing Mathematical Reasoning in Language Models with Fine-grained Rewards
    Hyeonbin Hwang, Doyoung Kim, Seungone Kim, Seonghyeon Ye, Minjoon Seo
    EMNLP 2024 Findings
    [paper] [code]
  • FLASK: Fine-grained Language Model Evaluation based on Alignment Skill Sets
    Seonghyeon Ye*, Doyoung Kim*, Sungdong Kim, Hyeonbin Hwang, Seungone Kim, Yongrae Jo, James Thorne, Juho Kim, Minjoon Seo
    ICLR 2024 Spotlight
    [paper] [code]
  • Improving Probability-based Prompt Selection Through Unified Evaluation and Analysis
    Sohee Yang, Jonghyeon Kim, Joel Jang, Seonghyeon Ye, Hyunji Lee, Minjoon Seo
    TACL 2024
    [paper] [code]
  • Investigating the Effectiveness of Task-Agnostic Prefix Prompt for Instruction Following
    Seonghyeon Ye, Hyeonbin Hwang, Sohee Yang, Hyeongu Yun, Yireun Kim, Minjoon Seo
    AAAI 2024
    [paper] [code]
  • Carpe Diem: On the Evaluation of World Knowledge in Lifelong Language Models
    Yujin Kim, Jaehong Yoon, Seonghyeon Ye, Sung Ju Hwang, Se-young Yun
    NAACL 2024
    [paper]
  • Instructir: A benchmark for instruction following of information retrieval models
    Hanseok Oh, Hyunji Lee, Seonghyeon Ye, Haebin Shin, Hansol Jang, Changwook Jun, Minjoon Seo
    KnowledgeNLP Workshop @ ACL24
    [paper]

2023

  • The CoT Collection: Improving Zero-shot and Few-shot Learning of Language Models via Chain-of-Thought Fine-tuning
    Seungone Kim, Se June Joo, Doyoung Kim, Joel Jang, Seonghyeon Ye, Jamin Shin, Minjoon Seo
    EMNLP 2023
    [paper] [code]
  • Efficiently Enhancing Zero-Shot Performance of Instruction Following Model via Retrieval of Soft Prompt
    Seonghyeon Ye, Joel Jang, Doyoung Kim, Yongrae Jo, Minjoon Seo
    EMNLP 2023 Findings
    [paper] [code]
  • Exploring the Benefits of Training Expert Language Models over Instruction Tuning
    Joel Jang, Seungone Kim, Seonghyeon Ye, Doyoung Kim, Lajanugen Logeswaran, Moontae Lee, Kyungjae Lee, Minjoon Seo
    ICML 2023
    [paper] [code]
  • Guess the Instruction! Flipped Learning Makes Language Models Stronger Zero-Shot Learners
    Seonghyeon Ye, Doyoung Kim, Joel Jang, Joongbo Shin, Minjoon Seo
    ICLR 2023
    [paper] [code]
  • SelFee: Iterative Self-Revising LLM Empowered by Self-Feedback Generation
    Seonghyeon Ye*, Yongrae Jo*, Doyoung Kim*, Sungdong Kim, Hyeonbin Hwang, Minjoon Seo
    Blog post
    [blog] [code]

2022

  • Can Large Language Models Truly Understand Prompts? A Case Study with Negated Prompts
    Joel Jang*, Seonghyeon Ye*, Minjoon Seo
    Transfer Learning for NLP Workshop @ NeurIPS 2022
    [paper] [code]
  • TemporalWiki: A Lifelong Benchmark for Training and Evaluating Ever-Evolving Language Models
    Joel Jang*, Seonghyeon Ye*, Chango Lee, Sohee Yang, Joongbo Shin, Janghoon Han, Gyeonghun Kim, Minjoon Seo
    EMNLP 2022
    [paper] [code]
  • Towards Continual Knowledge Learning of Language Models
    Joel Jang, Seonghyeon Ye, Sohee Yang, Joongbo Shin, Janghoon Han, Gyeonghun Kim, Stanley Jungkyu Choi, Minjoon Seo
    ICLR 2022
    [paper] [code]

2021

  • Efficient Contrastive Learning via Novel Data Augmentation and Curriculum Learning
    Seonghyeon Ye, Jiseon Kim, Alice Oh
    EMNLP 2021 (short)
    [paper] [code]
  • Dimensional Emotion Detection from Categorical Emotion
    Sungjoon Park, Jiseon Kim, Seonghyeon Ye, Jaeyeol Jeon, Hee Young Park, Alice Oh
    EMNLP 2021
    [paper] [code]

Education


  • KAIST AI
    M.S. & Ph.D. in Artificial Intelligence, 2022 - Present
    Advisor: Minjoon Seo, Kimin Lee
  • KAIST CS
    B.S. in Computer Science, 2017 - 2021
    Advisor: Alice Oh, Jong C. Park

Work Experience