Junyi Li (李俊奕)
Logo PhD student @ HKU CDS
Logo Undergrad student @ HUST EIC

Hi! I am a first year PhD student at The University of Hong Kong (HKU), School of Computing and Data Science, supervised by Prof. Hengshuang Zhao. Previously, I received my B.Eng. from the School of Electronic Information and Communications at Huazhong University of Science and Technology (HUST) with a GPA of 3.98/4.0 (Rank: 1/155). During my undergraduate studies, I was fortunate to be supervised by Prof. Xiang Bai. My current research interests primarily lie in Multimodal Large Language Models and Reinforcement Learning.


Education
  • The University of Hong Kong
    The University of Hong Kong
    School of Computing and Data Science
    Ph.D. Student
    Sept. 2025 - Present
  • Huazhong University of Science and Technology
    Huazhong University of Science and Technology
    B.S. in Electronic Information and Communications
    Sept. 2021 - Jul. 2025
Experience
  • Tiktok
    Tiktok
    Research Intern
    Nov. 2024 - Present
Honors & Awards
  • National Scholarship
    2024
  • National Scholarship
    2023
  • National Scholarship
    2022
  • Outstanding Undergraduate
    2023
  • Meritorious Winner in Mathematical Contest In Modeling (MCM)
    2023
News
2025
🔥 One paper accepted by NeurIPS2025. Congratulations!
Sep 19
🔥 We release Mini-o3 and a full training recipe to reproduce OpenAI o3-style thinking-with-images capability. Welcome to our [Project Page]
Sep 11
🔥 We release VisionThink, a Smart and Efficient Vision Language Model built upon Reinforcement Learning. Welcome to the GitHub Page.
Jul 18
🔥 We release TokBench, a Simple and Effective benchmark for evaluating your visual tokenizers. Feel free to check out our [Project Page].
May 19
2024
PartGLEE is accepted by ECCV2024. This is the first Foundation Model for recognizing and parsing any object! [Project Page].
Oct 02
Awarded the National Scholarship.
Oct 01
2023
Awarded the Outstanding Undergraduate.
Oct 31
Awarded the National Scholarship.
Oct 01
2022
Awarded the National Scholarship.
Oct 01
Selected Publications (view all )
Mini-o3: Scaling Up Reasoning Patterns and Interaction Turns for Visual Search
Mini-o3: Scaling Up Reasoning Patterns and Interaction Turns for Visual Search

Xin Lai*, Junyi Li*, Wei Li, Tao Liu, Tianjian Li, Hengshuang Zhao# (* equal contribution, # corresponding author)

Under review. 2025

A full training recipe to reproduce OpenAI o3-style thinking-with-images capability.

Mini-o3: Scaling Up Reasoning Patterns and Interaction Turns for Visual Search

Xin Lai*, Junyi Li*, Wei Li, Tao Liu, Tianjian Li, Hengshuang Zhao# (* equal contribution, # corresponding author)

Under review. 2025

A full training recipe to reproduce OpenAI o3-style thinking-with-images capability.

VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning
VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning

Senqiao Yang*, Junyi Li*, Xin Lai*, Bei Yu, Hengshuang Zhao, Jiaya Jia (* equal contribution)

Advances in Neural Information Processing Systems (NeurIPS) 2025 Poster

A new paradigm of efficient VLMs with token compression.

VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning

Senqiao Yang*, Junyi Li*, Xin Lai*, Bei Yu, Hengshuang Zhao, Jiaya Jia (* equal contribution)

Advances in Neural Information Processing Systems (NeurIPS) 2025 Poster

A new paradigm of efficient VLMs with token compression.

TokBench: Evaluating Your Visual Tokenizer before Visual Generation
TokBench: Evaluating Your Visual Tokenizer before Visual Generation

Junfeng Wu*, Dongliang Luo*, Weizhi Zhao, Zhihao Xie, Yuanhao Wang, Junyi Li, Xudong Xie, Yuliang Liu, Xiang Bai# (* equal contribution, # corresponding author)

Under review. 2025

A simple Benchmark for evaluating your visual tokenizer.

TokBench: Evaluating Your Visual Tokenizer before Visual Generation

Junfeng Wu*, Dongliang Luo*, Weizhi Zhao, Zhihao Xie, Yuanhao Wang, Junyi Li, Xudong Xie, Yuliang Liu, Xiang Bai# (* equal contribution, # corresponding author)

Under review. 2025

A simple Benchmark for evaluating your visual tokenizer.

PartGLEE: A Foundation Model for Recognizing and Parsing Any Objects
PartGLEE: A Foundation Model for Recognizing and Parsing Any Objects

Junyi Li*, Junfeng Wu*, Weizhi Zhao, Song Bai, Xiang Bai# (* equal contribution, # corresponding author)

Eurepean Conference on Computer Vision (ECCV) 2024 Poster

The first part-level foundation model for locating and identifying both objects and parts in images through a hierarchical framework.

PartGLEE: A Foundation Model for Recognizing and Parsing Any Objects

Junyi Li*, Junfeng Wu*, Weizhi Zhao, Song Bai, Xiang Bai# (* equal contribution, # corresponding author)

Eurepean Conference on Computer Vision (ECCV) 2024 Poster

The first part-level foundation model for locating and identifying both objects and parts in images through a hierarchical framework.

All publications