Skip to content
View Koorye's full-sized avatar
:octocat:
Studying in school
:octocat:
Studying in school

Highlights

  • Pro

Block or report Koorye

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Koorye/README.md

↑ This is Zhongguancun Science & Technology Park in Beijing where I interned in summer 2025 ↑

English | 中文

👨‍🔬 About Me

Hello! I am an MS student at University of Electronic Science and Technology of China (UESTC).

Focus: Embodied Intelligence (VLA), Vision-Language Models (VLM).


📢 News

  • [Mar. 2026] 🎉 Our RoboCOIN dataset has been rated as a EAI-100 Top-10 Dataset in 2025 by ModelScope and CCF TCIR!
  • [Mar. 2026] 🖋️ I was invited to serve as a reviewer for BMVC 2026.
  • [Mar. 2026] 📈 Our RoboCOIN dataset has reached 4,000,000+ total downloads!
  • [Feb. 2026] 🚀 Paper InSpire (Intrinsic Spatial Reasoning for VLAs) accepted by ICRA 2026!

🔍 Media Reports & Coverage


💼 Experience

  • 🏢 Research Intern · Beijing Academy of Artificial Intelligence (BAAI) · 2025.06 - Present
  • 🎓 Master's Student · UESTC, Computer Science · 2023.09 - Present
    • 🏆National Scholarship (2024), 🏅Sichuan Province Outstanding Graduate (2026)
  • 🎓 Bachelor of SE · UESTC, Software Engineering · 2019.09 - 2023.06
    • 🏅UESTC Outstanding Graduate (2023), 🏆"Shiqiang" Special Scholarship (2022)

📕 Publications

  • 🤖 [EAI-100 TOP-10 Datasets in 2025] RoboCOIN: An Open-Sourced Bimanual Robotic Data Collection for Integrated Manipulation
    [Project] [arXiv] [PDF] [Code] Stars

    • Open-sourced large-scale bimanual robotic dataset with 15 robotic platforms and 180K+ demonstrations, collaborated with 20 institutions.
  • 🤖 [ICLR 2026] Policy Contrastive Decoding for Robotic Foundation Models
    [Project] [arXiv] [PDF] [Code] Stars

    • Universal framework for multiple VLA architectures, achieving +8%~41% improvement without training.
  • 🤖 [ICRA 2026] InSpire: Vision-Language-Action Models with Intrinsic Spatial Reasoning
    [Project] [arXiv] [PDF] [Code] Stars

    • Reducing spurious correlations in VLAs, boosting performance on seen (+6.2%) and unseen (+10%) tasks.
  • 🖼️ [IJCV 2026] A Closer Look at Conditional Prompt Tuning for Vision-Language Models
    [arXiv] [PDF] [Code] Stars

    • Identified critical issues in existing conditional prompt tuning methods, outperforming the state-of-the-art by 3.49%.
  • 🖼️ [CVPR 2025] Skip Tuning: Pre-trained Vision-Language Models are Effective and Efficient Adapters
    [arXiv][PDF][Code] Stars

    • Parameter-free adaptation method, +1.04% accuracy with 15x speedup and 6.4x memory efficiency.
  • 🖼️ [CVPR 2024] DePT: Decoupled Prompt Tuning
    [arXiv] [PDF] [Code] Stars

    • Plug-and-play method providing +0.67%~2.65% gains across various prompt tuning baselines.

🛠️ Technical Arsenal

Category Skills & Frameworks
AI
Data Science
Languages
Web & Backend
Tools

📊 GitHub Activity

snake


🖼️ Gallery

BAAI

UESTC

Pinned Loading

  1. FlagOpen/RoboCOIN FlagOpen/RoboCOIN Public

    RoboCoin + LeRobot integration

    Python 173 11

  2. DePT DePT Public

    [CVPR 2024] Offical implemention of the paper "DePT: Decoupled Prompt Tuning"

    Jupyter Notebook 110 5

  3. SkipTuning SkipTuning Public

    [CVPR 2025] Offical implementation of the paper "Skip Tuning: Pre-trained Vision-Language Models are Effective and Efficient Adapters Themselves"

    Python 32 1

  4. Inspire Inspire Public

    [ICRA 2026] Official implemetation of the paper "InSpire: Vision-Language-Action Models with Intrinsic Spatial Reasoning"

    Python 48 6

  5. PCD PCD Public

    [ICLR 2026] Official implemetation of the paper "Policy Contrastive Decoding for Robotic Foundation Models"

    Python 26 2

  6. CaPT CaPT Public

    [IJCV 2026] Offical implemention of the paper "A Closer Look at Conditional Prompt Tuning for Vision-Language Models"

    Python 8 1