Email: [email protected]
Yujie's research focuses on vision language models (VLMs) and large language models (LLMs), with an emphasis on robust and efficient post-training (text, image, video), interpretable and faithful evaluation, task planning with LLM agents, and alignment with human preferences.
News! 12/18/2025 Excited about the launch of our Native Video-to-Video Generation Model!
News! 09/20/2025 MLR-Bench: Evaluating AI Agents on Open-Ended Machine Learning Research is accepted at NeurIPS 2025 D&B Track!
News! 04/05/2025 Checkout out our first The llama 4 herd: The beginning of a new era of natively multimodal ai innovation!
News! 02/26/2025 VITED: Video Temporal Evidence Distillation is accepted at CVPR 2025!
News! 01/24/2025 I successfully complete my PhD defense and will join Meta GenAI Llama Research Team!
News! 01/22/2025 MMWorld is accepted at ICLR 2025!
News! 09/26/2024 WildVision is accepted at NeurIPS 2024 D&B Track! T2IScoreScore is accepted at NeurIPS 2024 Main Track!
News! 09/20/2024 Multimodal Procedural Planning is accepted at EMNLP 2024!
News! 08/27/2024 We release WildVision datasets: WV-Chat, WV Battle, and WV Bench.
News! 06/24/2024 I started working on video large language models at Meta (FAIR Embodied AI) in NYC this Summer!
News! 03/01/2024 Check out our WildVision Arena demo on HuggingFace for live benchmark VLMs!
News! 09/21/2023 LLMScore is accepted at NeurIPS 2023!
News! 01/20/2023 Neuro-Symbolic Causal Procedural Planning with Commonsense Prompting is accepted at ICLR 2023!
News! 04/07/2022 Imagination-Augmented Natural Language Understanding is accepted at NAACL 2022!