Skip to content

[NeurIPS 2025 Poster] Official implementation of "Provable Ordering and Continuity in Vision-Language Pretraining for Generalizable Embodied Agents"

Notifications You must be signed in to change notification settings

Daisy-zzz/AcTOL

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

27 Commits
 
 
 
 
 
 
 
 

Repository files navigation

AcTOL: Provable Ordering and Continuity in Vision-Language Pretraining for Generalizable Embodied Agents

This repository provides the official implementation of NeurIPS 2025 Poster "Provable Ordering and Continuity in Vision-Language Pretraining for Generalizable Embodied Agents".

🔥Abstract

Pre-training vision-language representations on human action videos has emerged as a promising approach to reduce reliance on large-scale expert demonstrations for training embodied agents. However, prior methods often employ time contrastive learning based on goal-reaching heuristics, progressively aligning language instructions from the initial to the final frame. This overemphasis on future frames can result in erroneous vision-language associations, as actions may terminate early or include irrelevant moments in the end. To address this issue, we propose Action Temporal Coherence Learning (AcTOL) to learn ordered and continuous vision-language representations without rigid goal-based constraint. AcTOL treats a video as a continuous trajectory where it (1) contrasts semantic differences between frames to reflect their natural ordering, and (2) imposes a local Brownian bridge constraint to ensure smooth transitions across intermediate frames. Extensive imitation learning experiments across varying numbers of demonstrations show that the pretrained features significantly enhance downstream manipulation tasks with high robustness to different linguistic styles of instructions, offering a viable pathway toward generalized embodied agents.

Demo

🚀 Installation

Prerequisites

  • Python 3.8+
  • PyTorch 1.13.1+

Install AcTOL

pip install -e .

🎯 Training AcTOL

To train AcTOL on EPIC-KITCHENS-100, run:

python main.py --image_path /path/to/data \
               --meta_file assets/EpicKitchen-100-train.csv \
               --batch-size 128 --epochs 1001 \
               --input-size 224 --num_frames 10 \
               --model RN50 --vlo_temp 0.01 \
               --opt adam --lr 1e-5 \
               --output_dir ./checkpoints

Key Arguments:

Argument Description
--image_path Path to the dataset directory.
--meta_file Path to metadata CSV file.
--batch-size Number of videos per batch.
--num_frames Number of sampled frames per video.
--model CLIP model backbone (RN50).
--vlo_temp Temperature for VLO loss.
--lr Learning rate.
--output_dir Directory to save checkpoints.

Model Zoo

Models Pretaining Methods Params
(M)
Epochs Pretrain ckpt
RN50-CLIP AcTOL 386 1000 link

📊 Evaluation

To evaluate the language conditioned visual reward:

import AcTOL
import torch
from PIL import Image
# Load AcTOL model

device = "cuda" if torch.cuda.is_available() else "cpu"
model = AcTOL.load("AcTOL", device=device)

image = Image.open("Your Image Path Here")
text = "Your Instruction Here"

with torch.no_grad():
    image_features = model.encode_image(image)
    text_features = model.encode_text(text)
    reward = model.get_reward(image, text)

To evaluate the language-conditioned behavior cloning, please refer to the evaluation code of R3M.


Citation

Kindly cite our paper if you find it helpful:

@inproceedings{zhang2025actol,
  author       = {Zhizhen Zhang and Lei Zhu and Zhen Fang and Zi Huang and Yadan Luo},
  title        = {Provable Ordering and Continuity in Vision-Language Pretraining for Generalizable Embodied Agents},
  booktitle    = {Advances in Neural Information Processing Systems 39: Annual Conference
                  on Neural Information Processing Systems 2025, NeurIPS 2025, San Diego,
                  United States, December 2 - 7, 2025},
  year         = {2025},
}

Acknowledgements

Part of this code are adapted from DecisionNCE and RnC, thanks for their excellent work!

About

[NeurIPS 2025 Poster] Official implementation of "Provable Ordering and Continuity in Vision-Language Pretraining for Generalizable Embodied Agents"

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages