[2026/03/10]We will host the GigaBrain Challenge 2026 @ CVPR 2026 with three competition tracks: RoboTwin (simulation), GigaWorld (World Model), and RoboChallenge (real robot). We also have a call for papers on OpenReview and will select a Best Paper Award.[2026/02/13]Released GigaBrain-0.5M* technique report. GigaBrain-0.5M* is a VLA that learns from world model-based reinforcement learning.[2026/02/09]π GigaBrain-0.1 achieved 1st place on the RoboChallenge leaderboard.[2026/02/02]Released GigaBrain-0.1 model weights, which follow the same usage as GigaBrain-0 but achieve better performance.[2025/11/27]Released GigaBrain-0 model weights. This version of the model excludes depth images and intermediate 2D manipulation trajectories for more user-friendly use. However, the code supports these features β if your dataset contains them and you wish to use them, simply enable the corresponding options in the configuration.[2025/11/27]Released the model architecture, as well as the pre-training and post-training implementations.
Training Vision-Language-Action (VLA) models for generalist robots typically requires large-scale real-world robot data, which is expensive and time-consuming to collect. The inefficiency of data collection severely limits the scalability, and generalization capacity of current VLA systems. Therefore, we introduce GigaBrain-0, a novel VLA foundation model empowered by world model-generated data. By leveraging world models to generate diverse data at scale, GigaBrain-0 significantly reduces reliance on real robot data while improving cross-task generalization. Our approach further improves policy robustness through RGBD input modeling and embodied Chain-of-Thought (CoT) supervision, enabling the model to reason about spatial geometry, object states, and long-horizon dependencies during task execution. This leads to substantial gains in real-world performance on dexterous, long-horizon, and mobile manipulation tasks. Extensive experiments demonstrate that GigaBrain-0 achieves superior generalization across variations in appearances (e.g., textures, colors), object placements, and camera viewpoints.
GigaBrain-0 was trained using approximately 1k hours of real-world robot data together with a large amount of World Modelβgenerated data. GigaBrain-0.1 scales the training data to 10k hours, with the detailed data composition shown in the figure below.
Leveraging the efficient world-model data engine and innovations in model architecture, GigaBrain-0.1 has demonstrated rapid performance improvements. GigaBrain-0.1 outperforms GigaBrain-0 across all real-robot tasks and achieves performance on complex long-horizon tasks comparable to
- Paper Towel Preparation*: Compared to the release with GigaBrain-0, the "Paper Towel Preparation" task has been re-evaluated under a new setting.
Using GigaBrain-0.1 to train on RoboChallenge tasks, we achieved 1st place on the leaderboard.
GigaBrain-0 depends on the following three frameworks:
- GigaTrain: An Efficient and Scalable Training Framework for AI Models.
- GigaDatasets: A Unified and Lightweight Framework for Data Curation, Evaluation and Visualization.
- GigaModels: A Comprehensive Repository for Multi-modal, Generative, and Perceptual Models.
We recommend a fresh conda environment.
conda create -n giga_brain_0 python=3.11.10 -y
conda activate giga_brain_0
pip3 install giga-train
pip3 install giga-datasets
pip3 install lerobot==0.3.2
pip3 install matplotlib
pip3 install numpydantic
git clone https://github.com/open-gigaai/giga-models.git
cd giga-models
pip3 install -e .
git clone https://github.com/open-gigaai/giga-brain-0.git
cd giga-brain-0
To begin, convert your data to the LeRobot format. For reference, see scripts/convert_from_hdf5.py, which demonstrates how to convert AgileX data (HDF5 files) to LeRobotDataset.
python scripts/convert_from_hdf5.py \
--data-path /path/to/raw_hdf5_data_path \
--out-dir /path/to/lerobot_dataset \
--task "Task prompt here"If your dataset is already in LeRobot format, compute normalization stats for observation.state and action using our script:
python scripts/compute_norm_stats.py \
--data-paths /path/to/lerobot_dataset1 /path/to/lerobot_dataset2 \
--output-path /path/to/norm_stats.json \
--embodiment-id {embodiment-id} \
--delta-mask {delta-mask} \
--sample-rate 1.0 \
--action-chunk 50 \
--action-dim 32
For AgileX Cobot Magic:
- embodiment_id = 0
- delta_mask = [True, True, True, True, True, True, False, True, True, True, True, True, True, False]
For Agibot G1:
- embodiment_id = 1
- delta_mask = [True, True, True, True, True, True, True, False, True, True, True, True, True, True, True, False, True, True, True, True]
To support custom robot-type data, you can add a newly initialized action-specific linear and train the newly added linear only and freeze other weights.
Then point your training config to the produced norm_stats.json (see examples in configs).
| Model | HF Link | Description |
|---|---|---|
| GigaBrain-0.1-3.5B-Base | π€ Huggingface | More generalizable, more robust, and more powerful. |
| GigaBrain-0-3.5B-Base | π€ Huggingface | The current release of the model excludes depth images and intermediate 2D manipulation trajectories for more user-friendly use. |
We provide ready-to-use configs for GigaBrain-0. Adjust gpu_ids, batch_size_per_gpu, data_paths, and norm_stats_path as needed.
Logs, configs and checkpoints will be stored at the path project_dir
Pre-training:
python scripts/train.py --config configs.giga_brain_0_from_scratch.configFine-tuning:
python scripts/train.py --config configs.giga_brain_0_agilex_finetune.config # or
python scripts/train.py --config configs.giga_brain_0_agibot_finetune.configConfiguration details can be checked in configure_introduction.md
Run inference on a LeRobot dataset and optionally visualize predictions.
For the same model weights, we provide three different scripts to support different output information.
-
Inference continuous action:
python scripts/inference.py \ --model-path /path/to/giga_brain_0_checkpoints \ --data-path /path/to/lerobot_dataset \ --norm-stats-path /path/to/norm_stats.json \ --output-path /tmp/vis_path \ --delta-mask <DELTA_MASK> \ --embodiment-id <EMBODIMENT_ID> \ --action-chunk 50 \ --original-action-dim <ACTION_DIM> \ --tokenizer-model-path google/paligemma2-3b-pt-224 \ --fast-tokenizer-path physical-intelligence/fast \ --device cuda
-
Inference subgoal prediction:
python scripts/inference_task_planning.py \ --model-path /path/to/giga_brain_0_checkpoints \ --data-path /path/to/lerobot_dataset \ --norm-stats-path /path/to/norm_stats.json \ --delta-mask <DELTA_MASK> \ --embodiment-id <EMBODIMENT_ID> \ --original-action-dim <ACTION_DIM> \ --tokenizer-model-path google/paligemma2-3b-pt-224 \ --fast-tokenizer-path physical-intelligence/fast \ --device cuda
-
Inference discrete action in autoregressive mode (usually for debugging):
python scripts/inference_discrete_action.py \ --model-path /path/to/giga_brain_0_checkpoints \ --data-path /path/to/lerobot_dataset \ --norm-stats-path /path/to/norm_stats.json \ --output-path /tmp/vis_path \ --delta-mask <DELTA_MASK> \ --embodiment-id <EMBODIMENT_ID> \ --original-action-dim <ACTION_DIM> \ --tokenizer-model-path google/paligemma2-3b-pt-224 \ --fast-tokenizer-path physical-intelligence/fast \ --device cuda
-
Run the server:
python scripts/inference_server.py \ --model-path /path/to/giga_brain_0_checkpoints \ --tokenizer-model-path google/paligemma2-3b-pt-224 \ --fast-tokenizer-path physical-intelligence/fast \ --delta-mask <DELTA_MASK> \ --embodiment-id <EMBODIMENT_ID> \ --norm-stats-path /path/to/norm_stats.json \ --original-action-dim <ACTION_DIM> \ --autoregressive-mode-only False
-
Run the client:
python scripts/inference_client.py
This is a minimal client example. It generates random observations to demonstrate the end-to-end request/response flow with the server. You can copy the relevant client code onto your robot and replace the random inputs with real onboard sensor data (e.g., cameras, proprioception) and your robot's control interface. Ensure input shapes and field names remain consistent with the server's expectations.
We also provide an inference client script for AgileX robots: scripts/inference_agilex_client.py.
Make sure the host and port are the same in both server and client.
This project is licensed under the Apache License 2.0 - see the LICENSE file for details.
@article{gigaai2025gigabrain0,
title={GigaBrain-0: A World Model-Powered Vision-Language-Action Model},
author={GigaAI},
year={2025},
eprint={2510.19430},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2510.19430},
}



