Skip to content

williamtsai726/YAM

Repository files navigation

YAM

Teleoperation, Data Collection, and model evaluation on Bimanual YAM.

Motor Configuration

For long time bimanual teleop, data collection, or evaluation, the default timeout is too short and often causes abrupt collapse. To prevent that, we turn off the motor timeout for both arms.

python ../i2rt/i2rt/motor_config_tool/set_timeout.py --channel can_left &&
python ../i2rt/i2rt/motor_config_tool/set_timeout.py --channel can_right

Gello Configuration

Everything for gello is located in gello_software.

cd gello_software

Upon reconnecting the YAM to the PC, make sure to reset the CAN.

bash scripts/reset_all_can.sh

Configuration of the left arm is in configs/yam_left.yaml and configuration of the right arm is in configs/yam_right.yaml.

Reset gripper

python ../i2rt/i2rt/motor_config_tool/set_zero.py --channel=can_left --motor_id=7 &&
python ../i2rt/i2rt/motor_config_tool/set_zero.py --channel=can_right --motor_id=7

Teleoperation

To perform teleoperation, simply run

python experiments/launch_yaml.py --left_config_path=configs/yam_left.yaml --right_config_path=configs/yam_right.yaml

Data Collection

To perform data collection, update configs/yam_left.yaml, mainly sections storage and lerobot:

# Data storage configuration
storage:
  episodes: 30
  base_dir: "/home/sean/Desktop/YAM/gello_software/data_recurrent"
  task_directory: "test"
  language_instruction: "test"
  teleop_device: "oculus" # ["oculus", "keyboard", "gello", "none"]
  save_format: "json" # ["json", "npy"]
  old_format: false

# LeRobot conversion + upload pipeline
lerobot:
  auto_convert: true
  auto_upload: true
  hf_repo_id: "your_huggingface_user/your_dataset_name"
  delete_local_after_upload: true
  fps: 30
  robot_type: "molmoact_dual_arm"
  skip_initial_frames: 0
  action_mode: "next_joint_fields" # ["next_joint_fields", "next_state", "copy_state"]
  sanitize_online_viz_meta: true

storage.episodes is the maximum episode index to collect. The collection loop ends when this limit is reached.

storage.base_dir is the location to store collected raw json episodes.

storage.task_directory is the subdirectory name for that task.

storage.language_instruction is the instruction written into collected data.

lerobot.auto_convert controls whether post-collection conversion is run.

lerobot.auto_upload controls whether converted data is uploaded to Hugging Face after conversion. If upload is enabled, the script also tries to create dataset tag v3.0 and skips tag creation if it already exists.

lerobot.hf_repo_id is the destination Hugging Face dataset repo in the form username/dataset_name.

lerobot.delete_local_after_upload controls whether local raw json and local LeRobot output are deleted after successful upload/tagging.

lerobot.fps is the frame rate metadata written into the generated LeRobot dataset (set this to match your collection/control frequency).

lerobot.robot_type sets the robot metadata field saved in the LeRobot dataset.

lerobot.skip_initial_frames skips the first N frames of each episode during conversion (useful to remove startup transients).

lerobot.action_mode controls how action is derived:

  • next_joint_fields (recommended): use next_left_joint/next_right_joint from json.
  • next_state: use shifted joint state at t+1 as action.
  • copy_state: use current joint state at t as action.

lerobot.sanitize_online_viz_meta removes quantile-only metadata columns after conversion to improve compatibility with some online visualizers.

To perform data collection after configuration simply run:

python experiments/launch_yaml_collect_data.py --left_config_path=configs/yam_left.yaml --right_config_path=configs/yam_right.yaml

The program will launch a color pad to take keyboard input.

Press s to start collecting 1 episode of data.

Press a to end and save collected episode.

Press b to end and delete collected episode.

After all episodes are collected, the script runs post-collection pipeline based on config:

  • if auto_convert: true and auto_upload: false, it converts only.
  • if auto_convert: true and auto_upload: true, it converts, uploads, and tags.

If the LeRobot output directory already exists, it will ask:

Do you want to remove it and continue? (y/n)

Type y to remove and continue, or n to cancel the post-collection pipeline.

Important: pressing ctrl+c exits early and only performs robot/socket cleanup.
It does not run the convert/upload/tag pipeline.

Note: make sure you are on the color pad so it can take in the keyboard input (don't put it in the background).
To kill the program with ctrl+c, you will need to be on your IDE or Terminal.

Data Converstion

Manual conversion is still available if needed. Data is saved in json format and can be converted with molmoact_to_lerobot_v30.py.

By default, the script loads parameters from gello_software/configs/yam_left.yaml:

  • data_dir = storage.base_dir + storage.task_directory
  • output_dir = storage.base_dir + storage.task_directory + "_lerobot_v30"
  • upload behavior from lerobot.auto_upload

Field definitions used by conversion/upload in gello_software/configs/yam_left.yaml:

storage fields:

  • base_dir: root directory where collected json episodes are stored.
  • task_directory: task subfolder under base_dir (also used to derive output directory name).
  • language_instruction: default task text written into converted LeRobot episodes.

lerobot fields:

  • auto_convert: enables post-collection conversion in the launcher pipeline.
  • auto_upload: if true, converted data is uploaded to Hugging Face.
  • hf_repo_id: target Hugging Face dataset repo (format: username/dataset_name).
  • delete_local_after_upload: if true, remove local json + local LeRobot folder after successful upload/tag.
  • fps: frame rate metadata saved into the LeRobot dataset.
  • robot_type: robot metadata string saved into the LeRobot dataset.
  • skip_initial_frames: number of initial frames to skip per episode during conversion.
  • action_mode: how action is derived (next_joint_fields, next_state, copy_state).
  • sanitize_online_viz_meta: removes quantile-only metadata columns for better online visualizer compatibility.

So if your config is already set, you can run:

python molmoact_to_lerobot_v30.py

You can still override parameters manually:

python molmoact_to_lerobot_v30.py \
        --data_dir /path/to/molmoact \
        --output_dir /path/to/molmoact_lerobot_v30 \
        --repo_id your_huggingface_user/molmoact_v30 \
        --fps 10 \
        --upload_to_hf 1

When upload is enabled, the script uploads to Hugging Face and then adds tag v3.0 automatically.
If the tag already exists, it skips creating the duplicate tag.

Model Evaluation

Current evaluation supports two policy types in experiments/launch_yaml_eval.py:

  • dp (DiffusionPolicy)
  • pi05 (PI05Policy)

Set these fields in configs/yam_left.yaml under policy:

# Policy configuration
policy:
  type: "dp"   # ["dp", "pi05"]
  repo_id: "your_hf_user/your_dataset_or_local_dataset_path"
  checkpoint_path: "your_model_repo_or_local_checkpoint_path"

policy.type selects which evaluator path is used:

  • dp: runs run_control_loop_eval with diffusion policy.
  • pi05: runs run_control_loop_eval_pi with PI05 chunked actions.

policy.repo_id is used to load dataset metadata/statistics (HF dataset id or local dataset path).

policy.checkpoint_path is the policy checkpoint source (HF model id or local checkpoint path).

For pi05, task instruction is taken from storage.language_instruction.

After configuring policy, run:

python experiments/launch_yaml_eval.py --left_config_path=configs/yam_left.yaml --right_config_path=configs/yam_right.yaml

Notes: preprocess_observation in experiments/launch_yaml_eval.py converts robot observations to model inputs for the dp path. Make sure image size/camera mapping matches your model's expected input format.

# Define the target image size
TARGET_HEIGHT = 256
TARGET_WIDTH = 342

# Map cameras "observation : model"
camera_mapping = {"left_camera_rgb": 'left', "right_camera_rgb": 'right', "front_camera_rgb": 'front'}

Model Evaluation for MolmoAct2

Current evaluation supports remote inference only. The MolmoAct2 model should be hosted in a remote server.

Update the server url in experiments/molmoact.py

Update the task instruction storage.language_instruction in configs/yam_left.yaml.

After configuring policy, run:

python experiments/launch_yaml_eval_molmoact.py --left_config_path=configs/yam_left.yaml --right_config_path=configs/yam_right.yaml

About

YAM setup for both oculus quest and gello teleoperation and data collection

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors