Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
42 commits
Select commit Hold shift + click to select a range
076a27a
Update pick_place configs and tasks
HanchuZhou Dec 3, 2025
81db6d7
Fix bug that success rate is more than 100% in evaluate_lift.py; Fixe…
HanchuZhou Dec 6, 2025
17a74b3
add contributor
HanchuZhou Dec 6, 2025
c0699a8
[pre-commit.ci] auto fixes from pre-commit.com hooks
pre-commit-ci[bot] Dec 6, 2025
e0ceed6
Fixing success rate bug; Enable more flexible evaluation that user ca…
HanchuZhou Dec 11, 2025
58db815
Merge branch 'develop' of https://github.com/HanchuZhou/RoboVerse int…
HanchuZhou Dec 11, 2025
066596d
Make training script individual; Add gripper downward reward
HanchuZhou Dec 11, 2025
c13ef64
Update pick_place configs and tasks
HanchuZhou Dec 3, 2025
3196e18
Fix bug that success rate is more than 100% in evaluate_lift.py; Fixe…
HanchuZhou Dec 6, 2025
d67aab3
Fixing success rate bug; Enable more flexible evaluation that user ca…
HanchuZhou Dec 11, 2025
3ad659f
Make training script individual; Add gripper downward reward
HanchuZhou Dec 11, 2025
0355e60
Resolved merge conflicts by keeping local versions
Anchor1021 Dec 12, 2025
7e6fed2
"Refactor approach_grasp tasks, yaml, and track tasks yaml"
Anchor1021 Dec 12, 2025
849ce1d
"Refactor pick_place grasp logic and add object configs"
Anchor1021 Dec 12, 2025
6b49480
Refactor pick_place grasp logic and add object configs
Anchor1021 Dec 12, 2025
cb5635a
Adjust Spoon, two spoon class in one file, two spoon yaml.
Anchor1021 Dec 14, 2025
671aff4
Apply pre-commit fixes and add new pick_place_spoon2 config
Anchor1021 Dec 14, 2025
6f3913e
Local devcontainer config (ignored from push)
Anchor1021 Dec 14, 2025
e89acb4
Fix duplicated sim key in pick_place_spoon2.yaml
Anchor1021 Dec 14, 2025
5ced66f
Trigger CI
Anchor1021 Dec 14, 2025
dd932c5
[pre-commit.ci] auto fixes from pre-commit.com hooks
pre-commit-ci[bot] Dec 14, 2025
e96b4a2
update the position of banana and screw driver
HanchuZhou Dec 14, 2025
c2d2160
Merge branch 'develop' of https://github.com/HanchuZhou/RoboVerse int…
HanchuZhou Dec 14, 2025
92912be
add approach_grasp.py
HanchuZhou Dec 15, 2025
3066002
Merge branch 'RoboVerseOrg:develop' into develop
HanchuZhou Dec 15, 2025
f2853f4
[pre-commit.ci] auto fixes from pre-commit.com hooks
pre-commit-ci[bot] Dec 15, 2025
4fbe0b0
Roll back readme
HanchuZhou Dec 15, 2025
bbd2ba5
roll back
HanchuZhou Dec 15, 2025
342a44b
[pre-commit.ci] auto fixes from pre-commit.com hooks
pre-commit-ci[bot] Dec 15, 2025
7612ce4
roll back collect demo
HanchuZhou Dec 15, 2025
d8c9b24
roll back collect demo
HanchuZhou Dec 15, 2025
c0a8831
Revise devcontainer.json for development setup
HanchuZhou Dec 15, 2025
ae75807
Remove devcontainer.json from .gitignore
HanchuZhou Dec 15, 2025
2bd96c6
Merge upstream develop into fork develop
Anchor1021 Dec 16, 2025
3a5f9ec
hardcode approach_grasp; roll back rl/fast_td3/train and base.pyto pr…
Anchor1021 Dec 16, 2025
1d74b67
Apply pre-commit formatting and lint fixes
Anchor1021 Dec 16, 2025
a9b421e
In evaluate_lift.py, record the state that N frames after grasping; A…
HanchuZhou Dec 18, 2025
037e7dc
Change evaluate_lift.py, make sure the gripper holds the object; Add …
HanchuZhou Dec 18, 2025
8f1d867
Merge branch 'develop' of https://github.com/HanchuZhou/RoboVerse int…
HanchuZhou Dec 18, 2025
2beb889
[pre-commit.ci] auto fixes from pre-commit.com hooks
pre-commit-ci[bot] Dec 18, 2025
6a0cf99
Add several new tasks
HanchuZhou Dec 19, 2025
3fceee9
add single banana tasks
HanchuZhou Jan 14, 2026
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 5 additions & 1 deletion metasim/sim/isaacgym/isaacgym.py
Original file line number Diff line number Diff line change
Expand Up @@ -774,7 +774,11 @@ def _get_states(self, env_ids: list[int] | None = None) -> list[DictEnvState]:

# Apply GS background rendering if enabled
# TODO: Render with batch parallelization for efficiency
if self.scenario.gs_scene.with_gs_background and self.gs_background is not None:
if (
self.scenario.gs_scene is not None
and self.scenario.gs_scene.with_gs_background
and self.gs_background is not None
):
assert ROBO_SPLATTER_AVAILABLE, "RoboSplatter is not available. GS background rendering will be disabled."
camera_states = self._apply_gs_background_rendering(camera_states, env_ids)

Expand Down
89 changes: 89 additions & 0 deletions roboverse_learn/rl/fast_td3/configs/pick_place_banana2.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,89 @@
# Base Configuration for FastTD3 Training - Banana
# Default configuration with IsaacGym simulator and Franka robot

# -------------------------------------------------------------------------------
# Environment
# -------------------------------------------------------------------------------
sim: "isaacgym"
robots: ["franka"]
task: "pick_place.approach_grasp_simple_banana2"
decimation: 4
train_or_eval: "train"
headless: True

# -------------------------------------------------------------------------------
# Seeds & Device
# -------------------------------------------------------------------------------
seed: 1
cuda: true
torch_deterministic: true
device_rank: 0

# -------------------------------------------------------------------------------
# Rollout & Timesteps
# -------------------------------------------------------------------------------
num_envs: 400
num_eval_envs: 400
total_timesteps: 2000000
learning_starts: 10
num_steps: 1

# -------------------------------------------------------------------------------
# Replay, Batching, Discounting
# -------------------------------------------------------------------------------
buffer_size: 20480
batch_size: 32768
gamma: 0.99
tau: 0.1

# -------------------------------------------------------------------------------
# Update Schedule
# -------------------------------------------------------------------------------
policy_frequency: 2
num_updates: 5
# -------------------------------------------------------------------------------
# Optimizer & Network
# -------------------------------------------------------------------------------
critic_learning_rate: 0.0003
actor_learning_rate: 0.0003
weight_decay: 0.1
critic_hidden_dim: 512
actor_hidden_dim: 256
init_scale: 0.01
num_atoms: 101

# -------------------------------------------------------------------------------
# Value Distribution & Exploration
# -------------------------------------------------------------------------------
v_min: 0
v_max: 600.0
policy_noise: 0.001
std_min: 0.001
std_max: 0.4
noise_clip: 0.5

# -------------------------------------------------------------------------------
# Algorithm Flags
# -------------------------------------------------------------------------------
use_cdq: true
compile: true
obs_normalization: true
max_grad_norm: 0.0
amp: true
amp_dtype: "fp16"
disable_bootstrap: false
measure_burnin: 3

# -------------------------------------------------------------------------------
# Logging & Checkpointing
# -------------------------------------------------------------------------------
wandb_project: "get_started_fttd3"
exp_name: "get_started_fttd3_banana"
use_wandb: false
checkpoint_path: null
run_name: "pick_place.approach_grasp_simple_banana2" # Unique run name
model_dir: "models/banana2" # Separate directory for banana checkpoints
eval_interval: 5000
save_interval: 5000
video_width: 1024
video_height: 1024
89 changes: 89 additions & 0 deletions roboverse_learn/rl/fast_td3/configs/pick_place_banana3.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,89 @@
# Base Configuration for FastTD3 Training - Banana
# Default configuration with IsaacGym simulator and Franka robot

# -------------------------------------------------------------------------------
# Environment
# -------------------------------------------------------------------------------
sim: "isaacgym"
robots: ["franka"]
task: "pick_place.approach_grasp_simple_banana3"
decimation: 4
train_or_eval: "train"
headless: True

# -------------------------------------------------------------------------------
# Seeds & Device
# -------------------------------------------------------------------------------
seed: 1
cuda: true
torch_deterministic: true
device_rank: 0

# -------------------------------------------------------------------------------
# Rollout & Timesteps
# -------------------------------------------------------------------------------
num_envs: 400
num_eval_envs: 400
total_timesteps: 2000000
learning_starts: 10
num_steps: 1

# -------------------------------------------------------------------------------
# Replay, Batching, Discounting
# -------------------------------------------------------------------------------
buffer_size: 20480
batch_size: 32768
gamma: 0.99
tau: 0.1

# -------------------------------------------------------------------------------
# Update Schedule
# -------------------------------------------------------------------------------
policy_frequency: 2
num_updates: 5
# -------------------------------------------------------------------------------
# Optimizer & Network
# -------------------------------------------------------------------------------
critic_learning_rate: 0.0003
actor_learning_rate: 0.0003
weight_decay: 0.1
critic_hidden_dim: 512
actor_hidden_dim: 256
init_scale: 0.01
num_atoms: 101

# -------------------------------------------------------------------------------
# Value Distribution & Exploration
# -------------------------------------------------------------------------------
v_min: 0
v_max: 600.0
policy_noise: 0.001
std_min: 0.001
std_max: 0.4
noise_clip: 0.5

# -------------------------------------------------------------------------------
# Algorithm Flags
# -------------------------------------------------------------------------------
use_cdq: true
compile: true
obs_normalization: true
max_grad_norm: 0.0
amp: true
amp_dtype: "fp16"
disable_bootstrap: false
measure_burnin: 3

# -------------------------------------------------------------------------------
# Logging & Checkpointing
# -------------------------------------------------------------------------------
wandb_project: "get_started_fttd3"
exp_name: "get_started_fttd3_banana3"
use_wandb: false
checkpoint_path: null
run_name: "pick_place.approach_grasp_simple_banana3" # Unique run name
model_dir: "models/banana3" # Separate directory for banana checkpoints
eval_interval: 5000
save_interval: 5000
video_width: 1024
video_height: 1024
89 changes: 89 additions & 0 deletions roboverse_learn/rl/fast_td3/configs/pick_place_banana4.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,89 @@
# Base Configuration for FastTD3 Training - Banana
# Default configuration with IsaacGym simulator and Franka robot

# -------------------------------------------------------------------------------
# Environment
# -------------------------------------------------------------------------------
sim: "isaacgym"
robots: ["franka"]
task: "pick_place.approach_grasp_simple_banana4"
decimation: 4
train_or_eval: "train"
headless: True

# -------------------------------------------------------------------------------
# Seeds & Device
# -------------------------------------------------------------------------------
seed: 1
cuda: true
torch_deterministic: true
device_rank: 0

# -------------------------------------------------------------------------------
# Rollout & Timesteps
# -------------------------------------------------------------------------------
num_envs: 400
num_eval_envs: 400
total_timesteps: 2000000
learning_starts: 10
num_steps: 1

# -------------------------------------------------------------------------------
# Replay, Batching, Discounting
# -------------------------------------------------------------------------------
buffer_size: 20480
batch_size: 32768
gamma: 0.99
tau: 0.1

# -------------------------------------------------------------------------------
# Update Schedule
# -------------------------------------------------------------------------------
policy_frequency: 2
num_updates: 5
# -------------------------------------------------------------------------------
# Optimizer & Network
# -------------------------------------------------------------------------------
critic_learning_rate: 0.0003
actor_learning_rate: 0.0003
weight_decay: 0.1
critic_hidden_dim: 512
actor_hidden_dim: 256
init_scale: 0.01
num_atoms: 101

# -------------------------------------------------------------------------------
# Value Distribution & Exploration
# -------------------------------------------------------------------------------
v_min: 0
v_max: 600.0
policy_noise: 0.001
std_min: 0.001
std_max: 0.4
noise_clip: 0.5

# -------------------------------------------------------------------------------
# Algorithm Flags
# -------------------------------------------------------------------------------
use_cdq: true
compile: true
obs_normalization: true
max_grad_norm: 0.0
amp: true
amp_dtype: "fp16"
disable_bootstrap: false
measure_burnin: 3

# -------------------------------------------------------------------------------
# Logging & Checkpointing
# -------------------------------------------------------------------------------
wandb_project: "get_started_fttd3"
exp_name: "get_started_fttd3_banana4"
use_wandb: false
checkpoint_path: null
run_name: "pick_place.approach_grasp_simple_banana4" # Unique run name
model_dir: "models/banana4" # Separate directory for banana checkpoints
eval_interval: 5000
save_interval: 5000
video_width: 1024
video_height: 1024
89 changes: 89 additions & 0 deletions roboverse_learn/rl/fast_td3/configs/pick_place_banana5.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,89 @@
# Base Configuration for FastTD3 Training - Banana
# Default configuration with IsaacGym simulator and Franka robot

# -------------------------------------------------------------------------------
# Environment
# -------------------------------------------------------------------------------
sim: "isaacgym"
robots: ["franka"]
task: "pick_place.approach_grasp_simple_banana5"
decimation: 4
train_or_eval: "train"
headless: True

# -------------------------------------------------------------------------------
# Seeds & Device
# -------------------------------------------------------------------------------
seed: 1
cuda: true
torch_deterministic: true
device_rank: 0

# -------------------------------------------------------------------------------
# Rollout & Timesteps
# -------------------------------------------------------------------------------
num_envs: 400
num_eval_envs: 400
total_timesteps: 2000000
learning_starts: 10
num_steps: 1

# -------------------------------------------------------------------------------
# Replay, Batching, Discounting
# -------------------------------------------------------------------------------
buffer_size: 20480
batch_size: 32768
gamma: 0.99
tau: 0.1

# -------------------------------------------------------------------------------
# Update Schedule
# -------------------------------------------------------------------------------
policy_frequency: 2
num_updates: 5
# -------------------------------------------------------------------------------
# Optimizer & Network
# -------------------------------------------------------------------------------
critic_learning_rate: 0.0003
actor_learning_rate: 0.0003
weight_decay: 0.1
critic_hidden_dim: 512
actor_hidden_dim: 256
init_scale: 0.01
num_atoms: 101

# -------------------------------------------------------------------------------
# Value Distribution & Exploration
# -------------------------------------------------------------------------------
v_min: 0
v_max: 600.0
policy_noise: 0.001
std_min: 0.001
std_max: 0.4
noise_clip: 0.5

# -------------------------------------------------------------------------------
# Algorithm Flags
# -------------------------------------------------------------------------------
use_cdq: true
compile: true
obs_normalization: true
max_grad_norm: 0.0
amp: true
amp_dtype: "fp16"
disable_bootstrap: false
measure_burnin: 3

# -------------------------------------------------------------------------------
# Logging & Checkpointing
# -------------------------------------------------------------------------------
wandb_project: "get_started_fttd3"
exp_name: "get_started_fttd3_banana5"
use_wandb: false
checkpoint_path: null
run_name: "pick_place.approach_grasp_simple_banana5" # Unique run name
model_dir: "models/banana5" # Separate directory for banana checkpoints
eval_interval: 5000
save_interval: 5000
video_width: 1024
video_height: 1024
Loading