Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to construct an RLDS dataset? #144

Open
Loong1512 opened this issue Nov 25, 2024 · 0 comments
Open

How to construct an RLDS dataset? #144

Loong1512 opened this issue Nov 25, 2024 · 0 comments

Comments

@Loong1512
Copy link

I want to fine-tune the Octo model, but I don't know how to construct my own RLDS dataset. I have already built a reinforcement learning environment using dm_env, performed simulation with Isaac Gym, and generated an RLDS dataset using envlogger.

for i in range(FLAGS.num_episodes):
  timestep = env.reset()
  while True:
    # TODO: HOW TO GENERATE ACTION
    action = np.random.uniform(-3, 3, size=(9,)).astype(np.float32)
    timestep = env.step(action)
    gym_env.render()
gym_env.close()

However, for the action part, how should I control the robotic arm to generate grasping motion trajectories? Should it be learned through a reward function, controlled via a keyboard or gamepad, or through motion capture? I'm quite confused about this. Could someone tell me the general solution?

evelynmitchell pushed a commit to evelynmitchell/octo that referenced this issue Dec 31, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant