Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question: Task ordering strategy during training stage2, 3 #105

Open
shin-wn opened this issue Oct 22, 2024 · 0 comments
Open

Question: Task ordering strategy during training stage2, 3 #105

shin-wn opened this issue Oct 22, 2024 · 0 comments

Comments

@shin-wn
Copy link

shin-wn commented Oct 22, 2024

Thank you for the great work on this project.

I noticed that tasks are trained in a fixed order rather than being shuffled:

  • In stage 2, tasks follow the sequence t2m -> m2t -> predict for one epoch
  • In stage 3, tasks appear to be processed in the order defined in the JSON file

Since motion is treated as discrete data similar to text, using the same loss function across tasks should be possible. This makes me wonder about the following:

  1. Was there a specific reason for not shuffling tasks during training?
  2. Have you found better results with this fixed-order approach compared to random task selection?
  3. Did you experiment with randomly selecting tasks in pretraining (stage 2) and instruction tuning (stage 3)?

I'm curious to learn more about the design decisions behind this approach. Looking forward to hearing your insights!

@shin-wn shin-wn changed the title Question. Why don't you shuffle tasks during training? Question: Task ordering strategy during training stages Oct 22, 2024
@shin-wn shin-wn changed the title Question: Task ordering strategy during training stages Question: Task ordering strategy during training stage2, 3 Oct 22, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant