Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

lazy dataset loading to save RAM memory #12

Open
wants to merge 3 commits into
base: main
Choose a base branch
from

Conversation

alisher-ai
Copy link

@alisher-ai alisher-ai commented Mar 25, 2025

Implemented a lazy data loading inisaac_sim_dataset.py and made it configurable (check base_train_config.gin pretrained_gwm_train_config.gin)

previously;

  • the whole dataset was preloaded to the RAM before the training starts,
  • during the training, the data was sampled from preloaded data on RAM.

now;

  • the file paths are stored on the memory,
  • during the training, file paths are read and loaded to the memory by batch.

outcomes;

  • saves RAM space,
  • faster experimentation (data preloading takes time --> the actual training starts slower --> slower iterations),
  • probably a bit slower training, but I did not see any meaningful differences based on single epoch training times,
  • lazy_loading can be controlled in config (.gin) files, easy to switch between two modes.

@nv-liuw
Copy link
Collaborator

nv-liuw commented Apr 2, 2025

Thanks for the contribution @alisher-ai , LGTM, just a couple of small comments.

@nv-liuw nv-liuw self-requested a review April 2, 2025 19:06
@alisher-ai alisher-ai requested a review from nv-liuw April 3, 2025 03:03
Copy link
Collaborator

@nv-liuw nv-liuw left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM, thanks for the contribution!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants