-
Notifications
You must be signed in to change notification settings - Fork 61
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Memory issue? #11
Comments
You can try to i) reduce the number of training images ii) reduce feature dimensionality iii) store features on the disk and preprocess them by batches. |
Hi, thanks for the wonderful work and codes you provided. Can you explain how would I implement the last one (iii)? How do I process the features by batches if they are too huge that I cannot even store them on the disk? I have encounted the same problem as @HOMGH and the extracted pixel representations are so huge that before the prepare_data() function return the data, the process I'm running the script will be killed. For example, when I run ddpm on the cat_15 dataset using the original experiment setting with 30 training images, when the program tries to run the following two lines, the process will crash.
My solution is to write another prepare_data() function, which processes one image instead of all labelled training images at one time. Then during the training of the pixel classifier, at each epoch I will create a dataloader for one image and iterate through all training images. But there exists some gap between the final evaluation results I got and yours. Do you have any suggestions for that? Thanks a lot for your help. |
|
Hello ,@Yi-Lynn , can you send me your code ? |
@Yi-Lynn Could you send me your code? please? |
@SahadevPoudel Hi, Poudel! I had the same problem, and I succeeded in implementing it in the way shown above. The newly proposed method declares a class called 'DividedImageLabelDataset' from src/datasets.py , which is imported from the train code and used. Please comment if you still want code about it! |
Hello! I'm experiencing the same problem, but I don't know what I should do to fix it. Can you share the code you modified? Thanks a lot. |
Hi,
Thanks for sharing your code.
I got "scripts/ddpm/train_interpreter.sh: line 6: 3842 Killed python train_interpreter.py --exp experiments/${DATASET}/ddpm.json $MODEL_FLAGS" error.
I have ~65G RAM available on my Ubuntu. Considering your note that "it requires ~210Gb for 50 training images of 256x256."
Does it mean that it's not feasible to train the model on my system? How about evaluation?
Thanks in advance.
The text was updated successfully, but these errors were encountered: