You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I would kindly request the following as due to limitations on hardware i.e. 8gb vram gpu
1.) Fp8 support for the second inpainting step. Although I could run the fp16 but Fp8 should be faster on limited hardwares.
2.) any possible way to improve the inpainting quality when using lower resolution than 4k.
3.) Could we have output as frames sequence and input as frame sequence so that longer videos could be processed as when the frame is inpainted it will be written to out rather than holding in memory or something like the video which is saved at last.
4.) Is cpu offloading is enabled for second step ?
5.) Is it possible to use this technique with images (single image 2d to 3d) instead of video?
The text was updated successfully, but these errors were encountered:
We use the diffusers to define and load the model, which supports giving the torch_dtype parameters in the from_pretrained function to change the data type, but we have not tested it on fp8.
Our method supports running the video at lower resolution like 1080p and works fine in our examples.
Yes, the streaming way could be implemented during testing.
We do not use it in the second step, but you could use it by adding the line "pipe.enable_model_cpu_offload()" function in diffusers.
Yes, you could test it on a video with a single frame as the image input.
We use the diffusers to define and load the model, which supports giving the torch_dtype parameters in the from_pretrained function to change the data type, but we have not tested it on fp8.
Our method supports running the video at lower resolution like 1080p and works fine in our examples.
Yes, the streaming way could be implemented during testing.
We do not use it in the second step, but you could use it by adding the line "pipe.enable_model_cpu_offload()" function in diffusers.
Yes, you could test it on a video with a single frame as the image input.
Thank you for the response.
Unfortunately I don't have much technical background for coding so if you could share the script for second step with cpu offloading would be very helpful.
For point 2: yes the method works with lower resolution but the inpainted right side lacks so much details and looks like oil painting. Slight fix is to increase min and max guidance to 1.5-2. Proper fix would be great if any.
For point 5 I tried with single images but out of right side is corrupted/doesn't work. It is not for still images I guess.
Hello,
Thank you for this wonderful project.
I would kindly request the following as due to limitations on hardware i.e. 8gb vram gpu
1.) Fp8 support for the second inpainting step. Although I could run the fp16 but Fp8 should be faster on limited hardwares.
2.) any possible way to improve the inpainting quality when using lower resolution than 4k.
3.) Could we have output as frames sequence and input as frame sequence so that longer videos could be processed as when the frame is inpainted it will be written to out rather than holding in memory or something like the video which is saved at last.
4.) Is cpu offloading is enabled for second step ?
5.) Is it possible to use this technique with images (single image 2d to 3d) instead of video?
The text was updated successfully, but these errors were encountered: