-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] The file direct_url.json
is not found
#1
Comments
Pip automatically creates the This part of It's strange that you're missing the In the meantime, you can give a from noisebase import Noisebase
data_loader = Noisebase(
'sampleset_v1', # Dataset name
{
'data_path': '/mnt/data/noisebase' # Wherever you store our datasets
}
) Scripts like nb-download --data_path="/mnt/data/noisebase" sampleset_v1 |
Thank you so much for your helpful response.
|
I would like to try the train.py script in your nppd project, but I’m not sure how to configure from noisebase import Noisebase
data_loader = Noisebase(
'sampleset_v1', # Dataset name
{
'data_path': '/mnt/data/noisebase' # Wherever you store our datasets
}
) within the noisebase.lighting.Trainer class. It seems that this class does not provide an interface for the data path. |
NPPD uses Hydra to manage configurations. No need to do this now, but you could put this option in nppd/conf/base.yaml: ...
training_data:
data_path: /mnt/data/noisebase
samples: 8
batch_size: 8
... I think I fixed the issue in e57974c. Could you try updating the package ( Pip actually doesn't create the |
Dear balintio, I appreciate you providing a solution to the previous issue, even though I haven’t been able to test it yet as I’m still waiting for the training dataset to be downloaded. Additionally, I had a follow-up inquiry. I’ve noticed that the dataset you are using appears to be in a custom file format, such as .0 and .zarray. I was wondering if there is a possibility of using more common data formats, such as HDR, PNG, or other publicly available data storage formats? This would allow me to potentially test your solution with my own dataset. I’m curious to know if supporting more standard data storage formats is something you’ve considered or if there are any technical limitations that prevent that. I’m interested in exploring the potential of your solution, and having the flexibility to use my own dataset would be very helpful. Please let me know your thoughts on this. I’m grateful for your assistance and look forward to your response. Best regards. |
No worries, I'm happy to help. The point of publicly releasing a project is to make sure people can use it😉 What you're seeing are Zarr files. It's a very fast hierarchical data format, like HDF5, if you know that one. Initially, we tried storing everything in EXR files, but the speed of our data loaders quickly bottlenecked the training process. Loading hundreds of files for every batch just wasn't a scalable approach. If you already have everything in individual files, packing them into Zarr files following our format should be easy. We don't have such example scripts in the repository yet, but I can add some and help write the script for your specific case. Just make sure you have all the data listed here in some form. A large part of this difficulty comes down to using per-sample data. We plan to share per-pixel versions of our datasets in June. They will also use the Zarr format, but they should be even easier to convert between individual files. |
Dear balintio, We are currently on a tight schedule as our project group is aiming to submit a related paper in May. It would be immensely helpful if you could provide a more detailed demonstration or guidance on how to pack our data into Zarr format datasets in the near future. This would greatly facilitate our preparation for the submission. Additionally, I have noticed that the zeroday data mentioned in your paper seems to differ from the assets available at https://developer.nvidia.com/orca/beeple-zero-day. I am curious whether you rendered the zeroday dataset using Falcor? If the zeroday scenes were indeed rendered with Falcor, could you please provide us with the configuration parameters used at the time, including scene luminance among others? Warm regards, |
Dear author,
I am currently trying to use the noisebase library for my work, but I encountered an issue with the
resolve_data_path()
function in thedata.py
file.noisebase/noisebase/data.py
Line 13 in 0bd8e02
In the function, there is a reference to a
direct_url.json
file, which is used to determine the location of the dataset. However, I was unable to find this file in the repository.Could you please provide some additional information about this
direct_url.json
file? What is its purpose, and where can I find it? I would greatly appreciate if you could clarify this for me.Additionally, if there is an alternative way to specify the dataset path, could you please share the details? I want to ensure that I can properly configure the library to work with my dataset.
Thank you in advance for your assistance. I look forward to your response.
Best regards.
The text was updated successfully, but these errors were encountered: