-
Notifications
You must be signed in to change notification settings - Fork 6
Running on the new dataset #2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Hi @Huangmr0719 , thank you! It depends on the dataset, do the videos have ASR? And would you like to run the model with ASR only, captions only, ASR + captions, or even ASR + embeddings (SigLIP) + captions? If the videos already have ASR, the easiest and fastest way to get started is to run it with ASR only. You can find the relevant commands under If you want to use You’ll also need to create the following files under
Once you have the captions extracted and the files above ready, you can extract the chapters like this: python test.py subset=your-subset prompt=captions_asr I don’t have a script ready yet for running a single video with captions as I haven’t had the time, but it’s on my To Do list. Hope that helps! |
Thank you very much for your answer, it has well answered my question. I believe I need to use the ASR + captions setting on Once again, thank you for your answer, and I look forward to your future work! |
You're very welcome, glad it helped! Just a quick tip: before going all in with ASR + captions and Good luck with the experiments, and let me know if anything comes up! |
Thank you for your excellent work! I would like to use
chapter-llama
to conduct experiments on other datasets to explore its generalization performance withtest.py
. What pre-processing should I perform on these videos? After that, do I just need to replace the original content with annotations from the new dataset in theconfig
?The text was updated successfully, but these errors were encountered: