Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Requirements #48

Open
mateusztobiasz opened this issue Nov 22, 2024 · 2 comments
Open

Requirements #48

mateusztobiasz opened this issue Nov 22, 2024 · 2 comments

Comments

@mateusztobiasz
Copy link

Hello!

I am a beginner and I've wanted to play with your finetuning script so I decided to run it on Google Colab. Just to test it out, I prepared train, test and eval datasets which contained only one audio with caption. When I ran this script, it consumed about 15 GB of GPU vRAM and the process was killed. Is this a normal behaviour? If so, do you guys know how much vRAM do I need to finetune this model on more reasonable datasets (about 1000 rows)?

Thank you in advance for your reply!

@mateusztobiasz
Copy link
Author

I would like to also add that I haven't changed any hyperparameters and used audioldm_train/config/2023_08_23_reproduce_audioldm/audioldm_original_medium.yaml config. Also. here is the screenshot of the moment when process was killed:

image

@ahmetbekcan
Copy link

Hi, I also experience the same problem where T4 GPU is not sufficient. I don't know if it helps but I decreased the precision to "medium", decreased batch size to 1 but it is still not working. Could you find any solution to problem?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants