Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to train with large dataset #196

Open
Bach1502 opened this issue Oct 2, 2021 · 5 comments
Open

How to train with large dataset #196

Bach1502 opened this issue Oct 2, 2021 · 5 comments

Comments

@Bach1502
Copy link

Bach1502 commented Oct 2, 2021

Hello,
I believe that this is a fairly simple question but since I'm very new to ML in general, it still baffles me. I just followed the training instruction and has successfully trained my model on one pair of data (a clean speech.wav and a noise.wav) now I want to ask how can you repeat this process for larger dataset, I'm currently having a set of data with 300 files for these 2 categories and I don't think repeating this process 300 times is the way I should go.

Thanks.

@Zadagu
Copy link

Zadagu commented Oct 12, 2021

just concatenate the audio files.
But you need to be aware, that the input format is not .wav it's plain pcm without any header.

@Bach1502
Copy link
Author

thank you, I will try it to see if it works

@ZihCode
Copy link

ZihCode commented Aug 4, 2022

I want to know how to concatenate the audio files. Did you use any useful tools?Or just copy the RAW files and paste them into one file? How can I get a long RAW data? I would be very grateful if you could help me

@Zadagu
Copy link

Zadagu commented Aug 4, 2022

I wrote a python script to concatenate files. For reading audio files I used the soundfile package and resampled if needed using scipy.

@Zadagu
Copy link

Zadagu commented Aug 9, 2022

Sorry, but I think your behavior in the GitHub issues is somewhat inappropriate.
You spammed the very same question three times across multiple issues:
#208
#201 (comment)
#196
You can answer your question yourself, by reading the rnnoise paper and newer speech enhancement papers.
They all report numbers on how much data they are using.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants