-
Notifications
You must be signed in to change notification settings - Fork 7
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Train First Speaker Verification Model #19
Comments
Cannot directly use VoxCeleb1 data with SincNet. We need to split the samples in chunks, like they explain in the paper (at least at the beginning). @hbredin any way of chunking audio segments already implemented in pyannote? |
Not sure what you need.
Can you clarify your needs? |
Nevermind, I misunderstood the |
@hbredin Please tell me if I can do something to help solve the 3199 issue. |
I am not sure why this happens. One way of understanding this behavior is to create a simple script that simply iterates forever on SpeechSegmentGenerator and stops as soon as the number of samples is not 3200. You can edit SpeechSegmentGenerator temporarily so that it also returns the value of |
Starting from here, we will be able to investigate what is happening |
Got it. I'm switching to STS for the time being, to integrate the model and dataset. |
Will unblock and use 0-padding for samples with wrong dimensions while we still look for what's causing this problem. |
The validation code is too expensive to run after each epoch, will use a separate script to run validations in parallel when a model is saved. This will be done as part of another issue: #25 |
Update: Validation can be done in-training for VoxCeleb1, but it would be useful to parallelize anyway to tackle VoxCeleb2. |
Use SpeakerModel and VoxCeleb1 to train a first speaker verification model.
Cross entropy takes priority over the rest of the losses as discussed in previous meetings.
Make sure to use EER as the validation metric.
The text was updated successfully, but these errors were encountered: