You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Thanks for sharing the code and model. I'm trying to reproduce the SVAR results in the paper and find I can make the EER down to 1.57% with a threshold of 0.7896. However, with this threshold, I only got a SVAR around 25% on the conversion result.
For each of the 110 speakers (a newer version of VCTK was adopted, which contains 110 speakers), calculate a mean GE2E feature as the anchor by averaging all the features of that speaker.
For all utterances, calculate the cosine similiarity between its GE2E feature and the 110 anchors.
Do a binary search to find the threshold.
And after that, I got a EER of 1.57%, which is lower than the paper's 5.6%, the script I used is as follows:
Do you mind to share how do you build the SV system or the threshold adopted in the paper? It will be helpful.
The other question is that how do you split the training/testing data for the seen-to-seen. In my case, I only sampled 1000 utterances from the dataset randomly, it may by convered in the training. It will be good to get the same test set. Many thanks.
The text was updated successfully, but these errors were encountered:
enhuiz
changed the title
SVAR calculation and Dataset split
SVAR calculation and dataset split
Aug 3, 2021
Thanks for sharing the code and model. I'm trying to reproduce the SVAR results in the paper and find I can make the EER down to 1.57% with a threshold of 0.7896. However, with this threshold, I only got a SVAR around 25% on the conversion result.
My procedure of building the SV system is:
And after that, I got a EER of 1.57%, which is lower than the paper's 5.6%, the script I used is as follows:
Do you mind to share how do you build the SV system or the threshold adopted in the paper? It will be helpful.
The other question is that how do you split the training/testing data for the seen-to-seen. In my case, I only sampled 1000 utterances from the dataset randomly, it may by convered in the training. It will be good to get the same test set. Many thanks.
The text was updated successfully, but these errors were encountered: