You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have reviewed the paper, which asserts that there is no audio or information from unknown speakers during the training process. However, in your code, your validation set includes information from unknown speakers. Doesn't this imply that the model had already assimilated information from unknown speakers during training, and the reference speech used for inference corresponds to the speech encountered during training?
For speaker p261,the model already meet this speaker?
please help me!
The text was updated successfully, but these errors were encountered:
First of all, I am not the author of the paper. The purpose of this repo is to test the model, and the authors’ implementation and dataset configuration may differ. Additionally, in the current code, there is no bias in tuning the model because only a single test sample is generated (link), and there is no validation loss calculation during the training process.
I have reviewed the paper, which asserts that there is no audio or information from unknown speakers during the training process. However, in your code, your validation set includes information from unknown speakers. Doesn't this imply that the model had already assimilated information from unknown speakers during training, and the reference speech used for inference corresponds to the speech encountered during training?
For speaker p261,the model already meet this speaker?
please help me!
The text was updated successfully, but these errors were encountered: