https://uberduck.ai/ logo
Join Discord
Powered by
# talknet-support
  • w

    WeegeeFan1

    11/05/2022, 12:09 AM
    Since it's all limited
  • w

    WeegeeFan1

    11/05/2022, 12:09 AM
    For some reason my voice I'm doing now can't do anything exept regurgitate what I taught it
  • w

    WeegeeFan1

    11/05/2022, 12:09 AM
    Otherwise it has seizures
  • w

    WeegeeFan1

    11/05/2022, 12:09 AM
    I've been told it's because I don't have enough audio so i just added more
  • w

    WeegeeFan1

    11/05/2022, 12:09 AM
    like 15 minutes more (originaly had 13)
  • s

    Skyler

    11/05/2022, 12:10 AM
    might've fine tuned it too much as well, make sure you have a holdout set of wave files you use for a validation file
  • w

    WeegeeFan1

    11/05/2022, 12:10 AM
    What's the validation file actaully do?
  • w

    WeegeeFan1

    11/05/2022, 12:10 AM
    I just copy the list.txt filepath again
  • s

    Skyler

    11/05/2022, 12:11 AM
    basically that's data which is never shown to the network for training but periodically compared against your network during training to see a) that its learning the patterns and b) not just learning to mimic the exact training data
  • w

    WeegeeFan1

    11/05/2022, 12:12 AM
    So is that my issue?
  • w

    WeegeeFan1

    11/05/2022, 12:12 AM
    Do I just do the audio in the zip but seperate the transcriptions between two files?
  • s

    Skyler

    11/05/2022, 12:12 AM
    if you have no validation data that's most likely the issue, note: validation.txt and training.txt should have no overlaps
  • w

    WeegeeFan1

    11/05/2022, 12:12 AM
    ahhhhhhhhhh
  • w

    WeegeeFan1

    11/05/2022, 12:13 AM
    ive not been dfoing that at all
  • w

    WeegeeFan1

    11/05/2022, 12:13 AM
    I'll do this next model since im already training it. If there's issues I'll go back and take a third of it or so out
  • s

    Skyler

    11/05/2022, 12:13 AM
    basically yes, and make sure that you point to the validation script in your config files (or CLI inputs)
  • w

    WeegeeFan1

    11/05/2022, 12:13 AM
    What do I do though if some of it is singing and some of it is talking?
  • w

    WeegeeFan1

    11/05/2022, 12:13 AM
    Wouldn't it pick up on inconsistant things?
  • w

    WeegeeFan1

    11/05/2022, 12:13 AM
    Since one of them is natural, one is on beat
  • s

    Skyler

    11/05/2022, 12:14 AM
    I haven't worked with singing actually which is why I was asking here
  • w

    WeegeeFan1

    11/05/2022, 12:14 AM
    Ahh
  • w

    WeegeeFan1

    11/05/2022, 12:14 AM
    I've never worked with tacotron either
  • s

    Skyler

    11/05/2022, 12:17 AM
    @Justin actually I was looking at your offline talknet singing repo, i think its forked from the offline model I used to do general talknet training but can I use either of those repos to make a multipurpose voice and do I need to do anything specific (like make regular audio and singing into seperate speakers) or I can train it all as once. If I want to train both at once do I use this script https://github.com/justinjohn0306/ControllableTalkNet-Singer or the general offline talknet its forked from
  • w

    WeegeeFan1

    11/05/2022, 12:18 AM
    How do you train something for multiple speakers? What's that mean in techincal terms.
  • w

    WeegeeFan1

    11/05/2022, 12:18 AM
    I'm computer savvy just new to voice AI stuff
  • s

    Skyler

    11/05/2022, 12:18 AM
    you know how the dataset has fileloc|text spoken|id format
  • w

    WeegeeFan1

    11/05/2022, 12:18 AM
    Yes
  • w

    WeegeeFan1

    11/05/2022, 12:19 AM
    Welll id format isn't a thing in talknet as far as I can tell
  • w

    WeegeeFan1

    11/05/2022, 12:19 AM
    This is a thing that I do
  • s

    Skyler

    11/05/2022, 12:19 AM
    that id can correspond to different people for talknet2 models, and then you can train a multiple speaker dataset
1...363738...74Latest