Hello again!
I'm kind of feeling out the vibes of this before I commit to it, so I'd like some community opinion.
Firstly, I'm very happy with the current progress of 26's newest bank set, and I'm going to continue developing and recording it post-release in January.
Despite the aggressive accent on the JP bank (which I'm kind-of content with, as I'm not trying to explicitly imitate a native Japanese tone), the tone variance between the recorded samples make 26 sound MUCH more natural than previous versions. (Putting that BA in theatre to USE mama.)
I have no plans at this point to re-cast voice providers for the other Row 4 Banks at this time.
Secondly, I'm seeing a lot of "new" (as in, created within the time of my absence in the community whilst I was pursuing higher education, yeehaw) kinds of voicebanks using ENUNU/NNSVS AI model training.
I don't see a reason not to create a "26 AI" bank, but at the same time I'm kind of strong-arming a traditional bank out.
My personal pros and cons are this:
PROS:
Much more natural sounding banks
Would properly encapsulate the tones and ranges in vocals I have
Would become (if I'm understanding this correctly) compatible with other vocal synths like DeepVocal
CONS:
Whole lotta data to process
Not as streamlined for widespread use
With the above information in consideration, I still have a few questions.
What do you, as a regular or semi-regular UTAU user, feel the benefit of an AI-compatible/configured bank is?
How would/do you use regular banks in comparison to AI banks?
What do you prefer in terms of tone? (More power, character based, softer, etc)
I'm kind of feeling out the vibes of this before I commit to it, so I'd like some community opinion.
Firstly, I'm very happy with the current progress of 26's newest bank set, and I'm going to continue developing and recording it post-release in January.
Despite the aggressive accent on the JP bank (which I'm kind-of content with, as I'm not trying to explicitly imitate a native Japanese tone), the tone variance between the recorded samples make 26 sound MUCH more natural than previous versions. (Putting that BA in theatre to USE mama.)
I have no plans at this point to re-cast voice providers for the other Row 4 Banks at this time.
Secondly, I'm seeing a lot of "new" (as in, created within the time of my absence in the community whilst I was pursuing higher education, yeehaw) kinds of voicebanks using ENUNU/NNSVS AI model training.
I don't see a reason not to create a "26 AI" bank, but at the same time I'm kind of strong-arming a traditional bank out.
My personal pros and cons are this:
PROS:
Much more natural sounding banks
Would properly encapsulate the tones and ranges in vocals I have
Would become (if I'm understanding this correctly) compatible with other vocal synths like DeepVocal
CONS:
Whole lotta data to process
Not as streamlined for widespread use
With the above information in consideration, I still have a few questions.
What do you, as a regular or semi-regular UTAU user, feel the benefit of an AI-compatible/configured bank is?
How would/do you use regular banks in comparison to AI banks?
What do you prefer in terms of tone? (More power, character based, softer, etc)