The way SynthV works makes me feel like every voicebank basically "sings the same way" if you allow the model to 'drive'. If you turn down the influence of the AI, you regain some of the flexibility of the concatenative synths, but you do lose another element of the realism of the SynthV models.
On the other hand, if you let the AI do what it does best, you'll end up with an extremely convincing render, but the... quality? The... motif? I don't really know what vocabulary to use to describe what I'm trying to say, but hopefully it'll make sense, of the voice, will be the same throughout any voicebank you plug in there.
Yeah, it'll be significantly different sounding if you put a seriously different VB in there, like a bass voice (ASTERIAN) or a whisper voice (Chifuyu) or something of a different quality altogether (YUMA) but if they were actors in a play, I think the feeling is as if the singers are delivering their lines basically the same way, even in these edge cases. (Someone mentioned on the boards that they felt SOLARIA's release version felt like a disney princess-ified version of her voice provider/unfinished VB demo songs, and I think at least part of that effect can absolutely be owed to the current state of SynthV's model rendering. There really was, literally, less data for it to "drive" on back then, lol. Doesn't every singer kind of have a disney princess-type approach to singing? Even Feng Yi, who has an "opera" vocal mode?)
This is definitely SynthV working as intended. SynthV is hard-focusing a certain market that I think they feel does not always have overlapping sensibilities with the existing vocal synth doujin music contingent. So for these demos, I think partially at fault is the realism itself. There might be a different VB in the cartridge slot, but the model is singing each song the same way. And it's a model that they've definitely trained to sing the same kinds of songs they always demo realistically. They want to present these products under the most favorable conditions, so this totally makes sense.
But if Dreamtonics wants to express to the user how they can imbue their own emotion and artistic sensibilities into the words their voicebanks sing, it's definitely the wrong angle for all of their advertising, right?
The incredible quality of the AI has a 'wow' factor that can't be understated, but as an instrument, I think it'd be cool for Dreamtonics to go further in depth as to how SynthV works as an instrument instead of just as a simple plug-and-play AI model. But maybe that isn't where the money is right now?
I think whether or not they follow up on that will be a big factor as to whether or not SynthV serious long-term staying power among DTM people. (And I really do think it can but we're still in the beginning days in the big scheme of things)