I agree with a number of the points made here.
On the one hand, in working with Miku English V4, I've personally found her pronunciations quite understandable (though not always perfectly so) when I've just put the words in and let the dictionary do its work. She's good enough, in my mind, that in listening to some of the songs where she struggles, I wonder how it's happening. I think it may be due to a number of potential causes:
I don't think either approach is wrong. I'm personally pretty happy with what Miku does on her own, though as I have the need I will probably hope to learn to use phoneme replacement to help her pronunciations more than I currently do. Having said that, I think criticism against her bank could reach the point of being unfair if the complaint is that she doesn't have ultra smooth, native pronunciation. Just like Saki Fujita, she's a Japanese woman singing in English. That's a steeper hill to climb for her and for you as a producer than if you were working with a native Engloid. It's definitely doable--look at Lupin's songs, or "Thousand Little Voices" by Vault Kid and Flanger Moose--but it's more work.
On the one hand, in working with Miku English V4, I've personally found her pronunciations quite understandable (though not always perfectly so) when I've just put the words in and let the dictionary do its work. She's good enough, in my mind, that in listening to some of the songs where she struggles, I wonder how it's happening. I think it may be due to a number of potential causes:
- Possibly the producer of said song was using Miku V3 or Engrish. IMO, V4 does a much better job than V3, although V3 can be invaluable for providing a darker tone via XSY.
- Since I'm putting the words in during my usage, I know what the words are already, so it's harder for me to judge fairly.
- It's the work with the parameters that make the words in the songs less understandable. I think it's possible that there is a "best practice" workflow that produces better results. In my theorization, this is based on a general rule of making the biggest, most sweeping changes through effects, plugins, and DAW automation, and smaller changes via the vocal synth parameters. I think if you work with the parameters first and push them too hard, you may end up making a sound that's more distorted than artistically improved.
- Possibly the mixing/conflicts between the vocals and the background are to blame.
- XSY can cause weird issues that would seem difficult/impossible to fix.
I don't think either approach is wrong. I'm personally pretty happy with what Miku does on her own, though as I have the need I will probably hope to learn to use phoneme replacement to help her pronunciations more than I currently do. Having said that, I think criticism against her bank could reach the point of being unfair if the complaint is that she doesn't have ultra smooth, native pronunciation. Just like Saki Fujita, she's a Japanese woman singing in English. That's a steeper hill to climb for her and for you as a producer than if you were working with a native Engloid. It's definitely doable--look at Lupin's songs, or "Thousand Little Voices" by Vault Kid and Flanger Moose--but it's more work.