Replies: 19 comments
-
>>> carlfm01 |
Beta Was this translation helpful? Give feedback.
-
>>> alchemi5t |
Beta Was this translation helpful? Give feedback.
-
>>> carlfm01 |
Beta Was this translation helpful? Give feedback.
-
>>> alchemi5t |
Beta Was this translation helpful? Give feedback.
-
>>> erogol |
Beta Was this translation helpful? Give feedback.
-
>>> alchemi5t |
Beta Was this translation helpful? Give feedback.
-
>>> alchemi5t |
Beta Was this translation helpful? Give feedback.
-
>>> carlfm01 |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
>>> carlfm01 |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
>>> alchemi5t |
Beta Was this translation helpful? Give feedback.
-
>>> carlfm01 |
Beta Was this translation helpful? Give feedback.
-
>>> alchemi5t |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
>>> alchemi5t |
Beta Was this translation helpful? Give feedback.
-
>>> alchemi5t
[August 19, 2019, 8:15am]
Hello,
I have managed to train a model with 13hrs of annotated data. The
alignment is great and the words from the generated test sentences are
easily discernible. The only issue I have is that the generated audio is
not a 100% human-like( a hint of consistent robotic chop-ups in it). I
was wondering if I should fix that with post processing? or could I
handle it with hyper parameter tuning?
[This is an archived TTS discussion thread from discourse.mozilla.org/t/query-regarding-post-processing]
Beta Was this translation helpful? Give feedback.
All reactions