This guide is for integrating speech-to-text and text-to-speech functionality of Azure Cognitive Services.
We assume you have already set up a bot and have Web Chat running on a page.
Sample code in this article is optimized for modern browsers. You may need to use a transpiler (e.g. Babel) to target a broader range of browsers.
In order to use the speech functionality in Web Chat, the browser needs to provide minimal media capabilities, including recording from microphone and playing audio clips.
Internet Explorer 11 does not meet the basic requirements for both speech recognition and speech synthesis.
- Browser must support WebRTC API
- Available on most modern browsers on desktop and mobile platform
- WebRTC will not work on third-party apps or browsers on iOS
- Web page must be hosted over HTTPS
- User must explicitly grant permission for microphone access
Safari is the only browser that supports WebRTC API on iOS.
Chrome, Edge and native apps built using WKWebView
do not support WebRTC API. Apps based on Cordova/PhoneGap and React Native WebView might need additional plugins or custom code to support WebRTC API.
- Browser must support Web Audio API
- Available on all modern browsers on desktop and mobile platform
Safari requires additional permission granted implicitly by the user. The user needs to perform an interaction (click/tap/type) before any audio clips can be played during the browser session.
When the user taps on the microphone button for the first time, Web Chat will play a very short and silent audio clip. This will enable Web Chat to play any audio clip synthesized from bot messages.
If you customize Web Chat to perform any text-to-speech operations before user gestures, Web Chat will be blocked by Safari for audio playback. Thus, bot messages will not be synthesized.
You can present a splash screen with a tap-to-continue button, which will ready the engine by sending an empty utterance to unblock Safari.
To use Cognitive Services in Web Chat, you will need to add minimal setup code to set up Web Chat with your Cognitive Services subscription.
You will need to obtain a subscription key for your Azure Cognitive Services subscription. Please follow instructions on this page to obtain a subscription key.
To prevent exposing your subscription key, you should build/host a server that uses your subscription key to generate authorization tokens, and only send the authorization token to the client. You can find more information about authorization tokens on the Cognitive Services Authentication documentation. Never use subscription keys to access your Azure resources in a production environment.
To use new voices powered by deep neural network, you might need to have a subscription in "West US 2" region or other supported regions.
This integration code is excerpted from the sample named "Integrating with Cognitive Services Speech Services".
To bring more focus to the integration part, we simplified the original sample code by using subscription key instead of authorization token. You should always use authorization token for production environment.
const {
createCognitiveServicesSpeechServicesPonyfillFactory,
createDirectLine,
renderWebChat
} = window.WebChat;
renderWebChat({
directLine: createDirectLine({
secret: 'YOUR_DIRECT_LINE_SECRET'
}),
language: 'en-US',
webSpeechPonyfillFactory: await createCognitiveServicesSpeechServicesPonyfillFactory({
region: 'YOUR_REGION',
subscriptionKey: 'YOUR_SUBSCRIPTION_KEY'
})
}, document.getElementById('webchat'));
After adding the ponyfill factory, you should be able to see the microphone button in Web Chat, which indicates that speech is enabled.
These features are for improving the overall user experience while using speech in Web Chat.
- Using Speech Synthesis Markup Language
- Using input hint
- Selecting voice
- Custom Speech
- Custom Voice
- Text-to-speech audio format
- Text normalization options
- Disabling telemetry
- Using authorization token
- Using two subscription keys for speech-to-text and text-to-speech
Instead of synthesizing text, Web Chat can also synthesize Speech Synthesis Markup Language (or SSML). Cognitive Services supports SSML 1.0 with "mstts" extensions.
When the bot sends the activity, include the SSML in the speak
property.
<speak
version="1.0"
xmlns="https://www.w3.org/2001/10/synthesis"
xmlns:mstts="https://www.w3.org/2001/mstts"
xml:lang="en-US"
>
<voice name="en-US-JessaNeural">
<mstts:express-as type="cheerful">That'd be just amazing!</mstts:express-as>
</voice>
<voice name="ja-JP-Ayumi-Apollo">
<prosody pitch="+150%">素晴らしい!</prosody>
</voice>
</speak>
The SSML code snippet above is using neural voice "JessaNeural" for expression, which is only supported in some regions.
With "mstts" extension, you can also add speaking style (e.g. cheerful) and background audio to your synthesized speech.
The bot can set input hint when sending activity to the user to indicate whether the bot is anticipating user input. This can be used re-open the microphone if the last message was sent through microphone. You can set it to either expectingInput
, acceptingInput
, and ignoringInput
. If it is not defined, it will default to acceptingInput
.
"expectingInput"
: Web Chat will open the microphone after the bot's message is spoken and the last message was sent through microphone"acceptingInput"
: Web Chat will do nothing after the bot's message is spoken"ignoringInput"
: Web Chat will explicitly close the microphone
For more details, please follow this article on adding input hints to messages.
In issue #2022, it was brought to the Web Chat team's attention that the speech behavior of v3 and v4 of Web Chat do not match. In the 4.5.0 release, the expected behavior of a speech bot has been modified in order to bring parity to v3 behavior regarding input hint. This means the following:
- Expecting input will now be respected by Web Chat and open the microphone during a speech conversation. This is assuming that the user has given permission for the browser to use the mic.
- Accepting input will no longer open the mic after the bot has responded to a speech activity from the user. Instead, the user will have to press the microphone button again in order to further interact with the bot.
- Ignoring input will continue to not open the mic after a speech activity has been sent from the bot.
Different voices can be selected based on the synthesizing activity.
By default, we will use the following order to determine which voice to use. If available, we prioritize deep neural network-based voices over traditional voices.
- Locale specified in the activity
- Display language of Web Chat
- Browser language
- English (US)
- The first available voice
In the following code, voice is selected based on the language of the synthesizing activity. If the activity is in Cantonese (zh-HK), we will select the voice with keyword "TracyRUS". Otherwise, we will select the voice with keyword "Jessa24kRUS".
const {
createCognitiveServicesSpeechServicesPonyfillFactory,
createDirectLine,
renderWebChat
} = window.WebChat;
renderWebChat({
directLine: createDirectLine({
secret: 'YOUR_DIRECT_LINE_SECRET'
}),
language: 'en-US',
+ selectVoice: (voices, activity) => {
+ if (activity.locale === 'zh-HK') {
+ return voices.find(({ name }) => /TracyRUS/iu.test(name));
+ } else {
+ return voices.find(({ name }) => /Jessa24kRUS/iu.test(name));
+ }
+ }
webSpeechPonyfillFactory: await createCognitiveServicesSpeechServicesPonyfillFactory({
region: 'YOUR_REGION',
subscriptionKey: 'YOUR_SUBSCRIPTION_KEY'
})
}, document.getElementById('webchat'));
Custom Speech is a trained recognition model that improves recognition of words that are not in the default recognition model. For example, you can use it to improve accuracy when recognizing trademark words or the name of a person.
First, you need to set up a Custom Speech project. Please follow the article on creating a new Custom Speech project.
After your Custom Speech project is set up and a model is published to a deployment endpoint, in the "Deployment" tab, save your "Endpoint ID".
You will then need to modify your integration code as below.
renderWebChat({
directLine: createDirectLine({
secret: 'YOUR_DIRECT_LINE_SECRET'
}),
language: 'en-US',
webSpeechPonyfillFactory: await createCognitiveServicesSpeechServicesPonyfillFactory({
region: 'YOUR_REGION',
+ speechRecognitionEndpointId: 'YOUR_ENDPOINT_ID',
subscriptionKey: 'YOUR_SUBSCRIPTION_KEY'
})
}, document.getElementById('webchat'));
Custom Voice is a trained synthesis model for providing your user with an unique synthesized voice when performing text-to-speech.
First, you need to set up a Custom Voice project. Please follow the article on creating a new Custom Voice project.
After your Custom Voice project is set up and a model is published to a deployment endpoint, in the "Deployment" tab, save the "Model / Voice name" and the value of "deploymentId" inside the "Endpoint URL" field.
You will then need to modify your integration code as below. The selectVoice
function will be used to choose which trained synthesis model to use.
renderWebChat({
directLine: createDirectLine({
secret: 'YOUR_DIRECT_LINE_SECRET'
}),
language: 'en-US',
+ selectVoice: () => ({ voiceURI: 'YOUR_VOICE_MODEL_NAME' }),
webSpeechPonyfillFactory: await createCognitiveServicesSpeechServicesPonyfillFactory({
region: 'YOUR_REGION',
+ speechSynthesisDeploymentId: 'YOUR_DEPLOYMENT_ID',
subscriptionKey: 'YOUR_SUBSCRIPTION_KEY'
})
}, document.getElementById('webchat'));
Note: if you send SSML instead of plain text, make sure the voice model is correctly selected in your SSML.
To conserve bandwidth, you can set the text-to-speech audio format to one that consumes less bandwidth by modifying your integration code as below:
renderWebChat({
directLine: createDirectLine({
secret: 'YOUR_DIRECT_LINE_SECRET'
}),
language: 'en-US',
webSpeechPonyfillFactory: await createCognitiveServicesSpeechServicesPonyfillFactory({
region: 'YOUR_REGION',
+ speechSynthesisOutputFormat: 'audio-24khz-48kbitrate-mono-mp3',
subscriptionKey: 'YOUR_SUBSCRIPTION_KEY'
})
}, document.getElementById('webchat'));
Please refer to the following list of supported audio formats.
Text normalization is an ability to modify how the speech engine normalizes text. For example, when the user says, "I would like to order 2 4-piece chicken nuggets." It could be recognized as "two four piece" (default) or "2 four piece" (inverse text normalization, or ITN).
You can read more about various text normalization options.
To select different text normalization options, you will need to modify your integration code as below:
renderWebChat({
directLine: createDirectLine({
secret: 'YOUR_DIRECT_LINE_SECRET'
}),
language: 'en-US',
webSpeechPonyfillFactory: await createCognitiveServicesSpeechServicesPonyfillFactory({
region: 'YOUR_REGION',
subscriptionKey: 'YOUR_SUBSCRIPTION_KEY',
+ textNormalization: 'itn'
})
}, document.getElementById('webchat'));
Supported text normalization options are
"display"
(default),"itn"
,"lexical"
, and"maskeditn"
.
By default, Azure Cognitive Services will collect telemetry for service performance. If you prefer to disable telemetry, please modify the code as below:
renderWebChat({
directLine: createDirectLine({
secret: 'YOUR_DIRECT_LINE_SECRET'
}),
language: 'en-US',
webSpeechPonyfillFactory: await createCognitiveServicesSpeechServicesPonyfillFactory({
+ enableTelemtry: false,
region: 'YOUR_REGION',
subscriptionKey: 'YOUR_SUBSCRIPTION_KEY',
})
}, document.getElementById('webchat'));
This sample uses a subscription key for simplicity. If you need to use authorization token, you can pass a Promise
function to the authorizationToken
property. The function could potentially be a network call to your token server.
async function fetchAuthorizationToken() {
const res = await fetch('/api/speechtoken');
if (res.ok) {
return await res.text();
} else {
throw new Error('Failed to retrieve authorization token for Cognitive Services.');
}
}
renderWebChat({
directLine: createDirectLine({
secret: 'YOUR_DIRECT_LINE_SECRET'
}),
language: 'en-US',
webSpeechPonyfillFactory: await createCognitiveServicesSpeechServicesPonyfillFactory({
+ // Note we are passing the function, not the result of the function call, as there is no () appended to it.
+ // This function will be called every time the authorization token is being used.
+ authorizationToken: fetchAuthorizationToken,
region: 'YOUR_REGION',
- subscriptionKey: 'YOUR_SUBSCRIPTION_KEY',
})
}, document.getElementById('webchat'));
The function fetchAuthorizationToken
will be called every time a token is needed. For simplicity, token caching is not provided in this sample code. You should add caching based on the validity of the token.
In some cases, you may be using two different Cognitive Services subscriptions, one for speech-to-text and another one for text-to-speech. You could create two ponyfills and combine together to form a hybrid ponyfill.
+ const speechToTextPonyfillFactory = await createCognitiveServicesSpeechServicesPonyfillFactory({
+ region: 'YOUR_STT_REGION',
+ subscriptionKey: 'YOUR_STT_SUBSCRIPTION_KEY',
+ });
+ const textToSpeechPonyfillFactory = await createCognitiveServicesSpeechServicesPonyfillFactory({
+ region: 'YOUR_TTS_REGION',
+ subscriptionKey: 'YOUR_TTS_SUBSCRIPTION_KEY',
+ });
renderWebChat({
directLine: createDirectLine({
secret: 'YOUR_DIRECT_LINE_SECRET'
}),
language: 'en-US',
- webSpeechPonyfillFactory: await createCognitiveServicesSpeechServicesPonyfillFactory({
- region: 'YOUR_REGION',
- subscriptionKey: 'YOUR_SUBSCRIPTION_KEY',
- })
+ webSpeechPonyfillFactory: options => {
+ const { SpeechGrammarList, SpeechRecognition } = speechToTextPonyfillFactory(options);
+ const { speechSynthesis, SpeechSynthesisUtterance } = textToSpeechPonyfillFactory(options);
+
+ return {
+ SpeechGrammarList,
+ SpeechRecognition,
+ speechSynthesis,
+ SpeechSynthesisUtterance
+ };
+ }
}, document.getElementById('webchat'));
Note: it is correct that
speechSynthesis
is in camel-casing, while others are in Pascal-casing.
Using this approach, you can also combine two polyfills of different types. For example, speech recognition powered by Cognitive Services with browser-supported speech synthesis. You can refer to our Web Chat sample on hybrid speech to learn more.
- Try Cognitive Services
- List of browsers which support Web Audio API
- List of browsers which support WebRTC API
- Speech Synthesis Markup Language (SSML)
- Add input hints to messages
- Get started with Custom Voice
- What is Custom Speech
- Sample: Integrating with Cognitive Services Speech Services
- Sample: Using hybrid speech engine