The controller interface for the speech service this can be used to retrieve information about the synthesis voices available on the device, start and pause speech, and other commands besides. You can get these spoken by passing them to the SpeechSynthesis.speak() method.įor more details on using these features, see Using the Web Speech API. Speech synthesis is accessed via the SpeechSynthesis interface, a text-to-speech component that allows programs to read out their text content (normally via the device's default speech synthesiser.) Different voice types are represented by SpeechSynthesisVoice objects, and different parts of text that you want to be spoken are represented by SpeechSynthesisUtterance objects. Grammar is defined using JSpeech Grammar Format ( JSGF.) The SpeechGrammar interface represents a container for a particular set of grammar that your app should recognize. Generally you'll use the interface's constructor to create a new SpeechRecognition object, which has a number of event handlers available for detecting when speech is input through the device's microphone. Speech recognition is accessed via the SpeechRecognition interface, which provides the ability to recognize voice context from an audio input (normally via the device's default speech recognition service) and respond appropriately. The Web Speech API makes web apps able to handle voice data.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |