Last reviewed: 3/23/2024 10:29:23 AM
Speech Management Architecture
The SpeechKit class library includes classes for managing speech recognition and speech synthesis.
Speech Recognition Architecture
SpeechKit provides a productive way to develop software that listens. Applications set properties and invokes methods through a speech recognition management class. This class handles the low-level functions with speech recognition engines (i.e., recognizers).
Applications establish a session with a recognizer through which spoken language captured live via a microphone or from recorded audio is processed and converted to text. Applications use SpeechKit to manage the activities for speech recognition on their behalf. SpeechKit manages the resources and interacts directly with the speech application program interface (API). SpeechKit supports the following speech APIs for speech recognition:
- Apple Speech,
- Google android.speech,
- Microsoft Azure Speech,
- Microsoft SAPI 5,
- Microsoft Speech Platform,
- Microsoft System.Speech,
- Microsoft Microsoft.Speech,
- Microsoft WindowsMedia, and
- Nuance Dragon NaturallySpeaking.
Applications receive recognized speech as text and notification of other processing states through event callbacks.
SpeechKit encapsulates all of the technologies necessary to make the process of recognizing speech simple and efficient for applications.
SpeechKit simplifies the process of recognizing speech by handling the low-level activities directly with a recognizer.
Instantiate SpeechKit to recognize speech within the application and destroy SpeechKit to release its resources when speech recognition is no longer needed.
Speech Synthesis Architecture
SpeechKit provides a productive way to develop software that speaks. Applications set properties and invoke methods through the speech synthesis management class. This class handles the low-level functions with text-to-speech engines (i.e., synthesizers or voices).
Applications establish a session with a synthesizer through which speech is synthesized from text. Applications use SpeechKit to manage the synthesizer resources on their behalf. SpeechKit manages the resources and interacts directly with a speech application program interface (API). SpeechKit supports the following speech APIs for speech synthesis:
- Acapela TTS,
- Apple AVFoundation TTS,
- Cepstral Swift,
- CereProc CereVoice,
- Google android.speech.tts,
- Microsoft Azure Speech,
- Microsoft SAPI 5,
- Microsoft Speech Platform,
- Microsoft System.Speech,
- Microsoft Microsoft.Speech, and
- Microsoft WindowsMedia.
Application receive notification of other processing states through event callbacks.
SpeechKit simplifies the process of synthesizing speech by handling the low-level activities directly with a synthesizer.
Instantiate SpeechKit to synthesize speech within the application and destroy SpeechKit to release its resources when speech synthesis is no longer needed.