Rust bindings for Microsoft Cognitive Speech Services SDK. Provides thin abstraction around native C API. Heavily inspired by official Go library. Provides speech-to-text, text-to-speech and bot framework dialog management capabilities.
Pull requests welcome!
use cognitive_services_speech_sdk_rs as msspeech;
use log::*;
use std::env;
async fn speech_to_text() {
let filename = env::var("WAVFILENAME").unwrap();
let audio_config = msspeech::audio::AudioConfig::from_wav_file_input(&filename).unwrap();
let speech_config = msspeech::speech::SpeechConfig::from_subscription(
env::var("MSSubscriptionKey").unwrap(),
env::var("MSServiceRegion").unwrap(),
)
.unwrap();
let mut speech_recognizer =
msspeech::speech::SpeechRecognizer::from_config(speech_config, audio_config).unwrap();
speech_recognizer
.set_session_started_cb(|event| info!("set_session_started_cb {:?}", event))
.unwrap();
speech_recognizer
.set_session_stopped_cb(|event| info!("set_session_stopped_cb {:?}", event))
.unwrap();
speech_recognizer
.set_speech_start_detected_cb(|event| info!("set_speech_start_detected_cb {:?}", event))
.unwrap();
speech_recognizer
.set_speech_end_detected_cb(|event| info!("set_speech_end_detected_cb {:?}", event))
.unwrap();
speech_recognizer
.set_recognizing_cb(|event| info!("set_recognizing_cb {:?}", event.result.text))
.unwrap();
speech_recognizer
.set_recognized_cb(|event| info!("set_recognized_cb {:?}", event))
.unwrap();
speech_recognizer
.set_canceled_cb(|event| info!("set_canceled_cb {:?}", event))
.unwrap();
let result = speech_recognizer.recognize_once_async().await.unwrap();
info!("got recognition {:?}", result);
}
use cognitive_services_speech_sdk_rs as msspeech;
use log::*;
use std::env;
async fn text_to_speech() {
let pull_stream = msspeech::audio::PullAudioOutputStream::create_pull_stream().unwrap();
let audio_config = msspeech::audio::AudioConfig::from_stream_output(&pull_stream).unwrap();
let speech_config = msspeech::speech::SpeechConfig::from_subscription(
env::var("MSSubscriptionKey").unwrap(),
env::var("MSServiceRegion").unwrap(),
)
.unwrap();
let mut speech_synthesizer =
msspeech::speech::SpeechSynthesizer::from_config(speech_config, audio_config).unwrap();
speech_synthesizer
.set_synthesizer_started_cb(|event| info!("synthesizer_started_cb {:?}", event))
.unwrap();
speech_synthesizer
.set_synthesizer_synthesizing_cb(|event| info!("synthesizer_synthesizing_cb {:?}", event))
.unwrap();
speech_synthesizer
.set_synthesizer_completed_cb(|event| info!("synthesizer_completed_cb {:?}", event))
.unwrap();
speech_synthesizer
.set_synthesizer_canceled_cb(|event| info!("synthesizer_canceled_cb {:?}", event))
.unwrap();
match speech_synthesizer.speak_text_async("Hello Rust!").await {
Err(err) => error!("speak_text_async error {:?}", err),
Ok(speech_audio_bytes) => {
info!("speech_audio_bytes {:?}", speech_audio_bytes);
}
}
}
For more see github integration tests (tests folder) and samples (examples folder).
Currently build on Windows, Linux and MacOS is supported. Uses Clang and Microsoft Speech SDK shared libraries. Details can be found here here.
Install following prerequisites before running cargo build:
sudo apt-get update
sudo apt-get install clang build-essential libssl1.0.0 libasound2 wget
Build is generating Rust bindings for Speech SDK native functions. These are already prebuilt and put into ffi/bindings.rs file. In most cases it is not necessary to regenerate them. Set following to skip bindings regeneration:
export MS_COG_SVC_SPEECH_SKIP_BINDGEN=1
cargo build
Build process will download MS Speech SDK into target folder. From here you can copy it into other folder, e.g. ./SpeechSDK. When running compiled binary dynamic linking should be used:
Linux:
export LD_LIBRARY_PATH=/Users/xxx/cognitive-services-speech-sdk-rs/SpeechSDK/lib/x64 # or arm32, arm64
MacOS:
export DYLD_FALLBACK_FRAMEWORK_PATH=/Users/xxx/cognitive-services-speech-sdk-rs/SpeechSDK/macOS/sdk_output/MicrosoftCognitiveServicesSpeech.xcframework/macos-arm64_x86_64
Windows (pointing to SpeechSDK directly in target folder):
set PATH=%PATH%;"C:\Users\xxx\cognitive-services-speech-sdk-rs\target\debug\build\cognitive-services-speech-sdk-rs-b9c946c378fbb4f1\out\sdk_output\runtimes\win-x64\native"
We are supporting MacOS arm and aarch64 and x86_64 architectures.
Run following commands to build:
cargo build
Speech SDK libraries are linked dynamically during build and run. When running the application use following environment variable to point to custom library location:
export DYLD_FALLBACK_FRAMEWORK_PATH=/Users/xxx/cognitive-services-speech-sdk-rs/SpeechSDK/macOS/sdk_output/MicrosoftCognitiveServicesSpeech.xcframework/macos-arm64_x86_64
Then run your application utilizing cognitive-services-speech-sdk-rs or examples e.g.:
cargo run --example recognizer
See changelog