Real-time speech-to-text transcription and alignment with multi-language support, based on OpenAI's Whisper model. No python or any separated servers needed.
Demo video: Link
Free Demo project (exe): Link
This plugin allows you to recognize speech in 99 languages, just by adding one component to your blueprint, without relying on any separate servers or subscriptions.
Accuracy varies for each supported language. See the original paper for the accuracy of supported languages.
Prerequisite to use with GPU (CUDA)
To use this with a GPU, you need a supported NVIDIA GPU and to install the following versions of CUDA and cuDNN.
Number of Blueprints: 2
Number of C++ Classes: 13+
Network Replicated: No
Supported Development Platforms: Windows 64-bit
Supported Target Build Platforms: Windows 64-bit