[FREE] AzSpeech plugin: Text-to-Speech, Speech-to-Text and more with Microsoft Azure Cognitive Services

Hey,
Great plugin, love it.

I have a few problems though:

  1. when I use the text to speech function it only gets the second part of the sentence as result. E.g. if I say “lorem ipsum one two three four five” I get the string as “three four five”.

  2. the same problem occurs if I use the button to speak (when I press the button it calls the text to speech function)

  3. I decided to work around the truncated sentence problem by recording the audio using the audio component to a wav file. And then using the wav file to text function. Unfortunately in this case the wav file to text function does not work at all, i.e. I do not get any result.

The logic is as follows:
KeyPressed: Audio Capture Enable → Start Recording Output
KeyReleased: Audio Capture Disable → Finish Recording Output (export type wav file) → delay (1sec) → Wav file to text

To make sure the wav file was saved correctly, I changed the wav file to text function to convert wav file to usoundwave and then plugged its output into play sound 2d and the wav file was read correctly. Unfortunately the wav file to text function does not work for me at all.

Logs:
UE4.27
AzSpeech v1.4.3

PIE View:

Task: WavFileToText (155787)
ActivationTime: 6milliseconds
ActiveTime (102 second) -> Never stops
Current recognised string:

OutputLog:

LogAzSpeech: Display: Task: WavFileToText (155787); Function: Activate; Message: Activating task
LogAzSpeech_Internal: Display: Task: WavFileToText (155787); Function: StartAzureTaskWork; Message: Starting Azure SDK task
LogAzSpeech_Internal: Display: Thread: AzSpeech_WavFileToText_155787; Function: Init; Message: Initializing runnable thread
LogAzSpeech_Internal: Display: Thread: AzSpeech_WavFileToText_155787; Function: CanInitializeTask; Message: Checking if can initialize task in current context
LogAzSpeech_Internal: Display: Thread: AzSpeech_WavFileToText_155787; Function: Run; Message: Running runnable thread work
LogAzSpeech_Internal: Display: Thread: AzSpeech_WavFileToText_155787; Function: InitializeAzureObject; Message: Initializing Azure Object
LogAzSpeech_Internal: Display: Thread: AzSpeech_WavFileToText_155787; Function: InitializeAzureObject; Message: Creating recognizer object
LogAzSpeech_Internal: Display: Thread: AzSpeech_WavFileToText_155787; Function: CreateSpeechConfig; Message: Creating Azure SDK speech config
LogAzSpeech_Internal: Display: Thread: AzSpeech_WavFileToText_155787; Function: ApplySDKSettings; Message: Applying Azure SDK Settings
LogAzSpeech_Internal: Display: Thread: AzSpeech_WavFileToText_155787; Function: EnableLogInConfiguration; Message: Enabling Azure SDK log
LogAzSpeech_Internal: Display: Thread: AzSpeech_WavFileToText_155787; Function: ApplySDKSettings; Message: Using language: en-PL
LogAzSpeech_Internal: Display: Thread: AzSpeech_WavFileToText_155787; Function: Run; Message: Starting recognition
1 Like