A2F-Inworld and Metahuman. How to stream audio from UE to A2F and get back the resulting blend_shapes ?

Well, it’s all in the title. I started researching the topic but the confusion is getting bigger. I know how to have A2F into Unreal via connectors to animate a Metahuman with pre-recorded audio but I’d love to achieve a runtime scene where the audio coming from Inworld Characters get sent into A2F (using the streaming audio player) and the blend_shapes come back into UE for lipsync and facial animations. A2F docs are terrible and all other sources I found are vague and/or incomplete.

Is anyone tinkering with something similar?
Thanks.