To use Azure LUIS with voice do i need to get the text first? - azure

I think the title explains my doubt.
I've tried before the Speech to Text feature from Azure.
The question is:
Is there a way to use the sound binary to Azure LUIS instead of the text?

Yes, LUIS can accept speech input instead of text. LUIS provides this tutorial on how to setup speech services. The tutorial is in C#, however it appears their GitHub repo has samples in other languages, if of use.
Hope of help!

if you are creating a speech bot, here is a new approach.
so it leverages azure speech (TTS and SR), integrate with bot service in a more easy way
https://learn.microsoft.com/en-us/azure/bot-service/directline-speech-bot?view=azure-bot-service-4.0

Related

QnA maker And Dialogflow Knowledge Base

I have QnA Maker and Dialogflow knowledge base. I am trying to develop a faq bot. I need to know which is better to use the Dialogflow knowledge base or QnA Maker. Can some one tell me which is better?
Those two options are highly acceptable. As QnA is some easier method. While we need to accept in some cases like even QnA maker is created with Knowledge base. The best solution can be using Knowledge base to create FAQ bot.
Refer the links to create knowledge base and FAQ bot.
To create knowledge base, refer the below link
https://learn.microsoft.com/en-us/azure/cognitive-services/qnamaker/quickstarts/create-publish-knowledge-base
To create Bot, using the below link. But need to complete the procedure of link 1
https://learn.microsoft.com/en-us/azure/cognitive-services/qnamaker/tutorials/create-faq-bot-with-azure-bot-service

Is it possible to use Googles WaveNet Text-To-Speech model for the Actions-On-Google integration of a Dialogflow agent?

Google Clouds Text-To-Speech API has a WaveNet model whose output in my opinion sounds way better than the standard speech. This model can be used in Dialogflow agents (Settings > Speech > Text To Speech), which results in the generated speech being included in the DetectIntentResponse. However, I can find no way to use this speech with the Actions-On-Google integration, i.e. in an actual Google Assistant app. Have I overlooked this, or is this really not possible, and if so, does anyone know when they plan to enable this?
In the Actions console, going to the Invocation page lets you select a TTS voice.
All of the voices can be demoed on the Languages & Locales page of the docs, and the vast majority of them use WaveNet voices.

Is there a way to connect Custom Translator with Speech Translation?

I want to find the way how can I connect Speech Translation with Custom Translator.
On Custom Translator webpage there is mentioned that: "Custom Translator can be used for customizing text when using the Microsoft Translator Text API , and speech translation using the Microsoft Speech services."
Unfortunatelly, I didn't find any example of that usage.
Can anyone help with that?
Thank you!
If you build your application based on this example,
You can set the category id at the SpeechTranslationConfig as follows:
config->SetServiceProperty("category", "", ServicePropertyChannel.UriQueryParameter);
This is available since Microsoft Speech SDK 1.5. details

Can I use PubNub instead of DirectLine with BotBuilder?

Building a productivity chatbot. Hosting it on Heroku. Since I want to have very minimal dependency on Azure, I want to avoid directline APIs. Is it possible to use PubNub instead? Any insights / examples would be very helpful.
Building a Chatbot
You can build a chatbot with PubNub instead of DirectLine with BotBuilder. The following shows you how to build a chatbot using serverless techniques. The example also includes voice recognition using Google's voice API. So it's just a step up from what you're asking. You can include or exclude the voice recognition technology used.
Walkthrough: https://www.pubnub.com/blog/build-an-80s-chatbot-with-an-npm-package/
Try it live: https://stephenlb.github.io/artificial/
GitHub: https://github.com/stephenlb/artificial

how to integrate watson conversation with Speech-to-text and text-to-speech APIs

This is my first experience of using IBM watson and I am stuck with integrating watson conversation with speech-to-text and text-to-speech api services using node.js platform.
Done with conversation part but can't find a method to make
input speech ==> output of STT => input of conversation => output of conversation => input to TTS ==> output speech
I have tried multiple ways but still can't get even 1% of success. Followed multiple github repos even this one too with most forks https://github.com/watson-developer-cloud/node-sdk and multipletjbot recipes, etc still no results.
Can anyone here guide me with the right method?
the error with this link is attached below
does this help? I think this demo is similar to what you are doing. It is using STT, TTS, Conversation I believe.
https://github.com/watson-developer-cloud/speech-javascript-sdk/tree/master/examples
https://speech-dialog.mybluemix.net/
https://github.com/nfriedly/speech-dialog
There are some great examples that you can download, and play around with on the Watson Starter Kits Page.
Create a few of them and download the code, and then plunder what you need for your app or use one of the starter kits as the beginning of your app.
Starter kits on the page linked above that I think can help:
Watson Speech to Text Basic
Watson Assistant Basic
Watson Text to Speech Basic
Each of the starter kits listed above are available in Node and have README.md files to help you set everything up.

Resources