Google how can I create a music player for my Google Assistant - node.js

I’m wondering how I can create a music Player for my Google Assistant compatible devices (e.g. Google Home mini, my tablet, phone...). I’ve been researching about how I can do this, but I’ve just found things like using Dialogflow, node-js and/or Actions on Google using Google Firebase Cloud Functions. I’m new to all this, I was motivated by Spotify and Pandora and all those other services. So I also tried looking up how they do it, but I found nothing. If any of you Know how to do it, please help me.
In addition to all that, I am just a tad bit confused about the whole Dialogflow and Actions on Google integration, but that’s easier to fix than the overall question.
If this isn’t “solvable” is there a way to do it with Dialogflow Fulfillment’s?

In order to create something like Spotify or Pandora, you need to partner with Google to create a media action. These are different than the conversational actions that you can create using Actions on Google and Dialogflow.
If you want to create a conversational action with Actions on Google and Dialogflow that produce long-form audio results as part of the conversation, you will want to look into the Media response, which you can include in your replies.

Related

How to add google assistant or alexa to our own websites or apps using their sdk?

I just wanted to know if we can somehow integrate alexa or google assistant onto a website I made. I don't want to make completely new skills or apps. I just wanted to know if I type something, can I get a reply from google assistant or alexa and show on website.
I think Google has some google assistant sdk but then, it's written on python and even the node-js one is dependent on python environments
So is there any chance I can do this?
No.
The only way to initiate conversation with the smart speaker is by voice. No server side activation, sorry.

Triggering DialogFlow with Face detection

Does anyone know if there is a way to trigger DialogFlow from Face Detection API?
The DialogFlow conversation process is not very user friendly since you need to say :
"Ok Google, Talk to my app"
I've seen something about implicit invocations and deep links here:
https://blog.mirabeau.nl/nl/articles/creating_friendly_conversational_flows_using_google_deep_links/61fNoQEwS7WdUqRTMdo6J2
that provides a better approach
I'm trying to do something like this
https://www.forbes.com/sites/katiebaron/2018/06/07/ambient-tech-that-actually-works-hm-launches-a-voice-activated-mirror/#49b619634463
But with Google Assistant / Dialogflow / Vision API (Face detection)
Anyone has ideas how to do this in Google?
I am afraid that using face detection to trigger Google Assistant is not possible. Google requires you to use a trigger word such as "Ok, Google Talk to my app" when you build actions. This is done due to privacy for the user and makes sure that the app cannot be triggered without the user talking to the device.
Implicit invocations and deep links are shortcuts in your conversations, but they can only be used if you trigger the assistant first by saying "Okay Google..." Thanks for reading my blog by the way :)

custom built in intents with Google assistant SDK

I am trying to extend my voice technologies experience after working with Amazon Alexa.
Is it possible to use my Raspberry Pi based assistant handle the custom built-in intents like:
- Hey Google, play with the dog;
- Hey Google, make me a coffee;
- Hey Google, clean the room;
As I know the google has pretty the same way of deep linking of command and in the usual way user needs to say something like:
- Hey Google, talk to my concierge play with the dog;
- Hey Google, talk to my concierge make me a coffee;
- Hey Google, talk to my concierge clean the room;
Is there a way to make direct calls to some concrete action without calling its name using Actions SDK?
Yes, if you are using the Google Assistant SDK, you can register custom device actions which will allow you to get callbacks based on certain queries that you say, including the ability to pull parameters that may have been said.

Record audio using Google Actions

We are looking to build Google Action where it will record small snippets (like a voice TODO list) and can be played later.
Is there any documentation for this?
In short - no. Google does not provide access to the audio stream from the Assistant. You can get the Speech To Text (STT) processed by Google, however, using the Actions on Google API.

Possible to return an image in the Google Actions webhook response?

From the docs it seems like SpeechResponse is the only documented type of response you can return:
https://developers.google.com/actions/reference/conversation#SpeechResponse
Is it be possible to load an image or some other type of media in the assistant conversation via API.AI or the Actions SDK? Seems like this is supported with api.ai for FB, other messengers:
https://docs.api.ai/docs/rich-messages#image
Thanks!
As of today, Google Actions SDK supports Conversation Actions, by building a better Voice UI, which is integrated with Google Home.
Even API.AI integrations with Google Actions can be checked out here, which shows currently no support for images in the response.
When they provide integrations with Google Allo, then in the messaging interface, they might start supporting images, videos etc.
That feature seems to be present now. You can look it up in the docs at https://developers.google.com/actions/assistant/responses
Note: But images would be supported only on devices with a visual output. So Google Home would obviously not be able to do it. But the devices with screen do support a card with an image.
Pro Tip: Yes you can
What you want to do is represent your (image/video) as a URL within API.AI and render the URL as a (image/video) within your app
see working example

Resources