I want to try create simple bot for game(game has own client), which will be control character(may be mouse move and keys press) and read information from screen. Advice please ways/packets for node.js.
Any information will be useful
robotjs is a npm module that provides all of the functionality you described in the question. The documentation on its website goes over individual walkthroughs on how to handle mouse, keyboard, and screen reading functionality.
Related
Is there a way to register an interaction within a Adobe XD document?
For example if a user presses a button, can we send a response to an external server notifying that that event happened?
Reading the Adobe XD plugin documents it seems that most API's are used for read only.
Any help appreciated and maybe pointers on what's needed (websockets etc)
Yes you can the plugin API gives you several methods for interacting with objects in the scene and interface elements.
Read the docs for more info
i'm new in the chat bot programming . I would like to do exactly the command "ok google, take a picture", that the android open the camera and in 3 second take the pic. Dialogflow is a service from google, so I thinking that there are some library with some example of this, or if not, how I need to search to put this command in my action ?
PS: I'm making a location and opinion action that receive from the user the place and the opinion about the place, so i would like to ask if the user want to take a pic from the place using this, but a don't know how a search this!
Unfortunately, there is currently no library and no direct support for this type of thing. The Assistant does not give Action developers access to the camera. In fact, most of the work your Action does is on a cloud-based server, not on the device itself.
You can, in some cases, use something like the Android Link helper, but this requires the user to have installed your app on their phone, and doesn't quite do what it sounds like you want.
I am trying to write a simple chrome app to play a sequence of online pictures on my chromecast device.
I have looked at some examples, but could't find anything which I could tweak around to get the simple behavior i needed. Maybe someone here could help, by providing directions or advise on getting started with developing something like that for chromecast.
UPDATE:
To give you a better idea, about the specifics, let me add some more details to my requirements.
It needs to be controlled from chrome
I want to pass a playlist with 10s-100s of images so it can slide them in circles.
After receiving playlist chromecast device should be able to continue on its own, without continuously asking for next image.
This is actually similar to backdrop feature Google is planning to introduce, but I wanted to write something myself.
Thanks
If you don't want to develop your own Cast receiver, then you can use the media namespace channel and the Styled Media Receiver to display a photo at a time:
https://developers.google.com/cast/docs/styled_receiver
You will have to add the logic to advance from photo to photo in your sender app.
If you are willing to develop your own custom receiver, then you can start with this Cast sample app:
https://github.com/googlecast/CastHelloText-android
It allows you to send messages to a custom receiver. You can use that to send the URLs of the photos and then you can add JavaScript logic in the receiver to play a slideshow.
Just to let you know, I have tried various options and ended up writing custom receiver and Chrome sender applications. This was really straightforward and exactly what I wanted.
See the links above for guidance and also examples here.
I am using the Cast Companion Library and most of it is working well however I have not been able to figure out how to show the device chooser programmatically.
Basically when a user chooses a video on my app I would like to show the chooser if they haven't yet connected.
Thanks.
If you are using the MediaRouteButton and it's accessible:
mediaRouteButton.performClick();
you need to "Developing a custom UI with the MediaRouter API’s and MediaRouter.Callback" but I didn't find any example how to do it .
First, I suggest you do not do that. I don't know what you are planning to do in your app so I can't say much but don't do something custom in terms of user interaction/flow that is different from other apps and the apps offered by the framework; people will not expect that.
On the technical side, if you want to open the standard dialog, I am not aware of any trick but I haven't looked at the MediaRouter code in v7 support library for that; since that is open source, you are welcome to look in there and see if there is any simple way to do that. If you want to design your custom one, you should be able to do that; the CastCompanionLibrary provides a sample of the Callback dialog and the one for chooser follows a similar pattern.
As a side note, you mentioned you want to open the chooser when user selects a movie. So how can they select to play locally? Are you planning to add additional selection to that dialog for local playback?
Is there such a thing as a Spotify remote control API? I want to write an app which lets the user control the Spotify desktop client (play, pause, volume, playlists, etc.) but I can't seem to find any documentation anywhere.
There are iOS and Android remote control apps out there so I'm just wondering how they do it if there isn't an API.
On Mac OS X, Spotify can be controlled 'remotely' (i.e. from other apps) using AppleScript, either using the bundled API or the System Events service.
Here's an example of using the bundled API: https://github.com/activars/spotify-applescript/blob/master/spotify_control.applescript
And another one using System Events (it isn't quite as elegant): http://www.jacktams.net/2010/04/28/spotify-applescripts-version-0-4-3/
Official documentation is a bit thin, although you should be able to use the AppleScript editor's library feature to find out more about the methods offered.