Is there a way to register an interaction within a Adobe XD document?
For example if a user presses a button, can we send a response to an external server notifying that that event happened?
Reading the Adobe XD plugin documents it seems that most API's are used for read only.
Any help appreciated and maybe pointers on what's needed (websockets etc)
Yes you can the plugin API gives you several methods for interacting with objects in the scene and interface elements.
Read the docs for more info
Related
I want to try create simple bot for game(game has own client), which will be control character(may be mouse move and keys press) and read information from screen. Advice please ways/packets for node.js.
Any information will be useful
robotjs is a npm module that provides all of the functionality you described in the question. The documentation on its website goes over individual walkthroughs on how to handle mouse, keyboard, and screen reading functionality.
I am writing a Chrome extension and I would like to be able to get the current Gmail message ID (in either conversation view or not). I tried using the Gmail.js library but as mentioned here the observers quit firing after a while:
https://github.com/KartikTalwar/gmail.js/issues/304?_pjax=%23js-repo-pjax-container
So is there a different solution? I have searched around and do not see a clean, reliable solution.
Thanks for any tips or pointers.
Due to some reasons, I'm not allowed to touch Domino server but would like to make my plugin be able to modify email content right before it's being sent in client side. All comes up in my mind is to add a listener or somehow to hook the 'send' button of a message, which holds on when being clicked and runs my codes, and then sends the message. Is there anyone knowing about this trick, or any other way to meet the same purpose? Thanks.
BTW: I'm using Lotus Toolkit in Java.
Update:
I found a very similar question here:Lotus Notes - Java UI - Eclipse plugin
But I found Querysend only exists in NotesUIDocument of LotusScript but not Java. Strange.
You have two choices: install a DLL built using the Notes C API on every Notes client computer, or make modifications to the mail template. In the latter case, though, rather than implementing your code in the QuerySend event it probably makes more sense to implement to implement your code directly in the click event of the shared Send action - and that is just Lotus' formula language, so neither LotusScript nor Java is needed.
I am trying to write a simple chrome app to play a sequence of online pictures on my chromecast device.
I have looked at some examples, but could't find anything which I could tweak around to get the simple behavior i needed. Maybe someone here could help, by providing directions or advise on getting started with developing something like that for chromecast.
UPDATE:
To give you a better idea, about the specifics, let me add some more details to my requirements.
It needs to be controlled from chrome
I want to pass a playlist with 10s-100s of images so it can slide them in circles.
After receiving playlist chromecast device should be able to continue on its own, without continuously asking for next image.
This is actually similar to backdrop feature Google is planning to introduce, but I wanted to write something myself.
Thanks
If you don't want to develop your own Cast receiver, then you can use the media namespace channel and the Styled Media Receiver to display a photo at a time:
https://developers.google.com/cast/docs/styled_receiver
You will have to add the logic to advance from photo to photo in your sender app.
If you are willing to develop your own custom receiver, then you can start with this Cast sample app:
https://github.com/googlecast/CastHelloText-android
It allows you to send messages to a custom receiver. You can use that to send the URLs of the photos and then you can add JavaScript logic in the receiver to play a slideshow.
Just to let you know, I have tried various options and ended up writing custom receiver and Chrome sender applications. This was really straightforward and exactly what I wanted.
See the links above for guidance and also examples here.
What's the simplest way to setup a chrome extension to record audio from the microphone?
I see there is a working experimental speech input API but how come you don't have access to the recorded file? Seems like hooking up into that should be simple enough, as it's a step earlier in the process, no? Especially as there is also a text-to-speech API, so you could effectively record into text and then have the computer speak it back out, but unless you want a standard voice, how lame, redundant and prone to error is that?
Then there seem to be flash solutions like this but how can I use that in a chrome extension without having to setup anything server-side? (since I don't actually need to send anything to a server--it's all local and client-side)
Is NPAPI a possibility? Is there such a plugin ready-made?
Don't know of other possible alternatives (HTML5 isn't ready yet, it seems) but I welcome anything functional and simple to implement and hook into a chrome extension.
Finally a native solution appeared: Introducing getUserMedia
You cannot use the speech input API, since it will record only the microphone. Okay, you can grab the speakers like that, but it's clearly not the solution.
Using a NPAPI plugin is a solution. You'll can identify the sound made by a particular tab and after record that source, but it is no longer web dev.