I am using SKMaps v3.0 in my android application and drawing route using the below method.
SKRouteManager.getInstance().calculateRouteWithPoints(pointsList, route);
Everything is working fine, including navigation and voice suggestions. Now my task is to play my own audio file when user is reaches a particular lat, lng position. Is there a call back where I can identify user has reached that point and play my audio file there?
Thanks in Advance.
There is no default notification mechanism for "point along the route reached" (unless that point is a via point) but you can implement this functionality using the POI Tracking mechanism: just define which are the points you should be watching out for and start POI Tracking.
You can read more about this here and you can see it in action in the demo project, the POI Tracking area.
Related
I want to develop an invitation card maker app in flutter so the issue is that I don't know what kind of data I will need to import from the backend to make my cards editable. i have designed cards in photoshop but I don't know how to make them editable in a mobile app. if anyone has a suggestion please give me your suggestions
I'm glad to be able to help you.
In those cases you do the following:
Receive image to edit.
Use external library made to edit photos (https://pub.dev/packages/image_editor_pro)
Send edited image to the backend and replace the old one.
Hi all I am very happy that I finally found the solution.
So we have to first import the background image from API and load text data on it then we can use flutter packages to edit those texts.
But there is one thing you need to do before you import designs to the mobile app. You will need to have a perfect pixel size of font and background otherwise it will be overlapped.
I have images with dimensions of 3000*5000 so I have used aspect ratio for the responsiveness in every device.
And I have used Figma to design cards so we can get all CSS very fast for every line of text and then I am converting it to JSON using CSS to js converter and js to JSON converters.
I am making a game where I want to command the AI using word i speak.
Say for example I can say go and AI bot goes to certain distance.
Question is I am finding asset and no provider is giving me grantee that it is possible ?
What are the difficulties for doing it?
I am programmer so if some one suggest the way to handle it I can do it.
Should I make mic listener on all the time and read audio and then pass audio to some external sdk which can convert my voice to text ?
these are the asset provider i have contacted.
https://www.assetstore.unity3d.com/en/#!/content/73036
https://www.assetstore.unity3d.com/en/#!/content/45168
https://www.assetstore.unity3d.com/en/#!/content/47520
and few more !
If someone just explains the steps I need to follow then I can try it for sure.
I am currently using this external api for pretty much the same thing: https://api.ai/
It comes with a unity SDK that works quite well:
https://github.com/api-ai/api-ai-unity-sample#apiai-unity-plugin
You have to connect a audio source to the sdk, and tell it to start listening. It will then convert your voice audio to text, and even detect pre-selected intentions from your voice audio / text.
You can find all steps on how to integrate the unity plugin in the api.ai Unity SDK documentation on github.
EDIT: It's free too btw :)
If you want to recognize offline without sending data to the server, you need to try this plugin:
https://github.com/dimixar/unity3DPocketSphinx-android-lib
It uses open source speech recognition engine CMUSphinx
In Blogger I am trying to recreate the audio feature used in this web site http://www.talkenglish.com/lessondetails.aspx?ALID=2001 where you click on text and hear the audio (an mp3 file) instantly. When I embed identical code into a Blogger page and click it, it wants to play the mp3 file by kick starting Windows Media Player. I don't want a solution like SoundCloud that creates visible player controls. Can anyone suggest a solution. I am hoping to use it to create an online amateur speech therapy package to support some voluntary work I do in this area.
Thanks
That site is using a file called audioplayer.js. To be honest, I am not exactly sure if that's custom built or not. You can look at the source for that here. You have a bunch of options to get the functionality you're looking for. A couple are below.
You can use SoundManager2 which is a very robust JS Sound plugin.
You can also use HTML5's built in <audio> element. You can read more about that here
I am trying to write a simple chrome app to play a sequence of online pictures on my chromecast device.
I have looked at some examples, but could't find anything which I could tweak around to get the simple behavior i needed. Maybe someone here could help, by providing directions or advise on getting started with developing something like that for chromecast.
UPDATE:
To give you a better idea, about the specifics, let me add some more details to my requirements.
It needs to be controlled from chrome
I want to pass a playlist with 10s-100s of images so it can slide them in circles.
After receiving playlist chromecast device should be able to continue on its own, without continuously asking for next image.
This is actually similar to backdrop feature Google is planning to introduce, but I wanted to write something myself.
Thanks
If you don't want to develop your own Cast receiver, then you can use the media namespace channel and the Styled Media Receiver to display a photo at a time:
https://developers.google.com/cast/docs/styled_receiver
You will have to add the logic to advance from photo to photo in your sender app.
If you are willing to develop your own custom receiver, then you can start with this Cast sample app:
https://github.com/googlecast/CastHelloText-android
It allows you to send messages to a custom receiver. You can use that to send the URLs of the photos and then you can add JavaScript logic in the receiver to play a slideshow.
Just to let you know, I have tried various options and ended up writing custom receiver and Chrome sender applications. This was really straightforward and exactly what I wanted.
See the links above for guidance and also examples here.
Statement: I have a 3D model ready in Google SketchUp Pro.
Problem Statement:
I need to display the model in a webpage.
I need the model to be browsable (The user can navigate his way through it like on Google Earth)
I need the model to be update-able (If I update something in SketchUp, I should be able to easily port changes till the webpage)
Question: How can this be done?
Note: I figured out VRML could be an option, but I was unable to get that idea up and running.
Have you investigated the Sketchup Web Exporter plugin (still in beta at time of writing)?
If that doesn't do exactly what you're after, you can write a plugin yourself using the Sketchup API, allowing you to write plugins in Ruby or C++.
You can now use SketchUp in the browser, and it should be able to read anything coming from SketchUp Pro. Check out https://my.sketchup.com