I was wondering if someone knows about a project that provides YOLO as API service. I need to use image detection with object coordinates but I do not want to develop API backend.
Thank you, David
Related
I am using Azure Face API to tell two different persons' faces.
It was easy to use thanks to the good documentation on the MicroSoft Azure API website.
But the different confidence rate between API call and the demo on the webstie: https://azure.microsoft.com/en-us/services/cognitive-services/face/#demo
My code is simple.
First I get the face ids of two uploaded images using face detection API.
And I just send two face ids to face verify API. Then I get the result of confidence rate that means the similarity of two faces.
I always get less confidence rate from my API call than the demo of the Azure website. About 20% less.
ex) I get 0.65123 on API call while I get the higher number like 0.85121 on the demo.
This is the Azure face API specifications to verity two faces:
https://learn.microsoft.com/en-us/rest/api/cognitiveservices/face/face/verifyfacetoface
Since I got no clue why it happens. I don't resize or crop the images on uploading.
I use the exactly same images for this test.
Is it possible for MS Azure to manipulate the values for their own interests?
I wonder if anyone has the same issue? If yes, please share your experience with me.
Different 'detectionModel' values can be provided. To use and compare different detection models, please refer to How to specify a detection model.
'detection_02': Detection model released in 2019 May with improved accuracy compared to detection_01. When you use the Face - Detect API, you can assign the model version with the detectionModel parameter. The available values are:
detection_01
detection_02
I want to show image frames generated from OpenCV on react, I am using Django as server-side. I get a single frame from each request when I send requests from react to Django, so it constitutes a large number of requests at a time.
For now, I am just sending a request to API again and again which I think is not a proper way. I need resources like a tutorial , articles to learn the best way to achieve this thing.
I have seen like WebSockets will be good but I researched a lot but can't found a proper resource.
Can someone suggest to me or can share a good resource to learn?
I don't have much knowledge of react. But if I were to do the same task on a webpage then I would read images/frames from webcam using javascript.
Here is an example of it:
Read image from webcam using javascript.
After retrieving the image, you could send frames from js to Django and perform the OpenCV task there.
We have a web application which loads a 3D model(Build in MAYA) in the browser using Three.js. I need the following things to be added,
Text To Speech using Amazon Polly.
Get the Speech Marker from Amazon Polly and then the 3D model needs to use the Speech Marks and Whispering features that are available in Amazon Polly.
I have googled regarding it and found https://aws.amazon.com/sumerian/features/.
In Amazon sumerian we have a feature called Sumerian Hosts, but it has predefined models.
So my question is it possible to integrate TTS with speech marks using Amazon Polly for 3D model build in MAYA using Three.js for web application build in nodejs.
In Amazon Sumerian you can import .obj files, .mtl files, meshes, materials, and textures using a single drag and drop. Also, Sumerian can be integrated with Amazon services like Lex, Polly etc.
To have a brief idea about the functionality that can be done using Amazon Sumerian please go through https://docs.sumerian.amazonaws.com/
I guess what you are looking for could be achieved in some other way also without using Sumerian but you can try once by importing your model in Sumerian and see whether you have the functionality of controlling the model like you wanted or not.
Do let me know if you are able to integrate Sumerian with you web app.
Hi I am very interested in machine learning.
familiar with keras and tensorflow and know how to train model and save it.
I want to build node.js(loopback) web application that provides machine learning API.
I'm quite new to web application and don't know how to utilize python code in the node app.
To be specific, I'm trying to make deep style application like this site for practice purpose.
For codes that is open to public in github. most starred implementation is in torch and
I am currently looking (TF).
It has script work as "main.py" that consists of
1) loading pretrained VGG model(about 500Mb) to memory.
2) and using each framework's function, it create output photo inside folder
So I want to run this with GPU instance in AWS
problem is, I don't know how to call "main.py" inside javascript.
Q1 How did many machine learning based services achieved this??
did they implement their server in python like django to load trained model and inference from it?
But I want to integrate in node.js especially loopback. how can I achieve this in a most efficient and common way??
Q2 Is calling "main.py" efficient whenever there's request to server?
I mean, If the script is called everytime request is made, It loads 500Mb object to memory all the time and that's so inefficient.
and as far as I know AWS GPU instance can utilize 4 GPU at a time. I wonder how can I utilize all of it with most efficient and common way...
Thank you for reading!
I'm too frustrated with myself... I really want to know method deploy my model to server
and thank you in advance
You can communicate between NodeJS and e.g. Django with a REST interface. Deploy a Django server with the REST framework and call an url from NodeJS. This way, you can trigger the python code. You can also let the python code call your NodeJS with another REST interface on the javascript side if that is needed.
I was wondering is there an API or some material on how to implement handwriting recognition to webapp ? As in like you write with stylus and it stores it as font.
It would be awesome if someone has good reading materials on developing ocd webapps.
Thanks in advance.
I think https://code.google.com/p/tesseract-ocr/ is the best open source tool, is it's a good starting point.
MyScript Web Components leverage Polymer project components and MyScriptJS, an open source library that speeds up the development of handwriting interfaces for web applications. MyScriptJS supports common tasks involved in implementing the handwriting user interface, including ink capture, rendering, and requests to the MyScript Cloud recognition service. MyScriptJS also manages API calls, using HTTP/REST or WebSocket protocols. Check it out on http://myscriptwebcomponents.github.io/