Azure Form Recognizer pricing - azure

"I've got some questions about the Azure Form Recognizer pricing. The website says $50 per 1000 pages for the S0 Web custom document type. If I upload a 1000 page PDF but make 10 API calls on each page, will we be charged per API call to the uploaded document or based purely on the number of pages? So would we expect a $50 bill or a $500 bill for those 1000 pages?"

Well, I had this same doubt and drilled down a bit more to figure out that it is based on the number of pages you analyze using the API. So in your case, since, you are calling the API which is essentially analyzing the pages of the document, the bill would be $500.
However, it is always better to save the results and refer them later as with one model the result will always be the same for the same document/page.

Related

Is there a way to get earnings from paypal sdk python?

So I have my admin page for my website and I use Paypal to process payments and I was wondering if there is a way to get a report with the total earning. I read the documentation but I only found generating transaction reports.
In case I didn't make my self clear what I mean is I request the API and it returns a float or int with the total earnings so I can display the amount on my admin page.
You might be able to calculate a running total minus fees by setting up and listening for transaction Webhooks, specifically PAYMENT.CAPTURE.COMPLETED
I would just use the Reports features in PayPal.com rather than obsessing about building your own counter.

How can i get via AdSense Api the payment per Month and per Site

1) Hey i run API AdSense and im trying to get the total payment per month but filtered by site as well.
At the moment i can only extrect the amount of payment per month via https://www.googleapis.com/adsense/v1.4/accounts/accountId/payments but there is no way to filter there by Site
I can filter by site in this API https://www.googleapis.com/adsense/v1.4/accounts/accountId/reports but the payment that i get there are not accurate.
2) How can i get from the API the "InvalidTraffic" parameter per site ?
Thanks a lot in advance
1) Reporting is the correct way to get this information. Common causes of metric discrepancies are using the wrong time zone and using the wrong currency type. Also, be aware that "earnings up to yesterday are accurate, [but] more recent earnings are estimated due to the possibility of spam, or exchange rate fluctuations."
2) Invalid Traffic isn't exposed in the API.

Web scraping data which refreshes every minute

I am trying to extract some data from this website which refreshes every minute. I have tried researching about web scraping and tried chrome extensions but none seem to work for me.
Some background information about the website: it is a website where people go to monitor bid prices for COE (certificate of entitlement for cars in Singapore). Every alternate Wednesday, from 1430 to 1600, I would have to manually copy and paste the data into an Excel spreadsheet before it refreshes every minute.
Details for COE
I have attached screenshots to illustrate further.
This is the website to scrape; https://www.onemotoring.com.sg/1m/coe/coeDetail.html
You can get a very low cost with AWS Lambda with node-js.
Create a Lambda function and trigger it at your cron schedule you want to crawl the website. You can use library like
https://github.com/bda-research/node-crawler
to simplify crawling.
Also,
To get the exact nodes in the page use serverside jquery or any progressive script that can extract elements from the crawled page.
Once you have the details, you can store them under DynamoDB which is a nosql with very low latency.
You can use,
ODM like https://github.com/clarkie/dynogels to access DynamoDB with very less code.
Hope it helps.

Is it possible to use Instagram API in order to make a user list according to the number of followers?

Instagram API permission is strict now and I want to know whether the tool my company will make can get the permission from Instagram or not, because it may not suit with the 3 valid use case.
We are an offshore company which develops application.
One clients ( PR or Advertisement company) asked us to make an app which finds influencers(the user who has many followers) on Instagram so that the clients can ask the influencer to advertise their products.
We want to make a influencer searching tool. Lets, say if you search with #Chocolate, the list of users who have posted photos with #Chocolate will show up, and we want to sort out the list according to the number of followers.
Referring to the valid use case for Instagram, which case may be suitable for the app we want to make?
Also, by one-off project, what does it mean? We are an offshore company and get orders from a client which want us to make a tool for them. If we just make it, and will not sell or use the application again, would it be "one-off" project?
In order to avoid one-off project, do we need to keep selling or offering the app ?
There is no direct API call to make this. You have to make a series of API calls to get all the posts for hashtag, this will have user_id and name. Then for each user you have to make another API call to get followers count. And then you have sort the results and make query for >3000
For example #chocolate has 1 million posts, so to get all 1 million posts you have make about 50000 API calls (u get 20 post per API call), and then you have to make 1 million user_info API calls to get follower counts, so 1050000 API calls total and you have 5000/hr API calls limit. So to complete this operation it will take you 9 days.
This will be a one off project if u do for a client use, you have to make a platform or service for this feature and have your client (and others) login and use the service.
In any case, the project is unrealistic to implement due to the limitation of APIs

Why does Google Analytics show less visits than One&One stats?

Comparing google analytics results to one&one hosting monthly statics shows a huge discrepancy.
For last month:
Google shows 1046 visits.
One&one stats show 15304 unique visits.
The google code is in the footer which appears on every page.
I'm aware ga only works with js enabled but to assume that many non js users???
Google Analytics is a good indicator of how many humans are visiting your website.
Here are some things to check:
how many bots are in your monthly stats? You can usually find something that says User-Agent in your stats page. GoogleBot, Slurp, msnbot & others will be visiting every page on your site.
that you've read Google Analytics' definition of a visit.
that you have read what your statistics provider means by unique visit. Does that mean unique visitor, page view or something else?
Raw hits on servers can be misleading for a number of reasons..
If you have external style sheets & JavaScript etc, they could be counted as a hit in the webserver log
RSS feed readers will periodically update without being asked to by a human
Check the page views in Google Analytics - it's possible that 1&1 is tracking unique page views instead of the actual visits.
Google Analytics works for almost all users (I believe less than 5% have JS disabled). I have had the same discrepancy, in my case the difference was zeroed out when I took into account the bots (which server-side statistics often take into account, as they produce http-requests). You probably have the same "problem".
Neither stats are wrong, they just count different things. Google Analytics is the more "accurate", i.e. the numbers you want to take a look at. The hosting stats, which look only at http requests, often without filtering, are less interesting.
Blogger, and probably other sites, serve a different page template or skin to mobile visitors. In my case, that template didn't contain the google analytics snippet of code and so those hits were uncounted, until I noticed and fixed it.

Resources