Google maps API Map error: go.co/staticmaperror - node.js

everyone, I am trying to analyse the tweets about cats from Australia and it is all going well but some people are putting their location as: "NoWhere", "Mars" and other places, so I have been using google maps to confirm that the places are real and after many hours it worked perfectly, but when I analyze more than 50 tweets, the API gives me an image with API error and I can't find the problem
one of the things that people suggested that my link is over 2000 characters long but when I put the request in online tools, all of them say that my link is only 883 characters long. Any advice will be greatly appreciated:
Here are the locations I have been trying to make a map for:
(http://i.imgur.com/ckh6hcN.png)
and here is the map I get:
(http://i.imgur.com/Bl71jFC.png)
and here is the request link
https:// maps.googleapis.com/maps/api/staticmap?center=Australia&zoom=4&size=500x400&markers=Scarborough%20WA%206019%2C%20Australia&markers=Wandering%20WA%206308%2C%20Australia&markers=Australia&markers=West%20Coast%20TAS%207321%2C%20Australia&markers=Perth%20WA%2C%20Australia&markers=Baldivis%20WA%206171%2C%20Australia&markers=South%20Australia%2C%20Australia&markers=Alice%20Springs%20NT%200870%2C%20Australia&markers=a%2F11%20Talavera%20Rd%2C%20Macquarie%20Park%20NSW%202113%2C%20Australia&markers=Western%20Australia%2C%20Australia&markers=Boddington%20WA%206390%2C%20Australia&markers=Bunbury%20WA%206230%2C%20Australia&markers=Waroona%20WA%206215%2C%20Australia&markers=Melbourne%20VIC%2C%20Australia&markers=Geraldton%20WA%206530%2C%20Australia&markers=Osborne%20Park%20WA%206017%2C%20Australia&markers=Mandurah%20WA%2C%20Australia&key=[key]
These addresses are not invalid because I filtered them with google maps geocode to validate them.
note: I have separated the https in the request link because StackOverflow doesn't allow more than 2 links.

After talking to Google in twitter they informed me that there is nothing wrong in my link, is just that we are not allowed to request more than 15 markers and I am requesting 17.
Hope this helped someone :)

Related

Bing search API: specify filetype and site/domain?

Having had the luxury of having my fun and earning my bread well away from MS products for years, I am today trying to programmatically search with Bing Azure (wearing gloves) basically because I thought getting a google api was complex. So I headed down Data Market and issued this (let's say with perl's LWP which has been used to pass credentials):
https://api.datamarket.azure.com/Bing/Search/v1/Composite?%24skip=0&%24top=10&%24format=json&Sources=%27web%27&Query=%27abc%27
which works.
What I am now trying to find out is
1) how to tell the Bing search api to restrict results to a specific domain (e.g. ".org" or even a single website "www.wikipedia.org").
2) how to tell the search engine to restrict results to a specific filetype, e.g. 'PDF', 'XML' (or PDF and XML if that's possible)
3) if there is a simple list of the features/keywords in the GET request of the latest bing search API. Please no MS links if you please - i am really tired.
I have seen "site:.org" working on the bing search website when doing a manual search. And read about "filetype:pdf" working too.
Any hints?
bliako
cracked it:
... Query=%27abc site:.com filetype:pdf%27
at the point when m$ realises it costs to be clumsy
bliako

XPath Data Scraping From Online Community

I recently read this article on how to scrape the Inbound.org community members profile using Excel. And you can watch the video here if you prefer it that way.
Since the release of this tutorial, the Inbound website structure has changed a bit, as you can see at minute 11:00 in the video, if you attempt to copy the XPath of the social media icons it appears slightly different and because of this I haven't been able to extract that information.
Here's what I get now:
/html/body/div[3]/div/div/div[1]/div/div[2]/a[1]/i
This is how I wrote the syntax in Excel:
=XPathOnUrl(A2,"//a[#class='twitter']","href")
And then like this:
=XPathOnUrl(A2,"//a[contains(#class,twitter)]/#href")
Although I tried in many different ways, none of them showed me the link to the member's social media profile.
I even tried changing the xpath in multiple ways to get different data from the page, but none of it was the social media information:
=XPathOnUrl(A2,"//*[contains(#class,member-banner-tagline)]/div[2]/div/div/div[1]/div/div[1]")
=XPathOnUrl(A2,"//*[contains(#class,member-banner-tagline)]/div[2]/div/div/div[1]/div/h1")
I honestly don't know what to try anymore, something's wrong and I can't figure it out. Anybody have enough experience with this or can pinpoint the problem here with my syntax?
Thanks a lot
The first formula you tried looks fine, but this is the one that works for me (SEO Tools version 4.3.4) :
=Dump(XPathOnUrl(A2;"//a[#class='twitter']";"href";HttpSettings(TRUE)))

foursquare website venue search not matching up to api search

The answer just might be alluding me, but this bugs me
I've been doing a simple search trying to find all the outdoor picnic spots near me. So I search for picnic in my zip code, and I get a list of results.
https://foursquare.com/search?tab=tipResults&q=picnic&lat=&lng=&near=10001
I then try to do the same search using the api (the venues/explore endpoint) to see if my app works well, and I get a completely different list of places. The api url that I've been using is
https://api.foursquare.com/v2/venues/explore?near=10001&section=outdoors&query=picnic&radius=5000&client_id=XXXX&client_secret=YYYY
BTW, If I dont include a radius, it only returns 1 entry.
The web address you provided: https://foursquare.com/search?tab=tipResults&q=picnic&lat=&lng=&near=10001
Is tip results around 10001.
The API call you provided: https://api.foursquare.com/v2/venues/explore?near=10001&section=outdoors&query=picnic&radius=5000&client_id=XXXX&client_secret=YYYY
Is a single explore search api call.
These two are very different, and will yield different results
Try running a tip search via the API as such (API docs at this place):
https://api.foursquare.com/v2/tips/search?near=10001&query=picnic
(direct link for testing it)
Let us know if this this clears it up.

Google Search by Image API?

for my job, I'm looking into an idea in which people would use Google Search by Image and use any celebrity photo they find. Google would return the results and then on our end, a there'd be a database of professionals showing how to get that specific look.
I'm assuming this is extremely unlikely to do, based on that users could use ANY photo.
So, is there a way that I could have about 100 or so celebrity photos that Google Image results could compare to and then choose the one that is closest.
Basically:
Drag drop photo of Britney Spears
Google searches with that image
Google's results compare the top images with our 100, and selects the closest match.
User gets to see video of how to get Britney Spears look.
I'm not a programmer, but looking for some API or Search by Image extension that could make this remotely possible for the programmers here at my job. Does something like that (a search by image api) exist? The best I could find was just the support page, which is hardly of any help: http://support.google.com/images/bin/answer.py?hl=en&p=searchbyimagepage&answer=1325808
You can easily search by an existing image by inserting this into your address bar:
https://www.google.com/searchbyimage?site=search&sa=X&image_url=YOUR_IMAGE_URL
Example:
https://www.google.com/searchbyimage?site=search&sa=X&image_url=http://cdn.sstatic.net/Sites/stackoverflow/company/img/logos/so/so-icon.png
Sorry to say, but the Google image API is deprecated:
Important: The Google Image Search API has been officially deprecated as of May 26, 2011. It will continue to work as per our deprecation policy, but the number of requests you may make per day may be limited.
Quite sure there are some alternatives (http://www.tineye.com/ and http://mrisa.mage.me.uk)
Update (2013): There is now Google Custom Search which allows image searches.
These answers are quite obsolete, but the question comes up in searches. So, the Google Vision API has the "web detection" feature that does a reverse image search. First 1000 requests per month are free, $3.50/1000 afterwards.
I think Google Web Detection could be a solution for you. Google moved it permanently from Image search
You can do it via www.images.google.com but only from a browser (lets you upload your own image and compares it to similar).
I'm working on doing it from code (not from browser).
I had the same problem and came up with two solutions:
There are a number of APIs that give reverse image search results nowadays. The ones I used are https://reverseimageapi.com and TinEye.com.
As the selected answer mentions, you can easily scrape this information but will almost certainly need rotating proxies to prevent being banned by the search engine. There are plenty of proxy rotation services (Zyte, Oxylabs, ScrapingBee, etc.) to make you life easier.
I ended up going with option 1 due to the upkeep of scraping search engines and elements changing / breaking.

How to get a description of a URL

I have a list of URLs and am trying to collect their "descriptions." By description I mean what comes up, for example, if you Googled the link. For example, http://stackoverflow.com">Google: http://stackoverflow.com shows the description as
A language-independent collaboratively
edited question and answer site for
programmers. Questions and answers
displayed by user votes and tags.
This the data I'm trying to accumulate for the URLs I have.
I tried parsing the URL's meta-descriptions, however most of them are lacking a meta-description (yet Google and other search engines manage to get a description somehow).
Any ideas? Should I just "google" each link and scrape the data? I have a feeling Google wouldn't like this...
Thanks guys.
Different search engines have different algorithms to get the description out of the page if/when they are lacking the description meta tag. Some ignore the tag even it it's there.
If you want the description Google has, the most accurate way to get it would be to scrape it. Otherwise, you could write your own or look around on the web for code that does it.
These are called snippets.
Google use proprietary (and possibly patented) methods to garner this information, so there is no simple answer.
As you suggest, they will use meta-description information if it is there. (How to set the meta-information to help Google.)
They will also honour requests from the page authors to NOT include snippets. (How to prevent Google from displaying snippets) You should probably respect this too (as well as robots.txt, of course.)
You may have some luck with existing auto-summary packages, such as OTS.
You may want to check AboutUs.org (i.e. http://www.aboutus.org/StackOverflow.com).
But, there's little chance that the site will have an aboutus page and not have a meta description.
Some info that might explain how google does this:
Webmasters/Site owners Help
Adding a URL to google
I am not familiar with Google APIs, but perhaps there is an official way to get such information.
Interesting. some sources are better than others.
For "audiotuts.com" google has a worse description than AboutUs.com.
Google
Nov 18th in General by Joel Falconer ·
1. Recently, an AUDIOTUTS reader asked me about creative process. While this
is a topic that can’t be made into a
...
AboutUs.com:
AUDIOTUTS is a blog/tutorial site for
musicians, producers and audio
junkies! It is the sister site of the
popular PSDTUTS, VECTORTUTS and
NETTUTS.
I hate problems like these... they should be trivial but they aren't!
If you can assume English content, you can first look for Meta Description, and if that doesn't work, you can look for the first two or three sentence-like word sequences.
A product I worked on looked for the first P or DIV that contained more than one sequence of > n "words" delimited by periods. It would use the two or three sentence-like sequences, up to x total words, as a summary paragraph. It wasn't 100% accurate, but good enough for the average case. The number of words was adjusted a few times to eliminate things like navigation elements.

Resources