I have a website developed in asp.net MVC4 with Google analytics code. We noticed three issues while analyzed GA reports:
It displays admin pages (like "/Areas/Register/Views/Merchandiser/Create.cshtml", for example) and shouldn't be in analytics report
It displays duplicated (like "/myurl" and "/myURL", for example) urls and shouldn't be duplicated
It displays error pages (like "/Errors?aspxerrorpath=/Admin/Conteudos", for example) and shouldn't be in analytics report.
For #2, I have gone through this link but I did not get the results even after 72hrs.
For #1 & #3 - can I use robots.txt ? - if yes, how can I use this for asp.net mvc4 website?
Is there any way to resolve above issues?
Any help in this regard will be highly obliged.
The problem is that asp.net is allowing case-insensitive URLs while google analytics considers foo/bar and Foo/Bar to be different urls and is the correct behavior per the W3C: http://www.w3.org/TR/WD-html40-970708/htmlweb.html .
Related
I have one page on my site that uses SharedArrayBuffer, which means I need to use Cross-Origin-Opener-Policy - but this also means I can't use analytics to see if anyone actually uses that page or not.
So I wondered if there's a workaround for this?
I use Fathom Analytics for tracking, so have it analytics set to load from a sub-domain on my site.
I have set the Cross-Origin-Embedder-Policy to require-corp, and Cross-Origin-Opener-Policy to same-origin.
Is there a way to either iframe the analytics, or iframe the Cross-Origin-Opener-Policy app and have analytics on the parent? I've tried looking this stuff up but it's very much outside my realm of expertise. Thanks for any guidance you can offer.
I might get flagged down by this question.. but still will give it a shot..
Since Google Site Search is going out of business and we are not interested in the free version of it - We decided to go with the Amazon Cloud Search option. The challenge though is - it is not straight forward. We have to build a crawler and there are some features that needs to be custom built.
I am trying to see examples where websites have used ACS and worked but i am not able to find anything good.. Have anyone tried using Amazon Cloud search for their Website search. Our website has around 15000 plus pages.
We are .net based solution - so i am thinking to write a crawler.. extract content on nightly basis and send it to Amazon. Would it be the right way?
ACS is based on Solr. If your site is under your control, i think the first step is extracting all useful content out and generate them into xml/json files, then use AWS CLI upload these documents to ACS. ACS has REST APIs to let you to get the query result. You need to define indexes before uploading them.
I would like to scrape Content from a website with AlchemyAPI. I get informed about this feature on http://www.alchemyapi.com/api/scrape/qlang.html
I will implement it in the same way as in the example:"Querying Inside Tables (Selecting a Column Inside a Specific Row)".
Could somebody please help me, how to use this in Node.js and Cquery? Which parameters do I need to get specific fields like price as output?
No, it's not currently possible to do this. Since AlchemyAPI was acquired by IBM, the remaining services have been incorporated into Watson. Most of the AlchemyAPI services are now covered in the Natural Language Understanding (NLU) service: https://www.ibm.com/watson/developercloud/doc/natural-language-understanding/ but there is no feature that allows you to scrape content from a web site per se.
The NLU service does allow you to retrieve text from a web page using the analyze endpoint: https://www.ibm.com/watson/developercloud/natural-language-understanding/api/v1/#post-analyze
We upgraded to Universal Analytics about a year ago and since upgrading our Site Search isn't displaying the keywords within our Analytics account, or even tracking the number of times it is used.
This is a problem as a marketer as I like to know what people are looking for so that I can provide them relevant content.
All I can find on the Google blog is this:
https://support.google.com/customsearch/answer/2715510
'At this time, Google Custom Search Engine, is not compatible with Universal Analytics.'
Has anyone got any suggestions as to how I can get this data back? Are there any other custom site searches that I could use instead? Are Google even looking at this?
Google CSE is using old _gaq calls for analytics, it can be mimicked by adding following code line at the end of universal GA snippet:
_gaq={push:function(){ga('send','pageview',arguments[0][1]);}}
View - View settings - Site search settings is on and correctly tuned?
I have a simple HTML site with 100+ pages or so. I want to add a search bar at the top so the user can search the site. I know about Google Custom Search, but it shows ads unless you pay at least $100. Obviously I'd like ad-less search on my site for free if at all possible!
I've also heard about Lucene/Solr, but they do not actually crawl the site. For that I would apparently need Nutch.
Anyway, the site I have runs on a Microsoft IIS6 server, but I have basically no knowledge as to how Solr, Nutch, etc. gets "installed" on the server.
Also: I'd like to point out that I do have a local copy of the site. Perhaps I can do one big initial nutch "crawl" locally that will create an .xml for Solr?? That would help me get "up and running", but probably wouldn't be a good long-term solution.
..so should I just use Google Custom Search? or is there a not-extremely-painful-to-implement alternative? The brain hurts folks.
You did not mention how many search requests you want to handle but if you use the json-rest-api of google's custom search you have 100 searchqueries a day for free and you can display them without any ads on your page.
An simple example request can be found here.
Here is an easy way that works pretty well, although you may be looking for something more than this.
http://sitecomber.com/getsitecomber/
You can create code to paste into your site in about 2 minutes. It doesn't get easier than that. Search is powered by Google, but results are isolated to your website.
EDIT: This no longer works.