SEO Website with links only - search

I have a website that contains links to other sites only.
They link to an image gallery or a video.
Does google accepts this or will it penalize my site because I don't have any real content?
Thanks

Google does look for human generated content. They also look for the links to your site. I would add stuff to your site and also guest blog etc to get traffic to your site.

Related

Google do not index my posts

I'd like to know why google do not index my posts on my blog writes in NodeJS.
Link of a post : http://icecom.fr/articles-icecom/9
Anthony
There are several reasons why Google isn't indexing your website.
There are no links to your website. Google follows links on the internet to other pages. If there are no links to your website it won't find it.
You are denying access to Google through the robots meta-tag or robots.txt.
You haven't waited long enough yet, Google may take some time before it has indexed your website.
Of course you can supply Google with the proper URL's with a [sitemap]{https://support.google.com/webmasters/answer/156184?hl=en}. A good place to create this if you're new to it could be [here]{http://www.xml-sitemaps.com/}
#szenbalu already mentioned you can upload this sitemap.xml to Google Webmaster Tools and this way Google can index your site without the need of links. It is also faster most of the time.
Another way to get your website indexed through Google Webmaster Tools is the 'Fetch as Google' tool. In here you can tell Google to fetch and index your website. This is especially useful if you change content and want it reindexed.
About your specific case:
* You do not block Google with the meta robots tag
* I can not find a robots.txt file
* I can not find any links to your articles from [OpenSiteExplorer]{http://www.opensiteexplorer.org/}
I think that uploading a sitemap to Google Websmaster Tools + Using the Fetch as Google tool will get your site indexed within no time.
If you have any questions left, feel free to ask. :)
Do you have the robots.txt file and webmaster tools account joined to your page?
With webmaster tools you can upload sitemap that google will use to index pages.

Cross-domain sitemap in robots.txt

I want to submit a Google Webmaster Tools sitemap for a Blogger "blog" that only consists of static pages (using it as a website instead of a blog).
Since there's no way for me to host a file on the Blogger server, and therefore no way to submit a sitemap directly through Google Webmaster Tools, is it possible to host a sitemap.txt file on another server and reference it in robots.txt? Or would I run into some sort of cross-domain issue?
If you can managed to map robots.txt file to your root directory, Which is www.domain.com/robots.txt thats okay, but i don't think you have to do that. Just add your site to Google Webmasters Tools and use regular seo strategy. Google is smart enough to Index your static pages.
--Hope this helps!!

How to manage detailed description and links in google?

A client would like to have his website implemented like the following one, on google:
https://www.google.ch/search?q=mlzd&ie=utf-8&oe=utf-8&aq=t&rls=org.mozilla:fr:official&client=firefox-a
With:
a list of links to the differents sections of the website.
the google function with the arrow (on rollover of the links) to get previews of the website pages.
So, here are my questions:
Do you know to implement the several links? Is it tags?
Do I then have to do something to get the google images previews, or is it automatic?
Will the google images previews work, with a Flash website? If not, the preview will be a screenshot the website with no flash enabled?
Thank you a lot for your answers!!
David
The links are called Sitelinks, they are automatic. You cant do anything to sepefically enable them.
http://support.google.com/webmasters/bin/answer.py?hl=en&answer=47334
The preview is automatic, but you do need to make sure that 1) allow google to cache the url (ie avoid "nocache" and 2) not block in anyway the Previews https://sites.google.com/site/webmasterhelpforum/en/faq-instant-previews
See link in the previous answer

How do I check an entire website to see if any page in it links to a particular URL?

We have been hounded by an issue in our websites because web protection facility pages like ones from Norton keep on telling certain visitors in certain browsers that our websites are potential risks because we link to a certain http://something.abnormal.com/ (sample URL only).
I've been trying to scour the site page by page, to no avail.
My question, do you know any site that would be able to "crawl" into our website's pages and then check if any text, image, whatever in them links to the abnormal URL that keeps on bugging.
Thanks so much! :)
What you want is a 'spider' application. I use the spider in 'Burp Suite' but there are a range of free, cheap and expensive ones.
The good thing about Burp is you can get it to spider the entire site and then look at every page for whatever you want, whether it be something to match a regex or dynamic content etc.
If your websites consist of a small amount of static content pages, I would use wget to download all pages (ignoring images)
wget -r -np -R gif,jpg,png http://www.example.com
and then use a text search for the suspicious url on the result. If your websites are more complex, httrack might be easier to configure for a text-only download.

Search engine robot.txt

I want to add a robot.txt so my web page can be found...
So I have heard that putting a robot.txt with meta tags in the root of my site can do this.
Is this true?, if so,
What would be the steps to add or generate this robot.txt?
I have found this
Robots.txt is more for telling the crawlers where to and where to not go once they've already reached your site.
A better way to get crawlers onto your site is to build a sitemap for your site, then use Google Webmaster to submit this sitemap to google. You'll also want to include the sitemap on your site's root url and tell google where it is (all of this can be done in Google's Webmaster Tool linked to above).
No, it won't make your webpage suddenly visible. It just instructs web crawlers on how to index your site.
http://www.robotstxt.org/

Resources