Difference between subdomains `WWW` and `m` - web

What's the difference between these two subdomains WWW and m?
For instance, WWW.medium.com and m.medium.com
In my country, WWW.medium.com is blocked, but when I try using m.medium.com instead, It works well. So I want to know what's the difference between them? Can they have different content?
PS: I have searched about that, but haven't found anything related
Thanks in advance.

Yes, they can. In fact, the two domains www.medium.com and m.medium.com could point to two completely different servers. This is up to the owner of medium.com to decide.
Usually, a 'm' subdomain points to the version of the site that is intended for mobile devices, i.e. devices with a smaller screen size.
Pinging the domains, I get 162.159.153.4 for medium.com, 162.159.152.4 for www.medium.com and 162.159.152.4 for m.medium.com.

Related

How many number of subdomains can a domain have?

I am planning to build a service based on subdomain like wordpress or tumblr.
I want to know what is the maximum number of subdomains a domain can have.
Well. The absolute theoretical maximum is 2^504, but that assumes no limitations on the octets making up the names. If you want the names limited to ASCII letters and digits, the answer is 111444219848545291112918149658401217019177846881717006276548100629318214534968256903948922840416256 (that is, 36^63).
In another sense the answer is "Far, far more than you will ever need".
This would be more dependant on the DNS server than on the standard. BIND allows for a maximum of 16'777'216 objects per zone file while Microsoft DNS is reported to be stable up to 20'000 objects per zone. This does not mean however that you will be able to max out the DNS object limits as your average website owner is going to want to have a meaningful object name for their sites sub domain, additionally the maximum character count for fully qualified domain names is 255 characters with no individual segment (between dots) being longer than 63 characters as per the DNS specifications.
Effectively what this means is that while there are restrictions and limitations the practical answer is that you are unlikely to encounter limitations due to DNS specifications in any reasonable timeframe.

Sort domains by number of public web pages?

I'd like a list of the top 100,000 domain names sorted by the number of distinct, public web pages.
The list could look something like this
Domain Name 100,000,000 pages
Domain Name 99,000,000 pages
Domain Name 98,000,000 pages
...
I don't want to know which domains are the most popular. I want to know which domains have the highest number of distinct, publicly accessible web pages.
I wasn't able to find such a list in Google. I assume Quantcast, Google or Alexa would know, but have they published such a list?
For a given domain, e.g. yahoo.com you can google-search site:yahoo.com; at the top of the results it says "About 141,000,000 results (0.41 seconds)". This includes subdomains like www.yahoo.com, and it.yahoo.com.
Note also that some websites generate pages on the fly, so they might, in fact, have infinite "pages". A given page will be calculated when asked for, and forgotten as soon as it is sent. Each can have a link to the next page. Since many websites compose their pages on the fly, there is no real difference (except that there are infinite pages, which you can't find out unless you ask for them all).
Keep in mind a few things:
Many websites generate pages dynamically, leaving a potentially infinite number of pages.
Pages are often behind security barriers.
Very few companies are interested in announcing how much information they maintain.
Indexes go out of date as they're created.
What I would be inclined to do for specific answers is mirror the sites of interest using wget and count the pages.
wget -m --wait=9 --limit-rate=10K http://domain.test
Keep it slow, so that the company doesn't recognize you as a Denial of Service attack.
Most search engines will allow you to search their index by site, as well, though the information on result pages might be confusing for more than a rough order of magnitude and there's no way to know how much they've indexed.
I don't see where they keep or have access to the database at a glance, but down the search engine path, you might also be interested in the Seeks and YaCy search engine projects.
The only organization I can think of that might (a) have the information easily available and (b) be friendly and transparent enough to want to share it would be the folks at The Internet Archive. Since they've been archiving the web with their Wayback Machine for a long time and are big on transparency, they might be a reasonable starting point.

Limiting number of connections from bottlepy's `run`?

Is it possible to limit connections, restrict to certain number of unique IP addresses; with Bottle's run command?
All I could find was this: http://bottlepy.org/docs/dev/deployment.html
In a word, no.
There's no built-in way to do this. (You didn't include a lot of detail, so I'm making some assumptions about what I think you might have in mind.) In theory (and only in theory--please don't go this route), you could write a Bottle plug-in that tracks connections and denies requests depending on certain access patterns. But there are many reasons you shouldn't.
Instead, you should rely on your web server (the one in which Bottle will run in production--e.g., Apache, nginx) to handle this. For example, here's an SO discussion on rate limiting in Apache.
Sorry there's no "yes" answer; but hope this helps! Cheers.

Multiple domains for the same website

Short question: What is the optimal method for routing multiple Domains for the same website? Specifically, how to route a uri with an international tld (.рф or .срб) and an ascii tld (.ru or .rs respectively)?
Long question: I have two domain names for the same website, one ascii and one international (cyrillic), http://domain.rs and http://домен.срб pointing to the same website. On the one hand, I know many websites which use both domains equally and parallelly (such as for example http://rts.rs and http://ртс.срб), but on the other, I've been advised that it's a bad practice from SEO point of view, and that instead I should have one domain redirecting to the other. Аre there any advices, or resources where I could get informed about how to handle international domains alongside with ascii ones?
Using "parallel" domains, without some kind of canonicalization in-place, will result in duplicated content issue. So, I wouldn't suggest it at all.
(There is a "loop-hole", sort of speak, that allows different TLDs to appear independently for different locals but truly this gains you nothing at all, just removes some of the DC issues...)
If I understand you correctly, the right thing to do here is to stick to one Main Domain and use 301 redirects for all others (page to page preferably). Ascii or not, is irrelevant. For Main Domain, choose you "oldest" one or/and the one with most inbound links.
In the long run this is also most practical solution as it will allow you to concentrate your link-building efforts, focusing all inbound links around one Winner instead of just spreading them around among several Mediocres.

How Prevent Google Duplicate Content Problem | Multi Site

I'm about to launch a multi-domain affiliate sites which have one thing in common which is content. Reading about the problem with duplicate content and Google I'm a little worried that the parent domain or sub sites could get banned from the search engine for duplicated content.
If I have 100 sites with similar look and feel and basically same content with some minor element changes, how will I go on preventing banning, indexing these correctly?
Should I should just prevent sub-sites from been indexed completely with robots?
If so how will people be able to find their site... I actually think the parent is the one that should only be indexed to avoid, but will love to her other expert thoughts.
Google have recently released an update that will allow you to include a link tag in the head of pages that are using duplicated content that point to the original version, they're called canonical links and they exist for the exact reason you mention, to be able to use duplicated content without penalisation
For more information look here..
http://googlewebmastercentral.blogspot.com/2009/02/specify-your-canonical.html
This doesn't mean that your sites with duplicated content will be ranked well for the duplicated content but it does mean the original is "protected". For decent ranking in the duplicated sites you will need to provide unique content
If I have 100 sites with similar look
and feel and basically same content
with some minor element changes, how
will I go on preventing banning,
indexing these correctly?
Unfortunately for you, this is exactly what Google downgrades in its search listings, to make search results more relevant, and less rigged / gamed.
Fortunately for us (i.e. users of Google), their techniques generally work.
If you want 100s of sites, to be properly ranked, you'll need to make sure they each have unique content.
You won't get banned straight away. You will have to be reported by a person.
I would suggest launching with the duplicate content and then iterating over it in time, creating unique content that is dispersed across your network. This will ensure that not all sites are spammy copies of each other and will result in Google picking up the content as fresh.
I would say go ahead with it, but try to work in as much unique content as possible, especially where it matters most (page titles, headings, etc).
Even if the sites did get banned (more likely they would just have results omitted, but it is certainly possible they would be banned in your situation) you're now just at basicly the same spot you would have been if you decided to "noindex" all the sites.

Resources