Duplicate content
Google has a problem with duplicate content:
In order to serve our users, Google has to organize, sort and present relevant results.
The SEO wants positions in the "serp" and he knows that SEO a website necessarily requires the content.
The content is the resource coveted (because before even thinking about directing links, there must be something at the end of the link).
The human nature is lazy, and it will tend to take shortcuts - eg sting of the content in the neighbor ...
So on one side there is Google who wants to give the best experience to its users and on the other side there is the webmaster looking to give exposure to its customers ...
Google is simple:
If we can find the same information elsewhere, it is useless to repeat it, so we simply omits the results, see here I do not mean a penalty, but of omission, I have already mentioned this in a passed post of duplicate content.
Google has a problem with duplicate content:
In order to serve our users, Google has to organize, sort and present relevant results.
The SEO wants positions in the "serp" and he knows that SEO a website necessarily requires the content.
The content is the resource coveted (because before even thinking about directing links, there must be something at the end of the link).
The human nature is lazy, and it will tend to take shortcuts - eg sting of the content in the neighbor ...
So on one side there is Google who wants to give the best experience to its users and on the other side there is the webmaster looking to give exposure to its customers ...
Google is simple:
If we can find the same information elsewhere, it is useless to repeat it, so we simply omits the results, see here I do not mean a penalty, but of omission, I have already mentioned this in a passed post of duplicate content.
I will not go into details, but Google needs to store gargantuan amounts given on its own servers, which is to say that their costing $$$ (servers, electricity, technicians, air conditioning, local etc.) then it is clear that in order to reduce these costs, there is no point to keep "cache" the same content multiple times - not to mention that it is useless to present the same content to a user.
It must be remembered that phrase that I often say: "Google owes you nothing!", His only concern is to identify the Web as best he can. If he owes you nothing, why he'll pay to store your content in multiple copies?
Google now has to decide what it should prioritize websites (it does with its algorithm, and I hope it has developed a way to determine the content). But what if your content was indexed on a neighbor even before the "googlebot" have visited your site?
There are some people I call lazy web, who do not care to create content, and they are transplanting the content of others (scraping), if that site "scraping" has new content generated daily basis, it is normal that the "googlebot" go and visit more often, so it is likely that the content of your site to be indexed more quickly on a scraping site than yours!
Google certainly did not want to become police the web and decide who owns the content, but by the force of circumstances he was forced to. If your content is present in more than one place in your site or worse on other websites, Google needs to clean house.
For SEO is more complicated:
He must try to convince the owner of the website that he needs more content.
The answers that I usually get are:
I have no time to do this
It is you who knows that!
We will not re-write texts
Takes the texts of the brochure (which by the bandwidth a bit old)
Takes the text on the sites of my competitors and changes them a little.
I do not want it to cost me more (they think that hiring a SEO, the problem is solved)
You definitely have the best ...
As I mentioned above, a site needs content (it takes meat around the bone) and Google loves fresh meat!
I walk away a bit of the subject, because the article about today is not to solve the problem of creating content, but rather to avoid the content to be duplicated ...
I will humbly attempt to provide solutions for you to make corrections.
The causes of duplicated content
Mismanagement URLs INVOLUNTARY (CMS, index.php, session ID, RSS, Canonical, Printer friendly pages, etc.)
Scrapping VOLUNTARY (content thieves)
Solutions duplicate content
Decided whether you want or not WWW (.htaccess, Google webmaster tools)
www.2015TransAm.com or
2015TransAm.com
This is called "canonicalization"
Canonicalization in Webmaster Tools
Example: canonicalization in Webmaster Tools
Familiarize yourself with your CMS
Create duplicate content, every time you create an article an address is created, you must then link this article to a menu, when you create the menu, create another address is if you're not careful, you now have 4 copies of your article:
Article
the menu
.pdf Article
The pdf of the article related to the menu
Avoid putting a link on the index finger of your site from the Navigation Home button for example, because you could easily end up with duplicated content 3 times ...
www.2015TransAm.Com
www.2015TransAm.com/index.html
www.2015TransAm.com/accueil.html
Instead, put the link on the root <a href="/"> Home </a>
301 redirect
Google recommends using the 301 redirect with .htaccess
the Scraping
A solution that could be interesting for us to just blogs Google:
"However, it is useful to ensure that each site on which your content is broadcast includes a link back to your original article. »
It could also set the case of the web ... laziest ones that automates their Scraping.
If there is site Scraping which remove the links, your only recourse is to file a DMCA complaint
It must be remembered that phrase that I often say: "Google owes you nothing!", His only concern is to identify the Web as best he can. If he owes you nothing, why he'll pay to store your content in multiple copies?
Google now has to decide what it should prioritize websites (it does with its algorithm, and I hope it has developed a way to determine the content). But what if your content was indexed on a neighbor even before the "googlebot" have visited your site?
There are some people I call lazy web, who do not care to create content, and they are transplanting the content of others (scraping), if that site "scraping" has new content generated daily basis, it is normal that the "googlebot" go and visit more often, so it is likely that the content of your site to be indexed more quickly on a scraping site than yours!
Google certainly did not want to become police the web and decide who owns the content, but by the force of circumstances he was forced to. If your content is present in more than one place in your site or worse on other websites, Google needs to clean house.
For SEO is more complicated:
He must try to convince the owner of the website that he needs more content.
The answers that I usually get are:
I have no time to do this
It is you who knows that!
We will not re-write texts
Takes the texts of the brochure (which by the bandwidth a bit old)
Takes the text on the sites of my competitors and changes them a little.
I do not want it to cost me more (they think that hiring a SEO, the problem is solved)
You definitely have the best ...
As I mentioned above, a site needs content (it takes meat around the bone) and Google loves fresh meat!
I walk away a bit of the subject, because the article about today is not to solve the problem of creating content, but rather to avoid the content to be duplicated ...
I will humbly attempt to provide solutions for you to make corrections.
The causes of duplicated content
Mismanagement URLs INVOLUNTARY (CMS, index.php, session ID, RSS, Canonical, Printer friendly pages, etc.)
Scrapping VOLUNTARY (content thieves)
Solutions duplicate content
Decided whether you want or not WWW (.htaccess, Google webmaster tools)
www.2015TransAm.com or
2015TransAm.com
This is called "canonicalization"
Canonicalization in Webmaster Tools
Example: canonicalization in Webmaster Tools
Familiarize yourself with your CMS
Create duplicate content, every time you create an article an address is created, you must then link this article to a menu, when you create the menu, create another address is if you're not careful, you now have 4 copies of your article:
Article
the menu
.pdf Article
The pdf of the article related to the menu
Avoid putting a link on the index finger of your site from the Navigation Home button for example, because you could easily end up with duplicated content 3 times ...
www.2015TransAm.Com
www.2015TransAm.com/index.html
www.2015TransAm.com/accueil.html
Instead, put the link on the root <a href="/"> Home </a>
301 redirect
Google recommends using the 301 redirect with .htaccess
the Scraping
A solution that could be interesting for us to just blogs Google:
"However, it is useful to ensure that each site on which your content is broadcast includes a link back to your original article. »
It could also set the case of the web ... laziest ones that automates their Scraping.
If there is site Scraping which remove the links, your only recourse is to file a DMCA complaint