Google’s similarity patent makes me believe that in the near future the scenery of the SERPs will look quite different: hopefully no more scrapper sites and made for AdSense trash based on plagiarized materials.
With this new tool Google will be able to determine similarities between one web page and another and penalize the website that employs copied texts to rank high. Now the question is: how is Google to determine which pages are copied and which are the originals? Are they going to make their decisions on a “first indexed” basis? Then websites are in big trouble. If you publish a text on your website today and someone else copies it and publishes it in a blog, the blog will get indexed first.
Is Google to leave room for some similarity? Probably. Otherwise we’ll certainly assist at a new wave of spamdexing or we will see our pages vanishing because of what unscrupulous competitors might do.
Although Google’s similarity engine is not active yet, there are enough reasons why you should stay away from duplicate content: the supplemental results for instance. The similarity engine will push the penalties for duplicate content further: the “guilty” sites will not even get indexed anymore.
No matter how we look at it, the similarity engine is Google’s clear message to all the webmasters: no more duplicate content. Google wants quality unique content because Google wants happy users. So if you were a fan of “free articles” and “free content for your website”, I suggest you change your strategy. Fast! If you are not able, or if you don’t have the time, to write your texts alone, hire a web writer.