There has always been much debate about whether Google has an duplicate content penalty. Everyone agrees that that duplicate content is not a good thing. You want your website to look as credible to the search engines as you can. What in the blue blazes is duplicate content, anyway? It is essentially content that can be accessed by more than one URL.
Imagine that you have 50 pages of content on your website and that only 25 of those page have unique content. Spiders have to go through your site, prioritize and sort out your content, and will only index unique information.
The last thing you want is for your site to be competing against itself and dilute your results. If you have a large website, that can be especially challenging but can be done.
Here are few suggestions to help you along your way:
Rewrite Title Tags
Having good Title Tags is one of the most important ways to increase your website’s visibility. It is still very common to see titles like “Jim’s Carpet Company” or even worse; “Welcome to my site.” What makes a good page title is always up for debate but it can be safely said that if a page is important enough to exist it is important enough to have its own title that correlates well with the on page content.
Rewrite META Descriptions
While they are not as influential to the search engines as a pages title, the META Description is an important way for spiders to separate that page from the rest. Either take the time to write informative and unique page descriptions or be sure not to use the same description for all other pages.
Rewrite Page Copy
Duplicate content occurs when people include the same block of content on many pages or when content from another site has been lifted and placed on yours. If your website falls into this category, you are not alone. Several other websites have the same problem and the only way to remedy this is to take some time and rewrite the content on your site.
On a positive note – This will give you a chance to review the content and ensure that it is up-to-date.
Lighten Your Code
Everyone likes clean code and that includes the spiders that frequent your web site. Even though search engines have gotten a lot better at this, content that is buried deep into the code can be hard to find. We want to make it easy for search engines to find your unique content. Cleaning up the code on your web site will also help to speed it up which also factors into your rankings as well better response time and overal usability. I like fast websites. Don’t you?
Emphasize Unique Content
Remember those H tags that you could never find a use for when writing a Word Document? Well, they are important for SEO. Emphasizing certain content by using <h1>, <h2>, <strong>, and <em> tags help your visitors as well as search engines focus on specific content and will also help search engines to tell individual pages apart.
There are several names for this (Duplicate URLs, Canonical Formatting, Canonicalization) and this subject is worth several blog posts, but the idea is same. Dynamic sites (WordPress, Drupal, Joomla), when not set up correctly, suffer from content duplicates created by multiple URLs pointing to the same page. For example, you may have 1 page with the following 3 URLs:
In a perfect world just one of these URLs would show and the others would operate as hidden redirects, however, this is not always possible. To get around this pick the URL with the most descriptive format and block the rest by using using a nofollow or by placing the redundeant URLs in a robots.txt file.
Block Functional URLs
This is similar to identifying and resolving duplicate URLs (the post above) but is about URLs that only have a functional purpose; displaying a printable version of a page or passing along information such as a search string.
Removing these completely is time intensive and isn’t always possible, but can and should be blocked to search spiders. These page have no value to the spiders or SERPs.
For example, you may have something like: