#1
  1. No Profile Picture
    Registered User
    SEO Chat Explorer (0 - 99 posts)

    Join Date
    Nov 2008
    Posts
    11
    Rep Power
    0

    Usability vs Duplicate Content


    hello again,

    question: the site i'm currently helping out on has a lot of navigation code.
    after running similarity tests here:
    webconfs.com/similar-page-checker.php
    it looks like its going to be tough to get the ratio under 50% without major changes to the site.

    so given the options of html navigation vs flash vs recommendations? what does this group recommend in addition to more content to get the ratio lower.

    is there a preference to bot friendly vs non friendly navigation (assuming robots.txt & google sitemap are in place) and does the site need ANY html coded generic navigation?

    and
    if repetitive text (specifications, warranty information, etc) is replaced with images does this have a negative effect on optimization?

    reference file list
    original files @72% similar
    sdscooters.com/Electric_Scooters/Goped_Li-Ion.html
    vs
    sdscooters.com/Electric_Scooters/ESR750H_Hoverboard.html

    stripped of generic navigation and repetitive text @55% similar
    sdscooters.com/Electric_Scooters/Gopedtest2.html
    vs
    sdscooters.com/Electric_Scooters/ESR750Htest2.html

    thanks

    katie
  2. #2
  3. End game expert
    SEO Chat Good Citizen (1000 - 1499 posts)

    Join Date
    Aug 2007
    Location
    Cebu Philippines
    Posts
    1,272
    Rep Power
    33
    Originally Posted by katiem
    hello again,

    question: the site i'm currently helping out on has a lot of navigation code.
    after running similarity tests here:
    webconfs.com/similar-page-checker.php
    it looks like its going to be tough to get the ratio under 50% without major changes to the site.

    so given the options of html navigation vs flash vs recommendations? what does this group recommend in addition to more content to get the ratio lower.

    is there a preference to bot friendly vs non friendly navigation (assuming robots.txt & google sitemap are in place) and does the site need ANY html coded generic navigation?

    and
    if repetitive text (specifications, warranty information, etc) is replaced with images does this have a negative effect on optimization?

    reference file list
    original files @72% similar
    sdscooters.com/Electric_Scooters/Goped_Li-Ion.html
    vs
    sdscooters.com/Electric_Scooters/ESR750H_Hoverboard.html

    stripped of generic navigation and repetitive text @55% similar
    sdscooters.com/Electric_Scooters/Gopedtest2.html
    vs
    sdscooters.com/Electric_Scooters/ESR750Htest2.html

    thanks

    katie
    One of the immediate causes when you are talking about high similarity of web pages is that the navigation menu. IF you have a substantial long navigation menu, it will contains more words more than your actual content, so if you are going to compare page per page, it will still be similar because of your consistent navigation which occupies substantial amount of text.

    If you really need to drop the % , you may consider shortening your navigation and edit your content in such as way actual unique content is big compared to navigational text.

    This is not really a critical part, Google understands the importance of clear navigation and shortening it sacrifices the usability of your site.

    In this way, the best course of action is to add a substantially longer content which is both desirable to your users because it shows a more complete information at the same time dropping the % of the page similarity.

    Standard article pages consist of around 1500 words, it is good you aimed that amount of words.

    You can also eliminate repetitive text such as long copyright notices and unimportant/boilerplate repetition in your web page content. It may be better to put it into a single file and just a link to it.
  4. #3
  5. No Profile Picture
    Registered User
    SEO Chat Explorer (0 - 99 posts)

    Join Date
    Nov 2008
    Posts
    11
    Rep Power
    0
    Hi Emerson,

    Originally Posted by Emerson
    One of the immediate causes when you are talking about high similarity of web pages is that the navigation menu. IF you have a substantial long navigation menu, it will contains more words more than your actual content, so if you are going to compare page per page, it will still be similar because of your consistent navigation which occupies substantial amount of text.
    The site is very navigation heavy - anything on it can be reached in 3 clicks or less

    Originally Posted by Emerson
    If you really need to drop the % , you may consider shortening your navigation and edit your content in such as way actual unique content is big compared to navigational text.
    We don't want to loose the 3 click ability so were looking at other options - ie flash or other non spiderable options

    Originally Posted by Emerson
    This is not really a critical part, Google understands the importance of clear navigation and shortening it sacrifices the usability of your site.
    Does anyone here have any evidence that google removes the navigation code from their calculations when considering a page spam?

    Originally Posted by Emerson
    In this way, the best course of action is to add a substantially longer content which is both desirable to your users because it shows a more complete information at the same time dropping the % of the page similarity.
    Standard article pages consist of around 1500 words, it is good you aimed that amount of words.
    1500 words of product descriptions is an insane amount of fluff - we are lucky to get 100 words for most products and even luckier 20 words for basic items

    Originally Posted by Emerson
    You can also eliminate repetitive text such as long copyright notices and unimportant/boilerplate repetition in your web page content. It may be better to put it into a single file and just a link to it.
    hmmm, could be a solution google doesnt like ajax include files so it would need to be an external page. Wouldnt an image be easier, this is information that doesnt need to be indexed - recommendations?

    thanks

    katie
  6. #4
  7. End game expert
    SEO Chat Good Citizen (1000 - 1499 posts)

    Join Date
    Aug 2007
    Location
    Cebu Philippines
    Posts
    1,272
    Rep Power
    33
    Originally Posted by katiem
    Hi Emerson,



    The site is very navigation heavy - anything on it can be reached in 3 clicks or less


    We don't want to loose the 3 click ability so were looking at other options - ie flash or other non spiderable options


    Does anyone here have any evidence that google removes the navigation code from their calculations when considering a page spam?


    1500 words of product descriptions is an insane amount of fluff - we are lucky to get 100 words for most products and even luckier 20 words for basic items


    hmmm, could be a solution google doesnt like ajax include files so it would need to be an external page. Wouldnt an image be easier, this is information that doesnt need to be indexed - recommendations?

    thanks

    katie
    Try to see if it exceeds 100 links per page since it is navigation heavy. It will also increases usability since users are not overwhelmed with links.

    Those number of words in your product pages are ok. I thought you will go for the homepage which I preferred to have ample content.

    in this case, try to see if there are files you would not need to be indexed, block it in robots.txt and will make your site pretty crawable because Googlebots will focus on directories and files that will crawl, ignoring others.
  8. #5
  9. No Profile Picture
    Registered User
    SEO Chat Explorer (0 - 99 posts)

    Join Date
    Nov 2008
    Posts
    11
    Rep Power
    0
    Emerson,

    Looked into your question a bit further by examining highly regarded sites. #1 Shopping Site - Amazon - about 350 links, #1(shop favorite) Computer site New Egg - over 350, Highly regarded Computer Site (design/usability) Dell 115 - 150 links depending on page, and Victorias Secrete (bra page) - 78 - SDscooters ~200 depending on page.

    Don't really think we are going to worry about high link numbers any longer, we will add the google sitemap to make sure it all gets crawled. We are 150% crawled/indexed at the moment (extra 50% are old pages that are still directing traffic but unreachable from the site)

    Back to the topic - Any recommendations, comments etc on using flash navigation instead of html for the purpose of boosting page content relevancy percentages? Illegal?

    katie

    Originally Posted by Emerson
    Try to see if it exceeds 100 links per page since it is navigation heavy. It will also increases usability since users are not overwhelmed with links.

    Those number of words in your product pages are ok. I thought you will go for the homepage which I preferred to have ample content.

    in this case, try to see if there are files you would not need to be indexed, block it in robots.txt and will make your site pretty crawable because Googlebots will focus on directories and files that will crawl, ignoring others.
  10. #6
  11. End game expert
    SEO Chat Good Citizen (1000 - 1499 posts)

    Join Date
    Aug 2007
    Location
    Cebu Philippines
    Posts
    1,272
    Rep Power
    33
    Originally Posted by katiem
    Emerson,

    Looked into your question a bit further by examining highly regarded sites. #1 Shopping Site - Amazon - about 350 links, #1(shop favorite) Computer site New Egg - over 350, Highly regarded Computer Site (design/usability) Dell 115 - 150 links depending on page, and Victorias Secrete (bra page) - 78 - SDscooters ~200 depending on page.

    Don't really think we are going to worry about high link numbers any longer, we will add the google sitemap to make sure it all gets crawled. We are 150% crawled/indexed at the moment (extra 50% are old pages that are still directing traffic but unreachable from the site)

    Back to the topic - Any recommendations, comments etc on using flash navigation instead of html for the purpose of boosting page content relevancy percentages? Illegal?

    katie
    I prefer to use html text links to make those links crawlable. Not sure yet with flash.Though if you use any xml sitemap. I suggest not to include unimportant URLs.

    Block all of those in a robots.txt to save bandwidth and increase site crawlability.

    Try to use this tool: Xenu Sleuth.http://home.snafu.de/tilman/xenulink.html

    Maybe it can help you find all the URLs in your and helps you decide which of those you will prefer to be indexed and not.

    As to the number of links.Amazon got massive links pointing to the site. So if they have 350 links ,it can still be crawled because the frequency of crawling amazon is frequent due to the number of links pointing to it. The more links , increases your chances of indexation. So if you have high number of links, Google will find your site always and index it by gathering information about uncrawled sections of your site or page. So it increases chances of those high number of links to be found.

Similar Threads

  1. Duplicate content email
    By jperks1985 in forum Google Optimization
    Replies: 8
    Last Post: Nov 1st, 2007, 04:38 PM
  2. Would this be considered duplicate content?
    By Faith in forum Google Optimization
    Replies: 3
    Last Post: Mar 26th, 2007, 11:29 AM
  3. Duplicate Content and vBulletin Forums
    By joeychgo in forum Google Optimization
    Replies: 1
    Last Post: Apr 1st, 2006, 07:33 AM
  4. Removing Duplicate Content - The best way
    By Sniffer1 in forum Google Optimization
    Replies: 16
    Last Post: Sep 8th, 2005, 05:28 AM
  5. What is the best way to identify and remove duplicate content?
    By budbiss in forum Google Optimization
    Replies: 8
    Last Post: Jan 11th, 2005, 07:08 PM

IMN logo majestic logo threadwatch logo seochat tools logo