1. This site uses cookies. By continuing to use this site, you are agreeing to our use of cookies. Learn More.

Duplicate Content and Tables of Contents

Discussion in 'White Hat SEO' started by drumltd, Oct 13, 2013.

  1. drumltd

    drumltd Regular Member

    Joined:
    Dec 4, 2008
    Messages:
    467
    Likes Received:
    13
    I host a site which carries a lot of instruction manuals. Each page of the instruction manual is displayed with a table of contents for the manual, to aid easy navigation. The table of contents for all the manuals are all pretty similar but the content is mostly different.

    Running a test on copyscape, I was surprised to find that most of the pages that have "low content" i.e mainly images are coming up as duplicates because of the TOC displayed.

    Is there anyway I can tell google, that the TOC is a menu or similar, so that they don't think the pages are all duplicates?
     
  2. thedorf

    thedorf Senior Member

    Joined:
    Oct 1, 2008
    Messages:
    1,193
    Likes Received:
    700
    Occupation:
    what? I gotta have a job?
    Location:
    BHW - Where else?
    No way to do that I am aware of. You could try hiding the TOC in javascript but Google probably knows that trick.
     
  3. drumltd

    drumltd Regular Member

    Joined:
    Dec 4, 2008
    Messages:
    467
    Likes Received:
    13
    I was afraid of that.

    On a side note, due to the nature of the manuals, there are quite a few duplicate pages, is there anyway I can get a report, of the duplicate pages, so that I can add "noindex" to them, trying to avoid a panda slap.
     
  4. drumltd

    drumltd Regular Member

    Joined:
    Dec 4, 2008
    Messages:
    467
    Likes Received:
    13
    I've been reading up on this, and about the best I can come up with is, to put the TOC in an iframe, then put the source in a section, that is blocked by robots.txt

    This sounds good, but could it possibly break the flow of the website for the spiders, I do have next\prev buttons that will get to all pages eventually. So in theory it should work okay.

    Anybody any thoughts?
     
  5. drumltd

    drumltd Regular Member

    Joined:
    Dec 4, 2008
    Messages:
    467
    Likes Received:
    13
    Using an iframe to "hide" the toc from spiders, and hence make the pages less duplicate. Is there anyreason to use robots.txt to bloack access rather than simply put a met tag "noindex" in the toc html page??