1. This site uses cookies. By continuing to use this site, you are agreeing to our use of cookies. Learn More.

wordpress duplicate problem

Discussion in 'Blogging' started by c0ntenth|ef, Sep 10, 2009.

  1. c0ntenth|ef

    c0ntenth|ef Power Member

    Joined:
    May 20, 2009
    Messages:
    788
    Likes Received:
    118
    Location:
    california
    when i check all the indexed pages of my wordpress blog in google, there are some indexed pages that are like this www,blog,com/page/1,www,blog,com/page/2 and so on..... the indexed text after the title are the summary of my different posts but the title of all those page/1/2/3 is the same (main blog title) so i dont want my blog to be penalized for duplicate content,so is there anywayto the pages from being indexed like that? i have all-in-one-seo plugin too and as far as i know i set up most of the options that could make diplicate content to 'noindex'.:confused:
     
  2. xpressioniz

    xpressioniz Junior Member

    Joined:
    Jun 4, 2008
    Messages:
    121
    Likes Received:
    15
    You can use robots.txt to disallow search engine from reaching those pages. No need for plugins.

    Inside robots.txt:

    User-agent: *
    Disallow: /page/

    Create this file by using notepad and upload it to ftp to your public_html folder
     
    • Thanks Thanks x 1
  3. c0ntenth|ef

    c0ntenth|ef Power Member

    Joined:
    May 20, 2009
    Messages:
    788
    Likes Received:
    118
    Location:
    california
    yeah i was thinking that too just didn't know how to write the command in te robot.txt thanks :D
     
  4. keinehabe

    keinehabe Supreme Member

    Joined:
    Nov 4, 2008
    Messages:
    1,207
    Likes Received:
    472
    Gender:
    Male
    Occupation:
    -= CEO =-
    Location:
    Heaven
    Home Page:
    title of the page aren't means you have duplicate content :) spiders know ''how to read '' subsequent pages / multiple pages :) how and from where?:) ... guess yourself lol ...
     
  5. c0ntenth|ef

    c0ntenth|ef Power Member

    Joined:
    May 20, 2009
    Messages:
    788
    Likes Received:
    118
    Location:
    california
    hmm maybe but titles are part of the content too so if i add more posts and it adds up like this page/1/2/3/4 etc... and all have the same title google might see it as dublicates.
     
  6. boiler

    boiler Power Member

    Joined:
    Apr 27, 2009
    Messages:
    677
    Likes Received:
    127
    Home Page:
    correct because my site is like that which having domain.com/page/1 domain.com/page/2 and google even giving me PR 3 on root domain :D
     
  7. xpressioniz

    xpressioniz Junior Member

    Joined:
    Jun 4, 2008
    Messages:
    121
    Likes Received:
    15
    Yes, those are duplicate contents according to Google Webmaster Tools. The pages don't contain any useful info, just a couple of links and this might indicate a low quality content in the eyes of the big G. I'd suggest you do the robots.txt I told you above to remove the page from the index. You'll have a major problems on bigger sites (I've been there on 200k+ indexes) if you keep those "junk" pages. Problems such as sandboxes, slow crawling, wrong SERPs, etc. You definitely want to avoid that. I guess on very small blog with just a few posts this is not a big deal though.
     
  8. wowhaxor

    wowhaxor Executive VIP Premium Member

    Joined:
    Apr 28, 2007
    Messages:
    2,021
    Likes Received:
    3,353
    Location:
    ?¿?
    Home Page:
    I disagree. I think one of the many benefits of WP as a platform is that it creates all kinds of pages like these to help your site get crawled better, WP performs very will in SEs in terms of onsite SEO.
     
  9. xpressioniz

    xpressioniz Junior Member

    Joined:
    Jun 4, 2008
    Messages:
    121
    Likes Received:
    15
    You can still use category pages (still, add noindex to it) to backward-crawl google by using breadcrumb, but pages, don't bother. I do agree that WP performs very well, but I think onsite SEO depends on what template you are using.

    The problem with these kind of pages is, you can't add noindex to it by using plugins such as All-In-One SEO (as far as I know), but you can add noindex to category and tags pages without having to modify the template files. So robots.txt is one solution to it. By adding noindex, google still crawl and follow pages (assuming the meta tag: noindex, follow).