Noobie Scrapebox Questions

Discussion in 'Black Hat SEO Tools' started by PurpleSEO, Oct 18, 2011.

  1. PurpleSEO

    PurpleSEO Junior Member

    Joined:
    Sep 8, 2011
    Messages:
    169
    Likes Received:
    125
    Hello good folks,

    When scraping for blogs, how do you avoid posting your URL twice to the same blog?

    Also: What is the custom footprint to search multiple blogs as a time, as opposed to doing WP, MT one by one etc?

    Thanks in advance.
     
  2. HelloInsomnia

    HelloInsomnia Jr. Executive VIP Jr. VIP

    Joined:
    Mar 1, 2009
    Messages:
    1,852
    Likes Received:
    2,967
    You avoid posting to the same blogs twice by removing duplicate domains. You can also remove duplicate URLs to make sure you don't post to the same URL twice.

    I suppose if you really wanted to you can try a footprint like:

    "powered by wordpress" OR "powered by blogengine"

    You should check out some of these videos they will help you learn more about Scrapebox:
    Code:
    http://www.youtube.com/looplinescrapebox
    
     
  3. latestt181

    latestt181 Newbie

    Joined:
    May 22, 2011
    Messages:
    25
    Likes Received:
    23
    remove duplicate domains while in harvester...
    not just duplicate URLs..
     
  4. PurpleSEO

    PurpleSEO Junior Member

    Joined:
    Sep 8, 2011
    Messages:
    169
    Likes Received:
    125
    What about when you run a scrape on one day, and a few weeks later you run a different scrape, and may end up with the same domains again.

    How do you avoid that?
     
  5. HelloInsomnia

    HelloInsomnia Jr. Executive VIP Jr. VIP

    Joined:
    Mar 1, 2009
    Messages:
    1,852
    Likes Received:
    2,967
    Create a blacklist of all the sites you have posted to.
     
  6. apekillape

    apekillape Senior Member Premium Member

    Joined:
    Dec 13, 2007
    Messages:
    1,077
    Likes Received:
    1,441
    Occupation:
    Basically, I just walk the Earth.
    Location:
    The Layer Cake
    Keep all your "POSTED TO" sites in a separate text file, then when you're about to run a new list throw it into the Scrapebox Harvester window and "Import URL List ==> Compare on Domain Level" the old list.

    It'll remove the old urls from the list of new urls.

    N
     
    • Thanks Thanks x 1