1. This site uses cookies. By continuing to use this site, you are agreeing to our use of cookies. Learn More.

Why my scrapebox harvester harvests all those links I do not want?

Discussion in 'Black Hat SEO Tools' started by Chicilikit, May 23, 2013.

  1. Chicilikit

    Chicilikit Senior Member

    Joined:
    Dec 21, 2010
    Messages:
    873
    Likes Received:
    153
    Hello, I'm now using scrapebox for finding banned tumblr blogs, but when I run harvester it grabs so much domains I do not want and it is really pain in the ass. I do not know what I'm doing wrong. I use this footprint ( -site:"www.tumblr.com" inurl:tumblr.com ) and as keywords right now list of 30K most used english words.
    Also have another problems with scrapebox lately, very often after scraping few dozens of thousands links I'm not able to stop the harvester and it just runs forever. Also google very fast shows 503 error, I guess it is because of proxies. What are the best kind of proxies or proxy service when you want to use scrapebox for harvesting lots of links from google every day? Thanks a lot for answers all you scrapebox experts :)
     
  2. naweed

    naweed Junior Member

    Joined:
    Dec 25, 2011
    Messages:
    186
    Likes Received:
    37
    Your footprint is wrong. If you use the footprint -site:www.tumblr.com, you will get only results from tumblr main site. You should use site:tumblr.com. The 3 Ws make a big difference. Try it with both on google and you will see if you use site:www.tumblr.com, you won't get any tumblr blog,only results for the main tumblr site. But if you use site:tumblr.com then you will get a lot of tumblr blog.
    Hope it helps :)
     
    • Thanks Thanks x 1
  3. soothsayerpg

    soothsayerpg Power Member

    Joined:
    Feb 23, 2011
    Messages:
    584
    Likes Received:
    225
    Home Page:
    Also try inurl:.tumblr.com