1. This site uses cookies. By continuing to use this site, you are agreeing to our use of cookies. Learn More.

Scrapebox limit of 1,000,000

Discussion in 'Black Hat SEO Tools' started by lucabrasi, Nov 7, 2010.

  1. lucabrasi

    lucabrasi Regular Member

    Joined:
    Mar 22, 2010
    Messages:
    337
    Likes Received:
    17
    Occupation:
    Layed off from real job but working at a survival
    Location:
    Pacific Northwest, Oregon Washington border
    Home Page:
    If I am scraping and I get more than 1 million results (before duplicate domain removal, etc.) do I get all the results or does Scrapebox really limit it to 1,000,000.

    If thats the case I'll scrape all Google first then clean. Then add the next search engine and clean. And so on. Is this necessary?
     
  2. jakeruston

    jakeruston Junior Member

    Joined:
    Mar 29, 2009
    Messages:
    115
    Likes Received:
    70
    You'll still get all of the results, ScrapeBox saves all of the extra entries into text files in the ScrapeBox directory.
     
  3. Quo_Vadiz

    Quo_Vadiz Regular Member

    Joined:
    Mar 12, 2010
    Messages:
    259
    Likes Received:
    47
    scrapebox should REMOVE the 1 million limit. it really sucks..

    xrumer can handle list over 3 millions links

    cheers
     
    • Thanks Thanks x 1
  4. lucabrasi

    lucabrasi Regular Member

    Joined:
    Mar 22, 2010
    Messages:
    337
    Likes Received:
    17
    Occupation:
    Layed off from real job but working at a survival
    Location:
    Pacific Northwest, Oregon Washington border
    Home Page:
    Does that mean that I could be losing high PR URLs because I am over the limit.
    And thanks to both the people above for their responses.
     
  5. paincake

    paincake Power Member

    Joined:
    Aug 18, 2010
    Messages:
    716
    Likes Received:
    3,099
    Home Page:
    Like the other guy said, nothing is lost, all the urls are saved in Harvester_Sessions folder
     
    Last edited: Nov 7, 2010
  6. s4nt0s

    s4nt0s Jr. VIP Jr. VIP Premium Member

    Joined:
    Jul 10, 2009
    Messages:
    3,660
    Likes Received:
    1,940
    Location:
    Texas
    It's not a Scrapebox limitation ... txt files can only hold up to a million URL's a each.
     
  7. Quo_Vadiz

    Quo_Vadiz Regular Member

    Joined:
    Mar 12, 2010
    Messages:
    259
    Likes Received:
    47

    Nop i got txt files for xrumer with 8 million links in only 1 file , so it can be done.
     
    • Thanks Thanks x 1
  8. accelerator_dd

    accelerator_dd Jr. VIP Jr. VIP Premium Member

    Joined:
    May 14, 2010
    Messages:
    2,441
    Likes Received:
    1,005
    Occupation:
    SEO
    Location:
    IM Wonderland
    With scrapebox it saves more then 1 mil, go to Harvester sessions folder and you'll find your harvesting sessions, find the latest and inside are the txt files each with 1mil links (999927 or so) in them. Haven't tested with 3 or 4 mil but I'm sure it works.
     
  9. bobbys

    bobbys Regular Member

    Joined:
    Oct 29, 2008
    Messages:
    367
    Likes Received:
    412
    Occupation:
    Google Assault
    Location:
    127.0.0.1

    Stop making stuff up please.
     
    • Thanks Thanks x 1
  10. andreyg13

    andreyg13 Jr. VIP Jr. VIP

    Joined:
    Nov 13, 2009
    Messages:
    915
    Likes Received:
    1,774
    Occupation:
    SEO
    Location:
    http://seoshark.org
    Home Page:
    yeah, would be great if it could handle up to at least 3 mil
     
  11. lewi

    lewi Jr. VIP Jr. VIP Premium Member

    Joined:
    Aug 5, 2008
    Messages:
    2,309
    Likes Received:
    818
    Use the search function to find the thread i opened on the subject!

    Lewi
     
  12. softtouch2009

    softtouch2009 Senior Member

    Joined:
    Dec 2, 2009
    Messages:
    1,001
    Likes Received:
    225
    Occupation:
    Programming
    Location:
    ssdnet.biz
    Home Page:
    Just an explanation regarding the 1 mio. limit.
    The limit is due to the standard windows controls, which cannot hold more rows, so SB is forced to save the harvested urls into chunks with not more than 1 mio. rows per file, to allow loading them into the harvester grid.
    The control limit is not exactly 1 mio., but something around 1048560 rows.