1. This site uses cookies. By continuing to use this site, you are agreeing to our use of cookies. Learn More.

Scrapebox Question

Discussion in 'Black Hat SEO' started by datadyne, Aug 4, 2011.

  1. datadyne

    datadyne Junior Member

    Joined:
    Feb 12, 2011
    Messages:
    170
    Likes Received:
    78
    I know there is a guide out there, and I DID USE the search button but I couldn't figure one thing out....


    How do I extract all links from one domain?
    I have a list of 900 autoapprove domains and I want to get the the rest of the urls from those 900 domains.

    Thanks in advance guys!
     
  2. Roparadise

    Roparadise BANNED BANNED

    Joined:
    May 25, 2011
    Messages:
    786
    Likes Received:
    1,417
    You could use xenu to get all the urls from a domain,im not sure how it can be done n scrapebox.
     
  3. SuperLinks

    SuperLinks Elite Member

    Joined:
    Jul 14, 2008
    Messages:
    2,903
    Likes Received:
    847
    Location:
    New York
    Xenu would be the best way as long as you don't get banned by the domain for "ddos" protection.

    Unfortunately getting all the URLs via Scrapebox requires using the Scrapebox search engine functionality which can cause some headaches.

    Instead of doing a normal "scrape" session for keywrods, use the following to find the internal pages.

    site:domain.com

    Make sure that you aren't using footprints with that, otherwise you won't find all the URLs of that site. This should be a straight "search" within Scrapebox with no footprints.
     
    • Thanks Thanks x 1
  4. Tenshisendo

    Tenshisendo Registered Member

    Joined:
    Nov 20, 2010
    Messages:
    64
    Likes Received:
    24

    You can do it this way to find all indexed links.

    Or you can go to addons and use the link extractor and set it to internal. This will find all links on the sites indexed or not. I suggest to run all your links through a couple times to find all of them.

    Ex. Run your first pass save the list and load that list back in and run again--so on and so forth.
     
    • Thanks Thanks x 1
  5. datadyne

    datadyne Junior Member

    Joined:
    Feb 12, 2011
    Messages:
    170
    Likes Received:
    78
    It didnt work for me... I got 0 results, could you tell me what I did wrong?

    I picked custom footprint and put in site:domain.com
    And for my keywords I put in the 900 autoapprove domains, I got 0 results.
     
  6. dooogen

    dooogen Newbie

    Joined:
    Feb 11, 2010
    Messages:
    14
    Likes Received:
    0
    Cut all the autoapprove urls down to just the domain. Remove duplicates.

    Put site: in front of all the domains and paste them into the keywords spot without any custom footprint.

    Then scrape and you will get all of the indexed pages for each domain. You can also add a comment footprint if you want only commentable pages.
     
  7. hellohellosharp

    hellohellosharp Power Member

    Joined:
    Dec 8, 2010
    Messages:
    625
    Likes Received:
    552
    Occupation:
    CEO @ CLEANFILES LLC
    Location:
    USA
    Home Page:
    Not to steal your thread, but I also have a Scrapebox question...

    Do you guys always trim to root on your harvests? When i harvest I usually end up 5-6 posts per domain...

    I was thinking the best way to do it would be NOT trim to root on the first blast (get multiple posts from each domain) and THEN trim to root and remove duplicates for future use of the list. Is that right?
     
  8. Seo Lover

    Seo Lover Jr. Executive VIP Jr. VIP Premium Member

    Joined:
    Jan 30, 2011
    Messages:
    5,694
    Likes Received:
    4,117
    Gender:
    Male
    Occupation:
    Hanging Around Interwebs !
    Location:
    <-----------------Sin City
  9. Tenshisendo

    Tenshisendo Registered Member

    Joined:
    Nov 20, 2010
    Messages:
    64
    Likes Received:
    24
    What you need to do is trim all the sites to root domain then load them all in the keyword box.

    After that make a txt file with nothing in it but "site:"without quotes.

    Then press the little m next to footprint box and select the txt file you made.

    This will make every domain in your keyword box setup as site:domain.com
     
  10. trevormorley

    trevormorley Junior Member

    Joined:
    Feb 25, 2010
    Messages:
    170
    Likes Received:
    46
    Is there benefit in getting a link from every page on a site rather than just one? Surely big G doesn't count it multiple times?
     
  11. typeslowly

    typeslowly Registered Member

    Joined:
    Nov 30, 2008
    Messages:
    61
    Likes Received:
    9
    Location:
    United States
    I organize by domain, rather than individual URLS. Makes it much easier.
     
  12. TheMatrix

    TheMatrix BANNED BANNED

    Joined:
    Dec 20, 2008
    Messages:
    3,444
    Likes Received:
    7,279
    That's the worst lmgtfy I've ever seen.


    OP: I use the site:domain.com footprint to scrape all G indexed pages.