1. This site uses cookies. By continuing to use this site, you are agreeing to our use of cookies. Learn More.

ScrapeBox Domain Domination - My First "Give Back" to BHW ;)

Discussion in 'Black Hat SEO Tools' started by G0D0VERY0U, May 23, 2010.

  1. G0D0VERY0U

    G0D0VERY0U Regular Member

    Joined:
    Apr 6, 2010
    Messages:
    439
    Likes Received:
    282
    Hey guys,

    I don't think this has been posted here. I've never seen it anywhere so I thought I'd post another little trick I came across with ScrapeBox tonight. I'm always optimizing my usage, a lesson learned with Xrumer - and here's a great little trick to completely rape a blog - think .edu blogs :)

    I use this with blogengine blogs, but I'm sure it would work with everything.

    You will want to scrape, spam, and immediately check for your backlinks as soon as you are done posting. This will give you a list of blogs that autopost comments. Once you have a list of a couple thousand built up, or even if you are just starting to build a list but have a few high PR or EDU domains, you can use this.

    1. Open you Master Blog List or where ever you are saving your successful links. Paste all the URLS in the upper right scrapebox, box.(URL's harvested)

    2. Trim to Root

    3. Remove Duplicate Domains.

    4. Check PR and Filter out Blogs below a certain limit. For me it's PR4 or less. Some are more picky and will do PR6 or better. Others PR2 or better. The less picky you are, the more posts you will get.

    5. Copy all of those domains, and paste them in the keyword box on the left.

    6. Make sure the "Blogengine" is selected if they were BE blogs, or select whatever format the original blogs were in. If you forget, use blog analyzer to find out.

    7. Use proxies and hit "Start Harvesting"

    8. In the Harvested URL's box will appear all of the posts from those domains that are index by the search engines you selected.

    9. Remove duplicates and you should have a HUGE list of nearly 100% auto-posting blog posts to completely CRUSH a domain with backlinks to your sites. It's like being on the sites blog roll in effect.

    This may not be good to use if you are only backlinking to a couple URL's, but if you are like me and backlink to parasites than this is like Gold. I normally get a 10-40 fold increase in the number of posts I can comment on. In other words, 50 blogs produces nearly 500-2000 posts to dominate that are very likely to be auto-approve since that's what you started with. Imagine if you took all 500-1000 domains and did it no matter what their pagerank was. :) If nothing else you pick up posts that the normal harvester left behind.

    If you don't want to comment en' mass - you can then check PR of all of the posts and only comment on all of the PR3 and > posts for like... oh I don't know, .EDU blogs :)

    Lastly, since you are bombing just a single domain, this is one case where using proxies is a DAMN GOOD idea and keep the comments very random.

    Enjoy.
     
    • Thanks Thanks x 41
    Last edited: May 23, 2010
  2. boiler

    boiler Power Member

    Joined:
    Apr 27, 2009
    Messages:
    677
    Likes Received:
    127
    Home Page:
    cant get what you mean :)
     
  3. unkn0wn

    unkn0wn Junior Member

    Joined:
    Nov 18, 2009
    Messages:
    104
    Likes Received:
    20
    Nice tip. Iam going to use it =)
     
  4. willco

    willco Newbie

    Joined:
    Oct 14, 2009
    Messages:
    13
    Likes Received:
    2
    outstanding post! Why the hell didn't I think of this.
     
  5. baldaris

    baldaris Registered Member

    Joined:
    Apr 24, 2009
    Messages:
    76
    Likes Received:
    87
    this is a nice tip...i will definitely work this out...
     
  6. srb888

    srb888 Elite Member

    Joined:
    Jul 30, 2008
    Messages:
    3,260
    Likes Received:
    5,067
    Gender:
    Male
    Occupation:
    WebzSurfer
    Location:
    Sun, Mon, Tue, WTF, Sat!!! :)
    Thinking outside the box, and putting it back into the (scrape) box! Fantastic use of brain-power!! :)

    Rep+ and thanks given for giving your idea to this great community.


    ----
    Why do you say so? The OP is at least trying to give back his best shot/s to this great community, and that's more fun for him than to just take, take, and take and not give...
     
    Last edited: May 23, 2010
  7. markmaster

    markmaster Registered Member

    Joined:
    Apr 5, 2010
    Messages:
    53
    Likes Received:
    4
    about the sorting part (sorting out low Pagerank blogs.) i always wondered, is there some hidden special way to automatically remove all low-PR blogs/links? It is really time consuming to go through a very big list by hand and removing all the entries.. (i know it can be done with external tools, but can it be done with scrapebox?--> everyone acts like it can be done but i have no idea how to do it for a big list / instead of removing manually...)?


    EDIT: ok i just realized how to do it, if icheck for pagerank, and sort out the low-pagerank sites in the pagerank-checker.. :)
     
    Last edited: May 23, 2010
  8. srb888

    srb888 Elite Member

    Joined:
    Jul 30, 2008
    Messages:
    3,260
    Likes Received:
    5,067
    Gender:
    Male
    Occupation:
    WebzSurfer
    Location:
    Sun, Mon, Tue, WTF, Sat!!! :)
    You "Check Pagerank" your harvested list and sort it. Save the list with page-rank (export URL's & PR) to Excel, sort on PR column, and do whatever... It's easy and fast!
     
    Last edited: May 23, 2010
  9. G0D0VERY0U

    G0D0VERY0U Regular Member

    Joined:
    Apr 6, 2010
    Messages:
    439
    Likes Received:
    282
    I totally respect your look on that.

    My point of view is that so many of us have scrapebox anyways that eventually a lot of us are going to overlap comments eventually. It's inevitable.

    And to everyone else, I should be more clear on something - you don't get 100% success ratio on everything you scrape because the harvester will scrape all of the blog's pages, not just the post pages. I just did another run like this and kept just over 90% of all of blog comments though. Any increase to in your overall success ratio will lead to massive time savings and many many more links.
     
  10. GoogleAlchemist

    GoogleAlchemist Regular Member

    Joined:
    Nov 25, 2009
    Messages:
    249
    Likes Received:
    28
    Occupation:
    Bad Ass SEO Consultant
    Location:
    Wherever I want
    Home Page:
    BS Myth

    some cool tests done recently showing # of obl had no effect on so called 'dilutuion' of pr

    my main concern with this (and thanks by the way) is how much weight does a second comment using the same anchor/url have compared to the first one made,let alone the third,fourth etc. while the above obl myth has been debunked there still seems to be evidence that there is a steep drop off after the first link left from a domain going to the same domain using the same anchor text. ie the same domain using different profiles(user/pass) or in this case different blog posts on the same domain going to the same domain with same anchor

    thoughts?

    also, this search method of the root domains gets around the potential problem that using site:url has of not returning many results?
     
  11. G0D0VERY0U

    G0D0VERY0U Regular Member

    Joined:
    Apr 6, 2010
    Messages:
    439
    Likes Received:
    282
    I'm looking for a way to track down a sites sitemap.xml file. If anyone has an idea on how to scrape for them, that would be awesome.

    The secret to taking this to the next step and scraping a blog's self publish directory of posts instead of just getting the one's on the top of goggles index is getting those sitemaps.....

    I would really appreciate a share back on this one if anyone knows how to get or track down a websites sitemap file.... even if we have to do something really random like hope they have an ftp port open that is unsecured - love an app to do even that (and remote host it of course).
     
  12. demoniox

    demoniox Registered Member

    Joined:
    Mar 5, 2007
    Messages:
    98
    Likes Received:
    83
  13. G0D0VERY0U

    G0D0VERY0U Regular Member

    Joined:
    Apr 6, 2010
    Messages:
    439
    Likes Received:
    282
    Have you ever hated yourself for looking past such an evident idea.... Thanks for that - I think I'm going to go hang myself now, lol. I can't believe I didn't even think about that.
     
  14. TheMatrix

    TheMatrix BANNED BANNED

    Joined:
    Dec 20, 2008
    Messages:
    3,444
    Likes Received:
    7,279
    Thanks to OP for the tip. But this step won't work.

    It will return the URLs that mention the site name.

    A better tip is to harvest only BE URLs from your KWs. Then trim, check URL. Then remove unwanted ones. Export as txt. In notepad, replace "http" with "site:http".

    This will return only the URLs from the mentioned domains.

    TM
     
  15. G0D0VERY0U

    G0D0VERY0U Regular Member

    Joined:
    Apr 6, 2010
    Messages:
    439
    Likes Received:
    282
    Humm, I had it working. Let me re-try it. There has been about 8 million updates since I posted this, not that they should change anything but I'll give it a go and see if I get a problematic response from SB...
     
  16. Tighearna

    Tighearna Registered Member

    Joined:
    Apr 21, 2010
    Messages:
    53
    Likes Received:
    19
    Occupation:
    Bokor
    Location:
    739 Bourbon St, New Orleans, LA 70116
    Awesome post thank you, I have been doing something similar however, I was using the sit:url in the custom footprint. From what I'm reading it would be better to do it via a .txt doc and then specify the platform. Am I understanding correctly?
     
  17. Stu784

    Stu784 Regular Member

    Joined:
    Nov 2, 2009
    Messages:
    426
    Likes Received:
    106
    A bit of a newbie question but how do you do this?

    I have SB and see the 'Backlink Checker' but how does this differentiate between what you already have and what you've only just posted?

    Thanks.
     
  18. G0D0VERY0U

    G0D0VERY0U Regular Member

    Joined:
    Apr 6, 2010
    Messages:
    439
    Likes Received:
    282
    I'm not sure I clearly understand the question but I'll do my best to answer it.

    Basically once you post to 10 or 20k blogs, you try to immediately check to see what blogs automatically posted your link using the backlink checker. Then once the backlink check has produced a list of sites that have your link on them, save that list of blogs in a separate file because that's how you are going to build your list of sites to spam. Once you have a good sized list, you can go into it, pull the PR for the main domain of each blog and use let's say, blogs higher than a PR6 and try to get just not one or two of the pages in their blog, but all of the pages in their blog to comment spam on.

    You are taking a blog page on a high PR domain and getting your foot in the door once you found a link that was auto-approved.

    Searching for the rest of the pages on that blog is taking the door (that has your foot in it) and pushing it open to get all of the links and link juice you can out of it.

    I'm working on an update for this post that will describe how to do this using a pages sitemap file instead of depending on the search engines to report back what pages are on the blog. My suspicion is that using the sitemap will unleash a whole new wave of blogs posts that all have semi-low outgoing link counts since the search engines don't normally show those pages in their returned results.

    I'm also doing a comparison of ScrapeBox vs. MPP for blogengine posting. I wanted to include Pyro in it, but their support team wouldn't reply to my emails when I tried to buy a copy to use.
     
  19. Paper-Boy

    Paper-Boy Elite Member

    Joined:
    Jun 17, 2009
    Messages:
    5,116
    Likes Received:
    1,821
    looking forward to this post :)
     
  20. G0D0VERY0U

    G0D0VERY0U Regular Member

    Joined:
    Apr 6, 2010
    Messages:
    439
    Likes Received:
    282
    Well until I get it done, I've been able to scrape enough blogs to begin to look at them all and apparent blog engine calls it's sitemap "sitemap.axd" - so if you want to start playing wit that, feel free.

    I'm still working on everything and I'm not 100% sure of everything yet, but that file seems to be a trend.