1. This site uses cookies. By continuing to use this site, you are agreeing to our use of cookies. Learn More.

Scrapebox Folder Structure Organization Best Practice

Discussion in 'Black Hat SEO' started by Maddrax, Jul 22, 2011.

  1. Maddrax

    Maddrax Registered Member

    Joined:
    Jun 26, 2011
    Messages:
    71
    Likes Received:
    12
    Hey Guys,

    I am using Scrapebox now for a few weeks, or rather days. And I still have some troubles organizing my files well. I tend to loose track of what I harvested when because it's really confusing with all the exporting/importing etc...

    That's why I wanted to ask you experienced Scrapebox users how you organize your file structure?

    I mean I really like structures because I can work more efficient but this time I am having trouble establishing a working one.

    Right now I am having something like this:


    General - general files, such as footprints, common words, advanced operators

    Keywords - Keyword lists I harvest with. How do I keep track of the keywords I already harvested? What if Scrapebox crashes during a harvest how do I know which keywords have already been "crunched"?

    Generally I use a keyword, say, web hosting, get all the suggestion the adword tool gives me and run it through the Scrapebox Keyword Scraper or use Crazyflx's scraping tool. Would there be a better way I am not seeing?

    URLs - Harvested URLs containing folders:
    - Blogs
    - All harvested URLs - does that make sense?
    - Auto Approved
    - Failed Posts
    - Successful Posts
    - URLs for manual postings (Subfolders .edu, .gov, .mil and PR 4+)
    I don't know if I am overengeneering stuff here or if you came up with a simpler structure.

    Would be great to interchange some opinions on that matter.
     
  2. dowser

    dowser Power Member

    Joined:
    Jun 5, 2011
    Messages:
    685
    Likes Received:
    122
    Location:
    canada
    Actually you are better them me, so I better shut up :D

    I have the same problem, it's especially painful if you use a shared computer. I hope to ease the pain a bit with a dedicated computer.

    I'm all ears!
     
  3. Maruk

    Maruk Power Member

    Joined:
    Jun 15, 2009
    Messages:
    562
    Likes Received:
    898
    Home Page:
    Hey buddy,
    I got your email. I am not surprised to see you use the same folder structure I do ;)
    Anyway, I have nothing to add here, this is how I run my ops and I don't need anything else.
    I have about 20 urls that I consistently maintain backlinks for so for every url I have a folder like that.
    And of course the Excel sheet.

    Now for your questions:
    1) How do I keep track of the keywords I already harvested? What if Scrapebox crashes during a harvest how do I know which keywords have already been "crunched"?
    Ans: As far as I know there is no dump file for harvesting keywords, perhaps an option for a future update SF?
    Or do you mean when harvesting urls? There is no way to know which keywords have already been used to scrape urls, again, maybe an option for a future update?


    2 )Generally I use a keyword, say, web hosting, get all the suggestion the adword tool gives me and run it through the Scrapebox Keyword Scraper or use Crazyflx's scraping tool. Would there be a better way I am not seeing?
    Ans: That's about right, you can also check out my keywords thread. I have posted millions of keywords on 100+ topics in there.

    Anyway, to make your and perhaps someone else's life easier I uploaded a master folder structure for you to use.
    If you simply save files with dates in them you'll be able to keep great track.
    [Download]
     
    • Thanks Thanks x 1
  4. Maddrax

    Maddrax Registered Member

    Joined:
    Jun 26, 2011
    Messages:
    71
    Likes Received:
    12
    Hey Maruk,

    Thanks for your answer. It's quite hard to put my problems in words :) Do you have like a temp folder where you dump all the files in abd later move it to your final structure?

    I mean surely I understand the folder structure. Could you briefly outline an an example workflow? I think my main question would be how do you NOT get confused with the temp files that one doesn't have to do the same links all over again.

    I think I gotta practice it a little more to get my head around it.

    Thanks again.
     
  5. Kabeer

    Kabeer Newbie

    Joined:
    Dec 24, 2009
    Messages:
    49
    Likes Received:
    2
    That dude has got a good structure.
     
  6. Maruk

    Maruk Power Member

    Joined:
    Jun 15, 2009
    Messages:
    562
    Likes Received:
    898
    Home Page:
    Sure buddy, this is my day to day workflow:

    - Take a list of keywords or generate one (save a newly generated list to the appropriate folder named 'Keywords')
    - Find and test proxies until I have enough to start harvesting from Google (24hr old urls) and Yahoo and AOL (3mnd old urls)
    - Start harvesting (about 150 connections per search engine)
    - Remove dupes
    - Optional: Check for PR and save high PR urls to appropriate folder
    - Remove dupe domains
    - Load all the files (name, email, etc)
    - Start posting
    - When done, check for auto approved urls and save them to appropriate folder
    - Optional: Check auto approved high PR urls and save them to appropriate folder
    - Save all posted to urls to appropriate folder

    Next run, after removing dupe domains I remove any urls I posted to earlier. Import URL List > Select URL lists to compare > *URL text file I saved last time

    If you do this on a regular basis a couple things will happen:
    1) You'll post on fairly new urls only (last 24hrs option)
    2) You'll end up with a list of high PR urls to manually comment on twice a week or so
    3) You'll build a list of auto approve urls for you to use or sell
    4) You keep everything organized (especially with the Excel sheet)

    Hope this makes sense and I hope I didn't forget anything!

    Regards,
    Mark

    [edit]
    Also just for the fun of it:
    [​IMG]
     
    Last edited: Jul 22, 2011