Scrapebox Folder Structure Organization Best Practice

Discussion in 'Black Hat SEO' started by Maddrax, Jul 22, 2011.

  1. Maddrax

    Maddrax Registered Member

    Joined:
    Jun 26, 2011
    Messages:
    71
    Likes Received:
    12
    Hey Guys,

    I am using Scrapebox now for a few weeks, or rather days. And I still have some troubles organizing my files well. I tend to loose track of what I harvested when because it's really confusing with all the exporting/importing etc...

    That's why I wanted to ask you experienced Scrapebox users how you organize your file structure?

    I mean I really like structures because I can work more efficient but this time I am having trouble establishing a working one.

    Right now I am having something like this:


    General - general files, such as footprints, common words, advanced operators

    Keywords - Keyword lists I harvest with. How do I keep track of the keywords I already harvested? What if Scrapebox crashes during a harvest how do I know which keywords have already been "crunched"?

    Generally I use a keyword, say, web hosting, get all the suggestion the adword tool gives me and run it through the Scrapebox Keyword Scraper or use Crazyflx's scraping tool. Would there be a better way I am not seeing?

    URLs - Harvested URLs containing folders:
    - Blogs
    - All harvested URLs - does that make sense?
    - Auto Approved
    - Failed Posts
    - Successful Posts
    - URLs for manual postings (Subfolders .edu, .gov, .mil and PR 4+)
    I don't know if I am overengeneering stuff here or if you came up with a simpler structure.

    Would be great to interchange some opinions on that matter.
     
  2. dowser

    dowser Power Member

    Joined:
    Jun 5, 2011
    Messages:
    685
    Likes Received:
    122
    Location:
    canada
    Actually you are better them me, so I better shut up :D

    I have the same problem, it's especially painful if you use a shared computer. I hope to ease the pain a bit with a dedicated computer.

    I'm all ears!
     
  3. Maruk

    Maruk Power Member

    Joined:
    Jun 15, 2009
    Messages:
    562
    Likes Received:
    899
    Hey buddy,
    I got your email. I am not surprised to see you use the same folder structure I do ;)
    Anyway, I have nothing to add here, this is how I run my ops and I don't need anything else.
    I have about 20 urls that I consistently maintain backlinks for so for every url I have a folder like that.
    And of course the Excel sheet.

    Now for your questions:
    1) How do I keep track of the keywords I already harvested? What if Scrapebox crashes during a harvest how do I know which keywords have already been "crunched"?
    Ans: As far as I know there is no dump file for harvesting keywords, perhaps an option for a future update SF?
    Or do you mean when harvesting urls? There is no way to know which keywords have already been used to scrape urls, again, maybe an option for a future update?


    2 )Generally I use a keyword, say, web hosting, get all the suggestion the adword tool gives me and run it through the Scrapebox Keyword Scraper or use Crazyflx's scraping tool. Would there be a better way I am not seeing?
    Ans: That's about right, you can also check out my keywords thread. I have posted millions of keywords on 100+ topics in there.

    Anyway, to make your and perhaps someone else's life easier I uploaded a master folder structure for you to use.
    If you simply save files with dates in them you'll be able to keep great track.
    [Download]
     
    • Thanks Thanks x 1
  4. Maddrax

    Maddrax Registered Member

    Joined:
    Jun 26, 2011
    Messages:
    71
    Likes Received:
    12
    Hey Maruk,

    Thanks for your answer. It's quite hard to put my problems in words :) Do you have like a temp folder where you dump all the files in abd later move it to your final structure?

    I mean surely I understand the folder structure. Could you briefly outline an an example workflow? I think my main question would be how do you NOT get confused with the temp files that one doesn't have to do the same links all over again.

    I think I gotta practice it a little more to get my head around it.

    Thanks again.
     
  5. Kabeer

    Kabeer Registered Member

    Joined:
    Dec 24, 2009
    Messages:
    69
    Likes Received:
    3
    That dude has got a good structure.
     
  6. Maruk

    Maruk Power Member

    Joined:
    Jun 15, 2009
    Messages:
    562
    Likes Received:
    899
    Sure buddy, this is my day to day workflow:

    - Take a list of keywords or generate one (save a newly generated list to the appropriate folder named 'Keywords')
    - Find and test proxies until I have enough to start harvesting from Google (24hr old urls) and Yahoo and AOL (3mnd old urls)
    - Start harvesting (about 150 connections per search engine)
    - Remove dupes
    - Optional: Check for PR and save high PR urls to appropriate folder
    - Remove dupe domains
    - Load all the files (name, email, etc)
    - Start posting
    - When done, check for auto approved urls and save them to appropriate folder
    - Optional: Check auto approved high PR urls and save them to appropriate folder
    - Save all posted to urls to appropriate folder

    Next run, after removing dupe domains I remove any urls I posted to earlier. Import URL List > Select URL lists to compare > *URL text file I saved last time

    If you do this on a regular basis a couple things will happen:
    1) You'll post on fairly new urls only (last 24hrs option)
    2) You'll end up with a list of high PR urls to manually comment on twice a week or so
    3) You'll build a list of auto approve urls for you to use or sell
    4) You keep everything organized (especially with the Excel sheet)

    Hope this makes sense and I hope I didn't forget anything!

    Regards,
    Mark

    [edit]
    Also just for the fun of it:
    [​IMG]
     
    Last edited: Jul 22, 2011