1. This site uses cookies. By continuing to use this site, you are agreeing to our use of cookies. Learn More.

Saving a site from the web

Discussion in 'Cloaking and Content Generators' started by bournemarketer, Jul 20, 2017.

  1. bournemarketer

    bournemarketer Registered Member

    Joined:
    Mar 23, 2017
    Messages:
    88
    Likes Received:
    23
    Gender:
    Male
    I have a client that got scammed by a SEO agency. They hired them and signed a year contract and all they did was build a site and then put up a few hundred pages with dupe content. I want to take over and fix the dupe pages with original content.

    The problem is they wont give him the files for the pages except the main pages because they said "thats part of our seo package". I have the images/main html files but not the content pages they added. Is there a program to save them? I don't want to site to go from 200 indexed pages to 8 over night.
     
  2. MisterF

    MisterF Jr. VIP Jr. VIP

    Joined:
    Nov 29, 2009
    Messages:
    6,248
    Likes Received:
    4,740
    Occupation:
    Conference Organiser, Business Advisor.,
    Location:
    JADIP
    Home Page:
    What platform is it built on, WP, HTML, OR... ?

    If it's dupe content, are you planning on keeping it?
     
    • Thanks Thanks x 1
  3. interkul

    interkul Jr. VIP Jr. VIP

    Joined:
    May 21, 2017
    Messages:
    115
    Likes Received:
    31
    Gender:
    Male
    You can download website with a wget.
     
    • Thanks Thanks x 1
  4. bossofthebosses

    bossofthebosses Jr. VIP Jr. VIP

    Joined:
    Feb 7, 2015
    Messages:
    645
    Likes Received:
    247
    Don't u have access to the hosting? Why u need the files from them?
     
  5. StrixMedia

    StrixMedia Jr. VIP Jr. VIP

    Joined:
    Jul 25, 2013
    Messages:
    2,618
    Likes Received:
    356
    Gender:
    Male
    Occupation:
    BHW Seller
    Location:
    In a competitive world
  6. bournemarketer

    bournemarketer Registered Member

    Joined:
    Mar 23, 2017
    Messages:
    88
    Likes Received:
    23
    Gender:
    Male
    It's just HTML files. They put it on their own server and won't give him access. They just keep pushing him to sign a contract for another year. I was planning on taking the 200 pages loaded with dupe content and converting them to 100% unique content instead of just dropping the pages in general.
     
  7. bournemarketer

    bournemarketer Registered Member

    Joined:
    Mar 23, 2017
    Messages:
    88
    Likes Received:
    23
    Gender:
    Male
    httrack worked great, thanks for posting. Thanks for everyone elses comments too, I'm sure some of the other methods would have also worked if I tried them.
     
  8. Black Orange

    Black Orange Junior Member

    Joined:
    Jul 10, 2017
    Messages:
    133
    Likes Received:
    41
    Gender:
    Male
    Occupation:
    Social Media Marketing
    Location:
    Heaven :D
  9. iomatt

    iomatt Newbie

    Joined:
    Apr 23, 2017
    Messages:
    35
    Likes Received:
    1
    Gender:
    Male
    Wget is your best friend for this, then run some html parser to extract what you want.