1. This site uses cookies. By continuing to use this site, you are agreeing to our use of cookies. Learn More.

How does Google know you've updated your site?

Discussion in 'White Hat SEO' started by Water damage, Jul 23, 2011.

  1. Water damage

    Water damage Junior Member

    Joined:
    Nov 8, 2009
    Messages:
    173
    Likes Received:
    39
    How does Google know you've updated your site? Do they go off the sitemap timestamp or do they really keep a copy of your webpages and compare them. That seems like a hell of a lot of processing to me and seems like a bit too much work.

    So if they only go off the timestamp of the sitemap, can't you just keep re-submitting new sitemaps with current timestamps?

    Or do they have some sort of algorithm that takes a few things into account, like size of files, number of pics, number of videos etc and just compares the two values of the algorithm to see if they've changed?

    Just got me thinking....
     
  2. fanthomas

    fanthomas Registered Member

    Joined:
    Jan 25, 2010
    Messages:
    98
    Likes Received:
    31
    im pretty sure anytime it finds your link it will recrawl.

    for authority sites its a daily thing.
     
  3. flad051ja

    flad051ja Newbie

    Joined:
    Nov 6, 2009
    Messages:
    6
    Likes Received:
    0
    Use sitemap!
     
  4. purewealthinc

    purewealthinc Regular Member

    Joined:
    May 3, 2010
    Messages:
    427
    Likes Received:
    383
    Occupation:
    Web Fishing
    Location:
    World Wide Web City
    Ofcourse through Crawling -

    This is the process by which Googlebot discovers new and updated pages to be added to the Google index.
    Google bot use a huge set of computers to fetch (or "crawl") billions of pages on the web.

    The program that does the fetching is called Googlebot (also known as a robot, bot, or spider).
    Googlebot uses an algorithmic process: computer programs determine which sites to crawl, how often,
    and how many pages to fetch from each site.

    Google's crawl process begins with a list of web page URLs,
    generated from previous crawl processes, and augmented with Sitemap data provided by webmasters.

    As Googlebot visits each of these websites it detects links on each page and adds them to its list of pages to crawl.

    New sites, changes to existing sites, and dead links are noted and used to update the Google index.


    "If you hit thanks button.. i would highly appreciate that"
     
  5. Water damage

    Water damage Junior Member

    Joined:
    Nov 8, 2009
    Messages:
    173
    Likes Received:
    39
    Sorry, my original question wasn't very clear. Of course Google detects the websites and new pages by crawling - I understand that. My question is what do they use to determine whether a specific URL has been updated or changed it's content (without adding new pages). So if I have a page www.xyz.com/abc and add 2 more videos, 3 pics and a paragraph on abc marketing what does Google do to determine that the content has been updated?
     
  6. daymond

    daymond Junior Member Premium Member

    Joined:
    May 24, 2011
    Messages:
    185
    Likes Received:
    40
    gg have cache of each website in their servers, it compares your website with the cache and checks id something changed, if yes gg will learn that your website is sometimes updated and there is higher chance it will be made in future and ggbot will come back sooner. So with constantly updating content ggbot can crawl you website every day or few times a day, this is why forums, blogs are indexed very well
     
  7. SahL

    SahL Elite Member

    Joined:
    Jan 8, 2011
    Messages:
    1,594
    Likes Received:
    1,296
    Occupation:
    ★SEO expert and ★Sexpert
    Location:
    Bombay
    Home Page:
    Its crawls it. if new content found , it updates it. simple as that
     
  8. purewealthinc

    purewealthinc Regular Member

    Joined:
    May 3, 2010
    Messages:
    427
    Likes Received:
    383
    Occupation:
    Web Fishing
    Location:
    World Wide Web City

    Mate,

    i think your second question has already been answered..

    Okay let's recall again

    Sorry, my original question wasn't very clear. Of course Google detects the websites and new pages by crawling -
    I understand that.

    Your 2nd Question:

    My question is what do they use to determine whether a specific URL has been updated or changed it's content (without adding new pages).

    A: Googlebot visits your websites and it detects links on each page or even your single page like www.xyz.com/abc and adds them to its list of pages to crawl.

    So if I have a page www.xyz.com/abc and add 2 more videos, 3 pics and a paragraph on abc marketing

    what does Google do to determine that the content has been updated?

    A: Of course Google bot use a huge set of computers to determine the content you website when you have made changes or update and fetch (or "crawl") billions of pages on the web. Google's crawl process begins with a list of web page URLs, generated from previous crawl processes, and augmented with Sitemap data provided by webmasters.

    Googlebot uses an algorithmic process: computer programs determine which sites to crawl, how often, and how many pages to fetch from each site.