looking for ressources for big data SEO

aussiejack

Regular Member
Joined
Sep 23, 2014
Messages
493
Reaction score
62
Hi folks!

I am working on my first really BIG project. It is a job posting site for a special profession.

Part of it will be an api to other job posting sites. To make it short we will have massive amounts of content.

We are at a point where coding and SEO must go hand in hand to handle massive amounts of content.

My question is if you have any recommendation for this "big data seo"? I am interested in blogs, articles, gurus etc. All tipps are welcome.
 
What you want to do is more general and basic than that. What I do for engagements like that is prove the case by demonstrating what effect content tuning can have on one listing and then back it up with the tuning guidelines to reproduce that success on many pages. It is all empirically driven, but it starts by demonstrating and documenting so you are in a position to give your recommendations some authority within the organization...

You shouldn't write articles that point to job postings... that would be silly. And while resume tips and interview tips articles are useful they stand almost no chance to help your job postings rank... The only reason to write content like that would be to help the domain rank, which has the side effect of promoting awareness and inefficiently targeting job hunters. I think tuning your crawlable job categories and job postings is the best thing you can focus on if you are targeting active job hunters.

My advice is put all you time an energy into boosting one job posting or one regional job category then apply what you learned to all of the others.

I have also noticed that category pages are very easy to rank currently in e-com... the same might be true for job sites.
 
It's the same if you want to rank one article or the whole wikipedia. If Google detects a large portion of your content to be shallow and non-engaging, the Panda hits you. That's probably the only difference I see. If you put out 2 billion pages that nobody wants, your site will get hit.
 
Thanks for advice.

I probably was inprecise in my question. I am especially interested in the technical / coding issues. For example at first we had embedded the job posts in some scripts which could not be crawled by the search engine. Or: When the result pages are created dynamically should an offline version (cache) be created to safe time for the next query? How should the sitemap be created etc?

Any ideas?

What you want to do is more general and basic than that. What I do for engagements like that is prove the case by demonstrating what effect content tuning can have on one listing and then back it up with the tuning guidelines to reproduce that success on many pages. It is all empirically driven, but it starts by demonstrating and documenting so you are in a position to give your recommendations some authority within the organization...

You shouldn't write articles that point to job postings... that would be silly. And while resume tips and interview tips articles are useful they stand almost no chance to help your job postings rank... The only reason to write content like that would be to help the domain rank, which has the side effect of promoting awareness and inefficiently targeting job hunters. I think tuning your crawlable job categories and job postings is the best thing you can focus on if you are targeting active job hunters.

My advice is put all you time an energy into boosting one job posting or one regional job category then apply what you learned to all of the others.

I have also noticed that category pages are very easy to rank currently in e-com... the same might be true for job sites.
 
Yes cache... you might want to consider a graph database like Neo4J... you could also go the SQL database route. Either way you should definitely cache... I would only generate AMP output if I were engineering for today forward.

Break up site maps by page type and potentially catagories if there aren't too many. (and possibly geographies)

a.k.a. site maps like these

US_Engineering_Job_Listings.xml
US_Marketing_Job_Listings.xml
CA_Engineering_Job_Listings.xml
CA_Marketing_Job_Listings.xml
UK_Engineering_Job_Listings.xml
UK_Marketing_Job_Listings.xml

US_Job_Category_Pages.xml
CA_Job_Category_Pages.xml
UK_Job_Category_Pages.xml

Home_And_Support_Pages.xml

Blog_Pages.xml

This way Google Search Console will tell you if there is a problem indexing by particular types or locales of pages.
 
thats the kind of tipp I was looking for! Is there any ressource (blog, ebook, case study etc) where I can learn about these things??

Yes cache... you might want to consider a graph database like Neo4J... you could also go the SQL database route. Either way you should definitely cache... I would only generate AMP output if I were engineering for today forward.

Break up site maps by page type and potentially catagories if there aren't too many. (and possibly geographies)

a.k.a. site maps like these

US_Engineering_Job_Listings.xml
US_Marketing_Job_Listings.xml
CA_Engineering_Job_Listings.xml
CA_Marketing_Job_Listings.xml
UK_Engineering_Job_Listings.xml
UK_Marketing_Job_Listings.xml

US_Job_Category_Pages.xml
CA_Job_Category_Pages.xml
UK_Job_Category_Pages.xml

Home_And_Support_Pages.xml

Blog_Pages.xml

This way Google Search Console will tell you if there is a problem indexing by particular types or locales of pages.
 
Back
Top