The same way you clean up your house before your guests arrive, the same way you should get your website ready for Google’s crawler, as this is one of the most important guests you will ever have. According to that, here are 10 things you should double check before submitting your website to the index. If you want, you can view this article as the correction of the top 10 mistakes made by webmasters.
1. If you have a splash page on your website, make sure you have a text link that allows you to pass it.
I’ve seen many websites with a fancy flash introduction on the index and no other way to navigate around it. Well, Google can’t read into your flash page, and therefore it cannot bypass it. All you have to do is put a text link to your website’s second index, and the deed is done.
2. Make sure you have no broken links
I know this is kind of obvious, but you’ll be surprised to find out how many errors is the Google crawler experiencing daily due broken links. Therefore, you’d better check and double check every internal link of your webpage before submission. Don’t forget that your links are also your visitor’s paths to your content. It’s not all about Google, you know :)
3. Check the TITLE tags
Since you are able to search in title tags on Google and since the title tags is displayed in the top of your browser window, I’d say this is an important aspect you need to check. This doesn’t mean you have to compile a >20 keywords list there. Instead, make it a readable sentence since it’s viewable by both crawlers and surfers.
4. Check the META tags
Rumors about Google not caring about META tags are not 100% correct. Google relies on these tags to describe a site when there’s a lot of navigation code that wouldn’t make sense to a human searcher, so why not make sure you’re all in order and set up some valid KEYWORDS and a valid DESCRIPTION. You never know.
5. Check your ALT tags
The ALT tags are probably the most neglected aspect of a website since no one bothers to put them in order. It’s definitely a plus if you do, so Google spider can get a clue about all of your graphics. However, don’t go extreme and start explaining in an ALT tag that a list bullet is a list bullet.
6. Check your frames
If you use frames on your website, you might not be indexed 100%. Google actually recommends that you read an article of Danny Sullivan on Search Engines and Frames. You have to make sure that either Google can read from your frames, either that it has an alternative, defined via the NOFRAMES tag.
7. Do you have dynamically generated pages?
I know the web evolved so much in the last period of time, and more and more websites based on dynamic scripting languages (PHP, ASP, etc) are coming out every second, but Google said they are limiting the amount of dynamic webpages they’re indexing. It’s not too late to consider a compromise and include some static content in your pages. It helps.
8. Update your content regularly
This is an important aspect that you should consider, since Google indexes more quickly pages that get updated on a regular basis. You will notice that the number of pages indexed by the search engine will increase day by day if you update, but will stagnate or decrease if you don’t bring something new. I suggest setting up a META option in the header that will tell Google how frequently should it come back for a reindexing.
9. The robots.txt
This file is a powerful resource if used properly. You have the chance to filter out the bots that crawl your website, and you have the chance of restricting access to certain URL’s that should not be indexed (login pages, admin backends, etc).
10. To cache or not to cache?
Google caches some webpages for quick access, but some webmasters do not like that. The procedure is quite simple. All you have to do is write a line of code between your HEAD tags.
META NAME=”ROBOTS” CONTENT=”NOARCHIVE” - should be enough to stop all robots from caching and archiving the page where the code is embedded.
All these being said, you can now submit your website to Google’s index.