- crawling issues (you are risking not getting all of your content crawled)
– indexing issues (you are risking not getting all of your content indexed)
– broken website navigation and useless address bar (because all pages load under the same URL)
– useless back/next/reload buttons
– potential cloaking issues – which may be unintentional – where the user sees different content than the crawler
Here’s the difference in what the crawler can see without AJAX and can’t see with AJAX *:
Yet AJAX can and should be search engine friendly and following these simple best practices will get you there (or at least much closer to where you should be):
1) Decide where your website ends and your application starts: it does not make sense to make spiders crawl your drafts, but the documents made public and archived instead
2) Make sure you have a URL for each “page” you want to get crawled and indexed
3) Load the basic content containing the keywords at the beginning in a non-dynamic way
4) Cache dynamic pages and serve them as static ones, you could even take snapshots of “pages” which content changes all the time and publish the snapshots
5) Menu items, links and other crucial structural aspects of a site should work without AJAX and/or JavaScript
6) Do not use AJAX for effects but for extra functionality that truly enhances the user experience
Some additional tips:
Most people trigger their basic AJAX through “onClick” events on a given element. In this case the spider will not floor to the AJAX generated content. Another way to do it is using a regular anchor tag (
No comments:
Post a Comment