Googlebot queues for the web pages for rendering and crawling. The bot first checks for crawling permissions in the robots.txt file. If the URL is allowed to crawl, Googlebot adds it to the crawling queue.
Avoid ‘OnClick’ links
When a web page contains ‘onclick=” window.location=”‘ links, they are not treated as ordinary links by the search engine bots. In most cases, such links will be skipped entirely as they won’t be counted as part of the navigation or as internal link signals. This can result in poor indexing of your website and affect organic ranking.
Avoid # in URLs
Googlebot does not support the # fragment identifier in the URL structure. This is why you must stick to clean URL formatting.
Unsupported format: example.com/#url Supported format: example.com/url
Unique content must have unique URLs
Simply put, if you are dynamically changing the web page’s content without changing the URL, you are essentially preventing the search engine bots from accessing the unique content. If the page URL remains the same, bots will not crawl it again when dynamically new content is generated.
Avoid JS errors
Don’t block JS in robots.txt
“We recommend making sure Googlebot can access any embedded resource that meaningfully contributes to your site’s visible content or its layout…”
User-agent: * Disallow: /CSS Disallow: /js
Several external prerendering services like Prerender.io can be used to generate a prerendered version of webpages.