An XML standard sitemap is totally
different from the HTML standard sitemap. Yoast offer a free SEO plugin to
generate XML sitemap for Wordpress blogs, But Blogger does not support any
internal or external plugin to XML sitemaps. The XML Sitemap is for the search
engines, while the HTML standard sitemap is mostly designed for visitors. In other
words, XML sitemap is a list of your site or blog URLs that you submit to the
search engines.
XML standard sitemap serves two main
purposes:
- This helps search engines to find and crawl your web
pages more easily;
- Search engines can use XML sitemap as a reference
when choosing canonical URLs on your website.
The interesting thing we see on Blogger
and it was not before, is the XML index sitemap URL address (http://www.lablance.com/sitemap.xml)
at the end with the extension .xml(Chrome Browser Only), not atom.xml and tells crawlers
which direction our sitemap, a kind of map with which to help you to find every
corner of our website. It really is a list of all pages of the site (unique
addresses) which also includes some important data such as date of update. With
these data the search engines will know, as an index, such as the organization
of content to crawl more efficiently.
Therefore, submit your XML index sitemap
to various search engines (Google, Bing and Yandex) is an important part of
optimizing your website because not only you will facilitate to easily find all
the pages published, but also be downloaded faster when these change.
One last thinks. If you have many
entries is possible that your sitemap has been divided into several pages and
then not see it directly the relationship of all your URLs. In that case what
you see will be a Sitemap index or ratio of sitemaps partial, within each of
which yes you can see all the published articles:
http://www.lablance.com/sitemap.xml?page=1
http://www.lablance.com/sitemap.xml?page=2
http://www.lablance.com/sitemap.xml?page=3
You see that you also type
the main URL of your blog in the address bar of your browser and add /robots.txt.
What you're seeing is the text file that originally used for web crawlers do
not visit certain pages of your site. Its primary mission is to tell the
spiders which pages can recover (Allow) and what to avoid (Disallow). If you
have not customized your robots.txt from the option Setting>> Search Preferences control panel.
No comments