DynaBlogger allows you to configure various settings that help improve the search engines performance of your blog.
You can configure the following both for the blog in general and for articles:
- meta title and description: by default these are derived from the normal title and description of the blog or the article, but you can customise these. Search engines limit the length of the visible title and description shown in search results, so if you leave the meta title and description to be generated automatically from title and description of the blog, they will be truncated if too long according to SEO best practices.
- canonical URLs: with Liquid markup you can access the canonical URL of the page. If you add a `rel=canonical` link with the canonical URL for the href attribute, search engines will understand that this is the URL that must be indexed for that specific page. This way, if the same content happens to be reachable with multiple URLs (shouldn't happen on DynaBlogger), you avoid duplicate content issues with search engines.
Out of the box DynaBlogger helps you improve the SEO performance of your websites without requiring complicated plugins like you may have to with with other products.
By default the robots.txt file generated for your blog instructs search engines to not index it. This is because in most cases users want to finish customising themes and other settings before making the websites public. However in the blog settings you can enable indexing.
Archive/tag/author pages are not indexed by default to avoid duplicate content issues, because they might look too similar to each other, and search engines do not like this.
The XML sitemap compatible with search engines is available at the /sitemap.xml path and contains the full list of pages and posts of your blog in XML format.
The robots.txt file also includes a reference to the sitemap, so search engines can discover it easily to index your website.
Webpages often work both with and without a trailing slash, and this also can lead to duplicate content issues with search engines. Because of this, trailish slashes are enforced on all URLs automatically, and a 301 permanent redirect to the version of the URL with the trailing slash is done automatically when the version without the trailing slash is used.