Technical SEO are implementations which are done onsite. This builds a strong foundation for your content to rank for relevant keywords and phrases.
Here's a checklist on 5 basic but often overlooked, technical SEO aspects for Google to correctly find, crawl and index your webpages.
This configuration file informs web crawlers on which page or group of webpages is disallow for crawl or visit. By default, anything without
Disallowinstruction is considered as
Allow. The instructions can be for directory or url and can be ignored by the malware robots or similar.
// inform all bots NOT to crawl the specified url destination User-agent: * Disallow: /dir/url.html Disallow: /dir/img.png
Tells the search engines on whether to index your website or webpage and list in its public search results. The
noindexvalue can also be used for content with time expiry.
// do not show in search results but evaluate the links within meta name="robots" content="noindex, follow"
The preferred single URL destination for all URL variants of the same page.
rel=canonicalattribute to inform search engines like Google the given page is a copy of the specified original webpage. Alternatively, use 301 permanent redirect.
HTTPSare considered as 2 different URL by search engines. Other instance - prefix
WWWand naked url, trailing
/and without; are also considered differently.
Though duplicate content caused by different url versions does not directly hurt SEO, it wastes precious crawl budget.
The hreflang attribute informs search engine the language of your webpage. Only applicable if your website has similar content but in different languages.
link rel="alternate" href="example.com/de/" hreflang="de" // german link rel="alternate" href="example.com/cn/" hreflang="zh-Hans" // simplified Chinese
H1 Heading Element
The most important content element in your page. Any content wrap inside this
H1tag is considered top priority and should be used carefully.
Ideally should only be used once in a single webpage.