There are multiple ways to tell search engines how to behave on your site. These are called “crawl directives”. They allow you to:
- tell a search engine to not crawl a page at all;
- not to use a page in its index after it has crawled it;
- whether to follow or not to follow links on that page;
- a lot of “minor” directives.
We write a lot about these crawl directives as they are a very important weapon in an SEO’s arsenal. We try to keep these articles up to date as standards and best practices evolve.
You probably know that Yoast SEO helps you determine what should and should not be indexed. But did you know that it also checks if your site is indexable or not? Thanks to Ryte, we can check if your site is still reachable for both search engine bots and visitors. This is the indexability check.
The robots.txt file is a file you can use to tell search engines where they can and cannot go on your site. Learn how to use it to your advantage!
Must read articles about Crawl directives
Trying to prevent indexing of your site by using robots.txt is a no-go, use X-Robots-Tag or a meta robots tag instead! Here's why.
The canonical URL allows you to tell search engines that certain similar URLs are actually one and the same. Learn how to use rel=canonical!
Want to keep a page out of the search results? Ask yourself if it should be on your site anyways. If it should, use a robots meta tag to prevent it from being indexed.
Search engines need a bit of help to qualify links; use the nofollow, sponsored and UGC attribute to help them out. With Yoast SEO it's easy!
Recent Crawl directives articles
What do you know about bot traffic? Do you know that it affects the environment too? Read on to learn why you should care about bot traffic!
Yoast SEO Premium integrates with the IndexNow protocol to ping search engines like Bing whenever you post content. Why did we launch this?
IndexNow is a new protocol to get your website indexed by search engines. Read on to find out how IndexNow works and how it benefits you!