In this episode of SEO Fairy Tales, Jason Stevens, who works in SEO at Google, shares a story about a product team that reached out for an audit. After conducting the audit, they noticed that the site was getting “no information available for this site” snippets for many of its ranked queries, which was confusing users and impacting traffic. The team found that the site-wide crawling request had dropped to zero, which they determined was due to a disallow directive in the site’s robots.txt file. They were able to solve the issue by removing the directive. The team realized that the site’s traffic had declined because of the issue with the robots file, and the team was able to easily solve the issue by removing the directive.
Resolving rogue robots directives
by admin | Jan 28, 2023 | Other Tips | 0 comments