google search console on laptop
blog post | data + tech | marketing | news + business

You Don’t Have to Be a Webmaster to Use Google’s Webmaster Tools

Walt Irby Team Photo
Walt IrbyDirector of Development

17. Remove URLs From The SERPs, Temporarily

search-console-remove-urls

There are many ways to keep pages on your site from finding their way into SERPs (e.g. noindex, no-follow, sitemap exclusion, robots.txt directives, server authentication, etc.). But sometimes content just sneaks through.

Using the Remove URLs tool, you can (temporarily) remove pages from Google’s index on this page. If you don’t implement any of the aforementioned methods, however, the removal won’t last.

18. Crawl Errors On Your Website

search-console-url-errors

In the Crawl Errors report, you can view and export 404 errors (caused by broken links) detected by GoogleBot (from which, hopefully, you’ll then fix and/or create 301 redirects to resolve).

Fixing 404 errors (especially those that are linked from other websites) is a great way to build more authority for your site and its pages therein.

19. Crawl Stats About Your Website

search-console-crawl-stats

The Crawl Stats report allows you to see how often GoogleBot crawls your site. It’s helpful to put some context behind all that data.

To read more about crawl optimization and why you’re winning if your pages with lower PageRank are frequently crawled, check out AJ Kohn’s post on crawl optimization.

20. Fetch as Google To Test Crawlability

search-console-fetch-as-google

The Fetch As Google tool lets you run a test to ensure GoogleBot can effectively crawl particular pages on your site. Additionally, you can render the pages as GoogleBot to view which elements that might be seen or missed by bots.

Once you fetch a page, you can submit it (and its direct links) to Google’s index, so you don’t have to wait for Google to crawl your site on its own.

21. Robots.txt Tester For Blocking Crawlers

search-console-robots-txt-tester

In addition to noindex and no-follow tags, you can instruct GoogleBot to not index certain site content via a simple text file that resides in your site’s root directory and that follows the syntax described at http://www.robotstxt.org/. It’s a file that explicitly speaks to… you guessed it: robots.

In the robots.txt Tester section of Search Console, you can view the robots.txt file that GoogleBot discovered on your site and you can enter a URL into the test to see if GoogleBot will crawl it or not based on your robots.txt file directives.

22. Sitemaps You Submitted & Indexed

search-console-sitemaps

We’ve reviewed a lot of sections about blocking or removing pages from Google’s index, but what about manually adding pages?

You can submit sitemap files and view sitemaps you’ve already submitted in the Sitemaps section of Search Console. Sitemaps are XML files that you can write or generate using, for example, the Yoast WordPress plugin.

23. URL Parameters For Cleaner Data

search-console-url-params

Developers often append “GET parameters” to URLs, (e.g. example.com?utm_source=google&foo=bar), which can alter the content on the page. Sometimes it’s helpful to instruct GoogleBot what these parameters do to the page, or sometimes we don’t want GoogleBot to index pages with certain URL parameters or certain URL parameter values. The URL Parameters section in Search Console is the exact spot to do that.

24. Security Issues With Your Site’s Content

search-console-security-issues

Search Console’s Security Issues report will recognize most security issues with your site’s content that may prevent the effectiveness of your search listing.

25. Other Resources

search-console-other-resources

In the Other Resources section, Search Console provides several insightful articles for further maintaining your site’s presence in Google’s search results.

Conclusion

In conclusion, use Search Console! It’s free, robust and offered by Google— if there’s one expert source/tool for SEO, it’s going to be the #1 “SE” in the world…

But remember, technical SEO efforts only begin with Search Console. Here are just a few other strategies we at WebMechanix commonly deploy:

  • It’s important to create helpful (and branded) 400-level and 500-level error pages (learn more about all HTTP status codes here).
  • You should always test and optimize page size and load speed (Tools: GTMetrix, Pingdom).
  • Optimize for mobile and other slow-speed situations (see Chrome DevTools’ built-in device emulator and tools like Google’s mobile-friendly test).
  • Comb your site for (and remove) duplicate page content, as that’s a big SEO no-no.
  • Install a (non-self-signed) SSL certificate, which gives you a small boost on SERPs (free SSLs).
  • Before launching a new site or a redesign, go over all the items on this excellent Web Dev Checklist.
About the writer
Walt Irby Team Photo
Walt Irby | Director of Development
Walt has a 15-year track record of completing website redesigns on time with pixel-perfection. He also leads our team of developers to Internet glory.

Most newsletters suck...

So while we technically have to call this a daily newsletter so people know what it is, it's anything but.

You won't find any 'industry standards' or 'guru best practices' here - only the real stuff that actually moves the needle.

You may be interested in:

Using Conversational AI In Your Funnel

Using Conversational AI In Your Funnel

Marketers are trying to up their game, but their old tricks aren’t working. Even emulating competitors is leading to dismal results. What they need is something entirely new. Something no one has tried before. Digital marketing pioneer, Jason Goldsmith, joined WebMechanix CEO, Chris Mechanic, on the Performance Marketing Insiders podcast to emphasize the importance of...
Read this