We love digital

Call 0845 485 1219

We love digital - Call and say hello - Mon - Fri, 9am - 5pm

Using Webmaster Tools Search Query Data For SEO Action

Graeme Benge

by Graeme Benge on 6th February 2014

Search QueriesA post from the (ever useful and bookmark-worthy) DejanSEO blog back in January drew to my attention a revamp of the search query data reported on in Google’s Webmaster Tools. Since then I’ve been looking at its functionality and how it can be put to use for SEO purposes, the results of which I’m going to share in this post.

Gone is the rounding of number to the nearest bazillion and specific numbers now greet you. Bearing in mind that both Webmaster Tools and Google Analytics are plainly Google products, the use of Clicks in one tool (WMT) and Visits in another (GA) make me wary of what attribution model is being used to define the metrics on display. There are other  niggles I’ve got that I’ve yet to find further clarification on, maybe you can help me?

My Niggles

  • Comparing the data to rank tracking reports is likely to show disparities as different search configurations mean you’ll not be seeing a like for like comparison.
  • Keep in mind the terminology ‘Query’ does not reflect actual search terms used by users. You may be used to that type of granular data if you regularly use the AdWords Search Query Report but that still remains only available to those happy to buy their keyword data…
  • I’ve found when cross referencing the Click data against the Visits in the Organic Keyword report in Google Analytics different data is on show so take care

So diving in a bit deeper it’s worth noting that you’ve roughly 90 days of data which is a good sample, but nowhere near as comprehensive as Google Analytics clearly. The paranoid android in me has meant I’ve been archiving this data should the generosity be rescinded leaving us with absolutely no keyword friendly data sets to play with.

To view a comparison however you’ll need to limit yourself to up to 30 days’ worth of data. This is still a worthwhile exercise as CTR and Average position variances are worth tracking despite the short data sample demanding it be a regular manual job to add to the list.

Top tip: To display more than 500 rows, hack the URL like this:

URL hack

So that you get something like this:

Max Rows

Priority Queries

Use the star function to effectively ‘favourite’ priority search queries; these being brand terms or key product terms

Using the star filter, isolate those top terms and analyse their performance.

Webmaster Tools Favourites

  • If they have a high average position but low CTR – consider how effective your title and meta description tags are
  • If they have a low average position but high CTR – review existing content, how can this improved further? What sites could do with featuring your content? Use http://ubersuggest.org/ to look at how searchers are using that search query (as in how is that keyword being used). Research those SERPs and look for outreach opportunities

Moving outside of the priority queries, to make the data a bit more manageable you can exclude the queries with less than 10 impressions.

Exclude Low Impression Shares

Statistically, each niche is going to throw up its own positional traffic share. As a guide typically position one accounts for a third of all subsequent clicks with the drop off from 3 to 10 pretty drastic (this post suggests you’ll get over 50% of clicks appearing first in a SERP) so if you want natural traffic to become a significant traffic stream for your site you need to be aiming for those top three spots for the terms that make you money.

It can be a daunting task knowing where to start but I see there being three groups to focus on:

  • Positions 5-10
  • Second pagers
  • New Entries

Positions 5-10

Using the Top Pages tab in the WMT Search Query report you can assume that pages of your site that are sitting in position 5-10 clearly have been deemed authoritative. They combine relevant on page content with quality contextual back links (naive, moi?) so it stands to reason that there is momentum to be capitalised on.

Moving these pages into the top three positions is likely to reap traffic dividends. To make an impact here look at the following:

  • What is your CTR? Do the Meta tags need enhancing? Is your Call To Action strong enough?
  • Analyse those sites sitting above you. Is there content engaging in a different/better way than yours?
  • Take a look at the inbound links. Are there any link sources that could get some value from linking to your assets?

Second Pagers

Those pages languishing in positions between 11-20 are easy to overlook. But if they are duking it out in a set of results amounting to more than 6 figures, the likelihood again is that there is some quality here worth exploring.

Action: Export to CSV, filter by Avg. Position 11-20 and then find some candidates for a bit of TLC.

Average Position Filter

New Entries

Spying terms that don’t normally sit on your list of usual suspects should be a greeted with some glee. These are likely to be tricky to spot and will require a magnifying glass and your list of priority keywords. Spotting any new terms not on that list could mean you’ve found new opportunities.

Here you may be seeing the early benefit from new content you’ve pushed live. It can also mean that there is content enjoying some Hummingbird or LSI  benefits so scrutinize where your target keyword is and its relation to other terms supporting the overall theme of the page. You could very well have a new set of keyword or content opportunities to make the most of.

So there are several ways that the recent Webmaster Tools updates can benefit online marketers. How have you found the revamped data – still shaky or have you been able to successfully use the insight contained there? Let me know in the comments, I’d be very interested in your thoughts.

Graeme Benge

Graeme Benge

With five years background in Travel, Graeme has built up strong commercial experience alongside online and offline marketing skills working with a variety of Travel Agents and Cruise Specialists. A passionate advocate of SEO and Social Media, Graeme has a strong interest in ROI and analytics in order to deliver the best level of returns.

down arrow

Your Free Whitepaper

Getting Started With Search Engine Optimisation

Getting Started With Search Engine Optimisation

Download this whitepaper now and get a new one every month!


  • Pingback: SearchCap: The Day In Search, February 6, 2014

  • SEM 13th February 2014

    Great article, thanks. Btw do you know why google can not include to the index the page that is not closed from indexing in robots.txt, doesnt have noindex tag, has incoming links? Add url doesn’t help too. What can it be?

    Reply to this comment

    • Graeme Benge

      Graeme Benge 13th February 2014

      Thanks for stopping by.

      There could be a variety of reasons why your page isn’t appearing. Noindex tags and a robots exclusion are the obvious but as you’ve eliminated those it would be worth checking Webmaster Tools for:
      -any penalties
      -crawl errors
      -robot.txt errors

      Is it included in your XML sitemap?
      Does a 301 or canonical tag take bots away from that particular page?

      I’d start with those. Let us know how you get on. Good luck!

      Reply to this comment

Subscribe To The Koozai Blog