Google’s URL/Content Removal Tool Now A Wizard
Google has updated their URL removal tool to make it easier and smarter to remove content specifically from third-party web sites…
Matt Cutts: Google Tries To Minimize Search Updates Before Holidays
I am not sure if you noticed but all the Google search results tracking tools flared up yesterday reporting major changes in the Google search results. You can see for yourself at MozCast, SERPs.com…
Wikipedia and SEO: what every digital marketer needs to know
While everyone has heard of, and probably uses, Wikipedia, what fewer understand is the impact it has on search results.
Wikipedia ranks highly in responses to many search queries due to the high degree of trust the search engines have in its content, the breadth of information it contains and the global, multi-language scope of the site.
Our own SEO Ranking Factors Rank Correlation study shows this clearly. Typically Wikipedia ranks second in results for a large number of keywords, for example behind the respective brand or the specific URL for the search term itself.
However, if there is no significant brand that would match the search query, Wikipedia often ranks first. This is particularly true of informational keywords (such as definitions) that are unconnected to brands.
Looking at this graph from google.co.uk illustrates the correlation between Wikipedia URLs and search results:

So how can brands benefit from the correlation between Wikipedia and high search results? I’d say there are four key ways:
-
Keep pages up to dateWikipedia aims to be an independent information source, so over developed pages about your company or products will be frowned upon by the community and potentially removed. However you should ensure that there are up to date, factual pages to ensure that you rank highly in search results.
This is particularly true if you want to appear towards the top of results for informational keywords – make sure you are providing unbiased information that positions you as an expert.
-
Maximise your Google Knowledge Graph informationGoogle’s Knowledge Graph feature aims to enhance search results by providing immediate information on the results page. Normally on the right hand side, it gives either basic details and images/maps on a topic or company, or direct answers to very specific queries, such as What is the population of London?, without a user needing to click on other results.
Wikipedia is one of the major sources of this information – rather than your corporate website. So make sure information on your revenues, senior management, staff and products is always up to date on Wikipedia.
- Use photos within WikipediaKnowledge Graph also trawls Wikipedia for images to accompany results. Therefore make sure you include up to date images (such as your logo) on the site if you want to provide the latest information to browsers when they are searching.
-
Increase trust in your brandAs it has been vetted by the Wikipedia community, information about brands on the site is considered highly trustworthy and independent by consumers. It provides neutral, fact-based content, along with multiple links to specific information which encourage people to find out more.
Not having a Wikipedia page can therefore reduce trust in your brand.
Our Ranking Factors survey shows that while Google favours quality, substance, and ultimately, relevance in ordering its search results it always takes into account secondary factors such as Wikipedia.
The information on the site is interwoven into Knowledge Graph results, meaning it could be the first thing that a prospective customer sees about your company or brand. Therefore you need to ensure you have built the right presence on Wikipedia and are keeping it up to date if you want to benefit from its undoubted power when it comes to search.
One Year Without Google Analytics
In mid-2012 I decided I was fed up with Google Analytics’ bogus data and horrible interface. I replaced the free tool with my own self-served installation of Piwik. Unfortunately I run enough Websites that Piwik put a considerable load on my server. It didn’t help that constant robot probing from hackers and spammers kept knocking the server offline. In late November 2012 I decided to try the “less is more” approach to analytics reporting. Rather than spending hours every day poring over the minutiae of what people were doing on my sites, I decided to try just looking at the high-level data that might be reported by a simpler tool. As many of my sites are now built on WordPress the Jetpack module/plugin came to mind. So I installed Jetpack, connected my sites to a couple of WordPress.com accounts, and proceeded to depend on that dashboard for my personal analytics solution. To say that Jetpack offers fewer details and reporting options than Google Analytics should suffice. People who feel compelled to bury themselves in data every day (even if the data is nonsense) have no reason to look at Jetpack. It was designed for folks who want to focus on […]
Mastering PPC: Improving Conversion Rate
Impressions, Clicks, Cost Per Click, Average Position. These are just some of the 14 or so metrics we report on to our clients on a weekly basis. With so many different things to look at, it’s hard to grasp what is affecting your campaign’s performance short term and long term. Instead of looking at all […]
Creative Calling
Here at Distilled, we’ve always been big believers in the power of content. To this end, we’ve spent the last few years assembling a well-rounded team of designers, developers, data journalists and animators, as well as PR and outreach experts.
SMX West Rates Increase Saturday – Register Now & Save $300!
SMX West Super Early Bird rates – the lowest offered – expire Saturday, December 21st. Register now for an All Access pass and join us in San Jose March 11-13. Here’s what you get: Exceptional content: 60 tactic-packed sessions, inspiring keynotes and clinics Invaluable connections: Meet others…
Please visit Search Engine Land for the full article.
Google Denies A Major Update On December 17th
Yesterday, most of the Google search results tracking tools showed major changes in rankings within Google’s search results but during my analysis of webmaster communication within the discussion forums and social space, I saw very little discussion around changes in rankings at Google. Then,…
Please visit Search Engine Land for the full article.
A New Direction for SEO in 2014: The Secure Search Manifesto
Adapting to the epic changes and shift in SEO requires adapting to, and focusing on, a new direction. Here are four key truths about secure search you can use as a guide to formulate your SEO strategies, plus several tips and tactics for 2014.
Google Maps Pegman Is Now Santa For Christmas
If you go to Google Maps and look at the Pegman used for browsing street view images, you will see the Pegman is wearing a Santa hat. The Santa Pegman is not new, it was shown to us when Google had a welcome back party for the Pegman in the new version of Google Maps. […]
Please visit Search Engine Land for the full article.
What Will 2014 Bring For Local Search? 6 Predictions
Another year is closing and search marketers are anxiously looking ahead to what is on the horizon for local search in 2014. As opportunity grows for businesses to connect with customers via search, so do the challenges of determining the best approach…
Google’s Matt Cutts: Ideally ccTLDs Should Have IP In Country
This is probably a topic we’ve covered a dozen times here but it is always nice to have an official video from Matt Cutts on the topic.
Matt said…
Google AdSense Back To Square Arrows?
When Google introduced the arrows to the AdSense ads, they came with a colored square background button with an arrow overlaid on the square. Then a year ago…
Did The Authorship Reduction In Google Begin?
In October, Matt Cutts of Google at PubCon announced there would likely be a 15% reduction in authorship being displayed in the Google search results. Since then, I haven’t seen many – I’ve seen some – complaints about authorship going away…
Breaking up responsive design
Over the last couple of weeks I have been dealing with the fine art of CSS. Although that is not my daily business anymore – because I lead the website review team here at Yoast – I really enjoyed mastering SCSS and using that for an actual design. During this field trip, I encountered several…
This post first appeared on Yoast. Whoopity Doo!
6 Reasons Why BuzzFeed Destroys Your Content Strategy
Sure, it’s easy to disparage the content BuzzFeed spews out, but BuzzFeed is doing something that most brands completely fail to do – executing an effective content strategy. BuzzFeed’s content strategy is far, far better than yours. Here’s why.
Did Google Help or Hurt SEO in 2013?
Come, join us for a look back at some of the major SEO issues we faced heading into and during 2013, as well as what Google did for and to the SEO community. Then we’ll gaze into our crystal ball to see what’s likely to come in 2014, and why.
A Guide to all the Google AdWords PPC Settings
An in-depth guide into the various settings within Google AdWords, what they are and how to use them.
Post from Samantha Noble on State of Digital
A Guide to all the Google AdWords PPC Settings
I Am an Entity: Hacking the Knowledge Graph
Posted by Andrew_Isidoro
This post was originally in YouMoz, and was promoted to the main blog because it provides great value and interest to our community. The author’s views are entirely his or her own and may not reflect the views of Moz, Inc.
For a long time Google has algorithmically led users towards web pages based on search strings, yet over the past few years, we’ve seen many changes which are leading to a more data-driven model of semantic search.
In 2010 Google hit a milestone with its acquisition of Metaweb and its semantic database now known as Freebase. This database helps to make up the Knowledge Graph; an archive of over 570 million of the most searched-for people, places and things (entities), including around 18 billion cross-references. A truly impressive demonstration of what a semantic search engine with structured data can bring to the everyday user.
What has changed?
The surge of Knowledge Graph entries picked up by Dr Pete a few weeks ago indicates a huge change in the algorithm. Google has been attempting to establish a deep associative context around the entities to try and understand the query rather than just regurgitate what it believes is the closest result for some time, but this has been focused on a very tight dataset reserved for high profile people, places and things.
It seems that has changed.
Over the past few weeks, while looking into how the Knowledge Graph pulls data for certain sources, I have made a few general observations and have been tracking what, if any, impact certain practices have on the display of information panels.
If I’m being brutally honest, this experiment was to scratch a personal “itch.” I was interested in the constructs of the Knowledge Graph over anything else, which is why I was so surprised that a few weeks ago I began to see this:

It seems that anyone now wishing to find out “Andrew Isidoro’s Age” could now be greeted with not only my age but also my date of birth in an information panel. After a few well-planned boasts to my girlfriend about my new found fame (all of which were dismissed as “slightly sad and geeky”), I began to probe further and found that this was by no means the only piece of information that Google could supply users about me.
It also displayed data such as my place of birth and my Job. It could even answer natural language queries and connect me to other entities like in queries such as: “Where did Andrew Isidoro go to school?”

and somewhat creepily, “Who are Andrew Isidoro’s parents?“.
Many of you may now be a little scared about your own personal privacy, but I have a confession to make. Though I am by no means a celebrity, I do have a Freebase profile. The information that I have inputted into this is now available for all to see as a part of Google’s search product.
I’ve already written about the implications of privacy so I’ll gloss over the ethics for a moment and get right into the mechanics.
How are entities born?
Disclaimer: I’m a long-time user of and contributor to Freebase, I’ve written about its potential uses in search many times and the below represents my opinion based on externally-visible interactions with Freebase and other Google products.
After taking some time to study the subject, there seems to be a structure around how entities are initiated within the Knowledge Graph:
Affinity
As anyone who works with external data will tell you, one of the most challenging tasks is identifying the levels of trust within a data-set. Google is not different here; to be able to offer a definitive answer to a query, they must be confident of its reliability.
After a few experiments with Freebase data, it seems clear that Google are pretty damn sure the string “Andrew Isidoro” is me. There are a few potential reasons for this:
- Provenance
To take a definition from W3C:
“Provenance is information about entities, activities, and people involved in producing a piece of data or thing, which can be used to form assessments about its quality, reliability or trustworthiness.”
In summary, provenance is the ‘who’. It’s about finding the original author, editor and maintainer of data; and through that information Google can begin to make judgements about their data’s credibility.
Google has been very smart with their structuring of Freebase user accounts. To login to your account you are asked to sign in via Google; which of course gives the search giant access to your personal details, and may offer a source of data provenance from a user’s Google+ profile.
Freebase Topic pages also allow us to link a Freebase user profile through the “Users Who Say They Are This Person” property. This begins to add provenance to the inputted data and, depending on the source, could add further trust.
- External structured data
Recently an area of tremendous growth in material for SEOs has been structured data. Understanding the schema.org vocabulary has become a big part of our roles within search but there is still much that isn’t being experimented with.
Once Google crawls web pages with structured markup, it can easily extract and understand structured data based on the markup tags and add it to the Knowledge Graph.
No property has been more overlooked in the last few months than the sameAs relationship. Google has long used two-way verification to authenticate web properties, and even explicitly recommends using sameAs with Freebase within its documentation; so why wouldn’t I try and link my personal webpage (complete with person and location markup) to my Freebase profile? I used a simple itemprop to exhibit the relationship on my personal blog:
<link itemprop="sameAs" href="<a href="http://www.freebase.com/m/0py84hb" >http://www.freebase.com/m/0py84hb</a>">Andrew Isidoro</a>
Finally, my name is by no means common; according to howmanyofme.com there are just 2 people in the U.S. named Andrew Isidoro. What’s more, I am the only person with my name in the Freebase database, which massively reduces the amount of noise when looking for an entity related to a query for my name.
Data sources
Over the past few months, I have written many times about the Knowledge Graph and have had conversations with some fantastic people around how Google decides which queries to show information panels for.
Google uses a number of data sources and it seems that each panel template requires a number of separate data sources to initiate. However, I believe that it is less an information retrieval exercise and more of a verification of data.
Take my age panel example; this information is in the Freebase database yet in order to have the necessary trust in the result, Google must verify it against a secondary source. In their patent for the Knowledge Graph, they constantly make reference to multiple sources of panel data:
“Content including at least one content item obtained from a first resource and at least one second content item obtained from a second resource different than the first resource”
These resources could include any entity provided to Google’s crawlers as structured data, including code marked up with microformats, microdata or RDFa; all of which, when used to their full potential, are particularly good at making relationships between themselves and other resources.
The Knowledge Graph panels access several databases dynamically to identify content items, and it is important to understand that I have only been looking at initiating the Knowledge Graph for a person, not for any other type of panel template. As always, correlation ≠causation; however it does seem that Freebase is a major player in a number of trusted sources that Google uses to form Knowledge Graph panels.
Search behaviour
As for influencing what might appear in a knowledge panel, there are a lot of different potential sources that information might come from that go beyond just what we might think of when we think of knowledge bases.
Bill Slawski has written on what may affect data within panels; most notably that Google query and click logs are likely being used to see what people are interested in when they perform searches related to an entity. Google search results might also be used to unveil aspects and attributes that might be related to an entity as well.
For example, search for “David Beckham”, and scan through the titles and descriptions for the top 100 search results, and you may see certain terms and phrases appearing frequently. It’s probably not a coincidence that his salary is shown within the Knowledge Graph panel when “David Beckham Net Worth” is the top auto suggest result for his name.
Why now?
Dr Pete wrote a fantastic post a few weeks ago on “The Day the Knowledge Graph Exploded” which highlights what I am beginning to believe was a major turning point in the way Google displays data within panels.
However, where Dr Pete’s “gut feeling is that Google has bumped up the volume on the Knowledge Graph, letting KG entries appear more frequently,” I believe that there was a change in the way they determine the quality of their data. A reduction in affinity threshold needed to display information.
For example, not only did we see an increase in the number of panels displayed but we began to see a few errors in the data:

This error can be traced back to a rogue Freebase entry added in December 2012 (almost a year ago) that sat unnoticed until this “update” put it into the public domain. This suggests that some sort of editorial control was relaxed to allow this information to show, and that Freebase can be used as a single source of data.
For person-based panels, my inclusion seems to show a new era of Knowledge Graph that Dr Pete reported a few weeks ago. We can see that new “things” are being discovered as strings then, using data, free text extraction and natural language processing tools, Google is able to aggregate, clean, normalize and structure information from Freebase and the search index, with the appropriate schema and relational graphs, to create entities.
Despite the brash headline, this post is a single experiment and should not be treated as gospel. Instead, let’s use this as a chance to generate discussion around the changes to the Knowledge Graph, for us to start thinking about our own hypotheses and begin to test them. Please leave any thoughts or comments below.
Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you don’t have time to hunt down but want to read!
Facebook Likes, Shares Don’t Impact Google Search Rankings [Study]
Eric Enge conducted two studies on the impact of Facebook activity on SEO, and found no clear-cut evidence that Google uses Likes or shares for discovery, indexing, or ranking. But Google can figure out who you’re friends with and can index posts.
