Connect with us
Cloak And Track Your Affiliate Links With Our User-Friendly Link Cloaking Tool, Try It Free

SEO

What Are They And Why Do They Matter?

Published

on

What Are They And Why Do They Matter?

Celebrate the Holidays with some of SEJ’s best articles of 2023.

Our Festive Flashback series runs from December 21 – January 5, featuring daily reads on significant events, fundamentals, actionable strategies, and thought leader opinions.

2023 has been quite eventful in the SEO industry and our contributors produced some outstanding articles to keep pace and reflect these changes.

Catch up on the best reads of 2023 to give you plenty to reflect on as you move into 2024.


There is a lot of confusion about how SEO pros should both understand and, more importantly, leverage “entities” in SEO.

I understand where this comes from, especially with the traditional approach to SEO being around words and phrases.

Indeed, most of the algorithms that the first wave of SEO pros (like me) grew up with had no concept of an “entity” in search. SEO principals – from content writing to anchor text in links to SERPs tracking – were (and largely still are) keyword-driven, and many people still find it hard to understand what has changed.

But over the last decade, all search has been moving towards understanding the world as a string of words and as a series of interconnected entities.

Working with entities in SEO is the foundation for a future-proof search strategy.

They are also important for a future with generative AI and ChatGPT.

This article talks about why. It covers:

  • What are entities?
  • What is the Knowledge Graph?
  • A brief history of entities in search: Freebase, Wikidata, and entities.
  • How entities work and how they are used for ranking.
  • Examples of entities in Google.
  • How to optimize for entities.
  • Using Schema to help define entities.

What Are Entities?

SEOs often confuse entities with keywords.

An entity (in search terms) is a record in a database. An entity generally has a specific record identify.

In Google, that might be:

“MREID=/m/23456” or “KGMID=/g/121y50m4.”

It is certainly not a “word” or “phrase.” I believe that the confusion with keywords stems from two root causes:

  1. The first is that SEO pros learned their craft pre-2010 in terms of keywords and phrases. Many still do.
  2. The second is that every entity comes with a label – which is generally a keyword or descriptor.

So while “Eiffel Tower” might seem like a perfectly identifiable “entity” to us as humans, Google sees it as “KGMID=/m/02j81” and really doesn’t care if you call it “Eiffel Tower,” or “ Torre Eiffel,” or “ایفل بورجو” (Which is Azerbaijan for “Eiffel Tower”). It knows that you are probably referring to that underlying entity in its Knowledge Graph.

This comes on to the next point:

What Is “The Knowledge Graph”?

There are subtle but important differences between “a knowledge graph,” “The Knowledge Graph,” and “The Knowledge Panel.”

  • A knowledge graph is a semi-structured database containing entities.
  • The Knowledge Graph is generally the name given to Google’s Knowledge Graph, although thousands of others exist. Wikidata (itself a knowledge graph) attempts to cross-reference identifiers from different reputable data sources.
  • The Knowledge Panel is a specific representation of results from Google’s Knowledge Graph. It is the pane often showing on the right of the results (SERPs) in a desktop search, giving more details about a person, place, event, or other entity.

A Brief History Of Entities In Search

Metaweb

In 2005, Metaweb started to build out a database, then called Freebase, which it described as an “open, shared database of the world’s knowledge.”

I would describe it as a semi-structured encyclopedia.

It gave every “entity” (or article, to extend the metaphor) its own unique ID number – and from there, instead of a traditional article in words, the system tried to connect articles through their relationships with other ID numbers in the system.

Some $50 million dollars in capital funding, and 5 years later, the project was sold to Google.

No commercial product was ever built, but the foundation was set for a 10-year transition, for Google, from a keyword-based search engine to an entity-based one.

Wikidata

In 2016 – some six years after the purchase – Google formally closed down Freebase because it had migrated and developed the ideas into its own “knowledge graph,” the modern term for these databases.

At that time, it is useful to note that Google publicly stated that it had synced much of its entity data with Wikidata and that, moving forward, Wikidata (which underpins the data used in Wikipedia) was one way in which Google’s Knowledge Graph could interface with the outside world.

How Entities Work And How They Are Used For Ranking

Entities In The Core Algorithm

Entities are primarily used to disambiguate ideas, not to rank pages with the same ideas.

That is not to say that clever use of entities can’t help your site’s content rank more effectively. It can. But when Google tries to serve up results to a user search, it aims first and foremost for an accurate answer.

Not necessarily the most deserving.

Therefore, Google spends considerable time converting text passages into underlying entities. This happens both when indexing your site and when analyzing a user query.

For example, if I type in “The names of the restaurants underneath the Eiffel Tower,” Google knows that the searcher is not looking for “names” or the “Eiffel Tower.”

They are looking for restaurants. Not any restaurant, but ones in a specific location. The two relevant entities in this search are “restaurant” in the context of “Champ de Mars, 5 Av. Anatole France, Paris” (The address of the Eiffel Tower).

This helps Google to decide how to blend its various search results – images, Maps, Google businesses, adverts, and organic web pages, to name a few.

Most importantly, for the SEO pro, it is very important for (say) the Jules Verne restaurant’s site to talk about its spectacular view of the Eiffel Tower if it wants Google to recognize that the page is relevant to this search query.

This might be tricky since the Jules Verne restaurant is inside the Eiffel Tower.

Language Agnostic

Entities are great for search engines because they are language-agnostic. Moreover, that idea means that an entity can be described through multiple media.

An image would be an obvious way to describe the Eiffel Tower since it is so iconic. It might also be a speech file or the official page for the tower.

These all represent valid labels for the entity and, in some cases, valid identifiers in other knowledge graphs.

Connections Between Entities

The interplay between entities allows an SEO pro to develop coherent strategies for developing relevant organic traffic.

Naturally, the most “authoritative” page for the Eiffel Tower is likely to be the official page or Wikipedia. Unless you are literally the SEO pro for the Eiffel Tower, there is little that you can do to challenge this fact.

However, the interplay between entities allows you to write content that will rank. We already mentioned “restaurants” and “Eiffel Tower” – but what about “Metro” and “Eiffel Tower,” or “Discounts” and “Eiffel Tower”?

As soon as two entities come into play, the number of relevant search results drops dramatically. By the time you get to “Discounted Eiffel Tower tickets when you travel by Metro,” you become one of just a tiny selection of pages focusing on the juxtaposition between Metro tickets, Eiffel Tower tickets, and discounts.

Many fewer people type in this phrase, but the conversion rate will be much higher.

It may also prove a more monetizable concept for you! (This example is to explain the principle. I do not know if such discounts exist. But they should.)

This concept can be scaled to create exceptionally strong pages by first breaking all the competing pages for a search phrase into a table showing the underlying entities and their relative importance to the main query.

This can then act as a content plan for a writer to build up a new piece of content that is more authoritative than any of the other competing pieces.

So although a search engine may claim that entities are not a ranking factor, the strategy goes to the heart of the philosophy that “If you write good content, they will come.”

Examples Of Entities In Google

Entities In Image Search

Screenshot from search for [dog on a skateboard], Google, August 2023dog on a skateboard: Google search

Entities can also be very helpful in optimizing images.

Google has worked very hard to analyze images using machine learning. So typically, Google knows the main imagery in most photos.

So take [a dog on a skateboard] as a search term…making sure that your content fully supports the image can help your content be more visible, just when the user is searching for it.

Entities In Google Discover

One of the most underrated traffic sources for SEO professionals is Google Discover.

Google provides a feed of interesting pages for users, even when they are not actively looking for something.

This happens on Android phones and also in the Google app on iPhones. Whilst news heavily influences this feed, non-news sites can get traffic from “Discover.”

How? Well – I believe that entities play a big factor!

Google Discover data in GSCScreenshot from Google Search Console, August 2023Google Discover data in GSC

Don’t be disheartened if you do not see a “Discover” tab in your Google Search Console. But when you do, it can be a welcome sign that at least one of your web pages has aligned with entities enough that at least one person’s interests overlap with your content enough to have the page in a feed targeted specifically to the user.

In the example above, even though “Discover” results are not displayed at the exact time that a user is searching, there is still a 4.2% click-through rate.

This is because Google can align the interests and habits of many of its users to the content on the Internet by mapping entities.

Where a strong correlation occurs, Google can offer up a page for a user.

How To Optimize For Entities

Some Research From A Googler

In 2014, a paper came out that I find quite helpful in demonstrating that Google (or at least, its researchers) were keen to separate out the ideas of using keywords to understand topics vs. using entities.

In this paper, Dunietz and Gillick note how NLP systems moved towards entity-based processing. They highlight how a binary “salience” system can be used on large data sets to define the entities in a document (webpage).

A “binary scoring system” suggests that Google might decide that a document either IS or ISN’T about any given entity.

Later clues suggest that “salience” is now measured by Google on a sliding scale from 0 to 1 (for example, the scoring given in its NLP API).

Even so, I find this paper really helpful in seeing where Google’s research thinks “entities” should appear on a page to “count” as being salient.

I recommend reading the paper for serious research, but they list how they classified “salience as a study of ‘New York Times’ articles.”

Specifically, they cited:

1st-loc

This was the first sentence in which a mention of an entity first appears.

The suggestion is that mentioning the entity early in your web page might increase the chances of an entity being seen as “salient” to the article.

Head-count

This is basically the number of times the “head” word of the entity’s first mention appears.

“Head word” is not specifically defined in the article, but I take it to mean the word concatenated to its simplest form.

Mentions

This refers not just to the words/labels of the entity, but also to other factors, such as referrals of the entity (he/she/it)

Headline

Where when an entity appears in a headline.

Head-lex

Described as the “lowercased head word of the first mention.”

Entity Centrality

The paper also talks about using a variation of PageRank – where they switched out web pages for Freebase articles!

The example they shared was a Senate floor debate involving FEMA, the Republican Party, (President) Obama, and a Republican senator.

After applying a PageRank-like iterative algorithm to these entities and their proximity to each other in the knowledge graph, they were able to change the weightings of the importance of those entities in the document.

Putting These Entity Signals Together In SEO

Without being specific to Google, here, an algorithm would create values for all the above variables for every entity that an NLP or named entity extraction program (NEEP) finds on a page of text (or, for that matter, all the entities recognized in an image).

Then a weighting would be applied to each variable to give a score. In the paper discussed, this score turns into a 1 or 0 (salient or not salient), but a value from 0-1 is more likely.

Google will never share the details of those weightings, but what the paper also shows is that the weightings are determined only after hundreds of millions of pages are “read.”

This is the nature of large language learning models.

But here are some top tips for SEO pros who want to rank content around two or more entities. Returning to the example “restaurants near the Eiffel Tower”:

  • Decide on a “dead” term for each entity. I might choose “restaurant,” “Eiffel Tower,” and “distance” because distance has a valid meaning and article in Wikipedia. Cafe might be a suitable synonym for restaurant, as might “restaurants” in the plural.
  • Aim to have all three entities in the header and first sentence. For example: “Restaurants a small distance from the Eiffel Tower.”
  • Aim in the text to talk about the inter-relationship between these entities. For example: “The Jules-Verne restaurant is literally inside it.” Assuming “it” clearly refers to the Eiffel Tower in the context of the writing, it does not need to be written out every time. Keep the language natural.

Is This Enough For Entity SEO?

No. Probably not. (You are welcome to read my book!) However, not all factors are in your control as a writer or website owner.

Two ideas that do seem to have an impact, though, are linking content from other pages in context and adding schema to help with the definitions.

Using Schema To Help Define Entities

Further clarity might be given to search engines by using the “about” and “mentions” schema to help a search engine disambiguate content.

These two schema types help to describe what a page is talking about.

By making a page “about” one or two entities and “mentions” of maybe a few more, an SEO professional can quickly summarize a long piece of content into its key areas in a way that is ready-made for knowledge graphs to consume.

It should be noted, though, that Google has not expressly stated one way or another whether it uses this schema in its core algorithms.

I would probably add this schema to my article:

<script type=”application/ld+json”> {

“@context”: “https://schema.org”,

“@type”: “WebPage”,

“@id”: “https://www.yoursite.com/yourURL#ContentSchema”,

“headline”: “Restaurants a small distance from the Eiffel Tower”,

“url”: “https://www.yoursite.com/yourURL”,

“about”: [

   {“@type”: “Thing”, “name”: “Restaurant”, “sameAs”: “https://en.wikipedia.org/wiki/Restaurant”},

   {“@type”: “Place”, “name”: “Eiffel Tower”, “sameAs”: “https://en.wikipedia.org/wiki/Eiffel_Tower”}

],

“mentions”: [

   {“@type”: “Thing”, “name”: “distance”, “sameAs”: “https://en.wikipedia.org/wiki/Distance”},

   {“@type”: “Place”, “name”: “Paris”, “sameAs”: “https://en.wikipedia.org/wiki/Paris”}

]

} </script>

The exact choice of schema is as much a philosophical question as an SEO question.

But think of the schema you use as “disambiguating” your content rather than “optimizing your content,” and you will hopefully end up with more targeted search traffic.

Editor’s note: Dixon Jones is the author of Entity SEO: Moving from Strings to Things.

More resources:


Featured Image: optimarc/Shutterstock

Source link

Keep an eye on what we are doing
Be the first to get latest updates and exclusive content straight to your email inbox.
We promise not to spam you. You can unsubscribe at any time.
Invalid email address

SEO

Google Cautions On Blocking GoogleOther Bot

Published

on

By

Google cautions about blocking and opting out of getting crawled by the GoogleOther crawler

Google’s Gary Illyes answered a question about the non-search features that the GoogleOther crawler supports, then added a caution about the consequences of blocking GoogleOther.

What Is GoogleOther?

GoogleOther is a generic crawler created by Google for the various purposes that fall outside of those of bots that specialize for Search, Ads, Video, Images, News, Desktop and Mobile. It can be used by internal teams at Google for research and development in relation to various products.

The official description of GoogleOther is:

“GoogleOther is the generic crawler that may be used by various product teams for fetching publicly accessible content from sites. For example, it may be used for one-off crawls for internal research and development.”

Something that may be surprising is that there are actually three kinds of GoogleOther crawlers.

Three Kinds Of GoogleOther Crawlers

  1. GoogleOther
    Generic crawler for public URLs
  2. GoogleOther-Image
    Optimized to crawl public image URLs
  3. GoogleOther-Video
    Optimized to crawl public video URLs

All three GoogleOther crawlers can be used for research and development purposes. That’s just one purpose that Google publicly acknowledges that all three versions of GoogleOther could be used for.

What Non-Search Features Does GoogleOther Support?

Google doesn’t say what specific non-search features GoogleOther supports, probably because it doesn’t really “support” a specific feature. It exists for research and development crawling which could be in support of a new product or an improvement in a current product, it’s a highly open and generic purpose.

This is the question asked that Gary narrated:

“What non-search features does GoogleOther crawling support?”

Gary Illyes answered:

“This is a very topical question, and I think it is a very good question. Besides what’s in the public I don’t have more to share.

GoogleOther is the generic crawler that may be used by various product teams for fetching publicly accessible content from sites. For example, it may be used for one-off crawls for internal research and development.

Historically Googlebot was used for this, but that kind of makes things murky and less transparent, so we launched GoogleOther so you have better controls over what your site is crawled for.

That said GoogleOther is not tied to a single product, so opting out of GoogleOther crawling might affect a wide range of things across the Google universe; alas, not Search, search is only Googlebot.”

It Might Affect A Wide Range Of Things

Gary is clear that blocking GoogleOther wouldn’t have an affect on Google Search because Googlebot is the crawler used for indexing content. So if blocking any of the three versions of GoogleOther is something a site owner wants to do, then it should be okay to do that without a negative effect on search rankings.

But Gary also cautioned about the outcome that blocking GoogleOther, saying that it would have an effect on other products and services across Google. He didn’t state which other products it could affect nor did he elaborate on the pros or cons of blocking GoogleOther.

Pros And Cons Of Blocking GoogleOther

Whether or not to block GoogleOther doesn’t necessarily have a straightforward answer. There are several considerations to whether doing that makes sense.

Pros

Inclusion in research for a future Google product that’s related to search (maps, shopping, images, a new feature in search) could be useful. It might be helpful to have a site included in that kind of research because it might be used for testing something good for a site and be one of the few sites chosen to test a feature that could increase earnings for a site.

Another consideration is that blocking GoogleOther to save on server resources is not necessarily a valid reason because GoogleOther doesn’t seem to crawl so often that it makes a noticeable impact.

If blocking Google from using site content for AI is a concern then blocking GoogleOther will have no impact on that at all. GoogleOther has nothing to do with crawling for Google Gemini apps or Vertex AI, including any future products that will be used for training associated language models. The bot for that specific use case is Google-Extended.

Cons

On the other hand it might not be helpful to allow GoogleOther if it’s being used to test something related to fighting spam and there’s something the site has to hide.

It’s possible that a site owner might not want to participate if GoogleOther comes crawling for market research or for training machine learning models (for internal purposes) that are unrelated to public-facing products like Gemini and Vertex.

Allowing GoogleOther to crawl a site for unknown purposes is like giving Google a blank check to use your site data in any way they see fit outside of training public-facing LLMs or purposes related to named bots like GoogleBot.

Takeaway

Should you block GoogleOther? It’s a coin toss. There are possible potential benefits but in general there isn’t enough information to make an informed decision.

Listen to the Google SEO Office Hours podcast at the 1:30 minute mark:

Featured Image by Shutterstock/Cast Of Thousands

Source link

Keep an eye on what we are doing
Be the first to get latest updates and exclusive content straight to your email inbox.
We promise not to spam you. You can unsubscribe at any time.
Invalid email address
Continue Reading

SEO

AI Search Boosts User Satisfaction

Published

on

By

AI chat robot on search engine bar. Artificial intelligence bot innovation technology answer question with smart solution. 3D vector created from graphic software.

A new study finds that despite concerns about AI in online services, users are more satisfied with search engines and social media platforms than before.

The American Customer Satisfaction Index (ACSI) conducted its annual survey of search and social media users, finding that satisfaction has either held steady or improved.

This comes at a time when major tech companies are heavily investing in AI to enhance their services.

Search Engine Satisfaction Holds Strong

Google, Bing, and other search engines have rapidly integrated AI features into their platforms over the past year. While critics have raised concerns about potential negative impacts, the ACSI study suggests users are responding positively.

Google maintains its position as the most satisfying search engine with an ACSI score of 81, up 1% from last year. Users particularly appreciate its AI-powered features.

Interestingly, Bing and Yahoo! have seen notable improvements in user satisfaction, notching 3% gains to reach scores of 77 and 76, respectively. These are their highest ACSI scores in over a decade, likely due to their AI enhancements launched in 2023.

The study hints at the potential of new AI-enabled search functionality to drive further improvements in the customer experience. Bing has seen its market share improve by small but notable margins, rising from 6.35% in the first quarter of 2023 to 7.87% in Q1 2024.

Customer Experience Improvements

The ACSI study shows improvements across nearly all benchmarks of the customer experience for search engines. Notable areas of improvement include:

  • Ease of navigation
  • Ease of using the site on different devices
  • Loading speed performance and reliability
  • Variety of services and information
  • Freshness of content

These improvements suggest that AI enhancements positively impact various aspects of the search experience.

Social Media Sees Modest Gains

For the third year in a row, user satisfaction with social media platforms is on the rise, increasing 1% to an ACSI score of 74.

TikTok has emerged as the new industry leader among major sites, edging past YouTube with a score of 78. This underscores the platform’s effective use of AI-driven content recommendations.

Meta’s Facebook and Instagram have also seen significant improvements in user satisfaction, showing 3-point gains. While Facebook remains near the bottom of the industry at 69, Instagram’s score of 76 puts it within striking distance of the leaders.

Challenges Remain

Despite improvements, the study highlights ongoing privacy and advertising challenges for search engines and social media platforms. Privacy ratings for search engines remain relatively low but steady at 79, while social media platforms score even lower at 73.

Advertising experiences emerge as a key differentiator between higher- and lower-satisfaction brands, particularly in social media. New ACSI benchmarks reveal user concerns about advertising content’s trustworthiness and personal relevance.

Why This Matters For SEO Professionals

This study provides an independent perspective on how users are responding to the AI push in online services. For SEO professionals, these findings suggest that:

  1. AI-enhanced search features resonate with users, potentially changing search behavior and expectations.
  2. The improving satisfaction with alternative search engines like Bing may lead to a more diverse search landscape.
  3. The continued importance of factors like content freshness and site performance in user satisfaction aligns with long-standing SEO best practices.

As AI becomes more integrated into our online experiences, SEO strategies may need to adapt to changing user preferences.


Featured Image: kate3155/Shutterstock

Source link

Keep an eye on what we are doing
Be the first to get latest updates and exclusive content straight to your email inbox.
We promise not to spam you. You can unsubscribe at any time.
Invalid email address
Continue Reading

SEO

Google To Upgrade All Retailers To New Merchant Center By September

Published

on

By

Google To Upgrade All Retailers To New Merchant Center By September

Google has announced plans to transition all retailers to its updated Merchant Center platform by September.

This move will affect e-commerce businesses globally and comes ahead of the holiday shopping season.

The Merchant Center is a tool for online retailers to manage how their products appear across Google’s shopping services.

Key Changes & Features

The new Merchant Center includes several significant updates.

Product Studio

An AI-powered tool for content creation. Google reports that 80% of current users view it as improving efficiency.

This feature allows retailers to generate tailored product assets, animate still images, and modify existing product images to match brand aesthetics.

It also simplifies tasks like background removal and image resolution enhancement.

Centralized Analytics

A new tab consolidating various business insights, including pricing data and competitive analysis tools.

Retailers can access pricing recommendations, competitive visibility reports, and retail-specific search trends, enabling them to make data-driven decisions and capitalize on popular product categories.

Redesigned Navigation

Google claims the new interface is more intuitive and cites increased setup success rates for new merchants.

The platform now offers simplified website verification processes and can pre-populate product information during setup.

Initial User Response

According to Google, early adopters have shown increased engagement with the platform.

The company reports a 25% increase in omnichannel merchants adding product offers in the new system. However, these figures have yet to be independently verified.

Jeff Harrell, Google’s Senior Director of Merchant Shopping, states in an announcement:

“We’ve seen a significant increase in retention and engagement among existing online merchants who have moved to the new Merchant Center.”

Potential Challenges and Support

While Google emphasizes the upgrade’s benefits, some retailers, particularly those comfortable with the current version, may face challenges adapting to the new system.

The upgrade’s mandatory nature could raise concerns among users who prefer the existing interface or have integrated workflows based on the current system.

To address these concerns, Google has stated that it will provide resources and support to help with the transition. This includes tutorial videos, detailed documentation, and access to customer support teams for troubleshooting.

Industry Context

This update comes as e-commerce platforms evolve, with major players like Amazon and Shopify enhancing their seller tools. Google’s move is part of broader efforts to maintain competitiveness in the e-commerce services sector.

The upgrade could impact consumers by improving product listings and providing more accurate information across Google’s shopping services.

For the e-commerce industry as a whole, it signals a continued push towards AI-driven tools and data-centric decision-making.

Transition Timeline

Google states that retailers will be automatically upgraded by September if they still need to transition.

The company advises users to familiarize themselves with the new features before the busy holiday shopping period.


Featured Image: BestForBest/Shutterstock

Source link

Keep an eye on what we are doing
Be the first to get latest updates and exclusive content straight to your email inbox.
We promise not to spam you. You can unsubscribe at any time.
Invalid email address
Continue Reading

Trending