Tag Archive for: Google SEO

Every brand wants to put their best foot forward. If you want to do that online, you need to understand what canonical URLs are. So, today we are going to talk a bit about what canonical URLs are, why your pages may have a canonical version, and how Google chooses which page is the canonical page. 

What Are Canonical URLs?

A canonical URL or web page is the version of a page selected to be indexed by Google when there are multiple versions of the page. 

This version of the page is used by Google to rank the web page and be displayed in search results in order to prevent duplicate search listings. 

As the owner of the website, you have some control over which pages are chosen to be canonical URLs. As we will get into further down, though, Google doesn’t always select the page you believe should be the canonical version.

Before we get to that, let’s take a moment to talk about the legitimate reasons why you may have duplicate versions of a page.

5 Reasons For Having Duplicate Web Pages

According to Google’s official documentation and guidelines about canonical webpages, the search engine believes there are five legitimate reasons a webpage may have multiple versions. 

  1. Region variants: for example, a piece of content for the USA and the UK, accessible from different URLs, but essentially the same content in the same language
  2. Device variants: for example, a page with both a mobile and a desktop version
  3. Protocol variants: for example, the HTTP and HTTPS versions of a site
  4. Site functions: for example, the results of sorting and filtering functions of a category page
  5. Accidental variants: for example, the demo version of the site is accidentally left accessible to crawlers

How Google Chooses A Canonical Webpage 

Until very recently, it was unclear exactly how Google selected canonical pages. Website owners and managers could signal which version they wanted to appear in search results using the rel=”canonical” tag in the code of the page. 

However, this version wasn’t always the one that Google went with.

Gary Ilyes from Google cleared the mystery up (mostly) in a recent Google Search Central video. 

The process starts with finding the content and identifying the main content or “centerpiece of a page”. Then, it groups the pages with similar content in duplicate clusters. 

Then, Google uses a handful of pages to essentially rank each version of the page like it would a listing in search results. The page with the best ranking is selected as the canonical version and included in most search results. 

While he doesn’t list exactly what signals are used, Ilyes did say this:

“Some signals are very straightforward, such as site owner annotations in HTML like rel=”canonical”, while others, like the importance of an individual page on the internet, are less straightforward.”

Notably, this doesn’t mean that Google only indexes one version of the page to be used in all contexts. There are situations where Google may decide to show users a version of the page other than the canonical version.

“The other versions in the cluster become alternate versions that may be served in different contexts, like if the user is searching for a very specific page from the cluster.

To hear Gary Ilyes himself talk about the process, check out the full Google Search Central Video below:

Google is making a big change to its Core Web Vitals ranking signals soon, as the company announced that the new Interaction to Next Paint (INP) signal will replace the First Input Delay (FID) on March 12.

The new INP metric measures the amount of time between when a user interacts with a web page (for example, by clicking a button) to when a browser begins rendering pixels on the screen.

Though FID measured a similar time between user input and browser rendering, Google says INP captures interactivity in ways that were not possible previously.

The History Behind FID and INP Metrics

FID has been a metric used by Google to rank sites since the debut of Google’s Core Web Vitals in 2018. However, Google quickly began to see that this metric didn’t fully capture user interactions as they had hoped. 

This led to Google introducing INP as an experimental or “pending” metric in 2022. Now, almost 2 years later, Google has decided to fully replace FID with the INP metric in March.

What You Should Do

Before March, it is recommended that website managers ensure their site is meeting the threshold for a “good” INP performance. 

If you do not meet this mark, Google suggests optimizing your site with these strategies:

  • Evaluate your site’s performance using tools such as PageSpeed Insights or the Google Chrome User Experience Report.
  • Identify issues that may be slowing down INP, like extended JavaScript tasks, excessive main thread activity, or a large DOM. 
  • Optimize issues based on Google’s optimization guides for the specific issue.

As Google’s ranking algorithms evolve, this and other ranking signals will likely be updated or replaced. This emphasizes how important it is to use the latest optimization standards and to ensure a smooth user experience if you want your business to be easily found online.

The Google SEO Starter Guide is designed to help individuals and organizations quickly learn the most important steps necessary for getting their websites ranking within Google Search. 

While the guide reportedly maintains a 91% approval rating, it has largely gone without updates for several years but that will be changing soon.

In a recent episode of Google’s “Search Off The Record” podcast, the company’s Search Relations team discussed plans to update the SEO Starter Guide, including talking about what would and would not be included in the revised document. 

Discussions like this are great for seeing how SEO is talked about within the search engine and learning what the company prioritizes when ranking sites along with identifying SEO myths that might lead you astray when optimizing your own site. 

So, what’s changing in the revised SEO Starter Guide?

HTML Structure

One topic the group discussed was the importance (or lack thereof) of HTML structure when it comes to online rankings.

While the team agreed that using proper HTML structure can help with online rankings, they indicated the guide will clarify that these are not all that important in the grand scheme.

As Google’s Gary Ilyes said:

“Using headings and a good title element and having paragraphs, yeah, sure. It’s all great, but other than that it’s pretty futile to think about how the page… or the HTML is structured.”

Branded Domain Names vs Keyword Rich Domain Names

SEO experts have been increasingly debating whether it is better to focus on your existing branding when establishing a domain name, or if domains perform better when including specific keywords.

According to the Google team, the new guide will clarify this by indicating that brands should focus on including branding in their domains over using keywords. The thought process shared by those in the discussion was that establishing a memorable brand will have a more long-term impact than trying to optimize your domain specifically for search engines. 

Debunking SEO Myths

Lastly, the group said one thing they want to improve in the document was how it addressed widespread SEO myths and misconceptions. 

For example, everyone agreed that the SEO Starter Guide should specifically debunk the idea that using Google products while creating or optimizing your site will improve search rankings. 

They indicated they would address this myth and several others to prevent people from optimizing their site based on misinformation found elsewhere online. 

For more from the discussion, listen to the entire episode of “Search Off The Record” here.

Google is making some big changes to how it ranks results that aim to deliver more personalized search results and increase the prevalence of “first-hand knowledge”.

The search engine announced the changes earlier this month while spotlighting two specific updates that have recently come to users. 

Cathy Edwards, Vice President of Search at Google, says these updates will better connect humans with the topics and content that are most relevant to their interests and needs:

“Search has always been about connecting human curiosity with the incredible expanse of human wisdom on the net. These advancements will help users find the most helpful information just for them, no matter how specific their questions may be. 

Bringing First-Hand Knowledge To The Surface

Google has made adjustments to its ranking algorithm to show more first-person perspectives higher in search results. While the company didn’t tell us exactly how it tweaked the algorithm, Edwards emphasizes that it will help people find new individual experiences, advice, and opinions when searching. 

With this change, the company says it will hopefully show fewer repetitive pieces of content that don’t bring new perspectives or opinions in the first pages of results. 

The announcement says:

“As part of this work, we’ve also rolled out a series of ranking improvements to show more first-person perspectives in results, so it’s easier to find this content across Search.”

Follow Topics For More Curated Results

Google is giving you the ability to curate your own search results by following topics that are important to you. 

By following topics in search results, such as a favorite football team, style of restaurant, or genre of music, you can stay in touch with these topics naturally while you are searching. 

Follows not only impact what you see in typical search results but help highlight important topics in Discover and other areas of Google.

You can see an example of how this can shape your search results below. The first image shows what search results looked like before this update rolled out, and after.

Like most changes to the search results, however, it is unclear exactly how this affects optimization strategies going forward. We will know more as we get more data in the coming weeks.

Personalization Is The Future

Google has been increasingly customizing search results for users based on numerous factors including location, age, gender, demographics, and more. These latest updates continue this effort to ensure that the search results you see aren’t just the most relevant sites for anyone. They are the most relevant search results for you.

Google’s Search Liaison, Danny Sullivan, raised some eyebrows over the weekend by saying that “major changes” are coming to Google’s search results. 

The statement came during a live talk, where Sullivan reportedly told the crowd to “buckle up” because major changes were on the way.

As the public voice for Google’s Search team, Sullivan is uniquely positioned to speak on what the search engine’s developers are working on behind the scenes. For businesses, this means that he is one of the only people who can give advance notice about upcoming shifts to search results that could impact your online visibility and sales. 

What Did Sullivan Say?

Since it wasn’t livestreamed or recorded, there’s been some discussion about exactly what Sullivan told the crowd. Posts on X agree on a few details though. 

While attendees agree Sullivan specifically used the phrase “buckle up”, a few users provided longer versions of the quote that paint a slightly different picture. 

One person, Andy Simpson, says the entire quote was “There’s so much coming that I don’t want to say to buckle up because that makes you freak out because if you’re doing good stuff, it’s not going to be an issue for you.”

This is likely the case, as Sullivan has since clarified:

“I was talking about various things people have raised where they want to see our results improve, or where they think ‘sure, you fixed this but what about….’ And that these things all correspond to improvements we have in the works. That there’s so much coming that I don’t want to say buckle up, because those who are making good, people-first content should be fine. But that said, there’s a lot of improvements on the way.”

Either way, it is important for businesses to take note of these statements and watch their site’s search results performance for any signs of major shifts in the near future. 

Think using blogs to get to the top of the search engines is a thing of the past? Don’t be so quick to ditch your brand’s blog because a new study suggests that blog posts are the most common type of content found in the top 5 Google search results (excluding homepages). 

Even with low-quality AI-generated blog content on the rise, BrightEdge says that blogs are the leading type of content returned by Google – a strong indication that blogs with well-crafted content are one of the strongest search engine optimization tools available to brands today. 

About The Study

For the study, BrightEdge analyzed results for a dataset of 10,000 keywords of varying intent across 10 specific industries:

  • Banking
  • Insurance
  • Retail
  • Software
  • Higher Education
  • Real Estate
  • Advertising and Marketing
  • Manufacturing
  • Travel and Hospitality
  • Industrial

Using data collected during August of this year, the study then analyzed the content types of 23,785 pages ranking in the top 10 search positions. 

While the leading type of page found in the top search results were homepages, these were excluded because these are essentially the default type of page Google returns when it believes a site may be relevant but it does not know which specific page to recommend.

Once homepages have been accounted for and excluded, the leading type of content in top search results was blog posts – accounting for 19% of the top 10 search results. When you narrow the focus to just the top 5 search results, that climbs to 23% of search results. 

The Takeaway

Many brands have been moving away from traditional brand blogs because of a misguided notion that blogs were becoming irrelevant compared to more interactive or visual media like videos or user-generated content. This trend has only accelerated with the recent surge in lower-quality content pumped out by generative AI systems. 

As Jim Yu, founder of BrightEdge and executive chairman says, however, well-maintained blogs are still an essential tool for raising the visibility of your brand and educating consumers:

“The future is not just AI – it’s AI and human symbiosis. AI can inform and assist, but human creativity, expertise and skill sets are necessary to add the voice and trust of your brand. Success lies in the fusion of AI and human expertise throughout any content creation process,” 

Brands struggling to make progress in this area should likely re-evaluate their content and ensure their strategy is focused on delivering relevant, useful, and interesting information to your target market. 

Typically when a site starts ranking worse for one keyword, the effect is also seen for several of the other keywords it ranks for. So what does it mean when a website only loses rankings for one keyword? According to Google’s Gary Illyes, there are a few reasons a site might experience this rare problem. 

In a recent Google SEO Office Hours episode, Illyes addressed the issue while answering a question from a site owner who had effectively disappeared from the search results for a specific keyword – despite ranking at the top of results consistently in the past. 

The Most Likely Culprit

Unfortunately, the most common cause of an issue like this is simply that competitors have outranked your website, according to Illyes:

“It’s really uncommon that you would completely lose rankings for just one keyword. Usually, you just get out-ranked by someone else in search results instead if you did indeed disappear for this one particular keyword.”

Other Potential Causes

If you believe the drop in rankings for a specific keyword is the result of something other than increased competition, Illyes recommends investigating if the issue is isolated to a specific area or part of a larger ongoing global problem. 

“First, I would check if that’s the case globally. Ask some remote friends to search for that keyword and report back. If they do see your site, then it’s just a ‘glitch in the matrix.’”

Those without friends around the globe can effectively accomplish the same thing by using a VPN to change their search location.

On the other hand, if your site is absent from results around the globe, it may be indicative of a bigger issue – potentially the result of changes to your website:

“If they don’t [find your website], then next I would go over my past actions to see if I did anything that might have caused it.”

Lastly, Gary Illyes offers a few other potential causes of a sudden ranking drop.

Technical issues such as problems with crawling or indexing can prevent your website from appearing in search results. 

Sudden changes to your backlink profile – either through mass disavowing links or through the use of low-quality or spammy links can also trigger issues with Google. If you are hit with a manual penalty for low-quality links, it is highly likely your site will stop ranking for at least one keyword (if not several).

To hear the full discussion, check out the video below:

Google has confirmed it is rolling out its latest broad core algorithm update, signifying yet another potential shake-up for the search engine’s results.

Google’s broad core algorithm updates serve as some of the most significant updates for the search engine compared to the smaller updates that are happening multiple times a day. They can affect rankings for search engine results pages (SERPs) throughout Google’s entire platform.

As is usual with Google, the search company is being tight-lipped about specific details regarding the update, only going so far as to confirm the latest update. The update is also expected to take up to multiple weeks for the full impact to be obvious.

With this in mind, it is wise for brands to take note and monitor their own search performance in the coming weeks.

What Can You Do?

Aside from always striving to provide the best online experience possible with your website, there are a few specific steps you can take to safeguard your site from updates like these:

  • Monitor site performance regularly to identify early signs of issues with your site
  • Create content geared to your audience’s needs and interests
  • Optimize your site’s performance (including speed, mobile-friendliness, and user experience) to ensure your site isn’t off-putting to visitors

TL;DR

Google has launched its latest broad core algorithm update, which could potentially affect rankings for search engine results pages. The update may take several weeks to have full impact, so brands are advised to monitor their search performance. To safeguard your site, monitor its performance regularly, create audience-specific content, and optimize its performance for speed, mobile-friendliness, and user-experience.

If you’re still unclear on how Google thinks about marketing agencies that offer negative SEO linkbuilding services or link disavowal services, the latest comments from John Mueller should help clarify the company’s stance. 

In a conversation that popped up on Twitter between Mueller and several marketing experts, Mueller clearly and definitively slammed companies offering these types of services by saying that they are “just making stuff up and cashing in from those who don’t know better.”

This is particularly notable as some have accused Google of being unclear on their handling of link disavowal using their tools

The post that started it all came from Twitter user @RyanJones who said, “I’m still shocked at how many seos regularly disavow links. Why? Unless you spammed them or have a manual action you’re probably doing more harm than good.”

In response, one user began talking about negative SEO which caught the attention of Mueller. The user mentioned that “agencies know what kind of links hurt the website because they have been doing this for a long time. It’s only hard to down for very trusted sites. Even some agencies provide a money back guarantee as well. They will provide you examples as well with proper insights.”

In response, Mueller gave what is possibly his clearest statement on this type of “service” yet:

“That’s all made up & irrelevant. These agencies (both those creating, and those disavowing) are just making stuff up, and cashing in from those who don’t know better.”

Instead of spending time and effort on any of this, Mueller instead recommended something simple:

“Don’t waste your time on it; do things that build up your site instead.”

Google is encouraging brands to ensure content is properly dated in search engines by using multiple date indicators on each page. 

The recommendation came in the wake of an issue with Google News where the wrong dates were being shown.

In the response, Google’s Search Liaison, Danny Sullivan, emphasized that while many factors may have contributed in this specific situation, the lack of proper date signals made it difficult to show correct info in the search results. 

“That page is a particular challenge since the main story lacks a visible date (it only has a time), and the page contains multiple stories which do contain full dates. Our guidance warns about this.”

To prevent situations like this from arising, Sullivan says it is important to use several signals to clarify the date content is published:

“Understand that ideally, the meta data alone would seem to some to be enough, and we’ll keep working to improve. But there are good reasons why we like multiple date signals present.”

Why Does This Matter?

It may not seem like a big deal for the wrong date to occasionally get shown with content in the search results. However, these can undermine your authority, lead to confusion, and create a poor user experience. All of these can lead to decreased page performance and even demotions in Google’s search results.

On the other hand, situations like this also highlight the need for Google to deliver more consistent ways to signal a page’s publishing date. 

For now, the best recommendation Google has is to use a scattershot approach for the best chance of having your page correctly dated:

“Google doesn’t depend on a single date factor because all factors can be prone to issues. That’s why our systems look at several factors to determine our best estimate of when a page was published or significantly updated.”