Posts

Despite Google being very clear about its feelings on paying for SEO links (hint: it is not a fan), I still regularly come across stories of brands spending hundreds or even thousands of dollars on links that promise to increase their rankings.

Typically, these individuals have heard success stories from others who had recently bought a ton of SEO backlinks and saw their own site jump to the top of search results. Unfortunately, this is rarely the end of the story. 

Today, I wanted to highlight a more complete example of what happens when you pay for links and why.

The Full Story of Someone Who Spent $5,000 on SEO Links

In this instance, I came across someone who had spent thousands of dollars on links for SEO purposes through Search Engine Journal’s “Ask an SEO” column. In the most recent edition of this weekly article, a person named Marlin lays out their situation.

“I paid over $5,000 for SEO link building.”

From the outset, it is unclear if Marlin knew exactly what they had gotten into. While it is possible they directly purchased links from a website, there is also the potential that Marlin and their company put their trust in a questionable marketing agency that purchased or generated spammy links to “boost” rankings.

This is important because it is very common for online SEO packages to include “link building services” which are actually accomplished through link farms that will inevitably be identified and shut down. This is why it is crucial to know that the people handling your link-building efforts use proven, Google-approved strategies rather than cutting corners.

“At first, traffic was boosted.”

As promised, the initial result of buying links is frequently a quick spike in your search engine rankings. Even better, this payoff seems to come much more quickly than the rankings boosts seen from traditional link-building efforts. In some cases, you might even get a huge boost to your rankings within a week or two of paying for the service!

However, the story isn’t over.

“We then lost our rankings on those keywords and our traffic is gone!”

Despite the initially promising results, this is the inevitable conclusion of every story about paying for links.

In the best-case scenario, Google simply ignores your newly acquired low-quality links – putting you right back where you started. In some cases, depending on how widespread the link scheme appears to be, you can wind up even worse than when you began.

If Google believes you have a persistent habit of trying to manipulate search rankings, your site may receive a penalty that significantly impairs your rankings. In the worst cases, your site can be removed from search results entirely.

Why Paid Links Inevitably Fail

There is a very simple reason this story followed a predictable pattern. Google explicitly forbids any sort of “unnatural links” or link schemes. Additionally, the search engine has invested huge amounts of time and resources to identify these artificial links.

At the same time, Google is locked into a game of whack-a-mole where new link sellers are popping up all the time – which is why their links may help your rankings for a very short time.

In SEO, shortcuts are rarely as great as they appear. If you’re looking for long-term, sustainable success, the only option is to roll up your sleeves and build links the old-fashioned way: by creating great content and building real relationships with other members of your industry.

It won’t be quick and it won’t be easy, but it will be worth it in the long run.

Product pages may receive a temporary reduction in their visibility in Google search results if the product is listed as out of stock, according to Google’s Search Advocate John Mueller during the most recent Google Search Central SEO Office Hours session.

Surprisingly, though, this is not always the case.

As Mueller answered questions about how product stock affects rankings, he explained that Google has a few ways of handling out-of-stock product pages.

How Google Handles Out-of-Stock Products

Mueller says that, in most cases, Google treats out-of-stock listings as a soft redirect or unavailable page:

“Out of stock – it’s possible. That’s kind of simplified like that. I think there are multiple things that come into play when it comes to products themselves in that they can be shown as a normal search result.

They can also be shown as an organic shopping result as well. If something is out of stock, I believe the organic shopping result might not be shown – I’m not 100% sure.

And when it comes to the normal search results, it can happen that we when see that something is out of stock, we will assume it’s more like a soft 404 error, where we will drop that URL from the search results as well.

Theoretically, it could affect the visibility in search if something goes out of stock.”

In some situations, though, Google will essentially override this decision and continue to show a page if it is considered particularly relevant for users.

For example, if the product page also includes helpful information about the product in general, it may still be worth keeping in search results despite the lack of stock.

As Mueller explains”

“It doesn’t have to be the case. In particular, if you have a lot of information about that product anyway on those pages, then that page can still be quite relevant for people who are searching for a specific product. So it’s not necessarily that something goes out of stock, and that page disappears from search.”

Out-of-Stock Products Don’t Hurt Your Entire Site

While it is true that listing one product as unavailable can keep that specific page from appearing in search results, Mueller is sure to reassure you that this should not impact the rest of your website:

“The other thing that’s also important to note here is that even if one product goes out of stock, the rest of the site’s rankings are not affected by that.

So even if we were to drop that one specific product because we think it’s more like a soft 404 page, then people searching for other products on the site, we would still show those normally. It’s not that there would be any kind of negative effect that swaps over into the other parts of the site.”

You can watch the entire discussion with Google’s John Mueller in a recording of the SEO Office Hours session below:

Most people these days understand the general idea of how search engines work. Search engines like Google send out automated bots to scan or “crawl” all the pages on a website, before using their algorithms to sort through which sites are best for specific search queries. 

What few outside Google knew until recently, was that the search engine has begun using two different methods to crawl websites – one which specifically searches out new content and another to review content already within its search index.

Google Search Advocate John Mueller revealed this recently during one of his regular Search Central SEO office-hours chats on January 7th.

During this session, an SEO professional asked Mueller about the behavior he has observed from Googlebot crawling his website. 

Specifically, the user says Googlebot previously crawled his site daily when it was frequently sharing content. Since content publishing has slowed on this site, he has seen that Googlebot has been crawling his website less often.

As it turns out, Mueller says this is quite normal and is the result of how Google approaches crawling web pages.

How Google Crawls New vs. Old Content

While Mueller acknowledges there are several factors that can contribute to how often it crawls different pages on a website – including what type of pages they are, how new they are, and how Google understands your site.

“It’s not so much that we crawl a website, but we crawl individual pages of a website. And when it comes to crawling, we have two types of crawling roughly.

One is a discovery crawl where we try to discover new pages on your website. And the other is a refresh crawl where we update existing pages that we know about.”

These different types of crawling target different types of pages, so it is reasonable that they also occur more or less frequently depending on the type of content.

“So for the most part, for example, we would refresh crawl the homepage, I don’t know, once a day, or every couple of hours, or something like that.

And if we find new links on their home page then we’ll go off and crawl those with the discovery crawl as well. And because of that you will always see a mix of discover and refresh happening with regard to crawling. And you’ll see some baseline of crawling happening every day.

But if we recognize that individual pages change very rarely, then we realize we don’t have to crawl them all the time.”

The takeaway here is that Google adapts to your site according to your own publishing habits. Which type of crawling it is using or how frequently it is happening are not inherently good or bad indicators of your website’s health, and your focus should be (as always) on providing the smoothest online sales experience for your customers. 

Nonetheless, it is interesting to know that Google has made this adjustment to how it crawls content across the web and to speculate about how this might affect its ranking process.

To hear Mueller’s full response (including more details about why Google crawls some sites more often than others), check out the video below:

If your site is offline for more than a couple of days you could be at risk of having your pages deindexed, according to Google Search Advocate John Mueller.

It should go without saying that the less downtime your website experiences, the better. Still, some downtime is unavoidable thanks to maintenance, updates, redesigns, and other issues which can be entirely out of your hands.

This inevitably raises the question of exactly how long is too long for your site to be offline. At what point does this begin to hurt your rankings?

After years of debate, we finally have an official answer from Google courtesy of John Mueller during the most recent Google Search Central SEO office hours session.

How Long is Too Long to Be Offline?

The topic arose when an SEO specialist named Aakash Singh asked Mueller what can be done to minimize the loss of rankings or search performance while his client’s website undergoes an expected week of downtime.

The bad news is that a week is simply too long for a site to be offline without experiencing any negative side effects. In fact, Mueller says that sites can start having pages be de-indexed after being down for just a few days.

John Mueller On How Site Downtime Impacts Rankings

Beginning his response, Mueller explains how Google “sees” sites that are experiencing downtime.

“For an outage of maybe a day or so, using a 503 result code is a great way to tell us that we should check back. But after a couple of days we think this is a permanent result code, and we think your pages are just gone, and we will drop them from the index.”

“And when the pages come back we will crawl them again and we will try to index them again. But it’s essentially during that time we will probably drop a lot of the pages from the website from our index, and there’s a pretty good chance that it’ll come back in a similar way but it’s not always guaranteed.”

The general message is that sites should minimize downtime, even when using the proper redirects or site codes.

Mueller does leave us with a suggestion for avoiding the worst fallout from downtime, but he still emphasizes the importance of getting a site back up as quickly as possible:

“… that could be something like setting up a static version of the website somewhere and just showing that to users for the time being. But especially if you’re doing this in a planned way I would try to find ways to reduce the outage to less than a day if at all possible.”

To hear Mueller’s full explanation, check out the recording from the December 10th SEO office hours session below:

A few weeks ago, Google teased that it planned to refine its PageSpeed Insights tools to make data “more intuitive” and easy to understand. Now, that update has arrived.

What Is The PageSpeed Insights Tool?

If you’re unfamiliar, the PageSpeed Insights tool from Google evaluates your web pages to provide suggestions to improve how quickly content loads. 

The tool has been around in various forms since 2013 when it was a simple API webmaster could use to test their page speeds. Version 5, the most recent major update, arrived in 2018. However, smaller updates like this week’s happen somewhat regularly. 

Along with this new update, Google has moved the PageSpeed Insights tool to a new home at https://pagespeed.web.dev/.

What Is New In The PageSpeed Insights Tool?

The biggest focus of the new update is a change to the user interface to be more intuitive by “clearly differentiating between data derived from a synthetic environment and data collected from users in the field.”

To do this, Google has added dedicated sections for each type of data.

Where the tool used to include a label specifying which type of data you were viewing, Google has instead added information about what the data means for you and how it may be used to improve your performance.

Additionally, Google has shifted its emphasis to data collected from real users by moving field data to the top.

The Core Web Vitals assessment has also been expanded, with a label showing if your site has passed a Core Web Vitals assessment in the field and in-depth metrics from simulated environments.

Importantly, the PageSpeed Insights tool also includes details at the bottom of the page specifying how the data was collected in the field. This information includes:

  • Data collection period
  • Visit durations
  • Devices
  • Network connections
  • Sample size
  • Chrome versions

Lastly, Google has removed the previously included screenshot of the page as it indexed your content, replacing it with a series of images displaying the full loading sequence. 

For more, read the announcement for the update from Google’s Web.Dev blog.

Google’s Page Experience Algorithm update is officially coming to some desktop search results, beginning in February of next year.’

Google Search product manager Jeffrey Jose teased this news earlier this year at the annual I/O event. At the time, however, details about when it would be rolled out and how it would be implemented were scarce. Now, we have the full rundown.

What Is Google’s Page Experience Algorithm?

The Page Experience Algorithm was originally rolled out exclusively for searches coming from mobile devices earlier this year, but the search engine confirmed it will be bringing much of the algorithm to desktop searches. This includes the much-talked-about “Core Web Vitals” metrics which are intended to ensure a good user experience on sites.

As the announcement says:

“This means the same three Core Web Vitals metrics: LCP, FID, and CLS, and their associated thresholds will apply for desktop ranking. Other aspects of page experience signals, such as HTTPS security and absence of intrusive interstitials, will remain the same as well.”

However, one notable signal from the mobile Page Experience Algorithm will not be coming to desktop search results for obvious reasons: mobile-friendliness.

To accompany the new search signal, Google says it is working on a new Search Console report dedicated to showing how your desktop pages stack up when this algorithm is applied to them.; For now, the release date of that is unknown, but most believe the report will arrive before or at the same time as the algorithm update.

For more information, read the full announcement here.

It is no secret that Google knows the price you, your competitors, and even the shady third-party companies charge for your products or services. In some cases, you might even directly tell the company how much you charge through Google’s Merchant Center. So, it is reasonable to think that the search engine might also use that information when it is ranking brands or product pages in search results.

In a recent livestream, however, Google Webmaster Trends Analyst, John Mueller, denied the idea.

What John Mueller Has To Say About Price as a Google Ranking Signal

The question arose during an SEO Office-Hours hangout on October 8, which led to Mueller explaining that while Google can access this information, it does not use it when ranking traditional search results.

As he says in the recording of the discussion:

“Purely from a web search point of view, no, it’s not the case that we would try to recognize the price on a page and use that as a ranking factor.

“So it’s not the case that we would say we’ll take the cheaper one and rank that higher. I don’t think that would really make sense.”

At the same time, Mueller says he can’t speak on how products in shopping results (which may be shown in regular search results) are ranked. 

Within shopping search results, users can manually select to sort their results by price. Whether it is used as a factor the rest of the time isn’t something Mueller can answer:

“A lot of these products also end up in the product search results, which could be because you submit a feed, or maybe because we recognize the product information on these pages, and the product search results I don’t know how they’re ordered.

“It might be that they take the price into account, or things like availability, all of the other factors that kind of come in as attributes in product search.”

Price Is And Isn’t A Ranking Factor

At the end of the day, Mueller doesn’t work in the areas related to product search so he really can’t say whether price is a ranking factor within those areas of Google. This potentially includes when they are shown within normal search results pages.

What he can say for sure, is that within traditional web search results, Google does not use price to rank results:

“So, from a web search point of view, we don’t take price into account. From a product search point of view it’s possible.

“The tricky part, I think, as an SEO, is these different aspects of search are often combined in one search results page. Where you’ll see normal web results, and maybe you’ll see some product review results on the side, or maybe you’ll see some mix of that.”

You can hear Mueller’s full response in the recording from the October 8, 2021, Google SEO Office Hours hangout below:

Google has confirmed that it is sometimes replacing page titles in search results with other copy it finds more relevant. As public liaison for Google Search, Danny Sullivan, explains:

“Last week, we introduced a new system of generating titles for web pages. Before this, titles might change based on the query issued. This generally will no longer happen with our new system. This is because we think our new system is producing titles that work better for documents overall, to describe what they are about, regardless of the particular query.”

In plain English, this means that Google is rewriting the title tags accompanying web pages in some search results – often replacing it with other text from your page. This is not the first time Google has made adjustments to title tags being shown in search results, but it is definitely the most extensive rewriting the search engine has done. 

According to Sullivan, the goal of this is to highlight the most relevant content for users and focus on content that users can “visually see”: 

“Also, while we’ve gone beyond HTML text to create titles for over a decade, our new system is making even more use of such text. In particular, we are making use of text that humans can visually see when they arrive at a web page. We consider the main visual title or headline shown on a page, content that site owners often place within <H1> tags, within other header tags, or which is made large and prominent through the use of style treatments.”

Does This Mean HTML Title Tags Don’t Matter?

If Google is going to just replace the tags put on pages, why should we even bother? The answer is for a few reasons. 

Firstly, the title tags will still provide their traditional SEO value by helping the search engine understand your page.

Secondly, Google is not rewriting the majority of search results titles. According to Sullivan, Google will show the original HTML title tags in more than 80% of cases. The system will only revise title tags if it believes the current tags are either too long, stuffed with irrelevant keywords, or a generic boilerplate.

“In some cases, we may add site names where that is seen as helpful. In other instances, when encountering an extremely long title, we might select the most relevant portion rather than starting at the beginning and truncating more useful parts.”

What This Means For You

Since there is no way of opting out of this system, there is nothing for brands to change moving forward. 

The biggest changes from this will instead be in reporting, where some pages may see increased or decreased click-through rates due to changed titles in search results. 

For more, read the full statement from Google and Danny Sullivan here.

One of the most frustrating aspects of search engine optimization is the time it takes to see results. In some cases, you can see changes start to hit Google’s search engines in just a few hours. In others, you can spend weeks waiting for new content to be indexed with no indication when Google will get around to your pages.

In a recent AskGooglebot session, Google’s John Mueller said this huge variation in the time it takes for pages to be indexed is to be expected for a number of reasons. However, he also provides some tips for speeding up the process so you can start seeing the fruits of your labor as soon as possible.

Why Indexing Can Take So Long

In most cases, Mueller says sites that produce consistently high quality content should expect to see their new pages get indexed within a few hours to a week. In some situations, though, even high quality pages can take longer to be indexed due to a variety of factors.

Technical issues can pop up which can delay Google’s ability to spot your new pages or prevent indexing entirely. Additionally, there is always the chance that Google’s systems are just tied up elsewhere and need time to get to your new content.

Why Google May Not Index Your Page

It is important to note that Google does not index everything. In fact, there are plenty of reasons the search engine might not index your new content.

For starters, you can just tell Google not to index a page or your entire site. It might be that you want to prioritize another version of your site or that your site isn’t ready yet. 

The search engine also excludes content that doesn’t bring sufficient value. This includes duplicate content, malicious or spammy pages, and websites which mirror other existing sites.

How To Speed Up Indexing

Thankfully, Mueller says there are ways to help speed up indexing your content.

  • Prevent server overloading by ensuring your server can handle the traffic coming to it. This ensures Google can get to your site in a timely manner. 
  • Use prominent internal links to help Google’s systems navigate your site and understand what pages are most important.
  • Avoid unnecessary URLs to keep your site well organized and easy for Google to spot new content.
  • Google prioritizes sites which put out consistently quality content and provide high value for users. The more important Google thinks your site is for people online, the more high priority your new pages will be for indexing and ranking.

For more about how Google indexes web pages and how to speed up the process, check out the full AskGooglebot video below:

In a Google Search Central SEO session recently, Google’s John Mueller shed light on a way the search engine’s systems can go astray – keeping pages on your site from being indexed and appearing in search. 

Essentially the issue comes from Google’s predictive approach to identifying duplicate content based on URL patterns, which has the potential to incorrectly identify duplicate content based on the URL alone. 

Google uses the predictive system to increase the efficiency of its crawling and indexing of sites by skipping over content which is just a copy of another page. By leaving these pages out of the index, Google’s engine has less chances of showing repetitious content in its search results and allows its indexing systems to reach other, more unique content more quickly. 

Obviously the problem is that content creators could unintentionally trigger these predictive systems when publishing unique content on similar topics, leaving quality content out of the search engine. 

John Mueller Explains How Google Could Misidentify Duplicate Content

In a response to a question from a user whose pages were not being indexed correctly, Mueller explained that Google uses multiple layers of filters to weed out duplicate content:

“What tends to happen on our side is we have multiple levels of trying to understand when there is duplicate content on a site. And one is when we look at the page’s content directly and we kind of see, well, this page has this content, this page has different content, we should treat them as separate pages.

The other thing is kind of a broader predictive approach that we have where we look at the URL structure of a website where we see, well, in the past, when we’ve looked at URLs that look like this, we’ve seen they have the same content as URLs like this. And then we’ll essentially learn that pattern and say, URLs that look like this are the same as URLs that look like this.”

He also explained how these systems can sometimes go too far and Google could incorrectly filter out unique content based on URL patterns on a site:

“Even without looking at the individual URLs we can sometimes say, well, we’ll save ourselves some crawling and indexing and just focus on these assumed or very likely duplication cases. And I have seen that happen with things like cities.

I have seen that happen with things like, I don’t know, automobiles is another one where we saw that happen, where essentially our systems recognize that what you specify as a city name is something that is not so relevant for the actual URLs. And usually we learn that kind of pattern when a site provides a lot of the same content with alternate names.”

How Can You Protect Your Site From This?

While Google’s John Mueller wasn’t able to provide a full-proof solution or prevention for this issue, he did offer some advice for sites that have been affected:

“So what I would try to do in a case like this is to see if you have this kind of situations where you have strong overlaps of content and to try to find ways to limit that as much as possible.

And that could be by using something like a rel canonical on the page and saying, well, this small city that is right outside the big city, I’ll set the canonical to the big city because it shows exactly the same content.

So that really every URL that we crawl on your website and index, we can see, well, this URL and its content are unique and it’s important for us to keep all of these URLs indexed.

Or we see clear information that this URL you know is supposed to be the same as this other one, you have maybe set up a redirect or you have a rel canonical set up there, and we can just focus on those main URLs and still understand that the city aspect there is critical for your individual pages.”

It should be clarified that duplicate content or pages impacted by this problem will not hurt the overall SEO of your site. So, for example, having several pages tagged as being duplicate content won’t prevent your home page from appearing for relevant searches. 

Still, the issue has the potential to gradually decrease the efficiency of your SEO efforts, not to mention making it harder for people to find the valuable information you are providing. 

To see Mueller’s full explanation, watch the video below: