Q and A: Why is my site showing as redirected in the Google Search Console?

QuestionHi Kalena,

I just checked my Google Search Console and under Crawl, Fetch as Google I see my site status is showing as redirected, with a yellow sign beside it!

What does that mean and how can I fix it,  please?

Regards
Zara

————————————-

Hi Zara

First up, don’t panic. The *Fetch as Google* tool simply tries to imitate what Googlebot sees when it crawls your site code. It doesn’t mean that Google can’t index your site. If you are still seeing data in the Search Console relating to your site status, all is well. If there were major indexing issues, you would see errors in the Crawl Errors and Site Appearance categories.

As for WHY you are seeing a redirect status, without knowing your site URL, I can only guess. But I’m confident it will be one of these reasons:

1)  Domain canonicalization: Does your site load using both http:// and http://www. or does one version redirect to the other? Have you set a preferred domain in Google Search Console? To do this you need to set up views for both versions in Google Search Console and then set your preference. After you’ve done that, depending on which domain version you open within the Console, the *Fetch as Google* tool will show a different status.

complete-fetch

1) Fetch as Google results for http://www.searchenginecollege.com

2) Fetch as Google results for http://searchenginecollege.com

2) Fetch as Google results for http://searchenginecollege.com

For example, to avoid duplicate content issues in Google, I have set a) http://www.searchenginecollege.com as my preferred domain in my Search Console, but I also have a Console view for b) http://searchenginecollege.com. On the server side, I have set the non www version to redirect to the www version. This is known by several names, including URL redirecting, domain masking, piggy-backing and parking. If I fetch a) as Google, I see the status as shown in the first image. If I fetch b) as Google, I see the yellow *redirected* status as shown in the second image.

This is likely what you’re seeing and it simply means you have set up your domain redirect correctly. Learn more about domain canonicalization and how to set your preferred domain.

2) 301 or 302 redirects: Have you recently switched domains? Although Googlebot follows redirects, the *Fetch as Google* tool does not. So if your site was originally set up in the Google Search Console as one domain e.g. http://www.siteA.com but has now moved to http://www.siteB.com and you set up a 301 or 302 redirect server side, then if you are looking at the original site view in the Console, it will show up as redirected in the crawl tool. You can inspect the HTTP response on the fetch details page to see the redirect details. Learn how to do this.

3) Moving to SSL:  If you have recently updated your site from http:// to https:// and you’re seeing *redirected* in the crawl status, you may have the same domain canonicalization issue as 1). So you need to set up a view for the https:// version of your site in Google Search Console. More info on SSL issues here and here.

Hope this helps!

——————————————————————-

Like to teach yourself digital marketing online? Start here.

 

Spread the joy!

Q and A: How long will it take our site to be purged from Google?

QuestionHi Kalena,

It’s Tim here. I’m the developer for a website – [URL removed for privacy reasons] – and as of Thursday or Friday last week, Google has crawled my whole site. It shouldn’t have been able to do this, but it has.

Part of the site is written in PHP and Google has cached all the pages, several of which contain information that shouldn’t really be in the public domain.

I’ve submitted the FQDN to Google asking them to remove the URL which will hopefully prevent any results being shown in a Google search. The request is currently in a ‘pending’ state and I’m wondering how long this would actually take to be purged.

Thanks,

Tim

————————————-

Hi Tim

I’ve not personally lodged a take down request with Google, so I’m afraid I’m not speaking from experience, however I’ve had colleagues tell me this can take up to 3 months if a site has already been crawled.

Your email doesn’t make it clear what happened, but it may also depend on how sensitive the content is and why it was indexed in the first place.

A couple of things you can do while you’re waiting:

1) If Google managed to crawl your whole site, you might have conflicting instructions in your robots.txt file, your robots meta tags on those pages or you might be including content you don’t want public in your sitemap.xml file that Google is indexing. Check all those areas so the problem doesn’t re-occur.

2) Ask Google to remove content through the Webmaster Search Console. This is often faster than the formal take down request you submitted via email. It requires you to verify ownership / admin of the site via the Search Console first.

Keep in mind that even after you’ve blocked the pages from being indexed, they can take a while to fall off the Google search results, depending on the number of data-centers that have cached them and where they are serving results from.

Best of luck!

——————————————————————-

Like to teach yourself AdWords? Start here.

 

Spread the joy!

Q and A: Do regional domains constitute a duplicate content problem?

QuestionDear Kalena…

First of all I find the info on your site extremely useful –  I always look forward towards the newletter! I have been trying to find the time to do the SEO course but finding the time is always a problem! However, its still on my to do list.

I am trying to sort out a problem regarding duplicate content on my sites. We run local sites for each language/country we trade in (e.g. .fr for France and .co.uk for England). Unfortunately whilst growing the business I never had time to research SEO optimisation practices so I ended up with a lot of sites with the same duplicate content in them including title tags, descriptions etc. I had no idea how bad this was of course for organic ranking!

I have now created unique title tags and description for ALL the pages on ALL the sites. I have also changed the content into unique content for the home page and the paternity testing page (our main pages) for each site in English. The only site with complete unique content pages is .com and parts of .co.uk. For the rest of the pages that still have double content I have also put a NO INDEX, FOLLOW code on the pages that have duplicate content so that the spiders will not index the duplicate content pages. I did a FOLLOW as opposed to NO FOLLOW as I still want the internal links in the pages to be picked up – does this make sense ?

Also having made such changes how long does it normally take for Google to refresh its filters and starting ranking the site? The changes are now about a month old however the site is still not ranking.

Also should this not work – do you have any experience with submitting a re-consideration through the webmaster tools? What are the upside and downside of this?

Any advice would be greatly appreciated.

Regards
Kevin

Dear Kevin

Thanks for your coffee donation and I’m glad you like the newsletter. Now, about your tricky problem:

1) First up, take a chill pill. There’s no need to lodge a reinclusion request to Google. According to Google’s Site Status Tool, your main site is being indexed and hasn’t been removed from their datacenter results. A standard indexed page lookup shows 32 pages from your .com site have been indexed by Google, while a backward link lookup reveals at least 77 other sites are linking to yours. If you’ve put NoIndex tags on any dupe pages, you’ve covered yourself.

2) Next, pour yourself a drink and put your feet up. Your .fr site is also being indexed by Google, but there isn’t a dupe content issue because the site is in French, meaning that Googlebot sees the content as being completely different. Your .co.uk site is also being indexed by Google and again, there isn’t a dupe content issue because it looks like you have changed the content enough to ensure it doesn’t trip any duplicate content filters.

3) Now you’re relaxed, login to Google Webmaster Tools and make sure each of your domains are set to their appropriate regional search markets. To do this, click on each domain in turn and choose “Set Geographic Target” from the Tools menu. Your regional domains should already be associated with their geographic locations i.e. .co.uk should already be associated with the UK, meaning that Google will automatically be giving preference to your site in the SERPs shown to searchers in the UK. For your .com site, you can choose whether to associate it with the United States only (recommended as it is your main market), or not to use a regional association at all.

4) Now it’s time to do a little SEO clean up job on your HTML code. Fire or unfriend whoever told you to include all these unecessary META tags in your code:

  • Abstract
  • Rating
  • Author
  • Country
  • Distribution
  • Revisit-after

All these tags are un-supported by the major search engines and I really don’t know why programmers still insist on using them! All they do is clog up your code and contribute to excessive code bloat.

5) Finally, you need to start building up your site’s link popularity and boost your Google PageRank beyond the current 2 out of 10. And by link building, I mean the good old-fashioned type – seeking out quality sites in your industry and submitting your link request manually, NOT participating in free-for-all link schemes or buying text links on low quality link farms.

Good luck!

Spread the joy!

Confirmed: Google removes hacked sites from index

Well I’ve long suspected it and now Matt Cutts has confirmed it: sites that get hacked are removed from Google’s index.

Blogger and mountaineer Ryan Stewart (who I’ve long envied not just for his lack of acrophobia but also for his Google PageRank of 9/10) complained that Google had wiped his blog from the index.

Turns out the removal occurred for security reasons shortly after Ryan’s site was hacked. Here’s an extract from Matt’s comment on Ryan’s blog about the situation:

“Sorry to hear that your blog got hacked. I know that it’s disappointing if you don’t show up in Google, but there’s another way to look at it. It looks like your blog was hacked to show “buy pharmacy”-type links, but what if the hackers had hosted malware on your site? Then every user to your site might have gotten infected just by visiting your site. That danger to Google users is one of the reasons that we temporarily remove hacked sites from Google.”

So be warned, if your site is hacked, you might want to keep an eye on your Google Webmaster Tools account and submit a re-inclusion request as soon as things are back to normal.

Spread the joy!