How to handle Not found Crawl errors?
-
I'm using Google webmaster tools and able to see Not found Crawl errors. I have set up custom 404 page for all broken links. You can see my custom 404 page as follow.
http://www.vistastores.com/404
But, I have question about it.
Will it require to set 301 redirect for broken links which found in Google webmaster tools?
-
I agree with Ben on this one. There are plenty of 404s caused by scraper sites that don't and won't affect my time, especially on big sites.
Also, redirects aren't the only tool available. There are plenty of other ways to fix GWT 404 errors, particularly if there is a fundmental problem aside from the link in question.
-
Hi Commerce, I was certainly came across a blog post on this topic on Google's Webmaster Central blog, it covers most of the questions around 404 errors.
Generally speaking:
- If these are pages that you removed, then the 404 HTTP result code is fine.
- If these are pages that changed addresses, then you should 301 redirect to the new addresses. How you do this depends on your setup, for Apache-servers you may be able to use the .htaccess file for this.
- Unless these are pages that used to receive a lot of traffic from search, these 404s won't be the reason for your site's traffic dropping like that. Google understands that the web changes and that URLs disappear - that is not a reason for Google to stop showing your site.
So my recommendation would be to check the URLs that are listed as 404 crawl errors. If any are important, then set up redirects to the appropriate new URLs as soon as you can. If none of them are important, then keep this in mind as something worth cleaning up when you have time, but focus on the rest of your site first. Often drastic drops in traffic are due more to the general quality of the website, so that's what I'd recommend working on first.
For more deatails refer to How to Fix Crawl Errors.
I hope that your query had been solved.
-
Makes sense - in which case the homepage might not be the best place for you.
Another option for the custom 404 which works well in certain circumstances is to add a dynamic element to it.
For example, we know the referring URL has reference to product XYZ which may now be unavailable, but perhaps we can dynamically pull in other relevant products into the 404 page.
Thats something I am looking to do with hotels that become unavailable - pull in a dynamic element to the 404 which basically recommends some other hotels close by.
-
Well I would have to disagree with that principal. Sometimes you have to think a little broader than just SEO and ask yourself if it really makes commercial sense to redirect everything.
That's why I put a financial cost against each unique redirect. At the end of the day it requires someone to action it and that person has a cost associated with their time that may be better allocated working on something that will actually drive business uplift or improve customer experience.
Each to their own of course, but I see a lot of SEO's who don't think big picture and they up using up developer resource doing stuff that then has no impact. It just p!sses people off in my experience.
-
Hi Ben,
I agree with you that some links are not worth redirecting. However, in my experience a dead link never comes alone. Often there is some kind of reason that the link was created, and there might be others you don't know about.
For this reason I usually recommend redirecting all broken links, even if the individual link is not worth the trouble. Obviously there are exceptions to this rule, but most of the time it's worth your trouble.
Sven
-
Good to know! But, I have very bad experience to redirect such a strong page to home page. I have removed too many product pages for market umbrellas from my website and redirect it to home page. Because, I don't have specific landing page or inner level page for it. So, I'm able to see change over ranking for specific keywords. My home page is ranking well in Market Umbrellas keyword because too many external page link my product page with that keyword. It also create negative ranking impression for my actual targeted keyword which I'm using for my home page.
-
Yeah, which is basically what Kane is saying as well. If you don't have an appropriate internal page then you could send the 301 redirect to your homepage or if it was a specific product you might want to redirect it to the parent/child category.
If its a particularly strong URL that has been linked to from many good external sources then you might consider adding a replacement content page and redirecting to that.
Ben
-
Hi Ben,
I got your point. If my page is available on external page which have good value (Good page rank or heavy amount of traffic) so, I need to redirect it on specific internal page to save my page rank flow. Right?
-
Hopefully I am understanding your question correctly here....
The main benefit of the custom 404 page aside from the obvious improvement to user experience is that you provide additional links into content that otherwise wouldn't necessarily be available to the search bots.
In essence if you just had a standard 404 error page you'd send the search bots to a dead page where their only decision would be to leave the domain and go elsewhere.
Regards setting up 301 redirects I like to associate a cost to each 301 redirect. Imagine the time it will take you or someone else to set each redirect up (say $5 per redirect). Then consider the following:
Is the URL that is 404 worth redirecting?
(1) Does it hold some residual SEO value (i.e., is it present on external sites that is driving link equity? if so can you redirect that equity to somewhere more valuable?
(2) Is the URL present on an external site driving referral traffic? if so do you have a new content page that will still match the users intent?
if the URL(s) that are 404'ing have no real link equity associated to them and/or you don't have a genuinely useful page to redirect the user to then I would just let them hit the 404 page.
If in doubt put yourself in a users boots and ask yourself if the set-up you have done would offer a valuable experience? no point redirecting a user to something totally irrelevant to the original intent - it'll just p!ss them off most the time and increase your bounce rate.
-
If there is a link pointed at that 404 page, then I will almost always 301 it to regain that link value. If I control the source of the link, I'll change that instead. If the link is from a spammy or junky website, I don't worry about it.
Here is a worthwhile article on how to go about fixing GWT crawl errors:
http://www.seomoz.org/blog/how-to-fix-crawl-errors-in-google-webmaster-tools
I would suggest adding more content to your 404 page. Try to help people find what they're looking for by suggesting common pages, product segments, etc.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Google only crawling a small percentage of the sitemap
Hi, The company which I work for have developed a new website for a customer, there URL is https://www.wideformatsolutions.co.uk I've created a sitemap which has 25,555 URL's. I submitted this to Google around 4 weeks ago and the most crawls that have ever occurred has been 2,379. I've checked everything I can think of, including; Speed of website Canonical Links 404 errors Setting a preferred domain Duplicate content Robots Txt .htaccess Meta Tags I did read that Matt Cutts revealed in an interview with Eric Enge that the number of pages Google crawls is roughly proportional to your pagerank. But I'm sure it should crawl more than 2000 pages. The website is based on Opencart, if anyone has experienced anything like this I would love hear from you.
Technical SEO | | chrissmithps0 -
404s effecting crawl rate?
We made a change to our site where we all of a sudden we are creating a large number of 404 pages. Is this effecting the crawl/indexing rate? Currently we've submitted 3.4 million pages, have over 834K indexed but have over and 330K pages not found. Since the large increase in 404s we've noticed a decrease in pages crawled per day. I found this Q & A in Webmasters (http://googlewebmastercentral.blogspot.com/2011/05/do-404s-hurt-my-site.html) but it seems like the 404s should not have an effect. Is this article out of date? What do you think fellow Moz-ers? Is this a problem?
Technical SEO | | JoshKimber0 -
Why can no tool crawl this site?
I am trying to perform a crawl analysis on a client's website at https://www.bravosolution.com I have tried to crawl it with IIS for SEO, Sreaming Frog and Xenu and not one of them makes it further than the home page of the site. There is nothing I can see in the robots.txt that is blocking these agents. As far as I can see, Google is able to crawl the site although they have noticed a significant drop in organic traffic. Any advise would be very welcome Regards Danny
Technical SEO | | richdan0 -
Help Crawl friendliness for large site
After watching Rand's video I am trying to think of the best way to make my large site more crawl friendly. Background I have a large site with over 100k product skus and so when you get to a particular page of products there are tons of different refinements and options that help you sort the products. Most of these are noindex followed, but I was wondering if I should be nofollowing the internal links as well in order to keep bots out of those pages and going to the pages that I want them to go too. Is this a good way to handle it? Also, does anyone have good recommendations of links to posts that deal with helping the crawl friendliness of a large site? Thanks!
Technical SEO | | Gordian0 -
Are 404 Errors a bad thing?
Good Morning... I am trying to clean up my e-commerce site and i created a lot of new categories for my parts... I've made the old category pages (which have had their content removed) "hidden" to anyone who visits the site and starts browsing. The only way you could get to those "hidden" pages is either by knowing the URLS that I used to use or if for some reason one of them is spidering in Google. Since I'm trying to clean up the site and get rid of any duplicate content issues, would i be better served by adding those "hidden" pages that don't have much or any content to the Robots.txt file or should i just De-activate them so now even if you type the old URL you will get a 404 page... In this case, are 404 pages bad? You're typically not going to find those pages in the SERPS so the only way you'd land on these 404 pages is to know the old url i was using that has been disabled. Please let me know if you guys think i should be 404'ing them or adding them to Robots.txt Thanks
Technical SEO | | Prime850 -
Error Reporting
http://pro.seomoz.org/campaigns/33868/issues/18 Rel Canonical Found about 16 hours ago <dl> <dt>Tag value</dt> <dd>http://www.geeks.com/</dd> <dt>Description</dt> <dd>Using rel=canonical suggests to search engines which URL should be seen as canonical.</dd> <dd>We do have rel canonical on some of the pages this report is recommending that we "fix" this issue.</dd> <dd> Rel Canonical Found about 16 hours ago <dl> <dt>Tag value</dt> <dd>http://www.geeks.com/products.asp?cat=MBB</dd> <dt>Description</dt> <dd>Using rel=canonical suggests to search engines which URL should be seen as canonical.</dd> </dl> <a class="more expanded">Minimize</a> </dd> </dl>
Technical SEO | | JustinGeeks0 -
Pdf page titles and descriptions errors
In my weekly crawl report I suddenly have a huge number of title not found and missing page descriptions errors for all of my pdf files. The pdfs do have a page title (defined in the file/properties tab) However these page titles are not being picked up by the crawler or google. Any ideas how do I fix this? ( I am using the Adrobat 9 Distiller)
Technical SEO | | PerriCline0 -
404 Errors
Hello Team, I noticed that my site has 1,000s of 404 errors. Not sure how this happened, maybe when I updated our CMS. My question is, should I worry about them. Should I delete them or just leave them alone. Thank you for your feedback!
Technical SEO | | Dallas0