An immediate and long-term plan for expired Events?
-
Hello all, I've spent the past day scouring guides and walkthroughs and advice and Q&As regarding this (including on here), and while I'm pretty confident in my approach to this query, I wanted to crowd source some advice in case I might be way off base. I'll start by saying that Technical SEO is arguably my weakest area, so please bear with me. Anyhoozles, onto the question (and advance apologies for being vague):
PROBLEM
I'm working on a website that, in part, works with providers of a service to open their own programs/centers. Most programs tend to run their own events, which leads to an influx of Event pages, almost all of which are indexed. At my last count, there were approximately 800 indexed Event pages.
The problem? Almost all of these have expired, leading to a little bit of index bloat.
THINGS TO CONSIDER
-
A spot check revealed that traffic for each Event occurs for about a two-to-four week period then disappears completely once the Event expires.
-
About half of these indexed Event pages redirect to a new page. So the indexed URL will be /events/name-of-event but will redirect to /state/city/events/name-of-event.
QUESTIONS I'M ASKING
-
How do we address all these old events that provide no real value to the user?
-
What should a future process look like to prevent this from happening?
MY SOLUTION
Step 1: Add a noindex to each of the currently-expired Event pages. Since some of these pages have link equity (one event had 8 unique links pointing to it), I don't want to just 404 all of them, and redirecting them doesn't seem like a good idea since one of the goals is to reduce the number of indexed pages that provide no value to users.
Step 2: Remove all of the expired Event pages from the Sitemap and resubmit. This is an ongoing process due to a variety of factors, so we'd wrap this up into a complete sitemap overhaul for the client. We would also be removing the Events from the website so there are not internal links pointing to them.
Step 3: Write a rule (well, have their developers write a rule) that automatically adds noindex to each Event page once it's expired.
Step 4: Wait for Google to re-crawl the site and hopefully remove the expired Events from its index.
Thoughts? I feel like this is the simplest way to get things done quickly while preventing future expired events from being indexed. All of this is part of a bigger project involving the overhaul of the way Events are linked to on the website (since we wouldn't be 404ing them, I would simply suggest that they be removed entirely from all navigation), but ultimately, automating the process once we get this concern cleaned up is the direction I want to go.
Thanks. Eager to hear all your thoughts.
-
-
Great! Happy to help
-
Hi Robin, thanks for taking the time to write out such detailed and helpful responses. I think I've decided to go with the approach you're outlining above:
For those that are already indexed:
- Change the 302s to 301s (all of the expired events that are indexed are 302s for some reason)
- 404/410 those that don't have any equity
- Create a custom 404 page
- Wait for them to drop out of index
For future Expired Events
- Wait about one month, then apply a 404 with custom page
- Redirect any that have backlinks
It'll require a little more work, but it is, I think, the right thing to do in this very bizarre situation.
-
To be honest it sounds like you already have your plan.
One thing I'd bear in mind is a crawl you run of your site won't line up with the pages that Google is visiting. For one thing, the tools we use try to approximate Google but won't be exactly the same. More importantly, once Google knows of a page it'll come back and check on it to see if the content changed, the only way you'll see that is by looking at your log files.
Yea there's no point making it "noindex, follow", it's not that Google doesn't know what to do with the page, it's just that it's attitude to the page will change over time.
In terms of the large number of redirects, there is some risk that Google could see the large number of 301s as spammy but, to be honest, I've never directly seen evidence of that being a problem. The way I see it, the choice is fairly similar you could
-
404/410 that's the way the internet is meant to work when something no longer exists but you'll lose link equity.
-
301 to preserve link equity but you're essentially misusing the status code.
-
Do a monthly check, 301 any expired pages with discovered backlinks, 410 the rest. This is best of both worlds but is much more time consuming.
I think you can probably get away with the 301s but it all comes down to your appetite for risk.
Good luck!
-
-
Thanks for the detailed response and the suggestion. The problem is, I think, a little more complicated than that. So there are two main concerns:
**1. What do we do with the current expired pages? **
So one thing that happens is that the event pages are effectively orphaned once the event has passed. All trace of them is removed from the website, and if my previous crawl is to be believed, they don't get crawled. Right now, the majority of these expired and indexed event pages are actually 302 redirects. So we're getting a temporary redirect to a page that is expired. Hardly a good user experience.
I do know that since it's a 302, Google is thinking "Hey, the page is coming back, so we're going to index that but send visitors to the new pages." This would be why the 302 URL is indexed. Am I correct in assuming that updating all of these to a 301 would result in the URL ultimately being removed? If so, then I think the best course of action would simply be to 301 redirect all of the current 302 URLs, as well as the actual expired event pages to the relevant event host / program pages.
Also, I did not know that _noindex _was treated as noindex, nofollow after awhile. Would it be beneficial to make them _noindex, follow, _or would that still be a redundancy that Google will ultimately ignore? I also do not think a pop-up is the way to go. These are very short-term events, so the issue is _less _a user experience and more a means of preventing them from clogging up the index. Also it would just be more work for the client and I'm trying to keep things as simple as possible.
**2. What do we do with the future expired pages so they don't end up getting indexed? **
This is probably a more pressing question. So the main concern is we want the Event pages to be indexed while they're live then ultimately removed after they've expired. I'm okay with this process: write a script that auto-redirects and remove all internal links from the website and just simply be patient. My main concern is just having way too many 301 redirects in place.
I'm hoping that the 301 in place combined with the complete orphaning of the page will mean they simply won't be crawled and eventually dropped from the index and thus not accessible to Google or users, but I'm still a little wary. Thoughts? Is there any room for adding anything to robots.txt?
Thanks again for your help. It is much appreciated.
-
Hi there, thanks for posting!
I think my main question here is around the decision to note 404 or 301 these pages. I totally understand that you want to reduce the number of indexed pages which aren't providing value but also don't want to lose equity. I know you mention you're not super technical so I'm going to break down how I expect link equity to be passed around a site and therefore how I expect each of these techniques to impact the page.
Equity is passed from page to page via links so these events pages will pass equity to other pages on yo by Google having a record of the page and the equity of that page, then distributing that equity through links it can follow. Google representatives have said recently that, after a period of time, noindex pages are treated as noindex nofollow at which point we can't rely on equity being passed along any of the outbound links from these pages.
-
noindex: removes the page from the index, after a period of time no equity will be passed from the noindexed page. Initially Google will continue to crawl the page but that will reduce over time.
-
404: the page doesn't exist so will be removed from the index after a period of time. No equity will be passed from the page. Google should stop crawling the page fairly quickly.
-
410: more definitive than 404. Page should drop out of the index more quickly. No equity will be passed from the page. Google should should stop crawling the page fairly quickly.
-
301: we're telling Google that this address is no good any more and it should instead look at a different address. Again, the redirected page should drop out of the index and some proportion of the redirected page's equity should be transferred to the target page. Google should stop crawling the page more quickly than noindexed version but probably not as quickly as the 404/410.
Based on all that I don't think noindex is necessarily your best option. You'll still have a bunch of defunct pages, which Google may still spend time crawling, and you can't rely on them passing equity.
A custom 404/410 page explaining to users that the event has passed is probably a pretty good user experience and would be the most expected behaviour for a situation where content isn't there any more, but won't help you with equity.
I think what you could do is automatically 301 redirect to a relevant category page with a pop-up message that explains to users what's happened. Doesn't sound like you expect the event pages to pop in and out of existence so the logic should be fairly simple.
Hope that helps!
-
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Homepage Not Ranking in Google - How long do old (not current) bad SEO practices exert influence?
I'm trying to get to the bottom of a problem I have with the Google ranking for mauiactivities.com - it's far below what I would hope for. My research so far has uncovered the following, and any advice on where to go from here would be appreciated. _Edit:_No problems with Bing or Yahoo - the site is #1 for primary key word 'maui activities' 1. Running a site:http://www.mauiactivities.com search in Google reveals that the homepage doesn't rank. At all. I looked through the 17 pages of results and can't spot it. Edit: I have now, after fresh checks after submitting the homepage through Search Console, found it at #1 - still, the following applies ... 2. I've found that the domain (before it was purchased by my client in 2011) had some bad inbound links, specifically from scubamaui.com (no longer active). The links where white, on a white background. This web archive snapshot will reveal all. 3. Those bad links were 'cleaned up' (i.e. they don't show in the web archive) from 2014, and as mentioned above, the website is now 'down'. 4. Search Console doesn't have a manual penalty. 5. When I search for 'tropical divers maui' in Google I find www.mauiactivities.com is the 4th result. To me, this indicates a current relationship with the dead site (Tropical Divers Maui). No other term comes close to ranking to high for the homepage. So, to summarise - can the old, dead Tropical Divers Maui website still be affecting the Google ranking, and what would you suggest I do next?
Technical SEO | | jsherwin0 -
How long does it take before a site is back in the SERP after a manual spamaction is revoked
Hi, A client of ours has a website with a manual spam action (duplicate content). Because they made some mistakes with redirects while moving the site from a URL to another google penitalized the site. We fixed the errors and the spamction is revoked. My question is how long it ussualy takes before the first results are back in the SERP. In WMT Google says "some time". But has anyone got some more information on it? Best Regards, Sam
Technical SEO | | U-Digital0 -
How can I Style Long "List Posts" in Wordpress?
Hi All, I have been working on a list-post which spans over 100 items. Each item on the list has a quick blurb to explain it, an image and a few resource links. I am trying to find an attractive way to present this long list post in Wordpress. I have seen several sites with long list posts however; they place their items one on top of the other which yields a VERY long page and the end user has to do a lot of scrolling. Others turn their lists into slideshows, but I have no data on how slides perform against 10-mile-long-lists which load in 1 page. I would like to do something similar to what List25.com does as they present about 5-10 items per page and they seem to have pagination. The pagination part I understand however; is there a shortcode plugin to format lists in an attractive way just like list25?
Technical SEO | | IvanC0 -
What is the best way to deal with an event calendar
I have an event calendar that has multiple repeating items into the future. They are classes that typically all have the same titles but will occasionally have different information. I don't know what is the best way to deal with them and am open to suggestions. Currently Moz anayltics is showing multiple errors (duplicate page titles, descriptions and overly dynamic urls). I'm assuming that it's showing duplicate elements way into the future. I thought of having the calendar no followed at all but the content for the classes seems valuable. Thanks,
Technical SEO | | categorycode0 -
SEO is DEAD – Long live ADWORDS
SEO is dead and the winner is ….. Google Adwords Do you agree??? Shares in Google inc on October 24th were trading well over $1,000 per share, joining an elite club of super-shares having risen 42% since the start of 2013. And the reason is in one word ….. “Adwords” If your email inbox is anything like mine, you will have noticed a change in your “junk” emails that get past your filters. The spammers have given up on targeting “penial-enlargement” products (because every spam filter in the world is blocking them) to so-called-legitimate-business activities like SEO (Search Engine Optimisation) or “How to Get your company to #1 on Google” The truth is all those thousands of dollars and hundreds of hours “we”, “you” and your company spent on increasing your Google “organic” rankings over the last five years have been a total waste of money…… and Google are totally to blame … and are making a total gain Once-upon-a-time you could get #1 ranking on google by using clever and current SEO techniques such as link-building, elimination of duplicate content, keyword density etc etc … terms many of you in business will have heard so-called-SEO-experts spruiking too many times and persuading you to part with $$$$ in cash to fix. How many of those so-called-SEO-experts told you you could save thousands of dollars by adopting a simple a cost effective Google adwords campaign??? Not many, the SEO industry is self-promoting and full of smoke-and-mirror (snake-oil) salesman promises. In recent times Google have….. a) Increased the number of paid ads at the top of the page from 1 to 3
Technical SEO | | Webjobz
b) increased the number of lines their paid adverts take up
c) increased the % of the “above-the-fold” advertising space to paid adverts
d) reduced the distinguishment between paid and organic ads to such an extent that around 50% of us cannot determine what is a paid or organic ad anymore. Being #1 “organically” (un-paid) in Google will NOW only get you 9% of the clicks on that page. 42% of all clicks goes to the top three “paid” listings (Adwords) If you want to get traffic to your site so that people purchase your products, move your marketing budget AWAY from SEO and web design and invest in Google Adwords as soon as possible. It is very easy to setup and administer yourself [staff removed links to copy of blog post]0 -
To get homepage to rank highly for a given term, do you only add the term(and variations) to the homepage?
If adding the term on all pages then won't the pages start competing with each other in the SERP? I'm building a site for a client who wants to rank highly for 'Southern California general contractors' and its not clear to me whether to fit these keywords (and variations) only into the homepage or if I should try to fit them into all pages.
Technical SEO | | kennysmith0 -
How long does it take for an article or a page to be listed by google
Hi, my question is a two parter. I think i must be doing something wrong. With my site map, it is set to show different section of my site while on my old site the site map listed every single article - i am not sure if setting it to each section is correct, can someone please advise me on this. The second part of the question is, how long does it take for an article to be listed by google. This article on my site was written today http://www.in2town.co.uk/lifestyle/holidaymakers-ignore-the-importance-of-travel-insurance-according-to-survey Holidaymakers Ignore The Importance of Travel Insurance According To Survey but when i check to see if google has listed the article yet by putting in the whole title, it does not come up, i even added the website name at the end and still it did not come up. This is worrying me a bit as a lot of my articles are news stories which means they are current articles so if google is not picking them up then no one else will be. can anyone let me know what i should be doing so google picks them up quicker please.
Technical SEO | | ClaireH-1848860 -
Is Adobe Acrobat the best for making PDF documents in terms of seo and price?
As we add PDF documents to our website, I want to take it up a notch. In terms of seo and software price, is Adobe Acrobat the only choice? Thanks! No Mac here. I should clarify that I can convert files to PDFs with Microsoft Word and add some basic info for the search engines such as title, keywords, author, and links. This article inspired me: www.seomoz.org/ugc/how-to-optimize-pdf-documents-for-search I can add links back to the page when I create the PDF, but we also have specific product PDFs that suppliers let us copy and serve from our server--why use their bandwidth. Much as you would stamp your name on a hard copy brochure the vendor supplies, I want to add a link to our page from those PDFs. That makes me think I should ask our supplier to give me a version with a link to our page. Then there is the question: is that ok to do? In the meantime, I will check TriviaChicken's suggestions and dream about a Mac, Allan. Thanks
Technical SEO | | zharriet0