News Archive within a site - Detrimental?
-
I have a clients site, which we publish fresh news content daily and have done for some time. this has built up a huge archive of news over the past few years. All the news resides in a /news sub folder. The most interesting news articles are syndicated through facebook and twitter.
Note: All the news is original content.
The archive is based around a chronological filing system so /news/2010/December would retrieve news articles published in December 2010.
My first question is this, in the Post Panda seo world, since the news in the archives receives little to no traffic, will these be classed as low quality pages, even though they are generally informative and timely news articles?
My Second Question, It's my understanding that a site can be penalised for harbouring low quality pages, thus would I be better to do away with my news archive of around 600+ articles and impose a rule that news is removed from the site after a set period (say 6 -12 months)?
-
Thats why I ask on SEOmoz, a sensible answer with sensible advice that can be easily put into action.
This had crossed my mind, however I've been staring at the woods so long, I couldn't see the trees!
Thanks EGOL.
-
We have a blog that publishes about six to ten short posts per day, each linking to an external URL that has great content. Most of the content that we link to is newsy rather than evergreen.
Over time that content gets stale, the number of search queries drops and some of the links go dead. This content loses its value over time.
Once each year we do a major purge of old posts, deleting a couple thousand at a time.
Before this purge is done we look at analytics to determine if any of the posts are pulling traffic from search. When we find posts pulling lots of traffic we will create a new page of substantive content about that subject or identify an existing page that is relevant. The posts are then 301 redirected to those pages. All remaining posts are redirected to the homepage of the blog.
We have been doing this for a few years and it makes even more sense now that google seems to dislike pages with short content.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Google News problem
Hello to all. The latest Google algorithm changes have had a big impact on the way that Google news features stories, at least in my country. I've been featured heavily in Google News until about 6th of october, when the changes had the biggest impact, but since then, I haven't been featured at all. Prior to this, I would be featured for keywords on almost any article, not necessarily on the 1st position, but I was almost always there. Posts still show up in the dedicated News category, but not in the main search pages. I've seen a lot of websites being impacted, but some with lower ranks than mine still show up there. I haven't done any changes prior to the 6th of october, and I haven't done any link building campaings, just getting links from higher ranking news sites in my country, for articles I wrote. What I'd like to know is if there were any major changes for Google News and I'm not complying with any of them, or If i could check to see if there are any other problems. I don't have any penalties disclosed by Google, and no new errors in the Webmasters console, I'm just baffled by the fact that overnight the website was completely cut off from being featured in Google News. And one other strange thing, I'm now ranking better for searches that are kind of opposite to my website's main theme. Think about mainly writing about BMW, and less about AUDI, but ranking a lot better for the latter, and a lot less for the other. Thank you.
Technical SEO | | thefrost0 -
Should we dump the https from a client site?
We inherited a site that has both http and https. No e-commerce or data transfer...just html. Should we dump the https certificate? I think it might be causing issues with indexing and possible duplicate content. The https site has a certificate warning message...not good. The URL is www.charlottemechanical.com
Technical SEO | | theideapeople0 -
Site architecture & breadcrumbs
Hi A client hasn't structured site architecture in a silo type format so breadcrumbs are not predicating in a topical hierarchy as one would desire (or at least i think one would prefer) For example: say the site is called www.fruit.com and it has a category called 'types of fruit' and then sub/content pages called things like 'apples' and 'pears'. So in terms of architecture that should be: www.fruit.com/types-of-fruit/apples and www.fruit.com/types-of-fruit/pears etc etc The client has kept it all flat so instead architecture is: www.fruit.com/types-of-fruit and www.fruit.com/apples and www.fruit.com/pears As a result breadcrumbs follow suit and hence since also not employing logical predication dont reflect the topical & sub-topical hierarchy I have seen that some seo's at least used to think this was better for seo since kept the page/s nearer the root but surely its better to structure site architecture in a logical topical hierarchy so long as dont go beyond say 3 or 4 directories/forward slashes in the url's? Also is it theoretically possible to keep url structure as is (flat) and just edit/customise the breadcrumbs to reflect a topical hierarchy in a silo structure rather than change the entire site architecture & required 301'ing etc in order to do this (or is that misleading or just not possible?) Cheers Dan
Technical SEO | | Dan-Lawrence0 -
I noticed all my SEOed sites are getting attacked constantly by viruses. I do wordpress sites. Does anyone have a good recommendation to protect my clients sites? thanks
We have tried all different kinds of security plugins but none seem to work long term.
Technical SEO | | Carla_Dawson0 -
How to handle mobile site with less pages than the main site?
We are developing a mobile version of our website that will utilize responsive design/dynamic serving. About 70% of the main website will be included in the mobile version. What (if anything) should be the redirect for pages not included in the mobile version of the site? Also - for one specific section users will be redirected from that page to the homepage, what is the redirect that should be used for this? Thanks!
Technical SEO | | theLotter0 -
Ranking Multi-Language Site
Recently we updated our website to a new version. Our website has a structure in which the English page is our main page with about 50 subpages. All these pages are translated in 5 different languages. The different languages are divided into folders. For example www.ourdomain.com/de containts all german pages. The pages with products would be for example: www.ourdomain.com/products for english and www.ourdomain.com/de/produkte for the german page. On our previous website this used to be simililar. After the website update the SEOMoz crawls are showning duplicated page content/title errors for the pages saying that the pages in other languages have the same content/title as the basis English webpage. Any idea how I can solve these errors?
Technical SEO | | Exp0 -
Duplicate Content - Mobile Site
We think that a mobile version of our site is causing a duplicate content issue; what's the best way to stop the mobile version being indexed. Basically the site forwards mobile users to "/mobile" which is just a mobile optimised version of the original site. Is it best to block the /mobile folder from being crawled?
Technical SEO | | nsmith7870 -
Robots.txt blocking site or not?
Here is the robots.txt from a client site. Am I reading this right --
Technical SEO | | 540SEO
that the robots.txt is saying to ignore the entire site, but the
#'s are saying to ignore the robots.txt command? See http://www.robotstxt.org/wc/norobots.html for documentation on how to use the robots.txt file To ban all spiders from the entire site uncomment the next two lines: User-Agent: * Disallow: /0