Duplicate Forum Content
-
HI everyone,
great to be here, absolutely loving everything, please go easy on me I'm quite a noob when it comes to seo, but hopefully my question isn't too basic.
After running the initial checks on my websites, I found there are 7,646 duplicate pages? Some are easy fixes but the majority are not, being forum pages, the edit, quote and new post links are coming up as duplicates of the main post?
Does anyone know how to fix this?
Best Lee
-
How delightful that you refer to Invisionpower as I recently took the software for a test drive and plan to purchase it in January
A question though that has been troubling me. I use a paid for hosting service, which forwards to one of my sub-domains. The problem is I have nearly 7k pages of duplicate content (all from the forum), I really want to add a no follow and then remove the forum pages from the search index and thus remove all of the duplicate content. Though I'm worried that adding no follow to the sub domain may impact on the main site.
I have a lot of page one search positions, many long tails which combined result in a lot of visits, but I also have a couple of short phrases that get 50% of the visits alone, so I can't risk making any changes, actually I'm pretty terrified to make any changes, just in case all of my website pages are removed from the index.
-
invisionpower forum is good choice, as i knew. Found duplicated content pages here, it happens with most of forums, but there is full provision that this can resolved by canonical or stop crawling one you don't want in robots.txt.
-
Hey
just found some valuable info which has solved a lot of my problems, modified .htaccess which has removed all of the non www links, so thats a part solve.
I think what I'm going to do is purchase the new forum software, setup a robot.txt to not crawl the new forums and then once I have all of the posts transfered to the new forum, remove the old forums from the index.
That should solve it, the new forum is very well established and if need be I can pay one of their developers to incorporate some seo feature that will stop the duplicate content issue. Once thats solved I can remove the robot.txt
Many thanks, Lee
-
Hi, still one more thing, because you sound loss, and that's not the meaning of this Q&A. Contact the forumservice and ask for the canonical, this is a very simple and common thing to do.
-
Ok thanks Leonie,
still no nearer to knowing what to do, I know theres a problem and generally what I need to do, but I'm still at a complete loss as to how to do it
But thank you for trying to help,
Best, Lee
-
Seo stuff is'nt that difficult, though it can be comlicated
Anyway I wish you luck with the forum and hope you'll manage to get it the way you wanted.
Grtz, Leonie
-
Mmmm this is becoming more problematic by the minute, will need to implement robot.txt before removing the urls.
And there was me thinking this seo stuff would be easy.
-
Lol, tha'ts also an option
under url's index, though they have to be removed first, otherwise google crawl them again. but there you can remove complete directories at once.
-
or go somewhere else lol, have been considering buying invisionpower forum for a while, this might be the final push I needed.
Is there a way of mass deleting pages from the index in webmaster tools, i.e. all links that contain edit, quote and new ect?
-
Ah okay, that will be difficult than.
Maybe you can ask them to implement the canonical url. It will be helpful for you to avoid duplicates
-
is a paid forum service, but they give you the ability to add code snippets in the head.
-
Hi Lee,
What kind of CMS are you using?
In wmt under parameters you can configure parameters. If the ur's contain the same parameter it is a posibility.
-
Hi Leonie,
thanks for taking the time to answer
Not sure if that will work, there is only one head (hope that makes sense). Forums work pretty much the same as content management systems. I have the ability to change what's contained in the head, but adding rel="canonical" will have the same effect on all pages, even the duplicate one (I think).
Is there a way of removing all pages from the index (in webmaster tools) that contain the word edit, or quote or new? Kind of like using a wild card?
-
Hi Lee,
I'm not very familiar with forums. but i think you can solve it by putting a canonical in the head.
If you have a main page, which is the post, put a canonical url in it:
the duplicate pages need to have the same canonical.
I hope this work for you.
Grtz, Leonie
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Outsource Content Writing Services - Who Is The Best?
Hello fellow Mozers! We are currently in the process of hiring an additional content developer but he doesn’t start for a couple weeks. We have some work that is piling up and I am considering outsourcing some content development until the new hire gets here. I have looked into Scripted, TextMaster, CrowdContent and a few others. I stay away from sites like UpWork and Fiverr because I abuse them in the past and I don’t get quality work. Is there any company that you have used in the past that you really trust and who are familiar with SEO? Thank you in advance for all of your recommendations
Content Development | | InfinityDigital1 -
How do I split page content?
So I offer two services for which each has an FAQ page (let's call them S1 and S2). The problem is that I also have a longer FAQ page that covers both services (S1-2). I would like to eliminate the longer one and attach the relevant content to each of the shorter pages but i'm concerned that deleting pages with a lot of content might be a bad idea. I could redirect I suppose but I wouldn't know which page S1 or S2 to point redirect to. Any advice on this?
Content Development | | NationalPardon0 -
Duplicate page issue all from my website blog. How to i fix?
Crawl diagnosis indicates duplicate page content all from the blog on my website. What can i do to fix this?
Content Development | | skinbiz0 -
Possible to recover from Thin/Duplicate content penalties?
Hi all, first post here so sorry if in wrong section. After a little advice, if I may, from more experienced SEOers than myself with regards to writing off domains or keeping them. To cut a long story short I do a lot of affiliate marketing, back in the day (until the past 6 months or so) you could just take a merchant's datafeed and with some SEO outrank them for individual products. However, since the Google Panda update this hasn't worked as well and now it's much hard to do - which is better for the end user. The issue I have is that I got lazy and tried to see if I could still get some datafeeds to rank with only duplicate content. The sites ranked very well at first but after a couple of weeks died massively. They went from 0 to 300 hits a day in a matter of 24 hours and back to 2 hits a day. The sites now not rank for anything which is obviously because they are duplicate content. The question I have is are these domains dead, can they be saved? Not talking about duplicate content but as a domain itself. I used about 10 domains to test things, they ranged from DA 35 to DA 45 - one of the tests being can a domain with reasonable DA rank for duplicate content. Seeing as the test didn't work I want to use the domains for proper sites with proper unique content, however so far although the new unique content is getting indexed it is suffering from the same ranking penalties the duplicate (and now deleted content) pages had. Is it worth trying to use these domains, will Google finally remove the penalty when they notice that the bad content is no longer on the site or are the domains very much dead? Many thanks
Content Development | | ccgale0 -
What are the best content writer sites?
Hi, I'm doing some work on a new blog and wondered if anyone could recommend some low cost content writers? I have only justed started researching this service, so any advice the SEOmoz community could give would be grately appreciated. Thanks in advance.
Content Development | | RBH0 -
Duplicate Text on Blog & Internal News Page
I have two places I post news for our company. Our blog - typically more informal posts
Content Development | | seo-hunter
mycompany.wordpress.com & Our news page - typically more newsworthy than the blog
mycompany.com/news My question is, It is okay to just copy the exact text from my wordpress blog and paste to my news area of my site and vice versa? Does this hurt ranking potential for either page?0 -
Best way to resolve duplicate content issue?
Not sure about what to do about this - I have a client who has a ton of pages (around 1200) which are all City specific pages, for long-tail search. These are all written with paragraphs in the format such as: Order to [City] today. So every page has essentially the same content. The site also only has 1562 pages, so with 1200 of them being City-specific same-content pages, that can't be good. However the problem is that these pages still rank very well (usually Position 1 or 2) for the terms they're targeting, and bring in enough traffic and revenue to justify their purpose. We also have Country specific pages, and these are all with unique content, rather than the scripted content on the City pages. So for example, for Italy we might have: Italy Page (Unique Content) Rome (Duplicate Content) Milan (Duplicate Content) Venice (Duplicate Content) etc. (Duplicate Content) For a low traffic country (Austria), we tried to 301 the City pages to the Country page, but that only resulted in us seeing a drop in search results for the city keywords, from (usually) Position 1 to more like Page 3 or 4, so quite a drop. So, without writing 1200 pages worth of unique content, what would your advice be?
Content Development | | TME_Digital0 -
Displaying archive content articles in a writers bio page
My site has writers, and each has their own profile page (accessible when you click their name inside an article). We set up the code in a way that the bios, in addition to the actual writer photo/bio, would dynamically generate links to each article he/she produces. Figured that someone reading something by Bob Smith, might want to read other stuff by him. Which was fine, initially. Fast forward, and some of these writers have 3,4, even 15 pages of archives, as the archive system paginates every 10 articles (so www.example.com/bob-smith/archive-page3, etc) My thinking is that this is a bad thing. The articles are likely already found elsewhere in the site (under the content landing page it was written for, for example) and I visualize spiders getting sucked into these archive black holes, never to return. I also assume that it is just more internal mass linking (yech) and probably doesnt help the overall TOS/bounce/exit, etc. Thoughts?
Content Development | | EricPacifico0