Looking for list Pro's & Con's of removing Folder from URL?
-
Hi
We have a sub-folder ("/shop-by-department/") which is pretty much useless on our site and I'm looking to remove it. But the team want a list of the Pro's & Con's in doing so.
So for example I'll be changing www.example.ie/shop-by-department/furniture/beds/product-a to www.example.ie/furniture/beds/product-a
I know there will be an intial hit as Google adjusts to the change but think it's definitely the way to go. I was lookng for a complete list of the Pro's & Con's to send onto the team. It'll be going to the traditional marketing (print, radio, etc.) too so can ve top-level points too.
Hope you can help!
Thanks
-
There are no cons (that I can see) to losing the "/shop-by-department/" folder, the new URL structure you illustrated is better as it offers an opportunity to add pertinent keywords.
Just be sure you have your 301 redirects set up correctly.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Curious why site isn't ranking, rather seems like being penalized for duplicate content but no issues via Google Webmaster...
So we have a site ThePowerBoard.com and it has some pretty impressive links pointing back to it. It is obviously optimized for the keyword "Powerboard", but in no way is it even in the top 10 pages of Google ranking. If you site:thepowerboard.com the site, and/or Google just the URL thepowerboard.com you will see that it populates in the search results. However if you quote search just the title of the home page, you will see oddly that the domain doesn't show up rather at the bottom of the results you will see where Google places "In order to show you the most relevant results, we have omitted some entries very similar to the 7 already displayed". If you click on the link below that, then the site shows up toward the bottom of those results. Is this the case of duplicate content? Also from the developer that built the site said the following: "The domain name is www.thepowerboard.com and it is on a shared server in a folder named thehoverboard.com. This has caused issues trying to ssh into the server which forces us to ssh into it via it’s ip address rather than by domain name. So I think it may also be causing your search bot indexing problem. Again, I am only speculating at this point. The folder name difference is the only thing different between this site and any other site that we have set up." (Would this be the culprit? Looking for some expert advice as it makes no sense to us why this domain isn't ranking?
Web Design | | izepper0 -
Should Blog Category Archive URLs be Set to "No-Index" in Wordpress?
It appears that Google Webmaster Tools is listing about 120 blog archives URLs in Google Index>Index Status that should not be listed. Our site map contains 650 pages, but Google shows 860. Pages like: <colgroup><col width="464"></colgroup>
Web Design | | Kingalan1
| http://www.nyc-officespace-leader.com/blog/category/manhattan-office-space | With Titles Like: <colgroup><col width="454"></colgroup>
| Manhattan Office Space Archives - Metro Manhattan Office Space | Are listed when in the Rogerbot crawl report for the site. How can we remove such pages from Google Webmaster Tools, Index Status? Our site map shows about 650 pages, yet Google show these extra pages. We would prefer that they not be indexed. Note that these pages do not appear when we run a site:www.nyc-officespace-leader.com search. The site has suffered a drop in ranking since May and we feel it prudent to keep Google from indexing useless URLs. Before May 650 pages showed on the Webmaster Tools Index status, and suddenly in early June when we upgraded the site the index grew by about 175 pages. I suspect the 120 blog archives URLs may have something to do with it. How can we get them removed? Can we set them to "No-Index", or should the robot text be used to remove them? Or can some type of removal request be made to Google? My developers have been struggling with this issue since early June. The bloat on the site is about 175 URLs not on the site map. Is there any go to authority on this issue (it is apparently rather complicated) that can provide a definitive answer? Thanks!!
Alan0 -
Q&A DB Script For Client Questions
Our clients usually have a lot of questions about our industry. We implemented Facebook comments (see bottom of: http://www.jwsuretybonds.com/info/faq.htm) on many pages throughout the site to allow our clients to ask whatever they'd like and build additional content using copy on how our clients speak, not industry expert vocab... I have some problems with FB comments: Our answers only show to people logged into FB. For others it looks as if we don't reply. The data pulls from FB, so we don't own it and it is slower than our server. You need a Facebook, Yahoo, or Hotmail account...I want anyone to be able to ask a question. Can anyone recommend a script that accomplishes our Q&A functionality using our own database? Also, I'd like to allow anyone to post without a requirement of an account. If I had a WordPress site, I suppose this would be easier, as I'm sure there are various comment solutions, but I have a HTML\PHP based custom built site.
Web Design | | TheDude0 -
Check list for Manual penalty – Before going for reconsideration
Hi there, Any one please help us, in cross checking our recovery work before submitting to reconsideration request. Quality Check list for a Website SEO audit check list Why am requesting means, our previous request was came with a negative result. Thanking you in advance
Web Design | | Shashidhar.SEO1 -
Bizarre PDF URL string
Hey folks, I'm getting literally hundreds of Duplicate Title and Duplicate Content errors for a site, and most of them are a result of the same issue. The site uses javascript container pages a lot, but each gets their own URL. Unfortunately, it seems like each page is also loading all the content for all the other pages, or something. For instance, I have a section of the site under /for-institutions/, and then there are 5 container pages under that. Each container page has it's own URL, so when you select it, you get the URL /for-institutions/products/ or /for-institutions/services/ etc. However, the institutions container page doesn't change, just the content within. In my SEO results, I'm getting the following: /for-institutions/$%7Bpdf%7D/ /for-institutions/$%7Bpdf%7D/$%7Bpdf%7D/ etc, each as a duplicate title and content page. How can I eliminate this? Is there a regular expression that rewrites URL segments beginning with $ ? For your reference: The page is set up so that any URL that doesn't exist just refers to the subdirectory. /for-institutions/$%7Bpdf%7D/ displays /for-institutions/, but does not rewrite the URL. So too if I were to enter /for-institutions/dog.
Web Design | | SpokeHQ0 -
For A Corporation With 3 Distinct Business Divisions, Is It Better To Go With 1 Domain & 3 Sub-Domains, 1 Domain & 3 Folders, or 3 Domains for SEO Purposes?
Hi, I am working on a project right now for an existing client, we have one domain up and running well, they want to create an 'umbrella' site to cover three current business divisions and roll everything up under that main site, including the existing site on a totally different domain (would migrate over and 301 redirect from current domain). From what I've researched, I am inclined towards one main domain with three sub-domains due to the amount of content for each business division being significantly different enough that it seems to deserve separation from each other. However, in terms of SEO and maintaining consistent domain authority, would anyone recommend it be better to structure this as just folders/categories falling under the main domain instead of separate sub-domains for each division, and focus keyword targeting on pages tailored to that end within the main domain structure rather than spreading out link-juice to different sub-domains? Thanks!
Web Design | | Dan_InboundHorizons0 -
Ecommerce & Responsive design
Hi there, We are thinking to redevelope our ecommerce websites and thinking responsive design. Due to responsive design when the screen gets smaller to fit iphone and ipad we need to hide some content to make it more user friendly. My question is, how Google will treat hiding content with the smaller screens? Will this effect our rankings in a negative way? We really don't want to get punished by Google 🙂 Thank You
Web Design | | Jvalops0 -
The use of foreign characters and capital letters in URL's?
Hello all, We have 4 language domains for our website, and a number of our Spanish landing pages are written using Spanish characters - most notably: ñ and ó. We have done our research around the web and realised that many of the top competitors for keywords such as Diseño Web (web design) and Aplicaión iPhone (iphone application) DO NOT use these special chacracters in their URL structure. Here is an example of our URL's EX: http://www.twago.es/expert/Diseño-Web/Diseño-Web However when I simply copy paste a URL that contains a special character it is automatically translated and encoded. EX: http://www.twago.es/expert/Aplicación-iPhone/Aplicación-iPhone (When written out long had it appears: http://www.twago.es/expert/Aplicación-iPhone/Aplicación-iPhone My first question is, seeing how the overwhelming majority of website URL's DO NOT contain special characters (and even for Spanish/German characters these are simply written using the standard English latin alphabet) is there a negative effect on our SEO rankings/efforts because we are using special characters? When we write anchor text for backlinks to these pages we USE the special characteristics in the anchor text (so does most other competitors). Does the anchor text have to exactly I know most webbrowsers can understand the special characters, especially when returning search results to users that either type the special characters within their search query (or not). But we seem to think that if we were doing the right thing, then why does everyone else do it differently? My second question is the same, but focusing on the use of Capital letters in our URL structure. NOTE: When we do a broken link check with some link tools (such as xenu) the URL's that contain the special characters in Spanish are marked as "broken". Is this a related issue? Any help anyone could give us would be greatly appreciated! Thanks, David from twago
Web Design | | wdziedzic0