Magento creating odd URL's, no idea why. GWT reporting 404 errors
-
Hi Mozzes!
Problem 1
GWT and Moz, both are reporting approximately one hundred 404 errors for certain URL's. Examples shown below. We have no idea why or how these URL's are being created in Magento. Any hypothesis on the matter would be appreciated. The domain name in question is http://www.artorca.com/
These are valid URL's if /privacy is removed. The first URL is for a product, second for an artist profile and third for a CMS page
1. semi-abstract-landscape/privacy
2. jose-de-la-barra/privacy
3. seller-guide/privacy
What may be the source for these URL's?
What solution should we implement to fix existing 404's? 301 redirects should be fine?
Problem 2
Website pages seem to also be accessible with index.php in the domain name. Example Artorca.com/index.php/URL's. Will this cause a duplicate content issue? Should we implement 301's, canonicals, or just leave as is?
Cheers!
MozAddict
-
We run 2 sites on Magento and I've seen these problems before.
Problem 1 - this looks like a link to the privacy page (maybe in the footer?) is missing an initial /. Hence it's not being treated as relative and trying to build it onto the existing URL.
Problem 2 - shouldn't be an issue as long page has a canonical tag pointing to itself. I tried one of our pages with index.php/XXXX and it resolved and has a canonical tag pointing to it without index.php.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Solved Would my site's DA be transferred if I redirect to another?
Re: How to create link from google redirect? I am thinking of changing my domain name from https://experts.ng to https://expertsclan.com and wondering if my DA could be transferred to the new site
Moz Pro | | dodo1234 -
Large site with content silo's - best practice for deep indexing silo content
Thanks in advance for any advice/links/discussion. This honestly might be a scenario where we need to do some A/B testing. We have a massive (5 Million) content silo that is the basis for our long tail search strategy. Organic search traffic hits our individual "product" pages and we've divided our silo with a parent category & then secondarily with a field (so we can cross link to other content silo's using the same parent/field categorizations). We don't anticipate, nor expect to have top level category pages receive organic traffic - most people are searching for the individual/specific product (long tail). We're not trying to rank or get traffic for searches of all products in "category X" and others are competing and spending a lot in that area (head). The intent/purpose of the site structure/taxonomy is to more easily enable bots/crawlers to get deeper into our content silos. We've built the page for humans, but included link structure/taxonomy to assist crawlers. So here's my question on best practices. How to handle categories with 1,000+ pages/pagination. With our most popular product categories, there might be 100,000's products in one category. My top level hub page for a category looks like www.mysite/categoryA and the page build is showing 50 products and then pagination from 1-1000+. Currently we're using rel=next for pagination and for pages like www.mysite/categoryA?page=6 we make it reference itself as canonical (not the first/top page www.mysite/categoryA). Our goal is deep crawl/indexation of our silo. I use ScreamingFrog and SEOMoz campaign crawl to sample (site takes a week+ to fully crawl) and with each of these tools it "looks" like crawlers have gotten a bit "bogged down" with large categories with tons of pagination. For example rather than crawl multiple categories or fields to get to multiple product pages, some bots will hit all 1,000 (rel=next) pages of a single category. I don't want to waste crawl budget going through 1,000 pages of a single category, versus discovering/crawling more categories. I can't seem to find a consensus as to how to approach the issue. I can't have a page that lists "all" - there's just too much, so we're going to need pagination. I'm not worried about category pagination pages cannibalizing traffic as I don't expect any (should I make pages 2-1,000) noindex and canonically reference the main/first page in the category?). Should I worry about crawlers going deep in pagination among 1 category versus getting to more top level categories? Thanks!
Moz Pro | | DrewProZ1 -
Rogerbot's crawl behaviour vs google spiders and other crawlers - disparate results have me confused.
I'm curious as to how accurately rogerbot replicates google's searchbot I've currently got a site which is reporting over 200 pages of duplicate/titles content in moz tools. The pages in question are all session IDs and have been blocked in the robot.txt (about 3 weeks ago), however the errors are still appearing. I've also crawled the page using screaming frog SEO spider. According to Screaming Frog, the offending pages have been blocked and are not being crawled. Webmaster tools is also reporting no crawl errors. Is there something I'm missing here? Why would I receive such different results. Which one's should I trust? Does rogerbot ignore robot.txt? Any suggestions would be appreciated.
Moz Pro | | KJDMedia0 -
What's up with Yahoo and Bing in my keyword rankings?
SEOmoz is good I've learned a lot from it. I'm on my 3rd month and now have 8 keywords in top 50 for one site and 1 keyword in top 50 for another. This is for Google US. For Bing and Yahoo! US however I am still not in the top 50 for anything. I have read elsewhere that meta-tag page words can be regarded by these engines as a spam signal. In my meta I have used competative keywords that are relevant to my content, some of which are in the body, but are not my actual targeted keywords as a red-herring to potential competitors. Also might be usefull for other smaller places that still use meta-tags. Could not appearing in BingHoo! have something to do with their alogrithms and the meta-tags? or is it an SEOmoz issue? I have another question also. The " / " after a domain name doesn't appear in my Firefox address bar (I think I did a 301 redirect), but when I copy and paste it into an entry box (e.g. for a link submission) it puts the " / " there at the end which I take off. Does this make any difference?
Moz Pro | | Zoolander0 -
Broken links in the pdf of the On Page Report
Hi, I run an individual On Page report for a particular URL, then I export as pdf. The URL appears in the pdf and looks fine but when you click on it it goes to a 'page not found'. I know the URL is correct. When I hover over the URL in the pdf i notice that the word 'Good' is at the end of my URL but I did not put this in there. if I give the report to a client it doesn't look so good. http://www.narellanpools.com.au/local-contact/narellan-pools-alburywodongaGood Is this a bug? Cheers Virginia
Moz Pro | | VirginiaC0 -
I'm trying to get 'tigi bed head' up most of all...
I'm 87th ish with this term and I don't know why?! crap result I know. With every other phrase I use 'cheap tigi bed head' 'buy tigi bed head online' etc etc, we are on the first page all day long, pls help this worthy cause? I am www.thehairroom.co.uk, free hair products for the best results. Thank You.
Moz Pro | | smoki6660 -
SEOMOZ reports the statistics.
SEOMOZ reports the Statistics, but where do i manage & improve??? Simply Statistics is all about SEOMOZ??
Moz Pro | | webicers0 -
HTTP 404 for 404-page?
Hi Mozzers! SEOmoz just finished crawling one of my websites and this crawl found 3 errors. One of these errors was the (custom) 404-page, because of the http-status 404. What's you suggestion about this? Should a 404-page have a status 404? Thanks in advance for your suggestions!
Moz Pro | | Partouter0