Cloaking? Best Practices Crawling Content Behind Login Box
-
Hi-
I'm helping out a client, who publishes sale information (fashion sales etc.)
In order for the client to view the sale details (date, percentage off etc.) they need to register for the site.
If I allow google bot to crawl the content, (identify the user agent) but serve up a registration light box to anyone who isn't google would this be considered cloaking?
Does anyone know what the best practice for this is? Any help would be greatly appreciated.
Thank you,
Nopadon
-
Can I say I admire your inventiveness? You go to some lengths to not register and really, apart from the majority of people not knowing how to do a reverse image search, probably reflects people's attitude to those sorts of lightbox registration forms.
-
I'm going to respond from a human point of view and not a technical point of view.
I've been searching for houses recently on Craigslist. There are a couple of real estate agents who post ads on CL with a link to their site. When you click the link, you get a lightbox requiring that you fill out the lead form to be able to see the details of the house. I do one of two things:
-
I open up IE in private browsing mode and paste in the URL. The private browsing mode has something that prevents this script from running and I can see the house details just fine.
-
If the house address is not provided in the CL ad, I'll copy the image URL of one of the CL photos and put that into a Google reverse image search. I'll find a different website that has posted the same house and use their site that doesn't require me to register. (I realize this may not happen in your scenario above).
I agree what the other people say about not wanting provide one thing to Google and another to users, and wanted to add that people will try to find ways around the registration. I don't have a solution for you, sadly.
-
-
Heya there,
Thanks for asking your question here
My first point would be that human visitors don't like to be given forms when they first visit a site, so would suggest you don't do this.
My alternative strategy would be to provide a home page of good content talking about the data etc that is available on your site and then provide a button for people to register if they want to.
Don't detect the user agent and provide alternative content as, however good your intentions are, that could be considered cloaking. Google is against you providing Google different content to humans, so don't do it.
Do things differently
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Best Nofollow Practice Facing +500 instances of affiliate links
Hi. I have a set of affiliate links on my website. That's 50 unique links in total. They are placed in a separate folder /go/, and are all nofollowed. My website has 100 pages. When doing a site audit I get a report with 50 affiliate links pr. page x100. That's a huge report. Is this really necessary? I can ofc. choose to ignore the automated reports, but I can't help but to think that there must be an easier way to keep track of it? Like, if Link 1-50 is located in /go/, they should automatically be nofollowed, regardless of where they are located on the root/sub domain. Bonus question: The format of the link is https://exampledomain.com/go/brandname I am wondering if I should add that "/" at the end of it. If yes, why? If no, why?
Technical SEO | | Dan-Louis0 -
Best Practice for www and non www
How is the best way to handle all the different variations of a website in terms of www | non www | http | https? In Google Search Console, I have all 4 versions and I have selected a preference. In Open Site Explorer I can see that the www and non www versions are treated differently with one group of links pointing to each version of the same page. This gives a different PA score. eg. http://mydomain.com DA 25 PA 35 http://www.mydomain.com DA 19 PA 21 Each version of the home page having it's only set of links and scores. Should I try and "consolidate" all the scores into one page? Should I set up redirects to my preferred version of the website? Thanks in advance
Technical SEO | | I.AM.Strategist0 -
301 Clean-Up - Best Practices & Procedure?
Hello Again, I have taken over managing a website for about 2 months and have fixed a whole heap of problems. Im now turning my attention to the URL rewrites as there are ALOT of them. I have fixed the most problematic offenders that were blocking products and all sorts of mischief but I now want to clean them up. The website is on Magento, and there are 240 custom URL rewrites. Question 1: Am i correct that I should edit the links on my website so that they link directly to the new page instead of utilising the re-direct for best SEO results. Question 2: If my website doesn't utilise the URL rewrite (fixed in question 1) its only purpose is to transfer link juice from any external link the page had before. If this page didnt have any external inbound links then I can delete the URL rewrite as it serves no purpose. Question 3: If Q1 and Q2 are correct, what is the quickest way to check the inbound links to a page quickly so I can make a quick decision on if i should remove the re-write. Many Thanks in advance!
Technical SEO | | ATP1 -
Advice urgently needed on best practice for handling multiple product categories on Magento website
I have an ecommerce site built using Magento and urgently need advice on best practice for handling multiple product categories (where products appear in more than one category on the site creating multiple URLs to the same page). In April this year, based on advice from my SEO who felt that duplicate content issues were causing my rankings to be held back, I changed about 25% of the product categories to 'noindex, follow'. This has made organic traffic fall (obviously) as these pages fell out of Google's index. But, contrary to what I was hoping for, it didn't then improve rankings - not one iota, nothing - which was the ONLY reason why I did this. This has had a real negative impact on sales, so I'm starting to think this was actually an a terrible idea. Should I change them back? And to ask a wider question, what is best practice for this particular scenario?
Technical SEO | | Coraltoes770 -
Moving content
I have www.SiteA.com which contains a number of sections of content, a section of which (i.e. www.SiteA.com/sectionA), we would like to move to a new domain www.SiteB.com Definitely we will ensure that a redirect strategy is in place and that we submit a sitemap for SiteB Three Questions 1. Anything else I am missing from the migration plan? 2. Since we are only moving part of SiteA to SiteB, is there another way of telling Google that we changed address for that section or are the 301s enough? 3. Currently, Section A (under SiteA) contains a subsection where we were posting an article a day. In the new site (SiteB), we decided to drop this subsection and write content (but not "exactly" the same content) under a new section. During migration, how should we handle the subsection that we have decided to stop writing? Should we: A. Import the content into SiteB and call it archives and then redirect all the urls from subsection under SiteA to the archives under SiteB? OR B. Do not move the content but redirect all the pages (365 in total) to where we think the user would be more interested in going to on SiteB? Note: A colleague of mine is worried that since the subsection has good content he thinks its necessary to actually move the content to SiteB. But again, looking at the views for the archives it caters for 1% of the the total views of this section. In other words, people only view the article on the day it is written. I hope I was clear 🙂 Your help is appreciated Thank you
Technical SEO | | seo12120 -
Duplicate Page Content and Titles
A few weeks ago my error count went up for Duplicate Page Content and Titles. 4 errors in all. A week later the errors were gone... But now they are back. I made changes to the Webconfig over a month ago but nothing since. SEOmoz is telling me the duplicate content is this http://www.antiquebanknotes.com/ and http://www.antiquebanknotes.com Thanks for any advise! This is the relevant web.config. <rewrite><rules><rule name="CanonicalHostNameRule1"><match url="(.*)"><conditions><add input="{HTTP_HOST}" pattern="^www.antiquebanknotes.com$" negate="true"></add></conditions>
Technical SEO | | Banknotes
<action type="Redirect" url="<a href=" http:="" www.antiquebanknotes.com="" {r:1"="">http://www.antiquebanknotes.com/{R:1}" />
</action></match></rule>
<rule name="Default Page" enabled="true" stopprocessing="true"><match url="^default.aspx$"><conditions logicalgrouping="MatchAll"><add input="{REQUEST_METHOD}" pattern="GET"></add></conditions>
<action type="Redirect" url="/"></action></match></rule></rules></rewrite>0 -
I am Posting an article on my site and another site has asked to use the same article - Is this a duplicate content issue with google if i am the creator of the content and will it penalize our sites - or one more than the other??
I operate an ecommerce site for outdoor gear and was invited to guest post on a popular blog (not my site) for a trip i had been on. I wrote the aritcle for them and i also will post this same article on my website. Is this a dup content problem with google? and or the other site? Any Help. Also if i wanted to post this same article to 1 or 2 other blogs as long as they link back to me as the author of the article
Technical SEO | | isle_surf0 -
If I 301 re-direct a piece of content (A) to another piece of content (B) and B is unrelated in subject matter to A, will the referring search keywords to content piece A hold for content piece B?
For example, I have a piece of content about furniture and it ranks in top 5 in the SERPs for the phrase "furniture". If I were to 301 redirect that piece of furniture content to a piece of content about trucks, would the referring keyword "furniture" continue to rank over time for the trucks content? My instincts tell me that in the short term the content piece about trucks would receive traffic for the term "furniture", but over the long term, the trucks content would lose rankings for the term "furniture" since the piece has to do with trucks and not furniture. Any thoughts?
Technical SEO | | pbrothers240