Why isn't my uneven link flow among index pages causing uneven search traffic?
-
I'm working with a site that has millions of pages. The link flow through index pages is atrocious, such that for the letter A (for example) the index page A/1.html has a page authority of 25 and the next pages drop until A/70.html (the last index page listing pages that start with A) has a page authority of just 1. However, the pages linked to from the low page authority index pages (that is, the pages whose second letter is at the end of the alphabet) get just as much traffic as the pages linked to from A/1.html (the pages whose second letter is A or B). The site gets a lot of traffic and has a lot of pages, so this is not just a statistical biip. The evidence is overwhelming that the pages from the low authority index pages are getting just as much traffic as those getting traffic from the high authority index pages. Why is this? Should I "fix" the bad link flow problem if traffic patterns indicate there's no problem? Is this hurting me in some other way? Thanks
-
Thanks Everett, I appreciate it!
-
Hello Gil,
With regard to user-generated profile pages, I typically recommend to clients that they noindex,follow these until they reach a minimum threshold of completeness (e.g. 75% complete) to avoid filling the index with thin "stub" pages, or those created by spam profiles.
If these are local business type pages, as in the White Pages example, the more "supporting content" you customize those pages with the better. For example, a local business listing page could link to similar businesses in the area, provide star ratings, allow visitors to leave reviews/comments, share demographics data for the area, include links to the business' social profiles, embedded videos (commercials, etc...) for the business and many other things.
I realize these pages might be getting traffic at the moment, but as Google updates the machine learning algo to incorporate feedback from the quality raters, who are now being asked to look at supporting content, your client may find their traffic to those pages (and indeed the site as a whole) slowly declining over the next year or two.
That's about as far as I can take it without seeing the pages. Good luck and I hope we've been of some assistance!
-
Hi Travis,
Thanks for your reply.
As I just wrote to Everett, I can't share too many details for confidentiality reasons. My site is somewhat similar to WhitePages, where http://www.whitepages.com/ind/p-001 has a Moz Page Authority of 45, but http://www.whitepages.com/ind/p-150 has a Moz PA of 1. We have similar PA distribution among our index pages, but our organic search traffic is just as high when linked to from the PA 1 pages and when linked to from the PA 45 pages. So I don't know if my client should spend time fixing the problem.
Thanks
-
Thanks. I can't share too many details for confidentiality reasons. I realize that makes it hard / impossible to diagnose correctly, and I'm sorry about that.
These are person pages. The site's link structure naturally gives more link power to the people with the most connections. We could NoIndex (or mask links to) pages that don't have much information but I think such a system would probably be complex and may backfire.
So there's not the kind of taxonomy / directory / long-tail keyword structure that you would expect from a large product directory (for example).
Let's pretend we're discussing WhitePages.com where http://www.whitepages.com/ind/p-001 has a Moz Page Authority of 45, but http://www.whitepages.com/ind/p-150 has a Moz PA of 1. I can fix the problem and get the back pages to have higher PA, but I can't recommend that my client spend resources to fix this since the pages at the back of the index get just as much organic search traffic as the pages at the top.
Thanks
-
As others have stated, we can't really say much with certainty unless we view the site. However, here are my two pennies anyway...
The farther you go down into the directory structure (assuming you have a logical taxonomy and site architecture) the more long-tail and specific the keywords will be. The more long-tail and specific the topic, the less page authority is needed to rank.
With that said, if I was working on a site with millions of pages I'd look into doing a content audit to determine which ones even SHOULD be in the index. Very few sites can scale quality landing pages into the millions.
-
You shouldn't expect anyone to solve anything that technical, with any sort of certainty, without stating the actual domain.
If it's getting organic traffic, great. Could it get more? Maybe.
No one can speak with any sort of certainty based upon what you have written at this point.
Apologies if I appear a little cranky. I'm getting tired of all of these; "I have a problem with a bajillion possible issues, but I won't tell you what I'm looking at." questions.
So you can always PM me, I'm not coming after your client. The problem is more interesting.
-
Yes, an even distribution of organic search traffic seems to indicate that the pages are indexed and ranking. Gains might be made via external links, but as far as modifying your link flow goes, it doesn't seem like the site needs it based on what you've described.
-
Thanks. Sorry I wasn't clear, when I say the traffic is pretty evenly distributed among the pages, I'm referring specifically to organic traffic. I'm wondering if the relatively even distribution of organic traffic is proof that better balancing the link flow won't increase traffic.
-
If you're speaking in terms of just organic search visits it doesn't seem to be a problem, but "traffic" in your example is a little broad. There could be paid search being targeted to those pages, or some sort of social media mechanism that causes people to visit their specific page, or so on.
A segmented look at your analytics for the site (or site section) will give you a good idea of whether or not the pages have a problem getting organic search traffic. If they don't I wouldn't worry about link flow. Really the main reason to adjust it is if you're lacking indexation or rank, and so far from what you've described you're not.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
After hack and remediation, thousands of URL's still appearing as 'Valid' in google search console. How to remedy?
I'm working on a site that was hacked in March 2019 and in the process, nearly 900,000 spam links were generated and indexed. After remediation of the hack in April 2019, the spammy URLs began dropping out of the index until last week, when Search Console showed around 8,000 as "Indexed, not submitted in sitemap" but listed as "Valid" in the coverage report and many of them are still hack-related URLs that are listed as being indexed in March 2019, despite the fact that clicking on them leads to a 404. As of this Saturday, the number jumped up to 18,000, but I have no way of finding out using the search console reports why the jump happened or what are the new URLs that were added, the only sort mechanism is last crawled and they don't show up there. How long can I expect it to take for these remaining urls to also be removed from the index? Is there any way to expedite the process? I've submitted a 'new' sitemap several times, which (so far) has not helped. Is there any way to see inside the new GSC view why/how the number of valid URLs in the indexed doubled over one weekend?
Intermediate & Advanced SEO | | rickyporco0 -
Crawl and Indexation Error - Googlebot can't/doesn't access specific folders on microsites
Hi, My first time posting here, I am just looking for some feedback on a indexation issue we have with a client and any feedback on possible next steps or items I may have overlooked. To give some background, our client operates a website for the core band and a also a number of microsites based on specific business units, so you have corewebsite.com along with bu1.corewebsite.com, bu2.corewebsite.com. The content structure isn't ideal, as each microsite follows a structure of bu1.corewebsite.com/bu1/home.aspx, bu2.corewebsite.com/bu2/home.aspx and so on. In addition to this each microsite has duplicate folders from the other microsites so bu1.corewebsite.com has indexable folders bu1.corewebsite.com/bu1/home.aspx but also bu1.corewebsite.com/bu2/home.aspx the same with bu2.corewebsite.com has bu2.corewebsite.com/bu2/home.aspx but also bu2.corewebsite.com/bu1/home.aspx. Therre are 5 different business units so you have this duplicate content scenario for all microsites. This situation is being addressed in the medium term development roadmap and will be rectified in the next iteration of the site but that is still a ways out. The issue
Intermediate & Advanced SEO | | ImpericMedia
About 6 weeks ago we noticed a drop off in search rankings for two of our microsites (bu1.corewebsite.com and bu2.corewebsite.com) over a period of 2-3 weeks pretty much all our terms dropped out of the rankings and search visibility dropped to essentially 0. I can see that pages from the websites are still indexed but oddly it is the duplicate content pages so (bu1.corewebsite.com/bu3/home.aspx or (bu1.corewebsite.com/bu4/home.aspx is still indexed, similiarly on the bu2.corewebsite microsite bu2.corewebsite.com/bu3/home.aspx and bu4.corewebsite.com/bu3/home.aspx are indexed but no pages from the BU1 or BU2 content directories seem to be indexed under their own microsites. Logging into webmaster tools I can see there is a "Google couldn't crawl your site because we were unable to access your site's robots.txt file." This was a bit odd as there was no robots.txt in the root directory but I got some weird results when I checked the BU1/BU2 microsites in technicalseo.com robots text tool. Also due to the fact that there is a redirect from bu1.corewebsite.com/ to bu1.corewebsite.com/bu4.aspx I thought maybe there could be something there so consequently we removed the redirect and added a basic robots to the root directory for both microsites. After this we saw a small pickup in site visibility, a few terms pop into our Moz campaign rankings but drop out again pretty quickly. Also the error message in GSC persisted. Steps taken so far after that In Google Search Console, I confirmed there are no manual actions against the microsites. Confirmed there is no instances of noindex on any of the pages for BU1/BU2 A number of the main links from the root domain to microsite BU1/BU2 have a rel="noopener noreferrer" attribute but we looked into this and found it has no impact on indexation Looking into this issue we saw some people had similar issues when using Cloudflare but our client doesn't use this service Using a response redirect header tool checker, we noticed a timeout when trying to mimic googlebot accessing the site Following on from point 5 we got a hold of a week of server logs from the client and I can see Googlebot successfully pinging the site and not getting 500 response codes from the server...but couldn't see any instance of it trying to index microsite BU1/BU2 content So it seems to me that the issue could be something server side but I'm at a bit of a loss of next steps to take. Any advice at all is much appreciated!0 -
22 Pages 7 Indexed
So I submitted my sitemap to Google twice this week the first time everything was just peachy, but when I went back to do it again Google only indexed 7 out of 22. The website is www.theinboundspot.com. My MOZ Campaign shows no issues and Google Webmaster shows none. Should I just resubmit it?
Intermediate & Advanced SEO | | theinboundspot1 -
Big discrepancies between pages in Google's index and pages in sitemap
Hi, I'm noticing a huge difference in the number of pages in Googles index (using 'site:' search) versus the number of pages indexed by Google in Webmaster tools. (ie 20,600 in 'site:' search vs 5,100 submitted via the dynamic sitemap.) Anyone know possible causes for this and how i can fix? It's an ecommerce site but i can't see any issues with duplicate content - they employ a very good canonical tag strategy. Could it be that Google has decided to ignore the canonical tag? Any help appreciated, Karen
Intermediate & Advanced SEO | | Digirank0 -
Redirect Search Results to Category Pages
I am planning redirect the search results to it's matching category page to avoid having two indexed pages of essentially the same content. Example http://www.example.com/search/?kw=sunglasses
Intermediate & Advanced SEO | | WizardOfMoz
wil be redirected to
http://www.example.com/category/sunglasses/ Is this a good idea? What are the possible negative effect if I go this route? Thanks.0 -
How can i stop such links being indexed
Hi, How can i stop such links being indexed The first link is what i want to stop indexed. We have 1,000's of people writing articles and the below URl shows how many articles each did http://www.somename.com/article/15633 But this is the URl which shows the exact articlehttp://www.Somename.com/article/step-step-installation-ibm-lotus-notesAs both start as thishttp://www.Somename.com/article/How can i set noindex? Should we set for each URL manually one by oneThanks
Intermediate & Advanced SEO | | mtthompsons0 -
Our login pages are being indexed by Google - How do you remove them?
Each of our login pages show up under different subdomains of our website. Currently these are accessible by Google which is a huge competitive advantage for our competitors looking for our client list. We've done a few things to try to rectify the problem: - No index/archive to each login page Robot.txt to all subdomains to block search engines gone into webmaster tools and added the subdomain of one of our bigger clients then requested to remove it from Google (This would be great to do for every subdomain but we have a LOT of clients and it would require tons of backend work to make this happen.) Other than the last option, is there something we can do that will remove subdomains from being viewed from search engines? We know the robots.txt are working since the message on search results say: "A description for this result is not available because of this site's robots.txt – learn more." But we'd like the whole link to disappear.. Any suggestions?
Intermediate & Advanced SEO | | desmond.liang1 -
Anyone managed to change 'At a glance:' in local search results
On Google's local search results, i.e when the 'Google places' data is displayed along with the map on the right hand side of the search results, there is also an element 'At a glance:'
Intermediate & Advanced SEO | | DeanAndrews
The data that if being displayed is from some years ago and the client would if possible like it to reflect there current services, which they have been providing for some five years. According to Google support here - http://support.google.com/maps/bin/answer.py?hl=en&answer=1344353 this cannot be changed, they say 'Can I edit a listing’s descriptive terms or suggest a new one?
No; the terms are not reviewed, curated, or edited. They come from an algorithm, and we do not help that algorithm figure it out. ' My question is has anyone successfully influenced this data and if so how.0