Certain Product Pages Not Indexing
-
Hey All,
We discovered an issue where new product pages on our site were not getting indexed because a "noindex" tag was inadvertently being added to section when those pages were created.
We removed the noindex tag in late April and some of the pages that had not been previously indexed are now showing up, but others are still not getting indexed and I'd appreciate some help on why this could be.
Here is an example of a page that was not in the index but is now showing after removal of noindex:
http://www.cloud9living.com/san-diego/gaslamp-quarter-food-tour
And here is an example of a page that is still not showing in the index:
http://www.cloud9living.com/atlanta/race-a-ferrari
UPDATE: The above page is now showing after I manually submitted it in WMT. I had previously submitted another page like a month ago and it was still not indexing so I thought the manual submission was a dead end. However, it just so happens that the above URL just had its Page Title and H1 updated to something more specific and less duplicative so I am currently running a test to see if that's the problem with these pages not indexing. Will update this soon.
Any suggestions? Thanks!
-
Significantly changing the Page Title and H1 is working. Second page now indexing after not indexing for some time. Probably shoulda thought of that a long time ago but that noindex tag sidetracked me!
-
It's hard to say, and I admit that I would have also expected them to reindex the pages by now.
A while back I was working for a client who accidentally turned on noindex, nofollow in their WordPress SEO by Yoast plugin site-wide. I didn't catch it for a week, and after I turned it off it took an additional 3 weeks before a single page of the site was reindexed. Granted, this was a low ranked site, and it probably wasn't high on Google's priority, but it did take much longer than I hoped to recover from.
Unfortunately I think you just have to wait it out. Just keep doing what your doing, creating new content, etc. Maybe if you build a new link to the page, Google will recrawl it then?
-
Good point. But why then would they continue to not index a month after manual submission?
-
If the page was set to "noindex" for a long time, Google may have flagged the page as such and chosen to skip over it when it was crawling your site.
-
Also, historically indexation happened very quickly on this site (less than 24 hours) so that's why I think something else is afoot here. And it has been like 6 weeks... which I don't think makes sense for a site with this level of domain authority.
-
Hey Bradley,
Thanks for the response. Yes, I had manually fetched a few of these pages about a month back and that didn't change indexation so I thought it was a dead end. However, one I tried again this morning suddenly indexed and it just so happened to also have had its Page Title and H1 tag changed to be significantly more unique than they were previously so I am wondering if that is problem. I am currently running a test with another page that I manually submitted a month ago but without updating Page Title/H1 and now I have resubmitted with changed info.
We'll see if that does the trick.
Will let you know.
-
You can ask Google to crawl your page: http://support.google.com/webmasters/bin/answer.py?hl=en&answer=1352276
Ask Google to crawl a page or site:
- On the Webmaster Tools Home page, click the site you want.
- On the Dashboard, under Health, click Fetch as Google.
- In the text box, type the path to the page you want to check.
- In the dropdown list, select Web. (You can select another type of page, but currently we only accept submissions for our Web Search index.)
- Click Fetch. Google will fetch the URL you requested. It may take up to 10 or 15 minutes for Fetch status to be updated.
- Once you see a Fetch status of "Successful", click Submit to Index, and then click one of the following:
- To submit the individual URL to Google's index, select URL and click Submit. You can submit up to 500 URLs a week in this way.
- To submit the URL and all pages linked from it, click URL and all linked pages. You can submit up to 10 of these requests a month.
It's probably just Google slowly making their way around to re-crawling these pages. I would fetch the page, and just wait a little while longer.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Domain Authority Dropped and Indexed Pages Went Down on Google?
Hi there, We run an e-commerce site on Shopify. Our Domain Authority was 28 at the start of our campaign in May of this year. We also had 610 indexed pages on Google. We did some SEO work which included: Renaming Images for SEO Adding in alt tags Optimizing the meta title to "Product Name - Keyword - Brand Name" for products Optimizing meta descriptions Transition of Hubspot blog to Shopify (it was on a subdomain at Hubspot previously) Fixing some 404s Resubmitting site map after the changes Now it is almost at the 3-month mark and it looks like our Domain Authority has gone down 4 points to 24. The # of indexed pages has gone to down to 555. We made sure all our SEO updates weren't spammy or keyword-stuffed, but took a natural and helpful-sounding approach. We followed guidelines. So there shouldn't be any penalty right? I checked site traffic and it does not coincide with the drop. Our site traffic remains steady. I also looked at "site:" as well as conducted some test searches for the important pages (i.e. main pages, blog pages, and product pages) and they still come up on Google. So could it only be non-important pages being deindexed? My questions are: Why did both the Domain Authority and # of indexed pages go down? Is there any way to see which pages were deindexed? I checked Google Search Console, but couldn't find it. Thank you!
Intermediate & Advanced SEO | | kindalpaca70 -
Consolidating product pages during website migration
Hello, We are an e-commerce & content site undergoing a website migration and redesign in the coming months. We will be getting an entirely new website. Many of our URLs will be changing: Current URL setup: www.mysite.com/catalog/SKU12345/product-title-here
Intermediate & Advanced SEO | | katelynroberts
Future URL setup: www.mysite.com/catalog/product-title-here So we're aware we will be using plenty of 301 redirects to achieve this. Further to this though, we currently have a product page for each configuration of a product - for example, a single-sided bookmark has its own page and URL, and the double-sided version of the same bookmark has its own page and URL. In our site redesign, we are hoping to consolidate each of these instances into one product page where users can select single or double-sided and the price will update accordingly. The bookmark URLs would then go from:
_www.mysite.com/catalog/SKU12345/bookmark-single-sided _(call this URL A for simplicity)www.mysite.com/catalog/SKU67890/bookmark-double-sided (call this URL B) To (after migrating to the new URL structure for the new site, and the now-consolidated single- & double-sided product pages):
www.mysite.com/catalog/bookmark (call this URL C) What is the best way to make this transition without losing too much of our SEO value? I understand there is nearly always traffic loss with URL changes but I'd like to at least minimize the damage as best I can. We have backlinks and ranks for many product pages so I want to make sure we pass as much of this as we can. (And is this at all further complicated by the fact that URL A & B won't exist on the new site, and URL C doesn't exist on the current site? Does this impact the use of the 301 redirects and if so, how?) Are we better off to approach this page consolidation after the site migration and treat it as a separate project? This is something that is important to our user experience, and is definitely a change we want to make. Any advice is appreciated - thank you! I'm a fairly beginner-intermediate SEO so this is all somewhat new but I want to be able to at least convey some understanding to our developer of what we need to do. I was able to find this discussion (https://moz.com/community/q/merging-pages-and-seo) which describes a similar situation and solutions if we were just consolidating the pages but doesn't quite have the complicating factor of the entire site migration happening at the same time. Thanks so much!0 -
How do we decide which pages to index/de-index? Help for a 250k page site
At Siftery (siftery.com) we have about 250k pages, most of them reflected in our sitemap. Though after submitting a sitemap we started seeing an increase in the number of pages Google indexed, in the past few weeks progress has slowed to a crawl at about 80k pages, and in fact has been coming down very marginally. Due to the nature of the site, a lot of the pages on the site likely look very similar to search engines. We've also broken down our sitemap into an index, so we know that most of the indexation problems are coming from a particular type of page (company profiles). Given these facts below, what do you recommend we do? Should we de-index all of the pages that are not being picked up by the Google index (and are therefore likely seen as low quality)? There seems to be a school of thought that de-indexing "thin" pages improves the ranking potential of the indexed pages. We have plans for enriching and differentiating the pages that are being picked up as thin (Moz itself picks them up as 'duplicate' pages even though they're not. Thanks for sharing your thoughts and experiences!
Intermediate & Advanced SEO | | ggiaco-siftery0 -
Home page suddenly dropped from index!!
A client's home page, which has always done very well, has just dropped out of Google's index overnight!
Intermediate & Advanced SEO | | Caro-O
Webmaster tools does not show any problem. The page doesn't even show up if we Google the company name. The Robot.txt contains: Default Flywheel robots file User-agent: * Disallow: /calendar/action:posterboard/
Disallow: /events/action~posterboard/ The only unusual thing I'm aware of is some A/B testing of the page done with 'Optimizely' - it redirects visitors to a test page, but it's not a 'real' redirect in that redirect checker tools still see the page as a 200. Also, other pages that are being tested this way are not having the same problem. Other recent activity over the last few weeks/months includes linking to the page from some of our blog posts using the page topic as anchor text. Any thoughts would be appreciated.
Caro0 -
Google is indexing wrong page for search terms not on that page
I’m having a problem … the wrong page is indexing with Google, for search phrases “not on that page”. Explained … On a website I developed, I have four products. For example sake, we’ll say these four products are: Sneakers (search phrase: sneakers) Boots (search phrase: boots) Sandals (search phrase: sandals) High heels (search phrase: high heels) Error: What is going “wrong” is … When the search phrase “high heels” is indexed by Google, my “Sneakers” page is being indexed instead (and ranking very well, like #2). The page that SHOULD be indexing, is the “High heels” page (not the sneakers page – this is the wrong search phrase, and it’s not even on that product page – not in URL, not in H1 tags, not in title, not in page text – nowhere, except for in the top navigation link). Clue #1 … this same error is ALSO happening for my other search phrases, in exactly the same manner. i.e. … the search phrase “sandals” is ALSO resulting in my “Sneakers” page being indexed, by Google. Clue #2 … this error is NOT happening with Bing (the proper pages are correctly indexing with the proper search phrases, in Bing). Note 1: MOZ has given all my product pages an “A” ranking, for optimization. Note 2: This is a WordPress website. Note 3: I had recently migrated (3 months ago) most of this new website’s page content (but not the “Sneakers” page – this page is new) from an old, existing website (not mine), which had been indexing OK for these search phrases. Note 4: 301 redirects were used, for all of the OLD website pages, to the new website. I have tried everything I can think of to fix this, over a period of more than 30 days. Nothing has worked. I think the “clues” (it indexes properly in Bing) are useful, but I need help. Thoughts?
Intermediate & Advanced SEO | | MG_Lomb_SEO0 -
Glossary index and individual pages create duplicate content. How much might this hurt me?
I've got a glossary on my site with an index page for each letter of the alphabet that has a definition. So the M section lists every definition (the whole definition). But each definition also has its own individual page (and we link to those pages internally so the user doesn't have to hunt down the entire M page). So I definitely have duplicate content ... 112 instances (112 terms). Maybe it's not so bad because each definition is just a short paragraph(?) How much does this hurt my potential ranking for each definition? How much does it hurt my site overall? Am I better off making the individual pages no-index? or canonicalizing them?
Intermediate & Advanced SEO | | LeadSEOlogist0 -
Why are some pages indexed but not cached by Google?
The question is simple but I don't understand the answer. I found a webpage that was linking to my personal site. The page was indexed in Google. However, there was no cache option and I received a 404 from Google when I tried using cache:www.thewebpage.com/link/. What exactly does this mean? Also, does it have any negative implication on the SEO value of the link that points to my personal website?
Intermediate & Advanced SEO | | mRELEVANCE0 -
Why are new pages not being indexed, and old pages (now in robots.txt) remain in the index?
I currently have a site that was recently restructured, causing much of its content to be reposted, creating new URL's for each page. To avoid duplicates, all of the existing pages were added to the robots file. That said, it has now been over a week - I know Google has recrawled the site - and when I search for term X, it is stil the old page that is ranking, with the new one nowhere to be seen. I'm assuming it's a cached version, but why are so many of the old pages still appearing in the index? Furthermore, all "tags" pages (it's a Q&A site, like this one) were also added to the robots a few months ago, yet I think they are all still appearing in the index. Anyone got any ideas about why this is happening, and how I can get my new pages indexed?
Intermediate & Advanced SEO | | corp08030