Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Creating 100,000's of pages, good or bad idea
-
Hi Folks,
Over the last 10 months we have focused on quality pages but have been frustrated with competition websites out ranking us because they have bigger sites. Should we focus on the long tail again?
One option for us is to take every town across the UK and create pages using our activities. e.g.
Stirling
Stirling paintball
Stirling Go Karting
Stirling Clay shootingWe are not going to link to these pages directly from our main menus but from the site map.
These pages would then show activities that were in a 50 mile radius of the towns. At the moment we have have focused our efforts on Regions, e.g. Paintball Scotland, Paintball Yorkshire focusing all the internal link juice to these regional pages, but we don't rank high for towns that the activity sites are close to.
With 45,000 towns and 250 activities we could create over a million pages which seems very excessive! Would creating 500,000 of these types of pages damage our site? This is my main worry, or would it make our site rank even higher for the tougher keywords and also get lots of traffic from the long tail like we used to get.
Is there a limit to how big a site should be?
-
Hi Mark!
Thanks for asking this good question. While there is no limit to how big a website can be, I think you can see from the general response here that most members would encourage you to stick to manually developing quality pages rather than automating hundreds of thousands of pages, solely for ranking purposes. I second this advice.
Now, I would like to clarify your business model. Are you a physical, actual business that customers come to, either to buy paintball equipment or to play paintball in a gallery? Or, is your business virtual, with no in person transactions? I'm not quite understanding this from your description.
If the former, I would certainly encourage you to develop a very strong, unique page for each of your physical locations. If you have 10 locations (with unique street addresses and phone numbers), then that would be 10 pages. If you've got 20 locations, that would be 20 pages, etc. But don't approach these with a 'just switch out the city name in the title tags' mindset. Make these pages as exceptional as possible. Tell stories, show off testimonials, share pictures and videos, entertain, educate, inspire. These city landing pages will be intimately linked into your whole Local SEM campaign, provided they each represent a business location with a unique dedicated street address and unique local area code phone number.
But, if you are considering simply building a page for every city in the UK, I just can't see justification for doing so. Ask yourself - what is the value?
There are business models (such as carpet cleaners, chimney sweeps, general contractors, etc.) that go to their clients' locations to serve and for which I would be advising that they create city landing pages for each of their service cities, but this would be extremely regional...not statewide or national or International. A carpet cleaner might serve 15 different towns and cities in his region, and I would encourage him to start gathering project notes and testimonials, videos and photos to begin developing a body of content important enough for him to start creating strong, interesting and unique pages for each of these cities. But I've also had local business owners tell me they want to cover every city in California, for instance, because they think it will help them to do so, and I discourage this.
Even if the business is virtual and doesn't have any in-person transactions with clients or physical locations, I would still discourage this blanketing-the-whole-nation-with-pages approach. A national retailer needs to build up its brand so that it becomes known and visible organically for its products rather than your theoretical approach of targeting every city in the nation. In short order, the mindset behind that approach just doesn't make good horse sense.
And, as others have stated, adding thousands of thin, potentially duplicate pages to any site could definitely have a very negative effect on rankings.
My advice is to make the time to start developing a content strategy for cities in which you have a legitimate presence. If budget means you can't hire a copywriter to help you with this and to speed up the work, accept that this project deserves all the time you can give it and that a slow development of exceptional pages is better than a fast automation of poor quality pages.
Hope this helps!
-
Hi Mark,
If A,C, and E's page is similar to B,D, and F's page it is still consider dupllicate content. Based on Webmaster's definiton:
"Duplicate content generally refers to substantive blocks of content within or across domains that either completely match other content or are appreciably similar"
Each of your pages should be unique and different from other pages.
I suggest you to continue providing quality content and target the long tail keywords. That alone will help you generate more traffic. Furthermore, out ranking is not a problem. You should focus on getting to the frist page (providing quality content with long tail or regular keywords) and when you are on the first page, try to get searchers to click on your link using Title tag and Meta descriptions.
Out ranking just means they are ranked 4th and you are ranked 5th, 6th but as long as you have a better title tag and meta description. I believe searchers will click on the more attractive results.
-
Cookie cutter pages like these stopped working in Google about ten years ago.
If you toss them up I think that your entire site will tank.
I would go back to focusing on quality pages.
-
If the user experience awesome, and people are staying on your site and looking around, great. If you think the 100,000 pages will make search engines love you, machines can never provide the love users can give you.
-
Can you mix content up from your website e.g. paintball site A, C and E on one page and B,D and F on another if the towns are close together? What I'm not sure about is how different in % terms the content actually has to be.
If we have less written content then the amounts of words we have to actually change would be much less.
The challenge we have is we have build the site this time with filtering in mind, so rather than making customers navigate we allow them to be able to search which is much better in terms of getting the activities they want. The downside is now our site does not show for the long tail as we reduced the pages massively.
-
so we dont have the resources if we did it manually but what would happen is the content would be different on each page as we would only show activity sites within a 50 miles radius. And we would make certain text, h1 etc different and relate to the town.
Below are some examples of sites I see doing well ie number 1 using this method
Our content would be much better than say http://www.justpaintballDOTcoDOTuk/site_guide/Aberfeldy.htm or http://www.goballisticDOTcoDOTuk/paintball_in_/ABERFELDY.asp
But as you say getting this wrong is my worry.
-
Hi Mark,
Creating 100,000 pages is definitely good for Search Engine because you have a lot more contents for them to crawl and have more chances your pages might show up on related keywords. However, the problem is do you have enough unique contents you can post on all those 100,000 pages. If you use similar content, I am afraid it will be duplicate contents. You may think changing up the town names will be enough but it might be risky.
If you can create 100,000 unique contents, Sure go ahead. If not, don't take the risk of duplicate contents.
-
Do you have the resources to create unique content for all those pages? Because adding 500,000 pages of duplicate content will absolutely damage your site.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Change Google's version of Canonical link
Hi My website has millions of URLs and some of the URLs have duplicate versions. We did not set canonical all these years. Now we wanted to implement it and fix all the technical SEO issues. I wanted to consolidate and redirect all the variations of a URL to the highest pageview version and use that as the canonical because all of these variations have the same content. While doing this, I found in Google search console that Google has already selected another variation of URL as canonical and not the highest pageview version. My questions: I have millions of URLs for which I have to do 301 and set canonical. How can I find all the canonical URLs that Google has autoselected? Search Console has a daily quota of 100 or something. Is it possible to override Google's version of Canonical? Meaning, if I set a variation as Canonical and it is different than what Google has already selected, will it change overtime in Search Console? Should I just do a 301 to highest pageview variation of the URL and not set canonicals at all? This way the canonical that Google auto selected might get redirected to the highest pageview variation of the URL. Any advice or help would be greatly appreciated.
Intermediate & Advanced SEO | | SDCMarketing0 -
One of my Friend's website Domain Authority is Reducing? What could be the reason?
Hello Guys, One of my friend's website domain authority is decreasing since they have moved their domain from HTTP to https.
Intermediate & Advanced SEO | | Max_
There is another problem that his blog is on subfolder with HTTP.
So, can you guys please tell me how to fix this issue and also it's losing some of the rankings like 2-5 positions down. Here is website URL: myfitfuel.in/
here is the blog URL: myfitfuel.in/mffblog/0 -
Geo-Redirect: good idea or not?
Hi Mozzers, The background: I have this very corporate .com domain which is used worldwide. Next to that, we have another .com domain which is specifically created for the US visitors. Within the organic rankings, we notice that our corporate domain is ranking much better in the US. Many visitors are arriving on this domain. As it is a corporate domain being used worldwide, they get lost. My questions: I know there are ways to redirect by location. Would it be smart to automatically redirect US visitors for the corporate domain to the commercial US-specific domain? Is it possible to only redirect US visitors and leave the website as it is for visitors from other countries. Won't this harm the corporate website (organically) worldwide? If this would be a good idea, any recommended plugins or concrete procedures? Thank you so much for helping me out!
Intermediate & Advanced SEO | | WeAreDigital_BE
Sander0 -
Javascript to fetch page title for every webpage, is it good?
We have a zend framework that is complex to program if you ask me, and since we have 20k+ pages that we need to get proper titles to and meta descriptions, i need to ask if we use Javascript to handle page titles (basically the previously programming team had NOT set page titles at all) and i need to get proper page titles from a h1 tag within the page. current course of action which we can easily implement is fetch page title from that h1 tag being used throughout all pages with the help of javascript, But this does makes it difficult for engines to actually read what's the page title? since its being fetched with javascript code that we have put in, though i had doubts, is anyone one of you have simiilar situation before? if yes i need some help! Update: I tried the JavaScript way and here is what it looks like http://islamicencyclopedia.org/public/index/hadith/id/1/book_id/106 i know the fact that google won't read JavaScript like the way we have done with the website, But i need help on "How we can work around this issue" Knowing we don't have other options.
Intermediate & Advanced SEO | | SmartStartMediacom0 -
What's the deal with significantLinks?
http://schema.org/significantLink Schema.org has a definition for "non-navigation links that are clicked on the most." Presumably this means something like the big green buttons on Moz's homepage. But does anyone know how they affect anything? In http://moz.com/blog/schemaorg-a-new-approach-to-structured-data-for-seo#comment-142936, Jeremy Nelson says " It's quite possible that significant links will pass anchor text as well if a previous link to the page was set in navigation, effictively making obselete the first-link-counts rule, and I am interested in putting that to test." This is a pretty obscure comment but it's one of the only results I could find on the subject. Is this BS? I can't even make out what all of it is saying. So what's the deal with significantLinks and how can we use them to SEO?
Intermediate & Advanced SEO | | NerdsOnCall0 -
Targeting local areas without creating landing pages for each town
I have a large ecommerce website which is structured very much for SEO as it existed a few years ago. With a landing page for every product/town nationwide (its a lot of pages). Then along came Panda... I began shrinking the site in Feb last year in an effort to tackle duplicate content. We had initially used a template only changing product/town name. My first change was to reduce the amount of pages in half by merging the top two categories, as they are semantically similar enough to not need their own pages. This worked a treat, traffic didn't drop at all and the remaining pages are bringing in the desired search terms for both these products. Next I have rewritten the content for every product to ensure they are now as individual as possible. However with 46 products and each of those generating a product/area page we still have a heap of duplicate content. Now i want to reduce the town pages, I have already started writing content for my most important areas, again, to make these pages as individual as possible. The problem i have is that nobody can write enough unique content to target every town in the UK via an individual page (times by 46 products), so i want to reduce these too. QUESTION: If I have a single page for "croydon", will mentioning other local surrounding areas on this page, such as Mitcham, be enough to rank this page for both towns? I have approx 25 Google local place/map listings and grwoing, and am working from these areas outwards. I want to bring the site right down to about 150 main area pages to tackle all the duplicate content, but obviously don't want to lose my traffic for so many areas at once. Any examples of big sites that have reduced in size since Panda would be great. I have a headache... Thanks community.
Intermediate & Advanced SEO | | Silkstream0 -
What's the best way to redirect categories & paginated pages on a blog?
I'm currently re-doing my blog and have a few categories that I'm getting rid of for housecleaning purposes and crawl efficiency. Each of these categories has many pages (some have hundreds). The new blog will also not have new relevant categories to redirect them to (1 or 2 may work). So what is the best place to properly redirect these pages to? And how do I handle the paginated URLs? The only logical place I can think of would be to redirect them to the homepage of the blog, but since there are so many pages, I don't know if that's the best idea. Does anybody have any thoughts?
Intermediate & Advanced SEO | | kking41200 -
Xml sitemap advice for website with over 100,000 articles
Hi, I have read numerous articles that support submitting multiple XML sitemaps for websites that have thousands of articles... in our case we have over 100,000. So, I was thinking I should submit one sitemap for each news category. My question is how many page levels should each sitemap instruct the spiders to go? Would it not be enough to just submit the top level URL for each category and then let the spiders follow the rest of the links organically? So, if I have 12 categories the total number of URL´s will be 12??? If this is true, how do you suggest handling or home page, where the latest articles are displayed regardless of their category... so I.E. the spiders will find l links to a given article both on the home page and in the category it belongs to. We are using canonical tags. Thanks, Jarrett
Intermediate & Advanced SEO | | jarrett.mackay0