URLs: Removing duplicate pages using anchor?
-
I've been working on removing duplicate content on our website. There are tons of pages created based on size but the content is the same.
The solution was to create a page with 90% static content and 10% dynamic, that changed depending on the "size" Users can select the size from a dropdown box.
So instead of 10 URLs, I now have one URL.
- Users can access a specific size by adding an anchor to the end of the URL (?f=suze1, ?f=size2)
For e.g:
Old URLs.
- www.example.com/product-alpha-size1
- www.example.com/product-alpha-size2
- www.example.com/product-alpha-size3
- www.example.com/product-alpha-size4
- www.example.com/product-alpha-size5
New URLs
- www.example.com/product-alpha-size1
- www.example.com/product-alpha-size1?f=size2
- www.example.com/product-alpha-size1?f=size3
- www.example.com/product-alpha-size1?f=size4
- www.example.com/product-alpha-size1?f=size5
Do search engines read the anchor or drop them? Will the rank juice be transfered to just www.example.com/product-alpha-size1?
-
Thanks Everett,
- Rel="canonical" is in place, so that's covered
- The urls with the parameter are only accessible if you want to directly access a particular size. If you are on the default page and switch sizes from the dropdown, no URL change is presented.
- I have left webmaster to decide what should be crawled or not. The parameter has been mentioned though.
-
Cyto,
The Google Webmaster Tools parameter handling, in my opinion, is often best left up to Google. In other words, I rarely change it. Instead, I try to fix the issue itself. In your case, here is what I would advise:
Instead of using a parameter in the URL, use cookies or hidden divs to change the content on the page to the different size. Have a look at most major online retailers. You can select a size or color from the drop down and it never changes the URL.
If this is not possible, I recommend the following:
Ensure the rel = "canonical" tag on all of those pages references the canonical version (e.g. /product-alpha-size1) which will consolidate the link-related metrics like PageRank into the one page.
-
Please say YES
-
Thank you Celilcan2,
- I'll set it up as 'yes' and it 'narrows' the page
- What is the perk of doing this though? Will Google not count anything after the parameter as something or value, it would focus on just the single URL?
-
Go to google webmaster tools
- On the Dashboard, under Crawl, click URL Parameters.
- Next to the parameter you want, click Edit. (If the parameter isn’t listed, click Add parameter. Note that this tool is case sensitive, so be sure to type your parameter exactly as it appears in your URL.)
- If the parameter doesn't affect the content displayed to the user, select **No ... **in the Does this parameter change... list, and then click Save. If the parameter does affect the display of content, click Yes: Changes, reorders, or narrows page content, and then select how you want Google to crawl URLs with this parameter.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Should I use noindex or robots to remove pages from the Google index?
I have a Magento site and just realized we have about 800 review pages indexed. The /review directory is disallowed in robots.txt but the pages are still indexed. From my understanding robots means it will not crawl the pages BUT if the pages are still indexed if they are linked from somewhere else. I can add the noindex tag to the review pages but they wont be crawled. https://www.seroundtable.com/google-do-not-use-noindex-in-robots-txt-20873.html Should I remove the robots.txt and add the noindex? Or just add the noindex to what I already have?
Intermediate & Advanced SEO | | Tylerj0 -
Duplicate Page Due To Website Display Function
Hi Can anyone help with how I can rectify a duplicate issue? A high priority on my Moz report shows a duplicate issue however, this is due to the way the website is structured. For example. the below duplicate is created due to the website having a function to display all trips, so customers do not need to search page by page i.e: http://www.bikecation.co.uk/categories/cycling-climbs http://www.bikecation.co.uk/categories/cycling-climbs/page/2?showall=1 My question is, Will this format damage the SEO for this page? Is there a way to rectify? Would a canonical tag work in this case? Many Thanks Claire
Intermediate & Advanced SEO | | Strateji0 -
Need to change 1 million page URLs
Hey all, I have a community site where users are uploading photos and videos. Launched in 2003, back then it wasn't such a bad idea to use keywords/tags in the URLs, so I did that. All my content pages (individual photo/video) are looking like this: www.domain.com/12345-kw1-kw2-kw3-k4-k5 and so on. Where the 12345 is the unique content ID and the rest are keywords/tags added by the uploader. I would like to get rid of of the keywords after the ID in the URL. My site is well coded, so this can be easily done by changing a simple function, so my content page URLs become this: www.domain.com/ID What is the best course of action? 301 the KW URLs to non-KW version? Canonical? I really want to do this the proper way. Any advice is highly appreciated. Thanks in advance.
Intermediate & Advanced SEO | | mlqsko0 -
Duplicate content - how to diagnose duplicate content from another domain before publishing pages?
Hi, 🙂 My company is having new distributor contract, and we are starting to sell products on our own webshop. Bio-technology is an industry in question and over 1.000 products. Writing product description from scratch would take many hours. The plan is to re-write it. With permission from our contractors we will import their 'product description' on our webshop. But, I am concerned being penalies from Google for duplicate content. If we re-write it we should be fine i guess. But, how can we be sure? Is there any good tool for comparing only text (because i don't want to publish the pages to compare URLs)? What else should we be aware off beside checking 'product description' for duplicate content? Duplicate content is big issue for all of us, i hope this answers will be helpful for many of us. Keep it hard work and thank you very much for your answers, Cheers, Dusan
Intermediate & Advanced SEO | | Chemometec0 -
Duplicate Content for Deep Pages
Hey guys, For deep, deep pages on a website, does duplicate content matter? The pages I'm talk about are image pages associated with products and will never rank in Google which doesn't concern me. What I'm interested to know though is whether the duplicate content would have an overall effect on the site as a whole? Thanks in advance Paul
Intermediate & Advanced SEO | | kevinliao1 -
Using abbreviations in URL - Matching Keyword
We have a website that uses /us/, /ca/, /va/, etc for URLs of the different U.S. states. How much better is it (or is it at all better) to use /california/ or /virginia/ instead in our URLs to rank for searches that include the name of those states?
Intermediate & Advanced SEO | | Heydarian0 -
Removing a Page From Google index
We accidentally generated some pages on our site that ended up getting indexed by google. We have corrected the issue on the site and we 404 all of those pages. Should we manually delete the extra pages from Google's index or should we just let Google figure out that they are 404'd? What the best practice here?
Intermediate & Advanced SEO | | dbuckles0 -
Paging. is it better to use noindex, follow
Is it better to use the robots meta noindex, follow tag for paging, (page 2, page 3) of Category Pages which lists items within each category or just let Google index these pages Before Panda I was not using noindex because I figured if page 2 is in Google's index then the items on page 2 are more likely to be in Google's index. Also then each item has an internal link So after I got hit by panda, I'm thinking well page 2 has no unique content only a list of links with a short excerpt from each item which can be found on each items page so it's not unique content, maybe that contributed to Panda penalty. So I place the meta tag noindex, follow on every page 2,3 for each category page. Page 1 of each category page has a short introduction so i hope that it is enough to make it "thick" content (is that a word :-)) My visitors don't want long introductions, it hurts bounce rate and time on site. Now I'm wondering if that is common practice and if items on page 2 are less likely to be indexed since they have no internal links from an indexed page Thanks!
Intermediate & Advanced SEO | | donthe0