Should I check Use noindex for Tag Archives?
-
I have a page indexed > (http://mysite.com/mypost) and also http://mysite.com/tag/mypost The same post shows up twice, one with /tag/ one without when I search site:http://mysite.com
Is this a duplicate content?? Can I get penalized for this?In the All in one plugin should I check Use noindex for Tag Archives to avoid this or doesn't matter.
Thanks -
very good and well thought out answer.
-
thanks bemcapaz & Marcus for expert advice. wht u think if we control how much content to be shown on tag pages .. like this http://www.lancelhoff.com/change-the-excerpt-length-wordpress/. is it ok? actually i m getting traffic from TAG pages too... .
-
Most of my sites uses Wordpress, what I did to never have any duplicate content is the following:
First, I have the following plugins installed
- .html on PAGES
- All in ONe SEO Pack
- cbnet Ping Optimizer
- Google XML Sitemaps
Then, added the following text to the robots.txt
User-agent: *
Disallow: /wp-adminDisallow: /wp-includes
Disallow: /wp-content/plugins
Disallow: /wp-content/cache
Disallow: /wp-content/themes
Disallow: /trackbackDisallow: /comment
Disallow: /categoria//
Disallow: */trackback
Disallow: */comments
Disallow: /sem-categoria
Disallow: /pollsarchive
Disallow: /category
Disallow: /?
Disallow: /*?
Disallow: /*.php$
Disallow: /*.js$
Disallow: /*.inc$
Disallow: /*.css$
Disallow: /*.gz$
Disallow: /*.wmv$
Disallow: /*.cgi$
Disallow: /*.xhtml$
Allow: /wp-content/uploads
Google ImageUser-agent: Googlebot-ImageDisallow:Allow: /* # Google AdSenseUser-agent: Mediapartners-GoogleDisallow:Allow: / # digg mirrorUser-agent: duggmirrorDisallow: / Sitemap: http://www.YOURSITE.com.br/sitemap.xml
On the admin in the PERMALINK config tab, in the COMMON SETTINGS i choose
- Custom Structure -> /%postname%.html
In SETTINGS > cbnet Ping Optimizer I inserted
http://blogsearch.google.com/ping/RPC2
http://ping.weblogalot.com/rpc.php
http://ping.syndic8.com/xmlrpc.php
http://rpc.technorati.com/rpc/ping
http://rpc.reader.livedoor.com/ping
http://www.blogpeople.net/servlet/weblogUpdates
http://audiorpc.weblogs.com/RPC2
I use this plugin to make sure that wordpress will not ping those services everytime I change or update something on the page,actually I seted the plugin to ping only after 30 min after the last ping.
So If you just posted something on Wordpress it will ping all those services for a fast index, however, if for some reason you have to edit the post when you save this plugin will make sure the services will not get pingged again in such a short space of time.
In the All in One Seo config I have the following config
UNCHECKED
- Use noindex for Categories
- Use noindex for Tag Archives
- Autogenerate Description
I also set Exlcude Pages, Additional Post Headers, Additional Page Headers and Additional Home Headers all blank
With all this configuration if someone access any post of my blogs they get a page with http://domain/post permalink structure, doesnt matter if the access came from a TAG, Categorie or normal Search.
Aditionally the main Tag page and Categorie page are indexed, so i ended up using my tags for some sort of relevance to the posts that belongs to that Tag. In google search the same post could appear by many related TAGs filter but the content of the post appears in an unique page.
Hope that helps
PS.: Suggestions to improve this config are welcome
-
Hey, I am guessing this is a wordpress site?
You could solve this a couple of ways
- 301 redirect the duplicate page - not recommended as it is a valid page but it may work
- canonical link on both pages showing the main version of this content
- noindex the tag page
Alternatively, you can use options 2 & 3 and it will resolve it.
If it's wordpress, happy to take a look at the actual page if it helps? I spend a lot of time tinkering with wordpress so there maybe another way & this answer is based on some assumption without real links.
Marcus
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Using copy from a current site on a new one
I have a client who is closing down his local business because he'smoving to another state. When he gets there he will launch a new website.On his current website, he put in a lot of work and has a ton of good copy, including blog posts that have helped gain him excellent rankings.He's asking me if he can use that copy on his new site and get original author credit for that, like he did on his current site.Can he use the same copy from his current website on his new websitewithout any problems — and get original author credit for it?Would it be best to shut down the old site or to 301 all of the pages beingmoved to the new corresponding pages?If 301's are the way to go, how long should he leave those in place?Thanks!Kirk
Intermediate & Advanced SEO | | kbates1 -
How many images should I use in structured data for a product?
We have a basic printing website that offers business cards. Each type of business card has a few product images. Should we use structured data for all the images, or just the main image? What is your opinion about this? Thanks in advance.
Intermediate & Advanced SEO | | Choice0 -
Google SERPs displaying Tracking Tags
Hello, I'm hoping someone can help me! Can you tell me why Google would be displaying the tracking URLs in the SERPs (screenshot - http://i.imgur.com/gbskD26.jpg)? I'm thinking it may have to do with the canonical URLs, but I'm not sure.. Thanks in advance! gbskD26.jpg
Intermediate & Advanced SEO | | Mindstream_Media0 -
How do you check the google cache for hashbang pages?
So we use http://webcache.googleusercontent.com/search?q=cache:x.com/#!/hashbangpage to check what googlebot has cached but when we try to use this method for hashbang pages, we get the x.com's cache... not x.com/#!/hashbangpage That actually makes sense because the hashbang is part of the homepage in that case so I get why the cache returns back the homepage. My question is - how can you actually look up the cache for hashbang page?
Intermediate & Advanced SEO | | navidash0 -
Block in robots.txt instead of using canonical?
When I use a canonical tag for pages that are variations of the same page, it basically means that I don't want Google to index this page. But at the same time, spiders will go ahead and crawl the page. Isn't this a waste of my crawl budget? Wouldn't it be better to just disallow the page in robots.txt and let Google focus on crawling the pages that I do want indexed? In other words, why should I ever use rel=canonical as opposed to simply disallowing in robots.txt?
Intermediate & Advanced SEO | | YairSpolter0 -
Robots.txt vs noindex
I recently started working on a site that has thousands of member pages that are currently robots.txt'd out. Most pages of the site have 1 to 6 links to these member pages, accumulating into what I regard as something of link juice cul-d-sac. The pages themselves have little to no unique content or other relevant search play and for other reasons still want them kept out of search. Wouldn't it be better to "noindex, follow" these pages and remove the robots.txt block from this url type? At least that way Google could crawl these pages and pass the link juice on to still other pages vs flushing it into a black hole. BTW, the site is currently dealing with a hit from Panda 4.0 last month. Thanks! Best... Darcy
Intermediate & Advanced SEO | | 945010 -
Is This 301 Use Best Practice??
I know its effective practice cuz we're getting our arse kicked. I'm curious if its best practice (white, gray or black hat). I'm checking a competitors link profile on its landing page that is hitting the top of page 1 for several keywords. This competitor (national chain) has a strong domain authority (69). The particular landing page I'm checking in OSE has two 301 redirects from its own site among some other directory links to the page. The page shows 15 external links and half of them are very strong including it's own 301's. Aren't they essentially sending their own juice to the landing page to bolster page/domain authority to rank higher in the SERPS for those keywords? Is this a common practice using the 301's to a landing page? Is it white, gray or black hat? They are appearing suddenly appearing on the first page for several category keywords, so we're doing some snooping. Thanks.
Intermediate & Advanced SEO | | AWCthreads0