What is the best method for indexing blog pages?
-
I have a client whose blog has hundreds if not thousands of entries. My question is does it help his site if each unique blog entry becomes indexed on Google? Can we do this dynamically? And role does the canonical tag play in blog entries if at all?
Thanks,
Chris
-
is there a way to insert the title tag dynamically on each blog post via the cms?
CMS is software, and every software package is different. I will share there should be a way to do it, but you would need to search your CMS provider's site to get the details.
For your titles, I am not clear what you are asking. I would recommend the title tag for your blog matching your blog's title. You may want to add your site name or category name depending on the situation. For example if your site is "Chevyworld.com" and you have a blog entry titled "1982 Stingray, the end of an era" then the post title could be:
1982 Stingray, the end of an era
1982 Stingray, the end of an era | Chevyworld
1982 Stingray, the end of an era | Corvettes | Chevyworld
In the first example, your CMS would be adjusted to use the blog title for the title tag. In the second, the blog title plus your site name would be used for the title tag. The last example uses the blog title, the blog's main category tag and the site title.
Will google treat each entry as a unique page?
You need to ensure each page can only be accessed by one URL. For example, take a look at the following blog article's URL:
http://www.huffingtonpost.com/2011/06/25/rachel-weisz-daniel-craig-married_n_884653.html
Now try to access that same article with various other URLs such as without the www or with adding a trailing slash character:
http://huffingtonpost.com/2011/06/25/rachel-weisz-daniel-craig-married_n_884653.html
http://www.huffingtonpost.com/2011/06/25/rachel-weisz-daniel-craig-married_n_884653.html/
Notice that when you try to remove the "www' the same article appears but the site's redirect works and adjusts the URL by adding the "www'. Does your blog article redirect itself in this manner? Or does it display for both the www and non-www url?
Another example is the trailing slash. In this case the URL is adjusted and a question mark is added. If you View Page Source you will see there is a canonical meta tag which ensures the correct version of the page is consistently used by search engines.
-
Thanks for the reply. Follow up to your response, I am familiar with sitemaps, but what I meant by dynamically was is there a way to insert the title tag dynamically on each blog post via the cms? Also should each title tag for the blog follow a formula like: Client Blog: article a Client Blog: article b Client Blog: article c (and so on) Also thanks for the info on the canonical tags. So for a typical blog, will google treat each entry as a unique page? I want to make sure we don't get dinged for having duplicate pages.
-
**My question is does it help his site if each unique blog entry becomes indexed on Google? **
I am assuming that since each entry is unique content and is offered on your site that you desire people to read it. If that is the case then yes, it would be a tremendous help to be indexed by Google. If the pages are not indexed by Google, then how will people find these pages? It will only happen if they are either already on your site, or are told about the pages or linked to the pages.
Having a page indexed by Google allows people to find it through a normal search, which is how most pages on the web are discovered.
Can we do this dynamically?
Yes. You can submit a simple sitemap to Google, and they will try to crawl your site's pages if they have not already done so. It is important you do not block their efforts in your robots.txt file nor with any "noindex" meta tags.
** And role does the canonical tag play in blog entries if at all?**
Canonical tags help ensure the correct version of your blog entry, or any web page is indexed. If your page can be accessed through multiple URLs, then it should be canonicalized so the proper version of the page is indexed.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Inner pages of a directory site wont index
I have a business directory site thats been around a long time but has always been split into two parts, a subdomain and the main domain. The subdomain has been used for listings for years but just recently Ive opened up the main domain and started adding listings there. The problem is that none of the listing pages seem to be betting indexed in Google. The main domain is indexed as is the category page and all its pages below that eg /category/travel but the actual business listing pages below that will not index. I can however get them to index if I request Google to crawl them in search console. A few other things: I have nothing blocked in the robots.txt file The site has a DA over 50 and a decent amount of backlinks There is a sitemap setup also any ideas?
Technical SEO | | linklander0 -
Why blocking a subfolder dropped indexed pages with 10%?
Hy Guys, maybe you can help me to understand better: on 17.04 I had 7600 pages indexed in google (WMT showing 6113). I have included in the robots.txt file, Disallow: /account/ - which contains the registration page, wishlist, etc. and other stuff since I'm not interested to rank with registration form. on 23.04 I had 6980 pages indexed in google (WMT showing 5985). I understand that this way I'm telling google I don't want that section indexed, by way so manny pages?, Because of the faceted navigation? Cheers
Technical SEO | | catalinmoraru0 -
Best way to fix a whole bunch of 500 server errors that Google has indexed?
I got a notification from Google Webmaster tools saying that they've found a whole bunch of server errors. It looks like it is because an earlier version of the site I'm doing some work for had those URLs, but the new site does not. In any case, there are now thousands of these pages in their index that error out. If I wanted to simply remove them all from the index, which is my best option: Disallow all 1,000 or so pages in the robots.txt ? Put the meta noindex in the headers of each of those pages ? Rel canonical to a relevant page ? Redirect to a relevant page ? Wait for Google to just figure it out and remove them naturally ? Submit each URL to the GWT removal tool ? Something else ? Thanks a lot for the help...
Technical SEO | | jim_shook0 -
Container Page/Content Page Duplicate Content
My client has a container page on their website, they are using SiteFinity, so it is called a "group page", in which individual pages appear and can be scrolled through. When link are followed, they first lead to the group page URL, in which the first content page is shown. However, when navigating through the content pages, the URL changes. When navigating BACK to the first content page, the URL is that for the content page, but it appears to indexers as a duplicate of the group page, that is, the URL that appeared when first linking to the group page. The client updates this on the regular, so I need to find a solution that will allow them to add more pages, the new one always becoming the top page, without requiring extra coding. For instance, I had considered integrating REL=NEXT and REL=PREV, but they aren't going to keep that up to date.
Technical SEO | | SpokeHQ1 -
My blog page isn't ranking in Google
Hi, I noticed that my blog page on my site isn't in Google when i search for full URL link http://www.asggutter.com/blog/ instead i see page that isn't even working asggutter.com/sitemap.xml screen shot http://screencast.com/t/6OVFLwL8nTL How i can i fix that. Thanks
Technical SEO | | tonyklu0 -
De-indexing millions of pages - would this work?
Hi all, We run an e-commerce site with a catalogue of around 5 million products. Unfortunately, we have let Googlebot crawl and index tens of millions of search URLs, the majority of which are very thin of content or duplicates of other URLs. In short: we are in deep. Our bloated Google-index is hampering our real content to rank; Googlebot does not bother crawling our real content (product pages specifically) and hammers the life out of our servers. Since having Googlebot crawl and de-index tens of millions of old URLs would probably take years (?), my plan is this: 301 redirect all old SERP URLs to a new SERP URL. If new URL should not be indexed, add meta robots noindex tag on new URL. When it is evident that Google has indexed most "high quality" new URLs, robots.txt disallow crawling of old SERP URLs. Then directory style remove all old SERP URLs in GWT URL Removal Tool This would be an example of an old URL:
Technical SEO | | TalkInThePark
www.site.com/cgi-bin/weirdapplicationname.cgi?word=bmw&what=1.2&how=2 This would be an example of a new URL:
www.site.com/search?q=bmw&category=cars&color=blue I have to specific questions: Would Google both de-index the old URL and not index the new URL after 301 redirecting the old URL to the new URL (which is noindexed) as described in point 2 above? What risks are associated with removing tens of millions of URLs directory style in GWT URL Removal Tool? I have done this before but then I removed "only" some useless 50 000 "add to cart"-URLs.Google says themselves that you should not remove duplicate/thin content this way and that using this tool tools this way "may cause problems for your site". And yes, these tens of millions of SERP URLs is a result of a faceted navigation/search function let loose all to long.
And no, we cannot wait for Googlebot to crawl all these millions of URLs in order to discover the 301. By then we would be out of business. Best regards,
TalkInThePark0 -
Best method of redirecting http to https on homepage
Hi everyone, I'm looking to redirect all http requests to https for a site's homepage. It only needs to be for the homepage, not site wide. What's the best method of doing this without losing pagerank or ranking? I'm using IIS7.5 so I've been looking at a URL Rewrite or possibly this ASP.Net solution; http://www.xdevsoftware.com/blog/post/Redirect-from-Http-to-Https-in-ASPNET.aspx Or is a simple 301 or 302 (for some reason Microsoft's site says to do a 302 re-direct, though I'm not sure if this is great from an SEO perspective?) re-direct from http version to the https version the best method? Also if the solution retained the URL query string that would be even better! Any help appreciated! Thanks
Technical SEO | | PeterAlexLeigh0 -
Should I allow index of category / tag pages on Wordpress?
Quite simply, is it best to allow index of category / tag pages on a Wordpress blog or no index them? My thought is Google will / might see it as duplicate content? Thanks, K
Technical SEO | | SEOKeith0