Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Mobile URL parameter (Redirection to desktop)
-
Hello,
We have a parallel mobile website and recently we implemented a link pointing to the desktop website. This redirect is happening via a javascript code and results in a url followed by this paramenter: ?m=off
Example:
http://www.m.website.com redirects to:
http://www.website.com/?m=offQuestions:
- Will the "http://www.website.com/?m=off" be considered duplicate content with "http://www.website.com" since they both return the same content?
- Is there any possibility that Google will take into consideration the url ending in "/?m=off"? How should we treat this new url?
- The webmaster tools URL parameter configuration at the moment isn't experiencing problems but should we submit the parameter anyway in order not to be indexed or should we wait first and see the error response?
- In case we should submit this for removal... what's the best way to do it? Like this?
- Parameter: ?m=off
- Does this parameter change page content seen by the user? - doesn't affect page content
Any help is much appreciated.
Thank you! -
Hi again!
It's normal that you feature very similar or even, duplicate content through both your mobile and desktop Website versions, what it is fundamental though is that you use the rel alternate and canonical tags that I shared with you in my previous answer. I've just checked your site and you're not using them.
If you correctly add these tags in your mobile and desktop URLs as I described in my previous answer you shouldn't have any content duplication or cannibalization issues in Google's search results.
Thanks!
Aleyda
-
Thanks Aleyda, mobile website development is something new for and since we developed this in house, we are not sure if we have done the right thing, if we have duplicate content. Is there any way we can check for duplicate content? Could you take a look at our website and its mobile version and let us know what you think?
Thank you!
-
Hi there!
It's ok (and needed) in these circumstances to use a parameter in the URL since you cannot simply add a direct link otherwise your mobile users who select to switch to the desktop version will end-up in a redirection loop. What you can do to avoid any type of content duplication issues in this type of settings is:
- Add a link rel="alternate" tag in each of the desktop URLs referring to their specific mobile URL versions and a link rel="canonical" tag in all of the mobile URLs referring to their specific desktop URLS versions, as specified here by Google. Take a look at an example here
- Configure the appropriate canonical tags in your Desktop URLs as specified here by Google, so they always refer to their original URL versions (without parameters). This means that in http://www.website.com/?m=off you should add this canonical tag:```
pointing to its original URL version... and do this with every URL (referring to its own original URL version).
By doing this you won't have any content duplication issues. If you want to learn more about Mobile SEO take a look at the Webinar I did sometime ago in SEOmoz here and this post where I answer the most common Mobile SEO questions.
Thanks,
Aleyda
-
The page ending in ?m=off is non existent. We are using Joomla and such page does not exist and cannot be configured via canonical tags. Any other suggestion?
Thank you!
-
Query parameters can become indexed by Google and can lead to duplicate content. Just make sure that you are using the canonical tag in your header pointing to the non-query parameter page, that way Google will know to index the proper URL.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
I have a question about the impact of a root domain redirect on site-wide redirects and slugs.
I have a question about the impact (if any) of site-wide redirects for DNS/hosting change purposes. I am preparing to redirect the domain for a site I manage from https://siteImanage.com to https://www.siteImanage.com. Traffic to the site currently redirects in reverse, from https://www.siteImanage.com to https://siteImanage.com. Based on my research, I understand that making this change should not affect the site’s excellent SEO as long as my canonical tags are updated and a 301 redirect is in place. But I wanted to make sure there wasn’t a potential consequence of this switch I’m not considering. Because this redirect lives at the root of all the site’s slugs and existing redirects, will it technically produce a redirect chain or a redirect loop? If it does, is that problematic? Thanks for your input!
Technical SEO | | mollykathariner_ms0 -
URL too long. Shorten and redirect, or leave alone?
MOZ is indicating that i have several URLs that are too long. Should I shorten the URLs and redirect the long URLs to the new, shorter, URL? Or should i leave them alone, as I've been reading to avoid redirects.
Technical SEO | | Hanover4401 -
XML Sitemap and unwanted URL parameters
We currently don't have an XML sitemap for our site. I generated one using Screaming Frog and it looks ok, but it also contains my tracking url parameters (ref=), which I don't want Google to use, as specified in GWT. Cleaning it will require time and effort which I currently don't have. I also think that having one could help us on Bing. So my question is: Is it better to submit a "so-so" sitemap than having none at all, or the risks are just too high? Could you explain what could go wrong? Thanks !
Technical SEO | | jfmonfette0 -
Removing Redirected URLs from XML Sitemap
If I'm updating a URL and 301 redirecting the old URL to the new URL, Google recommends I remove the old URL from our XML sitemap and add the new URL. That makes sense. However, can anyone speak to how Google transfers the ranking value (link value) from the old URL to the new URL? My suspicion is this happens outside the sitemap. If Google already has the old URL indexed, the next time it crawls that URL, Googlebot discovers the 301 redirect and that starts the process of URL value transfer. I guess my question revolves around whether removing the old URL (or the timing of the removal) from the sitemap can impact Googlebot's transfer of the old URL value to the new URL.
Technical SEO | | RyanOD0 -
Does it really matter to maintain 301 redirect after de-indexing of old URLs?
Today, I was reading latest blog post on SEOmoz blog about. Uncrawled 301s - A Quick Fix for When Relaunches Go Too Well This is very interesting study about 301 & How it useful to maintain traffic. I'm working on eCommerce website and I have done similar stuff on my website. I have big confusion to manage 301 redirect. My website generates new URLs due to following actions. Re-write dynamic URLs. Re-launch entire website on different eCommerce platform. [osCommerce to Magento Commerce] Re-name category. Trasfer one product from one category to another category. I'm managing my 301 redirect with old practice. Excel sheet data from Google webmaster tools and set specific new URLs for redirect. Hoooo... Now, I have 8.5K redirect in htaccess... And, I'm thinking it's too much. Can we remove old 301 redirect from htaccess or not? This is big question for me. Because, all pages are not hyperlink on external website. Google have just de-indexed old URLs and indexed new URLs. So, Is it require to maintain 301 redirect after Google process?
Technical SEO | | CommercePundit0 -
What tools produce a complete list of all URLs for 301 redirects?
I am project managing the rebuild of a major corporate website and need to set up 301 redirects from the old pages to the new ones. The problem is that the old site sits on multiple CMS platforms so there is no way I can get a list of pages from the old CMS. Is there a good tool out there that will crawl through all the sites and produce a nice spreadsheet with all the URLs on it? Somebody mentioned Xenu but I have never used it. Any recommendations? Thanks -Adrian
Technical SEO | | Adrian_Kingwell0 -
Old URL redirect to New URL
Alright I did something dumb a year a go and I'm still paying for it. I changed my hyphenated URL to the non-hyphenated version when I redesigned my website. I say it was dumb because I lost most of my link juice even though I did 301 redirects (via the htaccess file) for almost all of the pages I could find in Google's index. Here's my problem. My new site took a huge hit in traffic (down 60%) when I made the change and even though I've done thousands of redirects my old site is still showing up in the SERPS and send much if not most of my traffic. I don't want to take the old site down in fear it will kill all of my traffic. What should I do? Is there a better method I should explore then 301 redirects? Could the other site be affecting my current rank since it's still there? (FYI...both sites are built on the WP platform). Any help or ideas are greatly appreciated. Thank you! Joe
Technical SEO | | kaje0 -
Blocking URL's with specific parameters from Googlebot
Hi, I've discovered that Googlebot's are voting on products listed on our website and as a result are creating negative ratings by placing votes from 1 to 5 for every product. The voting function is handled using Javascript, as shown below, and the script prevents multiple votes so most products end up with a vote of 1, which translates to "poor". How do I go about using robots.txt to block a URL with specific parameters only? I'm worried that I might end up blocking the whole product listing, which would result in de-listing from Google and the loss of many highly ranked pages. DON'T want to block: http://www.mysite.com/product.php?productid=1234 WANT to block: http://www.mysite.com/product.php?mode=vote&productid=1234&vote=2 Javacript button code: onclick="javascript: document.voteform.submit();" Thanks in advance for any advice given. Regards,
Technical SEO | | aethereal
Asim0