Where do these URL's come from?! (Indexation issues)
-
We have an international webshop with languages in the URLs. Our URLs are now set up as follows:
http://thermalunderwear.eu/eng/category/product
Now, we know that there's some kind of strange redirect problem causing problems with our indexation, this is a technical issue that should be fixed soon. But whether this is the cause of some other strange problems, I do not know. I'd be happy with any help/advice/tips.
1. The SEOmoz site crawler starts at http://thermalunderwear.eu. This currently does not yet redirect to http://thermalunderwear.eu/eng like we want it to, but all the links on the page do include the default language code. So all links on the page are http://thermalunderwear.eu/eng/category etc. However, apart from those URLs, the site crawler finds many URLs in the form http://thermalunderwear.eu/category/product etc., so not including the language variable. Where it gets these I do not know, and since these URLs dont exist and the webshop simply shows the homepage, these URLs all have 50+ duplicate titles/content. Why oh why?
2. If I do a Google search for indexed URL's with English as language, I get many results formatted like this:
Coldpruf Enthusiast mens thermal shirt - Thermal wear for men ...
thermalunderwear.eu/eng/men/coldpruf-enthusiast-mens-thermal-shirt 170+ items – Fine-ribbed longsleeve thermal shirt men from Enthusiast ... {$SCRIPT_NAME} eng/men/coldpruf-enthusiast-mens-the {$ajax_url} http://thermalunderwear.eu/ajaxWhat are those variables doing there? It looks like it's taking something from our Smarty debug console, which is hidden but still active in the source code, but also the ajax URL which is in a completely different location. What is Google trying to show here?
-
It sees it as a list, its like rich snipits , its a huge amount of your content, and things it is the main content.
see these reullts. 40+ is a list i have in my page, it shows a few samples
-
I guess that is the only solution then. I don't quite understand why Google picks that information to show in the SERP text (as well as the 170+ items) but we'll try disabling the Smarty debugging when we're not actively using it. I hope it helps!
-
I looked in the souce code of this page
http://thermalunderwear.eu/eng/men/devold-alpine-knee-thermal-socks-electric-blue
And i found {$SCRIPT_NAME} eng/men/coldpruf-enthusiast-mens-the
Your dubug code is in the souce code. you need to get rid of it, disable it or something. I have not used smarty debug, so I cant help much.
-
Ah thanks Alan! It looks like there is a problem in the code that generates the breadcrumb URLs. We will get that fixed asap, whicih should lower the number of duplicate content warnings considerably.
-
Your first problem
Look at this page,
http://thermalunderwear.eu/eng/kids-thermal-underwear/coldpruf-enthusiast-kids-thermal-shirt
you will see a link to http://thermalunderwear.eu/kids-thermal-underwear/coldpruf-enthusiast-kids-thermal-shirt
I will look at your other porblem in a few minutes
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Unsolved Ooops. Our crawlers are unable to access that URL
hello
Moz Pro | | ssblawton2533
i have enter my site faroush.com but i got an error
Ooops. Our crawlers are unable to access that URL - please check to make sure it is correct
what is problem ?0 -
Page with "Missing Title Tag" isn't a page
Hello, I am going through the various errors that the Moz Pro Crawl report and some non-existent pages keep coming up in the report. For example, one error category is "Missing Title Tag" with one page identified. But this page http://www.immigroup.com/news/“http%3A/crs.yorku.ca”?page=2 isn't real. It would have been a 404 were there not a redirect for everything that is /news/gobbledygook to /news. So my question is: when moz (or GA for that matter) identifies these pages as "real" and having errors, do I need to take this seriously? And what do I do about it? Thanks! George
Moz Pro | | canadageorge0 -
OK I'll try again.... Linking root domains and external links
I'm new to the crawl test tool, I think I know what is meant by external links and linking root domains but I need to check. If there are 500 external links and 50 linking root domains does that suggest the essentially those 500 links are coming from 50 root domains? If yes, then what would 1000 linking root domains and 100 links mean? Or am I completely wrong in my assumptions and if so, can somebody point me in the right direction? Thank you.
Moz Pro | | JemRobinson0 -
How do I delete labels from the list of labels we've used?
How do we delete labels from the pulldown in moz, that we already created, but no longer use. The function "delete labels" only deletes labels off of certain keywords. But we cant figure out how to remove unused labels completely from the list of labels. Thanks, Thanks, Andrew
Moz Pro | | jampaper0 -
SEOmoz report vs. Google's Algo
Hello, I got an SEOmoz Report for one of our clients and the report is showing these pages and many more others as duplicate page content. Thing is the page content is not duplicate however there is very little data differentiating the contents. My question is does Google see the following pages contents as duplicate? because seomoz does. http://dallastxlofts.com/blog/2012/06/using-a-loft-for-commercial-or-office-space.html/img_9632/ http://dallastxlofts.com/blog/2012/08/newly-renovated-in-victory-park.html/3-2/ http://dallastxlofts.com/blog/2012/08/pedestrian-friendly-uptown-west-village.html/attachment/18/ http://dallastxlofts.com/blog/2012/06/using-a-loft-for-commercial-or-office-space.html/img_4322/ http://dallastxlofts.com/blog/2012/08/historic-deep-ellum-lofts.html/2012-08-18-11-24-30/ http://dallastxlofts.com/blog/2012/08/pedestrian-friendly-uptown-west-village.html/attachment/13/ http://dallastxlofts.com/blog/2012/06/using-a-loft-for-commercial-or-office-space.html/842-4/
Moz Pro | | Bryan_Loconto0 -
What's the best research tool for measuring blogger outreach success?
I'm looking for the best MOZ Pro research tool to track the success of our fledgling blogger outreach/link building efforts. Should we use Open Site Explorer's "Full List of Link Metrics", or is there something better or more granular for counting and analyzing inbound blog links? Thanks!
Moz Pro | | MJOshea0 -
Why can't we access metrics in the API that are available in the mozbar?
For example, the Chrome/Firefox mozbar uses peid (Root Domain External Links) but the api doesn't allow this. I've done a little bit of reverse engeniering and it looks like the accessID for the mozbar (when logged in as a pro member) is "pro-xxxx". Does this mean that each pro account has its own accessid/secret key? If so, when will pro members have access to that? I've created a tool that I use for deciding which expired domains to buy and it uses all of the free metrics - but I find myself having to do one extra step to get the peid (navigating to ose). It doesn't make much sense that the metric is available to me when I'm using the mozbar logged in as a pro member (which I pay for), but not apart of the API.
Moz Pro | | SeanStewart811