How get rid of duplicate content, titles, etc on php cartweaver site?
-
my website http://www.bartramgallery.com was created using php and cartweaver 2.0 about five years ago by a web developer. I was really happy with the results of the design was inspired to get into web development and have been studying ever since. My biggest problem at this time is that I am not knowledgable with php and the cartweaver product but am learning as I read more. The issue is that seomoz tools are reporting tons of duplicate content and duplicate title pages etc. This is likely from the dynamic urls and same pages with secondary results etc. I just made a new sitemap with auditmypc I think it was called in an attempt to get rid of all the duplicate page titles but is that going to solve anything or do I need to find another way to configure the site? There are many pages with the same content competing for page rank and it is a bit frustrating to say the least. If anyone has any advice it would be greatly appreciated even pointing me in the right direction.
Thank you,
Jesse
-
I am still researching a bunch of sites trying to figure out a way to get the product ID name at the end which would be great as that is the page title. I just thought I would mention that I am working on it and see if you thought that it was not possible as you mentioned due to Cartweaver's limitations. It's funny that I have spent so much time trying to get my urls to show up how they should... seems this could have been configured into the original product. Beggars can't be choosers.
-
Yes I am going to take a look at that when I get home perhaps I have to change how a few things are referenced as well as create the change of address right? because if you type in the normal dynamic nasty url it still goes to the nasty url but if I select the url and paste it.. it brings up the page as I mentioned above. Basically stripped of images and styling.
I am wondering if it is possible to include that number at the end as it is the actual image and could potentially populate the title of the image at the end which would be sweet. Of course then I would have a new problem of too long of an url as I have the titles pretty keyword rich on a lot of them to make a proper title for the page.
If this all works out I have to create a link to your site at cartweaver and from a couple of my sites as you have been a great help and from what I can tell have been able to properly diagnose a fairly complex issue with php and cartweaver and even if some web page that I have not seen something similar enough you have been a great help. Thank you
-
I'm guessing the paths used to reference the images & css files are relative to the the results.php file.. now that there are "/"s the best thing to do is to change the template to either hard code an absolute path or use forward slash at the start to always start at the root.. eg
Old code:
New code
or
-
I tried the test example you did above and it was pretty cool. With the web address http://www.bartramgallery.com/photographer/charles-cramer/10.php it rendered a page with I believe everything except for design and styles as well as any imagery. Not sure what causes that to occur perhaps it is missing something but that was a pretty quick stab at fixing my url issue. I am too tired now and need to go to bed haha. Thanks
-
No worries
Look forward to seeing the site with the new URLs in place - a lot of great photos on that site that need to be shared with everyone
-
Yes it appears that this is a pretty good task to clean up this url issue but well worth it. I was surprised by the system moderators of Cartweaver discounting the url as if it were not important because they are very good developers however I think that the url is much more important than some realize as they are both keyword rich and more interesting to the customer. I am even less likely to click on some random url that has no meaning then if I saw one that clearly spelled out what the page was about.. Thanks Woj I am humbled and realize I have some studying to do.
-
There are 2 issues here:
-
Need to fix the URLs for better user experience & search engines and can do so by using rewrite rules in htaccess
The one suggested by the support forum (I've modified to better match your site but it's untested):
RewriteEngine on
RewriteRule ^photographer/([a-zA-Z0-9_-]+)/([0-9]+).php$ results.php?category=$2 The URLs would then be:
http://www.bartramgallery.com/photographer/charles-cramer/10.php (not ideal with "/10.php" at the end but may be best given the limitations of the cart)
rewrites to: http://www.bartramgallery.com/results.php?category=10 -
Clean up the Google index (remove old URLs & add new ones)
Since both URLs will render the same content we can fix by adding a
tag - attributing 1 source to the duplicate content - check if you can do this dynamically in the templates but be very careful not to canonical everything to the homepage or all your pages will be wiped out the index except the home page!)
-
-
when I read it it seems that the .htaccess was the way to go in that you can have the links appear to google as the old links but in presentation to the customer and keywords the new url would be used. The only thing I was confused about was that it seemed that it would not be good to do redirects but rewrites rather... or is it saying to do both?
-
Thanks
-
Great answer Woj!
-
My pleasure
If you set up redirects, you shouldn't loose any traffic
This can also be controlled via htaccess
In google, search for this "site:bartramgallery.com" (without the double quotes) & you will see all the pages you need to redirect
I see the Charles Cramer page as the first photographers page that comes up & the redirect would be something as simple as:
Redirect 301 /results.php?category=10 http://www.bartramgallery.com/charles-cramer
-
Thank you Woj for taking the time to look at my site and I like that organization method. I was not aware of the possibility of being able to reorganize my site like that. I will definately have to research and study a bit to be able to approach this and for awhile I will probably lose traffic but in the end after the changes it should be a much better foot going forward.
-
I'm not familiar with Cartweaver but these are just guides..
First define an organised URL structure - on bartramgallery.com, at a quick glance, a good one could be:
-
bartramgallery.com/photographer (e.g. bartramgallery.com/gordon-michael)
-
bartramgallery.com/photographer/photo (e.g. bartramgallery.com/gordon-michael/juniper-study-joshua-tree)
OR
bartramgallery.com/landscape-photography/photo (e.g. bartramgallery.com/landscape-photography/juniper-study-joshua-tree)
Keep in mind that the shorter URLs the better (could even have bartramgallery.com/photography/juniper-study-joshua-tree)
Second, rewrite the URLs using Rewrite Rules in the htaccess file (see this post: http://www.seomoz.org/blog/rewriterule-split-personality-explained)
I did a search on the Cartweaver support forums and found this:
http://forums.cartweaver.com/topic/google-analytics-identifying-products-and-categoriesOli, from the Cartweaver Support Team, seems to suggest the same "untested" approach as above
Let me know if you need any further help
-
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
International SEO and duplicate content: what should I do when hreflangs are not enough?
Hi, A follow up question from another one I had a couple of months ago: It has been almost 2 months now that my hreflangs are in place. Google recognises them well and GSC is cleaned (no hreflang errors). Though I've seen some positive changes, I'm quite far from sorting that duplicate content issue completely and some entire sub-folders remain hidden from the SERP.
Intermediate & Advanced SEO | | GhillC
I believe it happens for two reasons: 1. Fully mirrored content - as per the link to my previous question above, some parts of the site I'm working on are 100% similar. Quite a "gravity issue" here as there is nothing I can do to fix the site architecture nor to get bespoke content in place. 2. Sub-folders "authority". I'm guessing that Google prefers sub-folders over others due to their legacy traffic/history. Meaning that even with hreflangs in place, the older sub-folder would rank over the right one because Google believes it provides better results to its users. Two questions from these reasons:
1. Is the latter correct? Am I guessing correctly re "sub-folders" authority (if such thing exists) or am I simply wrong? 2. Can I solve this using canonical tags?
Instead of trying to fix and "promote" hidden sub-folders, I'm thinking to actually reinforce the results I'm getting from stronger sub-folders.
I.e: if a user based in belgium is Googling something relating to my site, the site.com/fr/ subfolder shows up instead of the site.com/be/fr/ sub-sub-folder.
Or if someone is based in Belgium using Dutch, he would get site.com/nl/ results instead of the site.com/be/nl/ sub-sub-folder. Therefore, I could canonicalise /be/fr/ to /fr/ and do something similar for that second one. I'd prefer traffic coming to the right part of the site for tracking and analytic reasons. However, instead of trying to move mountain by changing Google's behaviour (if ever I could do this?), I'm thinking to encourage the current flow (also because it's not completely wrong as it brings traffic to pages featuring the correct language no matter what). That second question is the main reason why I'm looking out for MoZ's community advice: am I going to damage the site badly by using canonical tags that way? Thank you so much!
G0 -
How to avoid duplicate content
Hi there, Our client has an ecommerce website, their products are also showing on an aggregator website (aka on a comparison website where multiple vendors are showing their products). On the aggregator website the same photos, titles and product descriptions are showing. Now with building their new website, how can we avoid such duplicate content? Or does Google even care in this case? I have read that we could show more product information on their ecommerce website and less details on the aggregator's website. But is there another or better solution? Many thanks in advance for any input!
Intermediate & Advanced SEO | | Gabriele_Layoutweb0 -
Internal Duplicate Content Question...
We are looking for an internal duplicate content checker that is capable of crawling a site that has over 300,000 pages. We have looked over Moz's duplicate content tool and it seems like it is somewhat limited in how deep it crawls. Are there any suggestions on the best "internal" duplicate content checker that crawls deep in a site?
Intermediate & Advanced SEO | | tdawson091 -
A lot of news / Duplicate Content - what to do?
Hi All, I have a blog with a lot of content (news and pr messages), I want to move my blog to new domain. What is your recommendation? 1. Keep it as is. old articles -> 301 -> same article different URL
Intermediate & Advanced SEO | | JohnPalmer
2. Remove all the duplicate content and create 301 from the old URL to my homepage.
3. Keep it as is, but add in the meta-tags NoIndex in duplicate articles. Thanks !0 -
PDF on financial site that duplicates ~50% of site content
I have a financial advisor client who has a downloadable PDF on his site that contains about 9 pages of good info. Problem is much of the content can also be found on individual pages of his site. Is it best to noindex/follow the pdf? It would be great to let the few pages of original content be crawlable, but I'm concerned about the duplicate content aspect. Thanks --
Intermediate & Advanced SEO | | 540SEO0 -
What's the best way to manage content that is shared on two sites and keep both sites in search results?
I manage two sites that share some content. Currently we do not use a cross-domain canonical URL and allow both sites to be fully indexed. For business reasons, we want both sites to appear in results and need both to accumulate PR and other SEO/Social metrics. How can I manage the threat of duplicate content and still make sure business needs are met?
Intermediate & Advanced SEO | | BostonWright0 -
Duplicate Content Error because of passed through variables
Hi everyone... When getting our weekly crawl of our site from SEOMoz, we are getting errors for duplicate content. We generate pages dynamically based on variables we carry through the URL's, like: http://www.example123.com/fun/life/1084.php
Intermediate & Advanced SEO | | CTSupp
http://www.example123.com/fun/life/1084.php?top=true ie, ?top=true is the variable being passed through. We are a large site (approx 7000 pages) so obviously we are getting many of these duplicate content errors in the SEOMoz report. Question: Are the search engines also penalizing for duplicate content based on variables being passed through? Thanks!0 -
Multiple cities/regions websites - duplicate content?
We're about to launch a second site for a different, neighbouring city in which we are going to setup a marketing campaign to target sales in that city (which will also have a separate office there as well). We are going to have it under the same company name, but different domain name and we're going to do our best to re-write the text content as much as possible. We want to avoid Google seeing this as a duplicate site in any way, but what about: the business name the toll free number (which we would like to have same on both sites) the graphics/image files (which we would like to have the same on both sites) site structure, coding styles, other "forensic" items anything I might not be thinking of... How are we best to proceed with this? What about cross-linking the sites?
Intermediate & Advanced SEO | | webdesignbarrie0