Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Has any one seen negative SEO effects from using Google Translate API
-
We have a site currently in development that is using the Google Translate API and I am having a massive issue getting screaming frog to crawl and all of our non-native English speaking employees have read through the translated copy in their native language and the general consensus is it reads at a 5th grade level at best. My questions to the community is, has anyone implemented this API on a site and has it a) helped with gaining traffic from other languages/countires and b) has it hurt there site from an SEO standpoint.
-
Hi Bernadette, I completely agree with that translation being human. Your are correct it wasn't google translate messing with the crawl, but it was a great argument to get removed
Where in screaming frog are you able to crawl slower? I have dug around the program and can't find the option.
-
Overall, if you are going to translate your website, it really should be translated by a human rather than an API. There are certain ways things should be translated and written that an API just cannot do properly. It's more of a user experience and readability issue than anything else.
It sounds as if the screaming frog crawling issue isn't related to the translation issue--it is a crawling issue. You may want to see if you can crawl much slower, which is a setting in screaming frog.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Advise on the right way to block country specific users but not block Googlebot - and not be seen to be cloaking. Help please!
Hi, I am working on the SEO of an online gaming platform - a platform that can only be accessed by people in certain countries, where the games and content are legally allowed.
International SEO | | MarkCanning
Example: The games are not allowed in the USA, but they are allowed in Canada. Present Situation:
Presently when a user from the USA visits the site they get directed to a restricted location page with the following message: RESTRICTED LOCATION
Due to licensing restrictions, we can't currently offer our services in your location. We're working hard to expand our reach, so stay tuned for updates! Because USA visitors are blocked Google which primarily (but not always) crawls from the USA is also blocked, so the company webpages are not being crawled and indexed. Objective / What we want to achieve: The website will have multiple region and language locations. Some of these will exist as standalone websites and others will exist as folders on the domain. Examples below:
domain.com/en-ca [English Canada]
domain.com/fr-ca [french Canada]
domain.com/es-mx [spanish mexico]
domain.com/pt-br [portugese brazil]
domain.co.in/hi [hindi India] If a user from USA or another restricted location tries to access our site they should not have access but should get a restricted access message.
However we still want google to be able to access, crawl and index our pages. Can i suggest how do we do this without getting done for cloaking etc? Would this approach be ok? (please see below) We continue to work as the present situation is presently doing, showing visitors from the USA a restricted message.
However rather than redirecting these visitors to a restricted location page, we just black out the page and show them a floating message as if it were a model window.
While Googlebot would be allowed to visit and crawl the website. I have also read that it would be good to put paywall schema on each webpage to let Google know that we are not cloaking and its a restricted paid page. All public pages are accessible but only if the visitor is from a location that is not restricted Any feedback and direction that can be given would be greatly appreciated as i am new to this angle of SEO. Sincere thanks,0 -
International SEO - how likely is it autoredirect via IP Address will impact rankings?
Hello, We're looking to internationalise our site so that US visitors will see the US branded version while everyone else will see the global version (currently at .com). This question specifically is about location-based auto-redirects. The literature I've read (including Google) recommends against auto-redirection: "Avoid automatic redirection based on the user’s perceived language. These redirections could prevent users (and search engines) from viewing all the versions of your site." Insofar as I understand it the theory goes as follows. Google crawls mainly from the US Auto-redirecting by US IP to the US domain will also redirect the Googlebot crawlers Because of this the crawlers will only see the US site / domain and not original .com website Crawlers can't index what they can't see Drop in rankings for the original site However, one of my colleagues has pointed out to me a company which does use auto-redirects. If a user is in the UK and type in their website they will be redirected to the UK version of the site, US will be US etc. I have checked their rankings and they are still ranking highly for relevant terms. I have been asked why they have been able to do this without impacting their visibility. Any ideas? Given their success have the risks of auto-redirecting have been overstated? How can we ensure US visitors land on the correct internationalised domain without auto-redirects in place? Looking forward to your thoughts on this as well as your experiences. Thanks in advance!
International SEO | | SEOCT0 -
How to Localise per Region (Europe, America, APAC, EMEI) and not per country as best SEO practise?
Hi SEO expertises! I am currently working with a client that initially have an English website targeting UK users but want to expand their market into four new regions (Europe, America, APAC and EMEI) keeping English as a main language. I would like to request your help here as I told the client ISO location and hreflang it will be just possible per language and they must need to localise each English region with local keywords, however I would like to double check if it will be any way (Sitemap, Hreflang) we can tell Google we are targeting per region and not per country? Thanks a lot!
International SEO | | Atalig20 -
What are the best practices for translation of city/state names for international SEO? (ie. New York in English vs. Nueva York in Spanish)
I'm working on international SEO / translation of a global travel site. While we have a global keyword research and translation strategy in process for each market they serve, I've run into a unique question. Overall, we are translating (and localizing) content for each market but aren't sure what to do with location names. Each country/state has cities and locations that have their own dedicated pages. I see three options for these location names (when titling a page and writing content): keep them in English, translate the names in the market languages, or use a combination of the two. The challenge with altering the location names to the market languages is that they are truly not known by those names. Though there are some instances where it may make sense…for instance **New York **in Spanish would be "Nueva York" with **‘**Nueva' being the Spanish translation of ‘new’. There are other instances, where no translation exists. If you’ve had a similar experience I'd love to hear your approach/recommendation.
International SEO | | JonClark150 -
Are my translated pages damaging my ranking?
Hi there, I have a site in English but with duplicates in different languages. The first problem is that these translated versions of my site receive no ranking on google stars (while the english does) - why is this? The second problem is that SEOmoz counts the errors on my site and then duplicates this error count for all the translated versions of my site - meaning I have a huge amount of errors (too many on-page links). Add this to the fact that I use affilite ID´s to track different types of traffic to my site - so all page urls in english and other languages, with an affiliate id on the end of the url, count as an error. This means I have a huge amount of on page errors indicated by SEOmoz, plus no ranking for my translated pages - I think this is really harming my overall ranking and site trust. What are your opinions on this?
International SEO | | sparkit0 -
SEO Company in Asia
Hi All. I have a client looking to expand their industrial services to southeast Asia (Vietnam and Indonesia specifically right now). Does anyone know of an SEO/Online Marketing firm local to that region that may be able to help them network with businesses and industries there? I've gone through the SEOmoz member database and reached out to a couple people with agencies in that area but never heard back from them. I personally thought a local firm would be more beneficial to the client but I'm also open to suggestions on ways that we might be able to help them market their services online from the US. Thanks so much! Megan
International SEO | | ILM_Marketing0 -
Non US site pages indexed in US Google search
Hi, We are having a global site wide issue with non US site pages being indexed by Google and served up in US search results. Conversley, we have US en pages showing in the Japan Google search results. We currently us IP detect to direct users to the correct regional site but it isn't effective if the users are entering through an incorrect regional page. At the top of each or our pages we have a drop down menu to allow users to manually select their preferred region. Is it possible that Google Bot is crawling these links and indexing these other regional pages as US and not detecting it due to our URL structure? Below are examples of two of our URLs for reference - one from Canada, the other from the US /ca/en/prod4130078/2500058/catalog50008/ /us/en/prod4130078/2500058/catalog20038/ If that is, in fact, what is happening, would setting the links within the drop down to 'no follow' address the problem? Thank you. Angie
International SEO | | Corel0 -
Google Webmaster Tools - International SEO Geo-Targeting site with Worldwide rankings
I have a client who already has rankings in the US & internationally. The site is broken down like this: url.com (main site with USA & International Rankings) url.com/de url.com/de-english url.com/ng url.com/au url.com/ch url.com/ch-french url.com/etc Each folder has it's own sitmap & relative content for it's respective country. I am reading in google webmaster tools > site config > settings, the option under 'Learn More': "If you don't want your site associated with any location, select Unlisted." If I want to keep my client's international rankings the way it currently is on url.com, do NOT geo target to United States? So I select unlisted, right? Would I use geo targeting on the url.com/de, url.com/de-english, url.com/ng, url.com/au and so on?
International SEO | | Francisco_Meza0