Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Is linking to search results bad for SEO?
-
If we have pages on our site that link to search results is that a bad thing? Should we set the links to "nofollow"?
-
No it will no be a problem. Even if the search results show duplicate content, to show a sample of pages is fine.
-
Those pages due to the nature of searching for them are the exact same same pages as your normal content just that the url has changed . The best way in my opinion to address this is by using the canonical element you can read more here . if your normal pages have the element embedded in the code any instance of duplicate content should not be an issue.
Hope this helps
-
So you have links to search results on your page and those links are followed and indexed by google. LInking to a search result is not a bad thing per se. Just make sure, that the indexed pages are not to similar (i.e. duplicate content).
An example:
if the indexed page for "widget" (domain.tld/search/widget) and the indexed page for "widgets" (domain.tld/search/widgets) are similar or nearly similar they might be considered duplicate content, resulting in none of them ranking. One way to test this is to use d/c checker or the seomoz campaign. You want at least 20% difference in any combination of pages.
If you manually set these links you might want to check whether the results are different enough. If you don't want these pages to be indexed at all just "follow, noindex" them.
Besides that it's not a problem.
-
Search results on our own domain
-
Hi michelleh,
to what pages specifically are you linking? Google results like http://www.google.com/search?q=query ? Pages on other domains? could you please give some more information?
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Is using a subheading to introduce a section before the main heading bad for SEO?
I have noticed a popular trend in web design which involves sections of content being started with what looks to be smaller sub heading something like <h3>, <h4> or <h5> and then followed by a bigger heading <h2>. My question is, what is the best way to deal with this visual structure and will having a structure like this hurt your SEO? <h5>Contact Us</h5> <h2>Get started with your next project in minutes!<h2> <p>Some text here ...</p> Here are some examples where the header structure is similar to above (smaller before bigger): https://www.snappr.com/ https://form.taxi/ https://fluz.app/ If that structure is bad for SEO, then it seems like a simple solution is to make it purely visual, mimicking a sub header with styling on a span or paragraph like these sites do: https://www.andrejilderda.nl/ https://nightwatch.io/ https://www.swingvy.com/ https://www.figma.com/ My only concern with that approach is because your section sub heading is no longer an actual header you will miss out on ranking important and relevant keyword information for that section. Is this correct something to be worried about? There is one last solution I stumbled upon that involves using headings for both but in reverse hierarchy so a <h3> is first but styled to be smaller, followed by a visually bigger <h4> which provides the addition context. https://avocode.com/ Anyone have thoughts, expertise or resources on the matter?
Intermediate & Advanced SEO | | si.analytics0 -
Does Disavowing Links Negate Anchor Text, or Just Negates Link Juice
I'm not so sure that disavowing links also discounts the anchor texts from those links. Because nofollow links absolutely still pass anchor text values. And disavowing links is supposed to be akin to nofollowing the links. I wonder because there's a potential client I'm working on an RFP for and they have tons of spammy directory links all using keyword rich anchor texts and they lost 98% of their traffic in Pengiun 1.0 and haven't recovered. I want to know what I'm getting into. And if I just disavow those links, I'm thinking that it won't help the anchor text ratio issues. Can anyone confirm?
Intermediate & Advanced SEO | | MiguelSalcido0 -
SEO Impact of High Volume Vertical and Horizontal Internal Linking
Hello Everyone - I maintain a site with over a million distinct pages of content. Each piece of content can be thought of like a node in graph database or an entity. While there is a bit of natural hierarchy, every single entity can be related to one or more other entities. The conceptual structure of the entities like so: Agency - A top level business unit ( ~100 pages/urls) Office - A lower level business unit, part of an Agency ( ~5,000 pages/urls) Person - Someone who works in one or more Offices ( ~80,000 pages/urls) Project - A thing one or more People is managing ( ~750,000 pages/urls) Vendor - A company that is working on one or more Projects ( ~250,000 pages/urls) Category - A descriptive entity, defining one or more Projects ( ~1,000 pages/urls) Each of these six entities has a unique (url) and content. For each page/url, there are internal links to each of the related entity pages. For example, if a user is looking at a Project page/url, there will be an internal link to one or more Agencies, Offices, People, Vendors, and Categories. Also, a Project will have links to similar Projects. This same theory holds true for all other entities as well. People pages link to their related Agencies, Offices, Projects, Vendors, etc, etc. If you start to do the math, there are tons of internal links leading to pages with tons of internal links leading to pages with tons of internal links. While our users enjoy the ability to navigate this world according to these relationships, I am curious if we should force a more strict hierarchy for SEO purposes. Essentially, does it make sense to "nofollow" all of the horizontal internal links for a given entity page/url? For search engine indexing purposes, we have legit sitemaps that give a simple vertical hierarchy...but I am curious if all of this internal linking should be hidden via nofollow...? Thanks in advance!
Intermediate & Advanced SEO | | jhariani2 -
Redirect Search Results to Category Pages
I am planning redirect the search results to it's matching category page to avoid having two indexed pages of essentially the same content. Example http://www.example.com/search/?kw=sunglasses
Intermediate & Advanced SEO | | WizardOfMoz
wil be redirected to
http://www.example.com/category/sunglasses/ Is this a good idea? What are the possible negative effect if I go this route? Thanks.0 -
Link Juice + multiple links pointing to the same page
Scenario
Intermediate & Advanced SEO | | Mark_Ch
The website has a menu consisting of 4 links Home | Shoes | About Us | Contact Us Additionally within the body content we write about various shoe types. We create a link with the anchor text "Shoes" pointing to www.mydomain.co.uk/shoes In this simple example, we have 2 instances of the same link pointing to the same url location.
We have 4 unique links.
In total we have 5 on page links. Question
How many links would Google count as part of the link juice model?
How would the link juice be weighted in terms of percentages?
If changing the anchor text in the body content to say "fashion shoes" have a different impact? Any other advise or best practice would be appreciated. Thanks Mark0 -
How to detect a bad neighborhood links?
I have the feeling that I am suffering from negative seo, so there is a way to get a list of links that should remove in the google disavow links tool ?
Intermediate & Advanced SEO | | Valarlf0 -
Best way to block a search engine from crawling a link?
If we have one page on our site that is is only linked to by one other page, what is the best way to block crawler access to that page? I know we could set the link to "nofollow" and that would prevent the crawler from passing any authority, and we can set the page to "noindex" to prevent it from appearing in search results, but what is the best way to prevent the crawler from accessing that one link?
Intermediate & Advanced SEO | | nicole.healthline0 -
Link Age as SEO factor?
Hi Guys
Intermediate & Advanced SEO | | VividLime
I have a client who ranks well within a competitive sector of the travel industry. They are planning CMS move which will involve changing from .cfm to .aspx We will be doing the standard redirects etc However Matt's statement here on 301 redirects got me thinking
http://www.youtube.com/watch?v=zW5UL3lzBOA&t=0m24s He says that basically you loose a bit of page rank when you do a 301 redirect. Now, we will be potentially redirecting 1000s of links and my thinking is 'a lot of a little, adds up to a lot' In other words, 1000s of redirects may have a big enough impact to loose some rankings in a very competitive and aggressive space. So recommended that we contact the sites who has the link highest value and ask them to manually change the links from cfm to aspx. This will then mean that there are no loss value as with a 301 redirect. -But now I have another dilemma which I'm unsure about. So the main question:
Is link age factor in rankings ? If I update any links, this will make said link new to Google, so if link age is a factor, would this also lessen the value passed initially?0