Is there a way to prevent Google Alerts from picking up old press releases?
-
I have a client that wants a lot of old press releases (pdfs) added to their news page, but they don't want these to show up in Google Alerts. Is there a way for me to prevent this?
-
Thanks for the post Keri.
Yep, the OCR option would still make the image option for hiding "moo"
-
Harder, but certainly not impossible. I had Google Alerts come up on scanned PDF copies of newsletters from the 1980s and 1990s that were images.
The files recently moved and aren't showing up for the query, but I did see something else interesting. When I went to view one of the newsletters (https://docs.google.com/file/d/0B2S0WP3ixBdTVWg3RmFadF91ek0/edit?pli=1), it said "extracting text" for a few moments, then had a search box where I could search the document. On the fly, Google was doing OCR work and seemed decently accurate in the couple of tests I had done. There's a whole bunch of these newsletters at http://www.modelwarshipcombat.com/howto.shtml#hullbusters if you want to mess around with it at all.
-
Well that is how to exclude them from an alert that they setup, but I think they are talking about anyone who would setup an alert that might find the PDFs.
One other idea I had, that I think may help. If you setup the PDFs as images vs text then it would be harder for Google to "read" the PDFs and therefore not catalog them properly for the alert, but then this would have the same net effect of not having the PDFs in the index at all.
Danielle, my other question would be - why do they give a crap about Google Alerts specifically. There has been all kinds of issues with the service and if someone is really interested in finding out info on the company, there are other ways to monitor a website than Google Alerts. I used to use services that simply monitor a page (say the news release page) and lets me know when it is updated, this was often faster than Google Alerts and I would find stuff on a page before others who did only use Google Alerts. I think they are being kind of myopic about the whole approach and that blocking for Google Alerts may not help them as much as they think. Way more people simply search on Google vs using Alerts.
-
The easiest thing to do in this situation would be to add negative keywords or advanced operators to your google alert that prevent the new pages from triggering the alert. You can do this be adding advanced operators that exclude an exact match phrase, a file type, the clients domain or just a specific directory. If all the new pdf files will be in the same directory or share a common url structure you can exclude using the "inurl:-" operator.
-
That also presumes Google Alerts is anything near accurate. I've had it come up with things that have been on the web for years and for whatever reason, Google thinks they are new.
-
That was what I was thinking would have to be done... It's a little complicated on why they don't want them showing up in Alerts. They do want them showing up on the web, just not as an Alert. I'll let them know they can't have it both ways!
-
Robots.txt and exclude those files. Note that this takes them out of the web index in general so they will not show up in searches.
You need to ask your client why they are putting things on the web if they do not want them to be found. If they do not want them found, dont put them up on the web.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Where does Google get its meta descriptions from?
We have a new client and they don't have meta descriptions yet. However, Google has assigned descriptions for them now appearing on the SERPs. The problem is that Google added a phone number that's totally not the client's and goes to a different unrelated business. Our plan is to update the meta to reflect the correct information, however, we're just perplexed as to how Google came up with the incorrect phone number. Where does it get its information from? The page currently has all the correct phone number, hours, and content. I've read that Google sometimes also doesn't recognise our meta descriptions if it thinks they could serve up a better one. My next question is, what if Google insists on showing the incorrect phone number. Is there a way we can fix this? Thanks!
On-Page Optimization | | nhhernandez2 -
Meta Titles On Site Different From Google Index Page
This is very embarrassing but I hope someone can help. The Meta Titles on my site are not being shown correctly on the Google site index. For example, when I got directly to the page the appropriate page title is shown obviously. However when I go view that page on Google, the title in completely different. Page title: <a class="attribute-value">Web Design, SEO, PPC, Mobile Development In Philadelphia, Bucks| Infinity Digital Agency</a> Google Shows: Our Services - Infinity Digital Agency Here is the result page. I am currently running WordPress and Yoast. Any thoughts would be greatly appreciated. https://www.google.com/search?q=site%3Awww.infinitydigitalagency.com%2Fservices%2F&ie=utf-8&oe=utf-8
On-Page Optimization | | infinitydigitalagency0 -
How to rank in Google during domain name search?
We have received one requirement for this website. http://www.coldcasebeer.com/ We would like to rank in Google during domain name search. If we are going to search following search terms in Google then We are not able to see this domain on first page of Google search result. cold case beer coldcasebeer We have done quick research on this issue. And, We have decided to implement following tasks to make it happen.But, We are quite excited to read certain inputs on this question from experts! 1. Upload default Robots.txt file 2. Verify website on Google webmaster tools 3. Set up USA region on Google webmaster tools 4. Submit Google crawl request on Google webmaster tools 5. Title tag and Meta description optimization for all pages (There are 5 pages only on website) 6. Audit Google local listing 7. Develop quality links with domain name
On-Page Optimization | | CommercePundit0 -
Transferring old articles to new site even if they are written horribly
OK my question for today is... If you currently have 300 articles on your current site but are building a new site would you transfer all of the articles over to the new site or focus on quality and rewrite the articles that had traffic? For example we have about 300 articles currently on our website 60 of which actually get traffic, We rewrote those articles to make sure they were written well. Someone thought that it would be best to simply transfer the other 240 articles over and rewrite them at another time to avoid 404 redirects. I would like your feedback on how you would approach this. Please be as detailed as possible explaining your thought process. Thanks!
On-Page Optimization | | PrintPlace.com0 -
90 days for Google
Hi, I'm new to Moz so still getting a feel of the forums. If my question has been answered then please point me in the right direction. I have noticed with many SEO companies they advertise that they can get you on google front page in 90 days. I'm not really interested in their techniques but more of why google takes 90+ to even appear. I have been working on my site for over a month, adding content, building good links, social media, blogs etc... but have not even come close to appearing in the top 50 pages for google. Is this normal? Is it just a matter of time before it starts to appear? Also, I have checked my backlinks and there is about 8 links that are coming from random pages in the US and some from China and india which i have no idea of. I tried to visit on of the sites but it had malware. I added all these back links to google disavow so hopefully that will fix it. Could that be the reason google would not even list my site? Thanks... Rick
On-Page Optimization | | pureozone0 -
Prevent Indexing of URLs Based on Tags
I started my website as a blog over at Posterous, but decided to turn it into a full scale business website with a self-hosted WordPress theme. Shortly after transitioning from Posterous to WordPress, I noticed that Google was indexing not only my old blog posts, but the URLs of my blog posts based on the tags they have. Is there any reason why this is a problem? I'm sure it shouldn't qualify as duplicate content, but for some reason it just feels a bit sloppy to me to have all of these pages indexed...Is this a non-issue? Should I just be more discriminating with my use of 'tags' if it bothers me? JiGLH.png
On-Page Optimization | | williammarlow0 -
Does Google give any (negative) weight to .info domains?
Hi, All! We're working with a client on setting up a support blog for their site (not hosted on-site because it's a related topic, not directly about their business). We're thinking about a few domain possibilities, but for some of the ones we'd like the .com is already taken. .info (as well as others) are available, and the client is interested in using a different TLD, but I'm shying away from it because of the concern that: a) people will look at it more suspiciously (exactmatch.info? must be spam) - but maybe that's just because I'm an SEO b) does Google have anything against exact match domains with endings like .info, .net, etc.? (I know there's never any guarantee that the exact match domain will continue to hold its weight in the algorithm at all, but taking that as a given for now - and we are planning on putting decent quality original content on it). Thanks in advance for the input! Aviva
On-Page Optimization | | debi_zyx0 -
What to do with old content in light of the Panda update?
Let's say you operate a laptop review website. After several years, the individual product review URL's (like site.com/dell/xp1234-review/) aren't receiving much traffic, they may have a few links here and there. In general and considering the panda update, would the best option be to 301 the old URL's back to the category page (site.com/dell/)or just keep them where they are? Any potential issues like having excessive 301's which could slow down the site or appear fishy to search engines?
On-Page Optimization | | BryanPhelps-BigLeapWeb0