I have a page in my website, It takes two parameters in query string.
I don't want to show this page on google search. I put meta tag as below:
<meta name="robots" content="noindex">
But It still shows results as Cached, I tried google webmaster to remove the url's. It removes requested url but not stopping index the url and I am getting the url with some other parameter value. I don't want to show this page on search completely.
I looked at remove pages from google dynamic url - robots.txt
But I am not getting my answer.
Thanks
It's known issue. After you made all the recommended steps, i.e. added 'noindex' meta tag, disallowed it in robots.txt and removed from search in Webmaster Tools it may take a couple of weeks until the page completely disappears from search results. It's Google.
So, an only way to prevent the page from visiting is catch the request on server and redirect.
Related
I have a page at:
https://www.luckycheckout.com/goto/282/cs/1?ct=1
which contains the following line of code in the head section:
<meta name="robots" content="noindex, nofollow" />
I also have "Disallow: /goto" in my "robots.txt" file.
However, despite this, Google Search Console is complaining that the page is:
Indexed, though blocked by robots.txt
As far as I can tell everything is both valid and correct and should not be indexed, can anybody explain what google is complaining about?
Thanks
You need to remove the Disallow for this site on the robots.txt file:
When Googlebot next crawls that page and see the tag or header, Googlebot will drop that page entirely from Google Search results, regardless of whether other sites link to it.
Important: For the noindex directive to be effective, the page must not be blocked by a robots.txt file. If the page is blocked by a robots.txt file, the crawler will never see the noindex directive, and the page can still appear in search results, for example if other pages link to it.
source: https://support.google.com/webmasters/answer/93710
You can also remove a site you own from the Google index, using the Google Search Console. You can find more information on the Google Webmaster documentation.
My site has webpage urls that use the following format:
www.mysite.com/id/pretty_title
The front page links to these pages, but the href actually contains some parameters:
www.mysite.com/id/?some_ugly_parameters_to_let_me_know_what_search_it_is_from
This then redirects to
www.mysite.com/id/pretty_title
which shows the page.
My issue is that Google's search results show the link to the page as the ugly one instead of the pretty redirected one. Besides looking ugly, it can cause errors because the parameters that are included in the ugly link are irrelevant to a user who directly enters a page from Google.
Can I get Google to only use the final redirect url? Or otherwise avoid this issue? I do need to pass these parameters along. Storing info in the session won't work because a user can have several tabs open.
On the "ugly URL page" you put <link rel="canonical" href="www.mysite.com/id/pretty_title"> which tells the search engine your preferred URL for that content.
Can also set www.mysite.com/id/?some_ugly_parameters_to_let_me_know_what_search_it_is_from to have a 301 redirect to your pretty URL
Canonical URLs: https://support.google.com/webmasters/answer/139066
301 Redirects: https://support.google.com/webmasters/answer/93633?hl=en
The company I work for has replaced our previously very flexible website with a much more restrictive "website in a box" technology. I have my web pages hosted on Google Sites and would like to redirect people to those pages. When I attempt to do this via javascript it gets stripped from the page when its saved. I do not have access to the section to attempt the depreciated method of redirecting.
Is there another method available to automatically redirect a customer other than just posting a link in a restricted environment like this?
If you're limited to using HTML to do the redirect, you can use a meta redirect:
<meta http-equiv="refresh" content="0; url=http://example.com/">
Though note that its use is deprecated because it may be disorienting to the user. In addition to the <meta> tag, you can add <link rel="canonical" href="http://example.com/"> to let search engines know that the targeted page is the canonical one.
Edit: if Google Sites won't allow you to change the <head> HTML, the Javascript, or the PHP, then it's time to go searching for solutions within Google Sites itself. One solution that pops up pretty frequently in searches seems to be using a URL Redirect Gadget.
On the page you want to redirect from, click the Edit Page button, then Insert Menu, then More Gadgets. Once there, search for "redirect gadgets" and some widgets that should help will show up.
These instructions are based on advice given in the Google Products forums. I don't have a Google Site myself, so I can't verify that they work.
I've been scouring the web and StackOverflow for an answer, but I've found no case that exactly applies to my situation. I'm using Facebook Linter to debug the way FB is scraping my meta tags. If I use it on a simple About page, it picks up everything fine, particularly the og:url meta tag.
See:
http://developers.facebook.com/tools/debug/og/object?q=http%3A%2F%2Felectionstats.com%2Fabout%2Fprivacy_policy
The trouble starts when I scrape my normal content pages. Although I've triple-checked that my tags are formed well, the FB Linter cuts the URI off the URL, so it reports that the og:url tag only has the domain name, electionstats.com/!
See:
http://developers.facebook.com/tools/debug/og/object?q=http%3A%2F%2Felectionstats.com%2Fsearch%2Fyear_from%3A2010%2Fyear_to%3A2010%2Foffice_id%3A6
The og:url tag that is actually on the page looks like this:
I am skeptical that it is an issue with FB caching the pages, because on my About pages I have made quick code changes that change the meta tag output, then re-run the same page through the Linter, and the Linter shows these quick changes, without fail, every time. But for some reason, when I try dozens of different URL combinations on the main content pages (the /search/ pages), I always get a cut-off URL and consequently only meta fields from my homepage.
I had even theorized that FB will ignore a URL that looks like a "search" page, so I re-routed the URL and the title tag to use the nomenclature "explore" instead of "search", but this still did nothing -- the URI would still get chopped off.
Oy, this is embarrassing.
I have code at the beginning of each page request that detects if the user's browser accepts cookies; if not, it kicks the user back to the homepage. The Facebook web crawler, like other web crawlers, does not use cookies. Thus, it kept ending up back on the homepage and reading the homepage's og/meta tags. The greater unintended consequence of my code was that it kicks out ALL web crawlers trying to get a sense of my website, including Google's.
The fix: skip the cookie-handling check if the user agent string matches part the UA provided by common web crawlers, e.g http://www.cult-f.net/detect-crawlers-with-php/
I've to correct the google search title and summary for a website having the following code as home page:
<?php
header("Location:/mil/index.php");
?>
It forwards the user to another page. I know this is not good, but I was wondering how to quickly fix it.
If I add etc... to this page, is enough ? Is google grabbing the information from this page ?
Or is it grabbing from the website pages and bypassing this page ?
thanks
At the moment you are performing a 302 redirect, which tells Google that content found at the first page has temporarily moved to the second. Because of this, Google will not update it's index, and will continue to treat the first page as the important one. If it's empty, then that is no good for your search rankings. Do this instead:
header ('HTTP/1.1 301 Moved Permanently');
header ("Location:/mil/index.php");
A 301 redirect tells Google that the content has permanently moved to the new location, and they will update their index appropriately.
If this is a permanant redirect, and you have access to the .htaccess file, then a faster, cleaner way of doing this would be to let Apache handle it. If your first script contains nothing but that redirect, then delete it completely and add this line to your .htaccess file:
Redirect 301 /the-first-script.php /mil/index.php