Google image crawler won’t respect my robots.txt entry to not crawl images

I was looking for a way to prevent reverse image searching (namely I didn’t want for people who had a copy of one of my images to upload it to google and discover where it originated from). I created the following robots.txt file at put it at the root of my blogspot blog:

User-agent: *
Disallow: /hide*.jpg$
Disallow: /hide*.jpeg$
Disallow: /hide*.png$

User-agent: Googlebot-Image
Disallow: /hide*.jpg$
Disallow: /hide*.jpeg$
Disallow: /hide*.png$

With it, I was expecting that all jpg and png image files that start with the word hide (eg. hide1023939.jpg) would not appear in Google Images (or any other search engine). I was inspired by the official documentation here and here.

However Google Images keeps showing them, both when reverse searching as well as searching sitewise for any images. I’ve added many new images since I implemented the robots directives but even these new files get crawled.

As an observation the images on blogspot/blogger.com are hosted on http://1.bp.blogspot.com/....file.jpg instead of my own subdomain (http://domain.blogspot.com) and I wonder if this is the cause of the issue?
Any ideas how to solve this?

Search engine indexing: Google says the URL is blocked by robots.txt but it is not

My google search console says my submitted url is blocked by robots.txt, but i checked the header, tags and the robots.txt file. My URL was at # 1 for my target keyword and I get 1000 clicks daily before this error. After trying everything, I have to duplicate the page and do a 301 redirect, but the problem has not disappeared.
Here is the error I receive:
enter the image description here
But when I test the live URL it shows everything fine:
enter the image description here
here is my robots.txt

User-agent:*
Disallow:/search
Allow:/

Sitemap:https://www.chanakyadrishti.com/sitemap.xml
Sitemap:https://www.chanakyadrishti.com/feeds/pages/default?orderby=UPDATED

Google search console: how do I fix the Indexed problem, although blocked by the robots.txt file if the reported pages don't exist?

I found related information such as How to solve Google "Indexed, although blocked by robots.txt", but it doesn't answer my question.

The pages I saw in the report do not exist at all.

enter the image description here

For example, the above two links do not exist at all. I don't know why people add some number or a image?url= to my URL It is so weird.

How can I tell Google to ignore these URLs?

Do not allow directory only at root, and not elsewhere using robots.txt

You almost have it.

This is the correct syntax:

User-agent: *
Disallow: /page

Explanation:

the Disallow: directive is relative to the root.

So the first / in that directive it indicates the root folder of your website.

Accordingly, the directive Disallow: /page will reject:

/page/alpha/
/page/beta/
/page/gamma/

but it will not to refuse:

/aleph/page/
/bet/page/
/gimel/page/

because the last three paths don't start with /page/.

2013 – How to hide the robots.txt file for SharePoint internet sites

I think based on Search Engine Site Map Settings disabled or enabled feature we can show or hide the robots.txt archive. However, if we want to restrict certain types of files, we can add the entry to the file as shown below:

User agent: *

Do not allow: / _layouts /

Do not allow: / _vti_bin /

Do not allow: / _catalogs /

If you want to allow SharePoint 2010 or 2013 to crawl your website, please add the following to your robots.txt file.

User agent: MS Search 6.0 Robot

To refuse:

Source:

The correct robots.txt settings to allow SharePoint to crawl your site

Forum robots.txt file | Forum promotion

Hello FP

We are all there when there are so many guests on your forum that you start to question what's going on. Of course, most of them are bots and some are harmful, mainly just looking for email addresses to send spam.

I found this beast from a robots.txt file and thought about sharing it. It is not made by me, all credit goes to mitchellkrogza.

The robots.txt file can be found at https://github.com/mitchellkrogza/apache-ultimate-bad-bot-blocker/blob/master/robots.txt/robots.txt

When will search engines ignore the rules in my robots.txt to NOT crawl my website and files?

I'm working with a marketer who wants to set up a website with a text field for people to type in their URL from another secret but publicly accessible website. Let's say the public but secret access website is located at https://you-are-the-winner-of-this-amazing-scavenger-hunt-and-now-you-will-get-a-free-trip-to-mars-with-space-x-accompanied-by-elon-musk.com.

The seller is concerned that someone may discover this super secret url by searching for it through search engines or other popular online search tools.

I imagine if I configure https://you-are-the-winner-of-this-amazing-scavenger-hunt-and-now-you-will-get-a-free-trip-to-mars-with-space-x-accompanied-by-elon-musk.comI just have to make sure that robots.txt it has the appropriate rules to disable.

But my question is under what conditions will search engines ignore my disablement rules? Is there any other way for people to discover the URL? https://you-are-the-winner-of-this-amazing-scavenger-hunt-and-now-you-will-get-a-free-trip-to-mars-with-space-x-accompanied-by-elon-musk.com through internet searches?

web crawlers – Ahrefs reports "Retrieving robots.txt file took too long"

My robots:

User-agent: *
Disallow: /wp-admin
Disallow: /sistema
Disallow: /site
Disallow: /old

Sitemap: http://www.example.com.br/page-sitemap.xml

In google test tool: https://www.google.com/webmasters/tools/robots-testing-tool it is becoming normal: & # 39; allowed & # 39;

But in ahref I have the comments:

Getting the robots.txt file took too long

I have already asked the hosting staff to see us if there are any tracker locks or ips used by ahrefs and there isn't.

I have already disabled all plugins on the site and nothing

Google Search Console: How to edit or delete robots.txt on a WordPress powered website

If this is a new installation of WordPress, it could be because you have configured the privacy settings to prevent search engines from crawling the site.

(and there will be no physical robots.txt on the server as WordPress creates it on the fly)

Go to your WordPress settings and see if this box is checked:

enter the image description here

If so, disable it and then the robots.txt file should change to

User-agent: *
Disallow: 

If you're still having trouble and the robots.txt file is still set to block crawlers, explore the other option as described in Facet.

301 redirect to the robots.txt file to a different domain, it can be to a front layer of the cloud

I would like to know if I can put a 301 redirect in the robots.txt file and navigate to a different location.

sample site = https://www.example.net/robots.txt redirects to https://differentdomain.net/example/robots.txt