Google Search Console is creating URLs that do not exist in my Sitemap and then complains that these pages have errors

You have a misconception of what a site map is.

The site map is used to audit the site crawl by the search engine bot. The site map and the crawling of a site are two different and independent things. Google will continue to track your site regardless of any sitemap. The site map will be used to audit / see if Google can properly track your site. For example, if pages are found on your site map and Google has not seen the page, Google can add the page to the search queue to include it.

The opposite is not true. If a page is not found on the site map, Google will not remove it from its index. Why? Because Google found it crawling the site.

What he seems to believe is that the site map is the total authority that Google uses to know which pages exist on a particular site. This is not the case. The tracking is. The site map only helps Google know Yes they can correctly track your site and, otherwise, what pages are missing from Google that should be added to the search queue.

Hope Google no longer attempts to access the pages because these pages are no longer on your site map is incorrect. Sitemaps are cached and only checked periodically. Why? Because it is an audit process.

You have a real problem that you must solve.

It is returning a 500 error for pages that are not found. This is bad. Your site should return a 404 error not found. Error 500 is a system error and Google will treat the condition as temporary. If your site returned a 404 error, Google will continue testing the page for several attempts over a period of time until you decide that the page no longer exists. If possible, you want to issue a 410 Deleted error for the pages you have deleted. If this is too much work or is not possible, 404 will equal the same over time.

You need to correct your 500 error.

Google Search Console, is creating URLs that do not exist in my Sitemap and then complains that these pages have errors

You have a misconception of what a site map is.

The site map is used to audit the site crawl by the search engine bot. The site map and the crawling of a site are two different and independent things. Google will continue to track your site regardless of any sitemap. The site map will be used to audit / see if Google can properly track your site. For example, if pages are found on your site map and Google has not seen the page, Google can add the page to the search queue to include it.

The opposite is not true. If a page is not found on the site map, Google will not remove it from its index. Why? Because Google found it crawling the site.

What he seems to believe is that the site map is the total authority that Google uses to know which pages exist on a particular site. This is not the case. The tracking is. The site map only helps Google know Yes they can correctly track your site and, otherwise, what pages are missing from Google that should be added to the search queue.

Hope Google no longer attempts to access the pages because these pages are no longer on your site map is incorrect. Sitemaps are cached and only checked periodically. Why? Because it is an audit process.

You have a real problem that you must solve.

It is returning a 500 error for pages that are not found. This is bad. Your site should return a 404 error not found. Error 500 is a system error and Google will treat the condition as temporary. If your site returned a 404 error, Google will continue testing the page for several attempts over a period of time until you decide that the page no longer exists. If possible, you want to issue a 410 Deleted error for the pages you have deleted. If this is too much work or is not possible, 404 will equal the same over time.

You need to correct your 500 error.

seo – Alexa-rank complains about meta-label of unsafe generator

The label itself is not insecure. It's just a tag in an HTML file.

However, what it does is provide direct / immediate information about the software you are using and, therefore, it can be used by hackers to select which of your hacking systems it will use to penetrate your system.

Without that information they would have to run probes and it could take a long time. Also with advanced web firewalls, I could detect those probes and block the IP address (some people say it's not a good idea to block IPs, my experience is that it works wonders).

Keep in mind that a label that only gives a name is quite safe (ie, content = "Wordpress"). If the label includes the name, the version, the creation date, etc., then it becomes really easy (ie, content = "Wordpress 1.2.3 May 13, 2019"). So in your example, it's perfectly safe. Especially if your generator is proprietary and uses it for one or two websites (unlike WordPress which is used for millions of websites).


What is a probe?

Each time you access a website, you can detect which CMS was used to generate it by looking at the HTML content.

Examples:

  • WordPress uses routes that include wp- as an introducer
  • Drupal uses ... / sites // records / ...

Each CMS will only do the same for all the websites that use them. However, the detection of the version is more difficult. You have subtle differences that will tell you but it makes the hacker's life. more interesting (to my point of view – but read Stronger).

So hiding the generatorThe tag does not make any difference if you only have the name of the CMS there.