seo: on a sitemap, should I update the lastmod tag of a URL based on text content or html content?

Imagine that I have this blog / e-commerce website with 1000 publications / products. And I created a site map for him, which is dynamically generated. It's basically a list with a lot of and Tags

I'm pretty sure the trackers expect me to update the dates for any product or blog post that you edit and change the content of the text (or change the images). Add something new, update information, etc. Basically, anything that users GO differently when they enter my page. This makes sense.

But my question is:

I have a dynamic single page website. So I don't save static pages. I generate and render them (server side) at runtime. So what happens if I decide that all my blog posts should now be displayed within a

or

tag instead of a div? Or what happens if I add some structured metadata to add prices and review the properties of my products, or if I add structured data for bread crumbs.

See what I mean? The content that the user sees has not changed. But I have updated some labels that CRAWLER will interpret differently. The text / image content is the same, but the HTML content has changed. And this could even have an impact on my ranking, since I am adding new tags that could improve my SEO.

But now, what should I do? The changes I made now will show the 1000 posts / products in a different way with the new tags (in the tracker's perspective). Should I update the tag ALL my 1000 URLs on my sitemap? The user will continue to see the same text / image content and will not notice any difference.

If I update all 1000 tags, doesn't the crawler think it's "weird" that now all my URLs have been updated the same day? Since everyone will have the same Tags Does it make sense?

Please, any help is appreciated.
Thank you

Are the links on canvas / WebGL with a sitemap good enough for SEO, or do I also need normal href links?

The main page of my site consists of great graphics and animations implemented with Canvas / WebGL using the pixi.js library. The problem is that all links are also implemented as interaction with the WebGL layer.

But I know that I can list my links on sitemap.xml. Is it enough to list the links on the site map for a good SEO in this situation?

Or should I somehow verify something and insert normal a instead?

robots.txt: Google Search Console and Sitemap settings for a website subfolder

You may have a different XML sitemap configured for your folder and it will not confuse search engines. Source.

Just keep in mind that they are correctly linked from the main site map.

But when it comes to GSC or Analytics, it is not possible to implement it in folders.

Google Search Console is creating URLs that do not exist in my Sitemap and then complains that these pages have errors

You have a misconception of what a site map is.

The site map is used to audit the site crawl by the search engine bot. The site map and the crawling of a site are two different and independent things. Google will continue to track your site regardless of any sitemap. The site map will be used to audit / see if Google can properly track your site. For example, if pages are found on your site map and Google has not seen the page, Google can add the page to the search queue to include it.

The opposite is not true. If a page is not found on the site map, Google will not remove it from its index. Why? Because Google found it crawling the site.

What he seems to believe is that the site map is the total authority that Google uses to know which pages exist on a particular site. This is not the case. The tracking is. The site map only helps Google know Yes they can correctly track your site and, otherwise, what pages are missing from Google that should be added to the search queue.

Hope Google no longer attempts to access the pages because these pages are no longer on your site map is incorrect. Sitemaps are cached and only checked periodically. Why? Because it is an audit process.

You have a real problem that you must solve.

It is returning a 500 error for pages that are not found. This is bad. Your site should return a 404 error not found. Error 500 is a system error and Google will treat the condition as temporary. If your site returned a 404 error, Google will continue testing the page for several attempts over a period of time until you decide that the page no longer exists. If possible, you want to issue a 410 Deleted error for the pages you have deleted. If this is too much work or is not possible, 404 will equal the same over time.

You need to correct your 500 error.

Google Search Console, is creating URLs that do not exist in my Sitemap and then complains that these pages have errors

You have a misconception of what a site map is.

The site map is used to audit the site crawl by the search engine bot. The site map and the crawling of a site are two different and independent things. Google will continue to track your site regardless of any sitemap. The site map will be used to audit / see if Google can properly track your site. For example, if pages are found on your site map and Google has not seen the page, Google can add the page to the search queue to include it.

The opposite is not true. If a page is not found on the site map, Google will not remove it from its index. Why? Because Google found it crawling the site.

What he seems to believe is that the site map is the total authority that Google uses to know which pages exist on a particular site. This is not the case. The tracking is. The site map only helps Google know Yes they can correctly track your site and, otherwise, what pages are missing from Google that should be added to the search queue.

Hope Google no longer attempts to access the pages because these pages are no longer on your site map is incorrect. Sitemaps are cached and only checked periodically. Why? Because it is an audit process.

You have a real problem that you must solve.

It is returning a 500 error for pages that are not found. This is bad. Your site should return a 404 error not found. Error 500 is a system error and Google will treat the condition as temporary. If your site returned a 404 error, Google will continue testing the page for several attempts over a period of time until you decide that the page no longer exists. If possible, you want to issue a 410 Deleted error for the pages you have deleted. If this is too much work or is not possible, 404 will equal the same over time.

You need to correct your 500 error.

google search console: sitemap could not be retrieved

Having examined the structure of your directory, it would appear that many (all?) Files at the root of the document (where you have uploaded sitemap.xml) are not accessible: all return 404 Not found when publicly accessed.

It seems that this could be a Laravel application and maybe you're rewriting everything in the /public subdirectory (a pattern of typical use with Laravel). In which case you should move sitemap.xml (and any other file that should be publicly accessible) to /public subdirectory.

To confirm this, we would need to see the content of its root .htaccess archive. (Edit your question to include this).

TO UPDATE: your .htaccess The file confirms the above.

RewriteRule ^$ public/ (L)
RewriteRule (.*) public/$1 (L)

From your .htaccess file at the root of the document we can see that all requests are rewritten in the /public subdirectory. So, a request for example.com/sitemap.xml is rewritten internally to example.com/public/sitemap.xml.

This means that the root of your document is "effectively" /home/winnerrs/public_html/publicno /home/winnerrs/public_html, as you thought for the first time.

You need to move the sitemap.xml archive of the public_html directory to the public_html/public directory.

Why doesn't the new Google Search Console read my sitemap?

You may not be doing anything wrong at all.

"General HTTP error" can mean many things.

Internet / Web is not a guaranteed delivery protocol. Things will go wrong occasionally. This is not uncommon even for Google.

If you can access the site map yourself, be patient. Many of us have experienced that Google has problems accessing our resources or pages periodically and it could take a few days before Google has access.

As a side note, the tradition is that sitemaps are at the root of the website as sitemap.xml. I am familiar with creating an index sitemap and additional sitemap files for large sites, but I am not familiar with providing site maps for different languages. I understand what you are doing. Google will try to find a site map in the root directory of the sites from time to time and why I mention it. Maybe you can take advantage of this fact.

seo – How many links per page maximum can we have on an HTML sitemap?

On the XML site map, there is a limit of 50,000 per XML site map. Is there a limit of links to be present in the HTML-based site map?

If I have more than 100k pages or publications, should I use them as a form of pagination configuration for HTML site maps?

PD: The XML sitemap is different from the HTML-based sitemap.

Sitemap best practices for multilingual websites: should I list my URLs in each language or is it enough to use rel = "alternate" + hreflang?

I am working on the site maps of a multilingual website and I have doubts about the best practices to refer to each language version of a page.

For a little background, the website refers to around 20,000 places with community comments and descriptions. The website is available in 5 languages ​​(website.com/fr; website.com/it …)

At the moment, my site map only refers to pages in English and on the site map for each page that I specify for each language (as well as English) as recommended by Google.

In Google Search Console, I see that approximately 75% of pages with valid coverage are described as "indexed, not submitted on Sitemap", which makes me think that the alternative link with the hreflang attribute is not enough to "send" the Google page to index it.

Should I list the pages in the 5 languages ​​on my site map and use them too? in each link?

How to create a 32 million page XML sitemap for the HTML website [on hold]

I have created an HTML website with 32 crore pages. This is a flight booking website. Now I want to create an xml site map for this and implement