Are the links on canvas / WebGL with a sitemap good enough for SEO, or do I also need normal href links?

The main page of my site consists of great graphics and animations implemented with Canvas / WebGL using the pixi.js library. The problem is that all links are also implemented as interaction with the WebGL layer.

But I know that I can list my links on sitemap.xml. Is it enough to list the links on the site map for a good SEO in this situation?

Or should I somehow verify something and insert normal a instead?

robots.txt: Google Search Console and Sitemap settings for a website subfolder

You may have a different XML sitemap configured for your folder and it will not confuse search engines. Source.

Just keep in mind that they are correctly linked from the main site map.

But when it comes to GSC or Analytics, it is not possible to implement it in folders.

Google Search Console is creating URLs that do not exist in my Sitemap and then complains that these pages have errors

You have a misconception of what a site map is.

The site map is used to audit the site crawl by the search engine bot. The site map and the crawling of a site are two different and independent things. Google will continue to track your site regardless of any sitemap. The site map will be used to audit / see if Google can properly track your site. For example, if pages are found on your site map and Google has not seen the page, Google can add the page to the search queue to include it.

The opposite is not true. If a page is not found on the site map, Google will not remove it from its index. Why? Because Google found it crawling the site.

What he seems to believe is that the site map is the total authority that Google uses to know which pages exist on a particular site. This is not the case. The tracking is. The site map only helps Google know Yes they can correctly track your site and, otherwise, what pages are missing from Google that should be added to the search queue.

Hope Google no longer attempts to access the pages because these pages are no longer on your site map is incorrect. Sitemaps are cached and only checked periodically. Why? Because it is an audit process.

You have a real problem that you must solve.

It is returning a 500 error for pages that are not found. This is bad. Your site should return a 404 error not found. Error 500 is a system error and Google will treat the condition as temporary. If your site returned a 404 error, Google will continue testing the page for several attempts over a period of time until you decide that the page no longer exists. If possible, you want to issue a 410 Deleted error for the pages you have deleted. If this is too much work or is not possible, 404 will equal the same over time.

You need to correct your 500 error.

Google Search Console, is creating URLs that do not exist in my Sitemap and then complains that these pages have errors

You have a misconception of what a site map is.

The site map is used to audit the site crawl by the search engine bot. The site map and the crawling of a site are two different and independent things. Google will continue to track your site regardless of any sitemap. The site map will be used to audit / see if Google can properly track your site. For example, if pages are found on your site map and Google has not seen the page, Google can add the page to the search queue to include it.

The opposite is not true. If a page is not found on the site map, Google will not remove it from its index. Why? Because Google found it crawling the site.

What he seems to believe is that the site map is the total authority that Google uses to know which pages exist on a particular site. This is not the case. The tracking is. The site map only helps Google know Yes they can correctly track your site and, otherwise, what pages are missing from Google that should be added to the search queue.

Hope Google no longer attempts to access the pages because these pages are no longer on your site map is incorrect. Sitemaps are cached and only checked periodically. Why? Because it is an audit process.

You have a real problem that you must solve.

It is returning a 500 error for pages that are not found. This is bad. Your site should return a 404 error not found. Error 500 is a system error and Google will treat the condition as temporary. If your site returned a 404 error, Google will continue testing the page for several attempts over a period of time until you decide that the page no longer exists. If possible, you want to issue a 410 Deleted error for the pages you have deleted. If this is too much work or is not possible, 404 will equal the same over time.

You need to correct your 500 error.

google search console: sitemap could not be retrieved

Having examined the structure of your directory, it would appear that many (all?) Files at the root of the document (where you have uploaded sitemap.xml) are not accessible: all return 404 Not found when publicly accessed.

It seems that this could be a Laravel application and maybe you're rewriting everything in the /public subdirectory (a pattern of typical use with Laravel). In which case you should move sitemap.xml (and any other file that should be publicly accessible) to /public subdirectory.

To confirm this, we would need to see the content of its root .htaccess archive. (Edit your question to include this).

TO UPDATE: your .htaccess The file confirms the above.

RewriteRule ^$ public/ (L)
RewriteRule (.*) public/$1 (L)

From your .htaccess file at the root of the document we can see that all requests are rewritten in the /public subdirectory. So, a request for example.com/sitemap.xml is rewritten internally to example.com/public/sitemap.xml.

This means that the root of your document is "effectively" /home/winnerrs/public_html/publicno /home/winnerrs/public_html, as you thought for the first time.

You need to move the sitemap.xml archive of the public_html directory to the public_html/public directory.

Why doesn't the new Google Search Console read my sitemap?

You may not be doing anything wrong at all.

"General HTTP error" can mean many things.

Internet / Web is not a guaranteed delivery protocol. Things will go wrong occasionally. This is not uncommon even for Google.

If you can access the site map yourself, be patient. Many of us have experienced that Google has problems accessing our resources or pages periodically and it could take a few days before Google has access.

As a side note, the tradition is that sitemaps are at the root of the website as sitemap.xml. I am familiar with creating an index sitemap and additional sitemap files for large sites, but I am not familiar with providing site maps for different languages. I understand what you are doing. Google will try to find a site map in the root directory of the sites from time to time and why I mention it. Maybe you can take advantage of this fact.

seo – How many links per page maximum can we have on an HTML sitemap?

On the XML site map, there is a limit of 50,000 per XML site map. Is there a limit of links to be present in the HTML-based site map?

If I have more than 100k pages or publications, should I use them as a form of pagination configuration for HTML site maps?

PD: The XML sitemap is different from the HTML-based sitemap.

Sitemap best practices for multilingual websites: should I list my URLs in each language or is it enough to use rel = "alternate" + hreflang?

I am working on the site maps of a multilingual website and I have doubts about the best practices to refer to each language version of a page.

For a little background, the website refers to around 20,000 places with community comments and descriptions. The website is available in 5 languages ​​(website.com/fr; website.com/it …)

At the moment, my site map only refers to pages in English and on the site map for each page that I specify for each language (as well as English) as recommended by Google.

In Google Search Console, I see that approximately 75% of pages with valid coverage are described as "indexed, not submitted on Sitemap", which makes me think that the alternative link with the hreflang attribute is not enough to "send" the Google page to index it.

Should I list the pages in the 5 languages ​​on my site map and use them too? in each link?

How to create a 32 million page XML sitemap for the HTML website [on hold]

I have created an HTML website with 32 crore pages. This is a flight booking website. Now I want to create an xml site map for this and implement

If my sitemap and robots.txt have http or https URL for the given scenario

Battery exchange network

The Stack Exchange network consists of 175 question and answer communities, including Stack Overflow, the largest and most reliable online community for developers to learn, share their knowledge and develop their careers.

Visit Stack Exchange