Google posted a public service announcement saying you should disallow Googlebot from crawling your action URLs. Gary Illyes from Google posted on LinkedIn, “You should really disallow crawling of your action URLs. Crawlers will not buy that organic non-GMO scented candle, nor do they care for a wishlist.”
I mean, this is not new advice. Why let a spider crawl pages where it cannot really take any actions. Googlebot can not make purchases, cannot sign up for your newsletter, and so on.
Gary wrote:
A common complaint we get about crawling is that were crawling too much, which uses too much of the server’s resources (though doesn’t cause problems otherwise). Looking at what we’re crawling from the sites in the complaints, way too often it’s action URLs such as “add to cart” and “add to wishlist”. These are useless for crawlers and you likely don’t want them to be crawled.
If you have URLs like:
https://example․com/product/scented-candle-v1?add_to_cart
and
https://example․com/product/scented-candle-v1?add_to_wishlist
How should you block Googlebot? He said, “You should probably add a disallow rule for them in your robots.txt file. Converting them to HTTP POST method also works, though many crawlers can and will make POST requests, so keep that in mind.”
Now, a few years ago, we reported that Googlebot can add products to your cart to verify your pricing is correct. It seems to be part of the merchant shopping experience score feature – so I’d be a tad careful with all of this.
Forum discussion at LinkedIn.
Note: This was pre-written and scheduled to be posted today, I am currently offline for Shavout.
Source link : Seroundtable.com