# Enables robots.txt rules for all crawlers User-agent: SemrushBot Disallow: / User-agent: * # Crawl delay Crawl-delay: 20 # Allow sitemap.xml Allow: /sitemaps/* # Allow Allow: /*?p= Allow: /*?variant # Disallow directories Disallow: /404/ Disallow: /app/ Disallow: /cgi-bin/ Disallow: /dev/ Disallow: /downloader/ Disallow: /errors/ Disallow: /feeds/ Disallow: /includes/ Disallow: /js/ Disallow: /lib/ Disallow: /magento/ Disallow: /pkginfo/ Disallow: /report/ Disallow: /scripts/ Disallow: /shell/ Disallow: /skin/ Disallow: /stats/ Disallow: /var/ # Disallow some paths (clean URLs) Disallow: /admin/ Disallow: /catalogsearch/result/ Disallow: /catalog/product_compare/ Disallow: /catalog/category/view/ Disallow: /catalog/product/view/ Disallow: /catalogsearch/ Disallow: /checkout/ Disallow: /control/ Disallow: /contacts/ Disallow: /customer/ Disallow: /customize/ Disallow: /newsletter/ Disallow: /poll/ Disallow: /review/ Disallow: /sales/ Disallow: /sendfriend/ Disallow: /tag/ Disallow: /wishlist/ # Disallow some paths (no clean URLs) Disallow: /*.php$ Disallow: /*.html$ Disallow: /*.js$ Disallow: /*.css$ Disallow: /*.sh$ Disallow: /*?*