# # robots.txt # # This file is to prevent the crawling and indexing of certain parts # of your site by web crawlers and spiders run by sites like Yahoo! # and Google. By telling these "robots" where not to go on your site, # you save bandwidth and server resources. # # This file will be ignored unless it is at the root of your host: # Used: http://example.com/robots.txt # Ignored: http://example.com/site/robots.txt # # For more information about the robots.txt standard, see: # http://www.robotstxt.org/wc/robots.html # # For syntax checking, see: # http://www.sxw.org.uk/computing/robots/check.html User-agent: URLAppendBot/1.0; +http://www.profound.net/urlappendbot.html Disallow: / User-agent: TweetmemeBot/3.0; +http://tweetmeme.com/ Disallow: / User-agent: MJ12bot/v1.4.3, http://www.majestic12.co.uk/bot.php?+ Disallow: / User-agent: Sosospider/2.0 Disallow: / User-agent: Python-urllib/2.6 Disallow: / User-agent: python-requests/1.2.3 CPython/2.7.3 Linux/3.3.8-gcg-201308121035 Disallow: / User-agent: python-requests/1.2.3 CPython/2.7.2+ Linux/3.0.0-16-virtual Disallow: / User-agent: python-requests/1.2.3 CPython/2.7.4 Linux/3.8.11-ec2 Disallow: / User-agent: Python-urllib/2.7 Disallow: / User-agent: Python-httplib2/0.7.7 (gzip) Disallow: / User-agent: python-requests/1.1.0 CPython/2.6.8 Linux/3.4.48-45.46.amzn1.x86_64 Disallow: / # Allow all other search engines and bots User-agent: * Allow:/ # sitemaps Sitemap: https://vikaspedia.in/sitemap-main.xml Sitemap: https://vikaspedia.in/sitemap_index.xml