Skip to content
/ nrobots Public
forked from sjdirect/nrobots

The Robots Exclusion Protocol or robots.txt protocol, is a convention to prevent cooperating web spiders and other web robots from accessing all or part of a website which is otherwise publicly viewable. This project provides an easy-to-use class, implemented in C#, to work with robots.txt files.

License

Notifications You must be signed in to change notification settings

staaam/nrobots

 
 

Repository files navigation

NRobots.txt

A robots.txt parser written in c#.

This is an unofficial fork of NRobots.txt on codeplex with extended functionality. For full documentation see https://nrobots.codeplex.com/.

Project Description:

The Robots Exclusion Protocol or robots.txt protocol, is a convention to prevent cooperating web spiders and other web robots from accessing all or part of a website which is otherwise publicly viewable. This project provides an easy-to-use class, implemented in C#, to work with robots.txt files.

Features:

  • Loading Robots.txt files by providing Url or file content
  • Easy-to-use and simple usage
  • Fluent interface
  • Supports multiple User-Agents
  • Supports different types of entries:
  • Disallow entries
  • Allow entries
  • Sitemap entries
  • Crawl-delay entries
  • Supports comments
  • Supports wild cards (both * and $)

About

The Robots Exclusion Protocol or robots.txt protocol, is a convention to prevent cooperating web spiders and other web robots from accessing all or part of a website which is otherwise publicly viewable. This project provides an easy-to-use class, implemented in C#, to work with robots.txt files.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • C# 100.0%