Skip to content

frakev/robotstxt

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 

Repository files navigation

Robots.txt checker

  • Pass an url as parameter
  • Get and parse robots.txt url
  • Check if user-agent is allowed to crawl the url passed as parameter
  • Return true or false, error if exists (HTTP 403 Forbidden for example)

Installation:

go get github.com/frakev/robotstxt

Example:

package main

import (
  rbt "github.com/frakev/robotstxt"
)

func main() {
  access, err := rbt.IsAllowed("http://www.google.com/search") // true or false, err if exists
 }

TO DO:

  • Pass user-agent as parameter

About

robots.txt parser

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages