3a9LL / gospidy

Concurrent web crawler on Go

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Spidy

Web crawler written on Go using Colly.

Usage

docker run --rm 3abpwasm/spidy -u "https://example.com"

Usage:

Usage of /bin/spidy:
  -d uint
        Maximum depth, 0 - no limit (default 2)
  -r uint
        Requests per second, 0 - no limit (default 5)
  -u string
        Target Uri, required

Build from source

Get requirements:

$ go get -u github.com/gocolly/colly/...

Download source code:

$ go github.com/3a9LL/spidy

Go to project directory like $GOPATH/github.com/3a9LL/spidy and build this:

$ go build -ldflags="-s -w"

Example

$ docker run --rm 3abpwasm/spidy -u "https://golang.org/doc/articles/go_command.html"
[+] Spidy started working
http://golang.org
https://golang.org/doc/articles/go_command.html
https://golang.org
https://golang.org/doc/tos.html
https://golang.org/blog/
https://tour.golang.org/
https://golang.org/
https://golang.org/doc/
https://golang.org/pkg/
https://golang.org/project/
https://golang.org/dl/
https://blog.golang.org/
https://golang.org/help/
https://golang.org/LICENSE
https://golang.org/doc/code.html
http://play.golang.org/
https://golang.org/cmd/go/
[+] Spidy finished

License

MIT

About

Concurrent web crawler on Go

License:MIT License


Languages

Language:Go 100.0%