simple web crawler to fetch links from a url
Ruby 2+
git clone https://github.com/vaibhavkhl/crawlrb.git
cd crawlrb
bundle
Run bin/crawl https://github.com(any url) 5
cmd from crawldb dir.
- last argument is optional which specifies the max no. of urls to visit to fetch links.
- once execution finishes or user stops it using
ctrl + c
, script exports all the links fetched to links.txt file.
run rspec spec/link_crawler_spec.rb
cmd to run specs.