Skip to content

cwrw/crawler

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

17 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Crawler

A web crawler using Depth First Search.

A sitemap is best represented as a graph. Algorithm chosen for now is Depth First Searching.

Installation

Add this line to your application's Gemfile:

gem 'crawler'

And then execute:

$ bundle

Or install it yourself as:

$ gem install crawler

Usage

#####Crawler allows you to crawl a url through the following rake task:

$ rake generate_sitemap["https://gocardless.com","/path/to/file","file_name"]

if you leave options empty ["https://gocardless.com"] will default to your desktop (OSx)

TODO

Handle redirects, Tests, Exception handling, Tidy up base class, fix algorithm as not performant

Development

After checking out the repo, run bundle install to install dependencies. Then, run rake to run the tests.

To install this gem onto your local machine, run gem build crawler.gemspec, then gem install ./crawler-0.0.0.gem with the relevant version number

Contributing

Bug reports and pull requests are welcome on GitHub at https://github.com/cwrw/crawler. This project is intended to be a safe, welcoming space for collaboration, and contributors are expected to adhere to the Contributor Covenant code of conduct.

License

The gem is available as open source under the terms of the MIT License.

Resources

algorithms, search with nokogiri, trees in ruby

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published