| Path: | README (CVS) |
| Last Update: | Sat Mar 22 23:02:18 -0700 2008 |
# $Id: README 16 2008-03-23 06:02:07Z warchild $
Hawler, the ruby HTTP crawler.
Written to ease satisfying curiousities about the way the web is woven.
Install by:
gem install --source http://spoofed.org/files/hawler/ hawler
Rough usage is… define a method that will take a URI, a referer, and an Net::HTTPResponse as arguments, and then does whatever it wants. Create a Hawler object, passing the start URI and the above method as arguments. Set Hawler options, then start it.
Basic usage is an implementation of htgrep:
#!/usr/bin/ruby
require 'rubygems'
require 'hawler'
require 'hawleroptions'
def grep (uri, referer, response)
line = 0
unless (response.nil?)
response.body.each do |l|
line += 1
if (l =~ /#{@match}/)
puts "#{uri}:#{line} #{l}"
end
end
end
end
options = HawlerOptions.parse(ARGV, "Usage: #{File.basename $0} [pattern] [uri] [options]")
if (ARGV.empty?)
puts @options.help
else
@match = ARGV.shift
ARGV.each do |site|
crawler = Hawler.new(site, method(:grep))
options.each_pair do |o,v|
crawler.send("#{o}=",v)
end
crawler.start
end
end