Wombat

CI Build Status Dependency Status

Generic Web crawler with an elegant DSL that parses structured data from web pages.

Usage:

gem install wombat

OBS: Requires ruby 1.9

Creating a crawler:

Create a class that includes Wombat::Crawler:

# => github_crawler.rb

#coding: utf-8
require 'wombat'

class GithubCrawler
  include Wombat::Crawler

  base_url "http://www.github.com"
  list_page "/"

  headline "xpath=//h1"

  what_is "css=.column.secondary p", :html

  explore "xpath=//ul/li[2]/a" do |e|
    e.gsub(/Explore/, "LOVE")
  end

  benefits do |b|
    b.first_benefit "css=.column.leftmost h3"
    b.second_benefir "css=.column.leftmid h3"
    b.third_benefit "css=.column.rightmid h3"
    b.fourth_benefit "css=.column.rightmost h3"
  end
end
Run it by calling the instance method crawl:
my_crawler = GithubCrawler.new
my_crawler.crawl

#=> the line above outputs: 

{
  "headline" => "1,316,633 people hosting over 3,951,378 git repositories", 
  "what_is" => "GitHub is the best way to collaborate with others.  Fork, send pull requests and manage all your <strong>public</strong> and <strong>private</strong> git repositories.",
  "explore" => "LOVE GitHub",
  "benefits" => {
    "first_benefit"  => "Team management", 
    "second_benefit" => "Code review", 
    "third_benefit"  => "Reliable code hosting", 
    "fourth_benefit" => "Open source collaboration"
  }
}

For the documentation, please check the project Wiki.

Contributing to Wombat

  • Check out the latest master to make sure the feature hasn't been implemented or the bug hasn't been fixed yet
  • Check out the issue tracker to make sure someone already hasn't requested it and/or contributed it
  • Fork the project
  • Start a feature/bugfix branch
  • Commit and push until you are happy with your contribution
  • Make sure to add tests for it. This is important so I don't break it in a future version unintentionally.
  • Please try not to mess with the Rakefile, version, or history. If you want to have your own version, or is otherwise necessary, that is fine, but please isolate to its own commit so I can cherry-pick around it.

Contributors

Changelog

version 0.3.1

  • Added the ability to provide a block to Crawler#crawl and override the default crawler properties for a one off run (thanks to @danielnc)

Copyright (c) 2012 Felipe Lima. See LICENSE.txt for further details.