Nokogiri (鋸) is an HTML, XML, SAX, and Reader parser. Among Nokogiri's many features is the ability to search documents via XPath or CSS3 selectors.
XML is like violence - if it doesn’t solve your problems, you are not using enough of it.
XPath 1.0 support for document searching
CSS3 selector support for document searching
Nokogiri parses and searches XML/HTML very quickly, and also has correctly implemented CSS3 selector support as well as XPath 1.0 support.
Before filing a bug report, please read our submission guidelines at:
The Nokogiri mailing list is available here:
The bug tracker is available here:
The IRC channel is #nokogiri on freenode.
require 'nokogiri' require 'open-uri' # Get a Nokogiri::HTML::Document for the page we’re interested in... doc = Nokogiri::HTML(open('http://www.google.com/search?q=sparklemotion')) # Do funky things with it using Nokogiri::XML::Node methods... #### # Search for nodes by css doc.css('h3.r a').each do |link| puts link.content end #### # Search for nodes by xpath doc.xpath('//h3/a').each do |link| puts link.content end #### # Or mix and match. doc.search('h3.r a.l', '//h3/a').each do |link| puts link.content end
ruby 1.9.3 or higher
in Nokogiri 1.6.0 and later libxml2 and libxslt are bundled with the gem, but if you want to use them installed on the system:
libxml2 >=2.6.17 and <2.9.0, with iconv support (libxml2-dev/-devel is required too)
libxslt, built with and supported by the given libxml2 (libxslt-dev/-devel is required too)
Strings are always stored as UTF-8 internally. Methods that return text values will always return UTF-8 encoded strings. Methods that return XML (like to_xml, to_html and inner_html) will return a string encoded like the source document.
Some documents declare one particular encoding, but use a different one. So, which encoding should the parser choose?
Remember that data is just a stream of bytes. Only we humans add meaning to that stream. Any particular set of bytes could be valid characters in multiple encodings, so detecting encoding with 100% accuracy is not possible. libxml2 does its best, but it can't be right 100% of the time.
If you want Nokogiri to handle the document encoding properly, your best bet is to explicitly set the encoding. Here is an example of explicitly setting the encoding to EUC-JP on the parser:
doc = Nokogiri.XML('<foo><bar /><foo>', nil, 'EUC-JP')
sudo gem install nokogiri
Binary packages are available for:
Developing on C Ruby (MRI)
Developing Nokogiri requires racc and rexical to generate the parser and tokenizer. To start development, make sure you have `libxml2` and `libxslt` installed.
Then install core gems and bootstrap:
$ gem install hoe rake-compiler mini_portile $ rake newb
Developing on JRuby
Currently, development with JRuby depends on CRuby being installed. With CRuby, install racc and rexical:
$ gem install racc rexical
Make sure hoe and rake compiler are installed with JRuby:
$ jgem install hoe rake-compiler
Then run rake:
$ jruby -S rake
(The MIT License)
Copyright © 2008 - 2013:
Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the 'Software'), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED 'AS IS', WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.