Skip to content
John Langford's original release of Vowpal Wabbit -- a fast online learning algorithm
Branch: master
Clone or download
Pull request Compare This branch is 66 commits behind VowpalWabbit:master.
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Permalink
Type Name Latest commit message Commit time
Failed to load latest commit information.
.github/ISSUE_TEMPLATE
.pipelines
.scripts
R
big_tests
c_test
cluster
cs
demo
doc
explore
java
library
logo_assets
python
rapidjson
reinforcement_learning
sdl
singularity
test
utl
vowpalwabbit
.clang-format
.editorconfig
.gitattributes
.gitignore
.travis.yml
AUTHORS
CMakeLists.txt
CMakeSettings.json
INSTALL
LICENSE
Makefile
README.Windows.md
README.md
appveyor.yml
build-linux.sh
build-windows-experimental.bat
deployvw.bat
libvw.pc.in
libvw_c_wrapper.pc.in
version.txt

README.md

Vowpal Wabbit

Build Status Windows Build status

Coverage Status Total Alerts Gitter chat

This is the Vowpal Wabbit fast online learning code. For Windows specific info, look at README.Windows.md

Why Vowpal Wabbit?

Vowpal Wabbit is a machine learning system which pushes the frontier of machine learning with techniques such as online, hashing, allreduce, reductions, learning2search, active, and interactive learning. There is a specific focus on reinforcement learning with several contextual bandit algorithms implemented and the online nature lending to the problem well. Vowpal Wabbit is a destination for implementing and maturing state of the art algorithms with performance in mind.

  • Input Format. The input format for the learning algorithm is substantially more flexible than might be expected. Examples can have features consisting of free form text, which is interpreted in a bag-of-words way. There can even be multiple sets of free form text in different namespaces.
  • Speed. The learning algorithm is fast -- similar to the few other online algorithm implementations out there. There are several optimization algorithms available with the baseline being sparse gradient descent (GD) on a loss function.
  • Scalability. This is not the same as fast. Instead, the important characteristic here is that the memory footprint of the program is bounded independent of data. This means the training set is not loaded into main memory before learning starts. In addition, the size of the set of features is bounded independent of the amount of training data using the hashing trick.
  • Feature Interaction. Subsets of features can be internally paired so that the algorithm is linear in the cross-product of the subsets. This is useful for ranking problems. The alternative of explicitly expanding the features before feeding them into the learning algorithm can be both computation and space intensive, depending on how it's handled.

Visit the wiki to learn more.

Getting Started

For the most up to date instructions for getting started on Windows, MacOS or Linux please see the wiki. This includes:

You can’t perform that action at this time.