-
Notifications
You must be signed in to change notification settings - Fork 0
pshizhsysu/dwarmstart
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
Install ======= in unix, run 'make' command to obtain 'train' `train' Usage ============= "Usage: train [options] training_set_file [model_file]\n" "options:\n" "-c cost : set the parameter C (default 1)\n" "-e epsilon : set termination criterion for TRON, default 0.01\n" "-g epsilon_cg : set termination criterion for CG, default 0.1\n" "-v k: k-fold cross validation\n" "-C : find the best C\n" Examples ======== > train data_file Train a logistic regression model. > train -v 5 -e 0.001 data_file Do five-fold cross-validation using LR. Use a smaller stopping tolerance 0.001 than the default 0.1 if you want more accurate solutions.
About
A distributed algorithm for parameter selection of logistic regression using warm start
Resources
Stars
Watchers
Forks
Releases
No releases published