Solver Prototxt
Pages 35
- Home
- AWS EC2 GPU enabled Caffe AMI
- Borrowing Weights from a Pretrained Network
- Caffe installing script for ubuntu 16.04 support Cuda 8
- Caffe on EC2 Ubuntu 14.04 Cuda 7
- Caffe Output: .caffemodel .solverstate
- Contributing
- Development
- Excluding Layers: Train and Test Phase
- Faster Caffe Training
- Fine Tuning or Training Certain Layers Exclusively
- GeForce GTX 1080, CUDA 8.0, Ubuntu 16.04, Caffe
- IDE Nvidia’s Eclipse Nsight
- Image Format: BGR not RGB
- Install Caffe on EC2 from scratch (Ubuntu, CUDA 7, cuDNN 3)
- Installation
- Installation (OSX)
- Making Prototxt Nets with Python
- Model Zo
- Model Zoo
- Models accuracy on ImageNet 2012 val
- OpenCV 3.2 Installation Guide on Ubuntu 16.04
- Python Layer Unit Tests
- Related Projects
- Reporting Bugs and Other Issues
- Simple Example: Sin Layer
- Solver Prototxt
- The Data Layer
- The Datum Object
- Training and Resuming
- Ubuntu 14.04 ec2 instance
- Ubuntu 14.04 VirtualBox VM
- Ubuntu 16.04 or 15.10 Installation Guide
- Using a Trained Network: Deploy
- Working with Blobs
- Show 20 more pages…
Clone this wiki locally
The solver.prototxt is a configuration file used to tell caffe how you want the network trained.
Parameters
base_lr
This parameter indicates the base (beginning) learning rate of the network. The value is a real number (floating point).
lr_policy
This parameter indicates how the learning rate should change over time. This value is a quoted string.
Options include:
- "step" - drop the learning rate in step sizes indicated by the gamma parameter.
- "multistep" - drop the learning rate in step size indicated by the gamma at each specified stepvalue.
- "fixed" - the learning rate does not change.
- "exp" - gamma^iteration
- "poly" -
- "sigmoid" -
gamma
This parameter indicates how much the learning rate should change every time we reach the next "step." The value is a real number, and can be thought of as multiplying the current learning rate by said number to gain a new learning rate.
stepsize
This parameter indicates how often (at some iteration count) that we should move onto the next "step" of training. This value is a positive integer.
stepvalue
This parameter indicates one of potentially many iteration counts that we should move onto the next "step" of training. This value is a positive integer. There are often more than one of these parameters present, each one indicated the next step iteration.
max_iter
This parameter indicates when the network should stop training. The value is an integer indicate which iteration should be the last.
momentum
This parameter indicates how much of the previous weight will be retained in the new calculation. This value is a real fraction.
weight_decay
This parameter indicates the factor of (regularization) penalization of large weights. This value is a often a real fraction.
solver_mode
This parameter indicates which mode will be used in solving the network.
Options include:
- CPU
- GPU
snapshot
This parameter indicates how often caffe should output a model and solverstate. This value is a positive integer.
snapshot_prefix:
This parameter indicates how a snapshot output's model and solverstate's name should be prefixed. This value is a double quoted string.
net:
This parameter indicates the location of the network to be trained (path to prototxt). This value is a double quoted string.
test_iter
This parameter indicates how many test iterations should occur per test_interval. This value is a positive integer.
test_interval
This parameter indicates how often the test phase of the network will be executed.
display
This parameter indicates how often caffe should output results to the screen. This value is a positive integer and specifies an iteration count.
type
This parameter indicates the back propagation algorithm used to train the network. This value is a quoted string.
Options include:
- Stochastic Gradient Descent "SGD"
- AdaDelta "AdaDelta"
- Adaptive Gradient "AdaGrad"
- Adam "Adam"
- Nesterov’s Accelerated Gradient "Nesterov"
- RMSprop "RMSProp"