Getting Started with SigOpt
Welcome to SigOpt Examples. These examples show you how to use SigOpt for model tuning tasks in various machine learning environments.
Most of these examples will run on any Linux or Mac OS X machine from the command line. Each example contains a README.md with specific setup instructions.
If this is your first time using SigOpt, we recommend you work through the Random Forest example. In this example, you will use a random forest to classify data from the iris dataset and use SigOpt to maximize the k-fold cross-validation accuracy by tuning the model's hyperparameters. This example is available in a wide variety of languages and integrations:
- ipython-notebook-example: Simple example of using SigOpt to optimize a 2D function with plots and comparisons in an iPython Notebook.
- sigopt-beats-vegas: Using SigOpt to tune a model to beat the Vegas odds in Python (blog post).
- text-classifier: Example using SigOpt to tune a text classifier in Python and R (blog post).
- unsupervised-model: Example using SigOpt and xgboost to tune a combined unsupervised and supervised model for optical character recognition (blog post)
- tensorflow-cnn: Example using SigOpt and TensorFlow to tune a convolutional neural network's structure and gradient descent algorithm (blog post)
- classifier: Using SigOpt to tune a machine learning classifier in Python (blog post).
- parallel: Examples of running SigOpt from multiple parallel processes in Python (blog post).
Any questions? Drop us a line at firstname.lastname@example.org.
To implement SigOpt for your use case, feel free to use or extend the code in this repository. Our core API can bolt on top of any complex model or process and guide it to its optimal configuration in as few iterations as possible.
With SigOpt, data scientists and machine learning engineers can build better models with less trial and error.
Machine learning models depend on hyperparameters that trade off bias/variance and other key outcomes. SigOpt provides Bayesian hyperparameter optimization using an ensemble of the latest research.
SigOpt can tune any machine learning model, including popular techniques like gradient boosting, deep neural networks, and support vector machines. SigOpt’s REST API, Python, and R libraries integrate into any existing ML workflow.
SigOpt augments your existing model training pipeline, suggesting parameter configurations to maximize any online or offline objective, such as AUC ROC, model accuracy, or revenue. You only send SigOpt your metadata, not the underlying training data or model.