Extensive EDA of the IBM telco customer churn dataset, implemented various statistical hypotheses tests and Performed single-level Stacking Ensemble and tuned hyperparameters using Optuna.
-
Updated
Nov 9, 2021 - HTML
Extensive EDA of the IBM telco customer churn dataset, implemented various statistical hypotheses tests and Performed single-level Stacking Ensemble and tuned hyperparameters using Optuna.
What important conclusion a company and an employee can take out of Analysis and Predicting Salary
Bayesian optimisation of prophet temperature model parameters with daily and yearly seasonalities plus extra regressors
This repository focuses on the Neural Networks and deep learning. It is a workbook you can refer it for a reference. I'll Include following content here: Neurons, perceptrons, weight and biases, learning rate, activation form, hyper parameters, RNN, CNN and other popular concepts.
Evolution-inspired optimisation algorithms
Practice hyper parameter tuning in Neural Network.
Recommender System. Politecnico di Milano, A.A. 2021-2022
Gauging how Support Vector Machine Algorithm behaves with Hyperparameter Tuning
Machine learning classification model with streamlit deployment.
Library for integrated use of H2O with Hyperopt
HOLA: Hyperparameter Optimization, Lightweight Asynchronous
Model to predict bank customer churn
The jupyter notebooks of the deep learning specialization by deeplearning.ai
AutoML Python package with ML models with builtin Hyperparameter Optimization and easy to use API.
A simple regression analysis of house prices in USA with 11 features selected on MECE Framework
This repository offers a robust solution for multilabel image classification. Utilizing advanced neural networks like VGG16, VGG19, ResNet50, InceptionV3, DenseNet121, and MobileNetV2, the project achieves precise classification across 107 diverse categories.
Add a description, image, and links to the hyperparameter-optimization topic page so that developers can more easily learn about it.
To associate your repository with the hyperparameter-optimization topic, visit your repo's landing page and select "manage topics."