No-Caffeine-No-Gain's Deep Knowledge Tracing (DKT)
-
Updated
Jul 29, 2021 - Python
No-Caffeine-No-Gain's Deep Knowledge Tracing (DKT)
Easy Custom Losses for Tree Boosters using Pytorch
Kaggle Competition - Analysis and prediction of PUBG players' finishing placement based on their final stats
This repository contains the code to build a prediction engine for London housing prices
Repository for the "Google Analytics Customer Revenue Prediction" Kaggle competition.
Project for applied classical ML course at the Weizmann institute
LGBM and logistic regression for prediction of customers' second time transaction for an online market app.
Find LGBM Hyperparams and train the model
Microsoft Malware Prediction Challenge. 8th position solution.
In this project I used basic classification algorithms including random forest, xgboost and decision tree to reach the best solution, finding out how many survived in the titanic accident. I use Kaggles free GPUs and Datasets in this project. I used different feature engineering techniques to clean my data
A project that demonstrates the use of the lgbm C++ API to perform inference without any python dependencies.
Music Genre Recommender website that can identify and recommend 10 different genres of music using Light Gradient Boosting Machine (LGBM). An accuracy of 90% was achieved on the test set by tuning the hyperparameters of the model with Optuna.
A collection of LightGBM callbacks. (DART early stopping, tqdm progress bar)
Add a description, image, and links to the lgbm topic page so that developers can more easily learn about it.
To associate your repository with the lgbm topic, visit your repo's landing page and select "manage topics."