Skip to content

Minyst/AI

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

🎓 Data Scientist Portfolio


📊 Projects

Project 1: Credit Card Fraud Detection

Objective
To determine which model performs best when data is reduced or augmented.

Technologies Used

  • Dimensionality Reduction: PCA, tSNE, UMAP
  • Dimensionality Augmentation: SMOTE, BorderLineSMOTE, ADASYN
  • Machine Learning Models: RandomForest, XGBoost, CatBoost, LightGBM
  • Deep Learning Models: TensorFlow, Pytorch

Key Results
To compare whether dimensionality reduction or augmentation improves model performance,
I used various machine learning and deep learning models.
As a result, I was able to create a ranking table showing which method and model combination yielded the best performance.
The accuracy was similar, so I ranked them based on the ROC_AUC_SCORE.

URL
https://www.kaggle.com/datasets/mlg-ulb/creditcardfraud


Project 2: YOLOv10 Pretrained vs Custom

Objective
To compare which performs better between the pretrained and custom YOLOv10 models.

Technologies Used

  • Model: YOLOv10
  • Package: ultralytics, supervision, cv2

Key Results
After capturing the video and creating multiple frames,
each frame was trained with the model, and then these frames were reassembled into a single video.
For the pretrained model, predictions were made directly using the model.
For the custom model, pre-prepared data was trained using the original YOLOv10 weights,
and the best weights obtained were selected as the final weights for the model, which was then used for predictions.
This process is similar to a relay race.

When comparing the pretrained and custom models, there was a significant difference.
The custom model, which was provided with images of various classes consistently, had a broader prediction range than the automatically recognizing pretrained model.
However, its accuracy was much lower compared to the pretrained model.

URL
https://github.com/THU-MIG/yolov10
https://docs.ultralytics.com/ko/models/yolov10


Project 3: Detectron2 Pretrained vs Custom

Objective
To compare which performs better between the pretrained and custom Detectron2 models.

Technologies Used

  • Model: Detectron2
  • Package: detectron2, cv2

Key Results
Detectron2 is almost identical to YOLOv10, but there are two key differences.
First, Detectron2 uses Faster RCNN weights, unlike YOLOv10.
Second, while YOLOv10 shows some differences in results between pretrained and custom models,
Detectron2 exhibits no noticeable differences."

URL
https://github.com/facebookresearch/detectron2/blob/main/README.md


Project 4: AI Cover - RVC

Objective
Using the RVC model to make one singer's voice sing another singer's song.

Technologies Used

  • Model: RVC

Key Results
This project can be explained in five steps.
First, split the downloaded YouTube music into vocals and background music.
Second, slice the vocals into multiple segments to enhance the model's learning.
Third, download the RVC_pretrained model.
Fourth, train the model.
Fifth, generate a music file where the singer performs a different song.

I was amazed at how natural the generated music sounded.
Detailed adjustments can be made, and having an expert involved could further improve the synchronization and overall quality.

URL
https://github.com/facebookresearch/demucs
https://github.com/openvpi/audio-slicer


Project 5: CNN - CIFAR-10

Objective
Using CIFAR-10 data, build a complex CNN with TensorFlow and PyTorch.

Technologies Used

  • Models : TensorFlow, Pytorch
  • CNN Process : Data Augmentation, Conv2d, Padding, Batch Normalization, Pooling, Dropout, Flatten

Key Results
All processes of the CNN with TensorFlow and PyTorch are included: Data Augmentation, Padding, Batch Normalization, Pooling, Dropout, Flatten.

URL
https://www.cs.toronto.edu/~kriz/cifar.html


Project 6: CLIP

Objective
To Find out how to use CLIP Model(Zero-shot image classification model) on Web Images and images from computer storage.

Technologies Used

  • Model : CLIP
  • Skill: Zero-shot image classification

Zero-shot image classification is a technique where the model can correctly classify new images even if it hasn't been directly trained on images of a specific class during training. The model leverages pre-learned knowledge and similarities or relationships between different classes it has learned to infer new classes.

CLIP (Contrastive Language-Image Pretraining) is a representative model for zero-shot image classification. It simultaneously learns from both images and text, allowing it to understand the relationship between the two. CLIP employs contrastive learning, where it pairs images with their corresponding text descriptions during training. This enables the model to associate previously unseen classes with appropriate text descriptions, allowing for effective zero-shot classification.

Key Results
The results of web images and images from computer storage predicted by CLIP.

URL
https://github.com/openai/CLIP


Project 7: SAM2

Objective
After detecting objects using YOLO, SAM2 is used to generate segmentation masks for the detected objects, and then the masks are overlaid with colors corresponding to each object to create an output video. YOLO generates bounding boxes, and SAM processes them to handle segmentation, integrating the two models.

Technologies Used
Models: SAM2, YOLO

Key Results
A new video detected by the model.

URL
https://github.com/facebookresearch/segment-anything-2
https://docs.ultralytics.com/ko/models/yolov10


📈 Skills

  • Programming Languages: Python
  • Data Preprocessing: Pandas, NumPy
  • Data Visualization: Matplotlib
  • Machine Learning & Deep Learning: Scikit-Learn, TensorFlow, Pytorch, OpenCV
  • Databases:
  • Tools: Jupyter Notebook, Google Colab

🛠️ Tools & Technologies

Tensorflow Pytorch OpenCV


About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published