Skip to content
View Egoluback's full-sized avatar
:octocat:
:octocat:

Highlights

  • Pro
Block or Report

Block or report Egoluback

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Egoluback/README.md

Hi there, I'm George Kokush

Computer science student, Intern ML-engineer from Russia 🇷🇺

Kaggle Badge

I'm George Kokush, 18 y. o. HSE&Yandex SHAD student and novice ML/DL-engineer from Russia.

Contact me on: Telegram VK
CV

PyTorch NumPy Pandas Sklearn

Python C++

My last projects

  • "Semantically-Informed Regressive Encoder Score" submission for WMT23 Shared task workshop [ Paper ] [ Repo ]
    • Our task was to develop NN-based metric for text evaluation(machine translation)
    • We improved our developments from AIRI research project trying different approaches(including use of additional vector representations and contrastive learning)
    • Our approach was on 5th place in chinese-english and hebrew-english language pairs and 11th on english-german language pair
    • Our paper was reviewed and we were invited to EMNLP 23 conference
  • Team submission for Eval4NLP Shared task workshop [ Paper ] [ Repo ]
    • Our task was to develop metric for text evaluation(MT&Summarization) only using prompt-engineering techniques and approaches
    • We tried the new approach based on AutoMQM work
    • Our paper was reviewed and we were invited to IJCNLP-AACL 23 conference
  • "Efficient LLM-based metrics for NLG" research project for AIRI Summer School [ Presentation ] [ Repo ]
    • Our task was to develop NN-based metric for text evaluation(machine translation)
    • We tried to beat GPT4-based GEMBA metric by fine-tuning LLMs for translation evaluation
    • I implemented LLM Encoder+MLP decoder architecture which got the best quality
  • "Multimodality in image2text tasks" research project for 1st year of HSE [ Poster ] [ Repo ]
    • Our task was to develop image2text model for russian language
    • We implemented the BLIP-2 architecture and tested it on various configurations
    • We adapted architecture for russian language and achieved tolerable quality
  • NTI ML contest, 2021 [ Repo ]
    • I used lots of classic ML algorithms(linear and logistic regression, trees, boostings, etc), web-scrapping for data extraction and grid-search for hyperparams search
    • We achieved one of the best scores in final rating
  • Toxic detector bot, pet project [ Repo ]
    • I trained CatBoostClassifiers for toxicity prediction using word2vec embeddings
  • Other pet-projects

Pinned

  1. Technolog796/image_captioning Technolog796/image_captioning Public

    Создание русскоязычной модели для image captioning

    Jupyter Notebook 5 1

  2. nti_ml_20-21 nti_ml_20-21 Public

    Code for 1th and 2th stage of 2020 NTI ML competition.

    Jupyter Notebook 3

  3. Rexhaif/airi_summer_llm_metrics Rexhaif/airi_summer_llm_metrics Public

    "Efficient LLM-based metrics for NLG" project at AIRI 2023 Summer School

    Jupyter Notebook 3

  4. neurowolf neurowolf Public

    Нейропацан, генерирующий демотиваторы с нейроцитатами и нейроволками(russian)

    Jupyter Notebook 5

  5. Toxic_Detector Toxic_Detector Public

    ML-bot that detects toxicity in russian texts.

    Jupyter Notebook 5

  6. nti_ar_19-20 nti_ar_19-20 Public

    Code for all tasks for 2th and final stage from 2019 NTI AR competition.

    Python 2