Skip to content

Extremesarova/ds_resources

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

22 Commits
 
 

Repository files navigation

Data Science Resources

For interview preparation and learning

Table of Contents:

Algorithms and Data Structures

Platforms

Courses

Resources

Articles

Books

Python

Clean Code

Theory

Questions

Other

Practice

SQL

Courses

Practice

Machine Learning

Sites

Courses

Books

Cheetsheets

Articles

Applied ML

Feature Engineering

Tutorials

Blog posts

Other

Deep Learning

Books

Courses

Tutorials

Blog posts

Other

NLP

Books

Courses

General

Large Language Models (LLMs) / Transformers

Reading papers with AI

Prompt Engineering

Tutorials

Blog posts

Articles

  • Word2Vec, Mikolov et al., Efficient Estimation of Word Representations in Vector Space
  • FastText, Bojanowski et al., Enriching Word Vectors with Subword Information
  • Attention, Bahdanau et al., Neural Machine Translation by Jointly Learning to Align and Translate
  • Transformers, Vaswani et al., Attention Is All You Need
  • BERT, Devlin et al., BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
  • GPT-2, Radford et al., Language Models are Unsupervised Multitask Learners
  • GPT-3, Brown et al, Language Models are Few-Shot Learners
  • LaBSE, Feng et al., Language-agnostic BERT Sentence Embedding
  • CLIP, Radford et al., Learning Transferable Visual Models From Natural Language Supervision
  • RoPE, Su et al., RoFormer: Enhanced Transformer with Rotary Position Embedding
  • LoRA, Hu et al., LoRA: Low-Rank Adaptation of Large Language Models
  • InstructGPT, Ouyang et al., Training language models to follow instructions with human feedback
  • Scaling laws, Hoffmann et al., Training Compute-Optimal Large Language Models
  • FlashAttention, Dao et al., FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness
  • NLLB, NLLB team, No Language Left Behind: Scaling Human-Centered Machine Translation
  • Q8, Dettmers et al., LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale
  • Self-instruct, Wang et al., Self-Instruct: Aligning Language Models with Self-Generated Instructions
  • Alpaca, Taori et al., Alpaca: A Strong, Replicable Instruction-Following Model
  • LLaMA, Touvron, et al., LLaMA: Open and Efficient Foundation Language Models

Computer Vision

Graphs

Reinforcement Learning

RecSys

Courses

Books

Other

Time Series

Big Data

Books

Other