Skip to content

SGT-Cho/LLM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

15 Commits
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

🧠 Large Language Models (LLMs): Fundamentals and Enhancement Techniques

License: MIT

LLM 기술의 핡심 κ°œλ…κ³Ό ν–₯상 기법을 μ²΄κ³„μ μœΌλ‘œ μ •λ¦¬ν•œ μ €μž₯μ†Œμž…λ‹ˆλ‹€. 이둠 μ„€λͺ…κ³Ό μ‹€μ œ κ΅¬ν˜„ μ˜ˆμ‹œλ₯Ό ν¬ν•¨ν•©λ‹ˆλ‹€.

πŸ“š Table of Contents

  1. LLM Overview
  2. Enhancement Techniques
  3. Contributing
  4. References

🧩 LLM Overview

What are LLMs?

λŒ€κ·œλͺ¨ μ–Έμ–΄ λͺ¨λΈ(LLM)은 λ°©λŒ€ν•œ ν…μŠ€νŠΈ λ°μ΄ν„°λ‘œ ν•™μŠ΅λœ λ”₯λŸ¬λ‹ λͺ¨λΈλ‘œ, μžμ—°μ–΄ 이해 및 생성 μž‘μ—…μ—μ„œ 인간 μˆ˜μ€€μ˜ μ„±λŠ₯을 λ‹¬μ„±ν•©λ‹ˆλ‹€.

Model Types

μ’…λ₯˜ μ˜ˆμ‹œ λͺ¨λΈ νŠΉμ§•
Autoregressive GPT-4, LLaMA 순차적 ν…μŠ€νŠΈ 생성
Autoencoder BERT, RoBERTa μ–‘λ°©ν–₯ λ¬Έλ§₯ 이해
Multimodal CLIP, Flamingo ν…μŠ€νŠΈ+이미지 처리

πŸ›  Enhancement Techniques

1. RAG (Retrieval-Augmented Generation)

πŸ” 지식 증강 생성

  • μ™ΈλΆ€ 지식 λ² μ΄μŠ€μ™€ κ²°ν•©ν•˜μ—¬ 정확도 ν–₯상
  • κ΅¬ν˜„ ν”„λ ˆμž„μ›Œν¬: LangChain, Haystack
  1. 지식 베이슀 λͺ…μ‹œμ  ν‘œκΈ°
  2. μ‹€μ œ λ™μž‘ 방식 반영
  3. 핡심 μ»΄ν¬λ„ŒνŠΈ κ°•μ‘°
graph TD
    A[User Question] --> B(Query Embedding)
    B --> C{Vector DB Search}
    C --> D[Top-k Documents]
    D --> E[Context Filtering]
    E --> F[Re-ranking]
    F --> G{LLM Generator}
    G --> H[Final Answer]
    
    subgraph Knowledge Base
        C -->|μ—°κ²°| I[(Chunk Storage)]
        I --> J[Metadata]
        I --> K[Text Embeddings]
    end
    
    G -->|μš”μ²­| M[External APIs]
    H -->|Feedback| A
Loading

2. Fine-tuning

🎯 도메인 νŠΉν™” ν•™μŠ΅

  • 사전 ν•™μŠ΅ λͺ¨λΈμ„ νŠΉμ • μž‘μ—…μ— 맞좰 μ‘°μ •
  • μ£Όμš” 기법:
    • Full Fine-tuning
    • LoRA (Low-Rank Adaptation)
    • Prompt Tuning

I tried the LoRA method to finetune llama 3.1 8b model. dataset is about korean food written in Korean. It worked well compared to the original llama 3.1

3. Quantization

βš–οΈ λͺ¨λΈ κ²½λŸ‰ν™”

  • FP32 β†’ INT8 λ³€ν™˜μœΌλ‘œ 4λ°° κ²½λŸ‰ν™”
  • μΆ”λ‘  속도 2-3λ°° ν–₯상

4. Multimodal Integration

🌐 λ©€ν‹° λͺ¨λ‹¬ 톡합

  • ν…μŠ€νŠΈ + 이미지/λΉ„λ””μ˜€/μ˜€λ””μ˜€ 처리
  • μ£Όμš” μ•„ν‚€ν…μ²˜:
    • Cross-modal Attention
    • Fusion Networks

πŸ“š References

About

Making my own LLM

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors