Skip to content

DSA-MLOPS/DSAA6000I

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

67 Commits
 
 
 
 

Repository files navigation

DSA 6000E: Model to Production: ML Ops for Real Industry

Spring 2024: Friday 9:00AM-11:50PM HKT @Rm202, W4 or Zoom

Co-instructors

  • Sung Kim
  • Jungwoo Ha (Adjunct Prof)

Course Overview:

This course will provide students with a comprehensive understanding of Large Language Models (LLMs) and their practical applications in a production environment, using LLM Ops methodology. We will explore the latest developments in LLMs and LLM Ops, as well as hands-on training in developing, deploying, and evaluating LLMs.

Potential Project Topics:

TBA

Grading breakdown:

  • Participations: 10%
  • Inclass test: 10%
  • Homework: 30%
  • Mid-project proposal: 20%
  • Final project: 30%

Course Outline:

Week 1 Jan 26 (No class)

Week 2 Feb 2: In-class finetuning and test - GZ

Spring Festival Break Feb 9

Spring Festival Break Feb 16

Feb 19: (Optional) Dinner GZ

Week 4 Feb 23: Introduction to Large Language Models (LLMs) and LLM Ops (Sung Kim) @GZ + Zoom

Week 5 Mar 1: A Survey of Large Language Models (John) @GZ

Week 6 Mar 8: Matthias Gallé from Cohere (7AM HKT) @GZ

Week 7 Mar 15: LLM project proposal.

Week 8 Mar 22: QWen 2.0 @Zoom

Week 9 Mar 29: LLM pre-training, ecosystem, and Sovereign LLM (Jung-Woo Ha, Head of Naver AI, Adjunct Prof at HKUST) @Zoom

Qingming Festival Break Apr 5: HW Read: LLM Evaluations, https://arxiv.org/abs/2307.03109

Week 10 Apr 12: Towards Sustainable Scale-Up of LLMs and Beyond, Prof. FU Jie

Increasing the model size, dataset size, and amount of compute for training has been shown to steadily improve the performance of Large Language Models (LLMs). However, unlike labs affiliated with companies like Google, which have access to vast computational resources, academic labs face the challenge of finding alternative and more sustainable ways of scaling up LLMs. In this talk, I will describe our journey of pre-training a 7B-parameter model from scratch. I will delve into the technical aspects of our approach, including the architecture of our model, the training dataset, and the optimization techniques employed. Furthermore, I will discuss the computational resources and infrastructure utilized, highlighting the challenges faced and the solutions implemented to overcome them within an academic setting. In addition to the practical experience of training a large-scale LLM, I will also share some of our ongoing investigations into modular design and continual learning as potential avenues for sustainable scale-up.

Week 11 Apr 19 LLM-as-a-Judge: A futuristic way of evaluating LLMs, Jamin Shin

Week 12 Apr 26: Project Presentations Poster session GZ

  • Students will learn about monitoring and maintenance techniques for LLM models in a production environment, and will set up monitoring and alerting mechanisms for their deployed models.
  • Final project presentations to the class, and feedback from instructors and peers.
  • Course review and wrap-up.

Overall, the course will provide students with a solid understanding of LLM Ops methodology, as well as hands-on experience in developing, deploying, and evaluating LLMs in a production environment. This skill set is essential for any career in the field of natural language processing. The course will be co-taught by Sung Kim and Jungwoo Ha, who have extensive experience in the field and have worked on a variety of LLM-based projects.

We encourage students with a background in natural language processing, machine learning, or data science to enroll in this course. Students should have experience with programming languages such as Python, and familiarity with deep learning frameworks such as TensorFlow or PyTorch.

References

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published