Skip to content
View JerryYin777's full-sized avatar
🎯
Focusing
🎯
Focusing

Highlights

  • Pro
Block or Report

Block or report JerryYin777

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
JerryYin777/README.md

Hi there(Update in 5.15) 👋

News:

  • One Paper Accepted by ACL2024 main conference
  • I am actively looking forward to 25Spring/25Fall Ph.D. Positions in USA.
  • I am located in Haidian District, Beijing at SenseTime Research from May to late August, feel free to contact me! (and eat) (Note: From Github)
JerryYin777

Biography

I am currently a junior Undergraduate Student pursuing a bachelor's Degree in computer science at College of Liberal Arts, University of Minnesota Twins Cities. (Wechat: JerryYin777)

My research interests lie in Large Multimodal Models (LMMs) and their application in diverse practical scenarios, such as biological and system large models. My focus also extends to developing efficient machine learning systems aimed at expediting the training and inference processing of LMMs (especially LLMs), leveraging expertise in high-performance computing and distributed systems.

Before transferring to the University of Minnesota, I studied at Nanchang University, majoring in Artificial Intelligence in a top-tier class with a School Academic Special Scholarship. I was honored to be advised by Professor Zichen Xu at GOOD LAB starting from March 2022, where my focus was on solving data-centric challenges and building efficient and reliable systems. I was the leader of Nanchang University Supercomputer Cluster Team (NCUSCC) Leader, with experience of ASC22 and SC23(IndySCC).

I was also fortunately recruited as a research assistant at TsinghuaNLP in Beijing from July to September 2023, advised by Professor Zhiyuan Liu, trying to build efficient distributed large language model training framework BMTrain and Develop 10B Chinese LLM CPM-Bee.

I am passionate about open source and firmly believe in its potential to disseminate knowledge widely, leverage technology to lead innovation to the world and contribute to the advancement of human society. I am proud to have garnered over 1000 stars and acquired 189 followers on GitHub. It is gratifying to know that my open-source projects have benefitted numerous individuals, and I have personally gained valuable knowledge from the open-source community.

Here is my CV in English Version, Template here Using Typst.

Contact

Education

2023.12 - 2024.12 (Expected), Computer Science, College of Liberal Arts, University of Minnesota Twin Cities.

2021.09 - 2023.12, Artificial Intelligence, School of Information Engineering, Nanchang University.

Research Interests

Natural Language Processing, Machine Learning System.

Skills

  • Natural Language Processing: Proficient in using the PyTorch framework, with the ability to reproduce mainstream large-scale models in the industry (such as Baichuan, llama2, Qwen). Proficient in using quantization and inference tools such as QLoRA, vLLM, and skilled in distributed parallel training (using training tools such as BMTrain, DeepSpeed).

  • High Performance Computing: Proficient in CUDA Programming, familiar with C++, knowledgeable in compiler optimization principles, and understanding of MPI, OpenMP, and SIMD acceleration optimization technologies.

  • Computer System Architecture: Familiar with GPU architecture and RISC-V instruction set, and has participated in the One Life, One Core project.

  • Other: Understanding of serverless computing architecture, experience in applying federated learning in network security, involvement in both front-end and back-end development, familiarity with Linux operating system and operation and maintenance ( maintained a large server cluster with 21 nodes in the GOOD LAB). Additionally, has researched computer vision and reinforcement learning in the field of artificial intelligence.

If you want to know more about me, feel free to click on my Personal Website.

Let's Accelerate!

My side projects

BMTrain: Efficient Training (including pre-training and fine-tuning) for Big Models

Nano-GPT-Pytorch2.0

NaturalCC: An Open-Source Toolkit for Code Intelligence

Nano-GPT-Pytorch2.0

Nano-GPT-Pytorch2.0 Implementation: Pytorch2.0-Implementation, faster and simpler

Nano-GPT-Pytorch2.0

Cr's Research Toolchain

Nano-GPT-Pytorch2.0

Intelligent Creation Platform Creator: Implementing GPT-2-based Title and Summary Generation Software

Intelligent Creation Platform Creator

Keep writing to share my knowledge about computer science and artificial intelligence in Chinese

Zhihu

Get in touch

GitHub Zhihu Google Scholar

Pinned Loading

  1. OpenBMB/BMTrain OpenBMB/BMTrain Public

    Efficient Training (including pre-training and fine-tuning) for Big Models

    Python 531 73

  2. CGCL-codes/naturalcc CGCL-codes/naturalcc Public

    NaturalCC: An Open-Source Toolkit for Code Intelligence

    Python 248 41

  3. NanoGPT-Pytorch2.0-Implementation NanoGPT-Pytorch2.0-Implementation Public

    This is a repo for my NanoGPT Pytorch2.0 Implementation when torch2.0 released soon, faster and simpler, a good tutorial learning GPT.

    Python 59 2

  4. Cr_Research_Toolchain Cr_Research_Toolchain Public

    Sharing my research toolchain

    66 6

  5. FPGA_Competition-RISC-V_Processor-in-PGL22G FPGA_Competition-RISC-V_Processor-in-PGL22G Public

    FPGA Innovation Design Competition:RISC-V Processor-based Hardware and Software Design in PGL22G

    Verilog 10 3

  6. IntelligentCreator IntelligentCreator Public

    Intelligent Creation Platform Creator: Implementing GPT-2-based Title and Summary Generation Software

    HTML 3 1