- Beijing, China
Stars
Contains the code examples from The UVM Primer Book sorted by chapters.
Reference examples and short projects using UVM Methodology
Build a SystemVerilog Environment for an ALU, using OOP testbench components as; stimulus generator, driver, monitor, scoreboard. ALU was verified using QuestaSim.
🧠 Guide to Building RAG (Retrieval-Augmented Generation) Applications
Simple Text-Generator with OpenAI gpt-2 Pytorch Implementation
Code for the paper "Language Models are Unsupervised Multitask Learners"
Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
Learn where some of the network sysctl variables fit into the Linux/Kernel network flow. Translations: 🇷🇺
A lightweight data processing framework built on DuckDB and 3FS.
A high-performance distributed file system designed to address the challenges of AI training and inference workloads.
DeepGEMM: clean and efficient FP8 GEMM kernels with fine-grained scaling
DeepEP: an efficient expert-parallel communication library
Reexamining Direct Cache Access to Optimize I/O Intensive Applications for Multi-hundred-gigabit Networks
Vendor-neutral programmable observability pipelines.
RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.
Fully-featured web interface for Ollama LLMs
🦜🔗 Build context-aware reasoning applications
面向开发者的 LLM 入门教程,吴恩达大模型系列课程中文版
code for benchmarking GPU performance based on cublasSgemm and cublasHgemm
Sample code to test and benchmark large CuFFTs on Nvidia GPUs
A Flexible Framework for Experiencing Cutting-edge LLM Inference Optimizations
Example models using DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
仅需Python基础,从0构建大语言模型;从0逐步构建GLM4\Llama3\RWKV6, 深入理解大模型原理