Block or Report
Block or report OpenMOSE
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abusePinned
-
RWKV-LM-LISA
RWKV-LM-LISA PublicLayerwise Importance Sampled AdamW for RWKV, aiming for RWKV-5 and 6. SFT, Aligning(DPO,ORPO). Cuda and Rocm6.0. can train 7b on 24GB GPU!
Python 11
-
RWKV5-LM-LoRA
RWKV5-LM-LoRA PublicRWKV v5,v6 LoRA Trainer on Cuda and Rocm Platform. RWKV is a RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN an…
-
RWKV-infctx-trainer-LoRA
RWKV-infctx-trainer-LoRA PublicRWKV v5, v6 infctx LoRA trainer with 4bit quantization,Cuda and Rocm supported, for training arbitary context sizes, to 10k and beyond!
-
RWKV-LM-RLHF-DPO-LoRA
RWKV-LM-RLHF-DPO-LoRA PublicForked from Triang-jyed-driung/RWKV-LM-RLHF-DPO
Direct Preference Optimization LoRA for RWKV, aiming for RWKV-5 and 6.
Python 1
-
RWKV-LM-State-4bit-Orpo
RWKV-LM-State-4bit-Orpo PublicState tuning with Orpo of RWKV v6 can be performed with 4-bit quantization. Every model can be trained with Orpo on Single 24GB GPU!
-
RWKV-Infer
RWKV-Infer PublicA large-scale RWKV v6 inference wrapper using the Cuda backend. Easy to deploy on docker. Supports multi-batch generation and dynamic State switching. Let's spread RWKV, which combines RNN technolo…
If the problem persists, check the GitHub status page or contact support.