Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support top chinese language models #116

Closed
wuxibin89 opened this issue Oct 10, 2023 · 15 comments
Closed

Support top chinese language models #116

wuxibin89 opened this issue Oct 10, 2023 · 15 comments
Assignees
Labels
enhancement New feature or request P0 High priority

Comments

@wuxibin89
Copy link
Collaborator

No description provided.

@catqaq
Copy link
Collaborator

catqaq commented Oct 13, 2023

List of expected supported models:

  • GLM
  • Baichuan2
  • Qwen
  • InternLM
  • mistral

@catqaq catqaq changed the title Support Baichuan2/Qwen chinese language model Support top chinese language models Oct 13, 2023
@catqaq
Copy link
Collaborator

catqaq commented Oct 15, 2023

Transformers officially supports Flash Attention 2, but it only covers models like Llama. Is it necessary for us to support Flash Attention 2 for other models?

@wuxibin89
Copy link
Collaborator Author

huggingface/transformers#26350 Transformers community is adding support for more architecture.

@catqaq
Copy link
Collaborator

catqaq commented Oct 19, 2023

List of expected supported models:

  • GLM
  • Baichuan2
  • Qwen
  • InternLM

Done for baichuan, e770404
but requires further performance testing, and more model testing is also welcome.

Data for test:

  • SFT: alpaca+alpaca_cn
  • RM: hh-rlhf
  • PPO: alpaca-cleaned+full-hh-rlhf+oasst1_pairwise_rlhf_reward
model SFT RM PPO Notes
Baichuan2-7B-Chat 1h 4h 71h
chatglm2-6b
Qwen-7B-Chat
internlm-chat-7b-v1_1

@catqaq catqaq self-assigned this Oct 19, 2023
@catqaq catqaq added the enhancement New feature or request label Oct 19, 2023
@catqaq
Copy link
Collaborator

catqaq commented Oct 23, 2023

List of expected supported models:

  • GLM
  • Baichuan2
  • Qwen
  • InternLM

Done, e770404 but requires further performance testing, and more model testing is also welcome.

Data for test:

  • SFT: alpaca+alpaca_cn
  • RM: hh-rlhf
  • PPO: alpaca-cleaned+full-hh-rlhf+oasst1_pairwise_rlhf_reward

model SFT RM PPO Notes
Baichuan2-7B-Chat 1h 4h 71h
chatglm2-6b
Qwen-7B-Chat
internlm-chat-7b-v1_1

more test please!

@catqaq catqaq added the P0 High priority label Oct 23, 2023
@pikaqqqqqq
Copy link
Contributor

pikaqqqqqq commented Oct 24, 2023

Qwen-14B is being supported:

  1. Different model architectures
  2. Qwen cannot add extra special tokens
  3. Qwen lacks eos_token setting, causing errors during dataset processing
  4. Qwen supports flash_attention, but the parameter passing method is different

@jovany-wang
Copy link
Contributor

What's the progress?

@Triang-jyed-driung
Copy link

Add RWKV

Would you like implementing RLHF for RWKV, here are some suggestions:

  1. As a non-transformer architecture, RWKV does not use the HF transformers library.
  2. RWKV uses a different tokenizer, called rwkv_vocab_v20230424, and tokenization process is based on their own tokenizer.
  3. The official repositories are https://github.com/BlinkDL/RWKV-LM and https://github.com/BlinkDL/ChatRWKV. See train.py for training process.
  4. RWKV contains tons of new tricks and would fail if any of these tricks is not set correctly.

@catqaq
Copy link
Collaborator

catqaq commented Oct 26, 2023

Add RWKV

Would you like implementing RLHF for RWKV, here are some suggestions:

  1. As a non-transformer architecture, RWKV does not use the HF transformers library.
  2. RWKV uses a different tokenizer, called rwkv_vocab_v20230424, and tokenization process is based on their own tokenizer.
  3. The official repositories are https://github.com/BlinkDL/RWKV-LM and https://github.com/BlinkDL/ChatRWKV. See train.py for training process.
  4. RWKV contains tons of new tricks and would fail if any of these tricks is not set correctly.

It seems that it is not easy to be compatible with RWKV in OpenRLHF.

@catqaq catqaq mentioned this issue Nov 29, 2023
@ftmtk
Copy link

ftmtk commented Dec 7, 2023

I am wondering if mistral has been supported?

@catqaq
Copy link
Collaborator

catqaq commented Dec 11, 2023

I am wondering if mistral has been supported?

Not yet

@hijkzzz
Copy link
Collaborator

hijkzzz commented Dec 27, 2023

I am wondering if mistral has been supported?

OpenRLHF supports mistral since it's model architecture is the same as llama2.

@hijkzzz hijkzzz closed this as completed Dec 27, 2023
@wxhheian
Copy link

现在qwen支持了吗

@wxhheian
Copy link

我看qwen的 特殊token <|endoftext|> 在代码里面没有体现?

@catqaq
Copy link
Collaborator

catqaq commented Feb 23, 2024

我看qwen的 特殊token <|endoftext|> 在代码里面没有体现?

应该是支持了,周末测测看~

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request P0 High priority
Projects
None yet
Development

No branches or pull requests

8 participants