Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Proposal for Integration of Gemma Model into Jan Project #2119

Closed
jerry-f opened this issue Feb 22, 2024 · 9 comments
Closed

Proposal for Integration of Gemma Model into Jan Project #2119

jerry-f opened this issue Feb 22, 2024 · 9 comments
Assignees
Labels
Milestone

Comments

@jerry-f
Copy link

jerry-f commented Feb 22, 2024

Dear Jan Project Team,

I hope this message finds you well. I am reaching out to propose the integration of the Gemma model into the Jan project. As Gemma has been gaining traction for its exceptional capabilities in areas such as natural language processing and image generation, I believe that incorporating it could significantly enhance the Jan project, especially in terms of processing efficiency and user engagement.

@jerry-f jerry-f added the type: feature request A new feature label Feb 22, 2024
@chenshaoju
Copy link

2024-02-22T02:53:20.198Z [NITRO]::Error: llama_model_load: error loading model: unknown model architecture: 'gemma'
llama_load_model_from_file: failed to load model
llama_init_from_gpt_params: error: failed to load model 'E:\AI\jan\datafolder\models\gemma-7b.Q4\gemma-7b.Q4_K_M.gguf'

2024-02-22T02:53:20.206Z [NITRO]::Debug: Load model success with response {}
2024-02-22T02:53:20.207Z [NITRO]::Debug: {"timestamp":1708570400,"level":"ERROR","function":"load_model","line":560,"message":"unable to load model","model":"E:\\AI\\jan\\datafolder\\models\\gemma-7b.Q4\\gemma-7b.Q4_K_M.gguf"}
20240222 02:53:20.198000 UTC 8276 ERROR Error loading the model - llamaCPP.cc:565

@jerry-f
Copy link
Author

jerry-f commented Feb 22, 2024

I have tried to add the model manually, but it has not been able to start,
error "Apologies, something's amiss!

Jan's in beta. Find troubleshooting guides here or reach out to us on Discord for assistance.”

@Van-QA Van-QA assigned Van-QA and tikikun and unassigned Van-QA Feb 23, 2024
@Van-QA
Copy link
Contributor

Van-QA commented Feb 23, 2024

Gemma will be supported after this PR is merged janhq/cortex.cpp#446

@lededev
Copy link

lededev commented Feb 29, 2024

@Van-QA I think v0.4.7 is already merged your PR, are there any example of config file to add gemma-7b and gemma-7b-it?

@Van-QA Van-QA added this to the v0.4.8 milestone Mar 1, 2024
@Van-QA
Copy link
Contributor

Van-QA commented Mar 1, 2024

hi @lededev, we tested the gemma using Jan nightly build ✅ https://discord.com/channels/1107178041848909847/1209906514069028885
Please follow this guideline to try it out on our latest nightly build 🙏
https://jan.ai/guides/using-models/import-manually/

@Van-QA Van-QA self-assigned this Mar 1, 2024
@Van-QA Van-QA closed this as completed Mar 1, 2024
@chenshaoju
Copy link

chenshaoju commented Mar 1, 2024

Version 0.4.7-293 with the Gemma model is working now.

Thanks.✋

image

@lededev
Copy link

lededev commented Mar 1, 2024

Jan v0.4.7-293 also work for me.
@chenshaoju 能把你model.json贴一下吗?我用的是gemma-7b-it.gguf 34.2GB,那个模型太大了,用RTX 4090只有0.62t/s 。

@chenshaoju
Copy link

@lededev 我用的是一个第三方版本,仅供参考: https://huggingface.co/mlabonne/gemma-7b-it-GGUF

以下是 model.json :

{
  "object": "model",
  "version": 1,
  "format": "gguf",
  "sources": [
    {
      "url": "gemma-7b-it.Q4_1.gguf",
      "filename": "gemma-7b-it.Q4_1.gguf"
    }
  ],
  "id": "gemma-7b-it.Q4_1",
  "name": "gemma-7b-it.Q4_1",
  "created": 1709277804515,
  "description": "",
  "settings": {
    "ctx_len": 4096,
    "embedding": false,
    "prompt_template": "{system_message}\n### Instruction: {prompt}\n### Response:",
    "llama_model_path": "gemma-7b-it.Q4_1.gguf"
  },
  "parameters": {
    "temperature": 0.7,
    "top_p": 0.95,
    "stream": true,
    "max_tokens": 2048,
    "stop": [
      "<endofstring>"
    ],
    "frequency_penalty": 0,
    "presence_penalty": 0
  },
  "metadata": {
    "size": 5496286176,
    "author": "User",
    "tags": []
  },
  "engine": "nitro"
}

@tsjslgy
Copy link

tsjslgy commented Mar 5, 2024

So when can use gemma, 4.8 ?

@Van-QA Van-QA mentioned this issue Mar 7, 2024
5 tasks
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
Archived in project
Development

No branches or pull requests

6 participants