Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Best Practices & ###### results problem #77

Closed
MyraBaba opened this issue Jan 2, 2024 · 2 comments
Closed

Best Practices & ###### results problem #77

MyraBaba opened this issue Jan 2, 2024 · 2 comments

Comments

@MyraBaba
Copy link

MyraBaba commented Jan 2, 2024

Hi @snexus

Thanks for your efforts and beatifull project.

1 . Would you mind to give more example for hugginaface llamacpp models ? ie what would be the best accurate result for multilingual or language different from the english ?

2 - What would be the choice for summarization document ?

3 - sometimes result adding many ##### and stopping

4 - Example config for mistral , mixtral and dolphin2

5 Do you know haystack looks commercialized version.

6 what will be the future for enterprise AI search for internal documents ? Is it worth to invest ? Can we talk ?

@snexus
Copy link
Owner

snexus commented Jan 3, 2024

Hi @MyraBaba

Would you mind to give more example for hugginaface llamacpp models ? ie what would be the best accurate result for multilingual or language different from the english ?

Unfortunately, it is hard to answer this question since it depends on the model and the task. The goal of the package is to enable users to choose the model they like without forcing any specific model, since "best" or "accurate" are subjective. A good place to look for new models or ask for the best model for the specific task would be https://www.reddit.com/r/LocalLLaMA/

What would be the choice for summarization document ?

Summarization is not in the scope of this project - it was built for question-answering.

3 - sometimes result adding many ##### and stopping

This is not coming from the package, but from the model itself - some models might require different prompt templates for asking the question. You can consult the model card e.g. Huggingface and update the prompt template in the config like specified here -

4 - Example config for mistral , mixtral and dolphin2

As long as it is supported by llamacpp, it should be similar - download the model (e.g. in gguf format), and specify the path in the config. Some models work better with different default parameters - you can configure it in the config.yaml, similar to here -

Do you know haystack looks commercialized version.

Can you clarify the question, please?

what will be the future for enterprise AI search for internal documents ? Is it worth to invest ? Can we talk ?

Quite hard to answer - some big players are entering the space, e.g. Microsoft already offers services for enterprise-grade document question answering, where models can be deployed to an organization's private network and data doesn't leave the perimeter. I think there will be a space for open-source projects like this for small organisations and privacy-conscious users, but hard to tell if it can be monetized.

@MyraBaba
Copy link
Author

MyraBaba commented Jan 3, 2024

@snexus
https://haystack.deepset.ai/ is the haystack

@snexus snexus closed this as completed Jan 6, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants