-
Notifications
You must be signed in to change notification settings - Fork 4.6k
-
Notifications
You must be signed in to change notification settings - Fork 4.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Batch Prediction API #6654
Comments
Exalate commented: sara-tagger commented: Thanks for submitting this feature request |
Exalate commented: sahinbatmaz commented: Any updates on this issue? |
Exalate commented: dakshvar22 commented: @sahinbatmaz Thanks for suggesting this. Although, this makes sense from a technical perspective as it could speed up inference in some cases but not really for use cases Rasa Open Source focusses on. What's your exact use case for doing batch inferencing in Rasa Open Source? |
Exalate commented: youmaa commented: Any updates? For a use case, in my case, I needed this to apply an interpretability model based on prediction probabilities |
Exalate commented: akelad commented: Could you elaborate a bit? As in, is it in the context of building a bot? |
Exalate commented: youmaa commented: @akelad sorry for the late reply it's actually after building a bot, I wanna apply an interpretability model (LIME for example) and it needs to make batch prediction from a performance perspective |
Exalate commented: Ghostvv commented: It is not a research issue, I moved it to |
➤ Maxime Verger commented: 💡 Heads up! We're moving issues to Jira: https://rasa-open-source.atlassian.net/browse/OSS. From now on, this Jira board is the place where you can browse (without an account) and create issues (you'll need a free Jira account for that). This GitHub issue has already been migrated to Jira and will be closed on January 9th, 2023. Do not forget to subscribe to the corresponding Jira issue! ➡️ More information in the forum: https://forum.rasa.com/t/migration-of-rasa-oss-issues-to-jira/56569. |
Description of Problem:
I am pointing my feature request in NLU part. '/model/parse' api produces results one by one.
To shorten the response time, a batch prediction can be used. Especially, classifier layers take the most of the time and these models are capable of predicting batches of data.
Overview of the Solution:
An additional function can be added to 'Component' class as 'process_batch'. It can take a list of text as input, and in the response it produces results for each.
Or, if the input of 'process' function is not only one message but a list of messages, the process function can produce outputs.
For these things, /model/parse api should be updated or a new api should be introduced.
The text was updated successfully, but these errors were encountered: