embeddings
endpoint understanding
#443
Unanswered
wei-ann-Github
asked this question in
Q&A
Replies: 1 comment 3 replies
-
I think this is probably a bug. It should return a list of two embedding representation. |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi,
I am trying to understand the output of the embeddings endpoint, and how it relates to the requests.
Deployed model: Llama-2-7b-chat
Request platform: http://localhost:3000/#operations-Service_APIs-llm-llama-service__embeddings_v1
I am making the reuqests through the BentoServer UI.
In one example, I used
as the request body. The output I got is
When my request body is just a single string in the list:
[ "Hey Jude, welcome to the jungle!" ]
The size of the embeddings is much longer than the request consisting of 2 string:
How does one interprete the embeddings and use it if it gives different size?
I was expecting the output from using 2 input strings to consists of a list of 2 sets of embeddings, but this does not seem to be the case.
Beta Was this translation helpful? Give feedback.
All reactions