You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
For now the token usage report is summarized per batch run.
However, it would be more helpful for users to estimate costs if promptflow could provide token usage reports for each AOAI connection deployment.
It would also be helpful if the user can read such report in json format via PF SDK.
The text was updated successfully, but these errors were encountered:
Hi @Wix97 . Thanks for reaching us, could you please elaborate more about the 'usage of each AOAI connection deployment'? Is that related to any run, for example a batch run with different deployment, to get the usage of each deployment, or just overall token usage of a deployment? For the latter one, seems that should be the capability of AzureOpenAI instead of promptflow.
Sorry for the confusion, the first case is what I wanted to mention here.
For example, there is a flow consuming 2 AOAI connections, each connection has 2 deployments. It would be helpful if promptflow can generate token usage report such like:
@Wix97 From pf perspective, it would be complicated to define such a contract to summarize tokens according to the tokens.
I have a proposal: what if we expose the ability to get the tokens in other nodes?
For example, we provide such an experession: {llm_node_1.system_metrics.usage} to get the usage dictionary of a node, then you can have a summary node to get the dictionary for yourself.
For now the token usage report is summarized per batch run.
However, it would be more helpful for users to estimate costs if promptflow could provide token usage reports for each AOAI connection deployment.
It would also be helpful if the user can read such report in json format via PF SDK.
The text was updated successfully, but these errors were encountered: