You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Describe the bug
Token estimation is wrong, as it doesn't take into account the functions/tools definition as part of the prompt.
In cases with heavy use of tools this can make a big diference and make you take a wrong decision regarding the token limit of a request.
How To Reproduce the bug
Steps to reproduce the behavior, how frequent can you experience the bug:
Create a prompty file with a large tools definition
Make an estimation with a small question
compare with the actual token consumption reported by oai response
Expected behavior
A clear and concise description of what you expected to happen.
Add an additional function to also estimate the tools definition token consumption
Screenshots
If applicable, add screenshots to help explain your problem.
Running Information(please complete the following information):
Promptflow Package Version using pf -v: [e.g. 0.0.102309906]
Operating System: [e.g. Ubuntu 20.04, Windows 11]: Debian
Python Version using python --version: [e.g. python==3.10.12]: 3.11.5
Additional context
Add any other context about the problem here.
The text was updated successfully, but these errors were encountered:
Additionally token from messages is also incorrect.
It doesn't take into account when an assistant's message has tool_calls
And when using Azure OpenAI the wrong model type may be selected. num_tokens_from_messages expects the model as string using one of OpenAI model's labels, but in this case the azure_deployment is used. This is a user generated label and not necessarily a valid model name, so the token count raises error and can´t be used.
Describe the bug
Token estimation is wrong, as it doesn't take into account the functions/tools definition as part of the prompt.
In cases with heavy use of tools this can make a big diference and make you take a wrong decision regarding the token limit of a request.
How To Reproduce the bug
Steps to reproduce the behavior, how frequent can you experience the bug:
Expected behavior
A clear and concise description of what you expected to happen.
Screenshots
If applicable, add screenshots to help explain your problem.
Running Information(please complete the following information):
pf -v
: [e.g. 0.0.102309906]python --version
: [e.g. python==3.10.12]: 3.11.5Additional context
Add any other context about the problem here.
The text was updated successfully, but these errors were encountered: