Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

benchmark locust tool feature request: update locust requests to match LPG requests #818

Open
annapendleton opened this issue Sep 16, 2024 · 0 comments

Comments

@annapendleton
Copy link
Collaborator

Currently Locust supports a simple format of taking in a list of prompts, output lengths are not considered in the input prompt format. This was done to make it easy to use different benchmarking data sets. LPG only works with the single dataset format, which enables LPG to send requests with various max output lengths.

LPG currently loads dataset directly into the container and at runtime will filter out any prompts with input len and output len that exceed the max.

When sending the request, LPG will input the prompt's output length as the max output length -

Locust requires these updates to match the LPG request behavior:

  1. upload the raw dataset to GCS bucket pathway in https://github.com/GoogleCloudPlatform/ai-on-gke/tree/main/benchmarks/benchmark/dataset/ShareGPT_v3_unflitered_cleaned_split
  2. in load_data.py, update the filtering to take the output_len into account. Save the prompt and output len in local dataset
  3. in tasks.py, load the prompt + output_len in load_dataset function, use the output_len in the request max_output_len field.
  4. (priority TBD) ensure continued support for simple list of prompts format (backwards compatibility to old locust request behavior) - eg. gate the above behavior behind a flag?
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant