-
Notifications
You must be signed in to change notification settings - Fork 42
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Setup at least one e2e test #77
Comments
/assign @liu-cong |
@liu-cong any status updates? |
/assign @kaushikmitr Example: https://github.com/kubernetes-sigs/jobset/blob/main/test/e2e/e2e_test.go |
@liu-cong: GitHub didn't allow me to assign the following users: kaushikmitr. Note that only kubernetes-sigs members with read permissions, repo collaborators and people who have commented on this issue/PR can be assigned. Additionally, issues/PRs can only have 10 assignees at the same time. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository. |
@kaushikmitr is picking this up |
/assign @danehans |
A Hugging Face Hub token is required to pull the model image used for e2e testing. How should this be handled when running e2e in CI, e.g. create a Hugging Face account used specifically for CI? |
If we can create one for CI, then that would be great. Perhaps we can ask on the sig-testing channel on how to handle such accounts? |
xref kubernetes/k8s.io#7698 for creating an HF account. |
@ahg-g @liu-cong regarding this issue, Mistral will not work without training Mistral-compatible LoRA weights. See this gist for details. |
Can we use a dummy LoRA? Check this in vllm: https://github.com/vllm-project/vllm/blob/28e0750847ded93158a66efdcbc869d87463b38f/vllm/lora/lora.py#L75 However, I am not sure if there is a way to configure vllm to actually set it up. |
I think we just need to find the right adapter compatible with Mistral, right? BTW, you can find compatible adapters in the HF page, like this Besides, the Qwen model doesn't require any HF login and it has adapters as well: https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct/discussions |
The model architecture also needs to be supported by vLLM. Qwen appears to be supported but I have not verified, have you? |
I didn't try but Qwen appears in the vLLM get started guide so I am pretty sure it's supported. |
To verify the end-to-end flow and CRD setup.
The text was updated successfully, but these errors were encountered: