-
Notifications
You must be signed in to change notification settings - Fork 82
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add base test for vLLM and its metrics #1438
Add base test for vLLM and its metrics #1438
Conversation
Signed-off-by: Luca Giorgi <lgiorgi@redhat.com>
namespace: vllm-gpt2 | ||
labels: | ||
gpt-download-pod: 'true' | ||
spec: |
Check warning
Code scanning / SonarCloud
Service account tokens should not be mounted in pods
- mountPath: "/mnt/models/" | ||
name: model-volume | ||
containers: | ||
- name: download-model |
Check warning
Code scanning / SonarCloud
Storage limits should be enforced
.../Tests/400__ods_dashboard/420__model_serving/LLMs/vllm/426__model_serving_vllm_metrics.robot
Fixed
Show fixed
Hide fixed
.../Tests/400__ods_dashboard/420__model_serving/LLMs/vllm/426__model_serving_vllm_metrics.robot
Dismissed
Show dismissed
Hide dismissed
.../Tests/400__ods_dashboard/420__model_serving/LLMs/vllm/426__model_serving_vllm_metrics.robot
Fixed
Show fixed
Hide fixed
.../Tests/400__ods_dashboard/420__model_serving/LLMs/vllm/426__model_serving_vllm_metrics.robot
Fixed
Show fixed
Hide fixed
.../Tests/400__ods_dashboard/420__model_serving/LLMs/vllm/426__model_serving_vllm_metrics.robot
Dismissed
Show dismissed
Hide dismissed
.../Tests/400__ods_dashboard/420__model_serving/LLMs/vllm/426__model_serving_vllm_metrics.robot
Fixed
Show fixed
Hide fixed
.../Tests/400__ods_dashboard/420__model_serving/LLMs/vllm/426__model_serving_vllm_metrics.robot
Fixed
Show fixed
Hide fixed
Robot Results
|
.../Tests/400__ods_dashboard/420__model_serving/LLMs/vllm/426__model_serving_vllm_metrics.robot
Outdated
Show resolved
Hide resolved
.../Tests/400__ods_dashboard/420__model_serving/LLMs/vllm/426__model_serving_vllm_metrics.robot
Show resolved
Hide resolved
Co-authored-by: Vedant Mahabaleshwarkar <vmahabal@redhat.com>
Signed-off-by: Luca Giorgi <lgiorgi@redhat.com>
.../Tests/400__ods_dashboard/420__model_serving/LLMs/vllm/426__model_serving_vllm_metrics.robot
Fixed
Show fixed
Hide fixed
.../Tests/400__ods_dashboard/420__model_serving/LLMs/vllm/426__model_serving_vllm_metrics.robot
Fixed
Show fixed
Hide fixed
.../Tests/400__ods_dashboard/420__model_serving/LLMs/vllm/426__model_serving_vllm_metrics.robot
Fixed
Show fixed
Hide fixed
.../Tests/400__ods_dashboard/420__model_serving/LLMs/vllm/426__model_serving_vllm_metrics.robot
Fixed
Show fixed
Hide fixed
Verify Vllm Metrics Are Present | ||
[Documentation] Confirm vLLM metrics are exposed in OpenShift metrics | ||
[Tags] Tier1 Sanity Resources-GPU ODS-XXX | ||
${host} = llm.Get KServe Inference Host Via CLI isvc_name=vllm-gpt2-openai namespace=${TEST_NS} |
Check warning
Code scanning / Robocop
The assignment sign is not consistent within the file. Expected '{{ expected_sign }}' but got '{{ actual_sign }}' instead Warning test
IF ${is_self_managed} | ||
Configure User Workload Monitoring | ||
Enable User Workload Monitoring | ||
#TODO: Find reliable signal for UWM being ready |
Check warning
Code scanning / Robocop
Missing blank space after comment character Warning test
Configure User Workload Monitoring | ||
Enable User Workload Monitoring | ||
#TODO: Find reliable signal for UWM being ready | ||
#Sleep 10m |
Check warning
Code scanning / Robocop
Missing blank space after comment character Warning test
END | ||
Load Expected Responses | ||
|
||
Suite Teardown |
Check warning
Code scanning / Robocop
Missing documentation in '{{ name }}' keyword Warning test
Signed-off-by: Luca Giorgi <lgiorgi@redhat.com>
Signed-off-by: Luca Giorgi <lgiorgi@redhat.com>
.../Tests/400__ods_dashboard/420__model_serving/LLMs/vllm/426__model_serving_vllm_metrics.robot
Outdated
Show resolved
Hide resolved
.../Tests/400__ods_dashboard/420__model_serving/LLMs/vllm/426__model_serving_vllm_metrics.robot
Show resolved
Hide resolved
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Approving, but we could re-use some keywords and file we already have (e.g., download model and inference serving runtime). I'd suggest raising a new PR to improve it besides considering adding support for Kserve raw deployment
Signed-off-by: Luca Giorgi <lgiorgi@redhat.com>
${rc} ${out}= Run And Return Rc And Output oc apply -f ${DL_POD_FILEPATH} | ||
Should Be Equal As Integers ${rc} ${0} | ||
Wait For Pods To Succeed label_selector=gpt-download-pod=true namespace=${TEST_NS} | ||
${rc} ${out}= Run And Return Rc And Output oc apply -f ${SR_FILEPATH} |
Check notice
Code scanning / Robocop
Variable '{{ name }}' is assigned but not used Note test
Wait For Pods To Succeed label_selector=gpt-download-pod=true namespace=${TEST_NS} | ||
${rc} ${out}= Run And Return Rc And Output oc apply -f ${SR_FILEPATH} | ||
Should Be Equal As Integers ${rc} ${0} | ||
#TODO: Switch to common keyword for model DL and SR deploy |
Check warning
Code scanning / Robocop
Missing blank space after comment character Warning test
${rc} ${out}= Run And Return Rc And Output oc apply -f ${SR_FILEPATH} | ||
Should Be Equal As Integers ${rc} ${0} | ||
#TODO: Switch to common keyword for model DL and SR deploy | ||
#Set Project And Runtime runtime=vllm namespace=${TEST_NS} |
Check warning
Code scanning / Robocop
Missing blank space after comment character Warning test
Should Be Equal As Integers ${rc} ${0} | ||
#TODO: Switch to common keyword for model DL and SR deploy | ||
#Set Project And Runtime runtime=vllm namespace=${TEST_NS} | ||
#... download_in_pvc=${DOWNLOAD_IN_PVC} model_name=gpt2 |
Check warning
Code scanning / Robocop
Missing blank space after comment character Warning test
#TODO: Switch to common keyword for model DL and SR deploy | ||
#Set Project And Runtime runtime=vllm namespace=${TEST_NS} | ||
#... download_in_pvc=${DOWNLOAD_IN_PVC} model_name=gpt2 | ||
#... storage_size=10Gi |
Check warning
Code scanning / Robocop
Missing blank space after comment character Warning test
Verify Vllm Metrics Are Present | ||
[Documentation] Confirm vLLM metrics are exposed in OpenShift metrics | ||
[Tags] Tier1 Sanity Resources-GPU RHOAIENG-6264 | ||
... Depends On Test Verify User Can Deploy A Model With Vllm Via CLI |
Check warning
Code scanning / Robocop
Tag '{{ tag }}' should not contain spaces Warning test
Verify Vllm Metrics Are Present | ||
[Documentation] Confirm vLLM metrics are exposed in OpenShift metrics | ||
[Tags] Tier1 Sanity Resources-GPU RHOAIENG-6264 | ||
... Depends On Test Verify User Can Deploy A Model With Vllm Via CLI |
Check warning
Code scanning / Robocop
Tag '{{ tag }}' should not contain spaces Warning test
Quality Gate passedIssues Measures |
* Add base test for vLLM and its metrics Signed-off-by: Luca Giorgi <lgiorgi@redhat.com> * Update ods_ci/tests/Resources/Files/llm/vllm/vllm_servingruntime.yaml Co-authored-by: Vedant Mahabaleshwarkar <vmahabal@redhat.com> * reimplement using common keywords Signed-off-by: Luca Giorgi <lgiorgi@redhat.com> * Better handle missing metrics from UWM, change expected response format Signed-off-by: Luca Giorgi <lgiorgi@redhat.com> * small cleanup Signed-off-by: Luca Giorgi <lgiorgi@redhat.com> * Move keyword, some cleanup, comments Signed-off-by: Luca Giorgi <lgiorgi@redhat.com> --------- Signed-off-by: Luca Giorgi <lgiorgi@redhat.com> Co-authored-by: Vedant Mahabaleshwarkar <vmahabal@redhat.com>
No description provided.