llmd run_guidellm_benchmark
Runs a Guidellm benchmark job against the LLM inference service
Parameters
endpoint_url
Endpoint URL for the LLM inference service to benchmark
name
Name of the benchmark job
default value:
guidellm-benchmark
namespace
Namespace to run the benchmark job in (empty string auto-detects current namespace)
image
Container image for the benchmark
default value:
ghcr.io/vllm-project/guidellm
version
Version tag for the benchmark image
default value:
v0.6.0
timeout
Timeout in seconds to wait for job completion
default value:
900
pvc_size
Size of the PersistentVolumeClaim for storing results
default value:
1Gi
guidellm_args
List of additional guidellm arguments (e.g., [”–rate=10”, “–max-seconds=30”])
run_as_root
Run the GuideLLM container as root user