Open Issues Need Help
View All on GitHub Add vllm CPU image for test workflows about 2 months ago
good first issue
Add quick start script for vLLM Deployment 4 months ago
AI Summary: Create a new shell script, `hack/deploy-vllm.sh`, that deploys a vLLM inference server to Kubernetes. The script should allow configuration of the vLLM image, include a health check to ensure the deployment is ready, and accept a parameter for specifying the model to use. This will mirror the existing `hack/deploy-ollama.sh` script, providing a streamlined deployment experience for vLLM users.
Complexity:
3/5
good first issue
Use cli.Path over naive cli.Update 4 months ago
AI Summary: The task involves refactoring the Llama Stack Kubernetes operator to use `cli.Path` instead of `cli.Update` for applying Kubernetes resources, improving code consistency and potentially performance. This is part of a larger effort to standardize resource application methods within the operator.
Complexity:
3/5
good first issue