Open Issues Need Help
View All on GitHub good first issue
A light weight vLLM simulator, for mocking out replicas.
Go
#incubating
Add retries to connect to ZMQ 2 months ago
good first issue
A light weight vLLM simulator, for mocking out replicas.
Go
#incubating
Support --max-model-len config parameter 3 months ago
AI Summary: Implement a new command-line parameter, `--max-model-len`, in the vLLM simulator. This parameter will define the maximum context window size (in tokens) for the model. Requests exceeding this limit should return a 400 Bad Request error with a specific error message indicating the context length exceeded.
Complexity:
4/5
enhancement good first issue
A light weight vLLM simulator, for mocking out replicas.
Go
#incubating