Running against local model service #1688
Labels
additions
New models or scenarios
competition
Support for the NeurIPS Large Language Model Efficiency Challenge
enhancement
New feature or request
models
p2
Priority 2 (Good to have for release)
For the neurips competetion we woud like to be able to run Helm against a model running local to the machine. I.e run helm in a container and have it send requests to another container. I started implementing a new local http client following the "adding a model" readme.
However I see that there already appears to be something pretty similiar here:
helm/src/helm/proxy/services/server_service.py
Line 38 in 05ce7e8
Do you think this would fit our needs? How exactly does one use this with helm-run?
I would imagine it to be something like:
The text was updated successfully, but these errors were encountered: