server : allow using LoRA adapters per-request#10994
Merged
Uh oh!
There was an error while loading. Please reload this page.
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Fix#10377
Example request POST
/completions:Example for
/v1/chat/completion:Please note that
/lora-adaptersendpoint now reflects the global value of LoRA adapter scales. Iflorais not specified per-request, we will use this global value.TODO:
SLOW_TESTS=1 ./examples/server/tests/tests.sh unit/test_lora.py -x -s -v