-
Notifications
You must be signed in to change notification settings - Fork 44
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
scheduling changes for lora affinity load balancing #423
base: main
Are you sure you want to change the base?
Conversation
[APPROVALNOTIFIER] This PR is NOT APPROVED This pull-request has been approved by: kaushikmitr The full list of commands accepted by this bot can be found here.
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
Hi @kaushikmitr. Thanks for your PR. I'm waiting for a kubernetes-sigs member to verify that this patch is reasonable to test. If it is, they should reply with Once the patch is verified, the new status will be reflected by the I understand the commands that are listed here. Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository. |
✅ Deploy Preview for gateway-api-inference-extension ready!
To edit notification comments on pull requests, go to your Netlify site configuration. |
/ok-to-test |
@kaushikmitr: The following tests failed, say
Full PR test history. Your PR dashboard. Please help us cut down on flakes by linking to an open issue when you hit one in your PR. Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository. I understand the commands that are listed here. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I didn't look at the algorithm change yet, left a couple of quick comments.
@@ -24,15 +24,23 @@ spec: | |||
- "1" | |||
- "--port" | |||
- "8000" | |||
- "--compilation-config" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
what does this do?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
we may not need this if using V0. It outputs the CUDA graph for optimization.
- "--lora-modules" | ||
- '{"name": "tweet-summary-0", "path": "vineetsharma/qlora-adapter-Llama-2-7b-hf-TweetSumm", "base_model_name": "llama-2"}' | ||
- '{"name": "tweet-summary-1", "path": "vineetsharma/qlora-adapter-Llama-2-7b-hf-TweetSumm", "base_model_name": "llama-2"}' | ||
env: | ||
- name: VLLM_USE_V1 | ||
value: "1" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The released vllm version doesn't support our metrics yet, right? if so, then we can't use it now.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes, that is why the tests are failing. I will switch back to V0
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I don't think that is, the integration test doesn't use this deployment yaml.
I think the test is failing because this PR introduces some randomness to the selection.
|
||
// Ignore metrics with both labels empty. | ||
if running == "" && waiting == "" { | ||
// continue |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
commented out code
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
this was a bug.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The algorithm is not using the waiting_lora_adapters metric, right?
name: "affinity LoRA", | ||
filter: toFilterFunc(loRAAffinityPredicate), | ||
nextOnSuccess: queueAndKVCacheFilter, | ||
nextOnFailure: &filter{ |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
was this never executed before?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
It was but we had "hard" affinity, which was optimized for throughput. This helped in lowering tail latency
@@ -37,6 +37,7 @@ import ( | |||
const ( | |||
LoraRequestInfoMetricName = "vllm:lora_requests_info" | |||
LoraRequestInfoRunningAdaptersMetricName = "running_lora_adapters" | |||
LoraRequestInfoWaitingAdaptersMetricName = "waiting_lora_adapters" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
can you please document the semantics of each metric?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We can update the protocol doc and reference the protocol doc here.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The changes here are not necessarily related to the algorithm change. Can we make this a separate PR?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think its related in the sense that the benchmark I did relies on both changes. But still can break it to two.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This is just an example manifest for startup guide.
@@ -37,6 +37,7 @@ import ( | |||
const ( | |||
LoraRequestInfoMetricName = "vllm:lora_requests_info" | |||
LoraRequestInfoRunningAdaptersMetricName = "running_lora_adapters" | |||
LoraRequestInfoWaitingAdaptersMetricName = "waiting_lora_adapters" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We can update the protocol doc and reference the protocol doc here.
@@ -37,6 +37,7 @@ import ( | |||
const ( | |||
LoraRequestInfoMetricName = "vllm:lora_requests_info" | |||
LoraRequestInfoRunningAdaptersMetricName = "running_lora_adapters" | |||
LoraRequestInfoWaitingAdaptersMetricName = "waiting_lora_adapters" | |||
LoraRequestInfoMaxAdaptersMetricName = "max_lora" | |||
// TODO: Replace these with the num_tokens_running/waiting below once we add those to the fork. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Can you clean up the TODOs and the metrics that are not currently used?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think the TODOs are still relevant. I will remove max token in KV Cache because its not being used.
// The value of 50 is arrived heuristicically based on experiments. | ||
queueingThresholdLoRA = 50 | ||
// The value of 128 is arrived heuristicically based on experiments. | ||
queueingThresholdLoRA = 128 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think we should make this configurable perhaps via a flag for now. Different environments will likely need different thresholds.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I would rather levarage this to make this configurable. #16
// Returns: | ||
// - Filtered slice of pod metrics based on affinity and availability | ||
// - Error if any issues occur during filtering | ||
func loRASoftAffinityPredicate(logger logr.Logger, req *LLMRequest, pods []*datastore.PodMetrics) ([]*datastore.PodMetrics, error) { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This is not a predicate, this is a filter, according to the current filter
and predicate
interface definition.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
agreed
// Categorize pods based on affinity and availability | ||
for _, pod := range pods { | ||
if pod == nil { | ||
continue |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
pls add a warning log here and state that this should never happen
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
removed this, as this scenario is captured already upstream
|
||
if _, exists := pod.ActiveModels[req.ResolvedTargetModel]; exists { | ||
filtered_affinity = append(filtered_affinity, pod) | ||
} else if len(pod.ActiveModels) < pod.MaxActiveModels { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This is essentially the canAcceptNewLoraPredicate
function below, are we still using canAcceptNewLoraPredicate
?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
we are not using canAcceptNewLoraPredicate any more. But would be good to keep I think.
} | ||
|
||
// Use crypto/rand for better randomization in production environments | ||
randSource := rand.NewSource(time.Now().UnixNano()) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This can be a follow up, but it sounds like we can extend the current filter framework to support such probability based filtering. So instead of having one base filter, we have a list of filters with weights. This way we can keep each filter very focused, and make them more reusable
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
agreed
queueingThresholdLoRA = 128 | ||
// TODO(https://github.com/kubernetes-sigs/gateway-api-inference-extension/issues/16) Make this configurable. | ||
// loraAffinityThreshold indicates the probability with which we prefer a pod with LoRA affinity over a pod without but having room to fit more LoRA adapters. | ||
loraAffinityThreshold = 0.999 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
do you have some insights to show why this is needed and why this value is picked?
@@ -37,6 +37,7 @@ import ( | |||
const ( | |||
LoraRequestInfoMetricName = "vllm:lora_requests_info" | |||
LoraRequestInfoRunningAdaptersMetricName = "running_lora_adapters" | |||
LoraRequestInfoWaitingAdaptersMetricName = "waiting_lora_adapters" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
On one hand, I can see why considering waiting
is useful, because waiting
loras are going to be served next. However, I have concerns of this weakening the lora affinity. running
is bound by the max lora
, waiting
is not bound. If we enter an unstable state with a long waiting
, we can lose the affinity benefit.
An improvement algorithm could be we prioritize waiting
over running
, what do you think?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
so using waiting + running for affinity is always superior to using just running. Because adapters get served in First come first serve basis. So we know for sure thar waiting if not available will get loaded for sure. But yes, within waiting + running prioritizing waiting over running makes sense I think, but need to test first.
It is, we are now checking for both waiting + running to determine affinity |
This pull request includes several changes to the deployment configuration, metrics collection, and scheduling logic. The most important changes include updating metrics collection to include waiting adapters, and implementing a new pod selection strategy that balances load while considering model affinity.
Scheduling Logic Enhancements:
pkg/epp/scheduling/filter.go
: Replaced theloRAAffinityPredicate
function with a newloRASoftAffinityPredicate
function that prioritizes pods with existing model affinity while allowing for load balancing through randomization (as long as there is room to fit another adapter in the pod).pkg/epp/scheduling/scheduler.go
: Updated the scheduling configuration to use the newloRASoftAffinityPredicate
function and increased thequeueingThresholdLoRA
value from 50 to 128. Added a newloraAffinityThreshold
constant to indicate the probability of preferring pods with model affinity. [1] [2] [3]Deployment Configuration Changes:
config/manifests/vllm/deployment.yaml
: Added new command-line arguments for--compilation-config
,--max-num-seqs
, and--max-lora-rank
. Added a new environment variableVLLM_USE_V1
. [1] [2]Metrics Collection Updates:
pkg/epp/backend/vllm/metrics.go
: Added a new metricLoraRequestInfoWaitingAdaptersMetricName
and updated thepromToPodMetrics
andgetLatestLoraMetric
functions to handle waiting adapters. Also pick the previous running + waiting adapters if there are no current running or waiting adapters [1] [2] [3]