@@ -116,10 +116,11 @@ func (r *ServiceReconciler) Reconcile(ctx context.Context, req ctrl.Request) (ct
116116 return ctrl.Result {}, err
117117 }
118118
119- workloadApplyConfiguration , err := buildWorkloadApplyConfiguration (service , models )
119+ workloadApplyConfiguration , err := buildWorkloadApplyConfiguration (service , models , configs )
120120 if err != nil {
121121 return ctrl.Result {}, err
122122 }
123+
123124 if err := setControllerReferenceForWorkload (service , workloadApplyConfiguration , r .Scheme ); err != nil {
124125 return ctrl.Result {}, err
125126 }
@@ -162,7 +163,7 @@ func (r *ServiceReconciler) SetupWithManager(mgr ctrl.Manager) error {
162163 Complete (r )
163164}
164165
165- func buildWorkloadApplyConfiguration (service * inferenceapi.Service , models []* coreapi.OpenModel ) (* applyconfigurationv1.LeaderWorkerSetApplyConfiguration , error ) {
166+ func buildWorkloadApplyConfiguration (service * inferenceapi.Service , models []* coreapi.OpenModel , configs * helper. GlobalConfigs ) (* applyconfigurationv1.LeaderWorkerSetApplyConfiguration , error ) {
166167 workload := applyconfigurationv1 .LeaderWorkerSet (service .Name , service .Namespace )
167168
168169 leaderWorkerTemplate := applyconfigurationv1 .LeaderWorkerTemplate ()
@@ -193,7 +194,7 @@ func buildWorkloadApplyConfiguration(service *inferenceapi.Service, models []*co
193194 leaderWorkerTemplate .WithWorkerTemplate (& podTemplateSpecApplyConfiguration )
194195
195196 // The core logic to inject additional configurations.
196- injectModelProperties (leaderWorkerTemplate , models , service )
197+ injectModelProperties (leaderWorkerTemplate , models , service , configs )
197198
198199 spec := applyconfigurationv1 .LeaderWorkerSetSpec ()
199200 spec .WithLeaderWorkerTemplate (leaderWorkerTemplate )
@@ -215,17 +216,17 @@ func buildWorkloadApplyConfiguration(service *inferenceapi.Service, models []*co
215216 return workload , nil
216217}
217218
218- func injectModelProperties (template * applyconfigurationv1.LeaderWorkerTemplateApplyConfiguration , models []* coreapi.OpenModel , service * inferenceapi.Service ) {
219+ func injectModelProperties (template * applyconfigurationv1.LeaderWorkerTemplateApplyConfiguration , models []* coreapi.OpenModel , service * inferenceapi.Service , configs * helper. GlobalConfigs ) {
219220 isMultiNodesInference := template .LeaderTemplate != nil
220221
221222 for i , model := range models {
222223 source := modelSource .NewModelSourceProvider (model )
223224 // Skip model-loader initContainer if llmaz.io/skip-model-loader annotation is set.
224225 if ! helper .SkipModelLoader (service ) {
225226 if isMultiNodesInference {
226- source .InjectModelLoader (template .LeaderTemplate , i )
227+ source .InjectModelLoader (template .LeaderTemplate , i , configs . InitContainerImage )
227228 }
228- source .InjectModelLoader (template .WorkerTemplate , i )
229+ source .InjectModelLoader (template .WorkerTemplate , i , configs . InitContainerImage )
229230 } else {
230231 if isMultiNodesInference {
231232 source .InjectModelEnvVars (template .LeaderTemplate )
0 commit comments