@@ -43,7 +43,7 @@ def __init__(
4343 self .dtype = dtype
4444 self .offload_mode = None
4545 self .model_names = []
46- self .models_offload_params = {}
46+ self ._models_offload_params = {}
4747
4848 @classmethod
4949 def from_pretrained (
@@ -289,10 +289,10 @@ def _enable_model_cpu_offload(self):
289289 model = getattr (self , model_name )
290290 if model is not None :
291291 model .to ("cpu" )
292- self .models_offload_params [model_name ] = {}
292+ self ._models_offload_params [model_name ] = {}
293293 for name , param in model .named_parameters (recurse = True ):
294294 param .data = param .data .pin_memory ()
295- self .models_offload_params [model_name ][name ] = param .data
295+ self ._models_offload_params [model_name ][name ] = param .data
296296 self .offload_mode = "cpu_offload"
297297
298298 def _enable_sequential_cpu_offload (self ):
@@ -327,7 +327,7 @@ def load_models_to_device(self, load_model_names: List[str] | None = None):
327327 if model_name not in load_model_names :
328328 model = getattr (self , model_name )
329329 if model is not None and (p := next (model .parameters (), None )) is not None and p .device != torch .device ("cpu" ):
330- param_cache = self .models_offload_params [model_name ]
330+ param_cache = self ._models_offload_params [model_name ]
331331 for name , param in model .named_parameters (recurse = True ):
332332 param .data = param_cache [name ]
333333 # load the needed models to device
0 commit comments