o
    h?E                     @   s   d Z ddlmZ ddlZddlZddlmZ ddl	m
Z
 ddlmZmZmZ e r3ddlZddlmZ eeZdd	 Ze rIe rIdd
lmZ nddlmZ G dd deZG dd deZdadd Zdd Zdd Zdd Z dd Z!dddZ"dd Z#dS ) z
Integration with Deepspeed
    N)partialmethod   )dep_version_check)is_accelerate_availableis_torch_availablelogging)get_schedulerc                  C   s@   t jdd u} | rztd}W dS  tjy   Y dS w d S )N	deepspeedTF)	importlibutil	find_specimportlib_metadatametadataPackageNotFoundError)package_exists_ r   Y/var/www/html/ai/venv/lib/python3.10/site-packages/transformers/integrations/deepspeed.pyis_deepspeed_available#   s   
r   )HfDeepSpeedConfig)objectc                       s    e Zd ZdZ fddZ  ZS )r   aJ  
    This object contains a DeepSpeed configuration dictionary and can be quickly queried for things like zero stage.

    A `weakref` of this object is stored in the module's globals to be able to access the config from areas where
    things like the Trainer object is not available (e.g. `from_pretrained` and `_get_resized_embeddings`). Therefore
    it's important that this object remains alive while the program is still running.

    [`Trainer`] uses the `HfTrainerDeepSpeedConfig` subclass instead. That subclass has logic to sync the configuration
    with values of [`TrainingArguments`] by replacing special placeholder values: `"auto"`. Without this special logic
    the DeepSpeed configuration is not modified in any way.

    Args:
        config_file_or_dict (`Union[str, Dict]`): path to DeepSpeed config file or dict.

    c                    s(   t |  td td t | d S )N
accelerater	   )set_hf_deepspeed_configr   super__init__selfconfig_file_or_dict	__class__r   r   r   I   s   zHfDeepSpeedConfig.__init__)__name__
__module____qualname____doc__r   __classcell__r   r   r   r   r   8   s    r   c                       sV   e Zd ZdZ fddZdd Zdd Zdd
dZeeddZ	dd Z
dd Z  ZS )HfTrainerDeepSpeedConfigz
    The `HfTrainerDeepSpeedConfig` object is meant to be created during `TrainingArguments` object creation and has the
    same lifespan as the latter.
    c                    s   t  | d | _g | _d S N)r   r   _dtype
mismatchesr   r   r   r   r   W   s   
z!HfTrainerDeepSpeedConfig.__init__c                 C   s   | j d u r	td| j S )Nz8trainer_config_process() wasn't called yet to tell dtype)r'   
ValueErrorr   r   r   r   dtype\   s   
zHfTrainerDeepSpeedConfig.dtypec                 C   s   |  |}|d u rdS |dkS )NFauto)	get_value)r   ds_key_longvalr   r   r   is_autoa   s   
z HfTrainerDeepSpeedConfig.is_autoNTc              
   C   s   |  |\}}|du rdS ||dkr|||< dS |sdS ||}|dur?||krA| jd| d| d| d|  dS dS dS )a  
        A utility method that massages the config file and can optionally verify that the values match.

        1. Replace "auto" values with `TrainingArguments` value.

        2. If it wasn't "auto" and `must_match` is true, then check that DS config matches Trainer
        config values and if mismatched add the entry to `self.mismatched` - will assert during
        `trainer_config_finalize` for one or more mismatches.

        Nr,   z- ds =z vs hf )find_config_nodegetr(   append)r   r.   hf_valhf_key
must_matchconfigds_keyds_valr   r   r   
fill_matchh   s   
(z#HfTrainerDeepSpeedConfig.fill_matchF)r7   c                 C   s  |j |j |j }| d|jd | d|jd | d|d | d|jd | d|jd	 | d
|j|jgd | d|jd | d|j	d | 
dd | d|jd	 |js_|jri|jdkrfdnd}nd}|jr| jdi | jd< |j| jd d< | d|js|jo|dkd | d|dkd | d|jd | d|jp|jd | drtj| _dS | drtj| _dS tj| _dS ) z
        Adjust the config with `TrainingArguments` values. This stage is run during `TrainingArguments` object
        creation.
        train_micro_batch_size_per_gpuper_device_train_batch_sizegradient_accumulation_stepstrain_batch_sizeztrain_batch_size (calculated)gradient_clippingmax_grad_normzoptimizer.params.lrlearning_ratezoptimizer.params.betaszadam_beta1+adam_beta2zoptimizer.params.epsadam_epsilonzoptimizer.params.weight_decayweight_decayzscheduler.params.warmup_min_lrr   zscheduler.params.warmup_max_lrapexampN
checkpointuse_node_local_storagezfp16.enabledz%fp16|fp16_full_eval+fp16_backend(amp)zamp.enabledzfp16+fp16_backend(apex)zamp.opt_levelfp16_opt_levelzbf16.enabledzbf16|bf16_full_eval)
world_sizer=   r>   r;   rA   rB   
adam_beta1
adam_beta2rC   rD   	fill_onlyfp16fp16_full_evalfp16_backendsave_on_each_noder8   r3   rI   bf16bf16_full_evalis_truetorchbfloat16r'   is_falsefloat32float16)r   argsr?   rP   r   r   r   trainer_config_process   s@   

z/HfTrainerDeepSpeedConfig.trainer_config_processc                    s   g d} fdd|D }t |dkrQt|jdr|jj}nt|jdr+t|jj}ntd| d d	||    rQ d
d| |   dd|   	d|d  	d|
|d t  jdkrwd j}td| ddS )z
        This stage is run after we have the model and know num_training_steps.

        Now we can complete the configuration process.
        )$zero_optimization.reduce_bucket_size-zero_optimization.stage3_prefetch_bucket_size4zero_optimization.stage3_param_persistence_thresholdc                    s   g | ]	}  |r|qS r   )r0   ).0xr*   r   r   
<listcomp>   s    zDHfTrainerDeepSpeedConfig.trainer_config_finalize.<locals>.<listcomp>r   hidden_sizehidden_sizeszThe model's config file has neither `hidden_size` nor `hidden_sizes` entry, therefore it's not possible to automatically fill out the following `auto` entries in the DeepSpeed config file: zb. You can fix that by replacing `auto` values for these keys with an integer value of your choice.r\   r]   g?r^   
   z scheduler.params.total_num_stepsznum_training_steps (calculated)z!scheduler.params.warmup_num_stepswarmup_steps
z]Please correct the following DeepSpeed config values that mismatch TrainingArguments values:
zF
The easiest method is to set these DeepSpeed config values to 'auto'.N)lenhasattrr8   rb   maxrc   r)   rM   is_zero3r;   get_warmup_stepsr(   join)r   rZ   modelnum_training_stepshidden_size_based_keyshidden_size_auto_keysrb   r(   r   r*   r   trainer_config_finalize   s4   	
z0HfTrainerDeepSpeedConfig.trainer_config_finalize)NT)r    r!   r"   r#   r   r+   r0   r;   r   rM   r[   rq   r$   r   r   r   r   r%   Q   s    
:r%   c                 C   s   t | ad S r&   )weakrefref_hf_deepspeed_config_weak_ref)hf_deepspeed_config_objr   r   r   r      s   r   c                   C   s   d a d S r&   )rt   r   r   r   r   unset_hf_deepspeed_config   s   rv   c                   C   s    t d urt  d urt   S dS )NF)rt   rj   r   r   r   r   is_deepspeed_zero3_enabled   s   
rw   c                   C   s   t d urt  d urt  jS d S r&   )rt   r8   r   r   r   r   deepspeed_config  s   rx   c                    s   ddl m}m} |j}d}d|v r|jrtd||d}n| r'td 	 }d|d	< d}	d
|v r=||}	||	fS t
||rS fdd}
|||
d}	||	fS j |d}	||	fS )zY
    A convenience wrapper that deals with optimizer and lr scheduler configuration.
    r   )
DummyOptimDummySchedulerN	optimizerz|--adafactor was passed, but also found `optimizer` configured in the DeepSpeed config. Only one optimizer can be configured.)paramszDetected ZeRO Offload and non-DeepSpeed optimizers: This combination should work as long as the custom optimizer has both CPU and GPU implementation (except LAMB)Tzero_allow_untested_optimizer	schedulerc                    s   t jj| j  dS )N)r{   num_warmup_stepsrn   )r   rZ   lr_scheduler_typerk   )r{   rn   trainerr   r   _lr_scheduler_callable@  s   
z5deepspeed_optim_sched.<locals>._lr_scheduler_callable)lr_scheduler_callable)rn   r{   )accelerate.utilsry   rz   r8   	adafactorr)   
is_offloadloggerinfocreate_optimizer
isinstancecreate_scheduler)r   hf_deepspeed_configrZ   rn   model_parametersry   rz   r8   r{   lr_schedulerr   r   r   r   deepspeed_optim_sched  s2   
r   Fc           
      C   s   ddl m} | j}| j}| jjjj}|||| |	|
  |r>| s*td|d |d d\}}d}	||fS d| _ttdd	 | }	t| ||||	\}}||fS )
aj  
    Init DeepSpeed, after updating the DeepSpeed configuration with any relevant Trainer's args.

    If `resume_from_checkpoint` was passed then an attempt to resume from a previously saved checkpoint will be made.

    Args:
        trainer: Trainer object
        num_training_steps: per single gpu
        resume_from_checkpoint: path to a checkpoint if to resume from after normal DeepSpeedEngine load
        inference: launch in inference mode (no optimizer and no lr scheduler)

    Returns: optimizer, lr_scheduler

    We may use `deepspeed_init` more than once during the life of Trainer, when we do - it's a temp hack based on:
    https://github.com/microsoft/DeepSpeed/issues/1394#issuecomment-937405374 until Deepspeed fixes a bug where it
    can't resume from a checkpoint after it did some stepping https://github.com/microsoft/DeepSpeed/issues/1612

    r   )r   zMZeRO inference only makes sense with ZeRO Stage 3 - please adjust your configr{   r   )NNNc                 S   s   | j S r&   )requires_grad)pr   r   r   <lambda>{  s    z deepspeed_init.<locals>.<lambda>)deepspeed.utilsr   rm   rZ   acceleratorstatedeepspeed_pluginhf_ds_configrq   setLevelget_process_log_levelrj   r)   del_config_sub_treer{   listfilter
parametersr   )
r   rn   	inference	ds_loggerrm   rZ   r   r{   r   r   r   r   r   deepspeed_initO  s(   


r   c                 C   st   dd l }t| | d}t|dkr3td|  | j|ddd\}}|d u r1td| d S td| )Nr   z/global_step*zAttempting to resume from T)load_optimizer_statesload_lr_scheduler_statesz-[deepspeed] failed to resume from checkpoint z!Can't find a valid checkpoint at )globsortedrg   r   r   load_checkpointr)   )deepspeed_enginecheckpoint_pathr   deepspeed_checkpoint_dirs	load_pathr   r   r   r   deepspeed_load_checkpoint  s   
r   )F)$r#   importlib.metadatar   r   importlib.utilr
   rr   	functoolsr   dependency_versions_checkr   utilsr   r   r   rU   optimizationr   
get_loggerr    r   r   accelerate.utils.deepspeedr   DeepSpeedConfigbuiltinsr   r%   rt   r   rv   rw   rx   r   r   r   r   r   r   r   <module>   s4   
 
A7