| from transformers.trainer import * |
|
|
|
|
| class MyTrainer(Trainer): |
| def _save(self, output_dir: Optional[str] = None, state_dict=None): |
| |
| output_dir = output_dir if output_dir is not None else self.args.output_dir |
| os.makedirs(output_dir, exist_ok=True) |
| logger.info(f"Saving model checkpoint to {output_dir}") |
| |
| |
| |
| |
| |
| |
| |
|
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| |
| |
| |
| |
| if not hasattr(self.model, "save"): |
| raise NotImplementedError( |
| f"MODEL {self.model.__class__.__name__} " |
| f"does not support save interface" |
| ) |
| else: |
| deepspeed = False |
| if self.deepspeed: |
| deepspeed = True |
| self.model.save(output_dir, deepspeed=deepspeed) |
| |
|
|
| |
| torch.save(self.args, os.path.join(output_dir, TRAINING_ARGS_NAME)) |
|
|
| def _save_checkpoint(self, model, trial, metrics=None): |
| |
| |
| |
|
|
| |
| checkpoint_folder = f"{PREFIX_CHECKPOINT_DIR}-{self.state.global_step}" |
|
|
| if self.hp_search_backend is None and trial is None: |
| self.store_flos() |
|
|
| run_dir = self._get_output_dir(trial=trial) |
| output_dir = os.path.join(run_dir, checkpoint_folder) |
| self.save_model(output_dir, _internal_call=True) |
| |
| |
| |
| |
| |
| |
|
|
| |
| if self.sharded_ddp == ShardedDDPOption.SIMPLE: |
| self.optimizer.consolidate_state_dict() |
|
|
| if self.fsdp: |
| |
| |
| |
| full_osd = self.model.__class__.full_optim_state_dict( |
| self.model, self.optimizer |
| ) |
|
|
| if is_torch_tpu_available(): |
| xm.rendezvous("saving_optimizer_states") |
| xm.save( |
| self.optimizer.state_dict(), os.path.join(output_dir, OPTIMIZER_NAME) |
| ) |
| with warnings.catch_warnings(record=True) as caught_warnings: |
| xm.save( |
| self.lr_scheduler.state_dict(), |
| os.path.join(output_dir, SCHEDULER_NAME), |
| ) |
| reissue_pt_warnings(caught_warnings) |
| elif is_sagemaker_mp_enabled(): |
| opt_state_dict = self.optimizer.local_state_dict(gather_if_shard=False) |
| smp.barrier() |
| if smp.rdp_rank() == 0 or smp.state.cfg.shard_optimizer_state: |
| smp.save( |
| opt_state_dict, |
| os.path.join(output_dir, OPTIMIZER_NAME), |
| partial=True, |
| v3=smp.state.cfg.shard_optimizer_state, |
| ) |
| if self.args.should_save: |
| with warnings.catch_warnings(record=True) as caught_warnings: |
| torch.save( |
| self.lr_scheduler.state_dict(), |
| os.path.join(output_dir, SCHEDULER_NAME), |
| ) |
| reissue_pt_warnings(caught_warnings) |
| if self.do_grad_scaling: |
| torch.save( |
| self.scaler.state_dict(), os.path.join(output_dir, SCALER_NAME) |
| ) |
| elif self.args.should_save and not self.is_deepspeed_enabled: |
| |
| if self.fsdp: |
| torch.save(full_osd, os.path.join(output_dir, OPTIMIZER_NAME)) |
| else: |
| torch.save( |
| self.optimizer.state_dict(), |
| os.path.join(output_dir, OPTIMIZER_NAME), |
| ) |
|
|
| with warnings.catch_warnings(record=True) as caught_warnings: |
| torch.save( |
| self.lr_scheduler.state_dict(), |
| os.path.join(output_dir, SCHEDULER_NAME), |
| ) |
| reissue_pt_warnings(caught_warnings) |
| if self.do_grad_scaling: |
| torch.save( |
| self.scaler.state_dict(), os.path.join(output_dir, SCALER_NAME) |
| ) |
|
|
| |
| if metrics is not None and self.args.metric_for_best_model is not None: |
| metric_to_check = self.args.metric_for_best_model |
| if not metric_to_check.startswith("eval_"): |
| metric_to_check = f"eval_{metric_to_check}" |
| metric_value = metrics[metric_to_check] |
|
|
| operator = np.greater if self.args.greater_is_better else np.less |
| if ( |
| self.state.best_metric is None |
| or self.state.best_model_checkpoint is None |
| or operator(metric_value, self.state.best_metric) |
| ): |
| self.state.best_metric = metric_value |
| self.state.best_model_checkpoint = output_dir |
|
|
| |
| if self.args.should_save: |
| self.state.save_to_json(os.path.join(output_dir, TRAINER_STATE_NAME)) |
|
|
| |
| rng_states = { |
| "python": random.getstate(), |
| "numpy": np.random.get_state(), |
| "cpu": torch.random.get_rng_state(), |
| } |
| if torch.cuda.is_available(): |
| if self.args.parallel_mode == ParallelMode.DISTRIBUTED: |
| |
| rng_states["cuda"] = torch.cuda.random.get_rng_state_all() |
| else: |
| rng_states["cuda"] = torch.cuda.random.get_rng_state() |
|
|
| if is_torch_tpu_available(): |
| rng_states["xla"] = xm.get_rng_state() |
|
|
| |
| |
| os.makedirs(output_dir, exist_ok=True) |
|
|
| if self.args.world_size <= 1: |
| torch.save(rng_states, os.path.join(output_dir, "rng_state.pth")) |
| else: |
| torch.save( |
| rng_states, |
| os.path.join(output_dir, f"rng_state_{self.args.process_index}.pth"), |
| ) |
|
|
| if self.args.push_to_hub: |
| self._push_from_checkpoint(output_dir) |
|
|
| |
| if self.args.should_save: |
| self._rotate_checkpoints(use_mtime=True, output_dir=run_dir) |