repo stringlengths 1 99 | file stringlengths 13 215 | code stringlengths 12 59.2M | file_length int64 12 59.2M | avg_line_length float64 3.82 1.48M | max_line_length int64 12 2.51M | extension_type stringclasses 1
value |
|---|---|---|---|---|---|---|
custom-diffusion | custom-diffusion-main/src/diffusers_training.py | # This code is built from the Huggingface repository: https://github.com/huggingface/diffusers/blob/main/examples/dreambooth/train_dreambooth.py, and
# https://github.com/huggingface/diffusers/blob/main/examples/textual_inversion/textual_inversion.py
# Copyright 2022- The Hugging Face team. All rights reserved.
# ... | 51,778 | 45.816456 | 253 | py |
custom-diffusion | custom-diffusion-main/src/compress.py | # Copyright 2022 Adobe Research. All rights reserved.
# To view a copy of the license, visit LICENSE.md.
import torch
import argparse
def compress(delta_ckpt, ckpt, diffuser=False, compression_ratio=0.6, device='cuda'):
st = torch.load(f'{delta_ckpt}')
if not diffuser:
compressed_key = 'state_dict'
... | 2,479 | 34.428571 | 107 | py |
custom-diffusion | custom-diffusion-main/src/diffusers_data_pipeline.py | # This code is built from the Huggingface repository: https://github.com/huggingface/diffusers/blob/main/examples/dreambooth/train_dreambooth.py, and
# https://github.com/huggingface/diffusers/blob/main/examples/textual_inversion/textual_inversion.py
# Copyright 2022- The Hugging Face team. All rights reserved.
# ... | 20,730 | 50.061576 | 161 | py |
custom-diffusion | custom-diffusion-main/src/diffusers_composenW.py | # Copyright 2022 Adobe Research. All rights reserved.
# To view a copy of the license, visit LICENSE.md.
import sys
import os
import argparse
import torch
from scipy.linalg import lu_factor, lu_solve
sys.path.append('./')
from diffusers import StableDiffusionPipeline
from src import diffusers_sample
def gdupdateWe... | 7,508 | 37.706186 | 172 | py |
custom-diffusion | custom-diffusion-main/src/model.py | # This code is built from the Stable Diffusion repository: https://github.com/CompVis/stable-diffusion.
# Copyright (c) 2022 Robin Rombach and Patrick Esser and contributors.
# CreativeML Open RAIL-M
#
# ==========================================================================================
#
# Adobe’s modifications... | 29,720 | 70.102871 | 1,097 | py |
custom-diffusion | custom-diffusion-main/src/custom_modules.py | # This code is built from the Huggingface repository: https://github.com/huggingface/transformers/tree/main/src/transformers/models/clip.
# Copyright 2018- The Hugging Face team. All rights reserved.
# Apache License
# Version 2.0, January 2004
# ... | 16,723 | 50.937888 | 137 | py |
custom-diffusion | custom-diffusion-main/src/finetune_data.py | # This code is built from the Stable Diffusion repository: https://github.com/CompVis/stable-diffusion.
# Copyright (c) 2022 Robin Rombach and Patrick Esser and contributors.
# CreativeML Open RAIL-M
#
# ==========================================================================================
#
# Adobe’s modifications... | 22,159 | 81.686567 | 1,097 | py |
custom-diffusion | custom-diffusion-main/src/get_deltas.py | # Copyright 2022 Adobe Research. All rights reserved.
# To view a copy of the license, visit LICENSE.md.
import os
import argparse
import glob
import torch
def main(path, newtoken=0):
layers = []
for files in glob.glob(f'{path}/checkpoints/*'):
if ('=' in files or '_' in files) and 'delta' not in file... | 2,047 | 36.925926 | 149 | py |
custom-diffusion | custom-diffusion-main/src/composenW.py | # Copyright 2022 Adobe Research. All rights reserved.
# To view a copy of the license, visit LICENSE.md.
import sys
import os
import argparse
import random
import torch
import torchvision
import numpy as np
from tqdm import tqdm
from scipy.linalg import lu_factor, lu_solve
sys.path.append('stable-diffusion')
sys.pa... | 11,722 | 38.738983 | 130 | py |
custom-diffusion | custom-diffusion-main/src/convert.py | # Copyright 2022 Adobe Research. All rights reserved.
# To view a copy of the license, visit LICENSE.md.
import os, sys
import argparse
import torch
from omegaconf import OmegaConf
from ldm.util import instantiate_from_config
sys.path.append('stable-diffusion')
sys.path.append('./')
from src.diffusers_model_pipeline i... | 7,914 | 47.858025 | 161 | py |
custom-diffusion | custom-diffusion-main/src/diffusers_sample.py | # ==========================================================================================
#
# MIT License. To view a copy of the license, visit MIT_LICENSE.md.
#
# ==========================================================================================
import argparse
import sys
import os
import numpy as np
impor... | 3,037 | 39.506667 | 122 | py |
custom-diffusion | custom-diffusion-main/customconcept101/evaluate.py | import argparse
import glob
import json
import os
import warnings
from pathlib import Path
import clip
import numpy as np
import pandas as pd
import sklearn.preprocessing
import torch
from packaging import version
from PIL import Image
from torchvision.transforms import CenterCrop, Compose, Normalize, Resize, ToTensor... | 10,450 | 33.953177 | 122 | py |
AnalyzeParameterEfficientFinetune | AnalyzeParameterEfficientFinetune-main/src/zlog.py | import os
import time
import torch
import traceback
from contextlib import contextmanager
from tensorboardX import SummaryWriter # maple
import jiant.utils.python.io as py_io
import jiant.utils.python.filesystem as filesystem
class BaseZLogger:
def log_context(self):
raise NotImplementedError()
def... | 7,042 | 27.864754 | 112 | py |
AnalyzeParameterEfficientFinetune | AnalyzeParameterEfficientFinetune-main/src/runscript.py | import os
import torch
from transformers import AutoConfig
import jiant.proj.main.write_task_configs as write_task_configs
import jiant.proj.main.export_model as export_model
#import jiant.proj.main.tokenize_and_cache as tokenize_and_cache
import tokenize_and_cache # maple
import jiant.proj.main.scripts.configurator... | 13,469 | 43.163934 | 201 | py |
AnalyzeParameterEfficientFinetune | AnalyzeParameterEfficientFinetune-main/src/evaluate.py | import json
import os
import torch
import jiant.utils.python.io as py_io
import jiant.proj.main.components.task_sampler as jiant_task_sampler
from jiant.proj.main.components.evaluate import *
def write_val_results(val_results_dict, metrics_aggregator, output_dir, verbose=True, result_file = "val_metrics.json"):
... | 1,180 | 38.366667 | 120 | py |
AnalyzeParameterEfficientFinetune | AnalyzeParameterEfficientFinetune-main/src/runner.py | from http.client import NotConnected
from typing import Dict
from dataclasses import dataclass
import torch
import math
import numpy as np
import copy
#from functorch import *
from torch.autograd.functional import *
import jiant.tasks.evaluate as evaluate
import jiant.utils.torch_utils as torch_utils
#from jiant.pro... | 45,479 | 47.177966 | 202 | py |
AnalyzeParameterEfficientFinetune | AnalyzeParameterEfficientFinetune-main/src/model_setup.py | import transformers
import torch
from jiant.ext.radam import RAdam
class OptimizerScheduler:
def __init__(self, optimizer, scheduler):
super().__init__()
self.optimizer = optimizer
self.scheduler = scheduler
def step(self):
self.optimizer.step()
self.scheduler.step()
... | 6,050 | 30.515625 | 112 | py |
AnalyzeParameterEfficientFinetune | AnalyzeParameterEfficientFinetune-main/src/main_runscript.py | import os
import torch
import datetime
import jiant.proj.main.modeling.model_setup as jiant_model_setup
import runner as jiant_runner
#import jiant.proj.main.components.container_setup as container_setup
#import jiant.proj.main.metarunner as jiant_metarunner
import metarunner as jiant_metarunner
#import jiant.proj.ma... | 15,719 | 41.95082 | 200 | py |
DeepSpeed | DeepSpeed-master/setup.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
"""
DeepSpeed library
To build wheel on Windows:
1. Install pytorch, such as pytorch 1.12 + cuda 11.6.
2. Install visual cpp build tool.
3. Include cuda toolkit.
4. Launch cmd console with Administrator privilege for creatin... | 11,499 | 36.216828 | 139 | py |
DeepSpeed | DeepSpeed-master/deepspeed/env_report.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
import deepspeed
import subprocess
import argparse
from .ops.op_builder.all_ops import ALL_OPS
from .git_version_info import installed_ops, torch_info
from deepspeed.accelerator import get_accelerator
GREEN = '... | 4,804 | 37.134921 | 136 | py |
DeepSpeed | DeepSpeed-master/deepspeed/git_version_info.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
try:
# This is populated by setup.py
from .git_version_info_installed import * # noqa: F401
except ModuleNotFoundError:
import os
if os.path.isfile('version.txt'):
# Will be missing from checkouts t... | 756 | 31.913043 | 88 | py |
DeepSpeed | DeepSpeed-master/deepspeed/__init__.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import sys
import types
import json
from typing import Optional, Union
import torch
from torch.optim import Optimizer
from torch.optim.lr_scheduler import _LRScheduler
from packaging import version as pkg_version
try:
i... | 14,510 | 41.06087 | 157 | py |
DeepSpeed | DeepSpeed-master/deepspeed/checkpoint/zero_checkpoint.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
from .constants import (BASE_OPTIMIZER_STATE, GROUP_PADDINGS, OPTIMIZER_STATE_DICT, PARTITION_COUNT)
from .reshape_utils import (basic_folder_validation, get_zero_files, merge_state)
from .reshape_3d_utils im... | 5,316 | 36.70922 | 110 | py |
DeepSpeed | DeepSpeed-master/deepspeed/checkpoint/utils.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import os
import torch
from .constants import (MODEL_FILE_PREFIX, MODEL_FILE_SUFFIX, OPTIM_FILE_SUFFIX, ZERO_FILE_PREFIX)
def get_model_ckpt_name_for_rank(base_folder, mp_rank_str):
ckpt_name = os.path.join(
ba... | 2,534 | 39.238095 | 111 | py |
DeepSpeed | DeepSpeed-master/deepspeed/checkpoint/reshape_utils.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import os
import torch
from collections import OrderedDict
from .constants import (ZERO_FILE_PREFIX, FP16_ZERO_FILE_PREFIX, BF16_ZERO_FILE_PREFIX)
def basic_folder_validation(dir):
assert os.path.exists(dir), f'{dir} p... | 2,888 | 28.783505 | 105 | py |
DeepSpeed | DeepSpeed-master/deepspeed/checkpoint/universal_checkpoint.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import os
import torch
import types
from .constants import (FP32_WEIGHT_KEY, PARAM, VOCAB_DIVISIBILITY_PADDING_TENSOR, CAT_DIM)
def load_hp_checkpoint_state(self, folder, tp_rank, tp_world_size):
hp_mapping = self._hp_... | 4,888 | 49.927083 | 141 | py |
DeepSpeed | DeepSpeed-master/deepspeed/checkpoint/deepspeed_checkpoint.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import os
from typing import Dict
import torch
from .reshape_3d_utils import model_3d_desc
from .reshape_utils import (basic_folder_validation, merge_state, partition_data, get_files, get_files_with_prefix)
from .constants... | 12,012 | 41.448763 | 119 | py |
DeepSpeed | DeepSpeed-master/deepspeed/profiling/flops_profiler/profiler.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import time
import torch
import torch.nn as nn
import torch.nn.functional as F
from functools import partial
from typing import List, Optional
from collections import OrderedDict
import numpy as np
from deepspeed.accelerator... | 51,583 | 39.26854 | 650 | py |
DeepSpeed | DeepSpeed-master/deepspeed/compression/basic_layer.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
import math
from torch import nn
from torch.nn import init
import deepspeed.comm as dist
from .utils import TopKBinarizer, SymQuantizer, AsymQuantizer, TernaryQuantizer, BinaryQuantizer
from deepspeed.utils impo... | 36,033 | 41.846611 | 169 | py |
DeepSpeed | DeepSpeed-master/deepspeed/compression/compress.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import re
from .helper import compression_preparation, fix_compression, recursive_getattr, is_module_compressible
from .config import get_compression_config
from ..runtime.config_utils import dict_raise_error_on_duplicate_ke... | 11,886 | 48.529167 | 167 | py |
DeepSpeed | DeepSpeed-master/deepspeed/compression/utils.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
from torch import autograd
import math
class TopKBinarizer(autograd.Function):
"""
Top-k Binarizer.
Computes a binary mask M from a real value matrix S such that `M_{i,j} = 1` if and only if `S_{i,... | 7,818 | 34.06278 | 108 | py |
DeepSpeed | DeepSpeed-master/deepspeed/compression/helper.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
from .basic_layer import Embedding_Compress, LinearLayer_Compress, Conv2dLayer_Compress, BNLayer_Compress, ColumnParallelLinear_Compress, RowParallelLinear_Compress
from .constants import *
from deepspeed.utils ... | 14,637 | 44.318885 | 165 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/lr_schedules.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
"""
Implementation of learning rate schedules.
Taken and modified from PyTorch v1.0.1 source
https://github.com/pytorch/pytorch/blob/v1.1.0/torch/optim/lr_scheduler.py
"""
import argparse
from torch.optim import Optimizer
i... | 33,541 | 42.903141 | 164 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/bf16_optimizer.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
from collections import OrderedDict
import torch
import sys
import os
from torch._utils import _flatten_dense_tensors, _unflatten_dense_tensors
from deepspeed import comm as dist
from deepspeed.runtime.constants import PIPE_... | 18,534 | 42.611765 | 118 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/engine.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import os
import re
import stat
import torch
import hashlib
from collections import defaultdict, OrderedDict, deque
from shutil import copyfile
import gc
from torch.nn.modules import Module
from torch.nn.parameter import Pa... | 157,235 | 45.12379 | 506 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/weight_quantizer.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
from ..module_inject.replace_policy import HFBertLayerPolicy, replace_policies
from deepspeed.accelerator import get_accelerator
class WeightQuantization(object):
def __init__(self, mlp_extra_grouping=Tru... | 7,027 | 44.636364 | 119 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/dataloader.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
from torch.utils.data import DataLoader, RandomSampler
from torch.utils.data.distributed import DistributedSampler
from deepspeed.accelerator import get_accelerator
from deepspeed.runtime.data_pipeline.data_sampling.data_sa... | 6,977 | 41.809816 | 110 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/eigenvalue.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
from deepspeed.utils import log_dist
import numpy as np
import logging
class Eigenvalue(object):
def __init__(self,
verbose=False,
max_iter=100,
tol=1e-2... | 5,625 | 36.506667 | 211 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/utils.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
"""
Copyright NVIDIA/Megatron
Helper functions and classes from multiple sources.
"""
from collections.abc import Iterable
from deepspeed.moe.utils import is_moe_param
import os
import psutil
import gc
from math import sqrt... | 35,805 | 35.686475 | 116 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/state_dict_factory.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
import os
import copy
import collections
import json
from abc import ABC, abstractmethod
from deepspeed.utils import logger
from deepspeed.runtime.checkpoint_engine.torch_checkpoint_engine import TorchCheckpoin... | 18,177 | 41.471963 | 153 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/config.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import os
from typing import Union
from enum import Enum
import torch
import json
import hjson
import copy
import base64
from .constants import *
from .fp16.loss_scaler import (
INITIAL_LOSS_SCALE,
SCALE_WINDOW,
... | 39,863 | 38.705179 | 174 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/sparse_tensor.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
"""
Implementation of a compressed sparse tensor. Similar in
functionality to TensorFlow's IndexedSlices implementation.
"""
import torch
class SparseTensor(object):
""" Compressed Sparse Tensor """
def __init__(s... | 2,416 | 34.028986 | 96 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/quantize.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
import math
from deepspeed.utils import logger
from deepspeed.ops.quantizer import ds_quantizer
TWO_D_PARAMS = 6
class Quantizer(object):
def __init__(self,
q_groups=1,
... | 7,699 | 41.541436 | 171 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/hybrid_engine.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
from deepspeed.inference.config import DeepSpeedInferenceConfig
from deepspeed.module_inject.replace_policy import replace_policies
from deepspeed.module_inject.utils import policy_to_ds_container
from .engine i... | 20,369 | 45.190476 | 126 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/activation_checkpointing/checkpointing.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
"""
Use to partition the activations stored for backward propagation
Therefore reduces the memory consumption
Also implements CPU checkpointing and contiguous memory checkpointing
Reduces memory consumption and memory fragmen... | 32,764 | 36.148526 | 168 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/fp16/fused_optimizer.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
"""
Copyright NVIDIA/apex
This file is adapted from FP16_Optimizer in NVIDIA/apex
"""
import torch
from torch._utils import _flatten_dense_tensors, _unflatten_dense_tensors
from deepspeed.runtime import DeepSpeedOptimizer
f... | 20,052 | 40.346392 | 126 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/fp16/loss_scaler.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
"""
Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of t... | 11,492 | 41.409594 | 325 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/fp16/unfused_optimizer.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
"""
Copyright NVIDIA/apex
This file is adapted from FP16_Optimizer in NVIDIA/apex
"""
from deepspeed.moe.utils import split_params_grads_into_shared_and_expert_params
import torch
from torch._utils import _flatten_dense_tens... | 18,076 | 41.039535 | 126 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/fp16/onebit/adam.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import types
import torch
import numpy as np
from deepspeed.accelerator import get_accelerator
from deepspeed import comm as dist
class OnebitAdam(torch.optim.Optimizer):
"""Implements the 1-bit Adam algorithm. Current... | 15,258 | 48.381877 | 196 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/fp16/onebit/zoadam.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import types
import torch
import numpy as np
from deepspeed.accelerator import get_accelerator
from deepspeed import comm as dist
class ZeroOneAdam(torch.optim.Optimizer):
"""Implements the 0/1 Adam algorithm. Currentl... | 19,112 | 51.798343 | 194 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/fp16/onebit/lamb.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import types
import torch
import numpy as np
from deepspeed import comm as dist
from torch._utils import _flatten_dense_tensors, _unflatten_dense_tensors
from deepspeed.accelerator import get_accelerator
class OnebitLamb(t... | 23,085 | 50.762332 | 196 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/swap_tensor/optimizer_utils.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
"""
Functionality of swapping tensors to/from (NVMe) storage devices.
"""
import os
import torch
from deepspeed import comm as dist
from deepspeed.utils.logging import logger
from deepspeed.runtime.swap_tensor.constants imp... | 18,967 | 39.271762 | 206 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/swap_tensor/utils.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
"""
Functionality of swapping tensors to/from (NVMe) storage devices.
"""
import torch
from deepspeed.utils.logging import logger
from deepspeed.accelerator import get_accelerator
from deepspeed import comm as dist
MIN_AIO... | 7,734 | 31.229167 | 117 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/swap_tensor/async_swapper.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
"""
Functionality of swapping tensors to/from (NVMe) storage devices.
"""
import torch
from deepspeed import comm as dist
from deepspeed.utils.logging import logger
from deepspeed.runtime.swap_tensor.utils import swap_out_te... | 6,282 | 34.902857 | 100 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/swap_tensor/partitioned_param_swapper.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
"""
Functionality of swapping tensors to/from (NVMe) storage devices.
"""
import os
import shutil
from enum import Enum
import torch
from deepspeed import comm as dist
from deepspeed.accelerator import get_accelerator
from d... | 17,684 | 42.774752 | 165 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/swap_tensor/partitioned_optimizer_swapper.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
"""
Functionality of swapping optimizer tensors to/from (NVMe) storage devices.
"""
import torch
from deepspeed.utils.logging import logger
from deepspeed.ops.op_builder import AsyncIOBuilder
from deepspeed import comm as d... | 9,654 | 43.086758 | 114 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/compression/cupy.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import cupy
from torch.utils.dlpack import to_dlpack
from torch.utils.dlpack import from_dlpack
class CupyBackend(object):
def __init__(self):
pass
def torch2cupy(self, tensor):
return cupy.fromDl... | 701 | 25 | 62 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/comm/nccl.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
from deepspeed import comm as dist
import cupy
import numpy as np
from deepspeed.runtime.compression.cupy import CupyBackend
from deepspeed.accelerator import get_accelerator
class NcclBackend(object):
d... | 7,712 | 44.370588 | 119 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/comm/coalesced_collectives.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
"""
batched collective operations for overhead amortization and better
bandwidth utilization
"""
import math
from typing import List
import torch
from torch import Tensor
from deepspeed import comm as dist
# NOTE: Use torch.... | 6,261 | 46.082707 | 119 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/comm/mpi.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
import cupy
import time
import numpy as np
from mpi4py import MPI
from deepspeed.runtime.compression.cupy import CupyBackend
class MpiBackend(object):
def __init__(self, cuda_aware):
self.comm = ... | 10,062 | 45.587963 | 119 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/data_pipeline/data_sampling/data_analyzer.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import os
from collections import defaultdict
import csv
import time
from multiprocessing import Process, Manager
import numpy as np
import torch
from torch.utils.data import BatchSampler, SequentialSampler, DataLoader, Subs... | 25,023 | 58.866029 | 191 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/data_pipeline/data_sampling/data_sampler.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
"""
coding=utf-8
Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obta... | 19,160 | 55.522124 | 177 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/data_pipeline/data_sampling/indexed_dataset.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
"""
Part of this code was adopted from https://github.com/NVIDIA/Megatron-LM/blob/main/megatron/data/indexed_dataset.py
"""
# Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT li... | 20,614 | 32.357605 | 115 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/data_pipeline/data_routing/basic_layer.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
from deepspeed.utils import logger
from torch import Tensor
from torch.nn import Module
from ..constants import *
from deepspeed.ops.random_ltd.dropping_utils import gpt_sample_tokens, bert_sample_tokens, GatherTokens, Scatt... | 5,638 | 48.464912 | 118 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/data_pipeline/data_routing/utils.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
def bsh_decoder_gather(reserved_length, hidden_states, mask):
# random-layer-token-drop
rand_list = []
part_hidden_states = [] # batch, seq, hidden ## different from megatron
for k in range(h... | 955 | 33.142857 | 100 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/zero/stage_1_and_2.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
import os
from deepspeed import comm as dist
from packaging import version as pkg_version
from collections import OrderedDict
from torch._utils import _flatten_dense_tensors, _unflatten_dense_tensors
from deeps... | 110,957 | 46.397693 | 237 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/zero/parameter_offload.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import sys
import torch
from collections import OrderedDict
from deepspeed.runtime.utils import see_memory_usage
from deepspeed.runtime.zero.offload_config import OffloadDeviceEnum
from deepspeed.runtime.zero.partition_param... | 22,985 | 42.369811 | 177 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/zero/test.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
from deepspeed.runtime.zero.contiguous_memory_allocator import ContiguousMemoryAllocator
def test1():
mem = ContiguousMemoryAllocator(1024, torch.half, 'cpu')
mem.print_allocation(resolution=100)
a... | 2,727 | 33.974359 | 97 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/zero/contiguous_memory_allocator.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
from deepspeed import comm as dist
def print_rank_0(message):
if dist.get_rank() == 0:
print(message)
class ContiguousMemoryAllocator(object):
def __init__(self, size, dtype, device):
... | 10,926 | 36.940972 | 147 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/zero/mics_utils.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
# Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
# SPDX-License-Identifier: Apache-2.0
import os
from dataclasses import dataclass
from typing import List
import numpy as np
import torch
from torch impo... | 7,500 | 35.769608 | 116 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/zero/partition_parameters.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import math
import os
import types
from typing import Callable, Iterable
from enum import Enum
import functools
import itertools
from typing import List
import logging
import torch
from torch import Tensor
from deepspeed imp... | 87,015 | 44.062662 | 182 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/zero/stage3.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import sys
import gc
import collections
from typing import Deque, Dict, Tuple
from deepspeed import comm as dist
from deepspeed.utils import groups
from torch._utils import _flatten_dense_tensors, _unflatten_dense_tensors
f... | 118,239 | 45.241689 | 219 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/zero/utils.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import os
from typing import List
import torch
from deepspeed import comm as dist
from deepspeed.utils import logger
from deepspeed.ops.adam import DeepSpeedCPUAdam
from deepspeed.ops.adagrad import DeepSpeedCPUAdagrad
from... | 3,033 | 33.477273 | 115 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/zero/tiling.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
import deepspeed
from deepspeed.runtime.utils import partition_uniform as partition
def split_tensor_along_last_dim(tensor, partitions, contiguous_split_chunks=False):
"""Split a tensor along its last dime... | 11,727 | 38.488215 | 115 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/zero/partitioned_param_coordinator.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
from dataclasses import dataclass
import collections
from collections import UserDict
from typing import Deque, Set
from deepspeed import comm as dist
from deepspeed.utils.logging import logger
from deepspeed.runtime.zero.o... | 25,138 | 48.978131 | 150 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/zero/linear.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
#Linear Module to use with ZeRO Stage 3 to allow for parameter memory release
#after the module execution during forward
#Instead of saving variables using save_for_backward, we save variable ids
#Allowing us to retrieve the... | 7,403 | 41.068182 | 162 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/zero/mics.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
# Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
# SPDX-License-Identifier: Apache-2.0
import sys
from typing import List
import deepspeed
import torch
from deepspeed import comm as dist
from deepspeed.... | 22,259 | 46.061311 | 138 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/checkpoint_engine/torch_checkpoint_engine.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
from deepspeed.utils import logger, log_dist
from deepspeed.runtime.checkpoint_engine.checkpoint_engine import \
CheckpointEngine
class TorchCheckpointEngine(CheckpointEngine):
def __init__(self, conf... | 1,060 | 29.314286 | 78 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/checkpoint_engine/nebula_checkpoint_engine.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import os
import torch
import torch_nebula
from deepspeed.runtime.checkpoint_engine.checkpoint_engine import \
CheckpointEngine
from deepspeed.utils import logger, log_dist
from deepspeed.nebula.constants import *
def... | 4,975 | 45.074074 | 150 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/pipe/engine.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
from types import MethodType
import torch
from deepspeed import comm as dist
from deepspeed.utils import logger
from deepspeed.utils.timer import ThroughputTimer
from deepspeed.accelerator import get_accelerator
from ..en... | 56,886 | 42.49159 | 128 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/pipe/p2p.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import pickle
import typing
import torch
from deepspeed import comm as dist
# To query whether we have send/recv support
from packaging.version import Version
from deepspeed.git_version_info import torch_info
from deepspee... | 5,477 | 28.294118 | 98 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/pipe/topology.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
from deepspeed import comm as dist
from collections import namedtuple
from itertools import product as cartesian_product
class ProcessTopology:
""" Manages the mapping of n-dimensional Cartesian coordinates to linear
... | 17,167 | 36.56674 | 116 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/pipe/module.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import os
import glob
import re as regex
from functools import partial
import torch
import torch.nn as nn
from deepspeed import comm as dist
from deepspeed.utils import logger
from .. import utils as ds_utils
from ..acti... | 27,190 | 42.998382 | 171 | py |
DeepSpeed | DeepSpeed-master/deepspeed/runtime/pipe/schedule.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
from ..utils import call_to_str
from abc import ABC, abstractmethod
class PipeSchedule(ABC):
"""Directs the execution of a pipeline engine by generating sequences of
:class:`PipeInstruction`.
Schedules are ge... | 15,546 | 30.408081 | 103 | py |
DeepSpeed | DeepSpeed-master/deepspeed/moe/layer.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
from deepspeed.utils import log_dist
from deepspeed.utils import groups
from .sharded_moe import MOELayer, TopKGate
from .experts import Experts
import typing
class MoE(torch.nn.Module):
"""Initialize an... | 6,082 | 47.664 | 151 | py |
DeepSpeed | DeepSpeed-master/deepspeed/moe/mappings.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
# The file has been adapted from the following Megatron-LM file:
# https://github.com/NVIDIA/Megatron-LM/blob/main/megatron/mpu/mappings.py
# Git commit hash: 9dc3c42a84aa656f583703cf8b6b4f79f712b796
# We retain the followin... | 3,529 | 31.990654 | 144 | py |
DeepSpeed | DeepSpeed-master/deepspeed/moe/sharded_moe.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
"""
The file has been adapted from two fairscale files:
(1) https://github.com/facebookresearch/fairscale/blob/master/fairscale/nn/moe/moe_layer.py
(2) https://github.com/facebookresearch/fairscale/blob/master/fairscale/nn/... | 20,621 | 36.494545 | 151 | py |
DeepSpeed | DeepSpeed-master/deepspeed/moe/utils.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
from typing import List, Tuple, Dict
import torch
from .layer import MoE
def has_moe_layers(m):
has_moe = False
num_experts = 0
for _, module in m.named_modules():
if isinstance(module, MoE):
... | 5,214 | 34.719178 | 98 | py |
DeepSpeed | DeepSpeed-master/deepspeed/moe/experts.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
import copy
class Experts(torch.nn.Module):
def __init__(self, expert, num_local_experts=1, expert_group_name=None):
super(Experts, self).__init__()
self.deepspeed_experts = torch.nn.Modu... | 1,223 | 33 | 111 | py |
DeepSpeed | DeepSpeed-master/deepspeed/launcher/launch.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
"""
DeepSpeed launcher, this is similar to torch's distributed.launch but supports
additional features such as arbitrary gpu exclusion.
deepspeed.launcher.launch is intended to be run on a single worker node and
will spawn s... | 14,654 | 40.16573 | 117 | py |
DeepSpeed | DeepSpeed-master/deepspeed/module_inject/module_quantize.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
def quantize_transformer_layer(orig_layer_impl, model, megatron=False, preln=False):
""" Quantize bert-style transformer layers with DeepSpeed's transformer layer
Arguments:
orig_layer_impl (to... | 3,107 | 41.575342 | 120 | py |
DeepSpeed | DeepSpeed-master/deepspeed/module_inject/load_checkpoint.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
from torch import nn
from deepspeed.model_implementations.transformers.ds_bloom import DeepSpeedBloomInference
from deepspeed.model_implementations.transformers.ds_gpt import DeepSpeedGPTInference
from deepspeed.model_implem... | 14,745 | 53.817844 | 134 | py |
DeepSpeed | DeepSpeed-master/deepspeed/module_inject/policy.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
from abc import ABC, abstractmethod
from deepspeed.utils.types import ActivationFuncType, NormType
import torch
from deepspeed.accelerator import get_accelerator
transformer_param_names = (
'attn_qkvw', \
'a... | 8,259 | 35.711111 | 155 | py |
DeepSpeed | DeepSpeed-master/deepspeed/module_inject/auto_tp.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
# Automatic Tensor Parallelism
import re
from torch import nn
from .replace_policy import replace_policies
class AutoTP():
def in_module_list(module, module_list):
for item in module_list:
if type... | 5,353 | 41.832 | 144 | py |
DeepSpeed | DeepSpeed-master/deepspeed/module_inject/layers.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import torch
from deepspeed import comm as dist
from torch import nn
from torch.nn import functional as F
from torch.nn.parameter import Parameter
from deepspeed.accelerator import get_accelerator
class LinearAllreduce(nn... | 4,592 | 33.276119 | 116 | py |
DeepSpeed | DeepSpeed-master/deepspeed/module_inject/inject.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import copy
import torch
from deepspeed.ops.transformer import DeepSpeedTransformerLayer, DeepSpeedTransformerConfig
def module_inject(layer_obj, model, config, micro_batch_size, max_seq_length, seed, preln, fp16=True):
... | 4,719 | 40.769912 | 108 | py |
DeepSpeed | DeepSpeed-master/deepspeed/module_inject/replace_module.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
import os
from typing import Optional
import torch
import tqdm
import deepspeed
import deepspeed.ops.transformer as transformer_inference
from deepspeed.ops.transformer.inference.diffusers_attention import DeepSpeedDiffusers... | 43,185 | 47.144928 | 156 | py |
DeepSpeed | DeepSpeed-master/deepspeed/module_inject/containers/distil_bert.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
from .base import *
from deepspeed.model_implementations.transformers.ds_bert import DeepSpeedBERTInference
import torch
from torch.nn.parameter import Parameter
from ..policy import TransformerPolicy
class DS_DistilBERTCo... | 3,188 | 37.421687 | 88 | py |
DeepSpeed | DeepSpeed-master/deepspeed/module_inject/containers/bert.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
from .base import *
from deepspeed.model_implementations.transformers.ds_bert import DeepSpeedBERTInference
import torch
from torch.nn.parameter import Parameter
from ..policy import TransformerPolicy
class DS_BERTContaine... | 3,769 | 39.106383 | 87 | py |
DeepSpeed | DeepSpeed-master/deepspeed/module_inject/containers/base.py | # Copyright (c) Microsoft Corporation.
# SPDX-License-Identifier: Apache-2.0
# DeepSpeed Team
# Create a container object to save model-specific tensors using the policy file above.
from abc import ABC
import torch
import deepspeed
from deepspeed.ops.transformer.inference.config import DeepSpeedInferenceConfig
from... | 13,564 | 41.523511 | 126 | py |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.