repo
stringlengths
7
90
file_url
stringlengths
81
315
file_path
stringlengths
4
228
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 14:38:15
2026-01-05 02:33:18
truncated
bool
2 classes
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/passes/passes_for_gpt2_test.py
colossalai/fx/passes/passes_for_gpt2_test.py
import inspect from typing import Any, Callable, Dict, List, Optional import torch from packaging import version from torch.fx._compatibility import compatibility from torch.fx.graph_module import GraphModule from colossalai.fx.passes.adding_split_node_pass import balanced_split_pass, pipe_split from colossalai.fx.pa...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/passes/experimental/adding_shape_consistency_pass.py
colossalai/fx/passes/experimental/adding_shape_consistency_pass.py
import builtins import operator from typing import List import torch from colossalai.tensor.shape_consistency import ShapeConsistencyManager from colossalai.tensor.sharding_spec import ShardingSpec def apply(*args, **kwargs): shape_consistency_manager = ShapeConsistencyManager() return shape_consistency_man...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/moe/_operation.py
colossalai/moe/_operation.py
from typing import Any, List, Optional, Tuple import torch import torch.distributed as dist from torch import Tensor from torch.cuda.amp import custom_bwd, custom_fwd from torch.distributed import ProcessGroup from colossalai.quantization.fp8 import all_to_all_single_fp8 MOE_KERNEL = None def load_moe(): globa...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/moe/__init__.py
colossalai/moe/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/weight_grad_store.py
colossalai/pipeline/weight_grad_store.py
import queue class WeightGradStore: cache = [] weight_grad_queue = [queue.Queue(), queue.Queue()] @classmethod def put(cls, total_input, grad_output, weight, func): cls.cache.append((total_input, grad_output, weight, func)) @classmethod def flush(cls, chunk=0): cls.weight_gr...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/stage_manager.py
colossalai/pipeline/stage_manager.py
import contextlib from typing import Dict, List, Optional, Tuple, Union import numpy as np import torch.distributed as dist from torch.distributed import ProcessGroup from colossalai.cluster import ProcessGroupMesh class PipelineStageManager: """PipelineStageManager is a helper class to manage pipeline stages. ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/__init__.py
colossalai/pipeline/__init__.py
from .p2p import PipelineP2PCommunication from .schedule import InterleavedSchedule, OneForwardOneBackwardSchedule, PipelineSchedule, ZeroBubbleVPipeScheduler from .stage_manager import PipelineStageManager __all__ = [ "PipelineSchedule", "OneForwardOneBackwardSchedule", "InterleavedSchedule", "ZeroBub...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/p2p.py
colossalai/pipeline/p2p.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- import io import pickle import re from collections import namedtuple from typing import Any, Callable, List, Optional, Tuple, Union import torch import torch.distributed as dist from packaging.version import Version from torch.distributed import ProcessGroup from torch....
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/schedule/one_f_one_b.py
colossalai/pipeline/schedule/one_f_one_b.py
from functools import partial from typing import Any, Callable, Dict, Iterable, List, Optional, Union import torch from torch.nn import Module from torch.utils._pytree import tree_map from colossalai.accelerator import get_accelerator from colossalai.interface import ModelWrapper, OptimizerWrapper from colossalai.pip...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/schedule/interleaved_pp.py
colossalai/pipeline/schedule/interleaved_pp.py
from functools import partial from typing import Any, Callable, Dict, Iterable, List, Optional, Tuple, Union import torch import torch.distributed from torch.nn import Module, ModuleList from torch.utils._pytree import tree_map from colossalai.accelerator import get_accelerator from colossalai.interface import Optimi...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/schedule/generate.py
colossalai/pipeline/schedule/generate.py
import time from functools import partial from typing import Any, Iterable, Optional, Union import torch import torch.cuda from torch.nn import Module from torch.utils._pytree import tree_map from colossalai.accelerator import get_accelerator from colossalai.inference.engine.microbatch_manager import MicroBatchManage...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/schedule/zero_bubble_pp.py
colossalai/pipeline/schedule/zero_bubble_pp.py
from functools import partial from typing import Any, Callable, Dict, Iterable, List, Optional, Union import torch import torch.cuda import torch.distributed from torch.nn import Module, ModuleList from torch.utils._pytree import tree_flatten, tree_map from colossalai.accelerator import get_accelerator from colossala...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/schedule/__init__.py
colossalai/pipeline/schedule/__init__.py
from .base import PipelineSchedule from .interleaved_pp import InterleavedSchedule from .one_f_one_b import OneForwardOneBackwardSchedule from .zero_bubble_pp import ZeroBubbleVPipeScheduler __all__ = [ "PipelineSchedule", "OneForwardOneBackwardSchedule", "InterleavedSchedule", "ZeroBubbleVPipeSchedule...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/schedule/base.py
colossalai/pipeline/schedule/base.py
from typing import Any, Callable, Iterable, Optional from torch import Tensor from torch.nn import Module from colossalai.interface import OptimizerWrapper from colossalai.pipeline.stage_manager import PipelineStageManager class PipelineSchedule: def __init__(self, stage_manager: PipelineStageManager) -> None: ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/schedule/v_schedule.py
colossalai/pipeline/schedule/v_schedule.py
# Refer from Zero Bubble Pipeline Parallelism. # Github: https://github.com/sail-sg/zero-bubble-pipeline-parallelism # Paper: https://arxiv.org/abs/2401.10241 # The following applies to all files unless otherwise noted: # Copyright (c) 2022, NVIDIA CORPORATION. All rights reserved. # # Redistribution and use in source ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/schedule/_utils.py
colossalai/pipeline/schedule/_utils.py
from collections import OrderedDict from typing import Any, List, Optional, Tuple import torch import torch.cuda from packaging.version import Version from torch.nn import Module from torch.utils._pytree import SUPPORTED_NODES, TreeSpec, tree_flatten, tree_map, tree_unflatten # this register are for torch under vers...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/__init__.py
examples/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/fp8/mnist/main.py
examples/community/fp8/mnist/main.py
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # See LICENSE for license information. import argparse import torch import torch.nn as nn import torch.nn.functional as F import torch.optim as optim from torch.optim.lr_scheduler import StepLR from torchvision import datasets, transfo...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/nvidia_bert_dataset_provider.py
examples/community/roberta/pretraining/nvidia_bert_dataset_provider.py
import os import random import time from concurrent.futures import ProcessPoolExecutor import h5py import numpy as np import torch import torch.distributed as dist from bert_dataset_provider import BertDatasetProviderInterface from torch.utils.data import DataLoader, Dataset from torch.utils.data.distributed import Di...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/pretrain_utils.py
examples/community/roberta/pretraining/pretrain_utils.py
import os import sys import torch import transformers from transformers import get_linear_schedule_with_warmup from colossalai.legacy.core import global_context as gpc from colossalai.nn.optimizer import HybridAdam sys.path.append(os.getcwd()) from collections import OrderedDict import torch.nn as nn from model.ber...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/arguments.py
examples/community/roberta/pretraining/arguments.py
import argparse __all__ = ["parse_args"] def parse_args(): parser = argparse.ArgumentParser() parser.add_argument( "--distplan", type=str, default="CAI_Gemini", help="The distributed plan [colossalai, zero1, zero2, torch_ddp, torch_zero].", ) parser.add_argument( ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/bert_dataset_provider.py
examples/community/roberta/pretraining/bert_dataset_provider.py
class BertDatasetProviderInterface: def get_shard(self, index, shuffle=True): raise NotImplementedError def release_shard(self, index): raise NotImplementedError def prefetch_shard(self, index): raise NotImplementedError def get_batch(self, batch_iter): raise NotImplem...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/evaluation.py
examples/community/roberta/pretraining/evaluation.py
import math import os import torch from nvidia_bert_dataset_provider import NvidiaBertDatasetProvider from tqdm import tqdm from utils.global_vars import get_tensorboard_writer, get_timers def evaluate(model, args, logger, global_step, criterion): evaluate_dataset_provider = NvidiaBertDatasetProvider(args, evalu...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/loss.py
examples/community/roberta/pretraining/loss.py
import torch __all__ = ["LossForPretraining"] class LossForPretraining(torch.nn.Module): def __init__(self, vocab_size): super(LossForPretraining, self).__init__() self.loss_fn = torch.nn.CrossEntropyLoss(ignore_index=-1) self.vocab_size = vocab_size def forward(self, prediction_scor...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/run_pretraining.py
examples/community/roberta/pretraining/run_pretraining.py
import math import os import time from functools import partial import torch from arguments import parse_args from evaluation import evaluate from loss import LossForPretraining from nvidia_bert_dataset_provider import NvidiaBertDatasetProvider from pretrain_utils import get_lr_scheduler, get_model, get_optimizer, sav...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/model/bert.py
examples/community/roberta/pretraining/model/bert.py
# coding=utf-8 # Copyright 2018 The Google AI Language Team Authors and The HuggingFace Inc. team. # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a cop...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/model/deberta_v2.py
examples/community/roberta/pretraining/model/deberta_v2.py
# coding=utf-8 # Copyright 2020 Microsoft and the Hugging Face Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless require...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/utils/global_vars.py
examples/community/roberta/pretraining/utils/global_vars.py
import time import torch from .WandbLog import TensorboardLog _GLOBAL_TIMERS = None _GLOBAL_TENSORBOARD_WRITER = None def set_global_variables(launch_time, tensorboard_path): _set_timers() _set_tensorboard_writer(launch_time, tensorboard_path) def _set_timers(): """Initialize timers.""" global _G...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/utils/logger.py
examples/community/roberta/pretraining/utils/logger.py
import logging import torch.distributed as dist logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s", datefmt="%m/%d/%Y %H:%M:%S", level=logging.INFO ) logger = logging.getLogger(__name__) class Logger: def __init__(self, log_path, cuda=False, debug=False): self.logger ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/utils/exp_util.py
examples/community/roberta/pretraining/utils/exp_util.py
import functools import os import shutil import psutil import torch from colossalai.legacy.core import global_context as gpc def logging(s, log_path, print_=True, log_=True): if print_: print(s) if log_: with open(log_path, "a+") as f_log: f_log.write(s + "\n") def get_logger(l...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/utils/WandbLog.py
examples/community/roberta/pretraining/utils/WandbLog.py
import os import time import wandb from torch.utils.tensorboard import SummaryWriter class WandbLog: @classmethod def init_wandb(cls, project, notes=None, name=time.strftime("%Y-%m-%d %H:%M:%S", time.localtime()), config=None): wandb.init(project=project, notes=notes, name=name, config=config) @...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/preprocessing/tokenize_mask.py
examples/community/roberta/preprocessing/tokenize_mask.py
import argparse import multiprocessing import os import time from random import shuffle import h5py import numpy as np import psutil from get_mask import PreTrainingDataset from tqdm import tqdm from transformers import AutoTokenizer def get_raw_instance(document, max_sequence_length=512): """ Get the initia...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/preprocessing/get_mask.py
examples/community/roberta/preprocessing/get_mask.py
import collections import logging import random import jieba jieba.setLogLevel(logging.CRITICAL) import re import mask import numpy as np PAD = 0 MaskedLMInstance = collections.namedtuple("MaskedLMInstance", ["index", "label"]) def map_to_numpy(data): return np.asarray(data) class PreTrainingDataset: de...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/preprocessing/sentence_split.py
examples/community/roberta/preprocessing/sentence_split.py
import argparse import functools import json import multiprocessing import os import re import time from typing import List from tqdm import tqdm def split_sentence(document: str, flag: str = "all", limit: int = 510) -> List[str]: sent_list = [] try: if flag == "zh": document = re.sub("(?...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/download_cifar10.py
examples/tutorial/download_cifar10.py
import os from torchvision.datasets import CIFAR10 def main(): dir_path = os.path.dirname(os.path.realpath(__file__)) data_root = os.path.join(dir_path, "data") dataset = CIFAR10(root=data_root, download=True) if __name__ == "__main__": main()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/large_batch_optimizer/train.py
examples/tutorial/large_batch_optimizer/train.py
import torch import torch.nn as nn from torchvision.models import resnet18 from tqdm import tqdm import colossalai from colossalai.legacy.core import global_context as gpc from colossalai.logging import get_dist_logger from colossalai.nn.lr_scheduler import CosineAnnealingWarmupLR from colossalai.nn.optimizer import L...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/large_batch_optimizer/config.py
examples/tutorial/large_batch_optimizer/config.py
from colossalai.legacy.amp import AMP_TYPE # hyperparameters # BATCH_SIZE is as per GPU # global batch size = BATCH_SIZE x data parallel size BATCH_SIZE = 512 LEARNING_RATE = 3e-3 WEIGHT_DECAY = 0.3 NUM_EPOCHS = 2 WARMUP_EPOCHS = 1 # model config NUM_CLASSES = 10 fp16 = dict(mode=AMP_TYPE.NAIVE) clip_grad_norm = 1.0...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/hybrid_parallel/train.py
examples/tutorial/hybrid_parallel/train.py
import os import torch from titans.model.vit.vit import _create_vit_model from tqdm import tqdm import colossalai from colossalai.legacy.context import ParallelMode from colossalai.legacy.core import global_context as gpc from colossalai.legacy.nn import CrossEntropyLoss from colossalai.legacy.pipeline.pipelinable im...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/hybrid_parallel/config.py
examples/tutorial/hybrid_parallel/config.py
from colossalai.legacy.amp import AMP_TYPE # hyperparameters # BATCH_SIZE is as per GPU # global batch size = BATCH_SIZE x data parallel size BATCH_SIZE = 4 LEARNING_RATE = 3e-3 WEIGHT_DECAY = 0.3 NUM_EPOCHS = 2 WARMUP_EPOCHS = 1 # model config IMG_SIZE = 224 PATCH_SIZE = 16 HIDDEN_SIZE = 128 DEPTH = 4 NUM_HEADS = 4 ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/opt/opt/colossalai_zero.py
examples/tutorial/opt/opt/colossalai_zero.py
try: from colossalai.zero.shard_utils import TensorShardStrategy except ImportError: # colossalai > 0.2.8 from colossalai.legacy.zero import TensorShardStrategy zero = dict( model_config=dict(shard_strategy=TensorShardStrategy(), tensor_placement_policy="auto", reuse_fp16_shard=True), optimizer_con...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/opt/opt/context.py
examples/tutorial/opt/opt/context.py
import torch.distributed as dist from colossalai.legacy.context import ParallelMode from colossalai.legacy.core import global_context as gpc class barrier_context: """ This context manager is used to allow one process to execute while blocking all other processes in the same process group. This is often ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/opt/opt/run_clm.py
examples/tutorial/opt/opt/run_clm.py
#!/usr/bin/env python # coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LI...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/opt/inference/opt_server.py
examples/tutorial/opt/inference/opt_server.py
import argparse import logging import random from typing import Optional from batch import BatchManagerForGeneration from cache import ListCache, MissCacheError from energonai import QueueFullError, launch_engine from energonai.model import opt_6B, opt_30B, opt_125M, opt_175B from pydantic import BaseModel, Field from...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/opt/inference/opt_fastapi.py
examples/tutorial/opt/inference/opt_fastapi.py
import argparse import logging import random from typing import Optional import uvicorn from batch import BatchManagerForGeneration from cache import ListCache, MissCacheError from energonai import QueueFullError, launch_engine from energonai.model import opt_6B, opt_30B, opt_125M, opt_175B from fastapi import FastAPI...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/opt/inference/cache.py
examples/tutorial/opt/inference/cache.py
from collections import OrderedDict from contextlib import contextmanager from threading import Lock from typing import Any, Dict, Hashable, List class MissCacheError(Exception): pass class ListCache: def __init__(self, cache_size: int, list_size: int, fixed_keys: List[Hashable] = []) -> None: """Ca...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/opt/inference/batch.py
examples/tutorial/opt/inference/batch.py
from typing import Any, Deque, Hashable, List, Tuple import torch from energonai import BatchManager, SubmitEntry, TaskEntry class BatchManagerForGeneration(BatchManager): def __init__(self, max_batch_size: int = 1, pad_token_id: int = 0) -> None: super().__init__() self.max_batch_size = max_batc...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/opt/inference/benchmark/locustfile.py
examples/tutorial/opt/inference/benchmark/locustfile.py
from locust import HttpUser, task class GenerationUser(HttpUser): @task def generate(self): prompt = "Question: What is the longest river on the earth? Answer:" for i in range(4, 9): data = {"max_tokens": 2**i, "prompt": prompt} with self.client.post("/generation", json...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/opt/inference/script/processing_ckpt_66b.py
examples/tutorial/opt/inference/script/processing_ckpt_66b.py
import os from multiprocessing import Pool import torch # download pytorch model ckpt in https://huggingface.co/facebook/opt-66b/tree/main # you can use whether wget or git lfs path = "/path/to/your/ckpt" new_path = "/path/to/the/processed/ckpt/" assert os.path.isdir(path) files = [] for filename in os.listdir(path...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/opt/inference/script/process-opt-175b/convert_ckpt.py
examples/tutorial/opt/inference/script/process-opt-175b/convert_ckpt.py
import argparse import json import os import re from collections import defaultdict import numpy as np import torch def load_json(path: str): with open(path) as f: return json.load(f) def parse_shape_info(flat_dir: str): data = load_json(os.path.join(flat_dir, "shape.json")) flat_info = default...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/new_api/cifar_vit/train.py
examples/tutorial/new_api/cifar_vit/train.py
import argparse import os from pathlib import Path import torch import torch.distributed as dist import torch.nn as nn import torchvision import torchvision.transforms as transforms from timm.models.vision_transformer import _cfg, _create_vision_transformer from torch.optim import Optimizer from torch.utils.data impor...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/new_api/glue_bert/finetune.py
examples/tutorial/new_api/glue_bert/finetune.py
import argparse from typing import List, Union import datasets import torch import torch.distributed as dist import torch.nn as nn from data import GLUEDataBuilder from torch.optim import Optimizer from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoConfig, BertForSequenceClassif...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/new_api/glue_bert/data.py
examples/tutorial/new_api/glue_bert/data.py
import datasets from transformers import AutoTokenizer, PreTrainedTokenizer from colossalai.booster.plugin.dp_plugin_base import DPPluginBase class GLUEDataBuilder: task_text_field_map = { "cola": ["sentence"], "sst2": ["sentence"], "mrpc": ["sentence1", "sentence2"], "qqp": ["que...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/new_api/cifar_resnet/train.py
examples/tutorial/new_api/cifar_resnet/train.py
import argparse import os from pathlib import Path import torch import torch.distributed as dist import torch.nn as nn import torchvision import torchvision.transforms as transforms from torch.optim import Optimizer from torch.optim.lr_scheduler import MultiStepLR from torch.utils.data import DataLoader from tqdm impo...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/new_api/cifar_resnet/eval.py
examples/tutorial/new_api/cifar_resnet/eval.py
import argparse import torch import torchvision import torchvision.transforms as transforms # ============================== # Parse Arguments # ============================== parser = argparse.ArgumentParser() parser.add_argument("-e", "--epoch", type=int, default=80, help="resume from the epoch's checkpoint") parse...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/train.py
examples/tutorial/sequence_parallel/train.py
import argparse import torch from data.bert_helper import SequenceParallelDataIterator, get_batch_for_sequence_parallel from data.dummy_dataloader import DummyDataloader from loss_func.bert_loss import BertLoss from lr_scheduler import AnnealingLR from model.bert import BertForPretrain, build_pipeline_bert import col...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/config.py
examples/tutorial/sequence_parallel/config.py
from colossalai.legacy.amp import AMP_TYPE # hyper-parameters TRAIN_ITERS = 10 DECAY_ITERS = 4 WARMUP_FRACTION = 0.01 GLOBAL_BATCH_SIZE = 32 # dp world size * sentences per GPU EVAL_ITERS = 10 EVAL_INTERVAL = 10 LR = 0.0001 MIN_LR = 1e-05 WEIGHT_DECAY = 0.01 SEQ_LENGTH = 128 # BERT config DEPTH = 4 NUM_ATTENTION_HEA...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/loss_func/bert_loss.py
examples/tutorial/sequence_parallel/loss_func/bert_loss.py
import torch import torch.nn as nn import torch.nn.functional as F from colossalai.legacy.context import ParallelMode from colossalai.legacy.core import global_context as gpc class BertLoss(nn.Module): def forward(self, lm_loss, sop_logits, loss_mask, sentence_order): lm_loss_ = lm_loss.float() l...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/loss_func/cross_entropy.py
examples/tutorial/sequence_parallel/loss_func/cross_entropy.py
import torch from torch.cuda.amp import custom_bwd, custom_fwd class _VocabCrossEntropy(torch.autograd.Function): @staticmethod @custom_fwd def forward(ctx, vocab_parallel_logits, target): # Maximum value along vocab dimension across all GPUs. logits_max = torch.max(vocab_parallel_logits, ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/loss_func/utils.py
examples/tutorial/sequence_parallel/loss_func/utils.py
import torch def ensure_divisibility(numerator, denominator): """Ensure that numerator is divisible by the denominator.""" assert numerator % denominator == 0, "{} is not divisible by {}".format(numerator, denominator) def divide(numerator, denominator): """Ensure that numerator is divisible by the deno...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/loss_func/__init__.py
examples/tutorial/sequence_parallel/loss_func/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/bert.py
examples/tutorial/sequence_parallel/model/bert.py
import inspect import torch import torch.nn as nn from colossalai.legacy.context import ParallelMode from colossalai.legacy.context.parallel_mode import ParallelMode from colossalai.legacy.core import global_context as gpc from colossalai.legacy.nn.layer.wrapper import PipelineSharedModuleWrapper from colossalai.lega...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/__init__.py
examples/tutorial/sequence_parallel/model/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/layers/pooler.py
examples/tutorial/sequence_parallel/model/layers/pooler.py
import torch import torch.nn as nn from .linear import Linear class Pooler(nn.Module): """Pooler layer. Pool hidden states of a specific token (for example start of the sequence) and add a linear transformation followed by a tanh. Arguments: hidden_size: hidden size init_method: wei...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/layers/init_method.py
examples/tutorial/sequence_parallel/model/layers/init_method.py
import math import torch def init_normal(tensor, sigma): """Init method based on N(0, sigma).""" torch.nn.init.normal_(tensor, mean=0.0, std=sigma) def output_init_normal(tensor, sigma, num_layers): """Init method based on N(0, sigma/sqrt(2*num_layers).""" std = sigma / math.sqrt(2.0 * num_layers) ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/layers/bert_layer.py
examples/tutorial/sequence_parallel/model/layers/bert_layer.py
import torch import torch.nn as nn from colossalai.kernel.jit import bias_dropout_add_fused_inference, bias_dropout_add_fused_train from colossalai.legacy.nn.layer.parallel_sequence import TransformerSelfAttentionRing from colossalai.nn.layer.layernorm import MixedFusedLayerNorm as LayerNorm from .dropout import get_...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/layers/mlp.py
examples/tutorial/sequence_parallel/model/layers/mlp.py
import torch.nn as nn import torch.nn.functional as F from colossalai.kernel.jit import bias_gelu_impl from .linear import Linear class TransformerMLP(nn.Module): """MLP. MLP will take the input with h hidden state, project it to 4*h hidden dimension, perform nonlinear transformation, and project the ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/layers/dropout.py
examples/tutorial/sequence_parallel/model/layers/dropout.py
import torch def bias_dropout_add(x, bias, residual, prob, training): # type: (Tensor, Tensor, Tensor, float, bool) -> Tensor out = torch.nn.functional.dropout(x + bias, p=prob, training=training) out = residual + out return out def get_bias_dropout_add(training): def _bias_dropout_add(x, bias, ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/layers/__init__.py
examples/tutorial/sequence_parallel/model/layers/__init__.py
from .bert_layer import BertLayer from .embedding import Embedding, VocabEmbedding from .head import BertDualHead from .preprocess import PreProcessor
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/layers/embedding.py
examples/tutorial/sequence_parallel/model/layers/embedding.py
import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.init as init class VocabEmbedding(torch.nn.Module): def __init__(self, num_embeddings, embedding_dim): super(VocabEmbedding, self).__init__() # Keep the input dimensions. self.num_embeddings = num_embeddings...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/layers/head.py
examples/tutorial/sequence_parallel/model/layers/head.py
import torch import torch.nn as nn import torch.nn.functional as F from loss_func.cross_entropy import vocab_cross_entropy from colossalai.legacy.context import ParallelMode from colossalai.legacy.core import global_context as gpc from colossalai.nn.layer.layernorm import MixedFusedLayerNorm as LayerNorm from .linear...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/layers/preprocess.py
examples/tutorial/sequence_parallel/model/layers/preprocess.py
import torch import torch.nn as nn from colossalai.legacy.context.parallel_mode import ParallelMode from colossalai.legacy.core import global_context as gpc class PreProcessor(nn.Module): def __init__(self, sub_seq_length): super().__init__() self.sub_seq_length = sub_seq_length def bert_pos...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/layers/linear.py
examples/tutorial/sequence_parallel/model/layers/linear.py
import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.init as init from torch.nn import Parameter class Linear(nn.Module): """Linear layer with column parallelism. The linear layer is defined as Y = XA + b. A is parallelized along its second dimension as A = [A_1, ..., A_p]. ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/lr_scheduler/annealing_lr.py
examples/tutorial/sequence_parallel/lr_scheduler/annealing_lr.py
# coding=utf-8 # Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless re...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/lr_scheduler/__init__.py
examples/tutorial/sequence_parallel/lr_scheduler/__init__.py
from .annealing_lr import AnnealingLR
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/bert_helper.py
examples/tutorial/sequence_parallel/data/bert_helper.py
import torch from colossalai.legacy.context import ParallelMode from colossalai.legacy.core import global_context as gpc _MAX_DATA_DIM = 5 def _build_key_size_numel_dictionaries(keys, data): """Build the size on rank 0 and broadcast.""" max_dim = _MAX_DATA_DIM sizes = [0 for _ in range(max_dim) for _ in...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/__init__.py
examples/tutorial/sequence_parallel/data/__init__.py
import torch from colossalai.legacy.context import ParallelMode from colossalai.legacy.context.parallel_context import ParallelContext from colossalai.legacy.core import global_context as gpc from colossalai.logging import get_dist_logger from .datasets.builder import build_train_valid_test_datasets from .datasets.da...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/dummy_dataloader.py
examples/tutorial/sequence_parallel/data/dummy_dataloader.py
import torch class DummyDataloader: def __init__(self, batch_size, vocab_size, seq_length): self.batch_size = batch_size self.vocab_size = vocab_size self.seq_length = seq_length self.step = 0 def generate(self): tokens = torch.randint( low=0, h...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/tokenizer/bert_tokenization.py
examples/tutorial/sequence_parallel/data/tokenizer/bert_tokenization.py
# coding=utf-8 # Copyright 2018 The Google AI Language Team Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/tokenizer/__init__.py
examples/tutorial/sequence_parallel/data/tokenizer/__init__.py
# coding=utf-8 # Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless re...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/tokenizer/tokenizer.py
examples/tutorial/sequence_parallel/data/tokenizer/tokenizer.py
# coding=utf-8 # Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless re...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/datasets/bert_dataset.py
examples/tutorial/sequence_parallel/data/datasets/bert_dataset.py
# coding=utf-8 # Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless re...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/datasets/ict_dataset.py
examples/tutorial/sequence_parallel/data/datasets/ict_dataset.py
import itertools import random import numpy as np from megatron import get_args, get_tokenizer from megatron.data.dataset_utils import get_indexed_dataset_ from megatron.data.realm_dataset_utils import get_block_samples_mapping from torch.utils.data import Dataset def make_attention_mask(source_block, target_block):...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/datasets/data_samplers.py
examples/tutorial/sequence_parallel/data/datasets/data_samplers.py
# coding=utf-8 # Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless re...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/datasets/dataset_utils.py
examples/tutorial/sequence_parallel/data/datasets/dataset_utils.py
# coding=utf-8 # Copyright 2018 The Google AI Language Team Authors, and NVIDIA. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/datasets/blendable_dataset.py
examples/tutorial/sequence_parallel/data/datasets/blendable_dataset.py
# coding=utf-8 # Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless re...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/datasets/__init__.py
examples/tutorial/sequence_parallel/data/datasets/__init__.py
from . import indexed_dataset
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/datasets/builder.py
examples/tutorial/sequence_parallel/data/datasets/builder.py
from colossalai.logging import get_dist_logger from .bert_dataset import BertDataset from .blendable_dataset import BlendableDataset from .dataset_utils import get_datasets_weights_and_num_samples, get_indexed_dataset_, get_train_valid_test_split_ DSET_TYPE_BERT = "standard_bert" DSET_TYPE_ICT = "ict" DSET_TYPE_T5 = ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/datasets/indexed_dataset.py
examples/tutorial/sequence_parallel/data/datasets/indexed_dataset.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. # copied from fairseq/fairseq/data/indexed_dataset.py # Removed IndexedRawTextDataset since it relied on Fairseq dictionary # other slight mod...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/datasets/test/test_indexed_dataset.py
examples/tutorial/sequence_parallel/data/datasets/test/test_indexed_dataset.py
# This file isn't really a formal automated test, it's just a place to # put some code used during development and manual testing of # indexed_dataset. import argparse import os import sys from megatron.data import indexed_dataset from megatron.tokenizer import build_tokenizer script_dir = os.path.dirname(os.path.re...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/auto_parallel/auto_ckpt_batchsize_test.py
examples/tutorial/auto_parallel/auto_ckpt_batchsize_test.py
from copy import deepcopy from functools import partial import torch import torchvision.models as tm from bench_utils import bench, data_gen_resnet import colossalai from colossalai.auto_parallel.checkpoint import CheckpointSolverRotor from colossalai.fx import metainfo_trace, symbolic_trace from colossalai.testing i...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/auto_parallel/setup.py
examples/tutorial/auto_parallel/setup.py
from setuptools import find_packages, setup setup( name="auto_parallel", version="0.0.1", description="", packages=find_packages(), install_requires=[ "torch", "numpy", "tqdm", ], )
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/auto_parallel/bench_utils.py
examples/tutorial/auto_parallel/bench_utils.py
import time from copy import deepcopy from typing import Callable, Tuple import numpy as np import torch import torch.nn as nn from transformers import GPT2Config, GPT2LMHeadModel from colossalai.auto_parallel.checkpoint import CheckpointSolverRotor from colossalai.fx import metainfo_trace def bench( gm: torch....
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/auto_parallel/auto_parallel_with_resnet.py
examples/tutorial/auto_parallel/auto_parallel_with_resnet.py
import torch from torchvision.models import resnet50 from tqdm import tqdm import colossalai from colossalai.auto_parallel.tensor_shard.initialize import initialize_model from colossalai.device.device_mesh import DeviceMesh from colossalai.legacy.core import global_context as gpc from colossalai.logging import get_dis...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/auto_parallel/auto_ckpt_solver_test.py
examples/tutorial/auto_parallel/auto_ckpt_solver_test.py
from argparse import ArgumentParser from functools import partial import matplotlib.pyplot as plt import torch import torchvision.models as tm from bench_utils import GPTLMLoss, bench_rotor, data_gen_gpt2, data_gen_resnet, gpt2_medium import colossalai from colossalai.fx import metainfo_trace, symbolic_trace from col...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/auto_parallel/config.py
examples/tutorial/auto_parallel/config.py
BATCH_SIZE = 32 NUM_EPOCHS = 2
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/inference/benchmark_ops/benchmark_context_attn_unpad.py
examples/inference/benchmark_ops/benchmark_context_attn_unpad.py
import torch from transformers.modeling_attn_mask_utils import AttentionMaskConverter from colossalai.inference.modeling.layers.attention import PagedAttention from colossalai.kernel.triton import context_attention_unpadded from colossalai.utils import get_current_device from tests.test_infer.test_kernels.triton.kerne...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/inference/benchmark_ops/benchmark_fused_rotary_embdding_unpad.py
examples/inference/benchmark_ops/benchmark_fused_rotary_embdding_unpad.py
import torch from colossalai.kernel.kernel_loader import InferenceOpsLoader from colossalai.kernel.triton import copy_kv_to_blocked_cache, decoding_fused_rotary_embedding, rotary_embedding from tests.test_infer.test_kernels.triton.kernel_utils import ( mock_alloc_block_table_and_kvcache_v2, mock_alloc_block_ta...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/inference/benchmark_ops/benchmark_decoding_attn.py
examples/inference/benchmark_ops/benchmark_decoding_attn.py
import torch from colossalai.kernel.triton import flash_decoding_attention from colossalai.utils import get_current_device from tests.test_infer.test_kernels.triton.kernel_utils import ( convert_kv_unpad_to_padded, create_attention_mask, generate_caches_and_block_tables_v2, generate_caches_and_block_ta...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/inference/benchmark_ops/benchmark_kv_cache_memcopy.py
examples/inference/benchmark_ops/benchmark_kv_cache_memcopy.py
import torch from colossalai.inference.modeling.layers.attention import copy_to_cache from colossalai.kernel.kernel_loader import InferenceOpsLoader from colossalai.kernel.triton import copy_kv_to_blocked_cache from colossalai.utils import get_current_device from tests.test_infer.test_kernels.cuda.test_kv_cache_memcpy...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/inference/benchmark_ops/benchmark_rmsnorm.py
examples/inference/benchmark_ops/benchmark_rmsnorm.py
import torch from colossalai.kernel.kernel_loader import InferenceOpsLoader from colossalai.kernel.triton import rms_layernorm try: import triton # noqa except ImportError: print("please install triton from https://github.com/openai/triton") inference_ops = InferenceOpsLoader().load() # Triton benchmark pl...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false