repo
stringlengths
7
90
file_url
stringlengths
81
315
file_path
stringlengths
4
228
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 14:38:15
2026-01-05 02:33:18
truncated
bool
2 classes
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/layer/colossalai_layer/__init__.py
colossalai/legacy/nn/layer/colossalai_layer/__init__.py
from ._utils import partition_batch from .dropout import Dropout from .embedding import Embedding, PatchEmbedding from .linear import Classifier, Linear from .normalization import LayerNorm __all__ = ["Linear", "Classifier", "Embedding", "PatchEmbedding", "LayerNorm", "Dropout", "partition_batch"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/layer/colossalai_layer/embedding.py
colossalai/legacy/nn/layer/colossalai_layer/embedding.py
import math from typing import Callable from torch import dtype, nn from colossalai.accelerator import get_accelerator from colossalai.nn import init from ..parallel_1d import Embedding1D, PatchEmbedding1D, VocabParallelEmbedding1D from ..parallel_2d import Embedding2D, PatchEmbedding2D, VocabParallelEmbedding2D fro...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/layer/colossalai_layer/normalization.py
colossalai/legacy/nn/layer/colossalai_layer/normalization.py
from torch import nn from colossalai.accelerator import get_accelerator from ..parallel_1d import LayerNorm1D from ..parallel_2d import LayerNorm2D from ..parallel_2p5d import LayerNorm2p5D from ..parallel_3d import LayerNorm3D from ..utils import get_tensor_parallel_mode from ..vanilla import VanillaLayerNorm from ....
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/layer/colossalai_layer/_utils.py
colossalai/legacy/nn/layer/colossalai_layer/_utils.py
import torch.nn as nn from torch import Tensor from ..parallel_2d._operation import split_batch_2d from ..parallel_2p5d._operation import split_batch_2p5d from ..parallel_3d._operation import split_batch_3d from ..utils import get_tensor_parallel_mode _parallel_split_batch = {"2d": split_batch_2d, "2.5d": split_batch...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/layer/colossalai_layer/linear.py
colossalai/legacy/nn/layer/colossalai_layer/linear.py
import inspect import math from typing import Callable from torch import dtype, nn from colossalai.nn import init from ..parallel_1d import * from ..parallel_2d import * from ..parallel_2p5d import * from ..parallel_3d import * from ..utils import get_tensor_parallel_mode from ..vanilla import * from ._utils import ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/layer/parallel_1d/_operation.py
colossalai/legacy/nn/layer/parallel_1d/_operation.py
import torch import torch.distributed as dist from colossalai.legacy.core import global_context as gpc try: import fused_mix_prec_layer_norm_cuda except: fused_mix_prec_layer_norm_cuda = None class FusedLayerNormAffineFunction1D(torch.autograd.Function): r"""Layernorm Args: input: input mat...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/layer/parallel_1d/__init__.py
colossalai/legacy/nn/layer/parallel_1d/__init__.py
from .layers import ( Classifier1D, Dropout1D, Embedding1D, LayerNorm1D, Linear1D, Linear1D_Col, Linear1D_Row, PatchEmbedding1D, VocabParallelClassifier1D, VocabParallelEmbedding1D, ) __all__ = [ "Linear1D", "Linear1D_Col", "Linear1D_Row", "Embedding1D", "Dro...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/layer/parallel_1d/_utils.py
colossalai/legacy/nn/layer/parallel_1d/_utils.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- import torch import torch.distributed as dist from colossalai.legacy.core import global_context as gpc from colossalai.legacy.global_variables import tensor_parallel_env as env from ..utils import divide def set_parallel_input(input_parallel: bool): env.parallel_...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/layer/parallel_1d/layers.py
colossalai/legacy/nn/layer/parallel_1d/layers.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- import math from collections import OrderedDict from typing import Callable, Tuple import torch import torch.nn.functional as F from torch import Tensor from torch.nn.parameter import Parameter from colossalai.accelerator import get_accelerator from colossalai.legacy.c...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/layer/parallel_3d/_operation.py
colossalai/legacy/nn/layer/parallel_3d/_operation.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- from typing import Optional, Tuple import torch from torch import Tensor from torch.cuda.amp import custom_bwd, custom_fwd from colossalai.legacy.communication import all_gather, all_reduce, broadcast, reduce, reduce_scatter from colossalai.legacy.constants import INPU...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/layer/parallel_3d/__init__.py
colossalai/legacy/nn/layer/parallel_3d/__init__.py
from ._operation import reduce_by_batch_3d, split_batch_3d, split_tensor_3d from .layers import ( Classifier3D, Embedding3D, LayerNorm3D, Linear3D, PatchEmbedding3D, VocabParallelClassifier3D, VocabParallelEmbedding3D, ) __all__ = [ "reduce_by_batch_3d", "split_tensor_3d", "spli...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/layer/parallel_3d/_utils.py
colossalai/legacy/nn/layer/parallel_3d/_utils.py
from collections import OrderedDict from functools import partial import torch from torch import Tensor from colossalai.legacy.constants import ( INPUT_GROUP_3D, INPUT_X_WEIGHT_3D, OUTPUT_GROUP_3D, OUTPUT_X_WEIGHT_3D, WEIGHT_GROUP_3D, ) from colossalai.legacy.core import global_context as gpc from...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/layer/parallel_3d/layers.py
colossalai/legacy/nn/layer/parallel_3d/layers.py
import math from collections import OrderedDict from typing import Callable import torch import torch.nn as nn import torch.nn.functional as F from torch import Tensor from torch.nn import Parameter from colossalai.accelerator import get_accelerator from colossalai.legacy.communication import all_reduce, broadcast fr...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/layer/parallel_2p5d/_operation.py
colossalai/legacy/nn/layer/parallel_2p5d/_operation.py
from typing import Any, Tuple import torch import torch.distributed as dist from torch import Tensor from torch.cuda.amp import custom_bwd, custom_fwd from colossalai.accelerator import get_accelerator from colossalai.legacy.communication.collective import all_gather, all_reduce, reduce_scatter from colossalai.legacy...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/layer/parallel_2p5d/__init__.py
colossalai/legacy/nn/layer/parallel_2p5d/__init__.py
from ._operation import reduce_by_batch_2p5d, split_batch_2p5d from .layers import ( Classifier2p5D, Embedding2p5D, LayerNorm2p5D, Linear2p5D, PatchEmbedding2p5D, VocabParallelClassifier2p5D, VocabParallelEmbedding2p5D, ) __all__ = [ "split_batch_2p5d", "reduce_by_batch_2p5d", "...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/layer/parallel_2p5d/_utils.py
colossalai/legacy/nn/layer/parallel_2p5d/_utils.py
from colossalai.legacy.context.parallel_mode import ParallelMode from colossalai.legacy.core import global_context as gpc from colossalai.legacy.global_variables import tensor_parallel_env as env def get_tesseract_dim_dep_from_env(): try: tesseract_dim = env.tesseract_dim tesseract_dep = env.tesse...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/layer/parallel_2p5d/layers.py
colossalai/legacy/nn/layer/parallel_2p5d/layers.py
import math from collections import OrderedDict from typing import Callable import torch import torch.nn as nn import torch.nn.functional as F from torch import Tensor from torch.nn import Parameter from colossalai.accelerator import get_accelerator from colossalai.legacy.communication import broadcast from colossala...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/metric/accuracy_2p5d.py
colossalai/legacy/nn/metric/accuracy_2p5d.py
import torch from torch import nn from colossalai.legacy.nn.layer.parallel_2p5d import reduce_by_batch_2p5d, split_batch_2p5d from ._utils import calc_acc class Accuracy2p5D(nn.Module): """Accuracy for 2p5D parallelism""" def __init__(self): super().__init__() def forward(self, logits, targets...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/metric/accuracy_3d.py
colossalai/legacy/nn/metric/accuracy_3d.py
import torch from torch import nn from colossalai.legacy.constants import INPUT_GROUP_3D, WEIGHT_GROUP_3D from colossalai.legacy.nn.layer.parallel_3d import reduce_by_batch_3d, split_tensor_3d from colossalai.legacy.nn.layer.parallel_3d._utils import get_parallel_mode_from_env from ._utils import calc_acc class Acc...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/metric/__init__.py
colossalai/legacy/nn/metric/__init__.py
from torch import nn from colossalai.legacy.nn.layer.utils import get_tensor_parallel_mode from ._utils import calc_acc from .accuracy_2d import Accuracy2D from .accuracy_2p5d import Accuracy2p5D from .accuracy_3d import Accuracy3D _parallel_accuracy = { "2d": Accuracy2D, "2.5d": Accuracy2p5D, "3d": Accu...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/metric/_utils.py
colossalai/legacy/nn/metric/_utils.py
import torch def calc_acc(logits, targets): preds = torch.argmax(logits, dim=-1) correct = torch.sum(targets == preds) return correct
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/metric/accuracy_2d.py
colossalai/legacy/nn/metric/accuracy_2d.py
import torch from torch import nn from colossalai.legacy.nn.layer.parallel_2d import reduce_by_batch_2d, split_batch_2d from ._utils import calc_acc class Accuracy2D(nn.Module): """Accuracy for 2D parallelism""" def __init__(self): super().__init__() def forward(self, logits, targets): ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/parallel/reducer.py
colossalai/legacy/nn/parallel/reducer.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the BSD license found in the # LICENSE file in the root directory of this source tree. import functools from typing import Callable, Dict, List, Optional, Tuple import torch import torch.distributed as dist from torch import Ten...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/parallel/data_parallel.py
colossalai/legacy/nn/parallel/data_parallel.py
from collections import OrderedDict from functools import partial from typing import Iterable, Optional, Set import torch import torch.distributed as dist from colossalai.legacy.tensor import ProcessGroup as ColoProcessGroup from colossalai.utils import is_ddp_ignored from .reducer import Reducer def free_storage(...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/parallel/__init__.py
colossalai/legacy/nn/parallel/__init__.py
from .data_parallel import ColoDDP __all__ = [ "ColoDDP", ]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/parallel/layers/module_utils.py
colossalai/legacy/nn/parallel/layers/module_utils.py
from typing import Dict import torch from colossalai.legacy.tensor import ComputeSpec, ProcessGroup from colossalai.tensor import ColoParameter from . import ColoModule _COLOSSAL_MODULES: Dict[type, ColoModule] = {} def register_colo_module(module_type: type, colo_module: ColoModule): global _COLOSSAL_MODULES...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/parallel/layers/__init__.py
colossalai/legacy/nn/parallel/layers/__init__.py
from .cache_embedding import ( CachedEmbeddingBag, CachedParamMgr, EvictionStrategy, LimitBuffIndexCopyer, ParallelCachedEmbeddingBag, ParallelCachedEmbeddingBagTablewise, ParallelCachedEmbeddingBagTablewiseSpiltCache, TablewiseEmbeddingBagConfig, ) from .colo_module import ColoModule fr...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/parallel/layers/colo_module.py
colossalai/legacy/nn/parallel/layers/colo_module.py
from typing import Dict, List from colossalai.legacy.tensor import ComputePattern from colossalai.legacy.tensor.distspec import _DistSpec class ColoModule(object): def __init__(self): self._shard_params: List[str] = [] self._allowed_patterns: Dict[ComputePattern, Dict[str, Dict[str, _DistSpec]]] ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/parallel/layers/embedding.py
colossalai/legacy/nn/parallel/layers/embedding.py
from colossalai.legacy.tensor import ComputePattern, ProcessGroup, ShardSpec from .colo_module import ColoModule class ColoEmbedding(ColoModule): def __init__(self): super(ColoEmbedding, self).__init__() self._register_shard_params(["weight"]) def register(self, compute_pattern, pg: ProcessG...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/parallel/layers/linear.py
colossalai/legacy/nn/parallel/layers/linear.py
from colossalai.legacy.tensor import ComputePattern, ProcessGroup, ShardSpec from .colo_module import ColoModule class ColoLinear(ColoModule): def __init__(self): super(ColoLinear, self).__init__() self._register_shard_params(["weight", "bias"]) def register(self, compute_pattern, pg: Proces...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/parallel/layers/cache_embedding/parallel_cached_embedding.py
colossalai/legacy/nn/parallel/layers/cache_embedding/parallel_cached_embedding.py
from typing import List, Optional, Tuple import torch import torch.nn.functional as F from colossalai.legacy.nn._ops._utils import dual_all_to_all from colossalai.legacy.tensor import ColoTensorSpec, ComputePattern, ProcessGroup, ShardSpec from colossalai.tensor import ColoTensor from .cache_mgr import EvictionStrat...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/parallel/layers/cache_embedding/cached_embedding.py
colossalai/legacy/nn/parallel/layers/cache_embedding/cached_embedding.py
from typing import Iterator, List, Optional, Tuple, Union import torch import torch.nn.functional as F from torch.nn.parameter import Parameter from .base_embedding import BaseEmbeddingBag from .cache_mgr import CachedParamMgr, EvictionStrategy class CachedEmbeddingBag(BaseEmbeddingBag): """CachedEmbeddingBag ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/parallel/layers/cache_embedding/copyer.py
colossalai/legacy/nn/parallel/layers/cache_embedding/copyer.py
import torch from torch import LongTensor class LimitBuffIndexCopyer(object): """LimitBuffIndexCopyer Index Copy using limited temp buffer on CUDA. Args: size (int): buffer size """ def __init__(self, size: int) -> None: self._buff_size = size @torch.no_grad() def index_...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/parallel/layers/cache_embedding/cache_mgr.py
colossalai/legacy/nn/parallel/layers/cache_embedding/cache_mgr.py
import sys from contextlib import contextmanager from enum import Enum from typing import List, Optional import numpy as np import torch from contexttimer import Timer from torch.profiler import record_function from .copyer import LimitBuffIndexCopyer class EvictionStrategy(Enum): LFU = 1 # dataset aware ev...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/parallel/layers/cache_embedding/parallel_cached_embedding_tablewise_split_cache.py
colossalai/legacy/nn/parallel/layers/cache_embedding/parallel_cached_embedding_tablewise_split_cache.py
import abc from typing import List import torch import torch.distributed as dist import torch.nn as nn from torch.profiler import record_function from colossalai.legacy.nn._ops._utils import dual_all_to_all_tablewise from colossalai.legacy.tensor import ProcessGroup from .cache_mgr import EvictionStrategy from .cach...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/parallel/layers/cache_embedding/embedding_config.py
colossalai/legacy/nn/parallel/layers/cache_embedding/embedding_config.py
import torch class TablewiseEmbeddingBagConfig: """ example: def prepare_tablewise_config(args, cache_ratio, ...): embedding_bag_config_list: List[TablewiseEmbeddingBagConfig] = [] ... return embedding_bag_config_list """ def __init__( self, num_embeddings:...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/parallel/layers/cache_embedding/parallel_cached_embedding_tablewise.py
colossalai/legacy/nn/parallel/layers/cache_embedding/parallel_cached_embedding_tablewise.py
from typing import List import torch import torch.distributed as dist import torch.nn.functional as F from colossalai.legacy.nn._ops._utils import dual_all_to_all_tablewise from colossalai.legacy.tensor import ProcessGroup from .cache_mgr import EvictionStrategy from .cached_embedding import CachedEmbeddingBag from ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/parallel/layers/cache_embedding/base_embedding.py
colossalai/legacy/nn/parallel/layers/cache_embedding/base_embedding.py
import abc import torch.nn as nn class BaseEmbeddingBag(abc.ABC, nn.Module): def __init__( self, num_embeddings, embedding_dim, padding_idx=None, max_norm=None, norm_type=2.0, scale_grad_by_freq=False, sparse=False, mode="mean", incl...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/nn/parallel/layers/cache_embedding/__init__.py
colossalai/legacy/nn/parallel/layers/cache_embedding/__init__.py
from .cache_mgr import CachedParamMgr, EvictionStrategy from .cached_embedding import CachedEmbeddingBag from .copyer import LimitBuffIndexCopyer from .embedding_config import TablewiseEmbeddingBagConfig from .parallel_cached_embedding import ParallelCachedEmbeddingBag from .parallel_cached_embedding_tablewise import P...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/engine/_base_engine.py
colossalai/legacy/engine/_base_engine.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- # this code is inspired by the DeepSpeed library and implemented with our own design from scratch from typing import Iterable, List, Optional, Type from torch import Tensor from torch.nn import Module from torch.nn.modules.loss import _Loss from colossalai.interface im...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/engine/__init__.py
colossalai/legacy/engine/__init__.py
from ._base_engine import Engine from .gradient_handler import * __all__ = ["Engine"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/engine/gradient_handler/_zero_gradient_handler.py
colossalai/legacy/engine/gradient_handler/_zero_gradient_handler.py
from colossalai.legacy.registry import GRADIENT_HANDLER from ._base_gradient_handler import BaseGradientHandler @GRADIENT_HANDLER.register_module class ZeROGradientHandler(BaseGradientHandler): """A helper class to handle all-reduce operations in a data parallel group. A all-reduce collective communication w...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/engine/gradient_handler/_moe_gradient_handler.py
colossalai/legacy/engine/gradient_handler/_moe_gradient_handler.py
from colossalai.context.moe_context import MOE_CONTEXT from colossalai.legacy.context.parallel_mode import ParallelMode from colossalai.legacy.core import global_context as gpc from colossalai.legacy.registry import GRADIENT_HANDLER from colossalai.utils.moe import get_moe_epsize_param_dict from ._base_gradient_handle...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/engine/gradient_handler/_sequence_parallel_gradient_handler.py
colossalai/legacy/engine/gradient_handler/_sequence_parallel_gradient_handler.py
from colossalai.legacy.context.parallel_mode import ParallelMode from colossalai.legacy.core import global_context as gpc from colossalai.legacy.registry import GRADIENT_HANDLER from ._base_gradient_handler import BaseGradientHandler from .utils import bucket_allreduce @GRADIENT_HANDLER.register_module class Sequenc...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/engine/gradient_handler/_pipeline_parallel_gradient_handler.py
colossalai/legacy/engine/gradient_handler/_pipeline_parallel_gradient_handler.py
#!/usr/bin/env python from collections import defaultdict import torch import torch.distributed as dist from torch._utils import _flatten_dense_tensors, _unflatten_dense_tensors from colossalai.legacy.core import global_context as gpc from colossalai.legacy.registry import GRADIENT_HANDLER from ._base_gradient_hand...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/engine/gradient_handler/utils.py
colossalai/legacy/engine/gradient_handler/utils.py
from typing import Iterable import torch.distributed as dist import torch.nn as nn from torch._utils import _flatten_dense_tensors, _unflatten_dense_tensors def bucket_allreduce(param_list: Iterable[nn.Parameter], group=None): # get communication world size comm_size = dist.get_world_size(group) # bucket...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/engine/gradient_handler/__init__.py
colossalai/legacy/engine/gradient_handler/__init__.py
from ._base_gradient_handler import BaseGradientHandler from ._data_parallel_gradient_handler import DataParallelGradientHandler from ._pipeline_parallel_gradient_handler import PipelineSharedModuleGradientHandler from ._sequence_parallel_gradient_handler import SequenceParallelGradientHandler from ._zero_gradient_hand...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/engine/gradient_handler/_data_parallel_gradient_handler.py
colossalai/legacy/engine/gradient_handler/_data_parallel_gradient_handler.py
from colossalai.legacy.context.parallel_mode import ParallelMode from colossalai.legacy.core import global_context as gpc from colossalai.legacy.registry import GRADIENT_HANDLER from ._base_gradient_handler import BaseGradientHandler from .utils import bucket_allreduce @GRADIENT_HANDLER.register_module class DataPar...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/engine/gradient_handler/_base_gradient_handler.py
colossalai/legacy/engine/gradient_handler/_base_gradient_handler.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- from abc import ABC, abstractmethod class BaseGradientHandler(ABC): """A basic helper class to handle all-reduce operations of gradients across different parallel groups before optimization. Args: model (Module): Model where the gradients accumulat...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/engine/schedule/_pipeline_schedule_v2.py
colossalai/legacy/engine/schedule/_pipeline_schedule_v2.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- from typing import Iterable, Tuple import torch.cuda import colossalai.legacy.communication.p2p_v2 as comm from colossalai.accelerator import get_accelerator from colossalai.legacy.context.parallel_mode import ParallelMode from colossalai.legacy.core import global_cont...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/engine/schedule/_pipeline_schedule.py
colossalai/legacy/engine/schedule/_pipeline_schedule.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- import inspect from typing import Callable, List, Tuple, Union import torch.cuda import colossalai.legacy.communication as comm from colossalai.accelerator import get_accelerator from colossalai.legacy.amp.naive_amp import NaiveAMPModel from colossalai.legacy.context.p...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/engine/schedule/_non_pipeline_schedule.py
colossalai/legacy/engine/schedule/_non_pipeline_schedule.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- import inspect from typing import Callable, Iterable import torch from colossalai.utils import conditional_context from ._base_schedule import BaseSchedule class NonPipelineSchedule(BaseSchedule): """A helper schedule class for no pipeline parallelism running en...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/engine/schedule/__init__.py
colossalai/legacy/engine/schedule/__init__.py
from ._base_schedule import BaseSchedule from ._non_pipeline_schedule import NonPipelineSchedule from ._pipeline_schedule import InterleavedPipelineSchedule, PipelineSchedule, get_tensor_shape __all__ = ["BaseSchedule", "NonPipelineSchedule", "PipelineSchedule", "InterleavedPipelineSchedule", "get_tensor_shape"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/engine/schedule/_base_schedule.py
colossalai/legacy/engine/schedule/_base_schedule.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- from abc import ABC, abstractmethod from typing import Callable, Iterable import torch from colossalai.accelerator import get_accelerator from colossalai.logging import get_dist_logger class BaseSchedule(ABC): """A basic helper class to control the process of tra...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/engine/gradient_accumulation/_gradient_accumulation.py
colossalai/legacy/engine/gradient_accumulation/_gradient_accumulation.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- from typing import Any, Iterable, Tuple, Union import torch.nn as nn from torch import Tensor from torch.nn.parallel.distributed import DistributedDataParallel from torch.optim import Optimizer from torch.optim.lr_scheduler import _LRScheduler from torch.utils.data impo...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/engine/gradient_accumulation/__init__.py
colossalai/legacy/engine/gradient_accumulation/__init__.py
from typing import Iterable, List import torch.nn as nn from torch.optim import Optimizer from torch.optim.lr_scheduler import _LRScheduler from colossalai.legacy.engine import BaseGradientHandler from ._gradient_accumulation import ( GradAccumDataloader, GradAccumGradientHandler, GradAccumLrSchedulerByS...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/async_engine.py
colossalai/legacy/inference/async_engine.py
import asyncio from colossalai.inference.dynamic_batching.ray_dist_init import Driver from .dynamic_batching.io_struct import RequestOutput from .dynamic_batching.sampling_params import SamplingParams class RequestTracker: """ A class for trace down all the requests, abstraction for async """ def _...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/async_manager.py
colossalai/legacy/inference/async_manager.py
from typing import List from .dynamic_batching.io_struct import Batch, Req, RequestOutput from .manager import DynamicBatchManager from .tensor_parallel import TPInferEngine class Async_DynamicBatchManager(DynamicBatchManager): def __init__( self, tp_engine: TPInferEngine, max_total_token...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/__init__.py
colossalai/legacy/inference/__init__.py
from .hybridengine import CaiInferEngine from .hybridengine.polices import LlamaModelInferPolicy __all__ = ["CaiInferEngine", "LlamaModelInferPolicy"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/manager.py
colossalai/legacy/inference/manager.py
# Adapted from https://github.com/ModelTC/lightllm import time from typing import List from .dynamic_batching.get_tokenizer import get_tokenizer from .dynamic_batching.infer_batch import InferBatch from .dynamic_batching.io_struct import Batch, Req from .dynamic_batching.req_queue import ReqQueue from .dynamic_batchi...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/quant/gptq/__init__.py
colossalai/legacy/inference/quant/gptq/__init__.py
from .cai_gptq import HAS_AUTO_GPTQ if HAS_AUTO_GPTQ: from .cai_gptq import CaiGPTQLinearOp, CaiQuantLinear
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/quant/gptq/cai_gptq/gptq_op.py
colossalai/legacy/inference/quant/gptq/cai_gptq/gptq_op.py
import torch from colossalai.kernel.triton import gptq_fused_linear_triton class CaiGPTQLinearOp(torch.nn.Module): def __init__(self, gptq_group_size, gptq_quant_bits): super(CaiGPTQLinearOp, self).__init__() self.group_size = gptq_group_size self.bits = gptq_quant_bits self.maxq ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/quant/gptq/cai_gptq/cai_quant_linear.py
colossalai/legacy/inference/quant/gptq/cai_gptq/cai_quant_linear.py
# Adapted from AutoGPTQ auto_gptq: https://github.com/PanQiWei/AutoGPTQ import math import warnings from typing import List, Union import numpy as np import torch import torch.distributed as dist import torch.nn as nn from torch.distributed import ProcessGroup from colossalai.lazy import LazyInitContext from colossa...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/quant/gptq/cai_gptq/__init__.py
colossalai/legacy/inference/quant/gptq/cai_gptq/__init__.py
import warnings HAS_AUTO_GPTQ = False try: import auto_gptq HAS_AUTO_GPTQ = True except ImportError: warnings.warn("please install auto-gptq from https://github.com/PanQiWei/AutoGPTQ") HAS_AUTO_GPTQ = False if HAS_AUTO_GPTQ: from .cai_quant_linear import CaiQuantLinear, ColCaiQuantLinear, RowCaiQ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/quant/smoothquant/__init__.py
colossalai/legacy/inference/quant/smoothquant/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/quant/smoothquant/models/base_model.py
colossalai/legacy/inference/quant/smoothquant/models/base_model.py
# Adapted from AutoGPTQ: https://github.com/PanQiWei/AutoGPTQ # Adapted from smoothquant: https://github.com/mit-han-lab/smoothquant/blob/main/smoothquant/calibration.py # Adapted from smoothquant: https://github.com/mit-han-lab/smoothquant/blob/main/smoothquant/smooth.py import os import warnings from abc import abst...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/quant/smoothquant/models/llama.py
colossalai/legacy/inference/quant/smoothquant/models/llama.py
import math import os import types from collections import defaultdict from functools import partial from typing import List, Optional, Tuple, Union import torch import torch.nn as nn import torch.nn.functional as F from torch_int.nn.bmm import BMM_S8T_S8N_F32T, BMM_S8T_S8N_S8T from transformers import PreTrainedModel...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/quant/smoothquant/models/__init__.py
colossalai/legacy/inference/quant/smoothquant/models/__init__.py
try: import torch_int HAS_TORCH_INT = True except ImportError: HAS_TORCH_INT = False raise ImportError( "Not install torch_int. Please install torch_int from https://github.com/Guangxuan-Xiao/torch-int" ) if HAS_TORCH_INT: from .llama import LLamaSmoothquantAttention, LlamaSmoothquantM...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/quant/smoothquant/models/linear.py
colossalai/legacy/inference/quant/smoothquant/models/linear.py
# modified from torch-int: https://github.com/Guangxuan-Xiao/torch-int/blob/main/torch_int/nn/linear.py import torch from torch_int._CUDA import linear_a8_w8_b8_o8, linear_a8_w8_bfp32_ofp32 from torch_int.functional.quantization import quantize_per_tensor_absmax try: from colossalai.kernel.op_builder.smoothquant ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/tensor_parallel/kvcache_manager.py
colossalai/legacy/inference/tensor_parallel/kvcache_manager.py
""" Refered/Modified from lightllm/common/mem_manager.py of the ModelTC/lightllm GitHub repository https://github.com/ModelTC/lightllm/blob/050af3ce65edca617e2f30ec2479397d5bb248c9/lightllm/common/mem_manager.py we slightly changed it to make it suitable for our colossal-ai shardformer TP-engine design. """ import tor...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/tensor_parallel/__init__.py
colossalai/legacy/inference/tensor_parallel/__init__.py
from .engine import TPInferEngine from .kvcache_manager import MemoryManager __all__ = ["MemoryManager", "TPInferEngine"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/tensor_parallel/batch_infer_state.py
colossalai/legacy/inference/tensor_parallel/batch_infer_state.py
# might want to consider combine with InferenceConfig in colossalai/ppinference/inference_config.py later from dataclasses import dataclass import torch from transformers.tokenization_utils_base import BatchEncoding from .kvcache_manager import MemoryManager # adapted from: lightllm/server/router/model_infer/infer_...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/tensor_parallel/engine.py
colossalai/legacy/inference/tensor_parallel/engine.py
from typing import Any, Callable, List, Optional, Union import torch import torch.nn as nn from transformers import BloomForCausalLM, LlamaForCausalLM from transformers.generation import GenerationConfig from transformers.generation.stopping_criteria import StoppingCriteriaList from transformers.tokenization_utils_bas...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/tensor_parallel/policies/llama.py
colossalai/legacy/inference/tensor_parallel/policies/llama.py
from functools import partial import torch from transformers.models.llama.modeling_llama import LlamaAttention, LlamaDecoderLayer, LlamaModel, LlamaRMSNorm from colossalai.shardformer.policies.base_policy import ModulePolicyDescription, SubModuleReplacementDescription # import colossalai from colossalai.shardformer....
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/tensor_parallel/policies/chatglm2.py
colossalai/legacy/inference/tensor_parallel/policies/chatglm2.py
from functools import partial from colossalai.shardformer.modeling.chatglm2_6b.modeling_chatglm import ( ChatGLMForConditionalGeneration, ChatGLMModel, GLMBlock, GLMTransformer, SelfAttention, ) # import colossalai from colossalai.shardformer.policies.chatglm2 import ChatGLMModelPolicy from ..mod...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/tensor_parallel/policies/bloom.py
colossalai/legacy/inference/tensor_parallel/policies/bloom.py
from functools import partial import torch from torch.nn import LayerNorm import colossalai.shardformer.layer as col_nn from colossalai.shardformer.policies.base_policy import ModulePolicyDescription, SubModuleReplacementDescription from colossalai.shardformer.policies.bloom import BloomForCausalLMPolicy from ..mode...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/tensor_parallel/policies/__init__.py
colossalai/legacy/inference/tensor_parallel/policies/__init__.py
from .bloom import BloomModelInferPolicy from .chatglm2 import ChatGLM2InferPolicy from .llama import LlamaModelInferPolicy __all__ = ["BloomModelInferPolicy", "LlamaModelInferPolicy", "ChatGLM2InferPolicy"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/tensor_parallel/modeling/llama.py
colossalai/legacy/inference/tensor_parallel/modeling/llama.py
import math from typing import List, Optional, Tuple import torch from transformers.modeling_outputs import BaseModelOutputWithPast from transformers.models.llama.modeling_llama import LlamaAttention, LlamaDecoderLayer, LlamaModel from colossalai.inference.tensor_parallel.batch_infer_state import BatchInferState from...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/tensor_parallel/modeling/chatglm2.py
colossalai/legacy/inference/tensor_parallel/modeling/chatglm2.py
import os from typing import Optional, Tuple import torch from torch.nn import CrossEntropyLoss from transformers.modeling_outputs import BaseModelOutputWithPast, CausalLMOutputWithPast from colossalai.inference.tensor_parallel.batch_infer_state import BatchInferState from colossalai.kernel.triton.token_attention_ker...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/tensor_parallel/modeling/bloom.py
colossalai/legacy/inference/tensor_parallel/modeling/bloom.py
import math import warnings from typing import Optional, Tuple, Union import torch import torch.distributed as dist from torch.nn import CrossEntropyLoss from torch.nn import functional as F from transformers.models.bloom.modeling_bloom import ( BaseModelOutputWithPastAndCrossAttentions, BloomAttention, Bl...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/tensor_parallel/modeling/__init__.py
colossalai/legacy/inference/tensor_parallel/modeling/__init__.py
from .bloom import BloomInferenceForwards from .chatglm2 import ChatGLM2InferenceForwards from .llama import LlamaInferenceForwards __all__ = ["BloomInferenceForwards", "LlamaInferenceForwards", "ChatGLM2InferenceForwards"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/tensor_parallel/modeling/_utils.py
colossalai/legacy/inference/tensor_parallel/modeling/_utils.py
""" Utils for model inference """ import os import torch from colossalai.kernel.triton.copy_kv_cache_dest import copy_kv_cache_to_dest def copy_kv_to_mem_cache(layer_id, key_buffer, value_buffer, context_mem_index, mem_manager): """ This function copies the key and value cache to the memory cache Args:...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/dynamic_batching/ray_init_config.py
colossalai/legacy/inference/dynamic_batching/ray_init_config.py
import logging import yaml from pydantic import BaseModel logger = logging.getLogger(__name__) class EngineArgsClass(BaseModel): """Config for Engine""" model: str tensor_parallel_size: int = 2 max_batch_size: int = 4 max_input_len: int = 128 max_output_len: int = 32 class RooterArgsClass...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/dynamic_batching/ray_dist_init.py
colossalai/legacy/inference/dynamic_batching/ray_dist_init.py
import logging import os from typing import List import ray import ray.util.collective as collective import torch from transformers import AutoModelForCausalLM import colossalai from colossalai.inference.async_manager import start_dynamic_batching from colossalai.inference.dynamic_batching.get_tokenizer import get_to...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/dynamic_batching/get_tokenizer.py
colossalai/legacy/inference/dynamic_batching/get_tokenizer.py
""" Motivated by VllM (https://github.com/vllm-project/vllm), This module is trying to resolve the tokenizer issue. license: MIT, see LICENSE for more details. """ from transformers import AutoTokenizer _FAST_LLAMA_TOKENIZER = "hf-internal-testing/llama-tokenizer" def get_tokenizer( tokenizer=None, tokeniz...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/dynamic_batching/__init__.py
colossalai/legacy/inference/dynamic_batching/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/dynamic_batching/infer_batch.py
colossalai/legacy/inference/dynamic_batching/infer_batch.py
# Adapted from https://github.com/ModelTC/lightllm import collections from dataclasses import dataclass from typing import Dict, List, Tuple import numpy as np import torch from colossalai.inference.tensor_parallel import MemoryManager # make batch infer state an attr of InferBatch class InferSamplingParams: d...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/dynamic_batching/io_struct.py
colossalai/legacy/inference/dynamic_batching/io_struct.py
# Adapted from https://github.com/ModelTC/lightllm from typing import Dict, List, Tuple from .sampling_params import SamplingParams class Req: def __init__(self, request_id, prompt_ids, sample_params: SamplingParams, prompts: str = ""): self.request_id = request_id self.prompt_ids = prompt_ids ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/dynamic_batching/stats.py
colossalai/legacy/inference/dynamic_batching/stats.py
# Adapted from https://github.com/ModelTC/lightllm import time class Stats: def __init__(self, log_status, log_stats_interval) -> None: self.log_stats = log_status self.log_stats_interval = log_stats_interval self.last_log_time = time.time() self.all_tokens = 0 self.output...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/dynamic_batching/sampling_params.py
colossalai/legacy/inference/dynamic_batching/sampling_params.py
# Adapted from https://github.com/ModelTC/lightllm """Sampling parameters for text generation.""" from typing import List, Optional, Union _SAMPLING_EPS = 1e-5 class SamplingParams: def __init__( self, do_sample: bool = False, presence_penalty: float = 0.0, frequency_penalty: flo...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/dynamic_batching/req_queue.py
colossalai/legacy/inference/dynamic_batching/req_queue.py
# Adapted from https://github.com/ModelTC/lightllm import uuid from typing import List import numpy as np from .io_struct import Batch, Req class ReqQueue: def __init__(self, max_total_tokens, batch_max_tokens, running_max_req_size, waiting_req_list=[]) -> None: self.max_total_tokens = max_total_tokens...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/serving/torch_serve/Colossal_Inference_Handler.py
colossalai/legacy/inference/serving/torch_serve/Colossal_Inference_Handler.py
import logging import os import zipfile from abc import ABC import torch import transformers from transformers import AutoTokenizer, BloomForCausalLM, BloomTokenizerFast, LlamaForCausalLM from ts.torch_handler.base_handler import BaseHandler import colossalai from colossalai.inference.tensor_parallel.engine import TP...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/serving/ray_serve/Colossal_Inference_rayserve.py
colossalai/legacy/inference/serving/ray_serve/Colossal_Inference_rayserve.py
import logging import os from typing import Any, List, Union import ray import ray.util.collective as collective import starlette import torch from pydantic import BaseModel from ray import serve from ray.serve import Application from transformers import AutoModelForCausalLM, AutoTokenizer import colossalai from colo...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/serving/ray_serve/send_requests.py
colossalai/legacy/inference/serving/ray_serve/send_requests.py
import ray import requests @ray.remote def send_query(text): resp = requests.get("http://localhost:8000/?text={}".format(text)) return resp.text test_sentences = [ "Introduce some landmarks in Beijing", "What is the weather today", "Coding requires practice and patience", "Rainy days inspire...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/serving/ray_serve/send_request.py
colossalai/legacy/inference/serving/ray_serve/send_request.py
import ray import requests @ray.remote def send_query(text): resp = requests.get("http://localhost:8000/?text={}".format(text)) return resp.text test_sentence = "Introduce some landmarks in Beijing" result = ray.get(send_query.remote(test_sentence)) print("Result returned:") print(result)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/hybridengine/__init__.py
colossalai/legacy/inference/hybridengine/__init__.py
from .engine import CaiInferEngine __all__ = ["CaiInferEngine"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/hybridengine/engine.py
colossalai/legacy/inference/hybridengine/engine.py
import torch import torch.distributed as dist import torch.nn as nn from transformers.tokenization_utils_base import BatchEncoding from colossalai.cluster import ProcessGroupMesh from colossalai.pipeline.schedule.generate import GenerateSchedule from colossalai.pipeline.stage_manager import PipelineStageManager from c...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/hybridengine/modeling/llama.py
colossalai/legacy/inference/hybridengine/modeling/llama.py
# This code is adapted from huggingface transformers: https://github.com/huggingface/transformers/blob/v4.34.1/src/transformers/models/llama/modeling_llama.py import math from typing import List, Optional, Tuple import torch from transformers.models.llama.modeling_llama import LlamaAttention, LlamaDecoderLayer, LlamaF...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/hybridengine/modeling/__init__.py
colossalai/legacy/inference/hybridengine/modeling/__init__.py
from .llama import LlamaInferenceForwards __all__ = ["LlamaInferenceForwards"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/legacy/inference/hybridengine/modeling/_utils.py
colossalai/legacy/inference/hybridengine/modeling/_utils.py
""" Utils for model inference """ import os import torch from colossalai.kernel.triton.copy_kv_cache_dest import copy_kv_cache_to_dest def copy_kv_to_mem_cache(layer_id, key_buffer, value_buffer, context_mem_index, mem_manager): """ This function copies the key and value cache to the memory cache Args:...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false