repo
stringlengths
7
90
file_url
stringlengths
81
315
file_path
stringlengths
4
228
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 14:38:15
2026-01-05 02:33:18
truncated
bool
2 classes
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_profiler/gpt_utils.py
tests/test_fx/test_profiler/gpt_utils.py
import torch.nn as nn from transformers import GPT2Config, GPT2LMHeadModel class GPTLMModel(nn.Module): def __init__( self, hidden_size=768, num_layers=12, num_attention_heads=12, max_seq_len=1024, vocab_size=50257, checkpoint=False, ): super()._...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_profiler/test_profiler_meta_info_prop.py
tests/test_fx/test_profiler/test_profiler_meta_info_prop.py
from typing import Tuple import torch import torch.fx import torchvision.models as tm from gpt_utils import gpt2_medium from torch.fx import symbolic_trace from colossalai.fx.passes.meta_info_prop import MetaInfoProp from colossalai.fx.profiler import calculate_fwd_out, calculate_fwd_tmp, is_compatible_with_meta, par...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_pipeline/test_timm_model/timm_utils.py
tests/test_fx/test_pipeline/test_timm_model/timm_utils.py
import inspect import random import numpy as np import torch from torch.fx import GraphModule from colossalai.fx import ColoTracer from colossalai.fx.passes.adding_split_node_pass import balanced_split_pass, split_with_split_nodes_pass MANUAL_SEED = 0 random.seed(MANUAL_SEED) np.random.seed(MANUAL_SEED) torch.manual...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_pipeline/test_timm_model/test_timm.py
tests/test_fx/test_pipeline/test_timm_model/test_timm.py
import pytest import timm.models as tm import torch from timm_utils import split_model_and_compare_output @pytest.mark.skip("balance split v2 is not ready") def test_timm_models_without_control_flow(): MODEL_LIST = [ tm.resnest.resnest50d, tm.beit.beit_base_patch16_224, tm.cait.cait_s24_22...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_pipeline/test_hf_model/test_t5.py
tests/test_fx/test_pipeline/test_hf_model/test_t5.py
import pytest import torch import transformers from hf_utils import split_model_and_compare_output BATCH_SIZE = 1 SEQ_LENGHT = 16 @pytest.mark.skip("balance split v2 is not ready") def test_t5(): MODEL_LIST = [ transformers.T5Model, transformers.T5ForConditionalGeneration, transformers.T5...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_pipeline/test_hf_model/test_albert.py
tests/test_fx/test_pipeline/test_hf_model/test_albert.py
import pytest import torch import transformers from hf_utils import split_model_and_compare_output BATCH_SIZE = 2 SEQ_LENGHT = 16 @pytest.mark.skip("balance split v2 is not ready") def test_single_sentence_albert(): MODEL_LIST = [ transformers.AlbertModel, transformers.AlbertForPreTraining, ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_pipeline/test_hf_model/hf_utils.py
tests/test_fx/test_pipeline/test_hf_model/hf_utils.py
import inspect import random import numpy as np import torch from torch.fx import GraphModule from colossalai.fx import ColoTracer from colossalai.fx.passes.adding_split_node_pass import balanced_split_pass, split_with_split_nodes_pass MANUAL_SEED = 0 random.seed(MANUAL_SEED) np.random.seed(MANUAL_SEED) torch.manual...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_pipeline/test_hf_model/test_opt.py
tests/test_fx/test_pipeline/test_hf_model/test_opt.py
import pytest import torch import transformers from hf_utils import split_model_and_compare_output BATCH_SIZE = 1 SEQ_LENGHT = 16 @pytest.mark.skip("balance split v2 is not ready") def test_opt(): MODEL_LIST = [ transformers.OPTModel, transformers.OPTForCausalLM, ] config = transformers....
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_pipeline/test_hf_model/test_gpt.py
tests/test_fx/test_pipeline/test_hf_model/test_gpt.py
import pytest import torch import transformers from hf_utils import split_model_and_compare_output BATCH_SIZE = 64 SEQ_LENGHT = 16 NUM_EPOCHS = 2 NUM_CHUNKS = 1 @pytest.mark.skip("balance split v2 is not ready") def test_gpt(): MODEL_LIST = [ transformers.GPT2Model, transformers.GPT2LMHeadModel, ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_pipeline/test_hf_model/test_bert.py
tests/test_fx/test_pipeline/test_hf_model/test_bert.py
import pytest import torch import transformers from hf_utils import split_model_and_compare_output BATCH_SIZE = 2 SEQ_LENGHT = 16 @pytest.mark.skip("balance split v2 is not ready") def test_single_sentence_bert(): MODEL_LIST = [ transformers.BertModel, transformers.BertForPreTraining, tra...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_pipeline/test_torchvision/test_torchvision.py
tests/test_fx/test_pipeline/test_torchvision/test_torchvision.py
import inspect import random import numpy as np import pytest import torch import torchvision import torchvision.models as tm from packaging import version from torch.fx import GraphModule from colossalai.fx import ColoTracer from colossalai.fx.passes.adding_split_node_pass import balanced_split_pass, split_with_spli...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_pipeline/test_topo/topo_utils.py
tests/test_fx/test_pipeline/test_topo/topo_utils.py
import random import numpy as np import torch from torch.fx import GraphModule from colossalai.fx import ColoTracer from colossalai.fx.passes.adding_split_node_pass import balanced_split_pass, split_with_split_nodes_pass from colossalai.legacy.pipeline.middleware import Partition, Topo from colossalai.legacy.pipeline...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_pipeline/test_topo/test_topo.py
tests/test_fx/test_pipeline/test_topo/test_topo.py
import pytest import torch import transformers from topo_utils import MLP, check_topo, split_model_and_get_DAG BATCH_SIZE = 1 SEQ_LENGHT = 16 @pytest.mark.skip("ShapeProp is not compatible with PyTorch 1.11.0") def test_opt(): MODEL_LIST = [ MLP, transformers.OPTModel, ] CONFIGS = [ ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_tracer/test_control_flow.py
tests/test_fx/test_tracer/test_control_flow.py
import torch import torch.nn as nn from torch.fx import GraphModule from colossalai.fx import ColoTracer as Tracer from colossalai.testing import clear_cache_before_run class ControlFlowModel(nn.Module): def __init__(self): super().__init__() self.linear1 = nn.Linear(10, 10) self.linear2 ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_tracer/test_functional_conv.py
tests/test_fx/test_tracer/test_functional_conv.py
import torch from torch.nn import functional as F from colossalai.fx.tracer.meta_patch import patched_function from colossalai.testing import clear_cache_before_run @clear_cache_before_run() def test_conv(): # test F.conv_1d data_1d = torch.rand(3, 16, 10) weight_1d = torch.rand(3, 16, 3) out_1d = F....
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_tracer/test_patched_module.py
tests/test_fx/test_tracer/test_patched_module.py
import torch from colossalai.fx.tracer.meta_patch import patched_module from colossalai.testing import clear_cache_before_run def _run(data, module, patch_fn): try: if isinstance(data, dict): output = patch_fn(module, **data) if isinstance(data, tuple) or isinstance(data, list): ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_tracer/test_patched_op.py
tests/test_fx/test_tracer/test_patched_op.py
from functools import partial import torch from colossalai.fx.tracer.meta_patch import patched_function from colossalai.testing import clear_cache_before_run def _run(data, patch_fn): try: output = patch_fn(data) return output except Exception as e: return e def _assert_output_shap...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_tracer/test_bias_addition_module.py
tests/test_fx/test_tracer/test_bias_addition_module.py
import torch from colossalai.fx import ColoGraphModule, ColoTracer from colossalai.testing import clear_cache_before_run class LinearModel(torch.nn.Module): def __init__(self, in_features, out_features): super().__init__() self.linear = torch.nn.Linear(in_features, out_features) def forward(...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_tracer/test_activation_checkpoint_annotation.py
tests/test_fx/test_tracer/test_activation_checkpoint_annotation.py
import torch from torch.fx import GraphModule from torch.utils.checkpoint import checkpoint from colossalai.fx import ColoTracer from colossalai.testing import clear_cache_before_run class MLP(torch.nn.Module): def __init__(self): super().__init__() self.linear1 = torch.nn.Linear(4, 4) se...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_tracer/test_timm_model/test_timm_model.py
tests/test_fx/test_tracer/test_timm_model/test_timm_model.py
import pytest import torch from packaging import version from colossalai._analyzer.fx import symbolic_trace from colossalai.testing import clear_cache_before_run from tests.kit.model_zoo import model_zoo def trace_and_compare(model_cls, data, output_transform_fn, meta_args=None): # trace model = model_cls() ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_tracer/test_torchvision_model/test_torchvision_model.py
tests/test_fx/test_tracer/test_torchvision_model/test_torchvision_model.py
import torch from colossalai._analyzer.fx import symbolic_trace from colossalai.testing import clear_cache_before_run from tests.kit.model_zoo import model_zoo @clear_cache_before_run() def test_torchvision_models(): torch.backends.cudnn.deterministic = True tv_sub_registry = model_zoo.get_sub_registry("torc...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_tracer/test_hf_model/test_hf_diffuser.py
tests/test_fx/test_tracer/test_hf_model/test_hf_diffuser.py
import pytest import torch from colossalai.fx import symbolic_trace from colossalai.testing import clear_cache_before_run from colossalai.testing.random import seed_all from tests.kit.model_zoo import model_zoo def assert_dict(da, db, assert_fn): assert len(da) == len(db) for k, v in da.items(): asse...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_tracer/test_hf_model/hf_tracer_utils.py
tests/test_fx/test_tracer/test_hf_model/hf_tracer_utils.py
from typing import List import torch # from colossalai.fx import symbolic_trace from colossalai._analyzer.fx import symbolic_trace def trace_model_and_compare_output(model, data_gen, ignore_data: List[str] = None): # must turn on eval mode to ensure the output is consistent model.eval() inputs = data_g...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_tracer/test_hf_model/test_hf_bert.py
tests/test_fx/test_tracer/test_hf_model/test_hf_bert.py
import pytest import torch from hf_tracer_utils import trace_model_and_compare_output from packaging import version from colossalai.testing import clear_cache_before_run from tests.kit.model_zoo import model_zoo @pytest.mark.skipif(version.parse(torch.__version__) < version.parse("1.12.0"), reason="torch version < 1...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_tracer/test_hf_model/test_hf_gpt.py
tests/test_fx/test_tracer/test_hf_model/test_hf_gpt.py
import pytest import torch from hf_tracer_utils import trace_model_and_compare_output from packaging import version from colossalai.testing import clear_cache_before_run from tests.kit.model_zoo import model_zoo @pytest.mark.skipif(version.parse(torch.__version__) < version.parse("1.12.0"), reason="torch version < 1...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_tracer/test_hf_model/test_hf_t5.py
tests/test_fx/test_tracer/test_hf_model/test_hf_t5.py
import pytest import torch from hf_tracer_utils import trace_model_and_compare_output from packaging import version from colossalai.testing import clear_cache_before_run from tests.kit.model_zoo import model_zoo @pytest.mark.skipif(version.parse(torch.__version__) < version.parse("1.12.0"), reason="torch version < 1...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_tracer/test_hf_model/test_hf_albert.py
tests/test_fx/test_tracer/test_hf_model/test_hf_albert.py
import pytest import torch from hf_tracer_utils import trace_model_and_compare_output from packaging import version from colossalai.testing import clear_cache_before_run from tests.kit.model_zoo import model_zoo BATCH_SIZE = 2 SEQ_LENGTH = 16 @pytest.mark.skipif(version.parse(torch.__version__) < version.parse("1.1...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_tracer/test_hf_model/test_hf_opt.py
tests/test_fx/test_tracer/test_hf_model/test_hf_opt.py
import pytest import torch from hf_tracer_utils import trace_model_and_compare_output from packaging import version from colossalai.testing import clear_cache_before_run from tests.kit.model_zoo import model_zoo @pytest.mark.skipif(version.parse(torch.__version__) < version.parse("1.12.0"), reason="torch version < 1...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_tracer/test_torchrec_model/test_deepfm_model.py
tests/test_fx/test_tracer/test_torchrec_model/test_deepfm_model.py
import torch from colossalai._analyzer.fx import symbolic_trace from colossalai.testing import clear_cache_before_run from tests.kit.model_zoo import model_zoo BATCH = 2 SHAPE = 10 def trace_and_compare(model_cls, data, output_transform_fn, meta_args=None): # trace model = model_cls() # convert to eval...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_tracer/test_torchrec_model/test_dlrm_model.py
tests/test_fx/test_tracer/test_torchrec_model/test_dlrm_model.py
import torch from colossalai._analyzer.fx import symbolic_trace from colossalai.testing import clear_cache_before_run from tests.kit.model_zoo import model_zoo BATCH = 2 SHAPE = 10 def trace_and_compare(model_cls, data, output_transform_fn, meta_args=None): # trace model = model_cls() # convert to eval...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_tracer/test_torchaudio_model/test_torchaudio_model.py
tests/test_fx/test_tracer/test_torchaudio_model/test_torchaudio_model.py
import pytest import torch from torchaudio_utils import trace_and_compare from colossalai.testing import clear_cache_before_run from tests.kit.model_zoo import model_zoo # We cannot handle the tensors constructed with constant during forward, such as ``torch.empty(0).to(device=Proxy.device)`` # TODO: We could handle...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_tracer/test_torchaudio_model/torchaudio_utils.py
tests/test_fx/test_tracer/test_torchaudio_model/torchaudio_utils.py
import torch from colossalai._analyzer.fx import symbolic_trace def trace_and_compare(model, data_gen, output_transform_fn, need_meta=False, need_concrete=False): data = data_gen() concrete_args = data if need_concrete else {} meta_args = {k: v.to("meta") for k, v in data.items()} if need_meta else {} ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_meta/test_backward.py
tests/test_fx/test_meta/test_backward.py
import pytest import timm.models as tmm import torch import torchvision.models as tm from colossalai.fx._compatibility import is_compatible_with_meta if is_compatible_with_meta(): from colossalai.fx.profiler import MetaTensor from colossalai.testing import clear_cache_before_run tm_models = [ tm.vgg11, ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_meta/test_aten.py
tests/test_fx/test_meta/test_aten.py
from typing import Any, Callable, Union import pytest import torch import torch.nn as nn from colossalai.fx._compatibility import is_compatible_with_meta from colossalai.testing import clear_cache_before_run if is_compatible_with_meta(): from colossalai.fx.profiler import MetaTensor aten = torch.ops.aten regis...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_meta/test_meta_trace.py
tests/test_fx/test_meta/test_meta_trace.py
import pytest import timm.models as tmm import torch import torchvision.models as tm from colossalai.fx._compatibility import is_compatible_with_meta if is_compatible_with_meta(): from colossalai.fx import meta_trace from colossalai.testing import clear_cache_before_run tm_models = [ tm.vgg11, tm.resnet...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_codegen/test_offload_codegen.py
tests/test_fx/test_codegen/test_offload_codegen.py
import copy import pytest import torch from torch.fx import GraphModule import colossalai from colossalai.fx import ColoTracer from colossalai.fx.graph_module import ColoGraphModule from colossalai.legacy.core import global_context as gpc from colossalai.testing import rerun_if_address_is_in_use, spawn try: from...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_codegen/test_nested_activation_checkpoint_codegen.py
tests/test_fx/test_codegen/test_nested_activation_checkpoint_codegen.py
import pytest import torch import colossalai from colossalai.fx import ColoTracer from colossalai.fx.graph_module import ColoGraphModule from colossalai.legacy.core import global_context as gpc from colossalai.testing import rerun_if_address_is_in_use, spawn try: from colossalai.fx.codegen import ActivationCheckp...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_codegen/test_activation_checkpoint_codegen.py
tests/test_fx/test_codegen/test_activation_checkpoint_codegen.py
import pytest import torch import torch.nn.functional as F from torch.utils.checkpoint import checkpoint import colossalai from colossalai.fx import ColoTracer from colossalai.fx.graph_module import ColoGraphModule from colossalai.legacy.core import global_context as gpc from colossalai.testing import rerun_if_address...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_lazy/lazy_init_utils.py
tests/test_lazy/lazy_init_utils.py
import random from copy import deepcopy from typing import Any, Callable, Optional, Tuple import numpy as np import torch from packaging import version from colossalai.device.device_mesh import DeviceMesh from colossalai.lazy.lazy_init import LazyInitContext, LazyTensor, _MyTensor from colossalai.tensor.d_tensor impo...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_lazy/test_models.py
tests/test_lazy/test_models.py
import pytest from lazy_init_utils import SUPPORT_LAZY, check_lazy_init from tests.kit.model_zoo import COMMON_MODELS, IS_FAST_TEST, model_zoo @pytest.mark.skipif(not SUPPORT_LAZY, reason="requires torch >= 1.12.0") @pytest.mark.parametrize( "subset", ( [COMMON_MODELS] if IS_FAST_TEST ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_lazy/test_ops.py
tests/test_lazy/test_ops.py
import copy import pytest import torch import torch.nn as nn from lazy_init_utils import SUPPORT_LAZY from torch.nn import Parameter from colossalai.lazy import LazyInitContext @pytest.mark.skipif(not SUPPORT_LAZY, reason="requires torch >= 1.12.0") def test_lazy_ops(): with LazyInitContext(): x = torch...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_lazy/test_from_pretrained.py
tests/test_lazy/test_from_pretrained.py
import os from transformers import BertForPreTraining, LlamaForCausalLM import colossalai.interface.pretrained as pretrained_utils from colossalai.lazy import LazyInitContext def test_lazy_from_pretrained(): # test from cached file, unsharded model = BertForPreTraining.from_pretrained("prajjwal1/bert-tiny")...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_checkpoint_io/test_safetensors_async_io.py
tests/test_checkpoint_io/test_safetensors_async_io.py
import tempfile import pytest import torch from safetensors.torch import load_file from colossalai.checkpoint_io.utils import create_pinned_state_dict from colossalai.testing import check_state_dict_equal, clear_cache_before_run from colossalai.utils import get_current_device from colossalai.utils.safetensors import ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_checkpoint_io/test_general_checkpoint_io.py
tests/test_checkpoint_io/test_general_checkpoint_io.py
import tempfile import pytest import torch from torch.optim import Adam from torchvision.models import resnet18 from colossalai.checkpoint_io import GeneralCheckpointIO from colossalai.nn.lr_scheduler import CosineAnnealingWarmupLR from colossalai.testing import check_state_dict_equal, clear_cache_before_run, paramet...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_checkpoint_io/test_plugins_huggingface_compatibility.py
tests/test_checkpoint_io/test_plugins_huggingface_compatibility.py
import pytest import torch import torch.distributed as dist from utils import shared_tempdir import colossalai from colossalai.booster import Booster from colossalai.booster.plugin import GeminiPlugin, LowLevelZeroPlugin, TorchDDPPlugin from colossalai.nn.optimizer import HybridAdam from colossalai.testing import ( ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_checkpoint_io/test_gemini_torch_compability.py
tests/test_checkpoint_io/test_gemini_torch_compability.py
import pytest import torch import torch.distributed as dist from torch.optim import Adam from utils import shared_tempdir import colossalai from colossalai.booster import Booster from colossalai.booster.plugin import GeminiPlugin, TorchDDPPlugin from colossalai.nn.optimizer import HybridAdam from colossalai.testing im...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_checkpoint_io/test_torch_ddp_checkpoint_io.py
tests/test_checkpoint_io/test_torch_ddp_checkpoint_io.py
import torch import torch.distributed as dist from torch.nn.parallel import DistributedDataParallel as DDP from torch.optim import SGD from torchvision.models import resnet18 from utils import shared_tempdir import colossalai from colossalai.booster import Booster from colossalai.booster.plugin import TorchDDPPlugin f...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_checkpoint_io/utils.py
tests/test_checkpoint_io/utils.py
import tempfile from contextlib import contextmanager, nullcontext from typing import Iterator import torch.distributed as dist @contextmanager def shared_tempdir() -> Iterator[str]: """ A temporary directory that is shared across all processes. """ ctx_fn = tempfile.TemporaryDirectory if dist.get_ra...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_checkpoint_io/test_gemini_checkpoint_io.py
tests/test_checkpoint_io/test_gemini_checkpoint_io.py
import os import pytest import torch import torch.distributed as dist from transformers import LlamaForCausalLM from utils import shared_tempdir import colossalai from colossalai.booster import Booster from colossalai.booster.plugin import GeminiPlugin from colossalai.lazy import LazyInitContext from colossalai.nn.op...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_checkpoint_io/test_hybrid_parallel_plugin_checkpoint_io.py
tests/test_checkpoint_io/test_hybrid_parallel_plugin_checkpoint_io.py
import pytest import torch import torch.distributed as dist from packaging.version import Version from torch.optim import Adam from utils import shared_tempdir import colossalai from colossalai.booster import Booster from colossalai.booster.plugin import HybridParallelPlugin from colossalai.shardformer.layer.utils imp...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_checkpoint_io/test_torch_fsdp_checkpoint_io.py
tests/test_checkpoint_io/test_torch_fsdp_checkpoint_io.py
import pytest import torch from packaging import version from torch.optim import SGD from torchvision.models import resnet18 from utils import shared_tempdir import colossalai from colossalai.booster import Booster if version.parse(torch.__version__) >= version.parse("1.12.0"): from colossalai.booster.plugin impo...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_checkpoint_io/test_low_level_zero_checkpoint_io.py
tests/test_checkpoint_io/test_low_level_zero_checkpoint_io.py
from copy import deepcopy from typing import Optional import torch import torch.distributed as dist from peft import LoraConfig from torchvision.models import resnet18 from utils import shared_tempdir import colossalai from colossalai.booster import Booster from colossalai.booster.plugin import LowLevelZeroPlugin fro...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_moe/test_kernel.py
tests/test_moe/test_kernel.py
import os import pytest import torch from colossalai.accelerator import get_accelerator from colossalai.moe._operation import MoeCombine, MoeDispatch, moe_cumsum NUM_EXPERTS = 4 BATCH_SIZE = 4 SEQ_LEN = 4 MOE_TENSOR_PATH = os.getenv("MOE_TENSOR_PATH") def check_equal(tensor_a, tensor_b, atol=1e-06): assert to...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_moe/moe_utils.py
tests/test_moe/moe_utils.py
import os import traceback from contextlib import contextmanager from time import sleep from typing import Callable, List, Optional import torch import torch.distributed as dist from torch.utils._pytree import tree_map def assert_loose_close(a, b, dtype: torch.dtype = torch.float32, name=""): assert loose_close(...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_moe/test_deepseek_layer.py
tests/test_moe/test_deepseek_layer.py
from copy import deepcopy import pytest import torch import torch.distributed as dist from torch.testing import assert_close from transformers import AutoConfig, AutoModel import colossalai from colossalai.booster.plugin.moe_hybrid_parallel_plugin import MoeHybridParallelPlugin from colossalai.shardformer.modeling.de...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_moe/test_mixtral_layer.py
tests/test_moe/test_mixtral_layer.py
from copy import deepcopy import pytest import torch import torch.distributed as dist from torch.testing import assert_close from transformers.models.mixtral.configuration_mixtral import MixtralConfig from transformers.models.mixtral.modeling_mixtral import MixtralSparseMoeBlock import colossalai from colossalai.boos...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_moe/test_moe_ep_zero.py
tests/test_moe/test_moe_ep_zero.py
from copy import deepcopy import pytest import torch import torch.distributed as dist from torch.nn.parallel import DistributedDataParallel as DDP from transformers.models.mixtral.configuration_mixtral import MixtralConfig from transformers.models.mixtral.modeling_mixtral import MixtralModel import colossalai from co...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_moe/test_moe_checkpoint.py
tests/test_moe/test_moe_checkpoint.py
import os import tempfile from contextlib import nullcontext from copy import deepcopy import pytest import torch import torch.distributed as dist from torch.optim import SGD, Adam from transformers.models.mixtral.configuration_mixtral import MixtralConfig from transformers.models.mixtral.modeling_mixtral import Mixtr...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_moe/test_moe_ep_tp.py
tests/test_moe/test_moe_ep_tp.py
from copy import deepcopy import pytest import torch import torch.distributed as dist from transformers.models.mixtral.configuration_mixtral import MixtralConfig from transformers.models.mixtral.modeling_mixtral import MixtralModel import colossalai from colossalai.booster.booster import Booster from colossalai.boost...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_device/test_extract_alpha_beta.py
tests/test_device/test_extract_alpha_beta.py
import pytest from colossalai.device import AlphaBetaProfiler from colossalai.initialize import launch from colossalai.logging import disable_existing_loggers from colossalai.testing import parameterize, rerun_if_address_is_in_use, spawn def check_extract_alpha_beta(rank, world_size, port, physical_devices): dis...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_device/test_device_mesh.py
tests/test_device/test_device_mesh.py
import pytest import torch import torch.distributed as dist import colossalai from colossalai.device.device_mesh import DeviceMesh from colossalai.testing import rerun_if_address_is_in_use, spawn def test_device_mesh(): physical_mesh_id = torch.arange(0, 16) mesh_shape = (4, 4) # [[0, 1, 2, 3], # [4...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_device/test_alpha_beta.py
tests/test_device/test_alpha_beta.py
import pytest from colossalai.device import AlphaBetaProfiler from colossalai.initialize import launch from colossalai.logging import disable_existing_loggers from colossalai.testing import parameterize, rerun_if_address_is_in_use, spawn def check_alpha_beta(rank, world_size, port, physical_devices): disable_exi...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_device/test_init_logical_pg.py
tests/test_device/test_init_logical_pg.py
import pytest import torch import torch.distributed as dist from torch.distributed import ReduceOp from colossalai.device.device_mesh import DeviceMesh from colossalai.initialize import launch from colossalai.testing import rerun_if_address_is_in_use, spawn def check_layer(rank, world_size, port): launch(rank=ra...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_device/test_search_logical_device_mesh.py
tests/test_device/test_search_logical_device_mesh.py
import pytest from colossalai.device import AlphaBetaProfiler from colossalai.initialize import launch from colossalai.logging import disable_existing_loggers from colossalai.testing import parameterize, rerun_if_address_is_in_use, spawn def check_alpha_beta(rank, world_size, port, physical_devices): disable_exi...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_tensor/test_mix_gather.py
tests/test_tensor/test_mix_gather.py
import pytest import torch import torch.distributed as dist from colossalai.device.device_mesh import DeviceMesh from colossalai.initialize import launch from colossalai.logging import disable_existing_loggers from colossalai.tensor.shape_consistency import CollectiveCommPattern, CommSpec from colossalai.tensor.shardi...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_tensor/test_sharding_spec.py
tests/test_tensor/test_sharding_spec.py
import torch from colossalai.device.device_mesh import DeviceMesh from colossalai.tensor.sharding_spec import ShardingSpec def test_sharding_spec(): physical_mesh_id = torch.arange(0, 16) mesh_shape = (4, 4) # [[0, 1, 2, 3], # [4, 5, 6, 7], # [8, 9, 10,11], # [12,13,14,15]] device_mesh...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_tensor/test_comm_spec_apply.py
tests/test_tensor/test_comm_spec_apply.py
import pytest import torch import torch.distributed as dist from colossalai.device.device_mesh import DeviceMesh from colossalai.initialize import launch from colossalai.logging import disable_existing_loggers from colossalai.tensor.shape_consistency import CollectiveCommPattern, CommSpec from colossalai.tensor.shardi...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_tensor/test_shape_consistency_apply.py
tests/test_tensor/test_shape_consistency_apply.py
import pytest import torch from colossalai.device.device_mesh import DeviceMesh from colossalai.initialize import launch from colossalai.logging import disable_existing_loggers from colossalai.tensor.shape_consistency import ShapeConsistencyManager from colossalai.tensor.sharding_spec import ShardingSpec from colossal...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_tensor/test_shape_consistency.py
tests/test_tensor/test_shape_consistency.py
import torch from colossalai.device.device_mesh import DeviceMesh from colossalai.tensor.shape_consistency import CollectiveCommPattern, ShapeConsistencyManager from colossalai.tensor.sharding_spec import ShardingSpec physical_mesh_id = torch.arange(0, 16) mesh_shape = (4, 4) # [[0, 1, 2, 3], # [4, 5, 6, 7], # [8, ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_tensor/test_padded_tensor.py
tests/test_tensor/test_padded_tensor.py
import torch from colossalai.device.device_mesh import DeviceMesh from colossalai.initialize import launch from colossalai.logging import disable_existing_loggers from colossalai.tensor.d_tensor import ShardingSpec, distribute_tensor, is_distributed_tensor, to_global from colossalai.tensor.padded_tensor import is_padd...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_tensor/test_dtensor/test_dtensor_sharding_spec.py
tests/test_tensor/test_dtensor/test_dtensor_sharding_spec.py
import operator from functools import reduce from colossalai.tensor.d_tensor.sharding_spec import ALLGATHER_COST, SHARD_COST, STEP_PENALTY, ShardingSpec def test_dtensor_sharding_spec(): dims = 4 dim_partition_dict_0 = {0: [0, 1]} # DistSpec: # shard_sequence: S01,R,R,R sharding_spec_0 = Shar...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_tensor/test_dtensor/test_comm_spec.py
tests/test_tensor/test_dtensor/test_comm_spec.py
import pytest import torch import torch.distributed as dist from colossalai.device.device_mesh import DeviceMesh from colossalai.initialize import launch from colossalai.logging import disable_existing_loggers from colossalai.tensor.d_tensor.comm_spec import CollectiveCommPattern, CommSpec from colossalai.testing impo...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_tensor/test_dtensor/test_layout_converter.py
tests/test_tensor/test_dtensor/test_layout_converter.py
import math import pytest import torch from colossalai.device.device_mesh import DeviceMesh from colossalai.initialize import launch from colossalai.logging import disable_existing_loggers from colossalai.tensor.d_tensor.comm_spec import CollectiveCommPattern from colossalai.tensor.d_tensor.layout import Layout from ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_tensor/test_dtensor/test_dtensor.py
tests/test_tensor/test_dtensor/test_dtensor.py
import torch from colossalai.device.device_mesh import DeviceMesh from colossalai.initialize import launch from colossalai.logging import disable_existing_loggers from colossalai.tensor.d_tensor import ShardingSpec, distribute_tensor, get_global_shape, redistribute, to_global from colossalai.testing import rerun_if_ad...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_booster/test_accelerator.py
tests/test_booster/test_accelerator.py
import torch.nn as nn from colossalai.booster.accelerator import Accelerator from colossalai.testing import clear_cache_before_run, parameterize @clear_cache_before_run() @parameterize("device", ["cpu", "cuda"]) def test_accelerator(device): accelerator = Accelerator(device) model = nn.Linear(8, 8) model...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_booster/test_plugin/test_torch_ddp_plugin.py
tests/test_booster/test_plugin/test_torch_ddp_plugin.py
from contextlib import nullcontext import torch import torch.distributed as dist import torch.nn as nn from torch.nn.parallel import DistributedDataParallel as DDP from torch.optim import SGD import colossalai from colossalai.booster import Booster from colossalai.booster.plugin import TorchDDPPlugin from colossalai....
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_booster/test_plugin/test_3d_plugin.py
tests/test_booster/test_plugin/test_3d_plugin.py
import copy from contextlib import nullcontext from typing import Optional import torch import torch.distributed as dist from torch.testing import assert_close from torch.utils.data import Dataset import colossalai from colossalai.accelerator import get_accelerator from colossalai.booster import Booster from colossal...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_booster/test_plugin/test_dp_plugin_base.py
tests/test_booster/test_plugin/test_dp_plugin_base.py
from typing import Callable, Dict, Iterator, List, Tuple, Union import torch import torch.distributed as dist import torch.nn as nn from torch.optim import Optimizer from torch.optim.lr_scheduler import _LRScheduler as LRScheduler from torch.utils.data import DataLoader, TensorDataset import colossalai from colossala...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_booster/test_plugin/test_low_level_zero_plugin.py
tests/test_booster/test_plugin/test_low_level_zero_plugin.py
from typing import Optional import torch import torch.distributed as dist from peft import LoraConfig from torch.optim import Adam import colossalai from colossalai.accelerator import get_accelerator from colossalai.booster import Booster from colossalai.booster.plugin import LowLevelZeroPlugin # from colossalai.nn....
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_booster/test_plugin/test_torch_fsdp_plugin.py
tests/test_booster/test_plugin/test_torch_fsdp_plugin.py
import pytest import torch from packaging import version from torch.optim import SGD import colossalai from colossalai.booster import Booster if version.parse(torch.__version__) >= version.parse("1.12.0"): from torch.distributed.fsdp import FullyShardedDataParallel as FSDP from colossalai.booster.plugin impor...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_booster/test_plugin/test_gemini_plugin.py
tests/test_booster/test_plugin/test_gemini_plugin.py
from contextlib import nullcontext from typing import Optional import torch import torch.distributed as dist import colossalai from colossalai.booster import Booster from colossalai.booster.plugin import GeminiPlugin from colossalai.fx import is_compatible_with_meta from colossalai.lazy.lazy_init import LazyInitConte...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_booster/test_mixed_precision/test_fp16_torch.py
tests/test_booster/test_mixed_precision/test_fp16_torch.py
import torch from torch.optim import Adam import colossalai from colossalai.booster.mixed_precision import FP16TorchMixedPrecision from colossalai.testing import rerun_if_address_is_in_use, spawn from tests.kit.model_zoo import model_zoo def run_torch_amp(rank, world_size, port): # init dist env colossalai.l...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_shardformer/test_with_torch_ddp.py
tests/test_shardformer/test_with_torch_ddp.py
from contextlib import nullcontext import pytest import torch import torch.distributed as dist from torch.nn.parallel import DistributedDataParallel as DDP import colossalai from colossalai.cluster import DistCoordinator from colossalai.lazy import LazyInitContext from colossalai.logging import disable_existing_logge...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_shardformer/test_flash_attention.py
tests/test_shardformer/test_flash_attention.py
import math from copy import copy import torch from torch.testing import assert_close from colossalai.kernel.kernel_loader import FlashAttentionLoader, FlashAttentionWithCustomMaskLoader from colossalai.shardformer.layer import AttnMaskType, ColoAttention from colossalai.shardformer.layer.attn import invert_mask from...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_shardformer/__init__.py
tests/test_shardformer/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_shardformer/test_shard_utils.py
tests/test_shardformer/test_shard_utils.py
import torch import torch.nn as nn from colossalai.shardformer.shard.utils import set_tensors_to_none class Net(nn.Module): def __init__(self) -> None: super().__init__() self.layers = nn.Sequential(nn.Linear(1, 2), nn.Linear(2, 3)) self.out = nn.Linear(3, 1) def test_release_layer(): ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_shardformer/test_layer/test_dropout.py
tests/test_shardformer/test_layer/test_dropout.py
import torch import torch.distributed as dist import torch.nn as nn import colossalai from colossalai.shardformer.layer import DropoutForParallelInput, DropoutForReplicatedInput from colossalai.testing import assert_equal, assert_not_equal, rerun_if_address_is_in_use, spawn def check_dropout_parallel_input(): dr...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_shardformer/test_layer/test_dist_crossentropy.py
tests/test_shardformer/test_layer/test_dist_crossentropy.py
import pytest import torch import torch.nn.functional as F import colossalai from colossalai.logging import disable_existing_loggers from colossalai.shardformer.layer import cross_entropy_1d from colossalai.testing import rerun_if_address_is_in_use, spawn CONFIG = dict( parallel=dict(data=1, pipeline=1, tensor=di...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_shardformer/test_layer/test_vocab_parallel_embedding_1d.py
tests/test_shardformer/test_layer/test_vocab_parallel_embedding_1d.py
from contextlib import nullcontext import torch import torch.distributed as dist import torch.nn as nn from torch.testing import assert_close import colossalai from colossalai.lazy import LazyInitContext from colossalai.shardformer.layer import VocabParallelEmbedding1D from colossalai.testing import parameterize, rer...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_shardformer/test_layer/test_gpt2_qkv_fused_linear_1d.py
tests/test_shardformer/test_layer/test_gpt2_qkv_fused_linear_1d.py
import os from contextlib import nullcontext import torch import torch.distributed as dist import torch.nn as nn from torch.testing import assert_close import colossalai from colossalai.lazy import LazyInitContext from colossalai.pipeline.weight_grad_store import WeightGradStore from colossalai.shardformer.layer impo...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_shardformer/test_layer/test_ring_attn.py
tests/test_shardformer/test_layer/test_ring_attn.py
import torch import torch.distributed as dist import torch.nn.functional as F from flash_attn import flash_attn_qkvpacked_func, flash_attn_varlen_qkvpacked_func from torch.testing import assert_close import colossalai from colossalai.cluster import ProcessGroupMesh from colossalai.shardformer.layer import AttnMaskType...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_shardformer/test_layer/test_embedding.py
tests/test_shardformer/test_layer/test_embedding.py
from contextlib import nullcontext import torch import torch.distributed as dist import torch.nn as nn from torch.testing import assert_close import colossalai from colossalai.lazy import LazyInitContext from colossalai.shardformer.layer import Embedding1D from colossalai.testing import parameterize, rerun_if_address...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_shardformer/test_layer/test_layernorm.py
tests/test_shardformer/test_layer/test_layernorm.py
from contextlib import nullcontext import torch import torch.nn as nn from torch.testing import assert_close import colossalai from colossalai.lazy import LazyInitContext from colossalai.shardformer.layer import FusedLayerNorm from colossalai.testing import parameterize, rerun_if_address_is_in_use, spawn @parameter...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_shardformer/test_layer/test_qkv_fused_linear_1d.py
tests/test_shardformer/test_layer/test_qkv_fused_linear_1d.py
import os from contextlib import nullcontext import torch import torch.nn as nn from torch.testing import assert_close import colossalai from colossalai.lazy import LazyInitContext from colossalai.shardformer.layer import FusedLinear, FusedLinear1D_Col, FusedLinear1D_Row from colossalai.shardformer.layer.qkv_fused_li...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_shardformer/test_layer/test_sequence_parallel.py
tests/test_shardformer/test_layer/test_sequence_parallel.py
import copy import torch import torch.distributed as dist import torch.nn as nn import torch.nn.functional as F from torch import Tensor from torch.testing import assert_close import colossalai from colossalai.shardformer.layer import all_to_all_comm from colossalai.testing import parameterize, rerun_if_address_is_in...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_shardformer/test_layer/test_linear_1d.py
tests/test_shardformer/test_layer/test_linear_1d.py
import os from contextlib import nullcontext import torch import torch.distributed as dist import torch.nn as nn from torch.testing import assert_close import colossalai from colossalai.lazy import LazyInitContext from colossalai.pipeline.weight_grad_store import WeightGradStore from colossalai.shardformer.layer impo...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_shardformer/test_layer/test_dist_log_prob.py
tests/test_shardformer/test_layer/test_dist_log_prob.py
import pytest import torch import colossalai from colossalai.logging import disable_existing_loggers from colossalai.shardformer.layer import dist_log_prob_1d from colossalai.testing import rerun_if_address_is_in_use, spawn CONFIG = dict( parallel=dict(data=1, pipeline=1, tensor=dict(size=2, mode="1d")), ) def ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_shardformer/test_hybrid_parallel_grad_clip_norm/test_amp_optimizer.py
tests/test_shardformer/test_hybrid_parallel_grad_clip_norm/test_amp_optimizer.py
import pytest import torch from torch.nn.utils.clip_grad import clip_grad_norm_ import colossalai from colossalai.logging import disable_existing_loggers from colossalai.shardformer.layer.utils import Randomizer from colossalai.tensor.d_tensor.api import clear_layout_converter from colossalai.testing import clear_cach...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_shardformer/test_hybrid_parallel_grad_clip_norm/test_naive_optimizer.py
tests/test_shardformer/test_hybrid_parallel_grad_clip_norm/test_naive_optimizer.py
import pytest import torch from torch.nn.utils.clip_grad import clip_grad_norm_ import colossalai from colossalai.logging import disable_existing_loggers from colossalai.shardformer.layer.utils import Randomizer from colossalai.tensor.d_tensor.api import clear_layout_converter from colossalai.testing import clear_cach...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_shardformer/test_hybrid_parallel_grad_clip_norm/test_zero_optimizer.py
tests/test_shardformer/test_hybrid_parallel_grad_clip_norm/test_zero_optimizer.py
import math import pytest import torch import torch.distributed as dist from torch.nn.utils.clip_grad import clip_grad_norm_ import colossalai from colossalai.logging import disable_existing_loggers from colossalai.shardformer.layer.utils import Randomizer from colossalai.tensor.d_tensor.api import clear_layout_conve...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false