repo
stringlengths
7
90
file_url
stringlengths
81
315
file_path
stringlengths
4
228
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 14:38:15
2026-01-05 02:33:18
truncated
bool
2 classes
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/__init__.py
colossalai/inference/__init__.py
from .config import InferenceConfig from .core import InferenceEngine __all__ = ["InferenceConfig", "InferenceEngine"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/batch_bucket.py
colossalai/inference/batch_bucket.py
from typing import Callable, List, Optional, Tuple, Union import torch from colossalai.inference.struct import Sequence from colossalai.utils import get_current_device class BatchBucket: """Container for a batch of Sequences, which is used to manage the batch of sequences. Attrs: _sequences_dict (D...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/kv_cache/kvcache_manager.py
colossalai/inference/kv_cache/kvcache_manager.py
from typing import List, Tuple import torch from transformers.configuration_utils import PretrainedConfig from colossalai.inference.config import InferenceConfig from colossalai.inference.struct import Sequence from colossalai.logging import get_dist_logger from colossalai.utils import get_current_device from .block...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/kv_cache/__init__.py
colossalai/inference/kv_cache/__init__.py
from .block_cache import CacheBlock from .kvcache_manager import KVCacheManager, RPCKVCacheManager __all__ = ["CacheBlock", "KVCacheManager", "RPCKVCacheManager"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/kv_cache/block_cache.py
colossalai/inference/kv_cache/block_cache.py
from typing import Any __all__ = ["CacheBlock"] class CacheBlock: """A simplified version of logical cache block used for Paged Attention.""" def __init__(self, block_id: int, block_size: int, elem_size: int, k_ptrs: Any = None, v_ptrs: Any = None): # Unique id of a cache block self.block_id...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/__init__.py
colossalai/inference/modeling/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/models/pixart_alpha.py
colossalai/inference/modeling/models/pixart_alpha.py
# Code adapted from: # https://github.com/huggingface/diffusers/blob/v0.29.0-release/src/diffusers/pipelines/pixart_alpha/pipeline_pixart_alpha.py from typing import Callable, List, Optional, Union import PIL.Image import torch from diffusers.pipelines.pixart_alpha.pipeline_pixart_alpha import ( ASPECT_RATIO_256_...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/models/stablediffusion3.py
colossalai/inference/modeling/models/stablediffusion3.py
# This code is adapted from huggingface diffusers: https://github.com/huggingface/diffusers/blob/v0.29.0-release/src/diffusers/pipelines/stable_diffusion_3/pipeline_stable_diffusion_3.py from typing import Any, Callable, Dict, List, Optional, Union import torch from diffusers.pipelines.stable_diffusion_3.pipeline_stab...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/models/__init__.py
colossalai/inference/modeling/models/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/models/glide_llama.py
colossalai/inference/modeling/models/glide_llama.py
# This is modified from huggingface transformers # https://github.com/huggingface/transformers/blob/v4.36.2/src/transformers/models/llama/modeling_llama.py import warnings from types import MethodType from typing import List, Optional, Tuple, Union import torch import torch.nn as nn from transformers.cache_utils impor...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/models/nopadding_baichuan.py
colossalai/inference/modeling/models/nopadding_baichuan.py
# This code is adapted from huggingface baichuan model: hhttps://huggingface.co/baichuan-inc/Baichuan2-13B-Base/blob/main/modeling_baichuan.py from typing import List, Optional, Tuple, Union import torch import torch.nn as nn from torch.distributed import ProcessGroup from colossalai.accelerator import get_accelerato...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/models/nopadding_llama.py
colossalai/inference/modeling/models/nopadding_llama.py
# This code is adapted from huggingface transformers: https://github.com/huggingface/transformers/blob/v4.34.1/src/transformers/models/llama/modeling_llama.py import itertools from typing import List, Optional, Tuple, Union import torch import torch.nn.functional as F from torch import nn from torch.distributed import...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/layers/baichuan_tp_linear.py
colossalai/inference/modeling/layers/baichuan_tp_linear.py
from typing import List, Union import torch.distributed as dist import torch.nn as nn from torch.distributed import ProcessGroup from colossalai.lazy import LazyInitContext from colossalai.shardformer.layer import Linear1D_Col from colossalai.shardformer.layer.parallel_module import ParallelModule class BaichuanLMH...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/layers/diffusion.py
colossalai/inference/modeling/layers/diffusion.py
import inspect import types import torch from torch import nn class DiffusionPipe(nn.Module): """ This Class convert a class of `DiffusionPipeline` into `nn.Module` and reserve most of origin attr,function and property. """ def __init__(self, source_obj) -> None: super(DiffusionPipe, self)._...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/layers/distrifusion.py
colossalai/inference/modeling/layers/distrifusion.py
# Code refer and adapted from: # https://github.com/huggingface/diffusers/blob/v0.29.0-release/src/diffusers # https://github.com/PipeFusion/PipeFusion import inspect from typing import Any, Dict, List, Optional, Tuple, Union import torch import torch.distributed as dist import torch.nn.functional as F from diffusers...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/layers/__init__.py
colossalai/inference/modeling/layers/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/layers/attention.py
colossalai/inference/modeling/layers/attention.py
import math import torch import torch.nn as nn import torch.nn.functional as F from transformers.modeling_attn_mask_utils import AttentionMaskConverter def copy_to_cache(source, cache, lengths, block_tables, type: str = "prefill"): """ Func: copy key/value into key/value cache. Args: key/value(source)...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/backends/__init__.py
colossalai/inference/modeling/backends/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/backends/attention_backend.py
colossalai/inference/modeling/backends/attention_backend.py
from abc import ABC, abstractmethod from dataclasses import dataclass import torch from colossalai.inference.config import ModelShardInferenceConfig from colossalai.kernel.kernel_loader import InferenceOpsLoader from colossalai.kernel.triton import context_attention_unpadded, flash_decoding_attention @dataclass cla...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/backends/pre_attention_backend.py
colossalai/inference/modeling/backends/pre_attention_backend.py
from abc import ABC, abstractmethod from colossalai.inference.config import ModelShardInferenceConfig from colossalai.inference.modeling.backends.attention_backend import AttentionMetaData from colossalai.kernel.kernel_loader import InferenceOpsLoader from colossalai.kernel.triton import copy_k_to_blocked_cache, decod...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/policy/pixart_alpha.py
colossalai/inference/modeling/policy/pixart_alpha.py
from diffusers.models.attention import BasicTransformerBlock from diffusers.models.transformers.pixart_transformer_2d import PixArtTransformer2DModel from torch import nn from colossalai.inference.config import RPC_PARAM from colossalai.inference.modeling.layers.diffusion import DiffusionPipe from colossalai.inference...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/policy/stablediffusion3.py
colossalai/inference/modeling/policy/stablediffusion3.py
from diffusers.models.attention import JointTransformerBlock from diffusers.models.transformers import SD3Transformer2DModel from torch import nn from colossalai.inference.config import RPC_PARAM from colossalai.inference.modeling.layers.diffusion import DiffusionPipe from colossalai.inference.modeling.layers.distrifu...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/policy/__init__.py
colossalai/inference/modeling/policy/__init__.py
from .glide_llama import GlideLlamaModelPolicy from .nopadding_baichuan import NoPaddingBaichuanModelInferPolicy from .nopadding_llama import NoPaddingLlamaModelInferPolicy from .pixart_alpha import PixArtAlphaInferPolicy from .stablediffusion3 import StableDiffusion3InferPolicy model_policy_map = { "nopadding_lla...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/policy/glide_llama.py
colossalai/inference/modeling/policy/glide_llama.py
from transformers.models.llama.modeling_llama import LlamaForCausalLM, LlamaModel from colossalai.inference.modeling.models.glide_llama import ( GlideLlamaDecoderLayer, glide_llama_causal_lm_forward, glide_llama_model_forward, ) from colossalai.inference.utils import init_to_get_rotary from colossalai.shar...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/policy/nopadding_baichuan.py
colossalai/inference/modeling/policy/nopadding_baichuan.py
from colossalai.inference.config import RPC_PARAM from colossalai.inference.modeling.layers.baichuan_tp_linear import BaichuanLMHeadLinear1D_Col from colossalai.inference.modeling.models.nopadding_baichuan import ( NopadBaichuanAttention, NopadBaichuanMLP, baichuan_rmsnorm_forward, ) from colossalai.inferen...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/modeling/policy/nopadding_llama.py
colossalai/inference/modeling/policy/nopadding_llama.py
from transformers.models.llama.modeling_llama import LlamaDecoderLayer, LlamaForCausalLM, LlamaModel, LlamaRMSNorm from colossalai.inference.config import RPC_PARAM from colossalai.inference.modeling.models.nopadding_llama import ( NopadLlamaAttention, NopadLlamaMLP, llama_causal_lm_forward, llama_deco...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/spec/struct.py
colossalai/inference/spec/struct.py
from dataclasses import dataclass from typing import Optional, Tuple import torch @dataclass class DrafterOutput: """ Dataclass for drafter model outputs. Args: speculated_length (int): Speculated length of the output sequence It is always less than or equal to spec_num during drafte...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/spec/drafter.py
colossalai/inference/spec/drafter.py
from typing import Optional, Tuple import torch import torch.nn as nn from transformers import PreTrainedTokenizer from transformers.cache_utils import DynamicCache from colossalai.utils import get_current_device from .struct import DrafterOutput, GlideInput class Drafter: """Container for the Drafter Model (A...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/spec/__init__.py
colossalai/inference/spec/__init__.py
from .drafter import Drafter from .struct import DrafterOutput, GlideInput __all__ = ["Drafter", "DrafterOutput", "GlideInput"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/executor/__init__.py
colossalai/inference/executor/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/executor/rpc_worker.py
colossalai/inference/executor/rpc_worker.py
from typing import List, Tuple, Union import rpyc import torch import torch.distributed as dist from torch import nn from transformers import AutoConfig, AutoModelForCausalLM from transformers.models.llama.modeling_llama import LlamaForCausalLM import colossalai from colossalai.accelerator import get_accelerator from...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/core/async_engine.py
colossalai/inference/core/async_engine.py
import asyncio import logging from functools import partial from typing import AsyncIterator, Dict, Iterable, List, Optional, Set, Tuple, Type from colossalai.inference.core.engine import InferenceEngine from colossalai.inference.sampler import search_tokens # CLI logger logging.basicConfig(level=logging.DEBUG, forma...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/core/llm_engine.py
colossalai/inference/core/llm_engine.py
import time from itertools import count from typing import Dict, List, Optional, Tuple, Type, Union import numpy as np import torch import torch.nn as nn from torch import distributed as dist from transformers import ( AutoConfig, AutoModelForCausalLM, GenerationConfig, PreTrainedTokenizer, PreTrai...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/core/plugin.py
colossalai/inference/core/plugin.py
import logging import os from functools import reduce from pathlib import Path from typing import Optional import torch from colossalai.checkpoint_io.general_checkpoint_io import GeneralCheckpointIO from colossalai.checkpoint_io.index_file import CheckpointIndexFile from colossalai.checkpoint_io.utils import is_safet...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/core/rpc_engine.py
colossalai/inference/core/rpc_engine.py
import asyncio from itertools import count from time import sleep from typing import List, Tuple, Union import rpyc import torch import torch.nn as nn from rpyc.utils.server import ThreadedServer from torch import multiprocessing as mp from transformers import AutoConfig, PreTrainedTokenizer, PreTrainedTokenizerFast f...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/core/request_handler.py
colossalai/inference/core/request_handler.py
from typing import Dict, List, Union import torch from transformers.configuration_utils import PretrainedConfig from transformers.generation import GenerationConfig from colossalai.inference.batch_bucket import BatchBucket from colossalai.inference.config import InferenceConfig from colossalai.inference.flash_decodin...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/core/__init__.py
colossalai/inference/core/__init__.py
from .engine import InferenceEngine from .request_handler import RequestHandler __all__ = ["InferenceEngine", "RequestHandler"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/core/base_engine.py
colossalai/inference/core/base_engine.py
from abc import ABC, abstractmethod import torch import torch.nn as nn from colossalai.cluster import ProcessGroupMesh from colossalai.inference.config import ModelShardInferenceConfig from colossalai.pipeline.stage_manager import PipelineStageManager from colossalai.shardformer import ShardConfig, ShardFormer from c...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/core/diffusion_engine.py
colossalai/inference/core/diffusion_engine.py
from itertools import count from typing import List, Tuple, Type, Union import numpy as np import PIL.Image import torch import torch.nn as nn from diffusers.pipelines.pipeline_utils import DiffusionPipeline from torch import distributed as dist from colossalai.accelerator import get_accelerator from colossalai.clust...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/core/engine.py
colossalai/inference/core/engine.py
from typing import List, Tuple, Type, Union import numpy as np import PIL.Image import torch.nn as nn from diffusers import DiffusionPipeline from transformers import PreTrainedTokenizer, PreTrainedTokenizerFast from colossalai.inference.config import InferenceConfig from colossalai.inference.utils import ModelType, ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/server/completion_service.py
colossalai/inference/server/completion_service.py
import asyncio from colossalai.inference.core.async_engine import AsyncInferenceEngine from .utils import id_generator class CompletionServing: def __init__(self, engine: AsyncInferenceEngine, served_model: str): self.engine = engine self.served_model = served_model try: asy...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/server/chat_service.py
colossalai/inference/server/chat_service.py
import asyncio import codecs import logging from fastapi import Request from colossalai.inference.core.async_engine import AsyncInferenceEngine from .utils import ChatCompletionResponseStreamChoice, ChatMessage, DeltaMessage, id_generator logger = logging.getLogger("colossalai-inference") class ChatServing: d...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/server/utils.py
colossalai/inference/server/utils.py
from typing import Any, Optional from pydantic import BaseModel # make it singleton class NumericIDGenerator: _instance = None def __new__(cls): if cls._instance is None: cls._instance = super(NumericIDGenerator, cls).__new__(cls) cls._instance.current_id = 0 return c...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/server/__init__.py
colossalai/inference/server/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/inference/server/api_server.py
colossalai/inference/server/api_server.py
""" Doc: Feature: - FastAPI based http server for Colossal-Inference - Completion Service Supported Usage: (for local user) - First, Lauch an API locally. `python3 -m colossalai.inference.server.api_server --model path of your llama2 model` - Second, you can turn to the page `http://127.0.0.1:8...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/__init__.py
colossalai/shardformer/__init__.py
from .shard import GradientCheckpointConfig, ModelSharder, PipelineGradientCheckpointConfig, ShardConfig, ShardFormer
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/_utils.py
colossalai/shardformer/_utils.py
import re def get_obj_list_element(obj, attr: str): r""" Get the element of the list in the object If the attr is a normal attribute, return the attribute of the object. If the attr is a index type, return the element of the index in the list, like `layers[0]`. Args: obj (Object): The ob...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/opt.py
colossalai/shardformer/policies/opt.py
import warnings from functools import partial from typing import Callable, Dict, List import torch.nn as nn from torch import Tensor, nn from colossalai.shardformer.layer import ( FusedLayerNorm, LayerNorm, Linear1D_Col, Linear1D_Row, LinearWithGradAccum, PaddingEmbedding, PaddingLMHead, ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/qwen2.py
colossalai/shardformer/policies/qwen2.py
from functools import partial from typing import Callable, Dict, List, Union import torch.nn as nn from torch import Tensor from torch.nn import Module from transformers.models.qwen2.modeling_qwen2 import ( Qwen2Attention, Qwen2DecoderLayer, Qwen2ForCausalLM, Qwen2ForSequenceClassification, Qwen2Mo...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/base_policy.py
colossalai/shardformer/policies/base_policy.py
# part of code modified from https://github.com/tunib-ai/parallelformers from abc import ABC, abstractmethod from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import torch.nn as nn from torch import Tensor from torch.nn import Module from colossalai.pipeline.stage_manage...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/auto_policy.py
colossalai/shardformer/policies/auto_policy.py
import importlib from dataclasses import dataclass import torch.nn as nn from .base_policy import Policy __all__ = ["PolicyLocation", "get_autopolicy", "import_policy"] @dataclass class PolicyLocation: """ PolicyLocation describes the location of a policy class. Args: file_name (str): The file...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/deepseek.py
colossalai/shardformer/policies/deepseek.py
from functools import partial from typing import Callable, Dict, List, Union import torch.nn as nn from torch import Tensor from torch.nn import Module from transformers.utils import is_flash_attn_greater_or_equal_2_10 from colossalai.shardformer.layer import FusedRMSNorm, Linear1D_Col, LinearWithGradAccum from colos...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/bert.py
colossalai/shardformer/policies/bert.py
import warnings from functools import partial from typing import Callable, Dict, List import torch.nn as nn from torch import Tensor from torch.nn import Module import colossalai.shardformer.layer as col_nn from ..modeling.bert import ( BertPipelineForwards, bert_sequence_parallel_forward_fn, get_bert_se...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/t5.py
colossalai/shardformer/policies/t5.py
from __future__ import annotations import warnings from functools import partial from typing import Callable, Dict, List, Tuple import numpy as np from torch import Tensor, nn from colossalai.shardformer.layer import ( DropoutForParallelInput, Embedding1D, FusedRMSNorm, Linear1D_Col, Linear1D_Row...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/whisper.py
colossalai/shardformer/policies/whisper.py
import warnings from functools import partial from typing import Callable, Dict, List, Tuple import numpy as np import torch.nn as nn from torch import Tensor import colossalai.shardformer.layer as col_nn from ..modeling.jit import get_jit_fused_dropout_add_func from ..modeling.whisper import ( WhisperPipelineFo...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/llama.py
colossalai/shardformer/policies/llama.py
from functools import partial from typing import Callable, Dict, List, Union import torch.nn as nn from torch import Tensor from torch.nn import Module from colossalai.shardformer.layer import ( FusedRMSNorm, Linear1D_Col, Linear1D_Row, LinearWithGradAccum, PaddingEmbedding, PaddingLMHead, ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/vit.py
colossalai/shardformer/policies/vit.py
import warnings from typing import Callable, Dict, List, Union import torch.nn as nn import colossalai.shardformer.layer as col_nn from colossalai.shardformer.layer import DropoutForReplicatedInput, Linear1D_Col from ..modeling.jit import get_jit_fused_dropout_add_func from ..modeling.vit import ( ViTForImageCla...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/falcon.py
colossalai/shardformer/policies/falcon.py
import warnings from functools import partial from typing import Callable, Dict, List from torch import Tensor, nn from torch.nn import Module import colossalai.shardformer.layer as col_nn from ..modeling.falcon import ( FalconPipelineForwards, build_falcon_alibi_tensor_fn, get_lm_forward_with_dist_cross...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/chatglm2.py
colossalai/shardformer/policies/chatglm2.py
import warnings from functools import partial from typing import Callable, Dict, List, Union import torch.nn as nn from torch import Tensor import colossalai.shardformer.layer as col_nn from colossalai.shardformer.modeling.chatglm2 import ChatGLMPipelineForwards from ..modeling.chatglm2 import ( get_chatglm_sequ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/qwen3.py
colossalai/shardformer/policies/qwen3.py
# Modifed from qwen2 policy from functools import partial from typing import Callable, Dict, List, Union import torch.nn as nn from torch import Tensor from torch.nn import Module from transformers.models.qwen3.modeling_qwen3 import ( Qwen3Attention, Qwen3DecoderLayer, Qwen3ForCausalLM, Qwen3ForSequenc...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/gptj.py
colossalai/shardformer/policies/gptj.py
import warnings from functools import partial from typing import Callable, Dict, List from torch import Tensor, nn import colossalai.shardformer.layer as col_nn from ..modeling.gptj import ( GPTJPipelineForwards, get_gptj_flash_attention_forward, gptj_model_forward_for_flash_attention, ) from .base_polic...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/bloom.py
colossalai/shardformer/policies/bloom.py
import warnings from functools import partial from typing import Callable, Dict, List import torch.nn as nn from torch import Tensor from torch.nn import Module import colossalai.shardformer.layer as col_nn from ..modeling.bloom import ( BloomPipelineForwards, build_bloom_alibi_tensor_fn, get_bloom_seque...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/command.py
colossalai/shardformer/policies/command.py
from functools import partial from typing import Callable, Dict, List, Union import torch.nn as nn from torch import Tensor from torch.nn import Module from colossalai.shardformer.layer import ( Linear1D_Col, Linear1D_Row, LinearWithGradAccum, PaddingEmbedding, PaddingLMHead, VocabParallelEmbe...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/deepseek_v3.py
colossalai/shardformer/policies/deepseek_v3.py
from functools import partial from typing import Callable, Dict, List, Union import torch.nn as nn from colossalai.shardformer.layer import FusedRMSNorm from colossalai.shardformer.modeling.deepseek_v3 import ( EpDeepseekV3MoE, deepseek_v3_for_causal_lm_forward, deepseek_v3_model_forward, ) from colossala...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/blip2.py
colossalai/shardformer/policies/blip2.py
import colossalai.shardformer.layer as col_nn from ..modeling.blip2 import ( forward_fn, get_blip2_flash_attention_forward, get_jit_fused_blip2_mlp_forward, get_jit_fused_blip2_QFormer_output_forward, get_jit_fused_blip2_QFormer_self_output_forward, ) from ..modeling.jit import get_jit_fused_dropou...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/gpt2.py
colossalai/shardformer/policies/gpt2.py
import warnings from functools import partial from typing import Callable, Dict, List from torch import Tensor, nn import colossalai.shardformer.layer as col_nn from ..modeling.gpt2 import GPT2PipelineForwards, get_gpt2_flash_attention_forward, get_jit_fused_gpt2_mlp_forward from .base_policy import ModulePolicyDesc...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/sam.py
colossalai/shardformer/policies/sam.py
import colossalai.shardformer.layer as col_nn from ..modeling.sam import forward_fn from .base_policy import ModulePolicyDescription, Policy, SubModuleReplacementDescription __all__ = ["SamPolicy", "SamModelPolicy"] class SamPolicy(Policy): def config_sanity_check(self): pass def preprocess(self): ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/__init__.py
colossalai/shardformer/policies/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/mixtral.py
colossalai/shardformer/policies/mixtral.py
import warnings from functools import partial from typing import Callable, Dict, List, Union import torch.nn as nn from torch import Tensor from torch.nn import Module from transformers.models.mixtral.modeling_mixtral import MixtralForCausalLM, MixtralModel from colossalai.shardformer.layer import ( FusedRMSNorm,...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/policies/mistral.py
colossalai/shardformer/policies/mistral.py
import warnings from functools import partial from typing import Callable, Dict, List, Union import torch.nn as nn from torch import Tensor from torch.nn import Module from colossalai.shardformer.layer import ( FusedRMSNorm, Linear1D_Col, Linear1D_Row, LinearWithGradAccum, PaddingEmbedding, Pa...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/layer/attn.py
colossalai/shardformer/layer/attn.py
from enum import Enum from typing import Callable, Dict, Optional, Tuple import torch import torch.distributed import torch.distributed as dist import torch.nn.functional as F from einops import rearrange from packaging import version from colossalai.kernel.kernel_loader import ( FlashAttentionDaoLoader, Flas...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/layer/_operation.py
colossalai/shardformer/layer/_operation.py
import functools import torch import torch.distributed as dist import torch.nn.functional as F from colossalai.pipeline.weight_grad_store import WeightGradStore from .utils import ( execute_conv1d_w_pass, execute_conv1d_w_pass_grad_accum, execute_w_pass, execute_w_pass_grad_accum, is_share_sp_tp,...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/layer/dropout.py
colossalai/shardformer/layer/dropout.py
from typing import List, Union import torch import torch.nn as nn from torch.distributed import ProcessGroup from .parallel_module import ParallelModule from .utils import create_randomizer_with_offset __all__ = ["DropoutForParallelInput", "DropoutForReplicatedInput"] class DropoutForParallelInput(ParallelModule, ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/layer/qkv_fused_linear.py
colossalai/shardformer/layer/qkv_fused_linear.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- import math from typing import Callable, List, Optional, Tuple, Union import torch import torch.distributed as dist import torch.nn as nn from torch import Tensor from torch.distributed import ProcessGroup from torch.nn.parameter import Parameter from colossalai.lazy i...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/layer/utils.py
colossalai/shardformer/layer/utils.py
from contextlib import contextmanager from typing import List, Optional, Union import torch import torch.distributed as dist from torch import nn from torch._utils import _flatten_dense_tensors, _unflatten_dense_tensors from torch.distributed import ProcessGroup, get_world_size from colossalai.accelerator import get_...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/layer/parallel_module.py
colossalai/shardformer/layer/parallel_module.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- import itertools from abc import ABC, abstractmethod from typing import List, Optional, Union import torch import torch.nn as nn from torch.distributed import ProcessGroup from torch.nn.modules.module import _EXTRA_STATE_KEY_SUFFIX, Module from colossalai.checkpoint_io...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/layer/loss.py
colossalai/shardformer/layer/loss.py
import torch import torch.distributed as dist from torch.autograd import Function from torch.distributed import ProcessGroup from torch.nn import CrossEntropyLoss from torch.nn.functional import log_softmax from colossalai.shardformer.layer._operation import reduce_forward from colossalai.shardformer.shard import Shar...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/layer/__init__.py
colossalai/shardformer/layer/__init__.py
from ._operation import all_to_all_comm from .attn import AttnMaskType, ColoAttention, RingAttention, get_pad_info from .dropout import DropoutForParallelInput, DropoutForReplicatedInput from .embedding import Embedding1D, PaddingEmbedding, VocabParallelEmbedding1D from .linear import Linear1D_Col, Linear1D_Row, Linear...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/layer/embedding.py
colossalai/shardformer/layer/embedding.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- from typing import Callable, List, Optional, Union import torch import torch.distributed as dist import torch.nn as nn import torch.nn.functional as F from torch import Tensor from torch.distributed import ProcessGroup from colossalai.lazy import LazyInitContext from c...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/layer/normalization.py
colossalai/shardformer/layer/normalization.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- import numbers import warnings from abc import ABC, abstractmethod import torch import torch.nn as nn from torch.nn import init from torch.nn.parameter import Parameter from colossalai.lazy import LazyInitContext from ._operation import hook_parameter_in_backward from ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/layer/linear.py
colossalai/shardformer/layer/linear.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- import math from typing import Callable, List, Optional, Tuple, Union import torch import torch.distributed as dist import torch.nn as nn import torch.nn.functional as F from torch import Tensor from torch.distributed import ProcessGroup from torch.nn.parameter import P...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/opt.py
colossalai/shardformer/modeling/opt.py
import random from typing import List, Optional, Tuple, Union import torch from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from transformers.modeling_attn_mask_utils import _prepare_4d_causal_attention_mask from transformers.modeling_outputs import ( BaseModelOutputWithPast, CausalLMOutputWit...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/qwen2.py
colossalai/shardformer/modeling/qwen2.py
import math from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from transformers.modeling_attn_mask_utils import ( _prepare_4d_causal_attention_mask, _prepare_4d_causal_attention_mask_for_sdpa, ) from transformers....
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/deepseek.py
colossalai/shardformer/modeling/deepseek.py
import warnings from typing import List, Optional, Tuple, Union import torch import torch.distributed as dist import torch.functional as F from torch.distributed import ProcessGroup from torch.nn import CrossEntropyLoss from transformers.cache_utils import Cache, DynamicCache from transformers.modeling_attn_mask_utils...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/bert.py
colossalai/shardformer/modeling/bert.py
import warnings from typing import List, Optional, Tuple, Union import torch from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from transformers.modeling_outputs import ( BaseModelOutputWithPoolingAndCrossAttentions, CausalLMOutputWithCrossAttentions, MaskedLMOutput, MultipleChoiceModel...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/t5.py
colossalai/shardformer/modeling/t5.py
import warnings from typing import Dict, List, Optional, Tuple, Union import torch from torch.nn import CrossEntropyLoss from transformers.modeling_outputs import ( BaseModelOutput, BaseModelOutputWithPastAndCrossAttentions, Seq2SeqLMOutput, Seq2SeqModelOutput, TokenClassifierOutput, ) from transfo...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/whisper.py
colossalai/shardformer/modeling/whisper.py
import logging import random from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from transformers.modeling_attn_mask_utils import ( _prepare_4d_causal_attention_mask, _prepare_4d_causal_attention_mask_for_sdpa, ) from transformers.modeling_ou...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/llama.py
colossalai/shardformer/modeling/llama.py
import math import warnings from typing import Dict, List, Optional, Tuple, Union import torch import torch.distributed import torch.utils.checkpoint from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from transformers.cache_utils import Cache, DynamicCache from transformers.modelin...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/vit.py
colossalai/shardformer/modeling/vit.py
from typing import List, Optional, Tuple, Union import torch from transformers.models.vit.modeling_vit import BaseModelOutput, ViTEncoder from transformers.utils import logging from colossalai.pipeline.stage_manager import PipelineStageManager from colossalai.shardformer.layer import ColoAttention def _encoder_forw...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/falcon.py
colossalai/shardformer/modeling/falcon.py
import warnings from typing import List, Optional, Tuple, Union import torch import torch.distributed as dist from torch.distributed import ProcessGroup from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from transformers.modeling_outputs import ( BaseModelOutputWithPastAndCrossAttentions, Causa...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/chatglm2.py
colossalai/shardformer/modeling/chatglm2.py
""" PyTorch ChatGLM model. """ from typing import List, Optional, Tuple import torch import torch.utils.checkpoint from transformers.modeling_outputs import BaseModelOutputWithPast, CausalLMOutputWithPast from transformers.utils import logging from colossalai.pipeline.stage_manager import PipelineStageManager from c...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/qwen3.py
colossalai/shardformer/modeling/qwen3.py
# Modifed from qwen2 modeling import math from typing import List, Optional, Tuple, Union import torch from torch import nn from transformers.modeling_attn_mask_utils import ( _prepare_4d_causal_attention_mask, _prepare_4d_causal_attention_mask_for_sdpa, ) from transformers.modeling_outputs import ( BaseMo...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/gptj.py
colossalai/shardformer/modeling/gptj.py
from typing import Dict, List, Optional, Tuple, Union import torch from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from transformers.cache_utils import Cache from transformers.modeling_outputs import ( BaseModelOutputWithPast, CausalLMOutputWithPast, QuestionAnsweringModelOutput, Sequ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/bloom.py
colossalai/shardformer/modeling/bloom.py
import warnings from typing import List, Optional, Tuple, Union import torch import torch.distributed as dist from torch.distributed import ProcessGroup from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from torch.nn import functional as F from transformers.cache_utils import Cache, DynamicCache from t...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/command.py
colossalai/shardformer/modeling/command.py
from typing import List, Optional, Tuple, Union import torch from torch import nn from transformers.cache_utils import Cache, DynamicCache from transformers.modeling_flash_attention_utils import FlashAttentionKwargs from transformers.modeling_outputs import BaseModelOutputWithPast, CausalLMOutputWithPast from transfor...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/deepseek_v3.py
colossalai/shardformer/modeling/deepseek_v3.py
from typing import List, Optional, Tuple, Union import numpy as np import torch import torch.distributed as dist from torch.distributed import ProcessGroup from torch.nn import CrossEntropyLoss from transformers.cache_utils import Cache, DynamicCache from transformers.modeling_attn_mask_utils import _prepare_4d_causal...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/jit.py
colossalai/shardformer/modeling/jit.py
import torch def get_dropout_add_func(): from transformers.models.bloom.modeling_bloom import dropout_add def self_dropout_add(self, x: torch.Tensor, residual: torch.Tensor, prob: float, training: bool) -> torch.Tensor: return dropout_add(x, residual, prob, training) return self_dropout_add de...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/blip2.py
colossalai/shardformer/modeling/blip2.py
from typing import Optional, Tuple import torch import torch.nn as nn from colossalai.shardformer.layer import ColoAttention def forward_fn(): def forward( self, hidden_states: torch.Tensor, head_mask: Optional[torch.Tensor] = None, output_attentions: Optional[bool] = False, ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/gpt2.py
colossalai/shardformer/modeling/gpt2.py
from typing import Dict, List, Optional, Tuple, Union import torch from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from transformers.modeling_outputs import ( BaseModelOutputWithPastAndCrossAttentions, CausalLMOutputWithCrossAttentions, QuestionAnsweringModelOutput, SequenceClassifier...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/sam.py
colossalai/shardformer/modeling/sam.py
import torch from torch import nn # Same as the SamVisionAttention forward method in the v4.51.3 transformers def forward_fn(): def forward(self, hidden_states: torch.Tensor, output_attentions=False) -> torch.Tensor: batch_size, height, width, _ = hidden_states.shape # qkv with shape (3, batch_siz...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false