repo
stringlengths
7
90
file_url
stringlengths
81
315
file_path
stringlengths
4
228
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 14:38:15
2026-01-05 02:33:18
truncated
bool
2 classes
psf/black
https://github.com/psf/black/blob/c3cc5a95d4f72e6ccc27ebae23344fce8cc70786/tests/data/gitignore_used_on_multiple_sources/dir1/a.py
tests/data/gitignore_used_on_multiple_sources/dir1/a.py
python
MIT
c3cc5a95d4f72e6ccc27ebae23344fce8cc70786
2026-01-04T14:40:23.735327Z
false
psf/black
https://github.com/psf/black/blob/c3cc5a95d4f72e6ccc27ebae23344fce8cc70786/action/main.py
action/main.py
import os import re import shlex import shutil import sys from pathlib import Path from subprocess import PIPE, STDOUT, run ACTION_PATH = Path(os.environ["GITHUB_ACTION_PATH"]) ENV_PATH = ACTION_PATH / ".black-env" ENV_BIN = ENV_PATH / ("Scripts" if sys.platform == "win32" else "bin") OPTIONS = os.getenv("INPUT_OPTION...
python
MIT
c3cc5a95d4f72e6ccc27ebae23344fce8cc70786
2026-01-04T14:40:23.735327Z
false
psf/black
https://github.com/psf/black/blob/c3cc5a95d4f72e6ccc27ebae23344fce8cc70786/profiling/list_huge.py
profiling/list_huge.py
config = some.Structure( value=set([u'some_rather_long_text_value', u'some_rather_long_text_value', u'some_rather_long_text_value', u'some_rather_long_text_value', u'some_rather_long_text_value', u'some_rather_long_text_value', u'some_rather_long_text_value', ...
python
MIT
c3cc5a95d4f72e6ccc27ebae23344fce8cc70786
2026-01-04T14:40:23.735327Z
true
psf/black
https://github.com/psf/black/blob/c3cc5a95d4f72e6ccc27ebae23344fce8cc70786/profiling/mix_big.py
profiling/mix_big.py
config = some.Structure( globalMap = { 103310322020340: [100000031211103,101042000320420,100100001202021,112320301100420,110101024402203,112001202000203,112101112010031,102130400200010,100401014300441,103000401422033], 110040120003212: [114413100031332,102101001412002,100210000032130,214000110100040...
python
MIT
c3cc5a95d4f72e6ccc27ebae23344fce8cc70786
2026-01-04T14:40:23.735327Z
true
psf/black
https://github.com/psf/black/blob/c3cc5a95d4f72e6ccc27ebae23344fce8cc70786/profiling/mix_huge.py
profiling/mix_huge.py
config = some.Structure( globalMap = { 103310322020340: [100000031211103,101042000320420,100100001202021,112320301100420,110101024402203,112001202000203,112101112010031,102130400200010,100401014300441,103000401422033], 110040120003212: [114413100031332,102101001412002,100210000032130,214000110100040...
python
MIT
c3cc5a95d4f72e6ccc27ebae23344fce8cc70786
2026-01-04T14:40:23.735327Z
true
psf/black
https://github.com/psf/black/blob/c3cc5a95d4f72e6ccc27ebae23344fce8cc70786/profiling/list_big.py
profiling/list_big.py
config = some.Structure( value=set([u'some_rather_long_text_value', u'some_rather_long_text_value', u'some_rather_long_text_value', u'some_rather_long_text_value', u'some_rather_long_text_value', u'some_rather_long_text_value', u'some_rather_long_text_value', ...
python
MIT
c3cc5a95d4f72e6ccc27ebae23344fce8cc70786
2026-01-04T14:40:23.735327Z
true
psf/black
https://github.com/psf/black/blob/c3cc5a95d4f72e6ccc27ebae23344fce8cc70786/profiling/dict_big.py
profiling/dict_big.py
config = some.Structure( some_mapping={ "00501": "AB890X", "00544": "AB890X", "01001": "AB889X", "01002": "AB889X", "01003": "AB889X", "01004": "AB889X", "01005": "AB889X", "01007": "AB889X", "01008": "AB889X", "01009": "AB889X", ...
python
MIT
c3cc5a95d4f72e6ccc27ebae23344fce8cc70786
2026-01-04T14:40:23.735327Z
true
psf/black
https://github.com/psf/black/blob/c3cc5a95d4f72e6ccc27ebae23344fce8cc70786/profiling/dict_huge.py
profiling/dict_huge.py
config = some.Structure( some_mapping={ "00501": "AB890X", "00544": "AB890X", "01001": "AB889X", "01002": "AB889X", "01003": "AB889X", "01004": "AB889X", "01005": "AB889X", "01007": "AB889X", "01008": "AB889X", "01009": "AB889X", ...
python
MIT
c3cc5a95d4f72e6ccc27ebae23344fce8cc70786
2026-01-04T14:40:23.735327Z
true
psf/black
https://github.com/psf/black/blob/c3cc5a95d4f72e6ccc27ebae23344fce8cc70786/profiling/mix_small.py
profiling/mix_small.py
config = some.Structure( globalMap = { 103310322020340: [100000031211103,101042000320420,100100001202021,112320301100420,110101024402203,112001202000203,112101112010031,102130400200010,100401014300441,103000401422033], 110040120003212: [114413100031332,102101001412002,100210000032130,214000110100040...
python
MIT
c3cc5a95d4f72e6ccc27ebae23344fce8cc70786
2026-01-04T14:40:23.735327Z
false
psf/black
https://github.com/psf/black/blob/c3cc5a95d4f72e6ccc27ebae23344fce8cc70786/docs/conf.py
docs/conf.py
# # Configuration file for the Sphinx documentation builder. # # This file does only contain a selection of the most common options. For a # full list see the documentation: # http://www.sphinx-doc.org/en/stable/config # -- Path setup -------------------------------------------------------------- # If extensions (or ...
python
MIT
c3cc5a95d4f72e6ccc27ebae23344fce8cc70786
2026-01-04T14:40:23.735327Z
false
zai-org/ChatGLM-6B
https://github.com/zai-org/ChatGLM-6B/blob/401bf3a8a7dd8a26fba189551dccfc61a7079b4e/web_demo_old.py
web_demo_old.py
from transformers import AutoModel, AutoTokenizer import gradio as gr tokenizer = AutoTokenizer.from_pretrained("THUDM/chatglm-6b", trust_remote_code=True) model = AutoModel.from_pretrained("THUDM/chatglm-6b", trust_remote_code=True).half().cuda() model = model.eval() MAX_TURNS = 20 MAX_BOXES = MAX_TURNS * 2 def pr...
python
Apache-2.0
401bf3a8a7dd8a26fba189551dccfc61a7079b4e
2026-01-04T14:40:23.749869Z
false
zai-org/ChatGLM-6B
https://github.com/zai-org/ChatGLM-6B/blob/401bf3a8a7dd8a26fba189551dccfc61a7079b4e/api.py
api.py
from fastapi import FastAPI, Request from transformers import AutoTokenizer, AutoModel import uvicorn, json, datetime import torch DEVICE = "cuda" DEVICE_ID = "0" CUDA_DEVICE = f"{DEVICE}:{DEVICE_ID}" if DEVICE_ID else DEVICE def torch_gc(): if torch.cuda.is_available(): with torch.cuda.device(CUDA_DEVIC...
python
Apache-2.0
401bf3a8a7dd8a26fba189551dccfc61a7079b4e
2026-01-04T14:40:23.749869Z
false
zai-org/ChatGLM-6B
https://github.com/zai-org/ChatGLM-6B/blob/401bf3a8a7dd8a26fba189551dccfc61a7079b4e/web_demo2.py
web_demo2.py
from transformers import AutoModel, AutoTokenizer import streamlit as st from streamlit_chat import message st.set_page_config( page_title="ChatGLM-6b 演示", page_icon=":robot:" ) @st.cache_resource def get_model(): tokenizer = AutoTokenizer.from_pretrained("THUDM/chatglm-6b", trust_remote_code=True) ...
python
Apache-2.0
401bf3a8a7dd8a26fba189551dccfc61a7079b4e
2026-01-04T14:40:23.749869Z
false
zai-org/ChatGLM-6B
https://github.com/zai-org/ChatGLM-6B/blob/401bf3a8a7dd8a26fba189551dccfc61a7079b4e/web_demo_vision.py
web_demo_vision.py
from transformers import AutoModel, AutoTokenizer import gradio as gr import mdtex2html tokenizer = AutoTokenizer.from_pretrained("THUDM/visualglm-6b", trust_remote_code=True) model = AutoModel.from_pretrained("THUDM/visualglm-6b", trust_remote_code=True).half().cuda() model = model.eval() """Override Chatbot.postpro...
python
Apache-2.0
401bf3a8a7dd8a26fba189551dccfc61a7079b4e
2026-01-04T14:40:23.749869Z
false
zai-org/ChatGLM-6B
https://github.com/zai-org/ChatGLM-6B/blob/401bf3a8a7dd8a26fba189551dccfc61a7079b4e/cli_demo_vision.py
cli_demo_vision.py
import os import platform import signal import sys from transformers import AutoTokenizer, AutoModel import readline tokenizer = AutoTokenizer.from_pretrained("THUDM/visualglm-6b", trust_remote_code=True) model = AutoModel.from_pretrained("THUDM/visualglm-6b", trust_remote_code=True).half().cuda() model = model.eval(...
python
Apache-2.0
401bf3a8a7dd8a26fba189551dccfc61a7079b4e
2026-01-04T14:40:23.749869Z
false
zai-org/ChatGLM-6B
https://github.com/zai-org/ChatGLM-6B/blob/401bf3a8a7dd8a26fba189551dccfc61a7079b4e/utils.py
utils.py
import os from typing import Dict, Tuple, Union, Optional from torch.nn import Module from transformers import AutoModel def auto_configure_device_map(num_gpus: int) -> Dict[str, int]: # transformer.word_embeddings 占用1层 # transformer.final_layernorm 和 lm_head 占用1层 # transformer.layers 占用 28 层 # 总共30层...
python
Apache-2.0
401bf3a8a7dd8a26fba189551dccfc61a7079b4e
2026-01-04T14:40:23.749869Z
false
zai-org/ChatGLM-6B
https://github.com/zai-org/ChatGLM-6B/blob/401bf3a8a7dd8a26fba189551dccfc61a7079b4e/web_demo.py
web_demo.py
from transformers import AutoModel, AutoTokenizer import gradio as gr import mdtex2html tokenizer = AutoTokenizer.from_pretrained("THUDM/chatglm-6b", trust_remote_code=True) model = AutoModel.from_pretrained("THUDM/chatglm-6b", trust_remote_code=True).half().cuda() model = model.eval() """Override Chatbot.postprocess...
python
Apache-2.0
401bf3a8a7dd8a26fba189551dccfc61a7079b4e
2026-01-04T14:40:23.749869Z
false
zai-org/ChatGLM-6B
https://github.com/zai-org/ChatGLM-6B/blob/401bf3a8a7dd8a26fba189551dccfc61a7079b4e/cli_demo.py
cli_demo.py
import os import platform import signal from transformers import AutoTokenizer, AutoModel import readline tokenizer = AutoTokenizer.from_pretrained("THUDM/chatglm-6b", trust_remote_code=True) model = AutoModel.from_pretrained("THUDM/chatglm-6b", trust_remote_code=True).half().cuda() model = model.eval() os_name = pla...
python
Apache-2.0
401bf3a8a7dd8a26fba189551dccfc61a7079b4e
2026-01-04T14:40:23.749869Z
false
zai-org/ChatGLM-6B
https://github.com/zai-org/ChatGLM-6B/blob/401bf3a8a7dd8a26fba189551dccfc61a7079b4e/ptuning/trainer_seq2seq.py
ptuning/trainer_seq2seq.py
# Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
python
Apache-2.0
401bf3a8a7dd8a26fba189551dccfc61a7079b4e
2026-01-04T14:40:23.749869Z
false
zai-org/ChatGLM-6B
https://github.com/zai-org/ChatGLM-6B/blob/401bf3a8a7dd8a26fba189551dccfc61a7079b4e/ptuning/arguments.py
ptuning/arguments.py
from dataclasses import dataclass, field from typing import Optional @dataclass class ModelArguments: """ Arguments pertaining to which model/config/tokenizer we are going to fine-tune from. """ model_name_or_path: str = field( metadata={"help": "Path to pretrained model or model identifier f...
python
Apache-2.0
401bf3a8a7dd8a26fba189551dccfc61a7079b4e
2026-01-04T14:40:23.749869Z
false
zai-org/ChatGLM-6B
https://github.com/zai-org/ChatGLM-6B/blob/401bf3a8a7dd8a26fba189551dccfc61a7079b4e/ptuning/main.py
ptuning/main.py
#!/usr/bin/env python # coding=utf-8 # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE...
python
Apache-2.0
401bf3a8a7dd8a26fba189551dccfc61a7079b4e
2026-01-04T14:40:23.749869Z
false
zai-org/ChatGLM-6B
https://github.com/zai-org/ChatGLM-6B/blob/401bf3a8a7dd8a26fba189551dccfc61a7079b4e/ptuning/trainer.py
ptuning/trainer.py
# coding=utf-8 # Copyright 2020-present the HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by ap...
python
Apache-2.0
401bf3a8a7dd8a26fba189551dccfc61a7079b4e
2026-01-04T14:40:23.749869Z
true
zai-org/ChatGLM-6B
https://github.com/zai-org/ChatGLM-6B/blob/401bf3a8a7dd8a26fba189551dccfc61a7079b4e/ptuning/web_demo.py
ptuning/web_demo.py
import os, sys import gradio as gr import mdtex2html import torch import transformers from transformers import ( AutoConfig, AutoModel, AutoTokenizer, AutoTokenizer, DataCollatorForSeq2Seq, HfArgumentParser, Seq2SeqTrainingArguments, set_seed, ) from arguments import ModelArguments, D...
python
Apache-2.0
401bf3a8a7dd8a26fba189551dccfc61a7079b4e
2026-01-04T14:40:23.749869Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/setup.py
setup.py
import os import sys from typing import List from setuptools import find_packages, setup try: import torch # noqa from torch.utils.cpp_extension import BuildExtension TORCH_AVAILABLE = True except ImportError: TORCH_AVAILABLE = False THIS_DIR = os.path.dirname(os.path.abspath(__file__)) BUILD_EXT =...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/.github/workflows/scripts/generate_leaderboard_and_send_to_lark.py
.github/workflows/scripts/generate_leaderboard_and_send_to_lark.py
import os from datetime import datetime, timedelta from typing import Any, Dict, List import matplotlib.pyplot as plt import pytz import requests import seaborn from requests_toolbelt import MultipartEncoder class Counter(dict): """ Dataclass for a github contributor. Args: name (str): name of t...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/.github/workflows/scripts/generate_release_draft.py
.github/workflows/scripts/generate_release_draft.py
#!/usr/bin/env python # coding: utf-8 import argparse import os import re import requests COMMIT_API = "https://api.github.com/repos/hpcaitech/ColossalAI/commits" TAGS_API = "https://api.github.com/repos/hpcaitech/ColossalAI/tags" def parse_args(): parser = argparse.ArgumentParser() parser.add_argument("--...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/.github/workflows/scripts/check_doc_i18n.py
.github/workflows/scripts/check_doc_i18n.py
import argparse import os def compare_dirs(dir1, dir2): # First, we need to check if the two directories exist if not os.path.exists(dir1) or not os.path.exists(dir2): return False # Now, we compare the list of items in each directory items1 = os.listdir(dir1) items2 = os.listdir(dir2) ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/.github/workflows/scripts/update_setup_for_nightly.py
.github/workflows/scripts/update_setup_for_nightly.py
from datetime import datetime def open_setup_file(): with open("setup.py", "r") as f: file_lines = f.readlines() return file_lines def replace_nightly_package_info(file_lines): version = datetime.today().strftime("%Y.%m.%d") package_name = "colossalai-nightly" for idx, line in enumerate...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/.github/workflows/scripts/send_message_to_lark.py
.github/workflows/scripts/send_message_to_lark.py
import argparse import requests def parse_args(): parser = argparse.ArgumentParser() parser.add_argument("-m", "--message", type=str) parser.add_argument("-u", "--url", type=str) return parser.parse_args() def send_message_to_lark(message, webhook_url): data = {"msg_type": "text", "content": {"...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/.github/workflows/scripts/example_checks/detect_changed_example.py
.github/workflows/scripts/example_checks/detect_changed_example.py
import argparse def main(): parser = argparse.ArgumentParser() parser.add_argument("-f", "--fileNameList", type=str, help="The list of changed files") args = parser.parse_args() name_list = args.fileNameList.split(":") folder_need_check = set() for loc in name_list: # Find only the sub...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/.github/workflows/scripts/example_checks/check_example_weekly.py
.github/workflows/scripts/example_checks/check_example_weekly.py
import os def show_files(path, all_files): # Traverse all the folder/file in current directory file_list = os.listdir(path) # Determine the element is folder or file. If file, pass it into list, if folder, recurse. for file_name in file_list: # Get the abs directory using os.path.join() and st...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/.github/workflows/scripts/example_checks/check_dispatch_inputs.py
.github/workflows/scripts/example_checks/check_dispatch_inputs.py
import argparse import os def check_inputs(input_list): for path in input_list: real_path = os.path.join("examples", path) if not os.path.exists(real_path): return False return True def main(): parser = argparse.ArgumentParser() parser.add_argument("-f", "--fileNameList",...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/cuda_extension.py
extensions/cuda_extension.py
import os import time from abc import abstractmethod from pathlib import Path from typing import List from .base_extension import _Extension from .cpp_extension import _CppExtension from .utils import check_pytorch_version, check_system_pytorch_cuda_match, set_cuda_arch_list __all__ = ["_CudaExtension"] # Some const...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/cpp_extension.py
extensions/cpp_extension.py
import importlib import os import time from abc import abstractmethod from pathlib import Path from typing import List from .base_extension import _Extension __all__ = ["_CppExtension"] class _CppExtension(_Extension): def __init__(self, name: str, priority: int = 1): super().__init__(name, support_aot=...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/base_extension.py
extensions/base_extension.py
import hashlib import os from abc import ABC, abstractmethod from typing import Callable, Union __all__ = ["_Extension"] class _Extension(ABC): def __init__(self, name: str, support_aot: bool, support_jit: bool, priority: int = 1): self._name = name self._support_aot = support_aot self._s...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/utils.py
extensions/utils.py
import os import re import subprocess import warnings from typing import List def print_rank_0(message: str) -> None: """ Print on only one process to avoid spamming. """ try: import torch.distributed as dist if not dist.is_initialized(): is_main_rank = True else: ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/__init__.py
extensions/__init__.py
from .pybind.cpu_adam import CpuAdamArmExtension, CpuAdamX86Extension from .pybind.flash_attention import ( FlashAttentionDaoCudaExtension, FlashAttentionNpuExtension, FlashAttentionSdpaCudaExtension, ) from .pybind.inference import InferenceOpsCudaExtension from .pybind.layernorm import LayerNormCudaExtens...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/triton_extension.py
extensions/triton_extension.py
from .base_extension import _Extension __all__ = ["_TritonExtension"] class _TritonExtension(_Extension): def __init__(self, name: str, priority: int = 1): super().__init__(name, support_aot=False, support_jit=True, priority=priority) def is_hardware_compatible(self) -> bool: # cuda extensio...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/csrc/__init__.py
extensions/csrc/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/pybind/__init__.py
extensions/pybind/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/pybind/softmax/scaled_masked_softmax_cuda.py
extensions/pybind/softmax/scaled_masked_softmax_cuda.py
from ...cuda_extension import _CudaExtension from ...utils import append_nvcc_threads class ScaledMaskedSoftmaxCudaExtension(_CudaExtension): def __init__(self): super().__init__(name="scaled_masked_softmax_cuda") def sources_files(self): ret = [self.csrc_abs_path(fname) for fname in ["kernel...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/pybind/softmax/__init__.py
extensions/pybind/softmax/__init__.py
from .scaled_masked_softmax_cuda import ScaledMaskedSoftmaxCudaExtension from .scaled_upper_triangle_masked_softmax_cuda import ScaledUpperTriangleMaskedSoftmaxCudaExtension __all__ = ["ScaledMaskedSoftmaxCudaExtension", "ScaledUpperTriangleMaskedSoftmaxCudaExtension"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/pybind/softmax/scaled_upper_triangle_masked_softmax_cuda.py
extensions/pybind/softmax/scaled_upper_triangle_masked_softmax_cuda.py
from ...cuda_extension import _CudaExtension from ...utils import append_nvcc_threads, get_cuda_cc_flag class ScaledUpperTriangleMaskedSoftmaxCudaExtension(_CudaExtension): def __init__(self): super().__init__(name="scaled_upper_triangle_masked_softmax_cuda") def sources_files(self): ret = [ ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/pybind/cpu_adam/cpu_adam_x86.py
extensions/pybind/cpu_adam/cpu_adam_x86.py
import platform from ...cuda_extension import _CudaExtension from ...utils import append_nvcc_threads class CpuAdamX86Extension(_CudaExtension): def __init__(self): super().__init__(name="cpu_adam_x86") def is_available(self) -> bool: return platform.machine() == "x86_64" and super().is_avai...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/pybind/cpu_adam/__init__.py
extensions/pybind/cpu_adam/__init__.py
from .cpu_adam_arm import CpuAdamArmExtension from .cpu_adam_x86 import CpuAdamX86Extension __all__ = ["CpuAdamArmExtension", "CpuAdamX86Extension"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/pybind/cpu_adam/cpu_adam_arm.py
extensions/pybind/cpu_adam/cpu_adam_arm.py
import platform from typing import List from ...cpp_extension import _CppExtension class CpuAdamArmExtension(_CppExtension): def __init__(self): super().__init__(name="cpu_adam_arm") def is_available(self) -> bool: # only arm allowed return platform.machine() == "aarch64" def as...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/pybind/layernorm/layernorm_cuda.py
extensions/pybind/layernorm/layernorm_cuda.py
from ...cuda_extension import _CudaExtension from ...utils import append_nvcc_threads, get_cuda_cc_flag class LayerNormCudaExtension(_CudaExtension): def __init__(self): super().__init__(name="layernorm_cuda") def sources_files(self): ret = [self.csrc_abs_path(fname) for fname in ["kernel/cud...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/pybind/layernorm/__init__.py
extensions/pybind/layernorm/__init__.py
from .layernorm_cuda import LayerNormCudaExtension __all__ = ["LayerNormCudaExtension"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/pybind/flash_attention/flash_attention_sdpa_cuda.py
extensions/pybind/flash_attention/flash_attention_sdpa_cuda.py
from ...base_extension import _Extension class FlashAttentionSdpaCudaExtension(_Extension): def __init__(self): super().__init__(name="flash_attention_sdpa_cuda", support_aot=False, support_jit=False) def is_available(self) -> bool: # cuda extension can only be built if cuda is available ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/pybind/flash_attention/flash_attention_dao_cuda.py
extensions/pybind/flash_attention/flash_attention_dao_cuda.py
from ...base_extension import _Extension class FlashAttentionDaoCudaExtension(_Extension): def __init__(self): super().__init__(name="flash_attention_dao_cuda", support_aot=False, support_jit=False, priority=10) def is_available(self) -> bool: # cuda extension can only be built if cuda is ava...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/pybind/flash_attention/flash_attention_npu.py
extensions/pybind/flash_attention/flash_attention_npu.py
import math from ...base_extension import _Extension class FlashAttentionNpuExtension(_Extension): def __init__(self): super().__init__(name="flash_attention_npu", support_aot=False, support_jit=False) def is_available(self) -> bool: try: import torch_npu return hasa...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/pybind/flash_attention/__init__.py
extensions/pybind/flash_attention/__init__.py
from .flash_attention_dao_cuda import FlashAttentionDaoCudaExtension from .flash_attention_npu import FlashAttentionNpuExtension from .flash_attention_sdpa_cuda import FlashAttentionSdpaCudaExtension try: # TODO: remove this after updating openmoe example import flash_attention # noqa HAS_FLASH_ATTN = Tr...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/pybind/inference/inference_ops_cuda.py
extensions/pybind/inference/inference_ops_cuda.py
from ...cuda_extension import _CudaExtension from ...utils import get_cuda_cc_flag class InferenceOpsCudaExtension(_CudaExtension): def __init__(self): super().__init__(name="inference_ops_cuda") def sources_files(self): ret = [ self.csrc_abs_path(fname) for fname in [...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/pybind/inference/__init__.py
extensions/pybind/inference/__init__.py
from .inference_ops_cuda import InferenceOpsCudaExtension __all__ = ["InferenceOpsCudaExtension"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/pybind/optimizer/fused_optimizer_cuda.py
extensions/pybind/optimizer/fused_optimizer_cuda.py
from ...cuda_extension import _CudaExtension from ...utils import get_cuda_cc_flag class FusedOptimizerCudaExtension(_CudaExtension): def __init__(self): super().__init__(name="fused_optim_cuda") def sources_files(self): ret = [ self.csrc_abs_path(fname) for fname in [...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/pybind/optimizer/__init__.py
extensions/pybind/optimizer/__init__.py
from .fused_optimizer_cuda import FusedOptimizerCudaExtension __all__ = ["FusedOptimizerCudaExtension"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/pybind/moe/moe_cuda.py
extensions/pybind/moe/moe_cuda.py
from ...cuda_extension import _CudaExtension from ...utils import append_nvcc_threads, get_cuda_cc_flag class MoeCudaExtension(_CudaExtension): def __init__(self): super().__init__(name="moe_cuda") def sources_files(self): ret = [self.csrc_abs_path(fname) for fname in ["kernel/cuda/moe_kernel...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/extensions/pybind/moe/__init__.py
extensions/pybind/moe/__init__.py
from .moe_cuda import MoeCudaExtension __all__ = ["MoeCudaExtension"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/conftest.py
tests/conftest.py
import gc from colossalai.accelerator import get_accelerator def pytest_runtest_setup(item): # called for running each test in 'a' directory accelerator = get_accelerator() accelerator.empty_cache() gc.collect()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/__init__.py
tests/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_optimizer/test_nvme.py
tests/test_optimizer/test_nvme.py
import pytest import torch from colossalai.nn.optimizer import CPUAdam, HybridAdam from colossalai.testing import clear_cache_before_run, parameterize from tests.kit.model_zoo import model_zoo def move_some_params_to_cuda(model, torch_model): model.embed.weight.data = model.embed.weight.cuda() torch_model.em...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_optimizer/test_adam_optim.py
tests/test_optimizer/test_adam_optim.py
from copy import deepcopy from typing import Type, Union import pytest import torch import torch.nn as nn from torch.optim import Adam, AdamW from colossalai.nn.optimizer import CPUAdam, FusedAdam, HybridAdam from tests.kit.model_zoo import model_zoo from tests.test_optimizer._utils import force_assign_grad, setup_pa...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_optimizer/test_adam_kernel.py
tests/test_optimizer/test_adam_kernel.py
# This test checks adam kernels # Baseline is pure fp32 torch adam optimizer import math from abc import abstractmethod from typing import Type import pytest import torch from torch import Tensor from colossalai.accelerator import get_accelerator from colossalai.utils import multi_tensor_applier _FUSED_ALLOWED_P_G_T...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_optimizer/test_dist_galore.py
tests/test_optimizer/test_dist_galore.py
"""Usage(requires 4 GPUs): python test_dist_galore.py""" import pytest import torch import torch.distributed as dist from torch.testing import assert_close import colossalai from colossalai.cluster import DistCoordinator, ProcessGroupMesh from colossalai.logging import disable_existing_loggers from colossalai.nn.opti...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_optimizer/test_lr_scheduler.py
tests/test_optimizer/test_lr_scheduler.py
import torch.nn as nn from torch.optim import Adam from colossalai.nn.lr_scheduler import CosineAnnealingWarmupLR def test_lr_scheduler_save_load(): model = nn.Linear(10, 10) optimizer = Adam(model.parameters(), lr=1e-3) scheduler = CosineAnnealingWarmupLR(optimizer, total_steps=5, warmup_steps=2) ne...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_optimizer/test_dist_came.py
tests/test_optimizer/test_dist_came.py
import pytest import torch import torch.distributed as dist from torch.testing import assert_close import colossalai from colossalai.cluster import ProcessGroupMesh from colossalai.logging import disable_existing_loggers from colossalai.nn.optimizer.came import CAME from colossalai.nn.optimizer.distributed_came import...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_optimizer/test_dist_lamb.py
tests/test_optimizer/test_dist_lamb.py
import pytest import torch import torch.distributed as dist import torch.nn as nn from torch.testing import assert_close import colossalai from colossalai.cluster import DistCoordinator, ProcessGroupMesh from colossalai.logging import disable_existing_loggers from colossalai.nn.optimizer import DistributedLamb, Lamb f...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_optimizer/_utils.py
tests/test_optimizer/_utils.py
import torch import torch.distributed as dist import torch.nn as nn from torch.testing import assert_close import colossalai from colossalai.shardformer.layer.utils import Randomizer from colossalai.tensor.d_tensor import get_layout, get_sharding_spec, is_distributed_tensor from colossalai.tensor.d_tensor.api import c...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_optimizer/test_dist_adafactor.py
tests/test_optimizer/test_dist_adafactor.py
import pytest import torch import torch.distributed as dist from torch import nn from torch.testing import assert_close import colossalai from colossalai.cluster import ProcessGroupMesh from colossalai.logging import disable_existing_loggers from colossalai.nn.optimizer.adafactor import Adafactor from colossalai.nn.op...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_analyzer/__init__.py
tests/test_analyzer/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_analyzer/test_fx/test_symbolic_profile.py
tests/test_analyzer/test_fx/test_symbolic_profile.py
import pytest import torch from packaging import version from colossalai.testing.utils import clear_cache_before_run, parameterize from tests.test_analyzer.test_fx.zoo import tm_models, tmm_models try: from colossalai._analyzer._subclasses import MetaTensorMode from colossalai._analyzer.fx import symbolic_pro...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_analyzer/test_fx/test_nested_ckpt.py
tests/test_analyzer/test_fx/test_nested_ckpt.py
import pytest import torch import torch.nn as nn from torch.utils.checkpoint import checkpoint from colossalai.testing import clear_cache_before_run try: from colossalai._analyzer.fx import symbolic_trace except: pass class MyModule(nn.Module): def __init__(self): super().__init__() self...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_analyzer/test_fx/test_mod_dir.py
tests/test_analyzer/test_fx/test_mod_dir.py
import pytest import torch from colossalai.testing import clear_cache_before_run, parameterize try: from colossalai._analyzer.fx import symbolic_trace except: pass class LinearModel(torch.nn.Module): def __init__(self, in_features, out_features, bias): super().__init__() self.linear = to...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_analyzer/test_fx/test_bias_addition.py
tests/test_analyzer/test_fx/test_bias_addition.py
import pytest import torch from packaging import version from torch.utils.checkpoint import checkpoint from colossalai.testing.utils import clear_cache_before_run, parameterize try: from colossalai._analyzer.fx import symbolic_trace except: pass class LinearModel(torch.nn.Module): def __init__(self, in_...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_analyzer/test_fx/zoo.py
tests/test_analyzer/test_fx/zoo.py
import timm.models as tmm import torchvision.models as tm # input shape: (batch_size, 3, 224, 224) tm_models = [ tm.alexnet, tm.convnext_base, tm.densenet121, # tm.efficientnet_v2_s, # tm.googlenet, # output bad case # tm.inception_v3, # bad case tm.mobilenet_v2, tm.mobilenet_v3_smal...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_analyzer/test_fx/test_shape_prop.py
tests/test_analyzer/test_fx/test_shape_prop.py
import pytest import torch from packaging import version from colossalai.testing.utils import clear_cache_before_run, parameterize from tests.test_analyzer.test_fx.zoo import tm_models, tmm_models try: from colossalai._analyzer._subclasses import MetaTensorMode from colossalai._analyzer.fx import symbolic_tra...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_analyzer/test_fx/__init__.py
tests/test_analyzer/test_fx/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_analyzer/test_subclasses/test_meta_mode.py
tests/test_analyzer/test_subclasses/test_meta_mode.py
import pytest import torch import torchvision.models as tm from packaging import version from colossalai.testing import clear_cache_before_run, parameterize try: from colossalai._analyzer._subclasses import MetaTensorMode except: pass from tests.test_analyzer.test_fx.zoo import tm_models, tmm_models def com...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_analyzer/test_subclasses/test_aten.py
tests/test_analyzer/test_subclasses/test_aten.py
from typing import Any, Callable, Union import pytest import torch import torch.nn as nn from colossalai.testing import clear_cache_before_run try: from colossalai._analyzer._subclasses import MetaTensor except: pass aten = torch.ops.aten registered_meta = { ("aten.convolution.default", True): [ # (at...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_analyzer/test_subclasses/test_flop_tensor.py
tests/test_analyzer/test_subclasses/test_flop_tensor.py
import pytest import torch import torch.nn.functional as F import torchvision.models as tm from packaging import version from tests.test_analyzer.test_fx.zoo import tm_models, tmm_models try: from colossalai._analyzer._subclasses import MetaTensorMode, flop_count except: pass @pytest.mark.skipif(version.par...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_analyzer/test_subclasses/__init__.py
tests/test_analyzer/test_subclasses/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_cluster/test_device_mesh_manager.py
tests/test_cluster/test_device_mesh_manager.py
from colossalai.cluster.device_mesh_manager import DeviceMeshInfo, DeviceMeshManager from colossalai.initialize import launch from colossalai.logging import disable_existing_loggers from colossalai.testing import rerun_if_address_is_in_use, spawn def check_device_mesh_manager(rank, world_size, port): disable_exis...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_cluster/test_process_group_mesh.py
tests/test_cluster/test_process_group_mesh.py
import pytest import torch.distributed as dist import colossalai from colossalai.cluster import ProcessGroupMesh from colossalai.testing import spawn def check_process_group_mesh_with_cases(): DP_DIM, PP_DIM, TP_DIM = 0, 1, 2 DP_SIZE, PP_SIZE, TP_SIZE = 1, 2, 2 RANK_TO_COORDINATE = { 0: (0, 0, 0)...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_lora/test_lora.py
tests/test_lora/test_lora.py
import copy import os from itertools import product import torch from peft import LoraConfig from torch import distributed as dist from torch.optim import AdamW import colossalai from colossalai.booster import Booster from colossalai.booster.plugin import HybridParallelPlugin, LowLevelZeroPlugin, TorchDDPPlugin from ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_pipeline/test_stage_manager.py
tests/test_pipeline/test_stage_manager.py
import pytest import torch.distributed as dist import colossalai from colossalai.cluster import ProcessGroupMesh from colossalai.pipeline.stage_manager import PipelineStageManager from colossalai.testing import rerun_if_address_is_in_use, spawn def check_stage_manager(): DP_DIM, PP_DIM = 0, 1 DP_SIZE, PP_SIZ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_pipeline/test_p2p_communication.py
tests/test_pipeline/test_p2p_communication.py
import pytest import torch import torch.distributed as dist import colossalai from colossalai.accelerator import get_accelerator from colossalai.cluster import ProcessGroupMesh from colossalai.pipeline.p2p import PipelineP2PCommunication, create_send_metadata from colossalai.pipeline.stage_manager import PipelineStage...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_pipeline/test_schedule/test_zerobubble_pp.py
tests/test_pipeline/test_schedule/test_zerobubble_pp.py
from contextlib import nullcontext from copy import deepcopy from functools import partial from typing import Tuple import pytest import torch import torch.distributed as dist import torch.nn as nn from torch.testing import assert_close from transformers.models.llama.configuration_llama import LlamaConfig from transfo...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_pipeline/test_schedule/test_interleaved.py
tests/test_pipeline/test_schedule/test_interleaved.py
import copy from functools import partial from types import MethodType import pytest import torch import torch.distributed as dist import torch.nn as nn from torch.testing import assert_close import colossalai from colossalai.cluster import ProcessGroupMesh from colossalai.interface import OptimizerWrapper from colos...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_pipeline/test_schedule/test_pipeline_schedule_utils.py
tests/test_pipeline/test_schedule/test_pipeline_schedule_utils.py
import torch from colossalai.pipeline.schedule._utils import get_batch_size, get_micro_batch, merge_batch def test_get_batch_size(): tensor = torch.rand(2, 3) assert get_batch_size(tensor) == 2 assert get_batch_size([tensor]) == 2 assert get_batch_size((1, tensor)) == 2 assert get_batch_size({"te...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_pipeline/test_schedule/test_oneF_oneB.py
tests/test_pipeline/test_schedule/test_oneF_oneB.py
import copy from functools import partial from types import MethodType import pytest import torch import torch.distributed as dist import torch.nn as nn from torch.testing import assert_close import colossalai from colossalai.cluster import ProcessGroupMesh from colossalai.interface import OptimizerWrapper from colos...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_pipeline/test_pipeline_utils/test_t5_pipeline_utils.py
tests/test_pipeline/test_pipeline_utils/test_t5_pipeline_utils.py
import random from colossalai.pipeline.stage_manager import PipelineStageManager from colossalai.shardformer.policies.t5 import T5BasePolicy from colossalai.shardformer.shard.shard_config import ShardConfig class _ShardConfig(ShardConfig): def __post_init__(self): pass class _PipelineStageManager(Pipel...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_pipeline/test_pipeline_utils/test_whisper_pipeline_utils.py
tests/test_pipeline/test_pipeline_utils/test_whisper_pipeline_utils.py
import random from colossalai.pipeline.stage_manager import PipelineStageManager from colossalai.shardformer.policies.whisper import WhisperPolicy from colossalai.shardformer.shard.shard_config import ShardConfig class _ShardConfig(ShardConfig): def __post_init__(self): pass class _PipelineStageManager...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_config/sample_config.py
tests/test_config/sample_config.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- train_data = dict( dataset=dict( type="CIFAR10Dataset", root="/path/to/data", download=True, transform_pipeline=[ dict(type="RandomResizedCrop", size=224), dict(type="RandomHorizontalFlip"), dict(typ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_config/test_load_config.py
tests/test_config/test_load_config.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- from pathlib import Path from colossalai.context.config import Config def test_load_config(): filename = Path(__file__).parent.joinpath("sample_config.py") config = Config.from_file(filename) assert config.train_data, "cannot access train data as attribut...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_comm_size_compute.py
tests/test_fx/test_comm_size_compute.py
import torch from torch.fx import symbolic_trace from colossalai.fx._compatibility import is_compatible_with_meta from colossalai.fx.passes.adding_split_node_pass import split_with_split_nodes_pass, uniform_split_pass from colossalai.fx.passes.meta_info_prop import MetaInfoProp from colossalai.fx.passes.utils import g...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_coloproxy.py
tests/test_fx/test_coloproxy.py
import torch import torch.nn as nn from torch.fx import GraphModule from colossalai.fx.proxy import ColoProxy from colossalai.fx.tracer.tracer import ColoTracer from colossalai.testing import clear_cache_before_run class Conv1D(nn.Module): def __init__(self, nf, nx): super().__init__() self.nf = ...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_parallel_1d.py
tests/test_fx/test_parallel_1d.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- import pytest import torch from torch.fx import symbolic_trace from colossalai.fx.passes import column_shard_linear_pass from colossalai.initialize import launch from colossalai.legacy.core import global_context as gpc from colossalai.logging import disable_existing_log...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_meta_info_prop.py
tests/test_fx/test_meta_info_prop.py
import torch from torch.fx import symbolic_trace from colossalai.fx._compatibility import is_compatible_with_meta from colossalai.fx.passes.meta_info_prop import MetaInfoProp, TensorMetadata from colossalai.testing import clear_cache_before_run if is_compatible_with_meta(): from colossalai.fx.profiler import Meta...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_pipeline_passes.py
tests/test_fx/test_pipeline_passes.py
import torch from torch.fx import symbolic_trace from colossalai.fx.passes.adding_split_node_pass import ( balanced_split_pass, balanced_split_pass_v2, split_with_split_nodes_pass, uniform_split_pass, ) from colossalai.testing import clear_cache_before_run MODEL_DIM = 16 BATCH_SIZE = 8 PIPELINE_SIZE =...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/tests/test_fx/test_graph_manipulation.py
tests/test_fx/test_graph_manipulation.py
import torch from colossalai.fx import ColoTracer from colossalai.fx.passes.utils import assign_bfs_level_to_nodes, get_leaf, get_top from colossalai.testing import clear_cache_before_run class MLP(torch.nn.Module): def __init__(self, dim: int): super().__init__() self.linear1 = torch.nn.Linear(d...
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false