model_name string | checkpoint string | date_released string | model_options list | original_modeling_code string | original_source string | current_modeling_code string | current_modular_code string | bases list |
|---|---|---|---|---|---|---|---|---|
arcee | arcee-ai/model-id | 2025-06-24 | [
"ariaaya_visionbambabiogptbitnetcamembertcoherecohere2colpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v3deformable_detrdiffllamadinov2_with_registersdptemu3erniefalcon_h1falcon_mambagemmagemma2gemma3glmglm4got_ocr2gpt_neoxgranitegranitemoegranitemoehybridgranitemoesharedheliumhgnet_v2hubertijepainforme... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/arcee/modular_arcee.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# ... | # Copyright 2025 Arcee AI and the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless req... | [
"llama",
"nemotron"
] |
aria | rhymes-ai/Aria | 2024-12-06 | [
"biogptcamembertcohereconditional_detrdata2vecdbrxdeformable_detrdpterniefalcon_mambagemmagemma2glmgpt_neoxgranitegranitemoehubertijepainformerinstructblipvideojambajetmoelayoutxlmllava_next_videollava_onevisionmask2formermistralmixtralolmoolmo2olmoeowlv2persimmonphiphi3phimoeplbartqwen2qwen2_moerobertart_detrsegfo... | # Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
import math
from dataclasses import dataclass
from functools import reduce
from math import gcd
from typing import Optional... | https://github.com/rhymes-ai/Aria/blob/9b25fecb02553a24832c26617d4e034517f71260/gptfast/model.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/aria/modular_aria.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# ... | # Copyright 2024 The Rhymes-AI Teams Authors and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE... | [
"llama",
"llava"
] |
bert_generation | google/bert_for_seq_generation_L-24_bbc_encoder | null | [] | # coding=utf-8
# Copyright 2020 The Google AI Language Team Authors and The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICEN... | https://github.com/huggingface/transformers/blob/c89bdfbe72/src/transformers/models/bert_generation/modeling_bert_generation.py | # Copyright 2020 The Google AI Language Team Authors and The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unle... | null | [] |
biogpt | microsoft/biogpt | 2022-12-05 | [
"camembertconditional_detrdata2vecdeformable_detrdpterniegpt_neoxhubertlayoutxlmplbartrobertasegformersewswitch_transformersunispeechwavlmyolos"
] | # coding=utf-8
# Copyright 2022 The HuggingFace Team and Microsoft Research AI4Science All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licen... | https://github.com/huggingface/transformers/blob/13e736685a/src/transformers/models/biogpt/modeling_biogpt.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/biogpt/modular_biogpt.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
... | # Copyright 2022 The HuggingFace Team and Microsoft Research AI4Science All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0... | [
"bart",
"opt"
] |
bitnet | microsoft/bitnet-b1.58-2B-4T | 2025-04-28 | [
"ariaaya_visionbambabiogptcamembertcoherecohere2colpaliconditional_detrdata2vecdbrxdeepseek_v3deformable_detrdiffllamadinov2_with_registersdptemu3erniefalcon_mambagemmagemma2gemma3glmglm4got_ocr2gpt_neoxgranitegranitemoegranitemoesharedheliumhubertijepainformerinstructblipvideointernvljambajanusjetmoelayoutxlmllava... | # Copyright (c) Facebook, Inc. and its affiliates. All rights reserved.
#
# This source code is licensed under the BSD license found in the
# LICENSE file in the root directory of this source tree.
from dataclasses import dataclass
from typing import Optional, Tuple, Union
import torch
from torch import nn
... | https://github.com/microsoft/BitNet/blob/01eb415772c342d9f20dc42772f1583ae1e5b102/gpu/model.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/bitnet/modular_bitnet.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
... | # Copyright 2025 The BitNet Team and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unl... | [
"gemma",
"llama"
] |
blt | itazap/blt-1b-hf | 2025-09-19 | [
"aimv2apertusarceeariaaya_visionbambabiogptbitnetcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptefficientloftremu3eomternieernie4_5ernie4_5_moeevollaexaone4falcon_h... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/blt/modular_blt.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# ... | # Copyright 2025 HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicab... | [
"cohere2",
"llama",
"mllama"
] |
camembert | almanach/camembert-base | 2020-11-16 | [] | # coding=utf-8
# Copyright 2019 Inria, Facebook AI Research and the HuggingFace Inc. team.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the... | https://github.com/huggingface/transformers/blob/c89bdfbe72/src/transformers/models/camembert/modeling_camembert.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/camembert/modular_camembert.py.
# Do NOT edit this file manually as any edits will be overwritten by the generati... | # Copyright 2019 Inria, Facebook AI Research and the HuggingFace Inc. team.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#... | [
"roberta"
] |
chameleon | facebook/chameleon-7b | null | [] | # coding=utf-8
# Copyright 2024 Meta Inc. and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.... | https://github.com/huggingface/transformers/blob/24cfcc2114/src/transformers/models/chameleon/modeling_chameleon.py | # Copyright 2024 Meta Inc. and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless re... | null | [] |
codegen | Salesforce/codegen-350M-mono | null | [] | # coding=utf-8
# Copyright 2021 The EleutherAI and HuggingFace Teams. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#... | https://github.com/salesforce/CodeGen/blob/6f09e495f41e006375055239e4bcb18539f3a38b/codegen1/jaxformer/hf/codegen/modeling_codegen.py | # Copyright 2022 Salesforce authors, The EleutherAI, and HuggingFace Teams. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE... | null | [] |
conditional_detr | microsoft/conditional-detr-resnet-50 | 2022-09-22 | [
"camembertdata2vecdeformable_detrdpterniegpt_neoxhubertlayoutxlmplbartrobertasegformersewunispeechwavlmyolos"
] | # coding=utf-8
# Copyright 2022 Microsoft Research Asia and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licen... | https://github.com/huggingface/transformers/blob/126a739058/src/transformers/models/conditional_detr/modeling_conditional_detr.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/conditional_detr/modular_conditional_detr.py.
# Do NOT edit this file manually as any edits will be overwritten b... | # Copyright 2022 Microsoft Research Asia and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0... | [
"deformable_detr",
"detr"
] |
cwm | facebook/cwm | 2025-10-09 | [
"aimv2apertusarceeariaaya_visionbambabiogptbitnetbltcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptedgetamedgetam_videoefficientloftremu3eomternieernie4_5ernie4_5_m... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/cwm/modular_cwm.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# ... | # Copyright 2025
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# d... | [
"llama",
"qwen2"
] |
data2vec_audio | hf-internal-testing/tiny-random-data2vec-seq-class | null | [] | null | null | null | null | [] |
deepseek_v2 | deepseek-ai/DeepSeek-V2-Lite | 2025-07-09 | [
"aimv2arceeariaaya_visionbambabiogptbitnetcamembertcoherecohere2colpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v3deformable_detrdiadiffllamadinov2_with_registersdogedots1dptemu3eomterniefalcon_h1falcon_mambagemmagemma2gemma3gemma3nglmglm4glm4vgot_ocr2gpt_neoxgranitegranitemoegranitemoehybridgranitemoe... | # coding=utf-8
# Copyright 2023 DeepSeek-AI and The HuggingFace Inc. team. All rights reserved.
#
# This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX
# and OPT implementations in this library. It has been modified from its
# original forms to accommodate minor architectural differences compared
# to ... | https://huggingface.co/deepseek-ai/DeepSeek-V2-Lite/blob/main/modeling_deepseek.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/deepseek_v2/modular_deepseek_v2.py.
# Do NOT edit this file manually as any edits will be overwritten by the gene... | # Copyright 2025 HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicab... | [
"llama",
"qwen2_moe"
] |
deepseek_v3 | bzantium/tiny-deepseek-v3 | 2025-03-28 | [
"ariaaya_visionbambabiogptcamembertcoherecohere2colpaliconditional_detrdata2vecdbrxdeformable_detrdiffllamadinov2_with_registersdptemu3erniefalcon_mambagemmagemma2gemma3glmgot_ocr2gpt_neoxgranitegranitemoegranitemoesharedheliumhubertijepainformerinstructblipvideojambajetmoelayoutxlmllava_next_videollava_onevisionma... | # coding=utf-8
# Copyright 2023 DeepSeek-AI and The HuggingFace Inc. team. All rights reserved.
#
# This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX
# and OPT implementations in this library. It has been modified from its
# original forms to accommodate minor architectural differences compared
# to ... | https://huggingface.co/bzantium/tiny-deepseek-v3/blob/main/modeling_deepseek.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/deepseek_v3/modular_deepseek_v3.py.
# Do NOT edit this file manually as any edits will be overwritten by the gene... | import math
from collections.abc import Callable
import torch
import torch.nn.functional as F
from torch import nn
from ... import initialization as init
from ...cache_utils import Cache
from ...modeling_flash_attention_utils import FlashAttentionKwargs
from ...modeling_layers import GenericForSequenceClassification,... | [
"llama",
"mixtral",
"qwen2_moe"
] |
deformable_detr | SenseTime/deformable-detr | 2022-09-14 | [
"camembertdata2vecdpterniegpt_neoxhubertlayoutxlmplbartrobertasegformersewunispeechwavlmyolos"
] | # coding=utf-8
# Copyright 2022 SenseTime and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.... | https://github.com/huggingface/transformers/blob/59407bbeb3/src/transformers/models/deformable_detr/modeling_deformable_detr.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/deformable_detr/modular_deformable_detr.py.
# Do NOT edit this file manually as any edits will be overwritten by ... | # Copyright 2022 SenseTime and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless re... | [
"detr"
] |
detr | facebook/detr-resnet-50 | null | [] | # Copyright 2025 Google LLC.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, ... | https://github.com/google-research/vision_transformer/blob/0d03f554b83af02550407da1e8c702a7cb75e74b/vit_jax/models.py | # Copyright 2021 Facebook AI Research The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Un... | null | [] |
diffllama | kajuma/DiffLlama-0.3B-handcut | 2025-01-07 | [
"ariabambabiogptcamembertcoherecohere2colpaliconditional_detrdata2vecdbrxdeformable_detrdinov2_with_registersdpterniefalcon_mambagemmagemma2glmgpt_neoxgranitegranitemoehubertijepainformerinstructblipvideojambajetmoelayoutxlmllava_next_videollava_onevisionmask2formermistralmixtralmodernbertolmoolmo2olmoeowlv2persimm... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/diffllama/modular_diffllama.py.
# Do NOT edit this file manually as any edits will be overwritten by the generati... | # Copyright 2024 weak-kajuma and the HuggingFace Inc. team. All rights reserved.
#
# This code is based on Llama implementations in this library and Microsoft's
# Differential Transformer implementations.
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance... | [
"gemma",
"llama",
"mistral"
] |
dit | microsoft/dit-base-finetuned-rvlcdip | null | [] | # --------------------------------------------------------
# BEIT: BERT Pre-Training of Image Transformers (https://arxiv.org/abs/2106.08254)
# Github source: https://github.com/microsoft/unilm/tree/master/beit
# Copyright (c) 2021 Microsoft
# Licensed under The MIT License [see LICENSE for details]
# By Hangbo Bao
# B... | https://github.com/microsoft/unilm/blob/833df7e7832e5064a281131ee64a481afa8e5b95/beit/modeling_discrete_vae.py | null | null | [] |
doge | SmallDoge/Doge-20M | 2025-07-08 | [
"arceeariaaya_visionbambabiogptbitnetcamembertcoherecohere2colpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v3deformable_detrdiadiffllamadinov2_with_registersdots1dptemu3eomterniefalcon_h1falcon_mambagemmagemma2gemma3gemma3nglmglm4glm4vgot_ocr2gpt_neoxgranitegranitemoegranitemoehybridgranitemoesharedhel... | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/doge/modular_doge.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# ... | https://github.com/SmallDoges/small-doge/blob/0e305a735708fd277f5a8e9fd36954492aba28f1/src/small_doge/models/doge/modeling_doge.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/doge/modular_doge.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# ... | # Copyright 2025 Jingze Shi and the HuggingFace Inc. team. All rights reserved.
#
# The Doge family of small language models is trained by SmallDoge Team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the... | [
"llama",
"mixtral"
] |
dots1 | redmoe-ai-v1/dots.llm1.test | 2025-06-25 | [
"arceeariaaya_visionbambabiogptbitnetcamembertcoherecohere2colpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v3deformable_detrdiffllamadinov2_with_registersdptemu3erniefalcon_h1falcon_mambagemmagemma2gemma3glmglm4got_ocr2gpt_neoxgranitegranitemoegranitemoehybridgranitemoesharedheliumhgnet_v2hubertijepain... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/dots1/modular_dots1.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# ... | # Copyright 2025 The rednote-hilab team and the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
... | [
"deepseek_v3",
"qwen3"
] |
edgetam | timm/repvit_m1.dist_in1k | 2025-09-29 | [
"aimv2apertusarceeariaaya_visionbambabiogptbitnetbltcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptefficientloftremu3eomternieernie4_5ernie4_5_moeevollaexaone4falco... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/edgetam/modular_edgetam.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation o... | # Copyright 2025 The Meta AI Authors and The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unle... | [
"sam2"
] |
edgetam_video | yonigozlan/EdgeTAM-hf | 2025-09-29 | [
"aimv2apertusarceeariaaya_visionbambabiogptbitnetbltcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptefficientloftremu3eomternieernie4_5ernie4_5_moeevollaexaone4falco... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/edgetam_video/modular_edgetam_video.py.
# Do NOT edit this file manually as any edits will be overwritten by the ... | # Copyright 2025 the HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | [
"sam2",
"sam2_video"
] |
emu3 | BAAI/Emu3-Chat-hf | 2025-01-10 | [
"ariabambabiogptcamembertcoherecohere2colpaliconditional_detrdata2vecdbrxdeformable_detrdiffllamadinov2_with_registersdpterniefalcon_mambagemmagemma2glmgpt_neoxgranitegranitemoehubertijepainformerinstructblipvideojambajetmoelayoutxlmllava_next_videollava_onevisionmask2formermistralmixtralmodernbertolmoolmo2olmoeowl... | # coding=utf-8
# Copyright 2024 The Emu team, BAAI and The HuggingFace Inc. team. All rights reserved.
#
# This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX
# and OPT implementations in this library. It has been modified from its
# original forms to accommodate minor architectural differences compare... | https://github.com/baaivision/Emu3/blob/dbbf9858194d70b8c58293e219ecffe22df0f9c7/emu3/mllm/modeling_emu3.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/emu3/modular_emu3.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# ... | # Copyright 2024 HuggingFace Inc. team. All rights reserved.
#
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applic... | [
"chameleon",
"llama",
"siglip"
] |
ernie4_5 | baidu/ERNIE-4.5-0.3B-PT | 2025-07-21 | [
"aimv2arceeariaaya_visionbambabiogptbitnetcamembertcoherecohere2colpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v2deepseek_v3deformable_detrdiadiffllamadinov2_with_registersdogedots1dptemu3eomterniefalcon_h1falcon_mambagemmagemma2gemma3gemma3nglmglm4glm4vgot_ocr2gpt_neoxgranitegranitemoegranitemoehybri... | # Copyright (c) 2025 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by appli... | https://github.com/PaddlePaddle/ERNIE/blob/790a50b045d1aca2753d5395d8bec0806b2e6925/ernie/modeling.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/ernie4_5/modular_ernie4_5.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation... | # Copyright (c) 2025 Baidu, Inc. and HuggingFace Inc. team. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless ... | [
"glm",
"llama",
"olmo"
] |
esmfold | facebook/esmfold_v1 | null | [] | null | null | null | null | [] |
evolla | westlake-repl/Evolla-10B-hf | 2025-07-26 | [
"aimv2arceeariaaya_visionbambabiogptbitnetcamembertcoherecohere2colpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptefficientloftremu3eomternieernie4_5ernie4_5_moefalcon_h1falcon_mambagemmagemma2gemma3gemm... | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from examples/modular-transformers/modular_add_function.py.
# Do NOT edit this file manually as any edits will be overwritten by the generatio... | https://github.com/zhoubay/transformers/blob/dd16acb8a3e93b643aa374c9fb80749f5235c1a6/examples/modular-transformers/modeling_add_function.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/evolla/modular_evolla.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
... | # Copyright 2025 Westlake Representational Learning Lab (Fajie Yuan Lab) team and the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http:... | [
"esm",
"llama"
] |
falcon | Rocketknight1/falcon-rw-1b | null | [] | # coding=utf-8
# Copyright 2023 the Falcon authors and HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICE... | https://github.com/huggingface/transformers/blob/b3ab3fac1d/src/transformers/models/falcon/modeling_falcon.py | # Copyright 2023 the Falcon authors and HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unl... | null | [] |
falcon_mamba | tiiuae/falcon-mamba-7b | 2024-08-12 | [
"biogptcamembertcohereconditional_detrdata2vecdbrxdeformable_detrdpterniegemmagemma2gpt_neoxhubertinformerinstructblipvideojambajetmoelayoutxlmllava_next_videomask2formermistralmixtralolmoowlv2persimmonphiphi3plbartqwen2qwen2_moerobertart_detrsegformersewstarcoder2switch_transformersunispeechvipllavawavlmyolos"
] | # coding=utf-8
# Copyright 2024 state-spaces/falcon_mamba org and HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
#... | https://github.com/huggingface/transformers/blob/7c11491208/src/transformers/models/falcon_mamba/modeling_falcon_mamba.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/falcon_mamba/modular_falcon_mamba.py.
# Do NOT edit this file manually as any edits will be overwritten by the ge... | # Copyright 2024 Tri Dao, Albert Gu, Technological Innovation Institute and HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICEN... | [
"mamba"
] |
fast_vlm | KamilaMila/FastVLM-0.5B | 2025-12-02 | [
"afmoeaimv2apertusarceeariaaudioflamingo3aya_visionbambabiogptbitnetbltcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmcwmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptedgetamedgetam_videoefficientloftremu3eomte... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/fast_vlm/modular_fast_vlm.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation... | # Copyright 2025 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | [
"llava"
] |
flex_olmo | shanearora/Flex-reddit-2x7B-1T | 2025-09-18 | [
"aimv2apertusarceeariaaya_visionbambabiogptbitnetcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptefficientloftremu3eomternieernie4_5ernie4_5_moeevollaexaone4falcon_h... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/flex_olmo/modular_flex_olmo.py.
# Do NOT edit this file manually as any edits will be overwritten by the generati... | # Copyright 2025 the HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | [
"mixtral",
"olmo2",
"olmoe"
] |
funnel | sgugger/funnel-random-tiny | null | [] | # coding=utf-8
# Copyright 2020-present Google Brain and Carnegie Mellon University Authors and the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.a... | https://github.com/huggingface/transformers/blob/c89bdfbe72/src/transformers/models/funnel/modeling_funnel.py | # Copyright 2020-present Google Brain and Carnegie Mellon University Authors and the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licen... | null | [] |
fuyu | adept/fuyu-8b | null | [] | # coding=utf-8
# Copyright 2023 HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requi... | https://github.com/huggingface/transformers/blob/caa0ff0bf1/src/transformers/models/fuyu/modeling_fuyu.py | # Copyright 2023 HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicab... | null | [] |
gemma3 | hf-internal-testing/tiny-random-Gemma3ForCausalLM | 2025-03-12 | [
"ariaaya_visionbambabiogptcamembertcoherecohere2colpaliconditional_detrdata2vecdbrxdeformable_detrdiffllamadinov2_with_registersdptemu3erniefalcon_mambagemmagemma2glmgot_ocr2gpt_neoxgranitegranitemoegranitemoesharedheliumhubertijepainformerinstructblipvideojambajetmoelayoutxlmllava_next_videollava_onevisionmask2for... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/gemma3/modular_gemma3.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
... | # Copyright 2025 Google Inc. HuggingFace Inc. team. All rights reserved.
#
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requir... | [
"gemma2",
"paligemma"
] |
glm46v | THUDM/GLM-4.1V-9B-Thinking | 2025-11-15 | [
"aimv2apertusarceeariaaudioflamingo3aya_visionbambabiogptbitnetbltcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmcwmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptedgetamedgetam_videoefficientloftremu3eomterniee... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/glm46v/modular_glm46v.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
... | # Copyright 2025 the HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | [
"glm4v"
] |
glm4_moe | zai-org/GLM-4.5 | 2025-07-21 | [
"aimv2arceeariaaya_visionbambabiogptbitnetcamembertcoherecohere2colpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v2deepseek_v3deformable_detrdiadiffllamadinov2_with_registersdogedots1dptemu3eomterniefalcon_h1falcon_mambagemmagemma2gemma3gemma3nglmglm4glm4vgot_ocr2gpt_neoxgranitegranitemoegranitemoehybri... | # Copyright (c) ModelScope Contributors. All rights reserved.
from swift.infer_engine import InferRequest, RequestConfig, TransformersEngine
from swift.model import Model, ModelGroup, ModelMeta, register_model
from swift.template import TemplateMeta, register_template
register_template(
TemplateMeta(
templ... | https://github.com/modelscope/ms-swift/blob/b6af05674945e6fae3cca8fc02698bbd97cef6d9/examples/custom/model.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/glm4_moe/modular_glm4_moe.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation... | # Copyright 2025 The ZhipuAI Inc. team and HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# U... | [
"cohere",
"deepseek_v3",
"glm",
"gpt_neox"
] |
glm4_moe_lite | zai-org/GLM-4.7-Flash | 2026-01-13 | [
"afmoeaimv2apertusarceeariaaudioflamingo3aya_visionbambabiogptbitnetbltcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmcwmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptedgetamedgetam_videoefficientloftremu3eomte... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/glm4_moe_lite/modular_glm4_moe_lite.py.
# Do NOT edit this file manually as any edits will be overwritten by the ... | # Copyright 2025 the HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | [
"deepseek_v3",
"glm4_moe"
] |
glm4v | THUDM/GLM-4.1V-9B-Thinking | 2025-06-25 | [
"arceeariaaya_visionbambabiogptbitnetcamembertcoherecohere2colpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v3deformable_detrdiffllamadinov2_with_registersdptemu3erniefalcon_h1falcon_mambagemmagemma2gemma3glmglm4got_ocr2gpt_neoxgranitegranitemoegranitemoehybridgranitemoesharedheliumhgnet_v2hubertijepain... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/glm4v/modular_glm4v.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# ... | # Copyright 2025 The ZhipuAI Inc. team and HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# U... | [
"glm4",
"qwen2_5_vl"
] |
glm4v_moe | zai-org/GLM-4.5V | 2025-08-08 | [
"aimv2arceeariaaya_visionbambabiogptbitnetcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptefficientloftremu3eomternieernie4_5ernie4_5_moeevollaexaone4falcon_h1falcon... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/glm4v_moe/modular_glm4v_moe.py.
# Do NOT edit this file manually as any edits will be overwritten by the generati... | # Copyright 2025 The ZhipuAI Inc. team and HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# U... | [
"deepseek_v3",
"glm4",
"glm4_moe",
"glm4v",
"gpt_neox",
"qwen3_vl_moe"
] |
glm_ocr | zai-org/GLM-OCR | 2026-01-27 | [
"afmoeaimv2apertusarceeariaaudioflamingo3aya_visionbambabiogptbitnetbltcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmcwmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptedgetamedgetam_videoefficientloftremu3eomte... | # -*- coding: utf-8 -*-
# @Time : 2019/8/23 21:57
# @Author : zhoujun
from addict import Dict
from paddle import nn
import paddle.nn.functional as F
from models.backbone import build_backbone
from models.neck import build_neck
from models.head import build_head
class Model(nn.Layer):
def __init__(self, model... | https://github.com/PaddlePaddle/PaddleOCR/blob/9f9d08f52a239b99b3db3a8c400dd08b25dfe7ea/benchmark/PaddleOCR_DBNet/models/model.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/glm_ocr/modular_glm_ocr.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation o... | # Copyright 2026 the HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | [
"glm4v"
] |
got_ocr2 | stepfun-ai/GOT-OCR-2.0-hf | 2025-01-31 | [
"ariabambabiogptcamembertcoherecohere2colpaliconditional_detrdata2vecdbrxdeformable_detrdiffllamadinov2_with_registersdptemu3erniefalcon_mambagemmagemma2glmgpt_neoxgranitegranitemoeheliumhubertijepainformerinstructblipvideojambajetmoelayoutxlmllava_next_videollava_onevisionmask2formermistralmixtralmodernbertmoonshi... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/got_ocr2/modular_got_ocr2.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation... | # Copyright 2024 HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicab... | [
"llava",
"sam"
] |
gpt2 | openai-community/gpt2 | null | [] | import numpy as np
import tensorflow as tf
from tensorflow.contrib.training import HParams
def default_hparams():
return HParams(
n_vocab=0,
n_ctx=1024,
n_embd=768,
n_head=12,
n_layer=12,
)
def shape_list(x):
"""Deal with dynamic shape in tensorflow cleanly."""
... | https://github.com/openai/gpt-2/blob/9b63575ef42771a015060c964af2c3da4cf7c8ab/src/model.py | # Copyright 2018 The OpenAI Team Authors and HuggingFace Inc. team.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# htt... | null | [] |
gpt_bigcode | bigcode/gpt_bigcode-santacoder | null | [] | # Copyright 2021 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | https://github.com/bigcode-project/transformers/blob/8b0cb2c6261e65d4d852d6813f071772c1b32665/src/transformers/modeling_flax_outputs.py | # Copyright 2023 The Bigcode team and HuggingFace Inc. team.
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | null | [] |
gpt_neo | valhalla/gpt-neo-random-tiny | null | [] | # coding=utf-8
# Copyright 2021 The Eleuther AI and HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-... | https://github.com/huggingface/transformers/blob/860264379f/src/transformers/models/gpt_neo/modeling_gpt_neo.py | # Copyright 2021 The Eleuther AI and HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless ... | null | [] |
gpt_neox | EleutherAI/pythia-410m-deduped | 2022-05-24 | [
"camembertdata2vecdpthubertlayoutxlmplbartrobertasegformersewunispeechwavlmyolos"
] | import torch
import transformers
from tqdm import tqdm
from lm_eval import utils
from lm_eval.base import BaseLM
class GPTNeoLM(BaseLM):
def __init__(
self,
device="cuda",
pretrained="EleutherAI/gpt-neo-1.3B",
revision="main",
subfolder=None,
tokenizer=None,
... | https://github.com/EleutherAI/pythia/blob/a19eecb807ec2c79a39ebf18108816e6ffffc1d5/case-studies/term_frequency/model.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/gpt_neox/modular_gpt_neox.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation... | from collections.abc import Callable
from typing import Optional
import torch
from torch import nn
from ...activations import ACT2FN
from ...cache_utils import Cache, DynamicCache
from ...generation import GenerationMixin
from ...masking_utils import create_causal_mask
from ...modeling_flash_attention_utils import Fl... | [
"llama"
] |
gptj | EleutherAI/gpt-j-6B | null | [] | # Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
"""
BART: Denoising Sequence-to-Sequence Pre-training for
Natural Language Generation, Translation, and Comprehension
"""
import logging
from t... | https://github.com/pytorch/fairseq/blob/3d262bb25690e4eb2e7d3c1309b1e9c406ca4b99/fairseq/models/bart/model.py | # Copyright 2021 The EleutherAI and HuggingFace Teams. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requi... | null | [] |
granite | ibm/PowerLM-3b | 2024-08-27 | [
"biogptcamembertcohereconditional_detrdata2vecdbrxdeformable_detrdpterniefalcon_mambagemmagemma2gpt_neoxhubertinformerinstructblipvideojambajetmoelayoutxlmllava_next_videomask2formermistralmixtralolmoowlv2persimmonphiphi3plbartqwen2qwen2_moerobertart_detrsegformersewstarcoder2switch_transformersunispeechvipllavawav... | # coding=utf-8
# Copyright 2024 EleutherAI and the HuggingFace Inc. team. All rights reserved.
#
# This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX
# and OPT implementations in this library. It has been modified from its
# original forms to accommodate minor architectural differences compared
# to G... | https://github.com/huggingface/transformers/blob/c35d2ccf5a/src/transformers/models/granite/modeling_granite.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/granite/modular_granite.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation o... | # Copyright 2024 IBM and the HuggingFace Inc. team. All rights reserved.
#
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requir... | [
"llama"
] |
granitemoe | ibm/PowerMoE-3b | 2024-09-20 | [
"biogptcamembertcohereconditional_detrdata2vecdbrxdeformable_detrdpterniefalcon_mambagemmagemma2gpt_neoxgranitehubertinformerinstructblipvideojambajetmoelayoutxlmllava_next_videollava_onevisionmask2formermistralmixtralolmoolmoeowlv2persimmonphiphi3plbartqwen2qwen2_moerobertart_detrsegformersewstarcoder2switch_trans... | # coding=utf-8
# Copyright 2024 IBM and the HuggingFace Inc. team. All rights reserved.
#
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
... | https://github.com/huggingface/transformers/blob/e472e077c2/src/transformers/models/granitemoe/modeling_granitemoe.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/granitemoe/modular_granitemoe.py.
# Do NOT edit this file manually as any edits will be overwritten by the genera... | # Copyright 2024 IBM and the HuggingFace Inc. team. All rights reserved.
#
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requir... | [
"granite",
"jetmoe",
"llama",
"mixtral"
] |
granitemoeshared | ibm/PowerMoE-3b | 2025-02-14 | [
"ariabambabiogptcamembertcoherecohere2colpaliconditional_detrdata2vecdbrxdeformable_detrdiffllamadinov2_with_registersdptemu3erniefalcon_mambagemmagemma2glmgot_ocr2gpt_neoxgranitegranitemoeheliumhubertijepainformerinstructblipvideojambajetmoelayoutxlmllava_next_videollava_onevisionmask2formermistralmixtralmodernber... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/granitemoeshared/modular_granitemoeshared.py.
# Do NOT edit this file manually as any edits will be overwritten b... | # Copyright 2024 IBM and the HuggingFace Inc. team. All rights reserved.
#
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requir... | [
"granitemoe"
] |
grounding_dino | IDEA-Research/grounding-dino-tiny | null | [] | # coding=utf-8
# Copyright 2024 IDEA Research and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENS... | https://github.com/huggingface/transformers/blob/b752ad3019/src/transformers/models/grounding_dino/modeling_grounding_dino.py | # Copyright 2024 IDEA Research and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unles... | # Copyright 2025 the HuggingFace Inc. team. All rights reserved.
#
# This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX
# and OPT implementations in this library. It has been modified from its
# original forms to accommodate minor architectural differences compared
# to GPT-NeoX and OPT used by the Me... | [] |
hubert | facebook/hubert-base-ls960 | 2021-06-16 | [
"camembertroberta"
] | # Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
"""
BART: Denoising Sequence-to-Sequence Pre-training for
Natural Language Generation, Translation, and Comprehension
"""
import logging
from t... | https://github.com/pytorch/fairseq/blob/3d262bb25690e4eb2e7d3c1309b1e9c406ca4b99/fairseq/models/bart/model.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/hubert/modular_hubert.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
... | # Copyright 2021 The Fairseq Authors and the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
#... | [
"wav2vec2"
] |
hunyuan_v1_moe | tencent/Hunyuan-A13B-Instruct | 2025-08-22 | [
"aimv2arceeariaaya_visionbambabiogptbitnetcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptefficientloftremu3eomternieernie4_5ernie4_5_moeevollaexaone4falcon_h1falcon... | # Copyright (C) 2024 THL A29 Limited, a Tencent company. All rights reserved.
#
# Licensed under the TENCENT HUNYUAN COMMUNITY LICENSE AGREEMENT (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://github.com/Tencent/Tencent-Hunyu... | https://huggingface.co/tencent/Hunyuan-A13B-Instruct/blob/main/modeling_hunyuan.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/hunyuan_v1_moe/modular_hunyuan_v1_moe.py.
# Do NOT edit this file manually as any edits will be overwritten by th... | # Copyright (C) 2025 THL A29 Limited, a Tencent company and the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licen... | [
"hunyuan_v1_dense",
"llama",
"mixtral"
] |
idefics3 | HuggingFaceM4/Idefics3-8B-Llama3 | null | [] | # Copyright 2020-2026 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by appl... | https://github.com/huggingface/trl/blob/1ad25f94e5556b2deb4869e11f62d3f91aa7bbac/trl/experimental/ppo/modeling_value_head.py | # Copyright 2024 the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by appl... | null | [] |
jetmoe | jetmoe/jetmoe-8b | 2024-05-14 | [
"biogptcamembertcohereconditional_detrdata2vecdbrxdeformable_detrdpterniegemmagpt_neoxhubertinformerjambalayoutxlmmask2formermistralmixtralolmoowlv2persimmonphiphi3plbartqwen2qwen2_moerobertasegformersewstarcoder2switch_transformersunispeechvipllavawavlmyolos"
] | # coding=utf-8
# Copyright 2024 JetMoE AI and the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.... | https://github.com/myshell-ai/JetMoE/blob/0a9bc5a32386af9ea7abe9e32298976a94da908a/jetmoe/modeling_jetmoe.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/jetmoe/modular_jetmoe.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
... | # Copyright 2024 JetMoe AI and the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless re... | [
"llama",
"mixtral"
] |
layoutlm | microsoft/layoutlm-base-uncased | null | [] | # coding=utf-8
# Copyright 2018 The Microsoft Research Asia LayoutLM Team Authors and the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/... | https://github.com/huggingface/transformers/blob/c89bdfbe72/src/transformers/models/layoutlm/modeling_layoutlm.py | # Copyright 2018 The Microsoft Research Asia LayoutLM Team Authors and the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENS... | null | [] |
lfm2_moe | LiquidAI/LFM2-8B-A1B | 2025-10-07 | [
"aimv2apertusarceeariaaya_visionbambabiogptbitnetbltcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptedgetamedgetam_videoefficientloftremu3eomternieernie4_5ernie4_5_m... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/lfm2_moe/modular_lfm2_moe.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation... | # Copyright 2025 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | [
"lfm2",
"llama",
"mixtral",
"qwen2_moe"
] |
lfm2_vl | LiquidAI/LFM2-VL-1.6B | 2025-09-18 | [
"aimv2apertusarceeariaaya_visionbambabiogptbitnetcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptefficientloftremu3eomternieernie4_5ernie4_5_moeevollaexaone4falcon_h... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/lfm2_vl/modular_lfm2_vl.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation o... | # Copyright 2025 the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by appl... | [
"llava"
] |
llama | meta-llama/Meta-Llama-3.1-8B-Instruct | null | [] | # Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the terms described in the LICENSE file in
# top-level folder for each specific model found within the models/ directory at
# the top-level of this source tree.
# Copyright (c) Meta Platforms, Inc. and af... | https://github.com/meta-llama/llama-models/blob/0e0b8c519242d5833d8c11bffc1232b77ad7f301/models/llama3/model.py | # Copyright 2022 EleutherAI and the HuggingFace Inc. team. All rights reserved.
#
# This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX
# and OPT implementations in this library. It has been modified from its
# original forms to accommodate minor architectural differences compared
# to GPT-NeoX and OPT... | null | [] |
llama4 | meta-llama/Llama-4-Scout-17B-16E | null | [] | # Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the terms described in the LICENSE file in
# top-level folder for each specific model found within the models/ directory at
# the top-level of this source tree.
# Copyright (c) Meta Platforms, Inc. and af... | https://github.com/meta-llama/llama-models/blob/0e0b8c519242d5833d8c11bffc1232b77ad7f301/models/llama3/model.py | # Copyright 2025 The LLAMA4 and HuggingFace Inc. team. All rights reserved.
#
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless req... | null | [] |
llava | llava-hf/bakLlava-v1-hf | null | [] | # coding=utf-8
# Copyright 2023 the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | https://github.com/huggingface/transformers/blob/44b5506d29/src/transformers/models/llava/modeling_llava.py | # Copyright 2023 the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by appl... | null | [] |
llava_next | llava-hf/llava-v1.6-mistral-7b-hf | null | [] | # coding=utf-8
# Copyright 2024 the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | https://github.com/huggingface/transformers/blob/d91fd7f92c/src/transformers/models/llava_next/modeling_llava_next.py | # Copyright 2024 the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by appl... | null | [] |
llava_next_video | llava-hf/LLaVA-NeXT-Video-7B-hf | 2024-06-26 | [
"biogptcamembertcohereconditional_detrdata2vecdbrxdeformable_detrdpterniegemmagpt_neoxhubertinformerinstructblipvideojambajetmoelayoutxlmmask2formermistralmixtralolmoowlv2persimmonphiphi3plbartqwen2qwen2_moerobertart_detrsegformersewstarcoder2switch_transformersunispeechvipllavawavlmyolos"
] | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/llava_next_video/modular_llava_next_video.py.
# Do NOT edit this file manually as any edits will be overwritten b... | # Copyright 2024 HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicab... | [] |
llava_onevision | llava-hf/llava-onevision-qwen2-0.5b-ov-hf | 2024-09-05 | [
"biogptcamembertcohereconditional_detrdata2vecdbrxdeformable_detrdpterniefalcon_mambagemmagemma2gpt_neoxgranitehubertinformerinstructblipvideojambajetmoelayoutxlmllava_next_videomask2formermistralmixtralolmoolmoeowlv2persimmonphiphi3plbartqwen2qwen2_moerobertart_detrsegformersewstarcoder2switch_transformersunispeec... | # coding=utf-8
# Copyright 2024 the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | https://github.com/huggingface/transformers/blob/43df47d8e7/src/transformers/models/llava_onevision/modeling_llava_onevision.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/llava_onevision/modular_llava_onevision.py.
# Do NOT edit this file manually as any edits will be overwritten by ... | # Copyright 2024 the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by appl... | [] |
longformer | patrickvonplaten/longformer-random-tiny | null | [] | # coding=utf-8
# Copyright 2020 The Allen Institute for AI team and The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2... | https://github.com/huggingface/transformers/blob/c89bdfbe72/src/transformers/models/longformer/modeling_longformer.py | # Copyright 2020 The Allen Institute for AI team and The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | null | [] |
mamba | hf-internal-testing/mamba-130m | null | [] | # coding=utf-8
# Copyright 2024 state-spaces/mamba org and HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless... | https://github.com/huggingface/transformers/blob/fb1c62e973/src/transformers/models/mamba/modeling_mamba.py | # Copyright 2024 state-spaces/mamba org and HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by ap... | null | [] |
markuplm | microsoft/markuplm-base | null | [] | # coding=utf-8
# Copyright 2022 Microsoft Research Asia and the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# U... | https://github.com/huggingface/transformers/blob/f3d2f7a6e0/src/transformers/models/markuplm/modeling_markuplm.py | # Copyright 2022 Microsoft Research Asia and the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required ... | null | [] |
ministral | mistralai/Ministral-8B-Instruct-2410 | 2025-09-11 | [
"aimv2apertusarceeariaaya_visionbambabiogptbitnetcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptefficientloftremu3eomternieernie4_5ernie4_5_moeevollaexaone4falcon_h... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/ministral/modular_ministral.py.
# Do NOT edit this file manually as any edits will be overwritten by the generati... | import torch
from torch import nn
from ...cache_utils import Cache, DynamicCache
from ...configuration_utils import PreTrainedConfig
from ...masking_utils import create_causal_mask, create_sliding_window_causal_mask
from ...modeling_outputs import BaseModelOutputWithPast
from ...modeling_rope_utils import RopeParamete... | [
"qwen2"
] |
ministral3 | mistralai/Ministral-3-3B-Instruct-2512 | 2025-12-01 | [
"afmoeaimv2apertusarceeariaaudioflamingo3aya_visionbambabiogptbitnetbltcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmcwmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptedgetamedgetam_videoefficientloftremu3eomte... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/ministral3/modular_ministral3.py.
# Do NOT edit this file manually as any edits will be overwritten by the genera... | from collections.abc import Callable
import torch
from ...cache_utils import Cache
from ...modeling_flash_attention_utils import FlashAttentionKwargs
from ...modeling_layers import (
GenericForQuestionAnswering,
GenericForSequenceClassification,
GenericForTokenClassification,
)
from ...modeling_utils impo... | [
"mistral"
] |
mistral | mistralai/Mistral-7B-v0.1 | 2023-09-27 | [
"biogptcamembertconditional_detrdata2vecdeformable_detrdpterniegpt_neoxhubertinformerlayoutxlmmask2formerpersimmonplbartrobertasegformersewswitch_transformersunispeechwavlmyolos"
] | # coding=utf-8
# Copyright 2023 Mistral AI and the HuggingFace Inc. team. All rights reserved.
#
# This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX
# and OPT implementations in this library. It has been modified from its
# original forms to accommodate minor architectural differences compared
# to G... | https://github.com/huggingface/transformers/blob/72958fcd3c/src/transformers/models/mistral/modeling_mistral.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/mistral/modular_mistral.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation o... | from collections.abc import Callable
import torch
from torch import nn
from ...cache_utils import Cache, DynamicCache
from ...masking_utils import create_causal_mask, create_sliding_window_causal_mask
from ...modeling_flash_attention_utils import FlashAttentionKwargs
from ...modeling_layers import (
GenericForQue... | [
"llama"
] |
mm_grounding_dino | openmmlab-community/mm_grounding_dino_tiny_o365v1_goldg_v3det | null | [] | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/mm_grounding_dino/modular_mm_grounding_dino.py.
# Do NOT edit this file manually as any edits will be overwritten... | # Copyright 2025 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed ... | [
"grounding_dino"
] |
mobilebert | google/mobilebert-uncased | null | [] | # MIT License
#
# Copyright (c) 2020 The Google AI Language Team Authors, The HuggingFace Inc. team and github/lonePatient
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restricti... | https://github.com/huggingface/transformers/blob/c89bdfbe72/src/transformers/models/mobilebert/modeling_mobilebert.py | # MIT License
#
# Copyright (c) 2020 The Google AI Language Team Authors, The HuggingFace Inc. team and github/lonePatient
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restricti... | null | [] |
modernbert_decoder | blab-jhu/test-32m-dec | null | [] | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/modernbert_decoder/modular_modernbert_decoder.py.
# Do NOT edit this file manually as any edits will be overwritt... | # Copyright 2025 Johns Hopkins University, LightOn, and the HuggingFace Inc. team. All rights reserved.
#
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/license... | [
"modernbert"
] |
moonshine | UsefulSensors/moonshine-tiny | 2025-01-10 | [
"ariabambabiogptcamembertcoherecohere2colpaliconditional_detrdata2vecdbrxdeformable_detrdiffllamadinov2_with_registersdpterniefalcon_mambagemmagemma2glmgpt_neoxgranitegranitemoehubertijepainformerinstructblipvideojambajetmoelayoutxlmllava_next_videollava_onevisionmask2formermistralmixtralmodernbertolmoolmo2olmoeowl... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/moonshine/modular_moonshine.py.
# Do NOT edit this file manually as any edits will be overwritten by the generati... | # Copyright 2025 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by appl... | [
"glm",
"llama",
"whisper"
] |
moonshine_streaming | UsefulSensors/moonshine-streaming-tiny | 2026-02-03 | [
"afmoeaimv2apertusarceeariaaudioflamingo3aya_visionbambabiogptbitnetbltcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmcwmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptedgetamedgetam_videoefficientloftremu3eomte... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/moonshine_streaming/modular_moonshine_streaming.py.
# Do NOT edit this file manually as any edits will be overwri... | # Copyright 2026 the HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | [
"llama",
"moonshine"
] |
mpnet | microsoft/mpnet-base | null | [] | # coding=utf-8
# Copyright 2018 The HuggingFace Inc. team, Microsoft Corporation.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License ... | https://github.com/huggingface/transformers/blob/df2af6d8b8/src/transformers/models/mpnet/modeling_mpnet.py | # Copyright 2018 The HuggingFace Inc. team, Microsoft Corporation.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http... | null | [] |
olmo | allenai/OLMo-1B-hf | 2024-04-17 | [
"biogptcamembertcohereconditional_detrdata2vecdeformable_detrdpterniegemmagpt_neoxhubertinformerlayoutxlmmask2formermistralmixtralowlv2persimmonphiplbartqwen2qwen2_moerobertasegformersewstarcoder2switch_transformersunispeechvipllavawavlmyolos"
] | # coding=utf-8
# Copyright 2024 EleutherAI and the HuggingFace Inc. team. All rights reserved.
#
# This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX
# and OPT implementations in this library. It has been modified from its
# original forms to accommodate minor architectural differences compared
# to G... | https://github.com/huggingface/transformers/blob/e4ea19b958/src/transformers/models/olmo/modeling_olmo.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/olmo/modular_olmo.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# ... | # Copyright 2024 HuggingFace Inc. team. All rights reserved.
#
# This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX
# and OPT implementations in this library. It has been modified from its
# original forms to accommodate minor architectural differences compared
# to GPT-NeoX and OPT used by the Meta A... | [
"llama"
] |
olmo2 | allenai/OLMo-2-0425-1B | 2024-11-25 | [
"biogptcamembertcohereconditional_detrdata2vecdbrxdeformable_detrdpterniefalcon_mambagemmagemma2glmgpt_neoxgranitegranitemoehubertinformerinstructblipvideojambajetmoelayoutxlmllava_next_videollava_onevisionmask2formermistralmixtralolmoolmoeowlv2persimmonphiphi3phimoeplbartqwen2qwen2_moerobertart_detrsegformersewsta... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/olmo2/modular_olmo2.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# ... | # Copyright 2024 HuggingFace Inc. team. All rights reserved.
#
# This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX
# and OPT implementations in this library. It has been modified from its
# original forms to accommodate minor architectural differences compared
# to GPT-NeoX and OPT used by the Meta A... | [
"llama",
"olmo"
] |
olmo3 | shanearora/2025-sep-a-base-model | 2025-09-16 | [
"aimv2apertusarceeariaaya_visionbambabiogptbitnetcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptefficientloftremu3eomternieernie4_5ernie4_5_moeevollaexaone4falcon_h... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/olmo3/modular_olmo3.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# ... | # Copyright 2025 the HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | [
"gemma2",
"olmo2"
] |
olmoe | allenai/OLMoE-1B-7B-0924 | 2024-09-03 | [
"biogptcamembertcohereconditional_detrdata2vecdbrxdeformable_detrdpterniefalcon_mambagemmagemma2gpt_neoxgranitehubertinformerinstructblipvideojambajetmoelayoutxlmllava_next_videomask2formermistralmixtralolmoowlv2persimmonphiphi3plbartqwen2qwen2_moerobertart_detrsegformersewstarcoder2switch_transformersunispeechvipl... | # Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under th... | https://github.com/huggingface/transformers/blob/ecd61c6286/src/transformers/models/olmoe/modeling_olmoe.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/olmoe/modular_olmoe.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# ... | # Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under th... | [
"gemma",
"llama",
"mixtral",
"qwen2_moe"
] |
omdet_turbo | omlab/omdet-turbo-swin-tiny-hf | null | [] | # coding=utf-8
# Copyright 2024 Om Research Lab and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICE... | https://github.com/huggingface/transformers/blob/94f18cf23c/src/transformers/models/omdet_turbo/modeling_omdet_turbo.py | # Copyright 2024 Om Research Lab and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unl... | null | [] |
ovis2 | thisisiron/Ovis2-2B-hf | 2025-08-18 | [
"aimv2arceeariaaya_visionbambabiogptbitnetcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptefficientloftremu3eomternieernie4_5ernie4_5_moeevollaexaone4falcon_h1falcon... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/ovis2/modular_ovis2.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# ... | # Copyright 2025 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by appl... | [
"aimv2",
"llama",
"llava",
"llava_next",
"siglip"
] |
paddleocr_vl | PaddlePaddle/PaddleOCR-VL | 2025-12-10 | [
"afmoeaimv2apertusarceeariaaudioflamingo3aya_visionbambabiogptbitnetbltcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmcwmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptedgetamedgetam_videoefficientloftremu3eomte... | # Copyright (c) 2025 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by appli... | https://huggingface.co/PaddlePaddle/PaddleOCR-VL/blob/main/modeling_paddleocr_vl.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/paddleocr_vl/modular_paddleocr_vl.py.
# Do NOT edit this file manually as any edits will be overwritten by the ge... | # Copyright 2025 The PaddlePaddle Team and The HuggingFace Inc. team. All rights reserved.
#
# This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX
# and OPT implementations in this library. It has been modified from its
# original forms to accommodate minor architectural differences compared
# to GPT-N... | [
"ernie4_5",
"qwen2_5_omni",
"qwen2_vl",
"siglip",
"video_llama_3"
] |
paligemma | google/paligemma-3b-pt-224 | null | [] | # coding=utf-8
# Copyright 2024 the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | https://github.com/huggingface/transformers/blob/1360801a69/src/transformers/models/paligemma/modeling_paligemma.py | # Copyright 2024 the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by appl... | null | [] |
pe_audio_video | facebook/pe-av-large | 2025-12-16 | [
"afmoeaimv2apertusarceeariaaudioflamingo3aya_visionbambabiogptbitnetbltcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmcwmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptedgetamedgetam_videoefficientloftremu3eomte... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/pe_audio_video/modular_pe_audio_video.py.
# Do NOT edit this file manually as any edits will be overwritten by th... | # Copyright 2025 the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by appl... | [
"qwen3"
] |
persimmon | adept/persimmon-8b-chat | 2023-09-12 | [
"biogptcamembertconditional_detrdata2vecdeformable_detrdpterniegpt_neoxhubertinformerlayoutxlmmask2formerplbartrobertasegformersewswitch_transformersunispeechwavlmyolos"
] | # coding=utf-8
# Copyright 2023 EleutherAI and the HuggingFace Inc. team. All rights reserved.
#
# This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX
# and OPT implementations in this library. It has been modified from its
# original forms to accommodate minor architectural differences compared
# to G... | https://github.com/huggingface/transformers/blob/9cccb3a838/src/transformers/models/persimmon/modeling_persimmon.py | # Copyright 2023 EleutherAI and the HuggingFace Inc. team. All rights reserved.
#
# This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX
# and OPT implementations in this library. It has been modified from its
# original forms to accommodate minor architectural differences compared
# to GPT-NeoX and OPT... | from collections.abc import Callable
from typing import Optional
import torch
from torch import nn
from ...activations import ACT2FN
from ...cache_utils import Cache, DynamicCache
from ...masking_utils import create_causal_mask
from ...modeling_flash_attention_utils import FlashAttentionKwargs
from ...modeling_output... | [
"llama"
] |
phi | microsoft/phi-1 | 2023-11-10 | [
"biogptcamembertconditional_detrdata2vecdeformable_detrdpterniegpt_neoxhubertinformerlayoutxlmmask2formermistralowlv2persimmonplbartrobertasegformersewswitch_transformersunispeechwavlmyolos"
] | from __future__ import annotations
import dataclasses
import itertools
import re
from dataclasses import dataclass
from enum import auto, Enum
from typing import TYPE_CHECKING
from typing_extensions import assert_never
from torchgen.utils import NamespaceHelper, OrderedSet
if TYPE_CHECKING:
from collections.abc... | https://github.com/pytorch/pytorch/blob/1de94372ec445490b2ad20b9376a84d9b6ef1f34/torchgen/model.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/phi/modular_phi.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# ... | from collections.abc import Callable
from typing import Optional
import torch
import torch.nn as nn
from ...cache_utils import Cache, DynamicCache
from ...masking_utils import create_causal_mask
from ...modeling_layers import GradientCheckpointingLayer
from ...modeling_outputs import (
BaseModelOutputWithPast,
)
... | [
"clip",
"llama"
] |
phi3 | microsoft/phi-3-mini-4k-instruct | 2024-04-24 | [
"biogptcamembertcohereconditional_detrdata2vecdbrxdeformable_detrdpterniegemmagpt_neoxhubertinformerjambalayoutxlmmask2formermistralmixtralolmoowlv2persimmonphiplbartqwen2qwen2_moerobertasegformersewstarcoder2switch_transformersunispeechvipllavawavlmyolos"
] | # coding=utf-8
# Copyright 2024 Microsoft and the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.... | https://huggingface.co/microsoft/phi-3-mini-4k-instruct/blob/main/modeling_phi3.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/phi3/modular_phi3.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# ... | # Copyright 2024 Microsoft and the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless re... | [
"mistral",
"phi"
] |
phimoe | microsoft/Phi-3.5-MoE-instruct | 2024-10-04 | [
"biogptcamembertcohereconditional_detrdata2vecdbrxdeformable_detrdpterniefalcon_mambagemmagemma2gpt_neoxgranitegranitemoehubertinformerinstructblipvideojambajetmoelayoutxlmllava_next_videollava_onevisionmask2formermistralmixtralolmoolmoeowlv2persimmonphiphi3plbartqwen2qwen2_moerobertart_detrsegformersewstarcoder2sw... | # coding=utf-8
# Copyright 2024 Microsoft and the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.... | https://huggingface.co/microsoft/Phi-3.5-MoE-instruct/blob/main/modeling_phimoe.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/phimoe/modular_phimoe.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
... | # Copyright 2024 Microsoft and the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless re... | [
"llama",
"mixtral"
] |
pixio | LiheYoung/pixio-vith16 | 2025-12-16 | [
"afmoeaimv2apertusarceeariaaudioflamingo3aya_visionbambabiogptbitnetbltcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmcwmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptedgetamedgetam_videoefficientloftremu3eomte... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/pixio/modular_pixio.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# ... | # Copyright 2025 Meta AI and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requ... | [
"dinov2",
"vit"
] |
qwen2 | Qwen/Qwen2-0.5B | 2024-01-17 | [
"biogptcamembertconditional_detrdata2vecdeformable_detrdpterniegpt_neoxhubertinformerlayoutxlmmask2formermistralmixtralowlv2persimmonphiplbartrobertasegformersewswitch_transformersunispeechvipllavawavlmyolos"
] | # coding=utf-8
# Copyright 2024 The Qwen team, Alibaba Group and the HuggingFace Inc. team. All rights reserved.
#
# This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX
# and OPT implementations in this library. It has been modified from its
# original forms to accommodate minor architectural differenc... | https://github.com/huggingface/transformers/blob/d6ffe74dfa/src/transformers/models/qwen2/modeling_qwen2.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/qwen2/modular_qwen2.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# ... | from collections.abc import Callable
import torch
from torch import nn
from ...cache_utils import Cache, DynamicCache
from ...masking_utils import create_causal_mask, create_sliding_window_causal_mask
from ...modeling_flash_attention_utils import FlashAttentionKwargs
from ...modeling_outputs import (
BaseModelOut... | [
"gemma2",
"llama",
"mistral"
] |
qwen2_moe | Qwen/Qwen1.5-MoE-A2.7B | 2024-03-27 | [
"biogptcamembertcohereconditional_detrdata2vecdeformable_detrdpterniegemmagpt_neoxhubertinformerlayoutxlmmask2formermistralmixtralowlv2persimmonphiplbartqwen2robertasegformersewstarcoder2switch_transformersunispeechvipllavawavlmyolos"
] | # coding=utf-8
# Copyright 2024 The Qwen team, Alibaba Group and the HuggingFace Inc. team. All rights reserved.
#
# This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX
# and OPT implementations in this library. It has been modified from its
# original forms to accommodate minor architectural differenc... | https://github.com/huggingface/transformers/blob/1c39974a4c/src/transformers/models/qwen2_moe/modeling_qwen2_moe.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/qwen2_moe/modular_qwen2_moe.py.
# Do NOT edit this file manually as any edits will be overwritten by the generati... | # Copyright 2024 The Qwen team, Alibaba Group and the HuggingFace Inc. team. All rights reserved.
#
# This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX
# and OPT implementations in this library. It has been modified from its
# original forms to accommodate minor architectural differences compared
# t... | [
"gemma",
"gemma2",
"llama",
"mixtral"
] |
qwen3 | Qwen/Qwen3-0.6B-Base | 2025-03-31 | [
"ariaaya_visionbambabiogptcamembertcoherecohere2colpaliconditional_detrdata2vecdbrxdeepseek_v3deformable_detrdiffllamadinov2_with_registersdptemu3erniefalcon_mambagemmagemma2gemma3glmgot_ocr2gpt_neoxgranitegranitemoegranitemoesharedheliumhubertijepainformerinstructblipvideojambajetmoelayoutxlmllava_next_videollava_... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/qwen3/modular_qwen3.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# ... | # Copyright 2025 The Qwen team, Alibaba Group and the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENS... | [
"gemma",
"llama",
"qwen2"
] |
qwen3_moe | Qwen/Qwen3-30B-A3B-Base | 2025-03-31 | [
"ariaaya_visionbambabiogptcamembertcoherecohere2colpaliconditional_detrdata2vecdbrxdeepseek_v3deformable_detrdiffllamadinov2_with_registersdptemu3erniefalcon_mambagemmagemma2gemma3glmgot_ocr2gpt_neoxgranitegranitemoegranitemoesharedheliumhubertijepainformerinstructblipvideojambajetmoelayoutxlmllava_next_videollava_... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/qwen3_moe/modular_qwen3_moe.py.
# Do NOT edit this file manually as any edits will be overwritten by the generati... | # Copyright 2025 The Qwen team, Alibaba Group and the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENS... | [
"llama",
"mixtral",
"qwen2_moe",
"qwen3"
] |
qwen3_vl_moe | Qwen/Qwen3-VL-30B-A3B-Instruct | 2025-09-15 | [
"aimv2apertusarceeariaaya_visionbambabiogptbitnetcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptefficientloftremu3eomternieernie4_5ernie4_5_moeevollaexaone4falcon_h... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/qwen3_vl_moe/modular_qwen3_vl_moe.py.
# Do NOT edit this file manually as any edits will be overwritten by the ge... | # Copyright 2025 The Qwen Team and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unles... | [
"qwen3_moe",
"qwen3_vl"
] |
sam2_video | facebook/sam2.1-hiera-tiny | 2025-08-14 | [
"aimv2arceeariaaya_visionbambabiogptbitnetcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptefficientloftremu3eomternieernie4_5ernie4_5_moeevollaexaone4falcon_h1falcon... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/sam2_video/modular_sam2_video.py.
# Do NOT edit this file manually as any edits will be overwritten by the genera... | # Copyright 2025 The Meta AI Authors and The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unle... | [
"sam2"
] |
sam3_tracker_video | facebook/sam3 | 2025-11-19 | [
"aimv2apertusarceeariaaudioflamingo3aya_visionbambabiogptbitnetbltcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmcwmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptedgetamedgetam_videoefficientloftremu3eomterniee... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/sam3_tracker_video/modular_sam3_tracker_video.py.
# Do NOT edit this file manually as any edits will be overwritt... | # Copyright 2025 the HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | [
"sam2_video"
] |
seed_oss | ByteDance-Seed/Seed-OSS-36B-Base | 2025-08-22 | [
"aimv2arceeariaaya_visionbambabiogptbitnetcamembertcoherecohere2cohere2_visioncolpalicolqwen2conditional_detrcsmd_finedata2vecdbrxdeepseek_v2deepseek_v3deepseek_vldeepseek_vl_hybriddeformable_detrdiadiffllamadinov2_with_registersdogedots1dptefficientloftremu3eomternieernie4_5ernie4_5_moeevollaexaone4falcon_h1falcon... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/seed_oss/modular_seed_oss.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation... | # Copyright 2025 Bytedance-Seed Ltd and the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# ... | [
"llama"
] |
sew | asapp/sew-tiny-100k | 2021-10-15 | [
"camemberthubertroberta"
] | # coding=utf-8
# Copyright 2021 ASAPP Inc. and the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2... | https://github.com/huggingface/transformers/blob/cd3166a8ed/src/transformers/models/sew/modeling_sew.py | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/sew/modular_sew.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# ... | # Copyright 2021 ASAPP Inc. and the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | [
"wav2vec2"
] |
sew_d | asapp/sew-d-tiny-100k | null | [] | # coding=utf-8
# Copyright 2021 ASAPP Inc. and the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2... | https://github.com/huggingface/transformers/blob/cd3166a8ed/src/transformers/models/sew_d/modeling_sew_d.py | # Copyright 2021 ASAPP Inc. and the HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | null | [] |
smolvlm | HuggingFaceTB/SmolVLM2-256M-Video-Instruct | 2025-02-20 | [
"ariabambabiogptcamembertcoherecohere2colpaliconditional_detrdata2vecdbrxdeformable_detrdiffllamadinov2_with_registersdptemu3erniefalcon_mambagemmagemma2glmgot_ocr2gpt_neoxgranitegranitemoegranitemoesharedheliumhubertijepainformerinstructblipvideojambajetmoelayoutxlmllava_next_videollava_onevisionmask2formermistral... | null | null | # 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/smolvlm/modular_smolvlm.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation o... | # Copyright 2025 the HuggingFace Inc. team. All rights reserved.
# Written by Orr Zohar
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# ... | [
"idefics3"
] |
stablelm | stabilityai/stablelm-3b-4e1t | null | [] | # coding=utf-8
# Copyright 2024 EleutherAI and the HuggingFace Inc. team. All rights reserved.
#
# This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX
# and OPT implementations in this library. It has been modified from its
# original forms to accommodate minor architectural differences compared
# to G... | https://github.com/huggingface/transformers/blob/de6029a059/src/transformers/models/stablelm/modeling_stablelm.py | # Copyright 2024 EleutherAI and the HuggingFace Inc. team. All rights reserved.
#
# This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX
# and OPT implementations in this library. It has been modified from its
# original forms to accommodate minor architectural differences compared
# to GPT-NeoX and OPT... | null | [] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.