Uploaded using `kernel-builder`.
Browse files- build/torch210-cxx11-cu126-aarch64-linux/{_flash_attn2_cuda_cf323a2.abi3.so → _flash_attn2_cuda_958fa11.abi3.so} +2 -2
- build/torch210-cxx11-cu126-aarch64-linux/_ops.py +3 -3
- build/torch210-cxx11-cu128-aarch64-linux/{_flash_attn2_cuda_cf323a2.abi3.so → _flash_attn2_cuda_958fa11.abi3.so} +2 -2
- build/torch210-cxx11-cu128-aarch64-linux/_ops.py +3 -3
- build/torch210-cxx11-cu130-aarch64-linux/{_flash_attn2_cuda_cf323a2.abi3.so → _flash_attn2_cuda_958fa11.abi3.so} +2 -2
- build/torch210-cxx11-cu130-aarch64-linux/_ops.py +3 -3
- build/torch211-cxx11-cu126-aarch64-linux/{_flash_attn2_cuda_cf323a2.abi3.so → _flash_attn2_cuda_958fa11.abi3.so} +2 -2
- build/torch211-cxx11-cu126-aarch64-linux/_ops.py +3 -3
- build/torch211-cxx11-cu128-aarch64-linux/_flash_attn2_cuda_958fa11.abi3.so +3 -0
- build/torch211-cxx11-cu128-aarch64-linux/_flash_attn2_cuda_cf323a2.abi3.so +0 -3
- build/torch211-cxx11-cu128-aarch64-linux/_ops.py +3 -3
- build/torch211-cxx11-cu130-aarch64-linux/_flash_attn2_cuda_958fa11.abi3.so +3 -0
- build/torch211-cxx11-cu130-aarch64-linux/_flash_attn2_cuda_cf323a2.abi3.so +0 -3
- build/torch211-cxx11-cu130-aarch64-linux/_ops.py +3 -3
- build/torch29-cxx11-cu129-aarch64-linux/_flash_attn2_cuda_958fa11.abi3.so +3 -0
- build/torch29-cxx11-cu129-aarch64-linux/_flash_attn2_cuda_cf323a2.abi3.so +0 -3
- build/torch29-cxx11-cu129-aarch64-linux/_ops.py +3 -3
build/torch210-cxx11-cu126-aarch64-linux/{_flash_attn2_cuda_cf323a2.abi3.so → _flash_attn2_cuda_958fa11.abi3.so}
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6fc8bd519ff37c3e79c24bd6bd252352b01f070fa8f7c1a8432c45f2bc22b52f
|
| 3 |
+
size 448533504
|
build/torch210-cxx11-cu126-aarch64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _flash_attn2_cuda_958fa11
|
| 3 |
+
ops = torch.ops._flash_attn2_cuda_958fa11
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_flash_attn2_cuda_958fa11::{op_name}"
|
build/torch210-cxx11-cu128-aarch64-linux/{_flash_attn2_cuda_cf323a2.abi3.so → _flash_attn2_cuda_958fa11.abi3.so}
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:97e690df2995d1d87a4cc3f8eb17f0d3ef80fc36796ef6d593f13c7e6aedf3fd
|
| 3 |
+
size 1037990952
|
build/torch210-cxx11-cu128-aarch64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _flash_attn2_cuda_958fa11
|
| 3 |
+
ops = torch.ops._flash_attn2_cuda_958fa11
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_flash_attn2_cuda_958fa11::{op_name}"
|
build/torch210-cxx11-cu130-aarch64-linux/{_flash_attn2_cuda_cf323a2.abi3.so → _flash_attn2_cuda_958fa11.abi3.so}
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:26dbdabe7ffa5834612cf37913f637c91aa991d757f4e42b4876c1921ecbeeb7
|
| 3 |
+
size 1008644888
|
build/torch210-cxx11-cu130-aarch64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _flash_attn2_cuda_958fa11
|
| 3 |
+
ops = torch.ops._flash_attn2_cuda_958fa11
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_flash_attn2_cuda_958fa11::{op_name}"
|
build/torch211-cxx11-cu126-aarch64-linux/{_flash_attn2_cuda_cf323a2.abi3.so → _flash_attn2_cuda_958fa11.abi3.so}
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4c35ba9816d590f803a80a45fcb5be581dfceb15af4f86daf3758ba46f56ef3b
|
| 3 |
+
size 448529832
|
build/torch211-cxx11-cu126-aarch64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _flash_attn2_cuda_958fa11
|
| 3 |
+
ops = torch.ops._flash_attn2_cuda_958fa11
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_flash_attn2_cuda_958fa11::{op_name}"
|
build/torch211-cxx11-cu128-aarch64-linux/_flash_attn2_cuda_958fa11.abi3.so
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cef2b4d449b3efb635e6226e2d0257b080677daa631e732c73905f94ea9bdb7c
|
| 3 |
+
size 1037987184
|
build/torch211-cxx11-cu128-aarch64-linux/_flash_attn2_cuda_cf323a2.abi3.so
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:d793922729ee7475e37c6a291ecb099a6e00d5dddf50de3c8546ec654ef0d664
|
| 3 |
-
size 1037987168
|
|
|
|
|
|
|
|
|
|
|
|
build/torch211-cxx11-cu128-aarch64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _flash_attn2_cuda_958fa11
|
| 3 |
+
ops = torch.ops._flash_attn2_cuda_958fa11
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_flash_attn2_cuda_958fa11::{op_name}"
|
build/torch211-cxx11-cu130-aarch64-linux/_flash_attn2_cuda_958fa11.abi3.so
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:27fa941b83482a166cbd246e95694161e89557eef2a5b284016d3e4765dce49a
|
| 3 |
+
size 1008641088
|
build/torch211-cxx11-cu130-aarch64-linux/_flash_attn2_cuda_cf323a2.abi3.so
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:7a25c1e807584eb384fbd7dcde4dfa0a71370f48905831435ca412a1ba34dae4
|
| 3 |
-
size 1008640568
|
|
|
|
|
|
|
|
|
|
|
|
build/torch211-cxx11-cu130-aarch64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _flash_attn2_cuda_958fa11
|
| 3 |
+
ops = torch.ops._flash_attn2_cuda_958fa11
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_flash_attn2_cuda_958fa11::{op_name}"
|
build/torch29-cxx11-cu129-aarch64-linux/_flash_attn2_cuda_958fa11.abi3.so
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b3363c49d10c61078a9a05359fef626f4ed3b0ee0f68eb5a88828823bfcfe651
|
| 3 |
+
size 1043164184
|
build/torch29-cxx11-cu129-aarch64-linux/_flash_attn2_cuda_cf323a2.abi3.so
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:8770d211cec7c9e9bf54015df64cef7c94e7641eb37b3a896d05d6d207fad9f6
|
| 3 |
-
size 1043164168
|
|
|
|
|
|
|
|
|
|
|
|
build/torch29-cxx11-cu129-aarch64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _flash_attn2_cuda_958fa11
|
| 3 |
+
ops = torch.ops._flash_attn2_cuda_958fa11
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_flash_attn2_cuda_958fa11::{op_name}"
|