Kernels
kernels-bot commited on
Commit
15c3ae3
·
verified ·
1 Parent(s): 0cf244b

Uploaded using `kernel-builder`.

Browse files
build/torch210-cxx11-cu126-aarch64-linux/{_flash_attn2_cuda_cf323a2.abi3.so → _flash_attn2_cuda_958fa11.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cde7daedb67d42ed96943d7ae16c3f6411749fa995a98294562a51f49c8743ca
3
- size 448533496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6fc8bd519ff37c3e79c24bd6bd252352b01f070fa8f7c1a8432c45f2bc22b52f
3
+ size 448533504
build/torch210-cxx11-cu126-aarch64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_attn2_cuda_cf323a2
3
- ops = torch.ops._flash_attn2_cuda_cf323a2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_attn2_cuda_cf323a2::{op_name}"
 
1
  import torch
2
+ from . import _flash_attn2_cuda_958fa11
3
+ ops = torch.ops._flash_attn2_cuda_958fa11
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_attn2_cuda_958fa11::{op_name}"
build/torch210-cxx11-cu128-aarch64-linux/{_flash_attn2_cuda_cf323a2.abi3.so → _flash_attn2_cuda_958fa11.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:07c63011fa51e209e141099a499f7f8a77b41532a05206c12d8c61456d1f55d3
3
- size 1037990944
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97e690df2995d1d87a4cc3f8eb17f0d3ef80fc36796ef6d593f13c7e6aedf3fd
3
+ size 1037990952
build/torch210-cxx11-cu128-aarch64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_attn2_cuda_cf323a2
3
- ops = torch.ops._flash_attn2_cuda_cf323a2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_attn2_cuda_cf323a2::{op_name}"
 
1
  import torch
2
+ from . import _flash_attn2_cuda_958fa11
3
+ ops = torch.ops._flash_attn2_cuda_958fa11
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_attn2_cuda_958fa11::{op_name}"
build/torch210-cxx11-cu130-aarch64-linux/{_flash_attn2_cuda_cf323a2.abi3.so → _flash_attn2_cuda_958fa11.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ea3363b843d2dcaf53d49de477864680412e731cb95f262701333d5a78d7f058
3
- size 1008644344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26dbdabe7ffa5834612cf37913f637c91aa991d757f4e42b4876c1921ecbeeb7
3
+ size 1008644888
build/torch210-cxx11-cu130-aarch64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_attn2_cuda_cf323a2
3
- ops = torch.ops._flash_attn2_cuda_cf323a2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_attn2_cuda_cf323a2::{op_name}"
 
1
  import torch
2
+ from . import _flash_attn2_cuda_958fa11
3
+ ops = torch.ops._flash_attn2_cuda_958fa11
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_attn2_cuda_958fa11::{op_name}"
build/torch211-cxx11-cu126-aarch64-linux/{_flash_attn2_cuda_cf323a2.abi3.so → _flash_attn2_cuda_958fa11.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1cb63e59f47fda4a50429592acfcb200e0acb248fc8deb431d2a0c77037c0a05
3
- size 448529824
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c35ba9816d590f803a80a45fcb5be581dfceb15af4f86daf3758ba46f56ef3b
3
+ size 448529832
build/torch211-cxx11-cu126-aarch64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_attn2_cuda_cf323a2
3
- ops = torch.ops._flash_attn2_cuda_cf323a2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_attn2_cuda_cf323a2::{op_name}"
 
1
  import torch
2
+ from . import _flash_attn2_cuda_958fa11
3
+ ops = torch.ops._flash_attn2_cuda_958fa11
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_attn2_cuda_958fa11::{op_name}"
build/torch211-cxx11-cu128-aarch64-linux/_flash_attn2_cuda_958fa11.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cef2b4d449b3efb635e6226e2d0257b080677daa631e732c73905f94ea9bdb7c
3
+ size 1037987184
build/torch211-cxx11-cu128-aarch64-linux/_flash_attn2_cuda_cf323a2.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d793922729ee7475e37c6a291ecb099a6e00d5dddf50de3c8546ec654ef0d664
3
- size 1037987168
 
 
 
 
build/torch211-cxx11-cu128-aarch64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_attn2_cuda_cf323a2
3
- ops = torch.ops._flash_attn2_cuda_cf323a2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_attn2_cuda_cf323a2::{op_name}"
 
1
  import torch
2
+ from . import _flash_attn2_cuda_958fa11
3
+ ops = torch.ops._flash_attn2_cuda_958fa11
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_attn2_cuda_958fa11::{op_name}"
build/torch211-cxx11-cu130-aarch64-linux/_flash_attn2_cuda_958fa11.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27fa941b83482a166cbd246e95694161e89557eef2a5b284016d3e4765dce49a
3
+ size 1008641088
build/torch211-cxx11-cu130-aarch64-linux/_flash_attn2_cuda_cf323a2.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a25c1e807584eb384fbd7dcde4dfa0a71370f48905831435ca412a1ba34dae4
3
- size 1008640568
 
 
 
 
build/torch211-cxx11-cu130-aarch64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_attn2_cuda_cf323a2
3
- ops = torch.ops._flash_attn2_cuda_cf323a2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_attn2_cuda_cf323a2::{op_name}"
 
1
  import torch
2
+ from . import _flash_attn2_cuda_958fa11
3
+ ops = torch.ops._flash_attn2_cuda_958fa11
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_attn2_cuda_958fa11::{op_name}"
build/torch29-cxx11-cu129-aarch64-linux/_flash_attn2_cuda_958fa11.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3363c49d10c61078a9a05359fef626f4ed3b0ee0f68eb5a88828823bfcfe651
3
+ size 1043164184
build/torch29-cxx11-cu129-aarch64-linux/_flash_attn2_cuda_cf323a2.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8770d211cec7c9e9bf54015df64cef7c94e7641eb37b3a896d05d6d207fad9f6
3
- size 1043164168
 
 
 
 
build/torch29-cxx11-cu129-aarch64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_attn2_cuda_cf323a2
3
- ops = torch.ops._flash_attn2_cuda_cf323a2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_attn2_cuda_cf323a2::{op_name}"
 
1
  import torch
2
+ from . import _flash_attn2_cuda_958fa11
3
+ ops = torch.ops._flash_attn2_cuda_958fa11
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_attn2_cuda_958fa11::{op_name}"