Kernels
kernels-bot commited on
Commit
7987398
·
verified ·
1 Parent(s): 15c3ae3

Uploaded using `kernel-builder`.

Browse files
Files changed (26) hide show
  1. build/torch210-cxx11-cpu-x86_64-linux/{_flash_attn2_cpu_cf323a2.abi3.so → _flash_attn2_cpu_958fa11.abi3.so} +1 -1
  2. build/torch210-cxx11-cpu-x86_64-linux/_ops.py +3 -3
  3. build/torch210-cxx11-cu126-x86_64-linux/{_flash_attn2_cuda_cf323a2.abi3.so → _flash_attn2_cuda_958fa11.abi3.so} +2 -2
  4. build/torch210-cxx11-cu126-x86_64-linux/_ops.py +3 -3
  5. build/torch210-cxx11-cu128-x86_64-linux/{_flash_attn2_cuda_cf323a2.abi3.so → _flash_attn2_cuda_958fa11.abi3.so} +2 -2
  6. build/torch210-cxx11-cu128-x86_64-linux/_ops.py +3 -3
  7. build/torch210-cxx11-cu130-x86_64-linux/{_flash_attn2_cuda_cf323a2.abi3.so → _flash_attn2_cuda_958fa11.abi3.so} +2 -2
  8. build/torch210-cxx11-cu130-x86_64-linux/_ops.py +3 -3
  9. build/{torch211-cxx11-xpu20253-x86_64-linux/_flash_attn2_xpu_cf323a2.abi3.so → torch210-cxx11-xpu20253-x86_64-linux/_flash_attn2_xpu_958fa11.abi3.so} +2 -2
  10. build/torch210-cxx11-xpu20253-x86_64-linux/_ops.py +3 -3
  11. build/torch211-cxx11-cpu-x86_64-linux/{_flash_attn2_cpu_cf323a2.abi3.so → _flash_attn2_cpu_958fa11.abi3.so} +1 -1
  12. build/torch211-cxx11-cpu-x86_64-linux/_ops.py +3 -3
  13. build/torch211-cxx11-cu126-x86_64-linux/_flash_attn2_cuda_958fa11.abi3.so +3 -0
  14. build/torch211-cxx11-cu126-x86_64-linux/_flash_attn2_cuda_cf323a2.abi3.so +0 -3
  15. build/torch211-cxx11-cu126-x86_64-linux/_ops.py +3 -3
  16. build/torch211-cxx11-cu128-x86_64-linux/_flash_attn2_cuda_958fa11.abi3.so +3 -0
  17. build/torch211-cxx11-cu128-x86_64-linux/_flash_attn2_cuda_cf323a2.abi3.so +0 -3
  18. build/torch211-cxx11-cu128-x86_64-linux/_ops.py +3 -3
  19. build/torch211-cxx11-cu130-x86_64-linux/_flash_attn2_cuda_958fa11.abi3.so +3 -0
  20. build/torch211-cxx11-cu130-x86_64-linux/_flash_attn2_cuda_cf323a2.abi3.so +0 -3
  21. build/torch211-cxx11-cu130-x86_64-linux/_ops.py +3 -3
  22. build/{torch210-cxx11-xpu20253-x86_64-linux/_flash_attn2_xpu_cf323a2.abi3.so → torch211-cxx11-xpu20253-x86_64-linux/_flash_attn2_xpu_958fa11.abi3.so} +2 -2
  23. build/torch211-cxx11-xpu20253-x86_64-linux/_ops.py +3 -3
  24. build/torch29-cxx11-cu129-x86_64-linux/_flash_attn2_cuda_958fa11.abi3.so +3 -0
  25. build/torch29-cxx11-cu129-x86_64-linux/_flash_attn2_cuda_cf323a2.abi3.so +0 -3
  26. build/torch29-cxx11-cu129-x86_64-linux/_ops.py +3 -3
build/torch210-cxx11-cpu-x86_64-linux/{_flash_attn2_cpu_cf323a2.abi3.so → _flash_attn2_cpu_958fa11.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:917cce036be12194bbc9d832021d2de136665d4d878b503e486cad869eb1591e
3
  size 1942496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74d828d5319877c6e85346b032a63dc69662c869f60a7a197347e3bcc988fdce
3
  size 1942496
build/torch210-cxx11-cpu-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_attn2_cpu_cf323a2
3
- ops = torch.ops._flash_attn2_cpu_cf323a2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_attn2_cpu_cf323a2::{op_name}"
 
1
  import torch
2
+ from . import _flash_attn2_cpu_958fa11
3
+ ops = torch.ops._flash_attn2_cpu_958fa11
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_attn2_cpu_958fa11::{op_name}"
build/torch210-cxx11-cu126-x86_64-linux/{_flash_attn2_cuda_cf323a2.abi3.so → _flash_attn2_cuda_958fa11.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:48ca91811a6a8b7c99e40b8f912af5b5d5dfe8ea7268cf139b2b24cfff944e66
3
- size 448709016
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e37bb2606f115a03cf4dd9f5cd9d5e06b9d023d537f387ef2b229ee02893fe90
3
+ size 448709032
build/torch210-cxx11-cu126-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_attn2_cuda_cf323a2
3
- ops = torch.ops._flash_attn2_cuda_cf323a2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_attn2_cuda_cf323a2::{op_name}"
 
1
  import torch
2
+ from . import _flash_attn2_cuda_958fa11
3
+ ops = torch.ops._flash_attn2_cuda_958fa11
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_attn2_cuda_958fa11::{op_name}"
build/torch210-cxx11-cu128-x86_64-linux/{_flash_attn2_cuda_cf323a2.abi3.so → _flash_attn2_cuda_958fa11.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a7c37d176fcc8b05c5cc2cfb5fdb078535777826972ad20df9567234fbda46a6
3
- size 1037803408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2aae1002c587f0abdf264b56aa1d0c47a64a313a3ceb2307f43f989f9a5a616c
3
+ size 1037803416
build/torch210-cxx11-cu128-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_attn2_cuda_cf323a2
3
- ops = torch.ops._flash_attn2_cuda_cf323a2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_attn2_cuda_cf323a2::{op_name}"
 
1
  import torch
2
+ from . import _flash_attn2_cuda_958fa11
3
+ ops = torch.ops._flash_attn2_cuda_958fa11
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_attn2_cuda_958fa11::{op_name}"
build/torch210-cxx11-cu130-x86_64-linux/{_flash_attn2_cuda_cf323a2.abi3.so → _flash_attn2_cuda_958fa11.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cf31f7666af3a48c3d4f4d15a11380f0643c02c619d05ad382ba50716ded45b4
3
- size 1009055088
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:759db342b8a9d98fb2d15631005f34010c5ba5f70897cfc90c06c3848a1a66f0
3
+ size 1008998032
build/torch210-cxx11-cu130-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_attn2_cuda_cf323a2
3
- ops = torch.ops._flash_attn2_cuda_cf323a2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_attn2_cuda_cf323a2::{op_name}"
 
1
  import torch
2
+ from . import _flash_attn2_cuda_958fa11
3
+ ops = torch.ops._flash_attn2_cuda_958fa11
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_attn2_cuda_958fa11::{op_name}"
build/{torch211-cxx11-xpu20253-x86_64-linux/_flash_attn2_xpu_cf323a2.abi3.so → torch210-cxx11-xpu20253-x86_64-linux/_flash_attn2_xpu_958fa11.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9294ec725a5bb1290f3ec8e3375647dc63c95aaf1e09fd0ec4b6d4a578ee062a
3
- size 17668088
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd11c6f87ce851a737789cbb6fc14a36f46b9236e7cee41a053162d97af27da0
3
+ size 17672184
build/torch210-cxx11-xpu20253-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_attn2_xpu_cf323a2
3
- ops = torch.ops._flash_attn2_xpu_cf323a2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_attn2_xpu_cf323a2::{op_name}"
 
1
  import torch
2
+ from . import _flash_attn2_xpu_958fa11
3
+ ops = torch.ops._flash_attn2_xpu_958fa11
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_attn2_xpu_958fa11::{op_name}"
build/torch211-cxx11-cpu-x86_64-linux/{_flash_attn2_cpu_cf323a2.abi3.so → _flash_attn2_cpu_958fa11.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:252ad0d0d4c5163348a4216968519b9de393026ad4307e8f305bc42f853fe775
3
  size 1942496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af048a61898598d46886d7da0dea02669ecd79f2e56959b859f500db6f2ca54f
3
  size 1942496
build/torch211-cxx11-cpu-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_attn2_cpu_cf323a2
3
- ops = torch.ops._flash_attn2_cpu_cf323a2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_attn2_cpu_cf323a2::{op_name}"
 
1
  import torch
2
+ from . import _flash_attn2_cpu_958fa11
3
+ ops = torch.ops._flash_attn2_cpu_958fa11
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_attn2_cpu_958fa11::{op_name}"
build/torch211-cxx11-cu126-x86_64-linux/_flash_attn2_cuda_958fa11.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c3eaf804444dcaa2e78153b868f1b4823d227527748593a55416ce95122f2fc
3
+ size 448702024
build/torch211-cxx11-cu126-x86_64-linux/_flash_attn2_cuda_cf323a2.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:14a9dd127d4fc02189877b4cf2c158c4a8a2e4356b9b251987c3bffa11d56c68
3
- size 448702016
 
 
 
 
build/torch211-cxx11-cu126-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_attn2_cuda_cf323a2
3
- ops = torch.ops._flash_attn2_cuda_cf323a2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_attn2_cuda_cf323a2::{op_name}"
 
1
  import torch
2
+ from . import _flash_attn2_cuda_958fa11
3
+ ops = torch.ops._flash_attn2_cuda_958fa11
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_attn2_cuda_958fa11::{op_name}"
build/torch211-cxx11-cu128-x86_64-linux/_flash_attn2_cuda_958fa11.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3dfdbe3091ddff2369e9e218d547414d4c9e804d5bd5d030a2b604d4004b215c
3
+ size 1037792312
build/torch211-cxx11-cu128-x86_64-linux/_flash_attn2_cuda_cf323a2.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b2c63a5d07c3f6188690d37a22e76801309337f472257b7c2463831ff94213c8
3
- size 1037788208
 
 
 
 
build/torch211-cxx11-cu128-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_attn2_cuda_cf323a2
3
- ops = torch.ops._flash_attn2_cuda_cf323a2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_attn2_cuda_cf323a2::{op_name}"
 
1
  import torch
2
+ from . import _flash_attn2_cuda_958fa11
3
+ ops = torch.ops._flash_attn2_cuda_958fa11
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_attn2_cuda_958fa11::{op_name}"
build/torch211-cxx11-cu130-x86_64-linux/_flash_attn2_cuda_958fa11.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a7bb5c7706908160b38eaa7a3dc9988cbfbae28a112ea5d47be2f49923a7b2e
3
+ size 1008991032
build/torch211-cxx11-cu130-x86_64-linux/_flash_attn2_cuda_cf323a2.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:66365d9913090aa9486ebae89e1ed7eabba1de23b6417b9c3a3d2a624520fffb
3
- size 1009048080
 
 
 
 
build/torch211-cxx11-cu130-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_attn2_cuda_cf323a2
3
- ops = torch.ops._flash_attn2_cuda_cf323a2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_attn2_cuda_cf323a2::{op_name}"
 
1
  import torch
2
+ from . import _flash_attn2_cuda_958fa11
3
+ ops = torch.ops._flash_attn2_cuda_958fa11
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_attn2_cuda_958fa11::{op_name}"
build/{torch210-cxx11-xpu20253-x86_64-linux/_flash_attn2_xpu_cf323a2.abi3.so → torch211-cxx11-xpu20253-x86_64-linux/_flash_attn2_xpu_958fa11.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:18d46588b4aef9cef5ff36e74b4984aeb04bf00a4eb695fa979b9225b4fcc5be
3
- size 17668088
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2467ca71c6d81a058217f1f15fe0932e0c59a178315dcfbfc45a2c6f92dbb3c
3
+ size 17672184
build/torch211-cxx11-xpu20253-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_attn2_xpu_cf323a2
3
- ops = torch.ops._flash_attn2_xpu_cf323a2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_attn2_xpu_cf323a2::{op_name}"
 
1
  import torch
2
+ from . import _flash_attn2_xpu_958fa11
3
+ ops = torch.ops._flash_attn2_xpu_958fa11
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_attn2_xpu_958fa11::{op_name}"
build/torch29-cxx11-cu129-x86_64-linux/_flash_attn2_cuda_958fa11.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23c6b54d6c7c8219b94b7e313eb11df57e191e394f3c59c270c7aab5ed5382f0
3
+ size 1042952368
build/torch29-cxx11-cu129-x86_64-linux/_flash_attn2_cuda_cf323a2.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6233725460d8f96cbfda48ecc7ea7f9d4fe3100d30ecae612fc206f8cdca774d
3
- size 1042948264
 
 
 
 
build/torch29-cxx11-cu129-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_attn2_cuda_cf323a2
3
- ops = torch.ops._flash_attn2_cuda_cf323a2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_attn2_cuda_cf323a2::{op_name}"
 
1
  import torch
2
+ from . import _flash_attn2_cuda_958fa11
3
+ ops = torch.ops._flash_attn2_cuda_958fa11
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_attn2_cuda_958fa11::{op_name}"