Upload transformer_engine_torch wheel
Browse files
.gitattributes
CHANGED
|
@@ -35,3 +35,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
flash-linear-attention/flash_linear_attention-0.4.2+cu126torch280-py3-none-any.whl filter=lfs diff=lfs merge=lfs -text
|
| 37 |
causal-conv1d/causal_conv1d-1.6.0+cu126torch280-cp311-cp311-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
flash-linear-attention/flash_linear_attention-0.4.2+cu126torch280-py3-none-any.whl filter=lfs diff=lfs merge=lfs -text
|
| 37 |
causal-conv1d/causal_conv1d-1.6.0+cu126torch280-cp311-cp311-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
|
| 38 |
+
transformer-engine-torch/transformer_engine_torch-2.10.0+cu126torch280cxx11abitrue-cp311-cp311-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
|
transformer-engine-torch/transformer_engine_torch-2.10.0+cu126torch280cxx11abitrue-cp311-cp311-linux_x86_64.whl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b4772e206f632a28df1310f63983b37317defc833fd518fc6cf5871eb9c4c9b3
|
| 3 |
+
size 638436
|