Upload flash-linear-attention/flash_linear_attention-0.4.2-py3-none-any.whl with huggingface_hub
Browse files
.gitattributes
CHANGED
|
@@ -40,3 +40,4 @@ vllm/vllm-0.17.2.dev0+g95c0f928c.d20260318.cu128-cp311-cp311-linux_x86_64.whl fi
|
|
| 40 |
causal_conv1d-1.6.1-cp312-cp312-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
|
| 41 |
flash_linear_attention-0.4.2-py3-none-any.whl filter=lfs diff=lfs merge=lfs -text
|
| 42 |
causal-conv1d/causal_conv1d-1.6.1-cp312-cp312-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 40 |
causal_conv1d-1.6.1-cp312-cp312-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
|
| 41 |
flash_linear_attention-0.4.2-py3-none-any.whl filter=lfs diff=lfs merge=lfs -text
|
| 42 |
causal-conv1d/causal_conv1d-1.6.1-cp312-cp312-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
|
| 43 |
+
flash-linear-attention/flash_linear_attention-0.4.2-py3-none-any.whl filter=lfs diff=lfs merge=lfs -text
|
flash-linear-attention/flash_linear_attention-0.4.2-py3-none-any.whl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c08be006ce4dbe1be81f54938ee8e6fc7968cfba397c8d06c7669e97b8c44c0d
|
| 3 |
+
size 284661
|