zbrl commited on
Commit
7d04250
·
verified ·
1 Parent(s): b208829

Upload flash-linear-attention/flash_linear_attention-0.4.2-py3-none-any.whl with huggingface_hub

Browse files
.gitattributes CHANGED
@@ -40,3 +40,4 @@ vllm/vllm-0.17.2.dev0+g95c0f928c.d20260318.cu128-cp311-cp311-linux_x86_64.whl fi
40
  causal_conv1d-1.6.1-cp312-cp312-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
41
  flash_linear_attention-0.4.2-py3-none-any.whl filter=lfs diff=lfs merge=lfs -text
42
  causal-conv1d/causal_conv1d-1.6.1-cp312-cp312-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
 
 
40
  causal_conv1d-1.6.1-cp312-cp312-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
41
  flash_linear_attention-0.4.2-py3-none-any.whl filter=lfs diff=lfs merge=lfs -text
42
  causal-conv1d/causal_conv1d-1.6.1-cp312-cp312-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
43
+ flash-linear-attention/flash_linear_attention-0.4.2-py3-none-any.whl filter=lfs diff=lfs merge=lfs -text
flash-linear-attention/flash_linear_attention-0.4.2-py3-none-any.whl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c08be006ce4dbe1be81f54938ee8e6fc7968cfba397c8d06c7669e97b8c44c0d
3
+ size 284661