Upload 2 files
Browse filessage_attention2 and radial attention
.gitattributes
CHANGED
@@ -34,3 +34,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
flash_attn-2.7.4+cu128torch2.7-cp312-cp312-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
flash_attn-2.7.4+cu128torch2.7-cp312-cp312-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
|
37 |
+
sageattention-2.2.0-cp311-cp311-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
|
38 |
+
spas_sage_attn-0.1.0-cp311-cp311-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
|
sageattention-2.2.0-cp311-cp311-linux_x86_64.whl
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:aa2559487fa36ef19d84d905d7b978754db973ae1a93e1160633172d19413b7e
|
3 |
+
size 7886038
|
spas_sage_attn-0.1.0-cp311-cp311-linux_x86_64.whl
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b97dbdba4ecfcf47aeb12b929257089cd767c6103ff6225b22721a862425a7b3
|
3 |
+
size 9239714
|