Skip to content

Commit 3eef68f

Browse files
committed
some intermittent issue with flex attention on sample, just disable and plan on writing own triton kernel
1 parent 9e884df commit 3eef68f

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

native_sparse_attention_pytorch/transformer.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -221,7 +221,7 @@ def sample(
221221
out,
222222
cache = cache,
223223
return_cache = True,
224-
disable_flex = not is_first,
224+
disable_flex = True,
225225
disable_triton_kernel = not is_first
226226
)
227227

pyproject.toml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
[project]
22
name = "native-sparse-attention-pytorch"
3-
version = "0.0.66"
3+
version = "0.0.68"
44
description = "Native Sparse Attention"
55
authors = [
66
{ name = "Phil Wang", email = "lucidrains@gmail.com" }

0 commit comments

Comments
 (0)