We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 39678de commit 1b3dd06Copy full SHA for 1b3dd06
native_sparse_attention_pytorch/native_sparse_attention.py
@@ -676,7 +676,8 @@ def forward(
676
fv = v
677
678
num_selected = min(num_selected, importance_scores.shape[-1] - 1)
679
-
+ has_selected_kv_for_fine_attn = num_selected > 0
680
+
681
if has_selected_kv_for_fine_attn:
682
683
# get the top-n kv segments for fine attention
pyproject.toml
@@ -1,6 +1,6 @@
1
[project]
2
name = "native-sparse-attention-pytorch"
3
-version = "0.1.11"
+version = "0.1.12"
4
description = "Native Sparse Attention"
5
authors = [
6
{ name = "Phil Wang", email = "lucidrains@gmail.com" }
0 commit comments