We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent dfd3b15 commit 463963bCopy full SHA for 463963b
native_sparse_attention_pytorch/native_sparse_attention.py
@@ -185,7 +185,7 @@ def __init__(
185
selection_block_size,
186
num_selected_blocks,
187
kv_heads = None,
188
- num_compressed_mem_kv = 4,
+ num_compressed_mem_kv = 1,
189
norm = True,
190
use_diff_topk = False,
191
interpolated_importance_score = False,
pyproject.toml
@@ -1,6 +1,6 @@
1
[project]
2
name = "native-sparse-attention-pytorch"
3
-version = "0.0.42"
+version = "0.0.43"
4
description = "Native Sparse Attention"
5
authors = [
6
{ name = "Phil Wang", email = "lucidrains@gmail.com" }
0 commit comments