Skip to content

Commit 463963b

Browse files
committed
default to one mem kv for compressed attn
1 parent dfd3b15 commit 463963b

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

native_sparse_attention_pytorch/native_sparse_attention.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -185,7 +185,7 @@ def __init__(
185185
selection_block_size,
186186
num_selected_blocks,
187187
kv_heads = None,
188-
num_compressed_mem_kv = 4,
188+
num_compressed_mem_kv = 1,
189189
norm = True,
190190
use_diff_topk = False,
191191
interpolated_importance_score = False,

pyproject.toml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
[project]
22
name = "native-sparse-attention-pytorch"
3-
version = "0.0.42"
3+
version = "0.0.43"
44
description = "Native Sparse Attention"
55
authors = [
66
{ name = "Phil Wang", email = "lucidrains@gmail.com" }

0 commit comments

Comments
 (0)