We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 1de7c94 commit 82a28beCopy full SHA for 82a28be
native_sparse_attention_pytorch/native_sparse_attention.py
@@ -225,7 +225,7 @@ def forward(
225
if self.use_diff_topk:
226
gates = selected_importance_values + (1. - selected_importance_values).detach()
227
228
- fmask = selected_importance_values > mask_value
+ fmask = selected_importance_values > 1e-10
229
230
fq = q
231
fk = k
pyproject.toml
@@ -1,6 +1,6 @@
1
[project]
2
name = "native-sparse-attention-pytorch"
3
-version = "0.0.5"
+version = "0.0.6"
4
description = "Native Sparse Attention"
5
authors = [
6
{ name = "Phil Wang", email = "lucidrains@gmail.com" }
0 commit comments