In the class AttentionLayer (models/attention/decoders/attention_layer.py), the initial parameter does not include "sigmoid_smoothing". But, in models/attention/attention_seq2seq.py, it calls AttentionLayer with sigmoid_smoothing=self.sigmoid_smoothing (line 338). Which file should I modify? Thx!