-
Notifications
You must be signed in to change notification settings - Fork 81
Open
Description
Hi there,
Thanks for the project firstly. I saw a method called get_padding_mask in tf_utils.py file. It is combined with attention mask. It is designed to resolve the padding issue in batch inference or training only?
I see that with very a few padding in the left can still make good predictions/generations, but if use a large batch for inference, which cause a long left padding, the predictions get very incorrect. Do you have any suggestions for that?
Looking forward to your response. Thanks.
Metadata
Metadata
Assignees
Labels
No labels