Skip to content

Unfair comparison between ProtBert and ESM #9

@ww-rm

Description

@ww-rm

In ProtTrans, the author says that:

No auxiliary tasks like BERT's next-sentence prediction were used for any model described here.

But in the PEER, the [CLS] token is used for ProtBert as a protein-level embedding representation. In this case the [CLS] token may not have the ability to represent sequence embedding.

For ProtBert, should we use the same strategy as for ESM (i.e., mean pooling over all residues) to get a fairer comparison?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions