Skip to content

questions about who2com. #12

@drowning-in-codes

Description

@drowning-in-codes

I find that the method of fusing the features is not the same as in the paper that relies on the attention mechanism score to select the features of the corresponding agent to fuse .
In the code, during training, the fused features are obtained directly using the attention mechanism, and it seems that the returned agent number, i.e., action_argmax , is not used subsequently.

outputs, log_action, action_argmax = self.model(images, training=True)

And you can see that the use of torch.argmax during training does not match the thesis
if training:
action = torch.argmax(prob_action, dim=2)
return pred, prob_action, action

https://github.com/GT-RIPL/MultiAgentPerception/blob/4ef300547a7f7af2676a034f7cf742b009f57d99/ptsemseg/trainer.py#L384C1-L398C1
And instead of using commun_label during training, it was used as a metric during validation and testing.
I feel that the log_action, action_argmax are really meaningless, and the features obtained during training are not all the features of a vehicle, but are obtained by using the attention method. Instead of selecting all the features of a certain car, in the test, new features are added after the attention is used.
https://github.com/GT-RIPL/MultiAgentPerception/blob/4ef300547a7f7af2676a034f7cf742b009f57d99/ptsemseg/models/agent.py#L629C1-L651C56
image

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions