-
Notifications
You must be signed in to change notification settings - Fork 12
Description
I find that the method of fusing the features is not the same as in the paper that relies on the attention mechanism score to select the features of the corresponding agent to fuse .
In the code, during training, the fused features are obtained directly using the attention mechanism, and it seems that the returned agent number, i.e., action_argmax , is not used subsequently.
MultiAgentPerception/ptsemseg/trainer.py
Line 391 in 4ef3005
outputs, log_action, action_argmax = self.model(images, training=True) |
And you can see that the use of torch.argmax during training does not match the thesis
MultiAgentPerception/ptsemseg/models/agent.py
Lines 625 to 627 in 4ef3005
if training: | |
action = torch.argmax(prob_action, dim=2) | |
return pred, prob_action, action |
https://github.com/GT-RIPL/MultiAgentPerception/blob/4ef300547a7f7af2676a034f7cf742b009f57d99/ptsemseg/trainer.py#L384C1-L398C1
And instead of using commun_label during training, it was used as a metric during validation and testing.
I feel that the log_action, action_argmax are really meaningless, and the features obtained during training are not all the features of a vehicle, but are obtained by using the attention method. Instead of selecting all the features of a certain car, in the test, new features are added after the attention is used.
https://github.com/GT-RIPL/MultiAgentPerception/blob/4ef300547a7f7af2676a034f7cf742b009f57d99/ptsemseg/models/agent.py#L629C1-L651C56