79688927

Date: 2025-07-03 13:37:34
Score: 3
Natty:
Report link

I got the answer here https://discuss.pytorch.org/t/multi-head-self-attention-in-transformer-is-permutation-invariant-or-equivariant-how-to-see-it-in-practice/221249/2

the correct evaluation is

torch.allclose(y0[1], y1[0], atol=1e-6)

wich evaluate as True

Reasons:
  • Probably link only (1):
  • Low length (1):
  • Has code block (-0.5):
  • Self-answer (0.5):
  • Low reputation (1):
Posted by: fenaux