cross_view_transformers
cross_view_transformers copied to clipboard
The softmax attention do not use a cosine similarity
Thanks for your work. according to your code, softmax attention do not use a cosine similarity,Did I get something wrong?
# Dot product attention along cameras
dot = self.scale * torch.einsum('b n Q d, b n K d -> b n Q K', q, k)
dot = rearrange(dot, 'b n Q K -> b Q (n K)')
att = dot.softmax(dim=-1)