multi-head-attention topic
DeepXi
Deep Xi: A deep learning approach to a priori SNR estimation implemented in TensorFlow 2/Keras. For speech enhancement and robust ASR.
TranAD
[VLDB'22] Anomaly Detection using Transformers, self-conditioning and adversarial training.
dodrio
Exploring attention weights in transformer-based models with linguistic knowledge.
attentions
PyTorch implementation of some attentions for Deep Learning Researchers.
Various-Attention-mechanisms
This repository contain various types of attention mechanism like Bahdanau , Soft attention , Additive Attention , Hierarchical Attention etc in Pytorch, Tensorflow, Keras
Attention-Visualization
Visualization for simple attention and Google's multi-head attention.
VRP_DRL_MHA
"Attention, Learn to Solve Routing Problems!"[Kool+, 2019], Capacitated Vehicle Routing Problem solver
multi-head_self-attention
A Faster Pytorch Implementation of Multi-Head Self-Attention
Diversify-MHA
EMNLP 2018: Multi-Head Attention with Disagreement Regularization; NAACL 2019: Information Aggregation for Multi-Head Attention with Routing-by-Agreement
point-transformer
This is the official repository of the original Point Transformer architecture.