Praveen Narayanan
Praveen Narayanan
I am wondering if the attention score calculation could be vectorized to run per batch instead of running two loops (batch size x encoder time steps) The sections of code...
It should be .tar. See my logs ''' **(base) pytorch@deepDSP01:~/projects/interspeech2019/tmp$ ls** speech_tools-2.4-release.tar.gz **(base) pytorch@deepDSP01:~/projects/interspeech2019/tmp$ tar -xvzf speech_tools-2.4-release.tar.gz** gzip: stdin: not in gzip format tar: Child returned status 1 tar: Error...
I am setting up embeddings for use with tacotron. For this, do we need to 'align_embeddings' in inference code? It seems to me that we simply store the output of...
@rosinality I am wondering if we could generate the results in section 4.4. We should create sequential context from latent frames, so we need a scheme to process latent frames...
Could you put up code (pytorch/chainer) for the argmax calculation during inference time? From Tachibana-Uenoyama: https://arxiv.org/pdf/1710.08969.pdf I am encountering the same problem during synthesis that the attention curve 'breaks' randomly...
### Prerequisite - [X] I have searched [Issues](https://github.com/open-mmlab/mmdetection3d/issues) and [Discussions](https://github.com/open-mmlab/mmdetection3d/discussions) but cannot get the expected help. - [X] I have read the [FAQ documentation](https://mmdetection3d.readthedocs.io/en/latest/notes/faq.html) but cannot get the expected help....
Could you post the reference to the image linear attention paper? Is it this? https://openreview.net/pdf?id=KiFeuZu24k Also, could you report performance of various implementations?