Cydral
Cydral
> Sorry I took so long to come back to this. Been a busy few weeks :| > > Anyway, I just pulled this and tried to run the unit...
I'm cancelling this PR because I've merged all the changes with the other PR containing the new definitions for multm_prev.
We can change the name without any problem. I am already dealing with compilation issues likely due to static uses of mult_prev_ (in the template part), and we will decide...
On the other hand, I was thinking of using the same convention for the transformation to be applied to softmax and thus have a special layer named softmaxm. So we...
It would be simpler to use for some people, but we would lose the flexibility to build attention in a potentially specific way (even though it currently follows fairly standard...
Indeed, I can add a high-level declaration in the layer definition file, similar to what was done for the inception layer, like : ``` template using attention_ = (...) ```
@davis, no you can do the merging. I think the conflicts with the master come from the fact that I created several branches from my own Dlib fork to be...
@Davis, could you please review this PR?
@Davis, I think I'm on the right track now (a lot of difficulty to find an enum shared by all the classes and accessible from the CPU and CUDA codes...
> @davis, I think I'm on the right track now (a lot of difficulty to find an enum shared by all the classes and accessible from the CPU and CUDA...