Grokking-Deep-Learning
Grokking-Deep-Learning copied to clipboard
this repository accompanies the book "Grokking Deep Learning"
See if any of these links make any sense for you to expand your coursework offerings a little bit: https://github.com/huggingface/tokenizers/issues/69 https://github.com/Kayzaks/HackingNeuralNetworks
 Hey, trask. There is a something wrong in Chapter 3 In[13], it should be 5 rows & 6 columns, right?
It looks like in Chapter 15, section "Secure Aggregation", there is a `deepcopy` made of the model both in receiving function `train_and_encrypt` and by the caller: ```python def train_and_encrypt(model, input,...
It is not clear to me why we are passing all ones tensors when calling `backward()` on a tensor.
Running first code snippet from master produce this error: `'weight_deltas' is not defined` Also function ele_mul is defined but never used
Please see https://github.com/suyash/grokking-deep-learning-rs I have implemented all the exercises in [Rust](https://www.rust-lang.org/). Rust is a new programming language with a focus on safety, speed and concurrency, while having a minimal runtime...
Code formatting in Chapter 11, section "Predicting Movie Reviews" is broken. I was not able to reproduce the results in the book, p. 196, and assume that there is some...
Why there was needed to define "new" and call it once after defining new. I think the Tensor data in this case is multiplied to the "other". But why it...
In Chapter 12, why are we giving a previous layer delta to embedded layer `embed_idx = sent[layer_idx]` `embed[embed_idx] -= layers[layer_idx]['hidden_delta'] * alpha / float(len(sent))` shouldn't it be `embed[embed_idx] -= layer['hidden_delta']...
Hey @iamtrask In chapter 10 , we were said to reuse weights but I cannot possibly comprehend how did the network reused it? The network looks same as chapter 9s...