Bill Ross
Bill Ross
Batch size 2048->512: OOM at same Iteration 155, slightly different stats: ``` Limit: 10916963943 InUse: 10899258624 MaxInUse: 10916435968 NumAllocs: 26261 MaxAllocSize: 55796736 ```
Keeping batch 512, going back to just greyscale+kwds, input=2980, dense_size=315: ``` Model parameter count: 1127264 Additional unused parameters: -126234 ``` http://phobrain.com/pr/home/gallery/t2_w_5e-05_1e-08.png
Going back to my GInfty-removed debug version (grey+keywords), batch_size 2048, it runs 10 epochs per Iteration, and winds up after 100 Iterations at loss: 0.0018 - binary_accuracy: 1.0000 - val_loss:...
I noticed the overfit.. it looks like validation accuracy just bounces around after the first few Iterations, like in the graphs above, but I haven't graphed yet. Adding BatchNorm+Dropout(.1) after...
Dropping the BatchNorm, top accuracy is 82%, so I'd run 100 times and use the top 20 models, and without playing with the model much, I've boosted top accuracy from...
I think the loss was changing more before I added the 6 Dense's at the end, or maybe it was the Dropout; from above: ``` loss: 2.3588e-04 - binary_accuracy: 1.0000...
88% is the highest yet. I'm grinding out predictions from my top 100 models, which will take 2 days. ``` Written: vert_v/m_model_28_2048_5_69_86.pairs (1482800) in 0:06:55.677205 ``` I might be getting...
You can try keywords and histograms directly, with green and yellow '+' options in search Mode: AI, and pure histogram nets in different groupings in the Sigma-[0|theta|0theta] options, act now...
I call it a crime against nature that I can't run my 95MB models on a 1030 Gtx or two, and let them chug away on predictions full-tilt, vs. wasting...
Maybe if you build a temporary global tree, and then descend it while building the final subtree arrays. The global tree could even be done with a different package in...