Pitfalls of Batch Norm in TensorFlow and Sanity Checks for Training Networks
Some more practical advices on #tensorflow training with source code and reference links to look up.
https://medium.com/@theshank/pitfalls-of-batch-norm-in-tensorflow-and-sanity-checks-for-training-networks-e86c207548c8
#beginner #novice #dl #tutorial
Some more practical advices on #tensorflow training with source code and reference links to look up.
https://medium.com/@theshank/pitfalls-of-batch-norm-in-tensorflow-and-sanity-checks-for-training-networks-e86c207548c8
#beginner #novice #dl #tutorial
Medium
Pitfalls of Batch Norm in TensorFlow and Sanity Checks for Training Networks
Caveats of Batch norm: Moving mean and variance update, sharing batch norm parameters, different behaviour at train and test
Most common pitfalls, you can encounter when training neural network.
http://telegra.ph/Most-common-neural-network-mistakes-07-01
#beginner #novice #dl #tutorial
http://telegra.ph/Most-common-neural-network-mistakes-07-01
#beginner #novice #dl #tutorial
Telegraph
Most common neural network mistakes
You didn't try to overfit a single batch first You forgot to toggle train/eval mode for the net You forgot to .zero_grad() (in pytorch) before .backward() You passed softmaxed outputs to a loss that expects raw logits You didn't use `bias=False` for your…
Deep Learning for Matching in Search and Recommendation
PDF: http://www.comp.nus.edu.sg/~xiangnan/sigir18-deep.pdf
#sigir2018 #Tutorial
PDF: http://www.comp.nus.edu.sg/~xiangnan/sigir18-deep.pdf
#sigir2018 #Tutorial