Plotting Activation Functions & Gradients in Neural Networks

Activation Functions: All neural network layers’ outputs must get passed through activation functions. These are usually non-linear and allow a neural network to learn even highly non-linear output mappings from input. This post tries to serve as a central place to get a graphical understanding of most common activation function in neural networks. I am…


TridentNet Explained

Share on Facebook Share on Twitter Share on Pinterest Share on Linkedin Share on Reddit TridentNet attempts to tackle the problem of multi-scale objects in 2D images through dilated convolutions. The changes are applied on Faster-RCNN, hence one must have at least a basic understanding of two-stage object detectors (e.g. Faster-RCNN) first to understand TridentNet….


Understanding ‘stateful’ option in Keras LSTM

I had a partially wrong understanding of the ‘stateful‘ option in LSTM while learning Keras. To confirm my knowledge, I did some searching and experimentation to make things clear. I have summarized my findings below. The source code is accessible in this Bitbucket repository. I would advise taking my conclusions with a grain of salt…