Skip to content

gcr01/deep-learning-notes

 
 

Repository files navigation

deep-learning-notes

Experiments with Deep Learning and other resources:

  • keras-capsule-pooling - an after-hours experiment in which I try to implement Capsule pooling for images.
  • max-normed-optimizer - an experimental implementation of an interesting gradient descent optimizer which normalizes gradients according to their norms. Contains various experiments which show potential power of this method.
  • selu-regularization - a Keras Regularizer Layer which allows for forcing SELU like regularization on the model weights (Dense and Conv2D versions are provided). Selu was introduced as an activation function with special initialization method, those regularizers can be add to force the weight to preserve self normalizing property during the training.
  • tf-oversampling - example with how to implement oversampling with tf.data.Dataset API.

Seminars on Deep Learning and Machine Learning

Seminars - contains a bunch of presentations I have gave at our company.

About

Experiments with Deep Learning

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 99.8%
  • Python 0.2%