ruder.io
  • About
  • Papers
  • Talks
  • FAQ
  • Newsletter
  • NLP Progress
  • Media
  • Contact
Sign in Subscribe

optimization

Posts about optimization with a focus on gradient-based algorithms.
Optimization for Deep Learning Highlights in 2017
optimization

Optimization for Deep Learning Highlights in 2017

Different gradient descent optimization algorithms have been proposed in recent years but Adam is still most commonly used. This post discusses the most exciting highlights and most promising recent approaches that may shape the way we will optimize our models in the future.
03 Dec 2017 15 min read
An overview of gradient descent optimization algorithms
optimization

An overview of gradient descent optimization algorithms

Gradient descent is the preferred way to optimize neural networks and many other machine learning algorithms but is often used as a black box. This post explores how many of the most popular gradient-based optimization algorithms such as Momentum, Adagrad, and Adam actually work.
19 Jan 2016 28 min read
Page 1 of 1
ruder.io © 2025
Powered by Ghost