吴正龙’s Bookmarks

07 JUN 2020
Visual walkthrough of how various gradient descent methods work. Assumes basic familiarity of why and how gradient descent is used in machine learning.
04 NOV 2019
Overview of different types of transfer learning techniques, and how they can be used to transfer knowledge to a different task, language or domain.
14 MAY 2019
A whirlwind tour of PyTorch's internals, covering tensors, autograd, and the overall structure of the project. Meant for aspiring OSS contributors.
24 DEC 2018
BERT was a model that broke several records for how well models could handle language-based tasks.
21 MAR 2017
Overview of transfer learning and discussion of practical applications and methods.
19 DEC 2016
Backpropagation is a leaky abstraction. It is a credit assignment scheme with non-trivial consequences. Ignoring how it works under the hood because "TensorFlow automagically makes my networks learn" means you won't be ready to wrestle with the dangers it presents.