Shawn's blog Shawn's blog
About Me
  • Category
  • Tag
  • Archive
GitHub (opens new window)

Shawn Jin

I am not a creator of knowledge, I am just a porter of knowledge.
About Me
  • Category
  • Tag
  • Archive
GitHub (opens new window)
  • linear-algebra

  • statistic

  • data-mining

  • machine-learning

    • linear-regression

      • Basic Terminologies in Linear Algebra
    • linear-modelling

    • nerual-networks

    • Difference between Training set, Validation set and Test Set
    • Regularization in Machine Learning
    • Learning Feedforward Neural Network Through XOR
    • Backprop Algorithm in Machine Learning
      • Computational Graph
      • Chain Rule of Calculus
  • Data Science or Information Science
  • Data-Science
  • machine-learning
Shawn Jin
2021-09-08

Backprop Algorithm in Machine Learning

# Backprop Algorithm in ML

Back propagation(Backprop) allows information from the cost to then flow backward through the network in order to compute the gradient.

The term back-propagation is often misunderstood as meaning the whole learning algorithm for multi layer neural networks. Actually, back-propagation refers only to the method for computing the gradient, while another algorithm, such as stochastic gradient descent, is used to perform learning using this gradient.

# Computational Graph

Many ways of formalizing computation as graphs are possible. Here, we use each node in the graph to indicate a variable. The variable may be a scalar, vector, matrix, tensor, or even a variable of another type.

To formalize our graphs, we also need to introduce the idea of an opera on . An operation is a simple function of one or more variables. Our graph language is accompanied by a set of allowable operations. Functions more complicated than the operations in this set may be described by composing many operations together.

image-20210908223916027

# Chain Rule of Calculus

The chain rule of calculus (not to be confused with the chain rule of probability) is used to compute the derivatives of functions formed by composing other functions whose derivatives are known. Back-propagation is an algorithm that computes the chain rule, with a specific order of operations that is highly efficient.

Jacobian Matrix

A Jacobian matrix, sometimes simply called a Jacobian, is a matrix (opens new window) of first order partial derivatives (opens new window) (in some cases, the term "Jacobian" also refers to the determinant (opens new window) of the Jacobian matrix).

#Machine Learning#Neural Network
Updated: 2021/09/15, 20:43:56
Learning Feedforward Neural Network Through XOR
Data Science or Information Science

← Learning Feedforward Neural Network Through XOR Data Science or Information Science→

最近更新
01
Python import files from different directories
12-31
02
Classmethod in Python
09-15
03
Single/Double Star (/*) Parameters in Python
09-15
更多文章>
Theme by Vdoing | Copyright © 2019-2021 Shawn Jin | MIT License
  • 跟随系统
  • 浅色模式
  • 深色模式
  • 阅读模式