Lecture 1:

Lecture 2:Image Classification Pipeline

Lecture 3:Loss Function and Optimization

Lecture 4:Neural Networks and Backpropagation

Summary for today:

  • (Fully-connected) Neural Networks are stacks of linear functions and nonlinear activation functions; they have much more representational power than linear classifiers
  • backpropagation = recursive application of the chain rule along a computational graph to compute the gradients of all inputs/parameters/intermediates
  • implementations maintain a graph structure, where the nodes implement the forward() / backward() API
  • forward: compute result of an operation and save any intermediates needed for gradient computation in memory
  • backward: apply the chain rule to compute the gradient of the loss function with respect to the inputs

截屏2024-02-18 21.21.22

非线性激活函数为神经网络带来非线性,否则堆多少个神经元都与单个线性层无异

截屏2024-02-18 21.21.22

截屏2024-02-18 21.23.39

截屏2024-02-18 21.24.33

截屏2024-02-18 21.25.00

截屏2024-02-18 21.25.26

截屏2024-02-18 21.41.07