3. Linear Neural Networks¶
Before we get into the details of deep neural networks, we need to cover the basics of neural network training. In this chapter, we will cover the entire training process, including defining simple neural network architectures, handling data, specifying a loss function, and training the model. In order to make things easier to grasp, we begin with the simplest concepts. Fortunately, classic statistical learning techniques such as linear and softmax regression can be cast as linear neural networks. Starting from these classic algorithms, we will introduce you to the basics, providing the basis for more complex techniques in the rest of the book.
- 3.1. Linear Regression
- 3.2. Linear Regression Implementation from Scratch
- 3.3. Concise Implementation of Linear Regression
- 3.4. Softmax Regression
- 3.4.1. Classification Problem
- 3.4.2. Network Architecture
- 3.4.3. Parameterization Cost of Fully-Connected Layers
- 3.4.4. Softmax Operation
- 3.4.5. Vectorization for Minibatches
- 3.4.6. Loss Function
- 3.4.7. Information Theory Basics
- 3.4.8. Model Prediction and Evaluation
- 3.4.9. Summary
- 3.4.10. Exercises
- 3.5. The Image Classification Dataset
- 3.6. Implementation of Softmax Regression from Scratch
- 3.7. Concise Implementation of Softmax Regression