Deep Learning
Basics
TBW
Data Preparation
Image
Information - How to Resize Input Images (e.g., AlexNet, ResNet)
Information - Automatic Image Augmentation
Information - Tensor Layouts in Memory: NCHW vs NHWC
https://docs.nvidia.com/deeplearning/performance/dl-performance-convolutional/index.html
https://gist.github.com/mingfeima/595f63e5dd2ac6f87fdb47df4ffe4772
https://www.intel.com/content/www/us/en/docs/onednn/developer-guide-reference/2023-1/understanding-memory-formats.html
- N: Batch
- C: Channels
- H: Height
- W: Width
Network
Design
- Input
- Hidden
- Output
Architecture
- MLP: Multi-Layer Perceptron
- CNN: Convolutional Neural Network
https://cs231n.github.io/convolutional-networks/
- RNN: Recurrent Neural Network
- Transformer
Layer
- Linear
- Convolution
- Dropout
- Pooling
- Max Pooling
- Adaptive Max Pooling
- Normalization
- Batch Normalization
https://cvml-expertguide.net/terms/dl/layers/batch-normalization-layer/
- Recurrent
- RNN
- LSTM
- Transformer
- Encoder
- Decoder
Node
- Number
Activation
- ReLU [Common Choice]
- Tanh
- Sigmoid
- SoftMax: Typically used in the output layer for (multi-class) classification problems
- GELU: Suitable for transformer networks [Smart Choice]
- Leaky ReLU
Loss Function
https://qiita.com/Hatomugi/items/d00c1a7df07e0e3925a8
- Regression
- MSE: More sensitive to outliers in the data [Common Choice]
- MAE: Less sensitive to outliers in the data
- Huber Loss: Switching between MSE and MAE with a threshold [Smart Choice]
- Classification
- Cross-Entropy: Prime candidate for classification
Optimizer
https://qiita.com/omiita/items/1735c1d048fe5f611f80
- SGD: Stochastic Gradient Descent
- Momentum: SGD + Moving Average
- RMSProp: Root Mean Square Propagation in 2012
- Adam: Momentum + RMSProp in 2014 [Common Choice]
- RAdam: Rectified Adam in 2020 [Smart Choice]
Hyperparameter
- Learning Rate: Initial / Final / Fixed or Time-Based Decay
- Batch Size: Greater leads faster training and avoiding trapping in local minima, but lower accuracy
Reinforcement Learning
- Gamma
- Epsilon
- Tau
- Entropy
Compiler
References
https://nonbiri-tereka.hatenablog.com/entry/2016/03/10/073633
https://qiita.com/omiita/items/d24568a835da6911b01e
https://acro-engineer.hatenablog.com/entry/2019/12/25/130000
https://medium.com/aureliantactics/ppo-hyperparameters-and-ranges-6fc2d29bccbe
https://stats.stackexchange.com/questions/153531/what-is-batch-size-in-neural-network
Acknowledgments
Daiphys is a professional-service company for research and development of leading-edge technologies in science and engineering.
Get started accelerating your business through our deep expertise in R&D with AI, quantum computing, and space development; please get in touch with Daiphys today!
Daiphys Technologies LLC - https://www.daiphys.com/