# Model Training Notes ## Validation Accuracy: `train0` *Note: "v1" = IMAGENET1K_V1, "v2" = V2* | Model | Run 0 | |---------------|----------| | ResNet50 v1 | 0.764273 | | ResNet50 v2 | 0.729282 | | ResNet101 v1 | 0.775936 | | ResNet101 v2 | 0.790055 | --- ## Validation Accuracy: `train1` *Utilizes new labeled test set from Stanford Cars for more training data!* | Model | Run 0 | |---------------|----------| | ResNet50 v1 | 0.848023 | | ResNet50 v2 | 0.833607 | | ResNet101 v1 | **0.867381** | | ResNet101 v2 | 0.861614 | --- ## Hyperparameterization: ResNet101v1 (`train1` best model) *Hyperparameters changed: optimizer and learning rate* | Description | Run 0 | |----------------|-----------| | Adam, lr=1e-4 | **0.867381** (baseline) ⭐ | | Adam, lr=3e-4 | 0.717875 | | Adam, lr=5e-5 | 0.841050 | | SGD, lr=1e-2 | 0.691104 | | SGD, lr=5e-3 | 0.417627 | --- ## Observations & Conclusions - **More data improves accuracy:** All models saw substantial gains in `train1` compared to `train0`. - **Deeper models help:** ResNet101 generally outperforms ResNet50. - **Optimizer matters:** Adam (`lr=1e-4`) yielded the highest accuracy; both lower/higher learning rates and SGD performed worse. - **IMAGENET v1 vs v2:** The difference between v1 and v2 initializations is minor compared to the effect of data volume and model size. - **Performance margins:** The right optimizer and learning rate can more than double validation accuracy for the same architecture.