giants19/Continuous-class-activation-generation-during-training-process-using-Matlab
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
This demo shows how to continuously creat a class activation mapping (CAM) during the traing process with a custom learning rate schedule with MATLAB. Automatic differentiation enables you to customize CNN as you want. This example trains a network to classify data and simulteniously compute the CAM (Class Activation Mapping) of the validation data with the weights during the training. This demo can visualize how the CNNs get to focus on the region in the image to classify which leads to the reability of the network and helps a lot in education of CNNs. Further, if the CNN is over-tuned to the dataset, the process also can be visualized. The class activation mapping was done referring to the paper below. Zhou, Bolei, Aditya Khosla, Agata Lapedriza, Aude Oliva, and Antonio Torralba. "Learning deep features for discriminative localization." In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2921-2929. 2016. This demo using the custom training loop was made with the official document below. https://jp.mathworks.com/help/deeplearning/ug/train-network-using-custom-training-loop.html
About
No description, website, or topics provided.
Resources
License
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published