Using .mat files to train convolutional NN - overfitting issue?

4 views (last 30 days)
Hi all,
I create .mat files out of my 3D dicom images. This is one step of my pipeline to work with a convolutional NN.
In the begining, I noticed that creating the .mat files once saves me time to train the NN with different kernel sizes.
Ex: for a 3x3 kernel size, I first create the imds with brand new .mat files.
Then, I clear all the variables and make the correspondent changes in my network to approach the imds with 5x5 kernel.
I only have to indicate the path to the imds (with the already created .mat files). Here, I noticed that the training step, even by clear all variables, it is faster than for 3x3.
Would it be the case MATLAB 'learned' the previous created .mat files, leading to an overfitting issue?
Thanks in advance.

Answers (1)

Rishik Ramena
Rishik Ramena on 13 Jan 2021
If you used a 'clear all' command, there's no way the previous created files and their contents(variables) are still in memory. The reason for a faster run in case of the 5x5 kernel could be different.

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by