Skip to content

Shaileshalluri/Deep-Learing-Project1

Repository files navigation

Deep-Learing-Homework-1-CPSC-8430

Shailesh Alluri

The files in this folder and their purpose are listed below. A description of the funtionality of each file can be found at the top of that file:

Part1-1.ipynb - source code for Part1.1 Simulate a Function (Task 1 of part 1).
Part1-2.ipynb - source code for Part1.2 Train on an Actual Task(Task 2 of part 1).
Part2-1.ipynb - source code for Part2.1 Visualize the Optimization Process.
Part2-2.ipynb - source code for Part2.2 Observe Gradient Norm during Training.
Part2-3.ipynb - souce code for Part2.3 Calculating minimal ratio at zero gradient.
Part3-1- Random Labels.ipynb - source code for Part3.1 Can Network fit Random Labels.
Part3-2-Number of Params.ipynb - source code for Part3.2 Number of Parameters VS.Generalization.
Part3-3-1 - Interpolation.ipynb - source code for Part3.3.1 Flatness VS. Generalization.
Part3-3-2-batch sz.ipynb - souce code for Part3.3.2 Flatness VS. Generalization.
Deep_Learning_HW1.pdf - Assignment report with analysis of code and results.

To run these files, the following libraries are required:
python 3.8
Pytorch 1.7
Tensorflow 2.4
Matplotlib 3.2.2
Numpy 1.19

The dataset required to run the source code is the MNIST dataset from the torchvision datasets library in pytorch. Once the dataset is downloaded to this pathfile, it can be used by every other file and is the only file you will need to download. In every file that requires the MNIST dataset, there is a line of code that grabs the dataset.
It is provided below:
trainingSet = datasets.MNIST('', train=True, download=False, ...)
To download the dataset, change the download option to True. This is done for you in the first source code file (Part1_TrainOnActualTask.ipynb). You will need internet connect to download the dataset.

To run the following .ipynb files, open juyper notbooks in an enviornment with the above packages installed and run the cells in the file from top to bottom.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published