- 1st May 2023
- 28th April 2023
- 20th April 2023
- 10th April 2023
- 4th April 2023
- 23th March 2023
- 20th March 2023
- 2nd March 2023
- 26th February 2023
- 23rd February 2023
- 13th February 2023
- Template
- Each other, trying to stay sane and not descend completely into panic.
- We spoke with Dovile as we weren't sure whether we were running her code correctly - the number of epoch seemed smaller than we expected ~she had similar numbers so that was reassuring.
- We have some carbon results. Not as many as we had hoped to have achieved, but we have decided that we will continue with what we have as time is running short.
- Have results from meta-analysis as to how many competitions, how many submissions etc.
- Starting writing draft paper
- Code not playing.
- Doubts over whether we have enough and whether it is reliable and generalisable enough.
- Writing paper up, finding the holes.
- Are we where we need to be, or have we missed something glaring - what we can do about it given the deadline is way way too soon.
We think you need at least 3 kaggle competitions to send a washed-up actor into space.
- Sanna helped Dee remember that we don't have to re-run all the code, but we can comment out the files that ran (there are 7) for training and pretraining, and the job will finish in less time without requiring more than 3 days.
- Joachim helped me figure out how to specify a desktop when submitting jobs.
- Made dataframes of all but 1 dataset (pcam-small)
- Ran training with CarbonTracker on all but 1 dataset with freeze=False for RevisitingTransfer (waiting on a few results to finish still)
- Rain training with freeze=True for RevisitingTransfer
- Have half of the results for pipeline1, both pretraining and training.
- Package version issues on the HPC
- Finish training on all datasets with freeze=True
- Enterpreting the results
- Writing
- Is it expected that training only runs for 5-7/200 epochs? Seems to be correct. When freeze=false, the first chunk of epochs is training on the classification layer, and the second chunk is training the weights.
- Verify that how we ran the training is correct
- Kimia dataset - she used it very little in the beginning. The pcam dataset is similar, but the kimia dataset is smaller. So she didn't run many experiments on it before dropping it completely.
- Pcam data - we are skipping for now
- mamms data - we are skipping for now
- Sanna helped Dee try to debug an issue with making the dataframes on the mammograms dataset in Dovile's project.
- Lottie helped us debug issue in the CarbonTracker package.
- Each other. We discussed how to handle data that is now coming in.
- For Dovile's project
- We now have results for training on 3 out of 8 of the datasets. (isic, kimia, and breast)
- Found that mammograms data has different structure on data_shares. Attempting to make dataframes work.
- We have results from pretraining in the Kaggle team's pipeline1.
- Fixed error in CarbonTracker
- Downloaded meta data from Kaggle and begun trying to work out what is useful (and what the various datasets actually include; there are no descriptions)
- We still have error's making some of the dataframes for Dovile's project. We cannot train until the dataframes are made.
- For kaggle project
- Run training job for pipeline1 to see carbon footprint
- Start looking into pipeline1
- Dovile's project:
- Finish debugging for mammograms dataset and make dataframes
- Make dataframes work for knee, chest, pcam-small, and thyroid
- Requesting red queue for running the training for the kaggle team.
--
--
- Continued working on the 2nd place team's code for the kaggle challenge
- Solved package and installation issues.
- Used the kaggle API to download a large dataset to the HPC
- Finished Step 2, data preparation for pipeline1
- Started Step 3, Training.
- Package/installation issues on the kaggle project
- Finish stage 3 Training for pipeline1 in the kaggle repo, shown here.
- Run training for Dovile's project
- Help solving the Cuda/pytorch issues on HPC
--
- Spent 14+ hours trying to problem solve environment issues on the HPC. Lottie helped solving problems. Solution was using pip and not anaconda, with module load for sklearn and tensorflow.
- Got a working environment for Dovile's project.
- Re-ran the code for Dovile's project to make the dataframes.
- Restarted work on the 2nd place team for the SIIM-FISABIO-RSNA COVID-19 Detection kaggle challenge. Their github is located here.
- Created a branch of the repo. Cloned branch locally and on the HPC in the shared "ccdd" folder.
- On the HPC, in pipeline1, completed part 1.Installation.
- In pipeline1, for part 2.Data Preparation, downloaded the first dataset (resized SIIM2021) to the HPC
- Getting working environments on the HPC. A workshop on setting up environments on HPC would have been amazing. Why does ITU not do this?
- For the second place team, pipeline1, finishing step 2 Data Preparation, and start step 3 Training.
Each other.
- Dovile helped us find the mammograms data, which had been named "cbis-ddsm" on the HPC, not "mammograms" as it was in her code.
- Lottie explained how to make a shared environment, and how to make a batch file to create an environment and install packages
- Added carbon tracker to the yaml file, and created the environment.
- Changed home path to data folders in the
make_dataframe.py
file - Changed name of the mammograms dataset folder to "cbis-ddsm" in the
data_paths.py
file - Make a job file to create the dataframes. Attempted to run job. Getting error related to
import cv2
- Comment out
import cv2
, as it doesn't seem needed to make the data splits. Getting error with tensorflow. runningconda install -c conda-forge tensorflow
in the environment. It has taken >1hour
- Conda installing packages on the HPC is taking >1hour each. It is excrutiating. Lottie says HPC is just slow today.
- We do not have access to conda envs that the other creates. This is inconvenient. We are running a batch file to create a shared env
- Can't get the environment to work with sklearn.
- We need to run the revi.job again and see if it gives the tensorflow related errors, or if it creates the dataframes
- Ideally we would make the data splits and train the model with carbontracker
- Work on getting 1 kaggle challenge participant's code to run
- Understanding HPC. For example, loading multiple modules. Can we load an environment and load tensorflow as shown on the HPC.itu.dk site?
--
- Dovile - talked us through her code from her Revisiting Hidden Representations in Transfer Learning for Medical Imaging project
--
- The real world. We already had delays due to my getting sick, then Dee had a family tragedy.
- Matching up datasets in the hpc. We haven't used the hpc much for a while and navigating can be interesting.
- Actually getting some code to run properly.
- if so, getting the carbontracker working.
--
--
--
--
- Uploaded COVID competition data to HPC
- Began working through the instructions for running team 1's code (their github). Got through stage 2.1.
- I tried running some teams' submission notebooks. While teams submit just one notebook, if often requires a larger repository to run. It doesn't seem sufficient to run their submission notebook alone.
- Download the rest of the external data for team one.
- Finish working on getting their code running (with a small sample of data)
--
Each other
Each other
- We learned that there are competitions on kaggle tagged as "code competitions". In these, the participants submit code, which kaggle will then run on their machines. We further noticed that code competitions tend to have a higher submission rate per team, with some teams up to 300 submissions, compared to competitions where a team submits an already trained model. It seems that this method of submission allows for a better tracking of the amount of times code is run, since the model-submission format does not indicate the amount of times a team trained a model for parameter tuning. Therefore, we believe we will get a more accurate result of the carbon footprint by looking into a code compeitions.
- Furthermore, code competitions are run by teams submitting code. Therefore the code is often publically available.
- The first competition we will be looking into is SIIM-FISABIO-RSNA COVID-19 Detection
- We are moving on from the Data Science Bowl 2017 due to issues finding runable publically available code (much of the code we found used legacy languages and packages that are no longer supported), and issues finding a trustworthy source of data.
- Working on a notebook to submit to the competition and run carbon-tracker on.
- Finding value in what we have done (or not done) so far
--
Sanna helped Dee find the kaggle datasets.
See above.
- Found the kaggle datasets (not so straightforward as expected)
- Found supplementary data
- We found competitions that are complete, have publically available code, and have publically available data:
- RSNA 2022 Cervical Spine Fracture Detection
- RSNA Breast Cancer Detection
- Might not have publically available code in time
- Is it too similar to the Cervical Spine Competition (<9hr run time restriction on code?)
- We learned that the category "code competition" on kaggle means teams submit code vs a trained model. Therefore competitions with this tag are more likely to have publically available code. We are thinking to narrow our investigation to competitions with the tag code competition. Furthermore, these types of competitions have restrictions regarding resources and environments, as well as maximum runtime. This might reduce the amount of confounding factors as well as make it easier to extrapolate on our findings.
- Finding the data for the competition.
-
The following items are needed for each project to work:
-
We are moving on from the lung cancer competition for now.
-
Download data for cervical spine fracture detection
-
Attempt running code for the cervical spine fracture detection competition
- Guidance on problem solving finding the data/missing elements
- Is it safe to download the data to the HPC from the sources we found?
- Next week... can we meet Thursday since the PURRlab event is also Thursday, and we will be at ITU?
Each other.
Google.
-
Found 3 github repos for high ranking teams for the Kaggle lung cancer challenge:
-
Started working on how to actually run the first one.
- Working out how to handle requirements for each repo, and looking at how to make the code run.
- Continue with how to run the first project
- Test the requirements.txt file on Sanna's computer
- Finish setting up a requirements.txt file for HPC
- Start looking at the next submission where code is available
- It seems some packages used are outdated. Is it ok if we use different versions of packages for the environment?
Replace this text with a one/two sentence description of who you helped this week and how.
Replace this text with a one/two sentence description of who helped you this week and how.
- Replace this text with a bullet point list of what you achieved this week.
- It's ok if your list is only one bullet point long!
- Replace this text with a bullet point list of where you struggled this week.
- It's ok if your list is only one bullet point long!
- Replace this text with a bullet point list of what you would like to work on next week.
- It's ok if your list is only one bullet point long!
- Try to estimate how long each task will take.
- Replace this text with a bullet point list of what you need help from Veronika on.
- It's ok if your list is only one bullet point long!
- Try to estimate how long each task will take.
This space is yours to add to as needed.