Autoencoder-based baseline system for DCASE2021 Challenge Task 2.
This system consists of two main scripts:
00_train.py
- "Development" mode:
- This script trains a model for each machine type by using the directory
dev_data/<machine_type>/train/
.
- This script trains a model for each machine type by using the directory
- "Evaluation" mode:
- This script trains a model for each machine type by using the directory
eval_data/<machine_type>/train/
. (This directory will be from the "additional training dataset".)
- This script trains a model for each machine type by using the directory
- "Development" mode:
01_test.py
- "Development" mode:
- This script makes a csv file for each section including the anomaly scores for each wav file in the directories
dev_data/<machine_type>/source_test/
anddev_data/<machine_type>/target_test/
. - The csv files are stored in the directory
result/
. - It also makes a csv file including AUC, pAUC, precision, recall, and F1-score for each section.
- This script makes a csv file for each section including the anomaly scores for each wav file in the directories
- "Evaluation" mode:
- This script makes a csv file for each section including the anomaly scores for each wav file in the directories
eval_data/<machine_type>/source_test/
andeval_data/<machine_type>/target_test/
. (These directories will be from the "evaluation dataset".) - The csv files are stored in the directory
result/
.
- This script makes a csv file for each section including the anomaly scores for each wav file in the directories
- "Development" mode:
Clone this repository from Github.
We will launch the datasets in three stages. So, please download the datasets in each stage:
- "Development dataset"
- Download
dev_data_<machine_type>.zip
from https://zenodo.org/record/4562016.
- Download
- "Additional training dataset", i.e. the evaluation dataset for training
- After April. 1, 2021, download
eval_data_train_<machine_type>.zip
from https://zenodo.org/record/4660992.
- After April. 1, 2021, download
- "Evaluation dataset", i.e. the evaluation dataset for test
- After June. 1, 2021, download
eval_data_test_<machine_type>.zip
from https://zenodo.org/record/4884786.
- After June. 1, 2021, download
Unzip the downloaded files and make the directory structure as follows:
- /dcase2021_task2_baseline_ae
- /00_train.py
- /01_test.py
- /common.py
- /keras_model.py
- /baseline.yaml
- /readme.md
- /dev_data
- /fan
- /train (Normal data in the source and target domains for all sections are included.)
- /section_00_source_train_normal_0000_.wav
- ...
- /section_00_source_train_normal_0999_.wav
- /section_00_target_train_normal_0000_.wav
- /section_00_target_train_normal_0001_.wav
- /section_00_target_train_normal_0002_.wav
- /section_01_source_train_normal_0000_.wav
- ...
- /section_02_target_train_normal_0002_.wav
- /source_test (Normal and anomaly data in the source domain for all sections are included.)
- /section_00_source_test_normal_0000.wav
- ...
- /section_00_source_test_normal_0099.wav
- /section_00_source_test_anomaly_0000.wav
- ...
- /section_00_source_test_anomaly_0099.wav
- /section_01_source_test_normal_0000.wav
- ...
- /section_02_source_test_anomaly_0099.wav
- /target_test (Normal and anomaly data in the target domain for all sections are included.)
- /section_00_target_test_normal_0000.wav
- ...
- /section_00_target_test_normal_0099.wav
- /section_00_target_test_anomaly_0000.wav
- ...
- /section_00_target_test_anomaly_0099.wav
- /section_01_target_test_normal_0000.wav
- ...
- /section_02_target_test_anomaly_0099.wav
- /train (Normal data in the source and target domains for all sections are included.)
- /gearbox (The other machine types have the same directory structure as fan.)
- /pump
- /slider
- /valve
- /ToyCar
- /ToyTrain
- /fan
- /eval_data (Add this directory after launch)
- /fan
- /train (Unzipped additional training dataset. Normal data in the source and target domains for all sections are included.)
- /section_03_source_train_normal_0000_.wav
- ...
- /section_03_source_train_normal_0999_.wav
- /section_03_target_train_normal_0000_.wav
- /section_03_target_train_normal_0001_.wav
- /section_03_target_train_normal_0002_.wav
- /section_04_source_train_normal_0000_.wav
- ...
- /section_05_target_train_normal_0002_.wav
- /source_test (Unzipped evaluation dataset. Normal and anomaly data in the source domain for all sections are included.)
- /section_03_source_test_0000.wav
- ...
- /section_03_source_test_0199.wav
- /section_04_source_test_0000.wav
- ...
- /section_05_source_test_0199.wav
- /target_test (Unzipped evaluation dataset. Normal and anomaly data in the target domain for all sections are included.)
- /section_03_target_test_0000.wav
- ...
- /section_03_target_test_0199.wav
- /section_04_target_test_0000.wav
- ...
- /section_05_target_test_0199.wav
- /train (Unzipped additional training dataset. Normal data in the source and target domains for all sections are included.)
- /gearbox (The other machine types have the same directory structure as fan.)
- /pump
- /slider
- /valve
- /ToyCar
- /ToyTrain
- /fan
You can change parameters for feature extraction and model definition by editing baseline.yaml
.
Run the training script 00_train.py
.
Use the option -d
for the development dataset dev_data/<machine_type>/train/
.
$ python3.6 00_train.py -d
Options:
Argument | Description | |
---|---|---|
-h |
--help |
Application help. |
-v |
--version |
Show application version. |
-d |
--dev |
Mode for the development dataset |
-e |
--eval |
Mode for the additional training and evaluation datasets |
00_train.py
trains a model for each machine type and store the trained models in the directory model/
.
Run the test script 01_test.py
.
Use the option -d
for the development dataset dev_data/<machine_type>/test/
.
$ python3.6 01_test.py -d
The options for 01_test.py
are the same as those for 00_train.py
.
01_test.py
calculates an anomaly score for each wav file in the directories dev_data/<machine_type>/source_test/
and dev_data/<machine_type>/target_test/
.
A csv file for each section including the anomaly scores will be stored in the directory result/
.
If the mode is "development", the script also outputs another csv file including AUC, pAUC, precision, recall, and F1-score for each section.
You can check the anomaly scores in the csv files anomaly_score_<machine_type>_section_<section_index>_<domain>_test.csv
in the directory result/
.
Each anomaly score corresponds to a wav file in the directories dev_data/<machine_type>/source_test/
and dev_data/<machine_type>/target_test/
:
anomaly_score_fan_section_00_source_test.csv
section_00_source_test_normal_0000.wav 9.020132729
section_00_source_test_normal_0001.wav 8.894007893
section_00_source_test_normal_0002.wav 9.085542421
section_00_source_test_normal_0003.wav 9.769387973
section_00_source_test_normal_0004.wav 8.656917097
section_00_source_test_normal_0005.wav 8.899467716
...
Also, anomaly detection results after thresholding can be checked in the csv files anomaly_score_<machine_type>_section_<section_index>_<domain>_test.csv
:
decision_result_fan_section_00_source_test.csv
section_00_source_test_normal_0000.wav 0
section_00_source_test_normal_0001.wav 0
section_00_source_test_normal_0002.wav 0
section_00_source_test_normal_0003.wav 1
section_00_source_test_normal_0004.wav 0
section_00_source_test_normal_0005.wav 0
...
Also, you can check performance indicators such as AUC, pAUC, precision, recall, and F1 score:
result.csv
fan
section domain AUC pAUC precision recall F1 score
0 source 0.6738 0.573157895 0.654545455 0.36 0.464516129
1 source 0.6786 0.505263158 0.625 0.3 0.405405405
2 source 0.6542 0.532631579 0.685714286 0.24 0.355555556
0 target 0.6955 0.548421053 0.657142857 0.46 0.541176471
1 target 0.4953 0.483157895 0.5 0.35 0.411764706
2 target 0.6441 0.559473684 0.497461929 0.98 0.65993266
arithmetic mean 0.64025 0.533684211 0.603310754 0.448333333 0.473058488
harmonic mean 0.631866714 0.531826049 0.592999386 0.367438982 0.453732583
...
valve
section domain AUC pAUC precision recall F1 score
0 source 0.5026 0.506842105 0.588235294 0.2 0.298507463
1 source 0.5404 0.494210526 0.512195122 0.21 0.29787234
2 source 0.596 0.516315789 0.615384615 0.24 0.345323741
0 target 0.4716 0.485263158 0.382352941 0.13 0.194029851
1 target 0.5857 0.548421053 0.519230769 0.81 0.6328125
2 target 0.5489 0.489473684 0.55 0.22 0.314285714
arithmetic mean 0.540866667 0.506754386 0.52789979 0.301666667 0.347138602
harmonic mean 0.537267925 0.505885378 0.515602416 0.218970897 0.307394569
AUC pAUC precision recall F1 score
arithmetic mean over all machine types, sections, and domains 0.634334516 0.537946639 0.511821964 0.448497815 0.439585489
harmonic mean over all machine types, sections, and domains 0.620051253 0.534340265 1.55E-15 1.55E-15 1.55E-15
After the additional training dataset is launched, download and unzip it.
Move it to eval_data/<machine_type>/train/
.
Run the training script 00_train.py
with the option -e
.
$ python3.6 00_train.py -e
Models are trained by using the additional training dataset eval_data/<machine_type>/train/
.
After the evaluation dataset for test is launched, download and unzip it.
Move it to eval_data/<machine_type>/source_test/
and eval_data/<machine_type>/target_test/
.
Run the test script 01_test.py
with the option -e
.
$ python3.6 01_test.py -e
Anomaly scores are calculated using the evaluation dataset, i.e., eval_data/<machine_type>/source_test/
and eval_data/<machine_type>/target_test/
.
The anomaly scores are stored as csv files in the directory result/
.
You can submit the csv files for the challenge.
From the submitted csv files, we will calculate AUC, pAUC, and your ranking.
We develop the source code on Ubuntu 16.04 LTS and 18.04 LTS. In addition, we checked performing on Ubuntu 16.04 LTS, 18.04 LTS, CentOS 7, and Windows 10.
- p7zip-full
- Python == 3.6.5
- FFmpeg
- Keras == 2.3.0
- Keras-Applications == 1.0.8
- Keras-Preprocessing == 1.0.5
- matplotlib == 3.0.3
- numpy == 1.18.1
- PyYAML == 5.1
- scikit-learn == 0.22.2.post1
- scipy == 1.1.0
- librosa == 0.6.0
- audioread == 2.1.5 (more)
- setuptools == 41.0.0
- tensorflow == 1.15.0
- tqdm == 4.43.0
If you use this baseline system, please cite all the following three papers:
- Yohei Kawaguchi, Keisuke Imoto, Yuma Koizumi, Noboru Harada, Daisuke Niizumi, Kota Dohi, Ryo Tanabe, Harsh Purohit, and Takashi Endo, "Description and Discussion on DCASE 2021 Challenge Task 2: Unsupervised Anomalous Sound Detection for Machine Condition Monitoring under Domain Shifted Conditions," in arXiv e-prints: 2106.04492, 2021. URL
- Noboru Harada, Daisuke Niizumi, Daiki Takeuchi, Yasunori Ohishi, Masahiro Yasuda, Shoichiro Saito, "ToyADMOS2: Another Dataset of Miniature-Machine Operating Sounds for Anomalous Sound Detection under Domain Shift Conditions," in arXiv e-prints: 2106.02369, 2021. URL
- Ryo Tanabe, Harsh Purohit, Kota Dohi, Takashi Endo, Yuki Nikaido, Toshiki Nakamura, and Yohei Kawaguchi, "MIMII DUE: Sound Dataset for Malfunctioning Industrial Machine Investigation and Inspection with Domain Shifts due to Changes in Operational and Environmental Conditions," in arXiv e-prints: 2105.02702, 2021. URL