diff --git a/autoencoders/conf/callbacks/default.yaml b/autoencoders/conf/callbacks/default.yaml index bc19b39..6db1064 100644 --- a/autoencoders/conf/callbacks/default.yaml +++ b/autoencoders/conf/callbacks/default.yaml @@ -13,7 +13,7 @@ early_stopping: _target_: pytorch_lightning.callbacks.EarlyStopping monitor: valid-loss min_delta: 0.001 - patience: 5 + patience: 10 check_on_train_epoch_end: True model_checkpoint: diff --git a/autoencoders/conf/data/sidae.yaml b/autoencoders/conf/data/sidae.yaml index c85365a..4971d32 100644 --- a/autoencoders/conf/data/sidae.yaml +++ b/autoencoders/conf/data/sidae.yaml @@ -12,7 +12,7 @@ train: num_ops: 1 loc: 0 scale: 1 - factor: 1.0 + factor: .2 batch_size: ${data.batch_size} shuffle: True num_workers: ${data.n_workers} diff --git a/outputs/SiDAE/train/2023-09-15/16-09-21/.hydra/config.yaml b/outputs/SiDAE/train/2023-09-15/16-09-21/.hydra/config.yaml new file mode 100644 index 0000000..e7e77a2 --- /dev/null +++ b/outputs/SiDAE/train/2023-09-15/16-09-21/.hydra/config.yaml @@ -0,0 +1,99 @@ +data: + batch_size: 256 + n_workers: 10 + name: mnist + train: + _target_: torch.utils.data.DataLoader + dataset: + _target_: autoencoders.data.SiDAEDataset + dataset: + _target_: autoencoders.data.get_mnist_dataset + train: true + num_ops: 1 + loc: 0 + scale: 1 + factor: 1.0 + batch_size: ${data.batch_size} + shuffle: true + num_workers: ${data.n_workers} + valid: + _target_: torch.utils.data.DataLoader + dataset: + _target_: autoencoders.data.SiDAEDataset + dataset: + _target_: autoencoders.data.get_mnist_dataset + train: false + num_ops: 1 + loc: 0 + scale: 1 + factor: 1.0 + batch_size: ${data.batch_size} + shuffle: false + num_workers: ${data.n_workers} +model: + optimizer: + _target_: torch.optim.Adam + _partial_: true + lr: 0.001 + betas: + - 0.9 + - 0.999 + weight_decay: 0 + scheduler: + _target_: torch.optim.lr_scheduler.ReduceLROnPlateau + _partial_: true + mode: min + factor: 0.1 + patience: 10 + name: SiDAE + nn: + _target_: autoencoders.models.sidae.SiDAE + encoder: + _target_: autoencoders.modules.CNNEncoderProjection + channels_in: 1 + base_channels: 32 + latent_dim: ${model.nn.dim} + decoder: + _target_: autoencoders.modules.CNNDecoder + channels_in: 1 + base_channels: 32 + latent_dim: ${model.nn.dim} + dim: 512 + pred_dim: 512 +trainer: + _target_: pytorch_lightning.Trainer + max_epochs: 100 + accelerator: mps + devices: 1 + logger: + _target_: pytorch_lightning.loggers.WandbLogger + project: autoencoders + name: null + id: null + group: null + job_type: null + save_dir: ${hydra:runtime.output_dir} + log_model: true + tags: ${tags} +callbacks: + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar + refresh_rate: 5 + leave: true + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: train-loss + min_delta: 0.001 + patience: 5 + check_on_train_epoch_end: true + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + dirpath: ${hydra:runtime.output_dir}/checkpoints + monitor: train-loss + save_top_k: 1 + save_on_train_epoch_end: true +tags: +- ${data.name} +- ${model.name} diff --git a/outputs/SiDAE/train/2023-09-15/16-09-21/.hydra/hydra.yaml b/outputs/SiDAE/train/2023-09-15/16-09-21/.hydra/hydra.yaml new file mode 100644 index 0000000..5b0d4f9 --- /dev/null +++ b/outputs/SiDAE/train/2023-09-15/16-09-21/.hydra/hydra.yaml @@ -0,0 +1,175 @@ +hydra: + run: + dir: outputs/${model.name}/${hydra.job.name}/${now:%Y-%m-%d}/${now:%H-%M-%S} + sweep: + dir: outputs/${model.name}/${hydra.job.name}/multirun + subdir: ${hydra.job.override_dirname}/${now:%Y-%m-%d}/${now:%H-%M-%S} + launcher: + _target_: hydra_plugins.hydra_joblib_launcher.joblib_launcher.JoblibLauncher + n_jobs: -1 + backend: null + prefer: processes + require: null + verbose: 0 + timeout: null + pre_dispatch: 2*n_jobs + batch_size: auto + temp_folder: null + max_nbytes: null + mmap_mode: r + sweeper: + _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper + max_batch_size: null + params: null + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][HYDRA] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: simple + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + loggers: + logging_example: + level: DEBUG + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: simple + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + mode: RUN + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: + - hydra.mode=RUN + task: + - data=sidae + - model=sidae + - callbacks=siam + job: + name: train + chdir: null + override_dirname: callbacks=siam,data=sidae,model=sidae + id: ??? + num: ??? + config_name: config + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.3.2 + version_base: '1.3' + cwd: /Users/chrissantiago/Dropbox/GitHub/autoencoders + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /Users/chrissantiago/Dropbox/GitHub/autoencoders/autoencoders/conf + schema: file + provider: main + - path: '' + schema: structured + provider: schema + output_dir: /Users/chrissantiago/Dropbox/GitHub/autoencoders/outputs/SiDAE/train/2023-09-15/16-09-21 + choices: + experiment: null + callbacks: siam + trainer: default + model: sidae + scheduler@model.scheduler: plateau + optimizer@model.optimizer: adam + data: sidae + hydra/env: default + hydra/callbacks: null + hydra/job_logging: default + hydra/hydra_logging: default + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: basic + hydra/launcher: joblib + hydra/output: default + verbose: false diff --git a/outputs/SiDAE/train/2023-09-15/16-09-21/.hydra/overrides.yaml b/outputs/SiDAE/train/2023-09-15/16-09-21/.hydra/overrides.yaml new file mode 100644 index 0000000..28db8c2 --- /dev/null +++ b/outputs/SiDAE/train/2023-09-15/16-09-21/.hydra/overrides.yaml @@ -0,0 +1,3 @@ +- data=sidae +- model=sidae +- callbacks=siam diff --git a/outputs/SiDAE/train/2023-09-15/16-09-21/checkpoints/best_k_models.yaml b/outputs/SiDAE/train/2023-09-15/16-09-21/checkpoints/best_k_models.yaml new file mode 100644 index 0000000..83de6c3 --- /dev/null +++ b/outputs/SiDAE/train/2023-09-15/16-09-21/checkpoints/best_k_models.yaml @@ -0,0 +1,2 @@ +? /Users/chrissantiago/Dropbox/GitHub/autoencoders/outputs/SiDAE/train/2023-09-15/16-09-21/checkpoints/epoch=29-step=7050.ckpt +: -0.9640730619430542 diff --git a/outputs/SiDAE/train/2023-09-15/20-15-48/.hydra/config.yaml b/outputs/SiDAE/train/2023-09-15/20-15-48/.hydra/config.yaml new file mode 100644 index 0000000..3d9275d --- /dev/null +++ b/outputs/SiDAE/train/2023-09-15/20-15-48/.hydra/config.yaml @@ -0,0 +1,99 @@ +data: + batch_size: 256 + n_workers: 10 + name: mnist + train: + _target_: torch.utils.data.DataLoader + dataset: + _target_: autoencoders.data.SiDAEDataset + dataset: + _target_: autoencoders.data.get_mnist_dataset + train: true + num_ops: 1 + loc: 0 + scale: 1 + factor: 1.0 + batch_size: ${data.batch_size} + shuffle: true + num_workers: ${data.n_workers} + valid: + _target_: torch.utils.data.DataLoader + dataset: + _target_: autoencoders.data.SiDAEDataset + dataset: + _target_: autoencoders.data.get_mnist_dataset + train: false + num_ops: 1 + loc: 0 + scale: 1 + factor: 0.2 + batch_size: ${data.batch_size} + shuffle: false + num_workers: ${data.n_workers} +model: + optimizer: + _target_: torch.optim.Adam + _partial_: true + lr: 0.001 + betas: + - 0.9 + - 0.999 + weight_decay: 0 + scheduler: + _target_: torch.optim.lr_scheduler.ReduceLROnPlateau + _partial_: true + mode: min + factor: 0.1 + patience: 10 + name: SiDAE + nn: + _target_: autoencoders.models.sidae.SiDAE + encoder: + _target_: autoencoders.modules.CNNEncoderProjection + channels_in: 1 + base_channels: 32 + latent_dim: ${model.nn.dim} + decoder: + _target_: autoencoders.modules.CNNDecoder + channels_in: 1 + base_channels: 32 + latent_dim: ${model.nn.dim} + dim: 512 + pred_dim: 512 +trainer: + _target_: pytorch_lightning.Trainer + max_epochs: 100 + accelerator: mps + devices: 1 + logger: + _target_: pytorch_lightning.loggers.WandbLogger + project: autoencoders + name: null + id: null + group: null + job_type: null + save_dir: ${hydra:runtime.output_dir} + log_model: true + tags: ${tags} +callbacks: + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar + refresh_rate: 5 + leave: true + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: train-loss + min_delta: 0.001 + patience: 5 + check_on_train_epoch_end: true + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + dirpath: ${hydra:runtime.output_dir}/checkpoints + monitor: train-loss + save_top_k: 1 + save_on_train_epoch_end: true +tags: +- ${data.name} +- ${model.name} diff --git a/outputs/SiDAE/train/2023-09-15/20-15-48/.hydra/hydra.yaml b/outputs/SiDAE/train/2023-09-15/20-15-48/.hydra/hydra.yaml new file mode 100644 index 0000000..4e02d06 --- /dev/null +++ b/outputs/SiDAE/train/2023-09-15/20-15-48/.hydra/hydra.yaml @@ -0,0 +1,175 @@ +hydra: + run: + dir: outputs/${model.name}/${hydra.job.name}/${now:%Y-%m-%d}/${now:%H-%M-%S} + sweep: + dir: outputs/${model.name}/${hydra.job.name}/multirun + subdir: ${hydra.job.override_dirname}/${now:%Y-%m-%d}/${now:%H-%M-%S} + launcher: + _target_: hydra_plugins.hydra_joblib_launcher.joblib_launcher.JoblibLauncher + n_jobs: -1 + backend: null + prefer: processes + require: null + verbose: 0 + timeout: null + pre_dispatch: 2*n_jobs + batch_size: auto + temp_folder: null + max_nbytes: null + mmap_mode: r + sweeper: + _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper + max_batch_size: null + params: null + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][HYDRA] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: simple + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + loggers: + logging_example: + level: DEBUG + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: simple + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + mode: RUN + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: + - hydra.mode=RUN + task: + - data=sidae + - model=sidae + - callbacks=siam + job: + name: train + chdir: null + override_dirname: callbacks=siam,data=sidae,model=sidae + id: ??? + num: ??? + config_name: config + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.3.2 + version_base: '1.3' + cwd: /Users/chrissantiago/Dropbox/GitHub/autoencoders + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /Users/chrissantiago/Dropbox/GitHub/autoencoders/autoencoders/conf + schema: file + provider: main + - path: '' + schema: structured + provider: schema + output_dir: /Users/chrissantiago/Dropbox/GitHub/autoencoders/outputs/SiDAE/train/2023-09-15/20-15-48 + choices: + experiment: null + callbacks: siam + trainer: default + model: sidae + scheduler@model.scheduler: plateau + optimizer@model.optimizer: adam + data: sidae + hydra/env: default + hydra/callbacks: null + hydra/job_logging: default + hydra/hydra_logging: default + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: basic + hydra/launcher: joblib + hydra/output: default + verbose: false diff --git a/outputs/SiDAE/train/2023-09-15/20-15-48/.hydra/overrides.yaml b/outputs/SiDAE/train/2023-09-15/20-15-48/.hydra/overrides.yaml new file mode 100644 index 0000000..28db8c2 --- /dev/null +++ b/outputs/SiDAE/train/2023-09-15/20-15-48/.hydra/overrides.yaml @@ -0,0 +1,3 @@ +- data=sidae +- model=sidae +- callbacks=siam diff --git a/outputs/SiDAE/train/2023-09-15/20-15-48/checkpoints/best_k_models.yaml b/outputs/SiDAE/train/2023-09-15/20-15-48/checkpoints/best_k_models.yaml new file mode 100644 index 0000000..4badc93 --- /dev/null +++ b/outputs/SiDAE/train/2023-09-15/20-15-48/checkpoints/best_k_models.yaml @@ -0,0 +1,2 @@ +? /Users/chrissantiago/Dropbox/GitHub/autoencoders/outputs/SiDAE/train/2023-09-15/20-15-48/checkpoints/epoch=28-step=6815.ckpt +: -0.9633734226226807 diff --git a/outputs/SiDAE/train/2023-09-15/20-28-23/.hydra/config.yaml b/outputs/SiDAE/train/2023-09-15/20-28-23/.hydra/config.yaml new file mode 100644 index 0000000..582aa83 --- /dev/null +++ b/outputs/SiDAE/train/2023-09-15/20-28-23/.hydra/config.yaml @@ -0,0 +1,99 @@ +data: + batch_size: 256 + n_workers: 10 + name: mnist + train: + _target_: torch.utils.data.DataLoader + dataset: + _target_: autoencoders.data.SiDAEDataset + dataset: + _target_: autoencoders.data.get_mnist_dataset + train: true + num_ops: 1 + loc: 0 + scale: 1 + factor: 0.2 + batch_size: ${data.batch_size} + shuffle: true + num_workers: ${data.n_workers} + valid: + _target_: torch.utils.data.DataLoader + dataset: + _target_: autoencoders.data.SiDAEDataset + dataset: + _target_: autoencoders.data.get_mnist_dataset + train: false + num_ops: 1 + loc: 0 + scale: 1 + factor: 1.0 + batch_size: ${data.batch_size} + shuffle: false + num_workers: ${data.n_workers} +model: + optimizer: + _target_: torch.optim.Adam + _partial_: true + lr: 0.001 + betas: + - 0.9 + - 0.999 + weight_decay: 0 + scheduler: + _target_: torch.optim.lr_scheduler.ReduceLROnPlateau + _partial_: true + mode: min + factor: 0.1 + patience: 10 + name: SiDAE + nn: + _target_: autoencoders.models.sidae.SiDAE + encoder: + _target_: autoencoders.modules.CNNEncoderProjection + channels_in: 1 + base_channels: 32 + latent_dim: ${model.nn.dim} + decoder: + _target_: autoencoders.modules.CNNDecoder + channels_in: 1 + base_channels: 32 + latent_dim: ${model.nn.dim} + dim: 512 + pred_dim: 512 +trainer: + _target_: pytorch_lightning.Trainer + max_epochs: 100 + accelerator: mps + devices: 1 + logger: + _target_: pytorch_lightning.loggers.WandbLogger + project: autoencoders + name: null + id: null + group: null + job_type: null + save_dir: ${hydra:runtime.output_dir} + log_model: true + tags: ${tags} +callbacks: + model_summary: + _target_: pytorch_lightning.callbacks.RichModelSummary + progress_bar: + _target_: pytorch_lightning.callbacks.RichProgressBar + refresh_rate: 5 + leave: true + early_stopping: + _target_: pytorch_lightning.callbacks.EarlyStopping + monitor: train-loss + min_delta: 0.001 + patience: 10 + check_on_train_epoch_end: true + model_checkpoint: + _target_: pytorch_lightning.callbacks.ModelCheckpoint + dirpath: ${hydra:runtime.output_dir}/checkpoints + monitor: train-loss + save_top_k: 1 + save_on_train_epoch_end: true +tags: +- ${data.name} +- ${model.name} diff --git a/outputs/SiDAE/train/2023-09-15/20-28-23/.hydra/hydra.yaml b/outputs/SiDAE/train/2023-09-15/20-28-23/.hydra/hydra.yaml new file mode 100644 index 0000000..24914f9 --- /dev/null +++ b/outputs/SiDAE/train/2023-09-15/20-28-23/.hydra/hydra.yaml @@ -0,0 +1,175 @@ +hydra: + run: + dir: outputs/${model.name}/${hydra.job.name}/${now:%Y-%m-%d}/${now:%H-%M-%S} + sweep: + dir: outputs/${model.name}/${hydra.job.name}/multirun + subdir: ${hydra.job.override_dirname}/${now:%Y-%m-%d}/${now:%H-%M-%S} + launcher: + _target_: hydra_plugins.hydra_joblib_launcher.joblib_launcher.JoblibLauncher + n_jobs: -1 + backend: null + prefer: processes + require: null + verbose: 0 + timeout: null + pre_dispatch: 2*n_jobs + batch_size: auto + temp_folder: null + max_nbytes: null + mmap_mode: r + sweeper: + _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper + max_batch_size: null + params: null + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][HYDRA] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: simple + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + loggers: + logging_example: + level: DEBUG + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: simple + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + mode: RUN + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: + - hydra.mode=RUN + task: + - data=sidae + - model=sidae + - callbacks=siam + job: + name: train + chdir: null + override_dirname: callbacks=siam,data=sidae,model=sidae + id: ??? + num: ??? + config_name: config + env_set: {} + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.3.2 + version_base: '1.3' + cwd: /Users/chrissantiago/Dropbox/GitHub/autoencoders + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: /Users/chrissantiago/Dropbox/GitHub/autoencoders/autoencoders/conf + schema: file + provider: main + - path: '' + schema: structured + provider: schema + output_dir: /Users/chrissantiago/Dropbox/GitHub/autoencoders/outputs/SiDAE/train/2023-09-15/20-28-23 + choices: + experiment: null + callbacks: siam + trainer: default + model: sidae + scheduler@model.scheduler: plateau + optimizer@model.optimizer: adam + data: sidae + hydra/env: default + hydra/callbacks: null + hydra/job_logging: default + hydra/hydra_logging: default + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: basic + hydra/launcher: joblib + hydra/output: default + verbose: false diff --git a/outputs/SiDAE/train/2023-09-15/20-28-23/.hydra/overrides.yaml b/outputs/SiDAE/train/2023-09-15/20-28-23/.hydra/overrides.yaml new file mode 100644 index 0000000..28db8c2 --- /dev/null +++ b/outputs/SiDAE/train/2023-09-15/20-28-23/.hydra/overrides.yaml @@ -0,0 +1,3 @@ +- data=sidae +- model=sidae +- callbacks=siam diff --git a/outputs/SiDAE/train/2023-09-15/20-28-23/checkpoints/best_k_models.yaml b/outputs/SiDAE/train/2023-09-15/20-28-23/checkpoints/best_k_models.yaml new file mode 100644 index 0000000..36640c7 --- /dev/null +++ b/outputs/SiDAE/train/2023-09-15/20-28-23/checkpoints/best_k_models.yaml @@ -0,0 +1,2 @@ +? /Users/chrissantiago/Dropbox/GitHub/autoencoders/outputs/SiDAE/train/2023-09-15/20-28-23/checkpoints/epoch=67-step=15980.ckpt +: -0.988296627998352 diff --git a/outputs/results.json b/outputs/results.json index fe3f6c8..43df27c 100644 --- a/outputs/results.json +++ b/outputs/results.json @@ -7,7 +7,6 @@ }, "ckpt": "/Users/chrissantiago/Dropbox/GitHub/autoencoders/outputs/autoencoder/train/2023-08-29/21-13-51/checkpoints/epoch=11-step=2820.ckpt" }, - { "model": "deep-autoencoder", "metrics": { @@ -135,5 +134,29 @@ "auc": 0.9842 }, "ckpt": "/Users/chrissantiago/Dropbox/GitHub/autoencoders/outputs/deep-DAE/train/2023-09-09/14-17-53/checkpoints/epoch=5-step=1410.ckpt" + }, + { + "model": "SiDAE", + "metrics": { + "acc": 0.9533, + "auc": 0.9741 + }, + "ckpt": "/Users/chrissantiago/Dropbox/GitHub/autoencoders/outputs/SiDAE/train/2023-09-15/16-09-21/checkpoints/epoch=29-step=7050.ckpt" + }, + { + "model": "SiDAE", + "metrics": { + "acc": 0.9664, + "auc": 0.9813 + }, + "ckpt": "/Users/chrissantiago/Dropbox/GitHub/autoencoders/outputs/SiDAE/train/2023-09-15/20-15-48/checkpoints/epoch=28-step=6815.ckpt" + }, + { + "model": "SiDAE", + "metrics": { + "acc": 0.9639, + "auc": 0.98 + }, + "ckpt": "/Users/chrissantiago/Dropbox/GitHub/autoencoders/outputs/SiDAE/train/2023-09-15/20-28-23/checkpoints/epoch=67-step=15980.ckpt" } ] \ No newline at end of file