Replies: 2 comments 1 reply
-
It is possible, but depends on the design of the algorithm. Some algorithms freeze the backbone and use it as a feature extractor, while others continue to fine-tune the backbone. For example, memory-bank based approaches such as PADIM and Patchcore do not train the backbone. anomalib/src/anomalib/models/patchcore/torch_model.py Lines 65 to 66 in 161d66b Some other models contiue to train the backbone. For example, knowledge distillation models such as student-teacher, train the backbone of the student model further, as shown below: anomalib/src/anomalib/models/stfpm/torch_model.py Lines 34 to 41 in 161d66b Overall, it is based on the design of the algorithm itself. Let us know if you have any other questions. Thanks! |
Beta Was this translation helpful? Give feedback.
-
I'm moving this to the Q&A section in Discussions. We could continue our discussion there. |
Beta Was this translation helpful? Give feedback.
-
I was just wondering if it's possible to train the backbone as well. In my understanding, when we set
pretrained: true
, pretrained imagenet weights are loaded in for eg. Resnet. And even if we setpretrained: false
, the weights seem to be randomly initialized and frozen, and we get a feature extractor that is not useful unless we separately load other weights offline. Is this accurate?Is it possible to train the backbone as well while training the anomaly model, i.e., backpropagate all the way through the backbone network during training?
Beta Was this translation helpful? Give feedback.
All reactions