Abstract: Attention–or attribution–maps methods are methods designed to highlight regions of the model’s input that were discriminative for its predictions. However, different attention maps methods ...
Abstract: Deep Neural Networks (DNNs) are vulnerable to backdoor attacks. In these attacks, an adversary inserts backdoor triggers during the training phase, causing misclassification of the model ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results