Abstract: Deep neural networks (DNNs) have achieved satisfactory performance in multiple fields. However, recent studies have shown that DNNs can be easily fooled by adversarial examples. To mitigate ...
Abstract: Recent studies have revealed that Deep Neural Networks (DNNs) are highly vulnerable to adversarial examples, which are generated by introducing imperceptible perturbations to clean images, ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results