기본 정보
연구 분야
프로젝트
논문
구성원
article|
인용수 54
·2022
GradDiv: Adversarial Robustness of Randomized Neural Networks via Gradient Diversity Regularization
Sungyoon Lee, Hoki Kim, Jaewook Lee
IF 23.6IEEE Transactions on Pattern Analysis and Machine Intelligence
초록

Deep learning is vulnerable to adversarial examples. Many defenses based on randomized neural networks have been proposed to solve the problem, but fail to achieve robustness against attacks using proxy gradients such as the Expectation over Transformation (EOT) attack. We investigate the effect of the adversarial attacks using proxy gradients on randomized neural networks and demonstrate that it highly relies on the directional distribution of the loss gradients of the randomized neural network. We show in particular that proxy gradients are less effective when the gradients are more scattered. To this end, we propose Gradient Diversity (GradDiv) regularizations that minimize the concentration of the gradients to build a robust randomized neural network. Our experiments on MNIST, CIFAR10, and STL10 show that our proposed GradDiv regularizations improve the adversarial robustness of randomized neural networks against a variety of state-of-the-art attack methods. Moreover, our method efficiently reduces the transferability among sample models of randomized neural networks.

키워드
Artificial neural networkMNIST databaseRobustness (evolution)Computer scienceArtificial intelligenceDeep neural networksRegularization (linguistics)Adversarial systemMachine learningPattern recognition (psychology)
타입
article
IF / 인용수
23.6 / 54
게재 연도
2022

주식회사 디써클

대표 장재우,이윤구서울특별시 강남구 역삼로 169, 명우빌딩 2층 (TIPS타운 S2)대표 전화 0507-1312-6417이메일 info@rndcircle.io사업자등록번호 458-87-03380호스팅제공자 구글 클라우드 플랫폼(GCP)

© 2026 RnDcircle. All Rights Reserved.