- Patent Title: Using multimodal model consistency to detect adversarial attacks
-
Application No.: US18196712Application Date: 2023-05-12
-
Publication No.: US11977625B2Publication Date: 2024-05-07
- Inventor: Ian Michael Molloy , Youngja Park , Taesung Lee , Wenjie Wang
- Applicant: International Business Machines Corporation
- Applicant Address: US NY Armonk
- Assignee: International Business Machines Corporation
- Current Assignee: International Business Machines Corporation
- Current Assignee Address: US NY Armonk
- Agent Edward J. Wixted, III
- Main IPC: G06F21/52
- IPC: G06F21/52 ; G06F21/64 ; G06N20/00

Abstract:
A method, apparatus and computer program product to defend learning models that are vulnerable to adversarial example attack. It is assumed that data (a “dataset”) is available in multiple modalities (e.g., text and images, audio and images in video, etc.). The defense approach herein is premised on the recognition that the correlations between the different modalities for the same entity can be exploited to defend against such attacks, as it is not realistic for an adversary to attack multiple modalities. To this end, according to this technique, adversarial samples are identified and rejected if the features from one (the attacked) modality are determined to be sufficiently far away from those of another un-attacked modality for the same entity. In other words, the approach herein leverages the consistency between multiple modalities in the data to defend against adversarial attacks on one modality.
Public/Granted literature
- US20230281298A1 USING MULTIMODAL MODEL CONSISTENCY TO DETECT ADVERSARIAL ATTACKS Public/Granted day:2023-09-07
Information query