Skip to main content

TAaMR

Targeted Adversarial Attack against Multimedia Recommender Systems

Research Questions

  1. Can targeted adversarial attacks against images of a low recommended category of products be exploited to modify the recommendation lists of multimedia recommender systems in terms of probability of being more recommended?
  2. What are the effects of adversarial perturbations against these attacked images for human-perceptions?

Definition of the Targeted Adversarial Attack

Let CC be a set of classes for a classifier FF. Let cCc \in C be the source class such that F(x)=cF(x) = c, and tCt \in C be a target class with tct \neq c. A Targeted Adversarial Attack finds the adversarial examples xx^∗ as following:

\begin{align*} min_{d\leq\epsilon}d(x,x^*)\\{such\ that\ F(x^*) = t} \end{align*}

Example of a product image before (a) and after (b) a PGD attack ( = 8) against VBPR on Amazon Men.