{ "id": "2105.12508", "version": "v1", "published": "2021-05-26T12:20:47.000Z", "updated": "2021-05-26T12:20:47.000Z", "title": "Adversarial robustness against multiple $l_p$-threat models at the price of one and how to quickly fine-tune robust models to another threat model", "authors": [ "Francesco Croce", "Matthias Hein" ], "categories": [ "cs.LG", "cs.CR", "cs.CV" ], "abstract": "Adversarial training (AT) in order to achieve adversarial robustness wrt single $l_p$-threat models has been discussed extensively. However, for safety-critical systems adversarial robustness should be achieved wrt all $l_p$-threat models simultaneously. In this paper we develop a simple and efficient training scheme to achieve adversarial robustness against the union of $l_p$-threat models. Our novel $l_1+l_\\infty$-AT scheme is based on geometric considerations of the different $l_p$-balls and costs as much as normal adversarial training against a single $l_p$-threat model. Moreover, we show that using our $l_1+l_\\infty$-AT scheme one can fine-tune with just 3 epochs any $l_p$-robust model (for $p \\in \\{1,2,\\infty\\}$) and achieve multiple norm adversarial robustness. In this way we boost the previous state-of-the-art reported for multiple-norm robustness by more than $6\\%$ on CIFAR-10 and report up to our knowledge the first ImageNet models with multiple norm robustness. Moreover, we study the general transfer of adversarial robustness between different threat models and in this way boost the previous SOTA $l_1$-robustness on CIFAR-10 by almost $10\\%$.", "revisions": [ { "version": "v1", "updated": "2021-05-26T12:20:47.000Z" } ], "analyses": { "keywords": [ "threat model", "quickly fine-tune robust models", "achieve adversarial robustness wrt single", "achieve multiple norm adversarial robustness" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }