Please use this identifier to cite or link to this item: https://scholarhub.balamand.edu.lb/handle/uob/6839
Title: A framework for evaluating image obfuscation under deep learning-assisted privacy attacks
Authors: Tekli, Jimmy
Al Bouna, Bechara
Tekli, Gilbert 
Couturier, Raphaël
Affiliations: Department of Mechatronics Engineering 
Keywords: Adversary model
Background knowledge
Deep learning-assisted attacks
Face obfuscation
Image transformation
Privacy-preserving techniques
Issue Date: 2023-04-11
Publisher: Springer nature
Part of: Multimedia Tools and Applications
Volume: 82
Start page: 42173
End page: 42205
Abstract: 
Image obfuscation techniques (e.g., pixelation, blurring and masking,..) have been developed to protect sensitive information in images (e.g. individuals’ faces). In a previous work, we designed a recommendation framework that evaluates the robustness of image obfuscation techniques and recommends the most resilient obfuscation against Deep-Learning assisted attacks. In this paper, we extend the framework due to two main reasons. First, to the best of our knowledge there is not a standardized evaluation methodology nor a defined model for adversaries when evaluating the robustness of image obfuscation and more specifically face obfuscation techniques. Therefore, we adapt a three-components adversary model (goal, knowledge and capabilities) to our application domain (i.e., facial features obfuscations) and embed it in our framework. Second, considering several attacking scenarios is vital when evaluating the robustness of image obfuscation techniques. Hence, we define three threat levels and explore new aspects of an adversary and its capabilities by extending the background knowledge to include the obfuscation technique along with its hyper-parameters and the identities of the target individuals. We conduct three sets of experiments on a publicly available celebrity faces dataset. Throughout the first experiment, we implement and evaluate the recommendation framework by considering four adversaries attacking obfuscation techniques (e.g. pixelating, Gaussian/motion blur and masking) via restoration-based attacks. Throughout the second and third experiments, we demonstrate how the adversary’s attacking capabilities (recognition-based and Restoration & Recognition-based attacks) scale with its background knowledge and how it increases the potential risk of breaching the identities of blurred faces.
URI: https://scholarhub.balamand.edu.lb/handle/uob/6839
ISSN: 13807501
DOI: 10.1007/s11042-023-14664-y
Open URL: Link to full text
Type: Journal Article
Appears in Collections:Department of Mechatronics Engineering

Show full item record

SCOPUSTM   
Citations

2
checked on Nov 16, 2024

Record view(s)

92
checked on Nov 21, 2024

Google ScholarTM

Check

Altmetric

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.