Keywords: neural networks, attacks on neural networks, adversarial attacks, resNet18, transformation matrix
The problem of compromising the image recognition system by purposefully falsifying the training set
UDC 004.83
DOI: 10.26102/2310-6018/2024.45.2.005
This work is devoted to the problem of the security of image recognition systems based on the use of neural networks. Such systems are used in various fields and it is extremely important to ensure their safety from attacks aimed at artificial intelligence methods. The convolutional neural network ResNet18, the ImageNet verification set for recognizing objects in an image and classifying it to a class, and adversarial attacks aimed at changing the image processed by this neural network are considered. Convolutional neural networks detect and segment the objects that are in the images. The attack was carried out at the detection stage in order not to recognize the presence of objects in the image, as well as at the segmentation stage, the modified image attributed the recognized object to another class. A series of experiments was implemented that showed how an adversarial attack changes different images. To do this, images with animals were taken and an adversarial attack was carried out on them, the analysis of their results allowed us to determine the number of iterations necessary to make a successful attack. The original images were also compared with their versions modified during the attack.
1. Murzina D.O., Dolzhenkova I.V. Application of artificial intelligence systems. Forum molodykh uchenykh. 2017;(12):1313–1316. (In Russ.).
2. Alikperova N.V. Artificial Intelligence in Healthcare: Risks and Opportunities. Zdorov'e megapolisa = City Healtyhcare. 2023;4(3):41–49. (In Russ.). https://doi.org/10.47619/2713-2617.zm.2023.v.4i3;41-49.
3. Prokopenya A.S., Azarov I.S. Overview of convolutional neural networks for image recognition. BIG DATA and Advanced Analytics. 2020;(6-1):271–280. (In Russ.).
4. Nazarov A.V., Marenkov A.N., Kaliev A.B. Detection of cryptographic viruses behavior signs in the work of the computer system. Prikaspiiskii zhurnal: upravlenie i vysokie tekhnologii = Caspian Journal: Management and High Technologies. 2018;(1):196–204. (In Russ.).
5. Marenkov A.N., Kuznetsova V.Yu., Gelagaev T.M. Application of face recognition technologies in control and access control systems. Prikaspiiskii zhurnal: upravlenie i vysokie tekhnologii = Caspian Journal: Management and High Technologies. 2021;(1):83–90. (In Russ.). https://doi.org/10.21672/2074-1707.2021.53.1.083-090.
6. Alekseenko Yu.V. Razrabotka sistemy raspoznavaniya izobrazhenii na osnove svertochnykh neironnykh setei. Evraziiskii Soyuz Uchenykh. 2017;(7-1):8–11. (In Russ.).
7. Demina R.Yu., Azhmukhamedov I.M. Increasing quality of classifying objects using new metrics of clustering. Vestnik Astrakhanskogo gosudarstvennogo tekhnicheskogo universiteta. Seriya: Upravlenie, vychislitel'naya tekhnika i informatika = Vestnik of Astrakhan State Technical University. Series: Management, Computer Science and Informatics. 2019;(4):106–114. (In Russ.). https://doi.org/10.24143/2072-9502-2019-4-106-114.
8. Chekhonina E.A., Kostyumov V.V. Overview of adversarial attacks and defenses for object detectors. International Journal of Open Information Technologies. 2023;11(7):11–20. (In Russ.).
9. Sai Abhishek A.V., Gurrala V.R., Sahoo L. Resnet18 Model With Sequential Layer For Computing Accuracy On Image Classification Dataset. International Journal of Creative Research Thoughts. 2022;10(5):176–181.
10. Sikorskii O.S. Obzor svertochnykh neironnykh setei dlya zadachi klassifikatsii izobrazhenii. In: Dvadtsatyi nauchno-prakticheskii seminar «Novye informatsionnye tekhnologii v avtomatizirovannykh sistemakh»: Materialy dvadtsatogo nauchno-prakticheskogo seminara, 20 April 2017, Moscow, Russia. Moscow: HSE Tikhonov Moscow Institute of Electronics and Mathematics; 2017. P. 37–42. (In Russ.).
Keywords: neural networks, attacks on neural networks, adversarial attacks, resNet18, transformation matrix
For citation: Khmeleva A.A., Demina R.Y., Azhmukhamedov I.M. The problem of compromising the image recognition system by purposefully falsifying the training set. Modeling, Optimization and Information Technology. 2024;12(2). URL: https://moitvivt.ru/ru/journal/pdf?id=1535 DOI: 10.26102/2310-6018/2024.45.2.005 (In Russ).
Received 04.04.2024
Revised 15.04.2024
Accepted 19.04.2024
Published 30.06.2024