Subj : This color manipulation attack reveals significant flaws in AI im To : All From : TechnologyDaily Date : Wed Nov 27 2024 23:00:05 This color manipulation attack reveals significant flaws in AI image handling Date: Wed, 27 Nov 2024 22:48:20 +0000 Description: UTSA researchers uncover alarming AI vulnerability linked to alpha channel oversight. FULL STORY ======================================================================AI can be manipulated by differences in the alpha channel of images, experts warn This can pose risks to medical diagnoses and autonomous driving Image recognition needs to adapt for the possibility of this attack While AI has the ability to analyze images, new research has revealed a significant oversight in modern image recognition platforms. Researchers at the University of Texas at San Antonio (UTSA), have claimed the alpha channel, which controls image transparency, is frequently ignored, which could open the door to cyberattacks with potentially dangerous consequences for the medial and autonomous driving industries. The UTSA research team, led by Assistant Professor Guenevere Chen, has developed a proprietary attack method named "AlphaDog" to exploit this overlooked vulnerability in AI systems. The alpha channel, a part of RGBA (red, green, blue, alpha) image data, controls the transparency of images and plays a crucial role in rendering composite images, and can cause a disconnect between how humans and AI systems perceive the same image. Vulnerability for cars, medical imaging, and facial recognition The AlphaDog attack is designed to target both human and AI systems, though in different ways. For humans, the manipulated images may appear relatively normal. However, when processed by AI systems, these images are interpreted differently, leading to incorrect conclusions or decisions. The researchers generated 6,500 images and tested them across 100 AI models, including 80 open-source systems and 20 cloud-based AI platforms such as ChatGPT. Their tests revealed AlphaDog performs particularly well when targeting grayscale regions of images. One of the most alarming findings of the study is the vulnerability of AI systems used in autonomous vehicles. Road signs, often containing grayscale elements, can be easily manipulated using the AlphaDog technique, misinterpreting road signs, potentially leading to dangerous outcomes. The research also highlights a critical issue in medical imaging, an area increasingly reliant on AI for diagnostics. X-rays, MRIs, and CT scans, which often contain grayscale images, can be manipulated using AlphaDog. In the wrong hands, this vulnerability could lead to misdiagnoses. Another area of concern is the potential manipulation of facial recognition systems, raising the possibility of security systems being bypassed or the misidentification of individuals, opening the door to both privacy concerns and security risks. The researchers are collaborating with major tech companies, including Google, Amazon, and Microsoft, to address the vulnerability in AI platforms. "AI is created by humans, and the people who wrote the code focused on RGB but left the alpha channel out. In other words, they wrote code for AI models to read image files without the alpha channel," said Chen. "That's the vulnerability. The exclusion of the alpha channel in these platforms leads to data poisoningAI is important. It's changing our world, and we have so many concerns," Chen added. Via TechXplore More from TechRadar Pro These are the best AI writers Microsoft's AI healthcare bots might have some flaws Take a look at the best mini-PCs ====================================================================== Link to news story: https://www.techradar.com/pro/this-color-manipulation-attack-reveals-significa nt-flaws-in-AI-image-handling --- Mystic BBS v1.12 A47 (Linux/64) * Origin: tqwNet Technology News (1337:1/100) .