Safeguarding AI medical imaging

Monday, September 04, 2023

Part of the promise of AI is that it will tap into vast repositories of data and deliver significant, untapped value. In the healthcare realm, AI promises to make patients, treatment teams, hospitals, and insurance companies healthier, better informed, and more efficient. But there’s a catch.

All of the data needed to power health and efficiency gains must be protected. Legislators the world over know this, which is why we have laws such as HIPAA and GDPR. And it is in adhering to strict privacy and security requirements that many AI for health projects fall flat as AI systems can be opaque, and their robustness in the face of hacking or attack can be questionable.

An MBZUAI team – comprising Associate Professors of Computer Vision Karthik Nandakumar and Mohammad Yaqub and research assistant Faris Almalik – aims to address security concerns around computer vision systems and the way they read, interpret, and report on medical images.

Their research: Self-Ensembling Vision Transformers for Robust Medical Image Classification (published at MICCAI), is helping to demonstrate the security and robustness of AI imaging systems. Their intention is to help safeguard the anonymity of the patients, as well as the validity of results, so that the revolution in AI for health can move forward.

A case of stop or go

Imagine, for a moment, that you are designing a computer vision system for a self-driving car. In research published recently, a team demonstrated that they could fool vision systems into “seeing” a stop sign, and through the introduction of visual noise, cause the system to instead interpret the sign as a yield, or as a sign that communicates the speed limit.

This work demonstrates how highly sophisticated systems can be disrupted cheaply and easily, with potentially deadly consequences. And it begs the question: who would want to disrupt self-driving cars from stopping at stop signs? The risk, while important, is probably quite low. When we move into the healthcare realm, however, the stakes get a lot higher — patients could suffer, doctors could act on bad information, and insurance fraud could skyrocket.

This is why the team from MBZUAI set out to test and resolve challenges with attacks in medical imaging systems. A hack that causes a medical imaging system to misinterpret a chest X-ray could have deadly and costly consequences. And holding such systems for ransom might likewise be profitable for capable and immoral hackers.

Protecting against such a scenario is one of the many reasons that mountains of healthcare regulations have been signed into law around the world. And also, as mentioned earlier, why developing and implementing such systems, while vital, is also so challenging.

Understanding computer vision

Funduscopy is the only way to directly inspect arteries, veins, and the central nervous system in an intact, living patient. Reading the visual outcomes of funduscopy accurately can help reveal a range of important things about a patient’s status related to endocarditis, anemia, diabetes, leukemia, hemorrhaging and more. The importance of funduscopy then, makes using AI to support treatment teams and patients a highly valuable undertaking and one that the research team were interested in investigating.

In the paper, the team propose a novel self-ensembling method to enhance the robustness of Vision Transformers (ViT) and Convolutional Neural Networks (CNN) — competitor computer vision technologies — for various computer vision tasks in medical imaging such as classification and segmentation. CNNs and ViTs have both been shown to have vulnerabilities to adversarial attack, which raises serious concerns about safety in clinical settings.

In response, the team propose Self-Ensembling Vision Transformers (SEViT) that make use of the fact that feature representations learned by initial blocks of a ViT are relatively unaffected by adversarial perturbations. Learning multiple classifiers based on these intermediate feature representations and combining these predictions with that of the final ViT classifier can provide robustness against adversarial attacks.

Measuring the consistency between the various predictions can also help detect adversarial samples. Experiments on two modalities (chest x-ray and funduscopy) demonstrate the efficacy of SEViT architecture to defend against various adversarial attacks in the gray-box (attacker has full knowledge of the target model, but not the defense mechanism) setting. Code: https://github.com/faresmalik/SEViT

Related

thumbnail
Wednesday, January 15, 2025

Cultural inclusivity in AI: A new benchmark dataset on 100 languages

Developed by MBZUAI scientists, the new dataset will enable greater cultural and linguistic inclusivity in multimodal LLMs.

  1. inclusivity ,
  2. linguistics ,
  3. benchmark ,
  4. languages ,
  5. multimodal ,
  6. inclusion ,
  7. llms ,
  8. dataset ,
  9. large language models ,
  10. computer vision ,
Read More
thumbnail
Monday, January 13, 2025

MBZUAI students win award for study presented at Asian Conference on Computer Vision

The students won the best student paper runners up award at ACCV for their new method called.....

  1. students ,
  2. ACCV ,
  3. award ,
  4. computer vision ,
  5. research ,
  6. student achievements ,
Read More
thumbnail
Wednesday, December 18, 2024

AI and the Arabic language: Preserving cultural heritage and enabling future discovery

The Arabic language is underrepresented in the digital world, making AI inaccessible for many of its 400.....

  1. large language models ,
  2. llms ,
  3. jais ,
  4. Arabic language ,
  5. atlas ,
  6. language ,
  7. Arabic LLM ,
  8. United Nations ,
Read More