TY - GEN
T1 - FETR
T2 - 2024 IEEE International Symposium on Medical Measurements and Applications, MeMeA 2024
AU - Khan, Ufaq
AU - Nawaz, Umair
AU - Khan, Mustaqeem
AU - El Saddik, Abdulmotaleb
AU - Gueaieb, Wail
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Weakly supervised object detection (WSOD) is a cutting-edge research field within computer vision that aims to detect objects in images with minimal or incomplete annotations. In this regard, we propose a novel WSOD architecture optimized for fetal ultrasound imaging. The model is designed to leverage the inherent capabilities of the convolutional neural network and the transformers to localize and classify fetal anatomical structures within ultrasound images without requiring extensive annotated datasets. Employing a class-agnostic Fetal Transformer (FETR) for generating high-quality object proposals, our approach integrates a Multiple Instance Learning (MIL) framework to enhance detection sensitivity. We conduct thorough experiments on the FPUS23 dataset, incorporating strategic data augmentation techniques to ensure model robustness while maintaining the diagnostic integrity of the images. The efficacy of our method is demonstrated through extensive evaluations, where it achieves superior performance against state-of-the-art models, not only on the FPUS23 dataset but also on the Fetal Plane DB dataset, showcasing its adaptability to various imaging conditions. The results from our ablation studies further validate the significance of each architectural component, with qualitative results emphasizing the model's precision and reliability. Our work sets the stage for advanced prenatal diagnostics, promising to elevate the standards of fetal health monitoring and care.
AB - Weakly supervised object detection (WSOD) is a cutting-edge research field within computer vision that aims to detect objects in images with minimal or incomplete annotations. In this regard, we propose a novel WSOD architecture optimized for fetal ultrasound imaging. The model is designed to leverage the inherent capabilities of the convolutional neural network and the transformers to localize and classify fetal anatomical structures within ultrasound images without requiring extensive annotated datasets. Employing a class-agnostic Fetal Transformer (FETR) for generating high-quality object proposals, our approach integrates a Multiple Instance Learning (MIL) framework to enhance detection sensitivity. We conduct thorough experiments on the FPUS23 dataset, incorporating strategic data augmentation techniques to ensure model robustness while maintaining the diagnostic integrity of the images. The efficacy of our method is demonstrated through extensive evaluations, where it achieves superior performance against state-of-the-art models, not only on the FPUS23 dataset but also on the Fetal Plane DB dataset, showcasing its adaptability to various imaging conditions. The results from our ablation studies further validate the significance of each architectural component, with qualitative results emphasizing the model's precision and reliability. Our work sets the stage for advanced prenatal diagnostics, promising to elevate the standards of fetal health monitoring and care.
KW - Instance Learning
KW - Object Detection
KW - Transformers
KW - Ultrasound Images
KW - Weakly Supervision
UR - http://www.scopus.com/inward/record.url?scp=85201165504&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85201165504&partnerID=8YFLogxK
U2 - 10.1109/MeMeA60663.2024.10596798
DO - 10.1109/MeMeA60663.2024.10596798
M3 - Conference contribution
AN - SCOPUS:85201165504
T3 - 2024 IEEE International Symposium on Medical Measurements and Applications, MeMeA 2024 - Proceedings
BT - 2024 IEEE International Symposium on Medical Measurements and Applications, MeMeA 2024 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 26 June 2024 through 28 June 2024
ER -