TY - GEN
T1 - Using media fusion and domain dimensions to improve precision in medical image retrieval
AU - Radhouani, Saïd
AU - Kalpathy-Cramer, Jayashree
AU - Bedrick, Steven
AU - Bakke, Brian
AU - Hersh, William
PY - 2010/11/5
Y1 - 2010/11/5
N2 - In this paper, we focus on improving retrieval performance, especially early precision, in the task of solving medical multimodal queries. The queries we deal with consist of a visual component, given as a set of image-examples, and textual annotation, provided as a set of words. The queries' semantic content can be classified along three domain dimensions: anatomy, pathology, and modality. To solve these queries, we interpret their semantic content using both textual and visual data. Medical images often are accompanied by textual annotations, which in turn typically include explicit mention of their image's anatomy or pathology. Annotations rarely include explicit mention of image modality, however. To address this, we use an image's visual features to identify its modality. Our system thereby performs image retrieval by combining purely visual information about an image with information derived from its textual annotations. In order to experimentally evaluate our approach, we performed a set of experiments using the 2009 ImageCLEFmed collection using our integrated system as well as a purely textual retrieval system. Our integrated approach consistently outperformed our text-only system by 43% in MAP and by 71% in precision within the top 5 retrieved documents. We conclude that this improved performance is due to our method of combining visual and textual features.
AB - In this paper, we focus on improving retrieval performance, especially early precision, in the task of solving medical multimodal queries. The queries we deal with consist of a visual component, given as a set of image-examples, and textual annotation, provided as a set of words. The queries' semantic content can be classified along three domain dimensions: anatomy, pathology, and modality. To solve these queries, we interpret their semantic content using both textual and visual data. Medical images often are accompanied by textual annotations, which in turn typically include explicit mention of their image's anatomy or pathology. Annotations rarely include explicit mention of image modality, however. To address this, we use an image's visual features to identify its modality. Our system thereby performs image retrieval by combining purely visual information about an image with information derived from its textual annotations. In order to experimentally evaluate our approach, we performed a set of experiments using the 2009 ImageCLEFmed collection using our integrated system as well as a purely textual retrieval system. Our integrated approach consistently outperformed our text-only system by 43% in MAP and by 71% in precision within the top 5 retrieved documents. We conclude that this improved performance is due to our method of combining visual and textual features.
KW - Domain Dimensions
KW - Image Classification
KW - Image Modality Extraction
KW - Media Fusion
KW - Medical Image Retrieval
KW - Performance Evaluation
UR - http://www.scopus.com/inward/record.url?scp=78049328603&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=78049328603&partnerID=8YFLogxK
U2 - 10.1007/978-3-642-15751-6_27
DO - 10.1007/978-3-642-15751-6_27
M3 - Conference contribution
AN - SCOPUS:78049328603
SN - 3642157505
SN - 9783642157509
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 223
EP - 230
BT - Multilingual Information Access Evaluation II
T2 - 10th Workshop of the Cross-Language Evaluation Forum, CLEF 2009
Y2 - 30 September 2009 through 2 October 2009
ER -