MedFuseNet: Attention-based Multimodal deep learning model for Visual Question Answering in the Medical Domain

dc.contributor.authorSharma, Dhruv
dc.contributor.authorPurushotham, Sanjay
dc.contributor.authorReddy, Chandan K.
dc.date.accessioned2021-09-15T18:02:28Z
dc.date.available2021-09-15T18:02:28Z
dc.date.issued2021
dc.description.abstractMedical images are difficult to comprehend for a person without expertise. The limited number of practitioners across the globe often face the issue of physical and mental fatigue due to the high number of cases, inducing human-errors during the diagnosis. In such scenarios, having an additional opinion can be helpful in boosting the confidence of the decision-maker. Thus, it becomes crucial to have a reliable Visual Question Answering (VQA) system to provide a ‘second opinion’ on medical cases. However, most of the VQA systems that work today cater to real-world problems and are not specifically tailored for handling medical images. Moreover, the VQA system for medical images needs to consider a limited amount of training data available in this domain. In this paper, we develop MedFuseNet, an attention-based multimodal deep learning model, for VQA on medical images taking the associated challenges into account. Our MedFuseNet aims at maximizing the learning with minimal complexity by breaking the problem statement into simpler tasks and predicting the answer. We tackle two types of answer prediction - categorization and generation. We conducted an extensive set of quantitative and qualitative analyses to evaluate the performance of MedFuseNet. Our experiments demonstrates that MedFuseNet outperforms the state-of-the-art VQA methods, and that visualization of the captured attentions showcases the intepretability of our model’s predicted results.en_US
dc.description.urihttps://people.cs.vt.edu/~reddy/papers/NSR21.pdfen_US
dc.format.extent20 pagesen_US
dc.genrejournal articlesen_US
dc.genrepreprintsen_US
dc.identifierdoi:10.13016/m2blqx-5unw
dc.identifier.citationSharma, Dhruv; Purushotham, Sanjay; Reddy, Chandan K.; MedFuseNet: Attention-based Multimodal deep learning model for Visual Question Answering in the Medical Domain; Virginia Polytechnic Institute and State University, 2021; https://people.cs.vt.edu/~reddy/papers/NSR21.pdfen_US
dc.identifier.urihttp://hdl.handle.net/11603/22992
dc.language.isoen_USen_US
dc.publisherVirginia Polytechnic Institute and State Universityen_US
dc.relation.isAvailableAtThe University of Maryland, Baltimore County (UMBC)
dc.relation.ispartofUMBC Information Systems Department Collection
dc.relation.ispartofUMBC Faculty Collection
dc.rightsThis item is likely protected under Title 17 of the U.S. Copyright Law. Unless on a Creative Commons license, for uses protected by Copyright Law, contact the copyright holder or the author.en_US
dc.titleMedFuseNet: Attention-based Multimodal deep learning model for Visual Question Answering in the Medical Domainen_US
dc.typeTexten_US

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
NSR21.pdf
Size:
6.66 MB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
2.56 KB
Format:
Item-specific license agreed upon to submission
Description: