1. <div id="8sgz1"><ol id="8sgz1"></ol></div>

        <em id="8sgz1"><label id="8sgz1"></label></em>
      2. <em id="8sgz1"><label id="8sgz1"></label></em>
        <em id="8sgz1"></em>
        <div id="8sgz1"><ol id="8sgz1"><mark id="8sgz1"></mark></ol></div>

        <button id="8sgz1"></button>
        west china medical publishers
        Keyword
        • Title
        • Author
        • Keyword
        • Abstract
        Advance search
        Advance search

        Search

        find Keyword "Medical visual question answering" 2 results
        • A medical visual question answering approach based on co-attention networks

          Recent studies have introduced attention models for medical visual question answering (MVQA). In medical research, not only is the modeling of “visual attention” crucial, but the modeling of “question attention” is equally significant. To facilitate bidirectional reasoning in the attention processes involving medical images and questions, a new MVQA architecture, named MCAN, has been proposed. This architecture incorporated a cross-modal co-attention network, FCAF, which identifies key words in questions and principal parts in images. Through a meta-learning channel attention module (MLCA), weights were adaptively assigned to each word and region, reflecting the model’s focus on specific words and regions during reasoning. Additionally, this study specially designed and developed a medical domain-specific word embedding model, Med-GloVe, to further enhance the model’s accuracy and practical value. Experimental results indicated that MCAN proposed in this study improved the accuracy by 7.7% on free-form questions in the Path-VQA dataset, and by 4.4% on closed-form questions in the VQA-RAD dataset, which effectively improves the accuracy of the medical vision question answer.

          Release date:2024-06-21 05:13 Export PDF Favorites Scan
        • Image-aware generative medical visual question answering based on image caption prompts

          Medical visual question answering (MVQA) plays a crucial role in the fields of computer-aided diagnosis and telemedicine. Due to the limited size and uneven annotation quality of the MVQA datasets, most existing methods rely on additional datasets for pre-training and use discriminant formulas to predict answers from a predefined set of labels. This approach makes the model prone to overfitting in low resource domains. To cope with the above problems, we propose an image-aware generative MVQA method based on image caption prompts. Firstly, we combine a dual visual feature extractor with a progressive bilinear attention interaction module to extract multi-level image features. Secondly, we propose an image caption prompt method to guide the model to better understand the image information. Finally, the image-aware generative model is used to generate answers. Experimental results show that our proposed method outperforms existing models on the MVQA task, realizing efficient visual feature extraction, as well as flexible and accurate answer outputs with small computational costs in low-resource domains. It is of great significance for achieving personalized precision medicine, reducing medical burden, and improving medical diagnosis efficiency.

          Release date:2025-06-23 04:09 Export PDF Favorites Scan
        1 pages Previous 1 Next

        Format

        Content

          1. <div id="8sgz1"><ol id="8sgz1"></ol></div>

            <em id="8sgz1"><label id="8sgz1"></label></em>
          2. <em id="8sgz1"><label id="8sgz1"></label></em>
            <em id="8sgz1"></em>
            <div id="8sgz1"><ol id="8sgz1"><mark id="8sgz1"></mark></ol></div>

            <button id="8sgz1"></button>
            欧美人与性动交α欧美精品