<?xml version="1.0" encoding="UTF-8"?>
<xml>
<records>
<record>
  <ref-type name="Journal Article">17</ref-type>
  <contributors>
    <authors>
      <author>Chatterjee, Soumick</author>
      <author>Yassin, Hadya</author>
      <author>Dubost, Florian</author>
      <author>Nürnberger, Andreas</author>
      <author>Speck, Oliver</author>
    </authors>
    <subsidiary-authors>
      <author>AG Speck</author>
    </subsidiary-authors>
  </contributors>
  <titles>
    <title>Weakly-supervised segmentation using inherently-explainable classification models and their application to brain tumour classification</title>
    <secondary-title>Neurocomputing</secondary-title>
  </titles>
  <periodical>
    <full-title>Neurocomputing</full-title>
  </periodical>
  <publisher>Elsevier</publisher>
  <pub-location>Amsterdam</pub-location>
  <isbn>0925-2312</isbn>
  <electronic-resource-num>10.1016/j.neucom.2026.133460</electronic-resource-num>
  <pages>133460</pages>
  <number/>
  <volume>682</volume>
  <abstract>Deep learning has demonstrated significant potential in medical imaging; however, the opacity of “black-box” models hinders clinical trust, while segmentation tasks typically necessitate laborious, hard-to-obtain pixel-wise annotations. To address these challenges simultaneously, this paper introduces a framework for three inherently explainable classifiers (GP-UNet, GP-ShuffleUNet, and GP-ReconResNet). By integrating a global pooling mechanism, these networks generate localisation heatmaps that directly influence classification decisions, offering inherent interpretability without relying on potentially unreliable post-hoc methods. These heatmaps are subsequently thresholded to achieve weakly-supervised segmentation, requiring only image-level classification labels for training. Validated on two datasets for multi-class brain tumour classification, the proposed models achieved a peak F1-score of 0.93. For the weakly-supervised segmentation task, a median Dice score of 0.728 (95% CI: 0.715–0.739) was recorded. Notably, on a subset of tumour-only images, the best model achieved an accuracy of 98.7%, outperforming state-of-the-art glioma grading binary classifiers. Furthermore, comparative Precision-Recall analysis validated the framework’s robustness against severe class imbalance, establishing a direct correlation between diagnostic confidence and segmentation fidelity. These results demonstrate that the proposed framework successfully combines high diagnostic accuracy with essential transparency, offering a promising direction for trustworthy clinical decision support.</abstract>
  <notes/>
  <label>PUB:(DE-HGF)16, ; 0, ; </label>
  <keywords/>
  <accession-num/>
  <work-type>Journal Article</work-type>
  <dates>
    <pub-dates>
      <year>2026</year>
    </pub-dates>
  </dates>
  <accession-num>DZNE-2026-00388</accession-num>
  <year>2026</year>
  <urls>
    <related-urls>
      <url>https://pub.dzne.de/record/286092</url>
      <url>https://doi.org/10.1016/j.neucom.2026.133460</url>
    </related-urls>
  </urls>
</record>

</records>
</xml>