JOURNAL ARTICLE

Kernel Self-Attention for Weakly-supervised Image Classification using Deep Multiple Instance Learning

Abstract

Not all supervised learning problems are described by a pair of a fixed-size input tensor and a label. In some cases, especially in medical image analysis, a label corresponds to a bag of instances (e.g. image patches), and to classify such bag, aggregation of information from all of the instances is needed. There have been several attempts to create a model working with a bag of instances, however, they are assuming that there are no dependencies within the bag and the label is connected to at least one instance. In this work, we introduce Self-Attention Attention-based MIL Pooling (SA-AbMILP) aggregation operation to account for the dependencies between instances. We conduct several experiments on MNIST, histological, microbiological, and retinal databases to show that SA-AbMILP performs better than other models. Additionally, we investigate kernel variations of Self-Attention and their influence on the results.

Keywords:
MNIST database Pooling Artificial intelligence Computer science Kernel (algebra) Pattern recognition (psychology) Contextual image classification Machine learning Image (mathematics) Tensor (intrinsic definition) Deep learning Mathematics

Metrics

42
Cited By
4.94
FWCI (Field Weighted Citation Impact)
55
Refs
0.96
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

AI in cancer detection
Physical Sciences →  Computer Science →  Artificial Intelligence
Digital Imaging for Blood Diseases
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Advanced Neural Network Applications
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
© 2026 ScienceGate Book Chapters — All rights reserved.