Recognition: unknown
Early Detection of Acute Myeloid Leukemia (AML) Using YOLOv12 Deep Learning Model
Pith reviewed 2026-05-10 08:29 UTC · model grok-4.3
The pith
YOLOv12 with Otsu thresholding on cell-based segmentation classifies AML cells at 99.3% validation and test accuracy.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
Our experiments demonstrate that YOLOv12 with Otsu thresholding on cell-based segmentation achieved the highest level of validation and test accuracy, both reaching 99.3%.
Load-bearing premise
The assumption that the (unspecified) dataset is representative of real-world patient variability and that the reported test accuracy reflects generalization rather than overfitting to the particular images used.
Figures
read the original abstract
Acute Myeloid Leukemia (AML) is one of the most life-threatening type of blood cancers, and its accurate classification is considered and remains a challenging task due to the visual similarity between various cell types. This study addresses the classification of the multiclasses of AML cells Utilizing YOLOv12 deep learning model. We applied two segmentation approaches based on cell and nucleus features, using Hue channel and Otsu thresholding techniques to preprocess the images prior to classification. Our experiments demonstrate that YOLOv12 with Otsu thresholding on cell-based segmentation achieved the highest level of validation and test accuracy, both reaching 99.3%.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript applies the YOLOv12 deep learning model to multiclass classification of Acute Myeloid Leukemia (AML) cells in blood-smear images. Two segmentation strategies (cell-based and nucleus-based) are tested with Hue-channel and Otsu-thresholding preprocessing; the authors report that YOLOv12 plus Otsu thresholding on cell-based segmentation reaches 99.3% accuracy on both validation and test sets.
Significance. If the accuracy claim is shown to be robust, reproducible, and free of leakage, the work could provide a practical high-accuracy pipeline for automated AML screening. The combination of a recent detection architecture with classical thresholding is a straightforward engineering contribution. However, the complete absence of dataset statistics, patient counts, split protocols, baselines, and error analysis prevents any assessment of whether the result reflects genuine generalization.
major comments (2)
- Abstract: The 99.3% validation and test accuracy is stated without any information on total images, number of patients, class distribution, dataset source, or train/validation/test partitioning strategy. Because multiple cells are routinely extracted from the same slide or patient, an image-level split risks leakage; the reported figure cannot be interpreted as evidence of reliable early detection until patient-stratified splits or external validation are demonstrated.
- Experiments/Results section: No baseline comparisons (other YOLO versions, standard CNN classifiers, or non-deep-learning methods) are provided, nor is any confusion matrix, per-class accuracy, or error analysis reported. For a multiclass problem whose difficulty stems from visual similarity between cell types, these omissions leave the superiority claim unsupported.
minor comments (1)
- Abstract: The number of AML cell classes and their identities are not stated, making it impossible to judge the difficulty of the reported multiclass task.
Simulated Author's Rebuttal
We thank the referee for the detailed and constructive review. The comments correctly identify omissions in dataset reporting and experimental comparisons that limit interpretability of the results. We will revise the manuscript to address both major points.
read point-by-point responses
-
Referee: Abstract: The 99.3% validation and test accuracy is stated without any information on total images, number of patients, class distribution, dataset source, or train/validation/test partitioning strategy. Because multiple cells are routinely extracted from the same slide or patient, an image-level split risks leakage; the reported figure cannot be interpreted as evidence of reliable early detection until patient-stratified splits or external validation are demonstrated.
Authors: We agree that these details are essential and were omitted from the abstract and main text. In the revised manuscript we will add a dedicated 'Dataset' subsection in Methods that reports the total number of images, number of patients, class distribution, dataset source, and the precise train/validation/test partitioning protocol. We will also update the abstract to summarize these elements. To address leakage, we will re-process the data using patient-stratified splits (all cells from one patient assigned to only one subset) and report the resulting accuracies; if the original experiments used image-level splits, the revised results will reflect the corrected protocol. revision: yes
-
Referee: Experiments/Results section: No baseline comparisons (other YOLO versions, standard CNN classifiers, or non-deep-learning methods) are provided, nor is any confusion matrix, per-class accuracy, or error analysis reported. For a multiclass problem whose difficulty stems from visual similarity between cell types, these omissions leave the superiority claim unsupported.
Authors: We acknowledge that the absence of baselines and error analysis weakens the claims. In the revised manuscript we will expand the Experiments section to include direct comparisons with YOLOv8, YOLOv11, ResNet50, EfficientNet-B0, and a traditional SVM baseline using hand-crafted features. We will also add a confusion matrix, per-class precision/recall/F1 scores, and a short error-analysis paragraph discussing misclassifications between visually similar cell types. These additions will be based on re-running the experiments under the same preprocessing and patient-stratified splits. revision: yes
Circularity Check
No circularity; empirical accuracy on held-out test set with no derivation chain
full rationale
The paper reports an experimental result: YOLOv12 combined with Otsu thresholding on cell-based segmentation yields 99.3% validation and test accuracy for AML cell classification. No equations, fitted parameters renamed as predictions, self-definitional constructs, or load-bearing self-citations appear in the abstract or described claims. The central claim is a direct empirical measurement on a test set rather than a quantity derived by construction from its own inputs. Data-split details are unspecified, but that is a generalization risk, not a circularity in any derivation. The result is self-contained against external benchmarks and does not reduce to tautology.
Axiom & Free-Parameter Ledger
free parameters (1)
- YOLOv12 training hyperparameters
axioms (1)
- domain assumption YOLOv12 architecture can learn discriminative features from pre-segmented blood-cell images
Reference graph
Works this paper leans on
-
[1]
Automatic detection of acute leukemia (all and aml) utilizing customized deep graph convolutional neural networks,
L. Zare, M. Rahmani, N. Khaleghi, S. Sheykhivand, and S. Dan- ishvar, “Automatic detection of acute leukemia (all and aml) utilizing customized deep graph convolutional neural networks,” Bioengineering, vol. 11, no. 7, p. 644, 2024
2024
-
[2]
Acute myeloid leukemia - cancer stat facts
SEER Program, National Cancer Institute, “Acute myeloid leukemia - cancer stat facts.” https://seer.cancer.gov/statfacts/ html/amyl.html, n.d
-
[3]
Acute myeloid leukemia (nursing),
A. Vakiti, S. B. Reynolds, P. Mewawalla, and S. Coleman, “Acute myeloid leukemia (nursing),” inStatPearls [Internet], StatPearls Publishing, 2024
2024
-
[4]
Leukemia: an overview for primary care,
A. S. Davis, A. J. Viera, and M. D. Mead, “Leukemia: an overview for primary care,”American family physician, vol. 89, no. 9, pp. 731–738, 2014
2014
-
[5]
Chapter 18 - acute myeloid leukemia—overview,
F. Naeim, P. Nagesh Rao, S. X. Song, and R. T. Phan, “Chapter 18 - acute myeloid leukemia—overview,” inAtlas of Hematopathology (Second Edition)(F. Naeim, P. Nagesh Rao, S. X. Song, and R. T. Phan, eds.), pp. 293–302, Academic Press, second edition ed., 2018
2018
-
[6]
Chapter 27 - granulocy- topoiesis and monocytopoiesis,
A. Khanna-Gupta and N. Berliner, “Chapter 27 - granulocy- topoiesis and monocytopoiesis,” inHematology (Seventh Edi- tion)(R. Hoffman, E. J. Benz, L. E. Silberstein, H. E. Heslop, J. I. Weitz, J. Anastasi, M. E. Salama, and S. A. Abutalib, eds.), pp. 321–333.e1, Elsevier, seventh edition ed., 2018
2018
-
[7]
Approach to acute myeloid leukemia with increased eosinophils and basophils,
S. Papadakis, I. Liapis, S. I. Papadhimitriou, E. Spanoudakis, I. Kotsianidis, and K. Liapis, “Approach to acute myeloid leukemia with increased eosinophils and basophils,”Journal of Clinical Medicine, vol. 13, no. 3, p. 876, 2024
2024
-
[8]
Detecting malignant leukemia cells using microscopic blood smear images: a deep learning approach,
R. Baig, A. Rehman, A. Almuhaimeed, A. Alzahrani, and H. T. Rauf, “Detecting malignant leukemia cells using microscopic blood smear images: a deep learning approach,”Applied Sci- ences, vol. 12, no. 13, p. 6317, 2022
2022
-
[9]
Ghost-resnext: An effective deep learning based on mature and immature wbc classification,
S. S. R. Bairaboina and S. R. Battula, “Ghost-resnext: An effective deep learning based on mature and immature wbc classification,”Applied Sciences, vol. 13, no. 6, p. 4054, 2023
2023
-
[10]
Cae-resvgg fusionnet: A feature extraction framework integrating convo- lutional autoencoders and transfer learning for immature white blood cells in acute myeloid leukemia,
T. Elhassan, A. H. Osman, M. S. M. Rahim, S. Z. M. Hashim, A. Ali, E. Elhassan, Y . Elkamali, and M. Aljurf, “Cae-resvgg fusionnet: A feature extraction framework integrating convo- lutional autoencoders and transfer learning for immature white blood cells in acute myeloid leukemia,”Heliyon, vol. 10, no. 19, 2024
2024
-
[11]
YOLOv12: Attention-Centric Real-Time Object Detectors
Y . Tian, Q. Ye, and D. Doermann, “Yolov12: Attention- centric real-time object detectors. arxiv 2025,”arXiv preprint arXiv:2502.12524, 2025
work page internal anchor Pith review arXiv 2025
-
[12]
You only look once: Unified, real-time object detection,
J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look once: Unified, real-time object detection,” inProceedings of the IEEE conference on computer vision and pattern recog- nition, pp. 779–788, 2016
2016
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.