<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Bioinform Biotech</journal-id><journal-id journal-id-type="publisher-id">bioinform</journal-id><journal-id journal-id-type="index">19</journal-id><journal-title>JMIR Bioinformatics and Biotechnology</journal-title><abbrev-journal-title>JMIR Bioinform Biotech</abbrev-journal-title><issn pub-type="epub">2563-3570</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v6i1e68848</article-id><article-id pub-id-type="doi">10.2196/68848</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Paper</subject></subj-group></article-categories><title-group><article-title>Lung Cancer Diagnosis From Computed Tomography Images Using Deep Learning Algorithms With Random Pixel Swap Data Augmentation: Algorithm Development and Validation Study</article-title></title-group><contrib-group><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Abe</surname><given-names>Ayomide Adeyemi</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1"/></contrib><contrib contrib-type="author"><name name-style="western"><surname>Nyathi</surname><given-names>Mpumelelo</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1"/></contrib></contrib-group><aff id="aff1"><institution>Department of Medical Physics, Sefako Makgatho Health Science University</institution><addr-line>Molotlegi St, Zone 1, Garankuwa</addr-line><addr-line>Pretoria</addr-line><country>South Africa</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Hacking</surname><given-names>Sean</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Guerrero-Contreras</surname><given-names>Gabriel</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Li</surname><given-names>Keqin</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Zhang</surname><given-names>Tianle</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Xia</surname><given-names>Yunfei</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Zhao</surname><given-names>Zihao</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Ayomide Adeyemi Abe, PhD, Department of Medical Physics, Sefako Makgatho Health Science University, Molotlegi St, Zone 1, Garankuwa, Pretoria, 0208, South Africa, 27 813208348; <email>saintcodded@gmail.com</email></corresp></author-notes><pub-date pub-type="collection"><year>2025</year></pub-date><pub-date pub-type="epub"><day>3</day><month>9</month><year>2025</year></pub-date><volume>6</volume><elocation-id>e68848</elocation-id><history><date date-type="received"><day>15</day><month>11</month><year>2024</year></date><date date-type="rev-recd"><day>26</day><month>05</month><year>2025</year></date><date date-type="accepted"><day>23</day><month>06</month><year>2025</year></date></history><copyright-statement>&#x00A9; Ayomide Adeyemi Abe, Mpumelelo Nyathi. Originally published in JMIR Bioinformatics and Biotechnology (<ext-link ext-link-type="uri" xlink:href="https://bioinform.jmir.org">https://bioinform.jmir.org</ext-link>), 3.9.2025. </copyright-statement><copyright-year>2025</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by/4.0/">http://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Bioinformatics and Biotechnology, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://bioinform.jmir.org/">https://bioinform.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://bioinform.jmir.org/2025/1/e68848"/><abstract><sec><title>Background</title><p>Deep learning (DL) shows promise for automated lung cancer diagnosis, but limited clinical data can restrict performance. While data augmentation (DA) helps, existing methods struggle with chest computed tomography (CT) scans across diverse DL architectures.</p></sec><sec><title>Objective</title><p>This study proposes Random Pixel Swap (RPS), a novel DA technique, to enhance diagnostic performance in both convolutional neural networks and transformers for lung cancer diagnosis from CT scan images.</p></sec><sec sec-type="methods"><title>Methods</title><p>RPS generates augmented data by randomly swapping pixels within patient CT scan images. We evaluated it on ResNet, MobileNet, Vision Transformer, and Swin Transformer models, using 2 public CT datasets (Iraq-Oncology Teaching Hospital/National Center for Cancer Diseases [IQ-OTH/NCCD] dataset and chest CT scan images dataset), and measured accuracy and area under the receiver operating characteristic curve (AUROC). Statistical significance was assessed via paired <italic>t</italic> tests.</p></sec><sec sec-type="results"><title>Results</title><p>The RPS outperformed state-of-the-art DA methods (Cutout, Random Erasing, MixUp, and CutMix), achieving 97.56% accuracy and 98.61% AUROC on the IQ-OTH/NCCD dataset and 97.78% accuracy and 99.46% AUROC on the chest CT scan images dataset. While traditional augmentation approaches (flipping and rotation) remained effective, RPS complemented them, surpassing the performance findings in prior studies and demonstrating the potential of artificial intelligence for early lung cancer detection.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>The RPS technique enhances convolutional neural network and transformer models, enabling more accurate automated lung cancer detection from CT scan images.</p></sec></abstract><kwd-group><kwd>lung cancer diagnosis</kwd><kwd>deep learning</kwd><kwd>data augmentation</kwd><kwd>convolutional neural network</kwd><kwd>transformer</kwd><kwd>random pixel swap</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><sec id="s1-1"><title>Background</title><p>Lung cancer is a lethal disease characterized by uncontrolled cell growth in the lungs [<xref ref-type="bibr" rid="ref1">1</xref>]. These malignant cells can proliferate, invade nearby tissues, and metastasize to other parts of the body [<xref ref-type="bibr" rid="ref2">2</xref>]. The disease progresses through distinct stages, with advanced stages often proving fatal [<xref ref-type="bibr" rid="ref3">3</xref>]. Lung cancer comprises multiple histological types and subtypes, affecting individuals regardless of gender [<xref ref-type="bibr" rid="ref4">4</xref>]. Globally, lung cancer remains the leading cause of cancer-related mortality [<xref ref-type="bibr" rid="ref5">5</xref>]. In 2020 alone, it accounted for 1.8 million deaths, ranking as the 6th leading cause of death worldwide among individuals younger than 70 years [<xref ref-type="bibr" rid="ref2">2</xref>]. A key contributor to this high mortality is the frequent absence of early symptoms, leading to late-stage diagnosis and poorer outcomes [<xref ref-type="bibr" rid="ref6">6</xref>]. The 5-year survival rate for lung cancer patients remains low, emphasizing the critical need for early detection [<xref ref-type="bibr" rid="ref7">7</xref>]. Early diagnosis significantly improves prognosis, reduces long-term treatment costs, expands therapeutic options, and alleviates the burden on caregivers and families [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref8">8</xref>-<xref ref-type="bibr" rid="ref10">10</xref>]. However, most cases are still detected at advanced stages, drastically limiting survival rates [<xref ref-type="bibr" rid="ref5">5</xref>]. These challenges underscore lung cancer as a major public health priority.</p><p>Computed tomography (CT) is a medical imaging technique that produces high-resolution cross-sectional images of the lungs, providing detailed anatomical information for clinical evaluation [<xref ref-type="bibr" rid="ref11">11</xref>]. As a noninvasive diagnostic tool, CT imaging has become indispensable for the early detection of lung cancer, offering superior sensitivity compared to conventional radiography [<xref ref-type="bibr" rid="ref12">12</xref>,<xref ref-type="bibr" rid="ref13">13</xref>]. However, the interpretation of CT scans presents significant challenges in clinical practice. The process demands considerable expertise from radiologists, as subtle early-stage malignancies may demonstrate imaging features that escape human detection, potentially leading to diagnostic oversights [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref15">15</xref>]. The subjective nature of image interpretation introduces variability in diagnostic accuracy among practitioners, which can result in false-positive identification of pulmonary nodules. Such errors may prompt unnecessary invasive procedures for confirmation, exposing patients to avoidable risks and health care systems to additional costs [<xref ref-type="bibr" rid="ref13">13</xref>]. Furthermore, the comprehensive evaluation of CT examinations is particularly demanding, as each study comprises numerous sequential slices, requiring both individual assessment and integrated analysis. This labor-intensive process frequently overwhelms available radiological resources, contributing to diagnostic delays and extended patient waiting periods [<xref ref-type="bibr" rid="ref15">15</xref>-<xref ref-type="bibr" rid="ref17">17</xref>]. To address these limitations, computer-assisted diagnostic systems have been developed to augment radiologists&#x2019; interpretive capabilities [<xref ref-type="bibr" rid="ref18">18</xref>]. These automated solutions employ advanced algorithms to analyze CT images, enhancing diagnostic accuracy while improving workflow efficiency [<xref ref-type="bibr" rid="ref19">19</xref>]. By integrating such technological advancements into clinical practice, health care providers can mitigate the current challenges associated with manual CT interpretation, ultimately improving patient outcomes through more timely and reliable diagnoses.</p><p>The application of computer algorithms for the automated early diagnosis of lung cancer from CT scan images has evolved considerably. Early approaches used radiomics and machine learning techniques, but recent advancements have established deep learning (DL) as the predominant methodology [<xref ref-type="bibr" rid="ref20">20</xref>]. Unlike traditional methods that depend on manually engineered features, a process prone to bias and time constraints, DL employs artificial neural networks to autonomously extract sophisticated features through training [<xref ref-type="bibr" rid="ref21">21</xref>]. Among DL architectures, both convolutional neural networks (CNNs) and Vision Transformers have demonstrated exceptional potential for the early detection of lung cancer [<xref ref-type="bibr" rid="ref22">22</xref>]. CNNs gained prominence after 2012, while Vision Transformers emerged in 2020 [<xref ref-type="bibr" rid="ref23">23</xref>], with both now leading innovations in automated CT scan analysis [<xref ref-type="bibr" rid="ref18">18</xref>,<xref ref-type="bibr" rid="ref19">19</xref>].</p><p>CNNs and transformers offer distinct advantages for medical image analysis. CNNs, with their inductive bias for spatial locality and translation invariance, benefit from a simpler, parameter-efficient architecture rooted in spatial priors, which is highly effective and easier to train on smaller datasets [<xref ref-type="bibr" rid="ref24">24</xref>,<xref ref-type="bibr" rid="ref25">25</xref>]. They specialize in extracting local features and understanding spatial relationships between adjacent pixels. In contrast, transformers excel at capturing long-range dependencies across the entire image [<xref ref-type="bibr" rid="ref26">26</xref>]. Vision Transformers are particularly scalable, maintaining image resolution better than CNNs during processing [<xref ref-type="bibr" rid="ref27">27</xref>]. Their parallel processing capability also enables faster training times compared to similarly complex CNNs [<xref ref-type="bibr" rid="ref28">28</xref>], although they typically require larger training datasets to achieve comparable performance [<xref ref-type="bibr" rid="ref29">29</xref>]. Recent developments have seen the rise of hybrid networks that combine CNN and transformer architectures, successfully integrating both local and global feature extraction to overcome the limitations of standalone approaches [<xref ref-type="bibr" rid="ref30">30</xref>,<xref ref-type="bibr" rid="ref31">31</xref>].</p><p>Despite their capabilities, DL models face significant data-related challenges. While these architectures proficiently automate nodule detection, classification, and segmentation in CT scans [<xref ref-type="bibr" rid="ref32">32</xref>], they demand extensive training data to outperform radiologist interpretations [<xref ref-type="bibr" rid="ref33">33</xref>]. The scarcity of annotated medical CT datasets presents a major constraint [<xref ref-type="bibr" rid="ref34">34</xref>], as creating such datasets requires time-consuming, expert-driven image labeling [<xref ref-type="bibr" rid="ref35">35</xref>]. Data augmentation (DA) has emerged as a crucial solution to expand dataset size and diversity [<xref ref-type="bibr" rid="ref36">36</xref>], enhancing both the quantity and quality of available training samples [<xref ref-type="bibr" rid="ref37">37</xref>]. However, selecting appropriate DA techniques for chest CT analysis remains challenging due to several factors, including the variable effectiveness of methods across different datasets and domains [<xref ref-type="bibr" rid="ref38">38</xref>], potential label distortions and crucial information loss caused by certain transformations [<xref ref-type="bibr" rid="ref39">39</xref>], and current limitations in improving performance for both CNN and transformer architectures [<xref ref-type="bibr" rid="ref37">37</xref>,<xref ref-type="bibr" rid="ref40">40</xref>]. To address these challenges, this study proposes the Random Pixel Swap (RPS) augmentation method, specifically designed to enhance the generalization capabilities of both architectural paradigms in lung cancer diagnosis from chest CT scan images.</p></sec><sec id="s1-2"><title>Related Work</title><p>The effectiveness of DA in training large neural networks was first conclusively demonstrated in 2012 [<xref ref-type="bibr" rid="ref41">41</xref>], sparking the development of numerous innovative techniques [<xref ref-type="bibr" rid="ref37">37</xref>]. These methods primarily fall into 2 categories: data synthesis and data transformation [<xref ref-type="bibr" rid="ref36">36</xref>]. Data synthesis techniques generate novel samples that maintain statistical similarity to the original training data, while data transformation techniques create variations by modifying existing training samples. Both approaches effectively increase training dataset size, quality, and diversity, although they differ significantly in implementation. Data synthesis typically requires parameter learning, a process that can prove computationally intensive and often demands substantial training data to achieve optimal results [<xref ref-type="bibr" rid="ref42">42</xref>]. In contrast, data transformation techniques generally avoid parameter learning and consequently require less computational resources. Traditional data transformation methods include fundamental image manipulations such as flipping, rotation, cropping, translation, and photometric adjustments (modifications to brightness, saturation, contrast, and hue) [<xref ref-type="bibr" rid="ref36">36</xref>]. More sophisticated approaches like Cutout [<xref ref-type="bibr" rid="ref43">43</xref>], Random Erasing [<xref ref-type="bibr" rid="ref44">44</xref>], MixUp [<xref ref-type="bibr" rid="ref45">45</xref>], and CutMix [<xref ref-type="bibr" rid="ref46">46</xref>] have subsequently emerged, achieving state-of-the-art performance across various domains. These advanced techniques have been employed in lung cancer diagnosis from CT scan images [<xref ref-type="bibr" rid="ref47">47</xref>-<xref ref-type="bibr" rid="ref49">49</xref>].</p><p>The following section provides a comprehensive examination of the Cutout, Random Erasing, MixUp, and CutMix techniques, analyzing their limitations in medical imaging applications and contrasting them with the proposed RPS method. This comparative analysis establishes the foundation rationale for developing specialized augmentation approaches optimized for medical image analysis challenges.</p><sec id="s1-2-1"><title>Cutout DA Technique</title><p>The Cutout technique randomly selects square regions within images and masks their pixel values [<xref ref-type="bibr" rid="ref43">43</xref>]. While effective for improving model robustness against occlusions in natural images, this approach presents significant limitations for medical CT scans. The method&#x2019;s potential to eliminate critical diagnostic information (such as cancerous regions) may degrade performance [<xref ref-type="bibr" rid="ref38">38</xref>]. Additionally, the masking process can inadvertently alter image labels, further limiting effectiveness [<xref ref-type="bibr" rid="ref39">39</xref>]. Unlike Cutout, our RPS approach avoids information loss. It preserves diagnostic information by replacing masked regions with pixel values that are derived from other areas within the same CT scan while maintaining original labels.</p></sec><sec id="s1-2-2"><title>Random Erasing DA Technique</title><p>Random Erasing extends Cutout&#x2019;s functionality by supporting both square and rectangular masks of varying sizes [<xref ref-type="bibr" rid="ref44">44</xref>]. This technique randomly selects image regions for erasure and replaces them with random pixel values. While the variable mask sizes increase dataset diversity compared to Cutout, the method still suffers from information loss and label alteration issues [<xref ref-type="bibr" rid="ref36">36</xref>,<xref ref-type="bibr" rid="ref40">40</xref>]. These limitations are particularly problematic for medical imaging, where preserving anatomical content is crucial.</p></sec><sec id="s1-2-3"><title>MixUp DA Technique</title><p>MixUp generates new samples through linear interpolation of pixel values and labels from 2 distinct images [<xref ref-type="bibr" rid="ref45">45</xref>]. This approach enhances model generalization by preventing label memorization and improving adversarial robustness. However, the technique&#x2019;s potential to blur important anatomical boundaries and the requirement of careful hyperparameter tuning can create a bottleneck in medical contexts [<xref ref-type="bibr" rid="ref47">47</xref>,<xref ref-type="bibr" rid="ref48">48</xref>]. Furthermore, its convergence speed is often suboptimal [<xref ref-type="bibr" rid="ref47">47</xref>]. RPS addresses these limitations by operating within single patient scans rather than mixing data across patients and employs a single hyperparameter for more efficient training.</p></sec><sec id="s1-2-4"><title>CutMix DA Technique</title><p>CutMix combines aspects of previous methods by cutting patches from one image and pasting them onto another while proportionally blending labels [<xref ref-type="bibr" rid="ref46">46</xref>]. Although this approach leverages the benefits from both Cutout and MixUp, the label blending can introduce noise that degrades model performance [<xref ref-type="bibr" rid="ref50">50</xref>]. For medical CT scans, combining patches from different patients may confuse learning models, particularly when dealing with subtle pathological features [<xref ref-type="bibr" rid="ref51">51</xref>]. RPS overcomes these challenges by performing pixel swaps exclusively within individual patient scans and preserving original labels without blending. <xref ref-type="fig" rid="figure1">Figure 1</xref> visually contrasts these techniques with the proposed RPS method.</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Computed tomography images for various data augmentation techniques. (A) MixUp; (B) CutMix; (C) Cutout; (D) Random Erasing; (E) Random Pixel Swap. The original image is in column 1, while the augmented images are in columns 2, 3, 4, and 5.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="bioinform_v6i1e68848_fig01.png"/></fig></sec></sec></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>RPS DA Technique</title><p>The RPS technique is a parameter-free DA algorithm that operates with a predefined transformation probability. This method partitions input images into 2 distinct regions that serve as source and target areas for patch selection and swapping operations. The study proposes 4 specific implementation approaches, designated as RPS<sub>H</sub> (vertical), RPS<sub>W</sub> (horizontal), RPS<sub>U</sub> (upper right diagonal), and RPS<sub>D</sub> (upper left diagonal) swap configurations, as illustrated in <xref ref-type="fig" rid="figure2">Figure 2</xref>. This multidirectional swapping mechanism provides several advantages: it generates diverse transformations within individual patient CT scans while maintaining pathological plausibility, introduces meaningful variability in the training dataset without requiring parameter learning, and preserves all critical diagnostic information by operating exclusively within each scan&#x2019;s original pixel values. The technique&#x2019;s ability to produce multiple distinct transformations from a single image significantly enhances dataset diversity while avoiding the label alteration and information loss issues associated with other augmentation methods.</p><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>Four possible swap approaches for the Random Pixel Swap (RPS) data augmentation technique. I is the original image. Areas A<sub>1</sub> and A<sub>2</sub> are the swap regions. RPS<sub>H</sub> (vertical), RPS<sub>W</sub> (horizontal), RPS<sub>U</sub> (upper right diagonal), and RPS<sub>D</sub> (upper left diagonal) are the possible swap configurations.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="bioinform_v6i1e68848_fig02.png"/></fig><p>RPS possesses distinct invariant properties compared to other techniques. For an image with <inline-formula><mml:math id="ieqn1"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:mstyle></mml:math></inline-formula> pixels and <inline-formula><mml:math id="ieqn2"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> intensity levels, the RPS transformation preserves global intensity, as shown in <xref ref-type="disp-formula" rid="E1 E2 E3">Equations (1)-(3)</xref>. The technique employs a controlled, systematic, random patch-based pixel swap, rather than a random point-based pixel swap, ensuring that image content is preserved. This approach generates meaningful variations while maintaining pathological truth, thereby retaining clinical relevance in the context of lung cancer diagnosis.</p><disp-formula id="E1"><label>(1)</label><mml:math id="eqn1"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msup><mml:mi>X</mml:mi><mml:mo>&#x2032;</mml:mo></mml:msup><mml:mo>=</mml:mo><mml:mi>T</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>X</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mstyle></mml:math></disp-formula><p>where T is permutation transform</p><disp-formula id="E2"><label>(2)</label><mml:math id="eqn2"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>p</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>n</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>i</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mi>N</mml:mi></mml:mfrac><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>n</mml:mi><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mi>i</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x2032;</mml:mo></mml:msup></mml:mrow><mml:msup><mml:mi>N</mml:mi><mml:mo>&#x2032;</mml:mo></mml:msup></mml:mfrac><mml:mo>;</mml:mo><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mo>.</mml:mo><mml:mo>.</mml:mo><mml:mo>.</mml:mo><mml:mo>,</mml:mo><mml:mi>L</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:mstyle></mml:math></disp-formula><disp-formula id="E3"><label>(3)</label><mml:math id="eqn3"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:msub><mml:mi>I</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:mrow><mml:mrow><mml:mi>L</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:munderover><mml:mfrac><mml:mi>i</mml:mi><mml:mi>N</mml:mi></mml:mfrac><mml:mo>=</mml:mo><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:mrow><mml:mrow><mml:mi>L</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:munderover><mml:mfrac><mml:msup><mml:mi>i</mml:mi><mml:mo>&#x2032;</mml:mo></mml:msup><mml:msup><mml:mi>N</mml:mi><mml:mo>&#x2032;</mml:mo></mml:msup></mml:mfrac></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula><p>where <italic>P</italic> is the probability of a pixel having intensity <italic>i</italic>; <inline-formula><mml:math id="ieqn3"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>n</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>i</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mi>n</mml:mi><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mi>i</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x2032;</mml:mo></mml:msup></mml:mrow></mml:mstyle></mml:math></inline-formula> is the number of pixels with intensity level <inline-formula><mml:math id="ieqn4"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mi>X</mml:mi><mml:mo>&#x2227;</mml:mo><mml:msup><mml:mi>X</mml:mi><mml:mo>&#x2032;</mml:mo></mml:msup></mml:mrow></mml:mstyle></mml:math></inline-formula>, respectively; <inline-formula><mml:math id="ieqn5"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>N</mml:mi><mml:mo>,</mml:mo><mml:msup><mml:mi>N</mml:mi><mml:mo>&#x2032;</mml:mo></mml:msup></mml:mrow></mml:mstyle></mml:math></inline-formula> is the total number of pixels in <inline-formula><mml:math id="ieqn6"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>X</mml:mi><mml:mo>&#x2227;</mml:mo><mml:msup><mml:mi>X</mml:mi><mml:mo>&#x2032;</mml:mo></mml:msup></mml:mrow></mml:mstyle></mml:math></inline-formula>, respectively; <inline-formula><mml:math id="ieqn7"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>L</mml:mi></mml:mrow></mml:mstyle></mml:math></inline-formula> is the intensity level; and <inline-formula><mml:math id="ieqn8"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>I</mml:mi><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> is the average global intensity.</p></sec><sec id="s2-2"><title>Implementation of RPS</title><p>The RPS technique is implemented by first randomly selecting 2 coordinate points (x&#x2081;, x&#x2082;) along the x-axis and 2 points (y&#x2081;, y&#x2082;) along the y-axis within the input image. These coordinates define 2 equal subswap regions: region <inline-formula><mml:math id="ieqn9"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:mstyle></mml:math></inline-formula> bounded by swap area <inline-formula><mml:math id="ieqn10"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>A</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula>:(x&#x2081;, y&#x2081;) and (x&#x2082;, y&#x2082;), and <inline-formula><mml:math id="ieqn11"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>A</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> bounded by swap area <inline-formula><mml:math id="ieqn12"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>A</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula>:(x&#x2081;, y&#x2081;)&#x2032; and (x&#x2082;, y&#x2082;)&#x2032;. The method incorporates a key hyperparameter called the swap area factor <inline-formula><mml:math id="ieqn13"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula>, which ranges from 0.1 to 1.0, to control the extent of augmentation. The actual swap areas <inline-formula><mml:math id="ieqn14"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> and <inline-formula><mml:math id="ieqn15"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> are derived by scaling the subswap regions using this factor, as specified in <xref ref-type="disp-formula" rid="E4 E5">Equations (4) and (5)</xref>. During the augmentation process, the contents of swap area <inline-formula><mml:math id="ieqn16"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> are cropped and pasted into swap area <inline-formula><mml:math id="ieqn17"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> while simultaneously transferring the contents of swap area <inline-formula><mml:math id="ieqn18"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> to swap area <inline-formula><mml:math id="ieqn19"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula>. This bidirectional swapping ensures comprehensive data transformation while preserving all original image information. The complete RPS procedure is formally described in <xref ref-type="other" rid="box1">Textbox 1</xref>.</p><disp-formula id="E4"><label>(4)</label><mml:math id="eqn4"><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>a</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2217;</mml:mo><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced></mml:math></disp-formula><disp-formula id="E5"><label>(5)</label><mml:math id="eqn5"><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>a</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2217;</mml:mo><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced></mml:math></disp-formula><boxed-text id="box1"><title> Algorithm 1: Random Pixel Swap data augmentation procedure.</title><p><bold>Input:</bold> data <inline-formula><mml:math id="ieqn20"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:mstyle></mml:math></inline-formula>; with shape <inline-formula><mml:math id="ieqn21"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>H</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>W</mml:mi></mml:mrow></mml:mstyle></mml:math></inline-formula></p><p><bold>Output:</bold> Augmented data <inline-formula><mml:math id="ieqn22"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>X</mml:mi><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:mstyle></mml:math></inline-formula></p><p>1: <inline-formula><mml:math id="ieqn23"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>A</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2208;</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:mfrac><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi mathvariant="normal">H</mml:mi></mml:mrow><mml:mo>&#x2217;</mml:mo><mml:mrow><mml:mi mathvariant="normal">W</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mstyle></mml:math></inline-formula></p><p>2: <bold>Init:</bold> All points <inline-formula><mml:math id="ieqn24"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>P</mml:mi></mml:mrow></mml:mstyle></mml:math></inline-formula> within <inline-formula><mml:math id="ieqn25"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>A</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula></p><p>3: <inline-formula><mml:math id="ieqn26"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> &#x2190; <inline-formula><mml:math id="ieqn27"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> &#x03F5; &#x211A; : <inline-formula><mml:math id="ieqn28"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> &#x03F5; [0.1, 1.0]</p><p>4: <bold>for</bold> <inline-formula><mml:math id="ieqn29"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mi>&#x03F5;</mml:mi><mml:mi>P</mml:mi><mml:mo>,</mml:mo><mml:mi>d</mml:mi><mml:mi>o</mml:mi></mml:mrow></mml:mstyle></mml:math></inline-formula></p><p>5: Randomly select <inline-formula><mml:math id="ieqn30"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> , <inline-formula><mml:math id="ieqn31"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> ,</p><p><inline-formula><mml:math id="ieqn32"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msubsup><mml:mi>P</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mo>&#x2032;</mml:mo></mml:msubsup></mml:mrow></mml:mstyle></mml:math></inline-formula> ,<inline-formula><mml:math id="ieqn33"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msubsup><mml:mi>P</mml:mi><mml:mrow><mml:mi>j</mml:mi></mml:mrow><mml:mo>&#x2032;</mml:mo></mml:msubsup></mml:mrow></mml:mstyle></mml:math></inline-formula> = <inline-formula><mml:math id="ieqn34"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2217;</mml:mo><mml:mn>2</mml:mn></mml:mrow></mml:mstyle></mml:math></inline-formula> , <inline-formula><mml:math id="ieqn35"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2217;</mml:mo><mml:mn>2</mml:mn></mml:mrow></mml:mstyle></mml:math></inline-formula></p><p>6: <inline-formula><mml:math id="ieqn36"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>A</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> = Area (<inline-formula><mml:math id="ieqn37"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> , <inline-formula><mml:math id="ieqn38"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> )</p><p>7: <inline-formula><mml:math id="ieqn39"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>A</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> = Area (<inline-formula><mml:math id="ieqn40"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msubsup><mml:mi>P</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mo>&#x2032;</mml:mo></mml:msubsup></mml:mrow></mml:mstyle></mml:math></inline-formula> , <inline-formula><mml:math id="ieqn41"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msubsup><mml:mi>P</mml:mi><mml:mrow><mml:mi>j</mml:mi></mml:mrow><mml:mo>&#x2032;</mml:mo></mml:msubsup></mml:mrow></mml:mstyle></mml:math></inline-formula> )</p><p>8: <inline-formula><mml:math id="ieqn42"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> = <inline-formula><mml:math id="ieqn43"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>A</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> * <inline-formula><mml:math id="ieqn44"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula></p><p>9: <inline-formula><mml:math id="ieqn45"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> = <inline-formula><mml:math id="ieqn46"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>A</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> * <inline-formula><mml:math id="ieqn47"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula></p><p>10: <inline-formula><mml:math id="ieqn48"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:mstyle></mml:math></inline-formula>* &#x2190;Replace <inline-formula><mml:math id="ieqn49"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> with <inline-formula><mml:math id="ieqn50"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> in <inline-formula><mml:math id="ieqn51"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:mstyle></mml:math></inline-formula> and <inline-formula><mml:math id="ieqn52"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> with <inline-formula><mml:math id="ieqn53"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> in <inline-formula><mml:math id="ieqn54"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:mstyle></mml:math></inline-formula></p><p>11: <bold>end for</bold></p><p>12: <bold>return</bold> <inline-formula><mml:math id="ieqn55"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:mstyle></mml:math></inline-formula>*</p></boxed-text></sec><sec id="s2-3"><title>Swap Area Factor</title><p>The swap area factor <inline-formula><mml:math id="ieqn56"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></inline-formula> is a crucial parameter in the RPS technique, representing the ratio between the subswap region and the total swap area as described in <xref ref-type="disp-formula" rid="E6">Equation (6)</xref>. This factor plays a vital role in the augmentation process for two key reasons: (1) it allows customization for different DL architectures that may benefit from varying swap region sizes, and (2) it helps maintain clinical relevance by limiting distortion of diagnostically important anatomical features. The study proposes two distinct implementations of this parameter: (1) single-value swap area factor (SVSF), which applies a fixed value throughout the augmentation process, and (2) multivalue swap area factor (MVSF), which uses multiple values to generate more diverse swap areas. In both implementations, the swap area factor operates within a defined range of 0.1 to 1.0, providing controlled flexibility for different medical imaging scenarios.</p><disp-formula id="E6"><label>(6)</label><mml:math id="eqn6"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:msub><mml:mi>A</mml:mi><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:mfrac></mml:mrow></mml:mstyle></mml:math></disp-formula></sec><sec id="s2-4"><title>Experimental Validation of the RPS Technique</title><p>We conducted comprehensive experiments to validate the effectiveness of the proposed RPS technique in enhancing DL model performance across both CNN and transformer architectures. For our evaluation, we selected 4 established models: ResNet-34 [<xref ref-type="bibr" rid="ref52">52</xref>], MobileNetV3 (small variant) [<xref ref-type="bibr" rid="ref53">53</xref>], Vision Transformer (base-16) [<xref ref-type="bibr" rid="ref23">23</xref>], and Swin Transformer (tiny version) [<xref ref-type="bibr" rid="ref29">29</xref>], all initialized with preactivated weights. These architectures were chosen based on three key criteria: (1) public availability for reproducible benchmarking, (2) widespread adoption in methodological comparisons [<xref ref-type="bibr" rid="ref29">29</xref>,<xref ref-type="bibr" rid="ref48">48</xref>], and (3) efficient training characteristics due to their relatively fewer trainable parameters compared to larger variants.</p><p>Our experimental design incorporated three key comparisons: (1) models trained without any augmentation, (2) models trained with RPS augmentation, and (3) models trained with 4 state-of-the-art DA techniques (Cutout [<xref ref-type="bibr" rid="ref43">43</xref>], Random Erasing [<xref ref-type="bibr" rid="ref44">44</xref>], MixUp [<xref ref-type="bibr" rid="ref45">45</xref>], and CutMix [<xref ref-type="bibr" rid="ref46">46</xref>]). These comparison techniques were selected because they represent current best practices in parameter-free augmentation methods that share conceptual similarities with RPS [<xref ref-type="bibr" rid="ref48">48</xref>]. We evaluated all models using two key metrics: (1) classification accuracy and (2) area under the receiver operating characteristic curve (AUROC), providing a comprehensive assessment of both overall performance and diagnostic discrimination capability.</p></sec><sec id="s2-5"><title>Experimental Setup and Implementation</title><p>All experiments were conducted using Python 3.12.2 (Python Software Foundation) and PyTorch 2.2.2+ cu118 (PyTorch Foundation) within Jupyter Notebook 7.0.8 (IPython Project), running on an NVIDIA Quadro RTX 3000 GPU (Nvidia Corporation). We adopted the AdamW optimizer with a cross-entropy loss function, using a batch size of 16. The StepLR scheduler was configured with a step size of 10 and a gamma value of 0.5 [<xref ref-type="bibr" rid="ref52">52</xref>]. Models were trained for 50 epochs, as additional training resulted in overfitting and performance degradation. After evaluating various learning rates, we selected 1&#x00D7;10&#x207B;&#x2074; as it yielded optimal results. Image normalization was applied with mean and SD values of 0.5 to enhance training stability and accelerate convergence [<xref ref-type="bibr" rid="ref53">53</xref>].</p><p>For RPS implementation, we used a swap area factor of 1.0 with an augmentation probability of 1.0 for all experiments. CNN models processed images at 512&#x00D7;512 and 224&#x00D7;224 resolutions, while transformer architectures used 224&#x00D7;224 resolution due to the Vision Transformer&#x2019;s input size limitations. Although the Swin Transformer supports 512&#x00D7;512 inputs, we maintained a consistent 224&#x00D7;224 resolution across all transformer experiments for fair comparison. All experiments were conducted with a random seed of 42 after verifying consistent performance patterns across 3 different seeds.</p></sec><sec id="s2-6"><title>Statistical Analysis</title><p>To evaluate our hypothesis that an effective DA technique should perform consistently across both CNN and transformer architectures, we treated each technique as an independent variable and considered model performance as the dependent variable. We used paired sample <italic>t</italic> tests [<xref ref-type="bibr" rid="ref54">54</xref>] to assess significant differences between techniques, considering <italic>P</italic> values &#x003C;.05 as statistically significant.</p><p>For comprehensive technique comparison, we implemented a ranking system based on cumulative scores <inline-formula><mml:math id="ieqn57"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>C</mml:mi></mml:mrow></mml:mstyle></mml:math></inline-formula> (<xref ref-type="disp-formula" rid="E7 E8">Equations (7) and (8)</xref>), where higher scores received lower rank numbers <inline-formula><mml:math id="ieqn58"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>R</mml:mi></mml:mrow></mml:mstyle></mml:math></inline-formula>. This approach enabled holistic performance benchmarking across all models and architectures.</p><disp-formula id="E7"><label>(7)</label><mml:math id="eqn7"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mi>C</mml:mi><mml:mo>=</mml:mo><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>m</mml:mi><mml:mi>o</mml:mi><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>l</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:munderover><mml:mi>m</mml:mi><mml:mi>o</mml:mi><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>l</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>A</mml:mi><mml:mo>+</mml:mo><mml:mi>A</mml:mi><mml:mi>U</mml:mi><mml:mi>R</mml:mi><mml:mi>O</mml:mi><mml:mi>C</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula><disp-formula id="E8"><label>(8)</label><mml:math id="eqn8"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mtable columnalign="left left" rowspacing="4pt" columnspacing="1em"><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:mi mathvariant="normal">R</mml:mi></mml:mrow><mml:mn>1</mml:mn></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mi mathvariant="normal">R</mml:mi></mml:mrow><mml:mn>2</mml:mn></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mi mathvariant="normal">R</mml:mi></mml:mrow><mml:mn>3</mml:mn></mml:msup><mml:mo>,</mml:mo><mml:mo>&#x2026;</mml:mo><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mi mathvariant="normal">R</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mn>1</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mn>2</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mn>3</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mo>&#x2026;</mml:mo><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>&#x2228;</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:msub><mml:mrow><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mn>1</mml:mn></mml:msub><mml:mo>&#x003E;</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mn>2</mml:mn></mml:msub><mml:mo>&#x003E;</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mn>3</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mo>&#x2026;</mml:mo><mml:mo>,</mml:mo><mml:mo>&#x003E;</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">C</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mtd></mml:mtr></mml:mtable></mml:mrow></mml:mstyle></mml:math></disp-formula><p>where <bold><inline-formula><mml:math id="ieqn59"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>C</mml:mi></mml:mrow></mml:mstyle></mml:math></inline-formula></bold> is cumulative score, <italic><bold><inline-formula><mml:math id="ieqn60"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>R</mml:mi></mml:mrow></mml:mstyle></mml:math></inline-formula></bold></italic> is rank, <bold><inline-formula><mml:math id="ieqn61"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:mstyle></mml:math></inline-formula></bold> is the total number of data augmentation techniques, <bold><inline-formula><mml:math id="ieqn62"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:mstyle></mml:math></inline-formula></bold> is the total number of selected models, <bold><inline-formula><mml:math id="ieqn63"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:mstyle></mml:math></inline-formula></bold> is accuracy, and <inline-formula><mml:math id="ieqn64"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>A</mml:mi><mml:mi>U</mml:mi><mml:mi>R</mml:mi><mml:mi>O</mml:mi><mml:mi>C</mml:mi></mml:mrow></mml:mstyle></mml:math></inline-formula> is the area under the receiver operating characteristic curve.</p></sec><sec id="s2-7"><title>Iraq-Oncology Teaching Hospital/National Center for Cancer Diseases Dataset</title><p>The Iraq-Oncology Teaching Hospital/National Center for Cancer Diseases (IQ-OTH/NCCD) dataset contains 1097 JPEG CT images collected from 110 patients [<xref ref-type="bibr" rid="ref35">35</xref>]. These images were obtained using a SOMATOM Siemens scanner (Siemens Healthineers) and encompass a diverse range of demographic characteristics. The dataset is organized into 3 categories: normal scans, benign tumor scans, and malignant tumor scans. Specifically, it includes 15 cases of benign tumors, totaling 120 images; 40 cases of malignant tumors, totaling 416 images; and 55 cases of normal findings, totaling 561 images. Each image has a resolution of 512&#x00D7;512 pixels. We divided the images in a ratio of 7:3 for training and testing.</p></sec><sec id="s2-8"><title>Chest CT Scan Images Dataset</title><p>The chest CT scan images dataset contains 1000 lung CT scans from patients diagnosed with 3 different types of lung cancers, as well as scans from healthy individuals, all in JPG format [<xref ref-type="bibr" rid="ref55">55</xref>]. The lung cancer types included in the dataset are adenocarcinoma, squamous cell carcinoma, and large cell carcinoma. The images are organized into training, testing, and validation sets for each lung cancer category.</p></sec><sec id="s2-9"><title>Ethical Considerations</title><p>Ethics approval was obtained from the Sefako Makgatho University Research Committee (ethics reference number: SMUREC/M/12/2022:PG).</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Average Training Time Overhead</title><p>To evaluate the computational impact of the RPS technique, we measured training duration for 4 architectures (ResNet-34, MobileNetV3 [small], Vision Transformer [base-16], and Swin Transformer [tiny]) with and without RPS implementation. The training time overhead was calculated as the difference between augmented and nonaugmented training times. Experiments were conducted on both the IQ-OTH/NCCD and chest CT scan datasets using 224&#x00D7;224 image resolution, with results averaged across 3 independent runs for reliability.</p><p>Our analysis included a comparative assessment of 4 established DA techniques: Cutout, Random Erasing, MixUp, and CutMix. Results demonstrated that while RPS increased training times across all models compared to nonaugmented training, this increase was not statistically significant (<italic>P</italic>=.07). Similarly, comparisons between RPS and other DA techniques revealed no statistically significant differences in computational overhead (Cutout: <italic>P</italic>=.06; Random Erasing: <italic>P</italic>=.17; MixUp: <italic>P</italic>=.49; CutMix: <italic>P</italic>=.16). Among all evaluated methods, RPS showed the highest training time overhead, followed sequentially by MixUp, CutMix, Random Erasing, and Cutout. Complete results are presented in <xref ref-type="fig" rid="figure3">Figure 3</xref>.</p><fig position="float" id="figure3"><label>Figure 3.</label><caption><p>Average training time overhead of 5 data augmentation techniques across 4 deep learning models.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="bioinform_v6i1e68848_fig03.png"/></fig></sec><sec id="s3-2"><title>Performance Comparison of RPS With State-of-the-Art DA Techniques for Lung Cancer Detection</title><p>To evaluate pulmonary nodule detection in chest CT scan images, the selected CNN and transformer models (ResNet-34, MobileNetV3 [small], Vision Transformer [base-16], and Swin Transformer [tiny]) were trained on the IQ-OTH/NCCD dataset to classify the scan images as normal or containing benign or malignant pulmonary nodules. Experimental results demonstrated that RPS significantly enhanced performance across all 4 architectures (<italic>P</italic>=.008). The MobileNetV3 model achieved particular success when combined with RPS using 512&#x00D7;512 image resolution, reaching a peak classification accuracy of 94.21%, representing a 1.22% accuracy improvement and 0.86% AUROC increase over the baseline model.</p><p>At 224&#x00D7;224 image resolution, our comprehensive comparison of RPS against the 4 established DA methods (Cutout: <italic>P</italic>=.03; Random Erasing: <italic>P</italic>=.008; MixUp: <italic>P</italic>=.02; CutMix: <italic>P</italic>=.02) revealed consistent superiority of the RPS technique (<italic>P</italic>&#x003C;.05). For ResNet-34, RPS exceeded CutMix (the best alternative) by 2.44% and Random Erasing (the least effective) by 5.49% in accuracy. MobileNetV3 showed a 0.3% improvement over Cutout (best alternative) and 1.83% over MixUp (least effective) in accuracy. Transformer architectures demonstrated even more pronounced benefits: Vision Transformer with RPS outperformed Random Erasing by 1.52% and MixUp by 16.77%, while Swin Transformer showed a 1.53% improvement over MixUp and 4.57% over Cutout in accuracy. Across all architectures, performance ranking was as follows: (1) RPS (best technique), (2) Random Erasing, (3) CutMix, (4) MixUp, and (5) Cutout. The detailed results are presented in <xref ref-type="table" rid="table1">Table 1</xref>.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Classification results of the IQ-OTH/NCCD<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup> dataset using preactivated deep learning models with various data augmentation techniques (224&#x00D7;224 image resolution).</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="top">Data augmentation</td><td align="left" valign="top">Rank<sup><xref ref-type="table-fn" rid="table1fn2">b</xref></sup></td><td align="left" valign="top" colspan="2">ResNet-34</td><td align="left" valign="top" colspan="2">MobileNetV3 (small)</td><td align="left" valign="top" colspan="2">Vision Transformer (base-16)</td><td align="left" valign="top" colspan="2">Swin Transformer (tiny)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top">Accuracy, %</td><td align="left" valign="top">AUROC<sup><xref ref-type="table-fn" rid="table1fn3">c</xref></sup>, %</td><td align="left" valign="top">Accuracy, %</td><td align="left" valign="top">AUROC, %</td><td align="left" valign="top">Accuracy, %</td><td align="left" valign="top">AUROC, %</td><td align="left" valign="top">Accuracy, %</td><td align="left" valign="top">AUROC, %</td></tr></thead><tbody><tr><td align="left" valign="top">Base model<sup><xref ref-type="table-fn" rid="table1fn4">d</xref></sup></td><td align="left" valign="top">6</td><td align="left" valign="top">85.98</td><td align="left" valign="top">83.39</td><td align="left" valign="top">86.59</td><td align="left" valign="top">93.16</td><td align="left" valign="top">57.62</td><td align="left" valign="top">64.88</td><td align="left" valign="top">85.67</td><td align="left" valign="top">89.06</td></tr><tr><td align="left" valign="top">Cutout<sup><xref ref-type="table-fn" rid="table1fn4">d</xref></sup></td><td align="left" valign="top">5</td><td align="left" valign="top">85.67</td><td align="left" valign="top">86.11</td><td align="left" valign="top">89.33</td><td align="left" valign="top">92.95</td><td align="left" valign="top">57.01</td><td align="left" valign="top">63.45</td><td align="left" valign="top">85.37</td><td align="left" valign="top">93.68</td></tr><tr><td align="left" valign="top">Random Erasing<sup><xref ref-type="table-fn" rid="table1fn4">d</xref></sup></td><td align="left" valign="top">2</td><td align="left" valign="top">82.62</td><td align="left" valign="top">91.23</td><td align="left" valign="top">88.72</td><td align="left" valign="top">90.10</td><td align="left" valign="top">71.65</td><td align="left" valign="top">75.41<sup><xref ref-type="table-fn" rid="table1fn5">e</xref></sup></td><td align="left" valign="top">86.89</td><td align="left" valign="top">91.42</td></tr><tr><td align="left" valign="top">MixUp<sup><xref ref-type="table-fn" rid="table1fn4">d</xref></sup></td><td align="left" valign="top">4</td><td align="left" valign="top">84.45</td><td align="left" valign="top">91.05</td><td align="left" valign="top">87.80</td><td align="left" valign="top">86.57</td><td align="left" valign="top">56.40</td><td align="left" valign="top">68.55</td><td align="left" valign="top">88.41</td><td align="left" valign="top">92.51</td></tr><tr><td align="left" valign="top">CutMix<sup><xref ref-type="table-fn" rid="table1fn4">d</xref></sup><sup>,</sup><sup><xref ref-type="table-fn" rid="table1fn6">f</xref></sup></td><td align="left" valign="top">3</td><td align="left" valign="top">85.67</td><td align="left" valign="top">88.34</td><td align="left" valign="top">88.41</td><td align="left" valign="top">93.02</td><td align="left" valign="top">68.90</td><td align="left" valign="top">69.68</td><td align="left" valign="top">88.41</td><td align="left" valign="top">92.05</td></tr><tr><td align="left" valign="top">Random Pixel Swap<sup><xref ref-type="table-fn" rid="table1fn6">f</xref></sup></td><td align="left" valign="top">1</td><td align="left" valign="top">88.11<sup><xref ref-type="table-fn" rid="table1fn5">e</xref></sup></td><td align="left" valign="top">93.70<sup><xref ref-type="table-fn" rid="table1fn5">e</xref></sup></td><td align="left" valign="top">89.63<sup><xref ref-type="table-fn" rid="table1fn5">e</xref></sup></td><td align="left" valign="top">93.80<sup><xref ref-type="table-fn" rid="table1fn5">e</xref></sup></td><td align="left" valign="top">73.17<sup><xref ref-type="table-fn" rid="table1fn5">e</xref></sup></td><td align="left" valign="top">74.64</td><td align="left" valign="top">89.94<sup><xref ref-type="table-fn" rid="table1fn5">e</xref></sup></td><td align="left" valign="top">94.79<sup><xref ref-type="table-fn" rid="table1fn5">e</xref></sup></td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup>IQ-OTH/NCCD: Iraq-Oncology Teaching Hospital/National Center for Cancer Diseases.</p></fn><fn id="table1fn2"><p><sup>b</sup>Rank represents the overall rating for each technique, with &#x201C;1&#x201D; indicating the best technique across all models.</p></fn><fn id="table1fn3"><p><sup>c</sup>AUROC: area under the receiver operating characteristic curve.</p></fn><fn id="table1fn4"><p><sup>d</sup>Significant difference between an augmentation technique and the Random Pixel Swap technique across all models.</p></fn><fn id="table1fn5"><p><sup>e</sup>Highest value in the column.</p></fn><fn id="table1fn6"><p><sup>f</sup>Significant difference between training using an augmentation technique and the base model across all models.</p></fn></table-wrap-foot></table-wrap><p>At 512&#x00D7;512 image resolution, ResNet-34 exhibited nuanced performance differences between augmentation techniques: while CutMix achieved a marginal 0.31% higher accuracy than RPS, RPS demonstrated significantly superior diagnostic capability with a 5.31% improvement in AUROC. Furthermore, RPS outperformed the least effective technique (Random Erasing) by 2.13% in accuracy and 3.17% in AUROC. For MobileNetV3, RPS dominated all comparative techniques in both accuracy and AUROC, except for a 1.23% AUROC advantage by CutMix. Specifically, RPS exceeded Cutout (the best alternative technique) by 0.61% and surpassed MixUp (the least effective) by 4.58% in accuracy. Across all evaluated methods, the overall performance ranking was as follows: (1) RPS (best technique), (2) Cutout, (3) CutMix, (4) MixUp, and (5) Random Erasing. The detailed results are presented in <xref ref-type="table" rid="table2">Table 2</xref>.</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Classification results of the IQ-OTH/NCCD<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup> dataset using preactivated deep learning models with various data augmentation techniques (512&#x00D7;512 image resolution).</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="top">Data augmentation</td><td align="left" valign="top">Rank<sup><xref ref-type="table-fn" rid="table2fn2">b</xref></sup></td><td align="left" valign="top" colspan="2">ResNet-34</td><td align="left" valign="top" colspan="2">MobileNetV3 (small)</td></tr><tr><td align="left" valign="bottom"/><td align="left" valign="bottom"/><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC<sup><xref ref-type="table-fn" rid="table2fn3">c</xref></sup>, %</td><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC, %</td></tr></thead><tbody><tr><td align="left" valign="top">Base model<sup><xref ref-type="table-fn" rid="table2fn4">d</xref></sup></td><td align="left" valign="top">6</td><td align="left" valign="top">88.72</td><td align="left" valign="top">78.51</td><td align="left" valign="top">92.99</td><td align="left" valign="top">94.81</td></tr><tr><td align="left" valign="top">Cutout<sup><xref ref-type="table-fn" rid="table2fn4">d</xref></sup></td><td align="left" valign="top">2</td><td align="left" valign="top">90.24</td><td align="left" valign="top">93.25</td><td align="left" valign="top">93.60</td><td align="left" valign="top">95.42</td></tr><tr><td align="left" valign="top">Random Erasing<sup><xref ref-type="table-fn" rid="table2fn4">d</xref></sup></td><td align="left" valign="top">5</td><td align="left" valign="top">89.94</td><td align="left" valign="top">91.85</td><td align="left" valign="top">90.85</td><td align="left" valign="top">92.19</td></tr><tr><td align="left" valign="top">MixUp<sup><xref ref-type="table-fn" rid="table2fn4">d</xref></sup></td><td align="left" valign="top">4</td><td align="left" valign="top">89.94</td><td align="left" valign="top">96.13<sup><xref ref-type="table-fn" rid="table2fn5">e</xref></sup></td><td align="left" valign="top">89.63</td><td align="left" valign="top">95.18</td></tr><tr><td align="left" valign="top">CutMix</td><td align="left" valign="top">3</td><td align="left" valign="top">92.38<sup><xref ref-type="table-fn" rid="table2fn5">e</xref></sup></td><td align="left" valign="top">89.71</td><td align="left" valign="top">92.68</td><td align="left" valign="top">96.90<sup><xref ref-type="table-fn" rid="table2fn5">e</xref></sup></td></tr><tr><td align="left" valign="top">Random Pixel Swap</td><td align="left" valign="top">1</td><td align="left" valign="top">92.07</td><td align="left" valign="top">95.02</td><td align="left" valign="top">94.21<sup><xref ref-type="table-fn" rid="table2fn5">e</xref></sup></td><td align="left" valign="top">95.67</td></tr></tbody></table><table-wrap-foot><fn id="table2fn1"><p><sup>a</sup>IQ-OTH/NCCD: Iraq-Oncology Teaching Hospital/National Center for Cancer Diseases.</p></fn><fn id="table2fn2"><p><sup>b</sup>Rank represents the overall rating for each technique, with &#x201C;1&#x201D; indicating the best technique across all models.</p></fn><fn id="table2fn3"><p><sup>c</sup>AUROC: area under the receiver operating characteristic curve.</p></fn><fn id="table2fn4"><p><sup>d</sup>Significant difference between an augmentation technique and the Random Pixel Swap technique across all models.</p></fn><fn id="table2fn5"><p><sup>e</sup>Highest value in the column.</p></fn></table-wrap-foot></table-wrap></sec><sec id="s3-3"><title>Performance Comparison of RPS With State-of-the-Art DA Techniques for Lung Cancer Classification From CT Scan Images Using DL Architectures</title><p>We evaluated the effectiveness of the RPS technique for lung cancer classification using the chest CT scan images dataset across multiple DL architectures. The experimental results demonstrated that RPS significantly enhanced classification performance for all architectures (<italic>P</italic>=.008). RPS combined with ResNet-34 at 512&#x00D7;512 image resolution achieved optimal performance, reaching 97.78% accuracy and 99.46% AUROC.</p><p>At 224&#x00D7;224 image resolution, RPS consistently outperformed competing techniques across most models (Cutout: <italic>P</italic>=.001; Random Erasing: <italic>P</italic>=.02; MixUp: <italic>P</italic>=.047; CutMix: <italic>P</italic>=.18). For ResNet-34, RPS exceeded CutMix (the best alternative) by 0.64% and Random Erasing (the least effective) by 5.08% in accuracy. MobileNetV3 showed even greater improvements over other methods, with RPS surpassing CutMix by 3.49% and MixUp by 9.21% in accuracy. For the implementation with Vision Transformer, RPS surpassed Random Erasing (the best alternative) by 1.91% and MixUp (the least effective) by 18.85% in accuracy. While CutMix showed a 2.22% accuracy advantage over RPS for the Swin Transformer, RPS maintained superior performance against all other techniques, exceeding Cutout by 7.3% (the least effective). Across all architectures, the overall performance ranking was as follows: (1) RPS (best technique), (2) CutMix, (3) Random Erasing, (4) Cutout, and (5) MixUp. The detailed results are presented in <xref ref-type="table" rid="table3">Table 3</xref>.</p><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>Classification results of the chest CT<sup><xref ref-type="table-fn" rid="table3fn1">a</xref></sup> scan images dataset using preactivated deep learning models with various data augmentation techniques (224&#x00D7;224 image resolution).</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="top">Data augmentation</td><td align="left" valign="top">Rank<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top" colspan="2">ResNet-34</td><td align="left" valign="top" colspan="2">MobileNetV3 (small)</td><td align="left" valign="top" colspan="2">Vision Transformer (base-16)</td><td align="left" valign="top" colspan="2">Swin Transformer (tiny)</td></tr><tr><td align="left" valign="bottom"/><td align="left" valign="bottom"/><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC<sup><xref ref-type="table-fn" rid="table3fn3">c</xref></sup>, %</td><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC, %</td><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC, %</td><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC, %</td></tr></thead><tbody><tr><td align="left" valign="top">Base model<sup><xref ref-type="table-fn" rid="table3fn4">d</xref></sup></td><td align="left" valign="top">5</td><td align="left" valign="top">93.33</td><td align="left" valign="top">99.00</td><td align="left" valign="top">87.30</td><td align="left" valign="top">97.09</td><td align="left" valign="top">82.86</td><td align="left" valign="top">95.84</td><td align="left" valign="top">84.76</td><td align="left" valign="top">96.92</td></tr><tr><td align="left" valign="top">Cutout<sup><xref ref-type="table-fn" rid="table3fn4">d</xref></sup><sup>,</sup><sup><xref ref-type="table-fn" rid="table3fn5">e</xref></sup></td><td align="left" valign="top">4</td><td align="left" valign="top">93.02</td><td align="left" valign="top">98.94</td><td align="left" valign="top">85.71</td><td align="left" valign="top">97.62</td><td align="left" valign="top">80.63</td><td align="left" valign="top">94.35</td><td align="left" valign="top">84.13</td><td align="left" valign="top">96.05</td></tr><tr><td align="left" valign="top">Random Erasing<sup><xref ref-type="table-fn" rid="table3fn4">d</xref></sup></td><td align="left" valign="top">3</td><td align="left" valign="top">90.48</td><td align="left" valign="top">98.54</td><td align="left" valign="top">88.89</td><td align="left" valign="top">97.45</td><td align="left" valign="top">84.76</td><td align="left" valign="top">96.72<sup><xref ref-type="table-fn" rid="table3fn6">f</xref></sup></td><td align="left" valign="top">88.25</td><td align="left" valign="top">97.28</td></tr><tr><td align="left" valign="top">MixUp<sup><xref ref-type="table-fn" rid="table3fn4">d</xref></sup></td><td align="left" valign="top">6</td><td align="left" valign="top">91.43</td><td align="left" valign="top">98.57</td><td align="left" valign="top">83.49</td><td align="left" valign="top">96.85</td><td align="left" valign="top">67.82</td><td align="left" valign="top">86.97</td><td align="left" valign="top">90.79</td><td align="left" valign="top">97.87</td></tr><tr><td align="left" valign="top">CutMix</td><td align="left" valign="top">2</td><td align="left" valign="top">94.92</td><td align="left" valign="top">98.69</td><td align="left" valign="top">89.21</td><td align="left" valign="top">97.80</td><td align="left" valign="top">76.82</td><td align="left" valign="top">92.60</td><td align="left" valign="top">93.65<sup><xref ref-type="table-fn" rid="table3fn6">f</xref></sup></td><td align="left" valign="top">98.74<sup><xref ref-type="table-fn" rid="table3fn6">f</xref></sup></td></tr><tr><td align="left" valign="top">Random Pixel Swap<sup><xref ref-type="table-fn" rid="table3fn5">e</xref></sup></td><td align="left" valign="top">1</td><td align="left" valign="top">95.56<sup><xref ref-type="table-fn" rid="table3fn6">f</xref></sup></td><td align="left" valign="top">99.15<sup><xref ref-type="table-fn" rid="table3fn6">f</xref></sup></td><td align="left" valign="top">92.70<sup><xref ref-type="table-fn" rid="table3fn6">f</xref></sup></td><td align="left" valign="top">98.02<sup><xref ref-type="table-fn" rid="table3fn6">f</xref></sup></td><td align="left" valign="top">86.67<sup><xref ref-type="table-fn" rid="table3fn6">f</xref></sup></td><td align="left" valign="top">96.32</td><td align="left" valign="top">91.43</td><td align="left" valign="top">98.45</td></tr></tbody></table><table-wrap-foot><fn id="table3fn1"><p><sup>a</sup>CT: computed tomography.</p></fn><fn id="table3fn2"><p><sup>b</sup>Rank represents the overall rating for each technique, with &#x201C;1&#x201D; indicating the best technique across all models.</p></fn><fn id="table3fn3"><p><sup>c</sup>AUROC: area under the receiver operating characteristic curve.</p></fn><fn id="table3fn4"><p><sup>d</sup>Significant difference between an augmentation technique and the Random Pixel Swap technique across all models.</p></fn><fn id="table3fn5"><p><sup>e</sup>Significant difference between training using an augmentation technique and the base model across all models.</p></fn><fn id="table3fn6"><p><sup>f</sup>Highest value in the column.</p></fn></table-wrap-foot></table-wrap><p>At 512&#x00D7;512 image resolution, the RPS technique demonstrated superior performance compared to all evaluated DA methods (Cutout: <italic>P</italic>=.13; Random Erasing: <italic>P</italic>=.27; MixUp: <italic>P</italic>=.13; CutMix: <italic>P</italic>=.31). For ResNet-34, RPS matched the accuracy of the top-performing alternative (CutMix) while achieving a 0.21% improvement in AUROC. Furthermore, RPS showed significant gains over the least effective technique (MixUp), with a 7.74% accuracy performance advantage. The MobileNetV3 architecture exhibited even more pronounced benefits, where RPS outperformed CutMix (the best alternative) by 2.23% and surpassed MixUp by 4.45% in accuracy. Across all techniques, the performance ranking was as follows: (1) RPS (best technique), (2) CutMix, (3) Cutout, (4) Random Erasing, and (5) MixUp. The detailed results are presented in <xref ref-type="table" rid="table4">Table 4</xref>.</p><table-wrap id="t4" position="float"><label>Table 4.</label><caption><p>Classification results of the chest CT<sup><xref ref-type="table-fn" rid="table4fn1">a</xref></sup> scan images dataset using preactivated deep learning models with various data augmentation techniques (512&#x00D7;512 image resolution).</p></caption><table id="table4" frame="hsides" rules="groups"><thead><tr><td align="left" valign="top">Data augmentation</td><td align="left" valign="top">Rank<sup><xref ref-type="table-fn" rid="table4fn2">b</xref></sup></td><td align="left" valign="top" colspan="2">ResNet-34</td><td align="left" valign="top" colspan="2">MobileNetV3 (small)</td></tr><tr><td align="left" valign="bottom"/><td align="left" valign="bottom"/><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC<sup><xref ref-type="table-fn" rid="table4fn3">c</xref></sup>, %</td><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC, %</td></tr></thead><tbody><tr><td align="left" valign="top">Base model</td><td align="left" valign="top">5</td><td align="left" valign="top">96.83</td><td align="left" valign="top">99.25</td><td align="left" valign="top">93.02</td><td align="left" valign="top">98.27</td></tr><tr><td align="left" valign="top">Cutout</td><td align="left" valign="top">3</td><td align="left" valign="top">96.51</td><td align="left" valign="top">99.35</td><td align="left" valign="top">94.60</td><td align="left" valign="top">98.39</td></tr><tr><td align="left" valign="top">Random Erasing</td><td align="left" valign="top">4</td><td align="left" valign="top">96.83</td><td align="left" valign="top">99.42</td><td align="left" valign="top">93.65</td><td align="left" valign="top">98.82<sup><xref ref-type="table-fn" rid="table4fn4">d</xref></sup></td></tr><tr><td align="left" valign="top">MixUp</td><td align="left" valign="top">6</td><td align="left" valign="top">92.38</td><td align="left" valign="top">98.64</td><td align="left" valign="top">92.38</td><td align="left" valign="top">98.51</td></tr><tr><td align="left" valign="top">CutMix</td><td align="left" valign="top">2</td><td align="left" valign="top">97.78<sup><xref ref-type="table-fn" rid="table4fn4">d</xref></sup></td><td align="left" valign="top">99.25</td><td align="left" valign="top">94.60</td><td align="left" valign="top">98.61</td></tr><tr><td align="left" valign="top">Random Pixel Swap</td><td align="left" valign="top">1</td><td align="left" valign="top">97.78<sup><xref ref-type="table-fn" rid="table4fn4">d</xref></sup></td><td align="left" valign="top">99.46<sup><xref ref-type="table-fn" rid="table4fn4">d</xref></sup></td><td align="left" valign="top">96.83<sup><xref ref-type="table-fn" rid="table4fn4">d</xref></sup></td><td align="left" valign="top">98.75</td></tr></tbody></table><table-wrap-foot><fn id="table4fn1"><p><sup>a</sup>CT: computed tomography.</p></fn><fn id="table4fn2"><p><sup>b</sup>Rank represents the overall rating for each technique, with &#x201C;1&#x201D; indicating the best technique across all models.</p></fn><fn id="table4fn3"><p><sup>c</sup>AUROC: area under the receiver operating characteristic curve.</p></fn><fn id="table4fn4"><p><sup>d</sup>Highest value in the column.</p></fn></table-wrap-foot></table-wrap></sec><sec id="s3-4"><title>Performance Analysis of Swap Area Factors for Lung Cancer Diagnosis</title><p>The swap area factor serves as a critical hyperparameter in RPS implementation. We systematically evaluated its influence using both SVSF and MVSF configurations across the 0.1 to 1.0 range on the IQ-OTH/NCCD dataset. MVSF provides over 100 possible combinations of lower and upper bounds (eg, 0.1&#x2010;0.5 and 0.4&#x2010;0.8); however, our experimental configurations maintained a fixed lower bound of 0.1. Experimental results revealed distinct optimal configurations for each architecture. For SVSF implementations, ResNet-34, Vision Transformer, and Swin Transformer achieved peak performance at 1.0, while MobileNetV3 performed best at 0.9. For MVSF implementations, ResNet-34 showed optimal results within 0.1&#x2010;0.9, MobileNetV3 performed best at 0.1&#x2010;0.7, Vision Transformer excelled at 0.1&#x2010;0.3, and Swin Transformer achieved peak performance at 0.1&#x2010;0.5.</p><p>Comparative analysis demonstrated that SVSF generally outperformed MVSF configurations for a fixed 0.1 lower bound across most architectures, with the notable exception of ResNet-34. For this model, MVSF (0.1&#x2010;0.9) surpassed SVSF (1.0) by 0.61% in accuracy and 1.08% in AUROC. The most effective overall configuration combined MobileNetV3 with RPS using an SVSF of 0.9, achieving 94.51% accuracy and 95.77% AUROC. The detailed results are presented in <xref ref-type="table" rid="table5">Table 5</xref>.</p><table-wrap id="t5" position="float"><label>Table 5.</label><caption><p>Analysis of the IQ-OTH/NCCD<sup><xref ref-type="table-fn" rid="table5fn1">a</xref></sup> dataset using different deep learning architectures and Random Pixel Swap data augmentation with single-value and multivalue swap area factors (224&#x00D7;224 image resolution).</p></caption><table id="table5" frame="hsides" rules="groups"><thead><tr><td align="left" valign="top">Swap factor</td><td align="left" valign="top" colspan="2">ResNet-34</td><td align="left" valign="top" colspan="2">MobileNetV3 (small)</td><td align="left" valign="top" colspan="2">Vision Transformer (base-16)</td><td align="left" valign="top" colspan="2">Swin Transformer (tiny)</td></tr><tr><td align="left" valign="bottom"/><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC<sup><xref ref-type="table-fn" rid="table5fn2">b</xref></sup>, %</td><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC, %</td><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC, %</td><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC, %</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="9">Single value</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.1</td><td align="left" valign="top">89.02</td><td align="left" valign="top">92.26</td><td align="left" valign="top">93.90</td><td align="left" valign="top">95.10</td><td align="left" valign="top">64.02</td><td align="left" valign="top">74.97<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup></td><td align="left" valign="top">86.28</td><td align="left" valign="top">92.02</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.2</td><td align="left" valign="top">91.16</td><td align="left" valign="top">94.62</td><td align="left" valign="top">92.99</td><td align="left" valign="top">95.04</td><td align="left" valign="top">60.98</td><td align="left" valign="top">63.20</td><td align="left" valign="top">83.23</td><td align="left" valign="top">93.59</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.3</td><td align="left" valign="top">90.85</td><td align="left" valign="top">93.48</td><td align="left" valign="top">92.07</td><td align="left" valign="top">95.38</td><td align="left" valign="top">64.02</td><td align="left" valign="top">67.95</td><td align="left" valign="top">89.02</td><td align="left" valign="top">91.63</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.4</td><td align="left" valign="top">89.63</td><td align="left" valign="top">92.66</td><td align="left" valign="top">92.99</td><td align="left" valign="top">95.24</td><td align="left" valign="top">69.21</td><td align="left" valign="top">72.59</td><td align="left" valign="top">84.76</td><td align="left" valign="top">92.81</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.5</td><td align="left" valign="top">90.55</td><td align="left" valign="top">90.96</td><td align="left" valign="top">92.68</td><td align="left" valign="top">95.01</td><td align="left" valign="top">68.60</td><td align="left" valign="top">72.47</td><td align="left" valign="top">87.80</td><td align="left" valign="top">84.13</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.6</td><td align="left" valign="top">90.55</td><td align="left" valign="top">94.76</td><td align="left" valign="top">92.99</td><td align="left" valign="top">95.24</td><td align="left" valign="top">69.21</td><td align="left" valign="top">72.36</td><td align="left" valign="top">84.76</td><td align="left" valign="top">89.63</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.7</td><td align="left" valign="top">91.46</td><td align="left" valign="top">95.23</td><td align="left" valign="top">92.68</td><td align="left" valign="top">95.12</td><td align="left" valign="top">67.99</td><td align="left" valign="top">66.48</td><td align="left" valign="top">83.54</td><td align="left" valign="top">90.65</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.8</td><td align="left" valign="top">89.63</td><td align="left" valign="top">92.22</td><td align="left" valign="top">93.60</td><td align="left" valign="top">95.60</td><td align="left" valign="top">67.99</td><td align="left" valign="top">69.83</td><td align="left" valign="top">89.94<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup></td><td align="left" valign="top">95.95<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup></td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.9</td><td align="left" valign="top">90.85</td><td align="left" valign="top">94.42</td><td align="left" valign="top">94.51<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup></td><td align="left" valign="top">95.77<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup></td><td align="left" valign="top">71.65</td><td align="left" valign="top">72.79</td><td align="left" valign="top">83.84</td><td align="left" valign="top">90.99</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>1.0</td><td align="left" valign="top">92.07<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup></td><td align="left" valign="top">95.02<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup></td><td align="left" valign="top">94.21</td><td align="left" valign="top">95.67</td><td align="left" valign="top">73.17<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup></td><td align="left" valign="top">74.64</td><td align="left" valign="top">89.94<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup></td><td align="left" valign="top">94.79</td></tr><tr><td align="left" valign="top" colspan="9">Multivalue</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.1&#x2010;0.2</td><td align="left" valign="top">90.55</td><td align="left" valign="top">93.80</td><td align="left" valign="top">93.90<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup></td><td align="left" valign="top">94.83</td><td align="left" valign="top">66.16</td><td align="left" valign="top">69.14</td><td align="left" valign="top">85.98</td><td align="left" valign="top">94.29</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.1&#x2010;0.3</td><td align="left" valign="top">89.33</td><td align="left" valign="top">90.53</td><td align="left" valign="top">93.29</td><td align="left" valign="top">95.18</td><td align="left" valign="top">72.56<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup></td><td align="left" valign="top">77.93<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup></td><td align="left" valign="top">84.76</td><td align="left" valign="top">91.84</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.1&#x2010;0.4</td><td align="left" valign="top">89.33</td><td align="left" valign="top">90.18</td><td align="left" valign="top">93.60</td><td align="left" valign="top">95.03</td><td align="left" valign="top">60.98</td><td align="left" valign="top">73.15</td><td align="left" valign="top">87.50</td><td align="left" valign="top">94.30</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.1&#x2010;0.5</td><td align="left" valign="top">91.16</td><td align="left" valign="top">95.93<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup></td><td align="left" valign="top">93.60</td><td align="left" valign="top">94.84</td><td align="left" valign="top">59.15</td><td align="left" valign="top">68.80</td><td align="left" valign="top">88.11<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup></td><td align="left" valign="top">94.94<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup></td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.1&#x2010;0.6</td><td align="left" valign="top">90.55</td><td align="left" valign="top">93.26</td><td align="left" valign="top">92.38</td><td align="left" valign="top">94.60</td><td align="left" valign="top">62.20</td><td align="left" valign="top">59.86</td><td align="left" valign="top">87.20</td><td align="left" valign="top">93.45</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.1&#x2010;0.7</td><td align="left" valign="top">89.94</td><td align="left" valign="top">90.99</td><td align="left" valign="top">93.90<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup></td><td align="left" valign="top">95.33</td><td align="left" valign="top">61.28</td><td align="left" valign="top">70.73</td><td align="left" valign="top">88.11<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup></td><td align="left" valign="top">92.53</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.1&#x2010;0.8</td><td align="left" valign="top">87.80</td><td align="left" valign="top">93.13</td><td align="left" valign="top">93.29</td><td align="left" valign="top">95.00</td><td align="left" valign="top">66.77</td><td align="left" valign="top">76.81</td><td align="left" valign="top">86.28</td><td align="left" valign="top">85.93</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.1&#x2010;0.9</td><td align="left" valign="top">92.68<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup></td><td align="left" valign="top">95.29</td><td align="left" valign="top">93.60</td><td align="left" valign="top">95.29</td><td align="left" valign="top">68.29</td><td align="left" valign="top">64.98</td><td align="left" valign="top">86.59</td><td align="left" valign="top">92.69</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.1&#x2010;1.0</td><td align="left" valign="top">89.02</td><td align="left" valign="top">92.53</td><td align="left" valign="top">93.60</td><td align="left" valign="top">95.71<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup></td><td align="left" valign="top">62.80</td><td align="left" valign="top">69.05</td><td align="left" valign="top">83.54</td><td align="left" valign="top">94.35</td></tr></tbody></table><table-wrap-foot><fn id="table5fn1"><p><sup>a</sup>IQ-OTH/NCCD: Iraq-Oncology Teaching Hospital/National Center for Cancer Diseases.</p></fn><fn id="table5fn2"><p><sup>b</sup>AUROC: area under the receiver operating characteristic curve.</p></fn><fn id="table5fn3"><p><sup>c</sup>Highest value in the column.</p></fn></table-wrap-foot></table-wrap><p>Our evaluation of the chest CT scan images dataset using different swap area factor configurations revealed architecture-specific optimal settings. SVSF demonstrated superior performance at 1.0 for both ResNet-34 and MobileNetV3, while Vision Transformer achieved peak accuracy with an SVSF of 0.1. For Swin Transformer, MVSF configurations between 0.1 and 0.6 yielded optimal results. Among all tested combinations, ResNet-34 paired with RPS using an SVSF of 1.0 delivered the highest classification performance, reaching 97.78% accuracy and 99.46% AUROC. The detailed results are presented in <xref ref-type="table" rid="table6">Table 6</xref>.</p><table-wrap id="t6" position="float"><label>Table 6.</label><caption><p>Analysis of the chest CT<sup><xref ref-type="table-fn" rid="table6fn1">a</xref></sup> scan images dataset using different deep learning architectures and Random Pixel Swap data augmentation with single-value and multivalue swap area factors (224&#x00D7;224 image resolution).</p></caption><table id="table6" frame="hsides" rules="groups"><thead><tr><td align="left" valign="top">Swap factor</td><td align="left" valign="top" colspan="2">ResNet-34</td><td align="left" valign="top" colspan="2">MobileNetV3 (small)</td><td align="left" valign="top" colspan="2">Vision Transformer (base-16)</td><td align="left" valign="top" colspan="2">Swin Transformer (tiny)</td></tr><tr><td align="left" valign="bottom"/><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC<sup><xref ref-type="table-fn" rid="table6fn2">b</xref></sup>, %</td><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC, %</td><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC, %</td><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC, %</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="9">Single value</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.1</td><td align="left" valign="top">96.19</td><td align="left" valign="top">99.13</td><td align="left" valign="top">94.60</td><td align="left" valign="top">98.55</td><td align="left" valign="top">86.67<sup><xref ref-type="table-fn" rid="table6fn3">c</xref></sup></td><td align="left" valign="top">96.32</td><td align="left" valign="top">94.29<sup><xref ref-type="table-fn" rid="table6fn3">c</xref></sup></td><td align="left" valign="top">98.72</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.2</td><td align="left" valign="top">97.46</td><td align="left" valign="top">99.27</td><td align="left" valign="top">94.60</td><td align="left" valign="top">98.60</td><td align="left" valign="top">81.27</td><td align="left" valign="top">94.37</td><td align="left" valign="top">92.06</td><td align="left" valign="top">98.78<sup><xref ref-type="table-fn" rid="table6fn3">c</xref></sup></td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.3</td><td align="left" valign="top">96.19</td><td align="left" valign="top">99.22</td><td align="left" valign="top">95.24</td><td align="left" valign="top">98.61</td><td align="left" valign="top">78.73</td><td align="left" valign="top">93.98</td><td align="left" valign="top">93.65</td><td align="left" valign="top">98.65</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.4</td><td align="left" valign="top">97.46</td><td align="left" valign="top">99.20</td><td align="left" valign="top">95.24</td><td align="left" valign="top">98.65</td><td align="left" valign="top">82.54</td><td align="left" valign="top">96.10</td><td align="left" valign="top">92.06</td><td align="left" valign="top">98.46</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.5</td><td align="left" valign="top">97.14</td><td align="left" valign="top">99.41</td><td align="left" valign="top">94.92</td><td align="left" valign="top">98.74</td><td align="left" valign="top">85.08</td><td align="left" valign="top">96.31</td><td align="left" valign="top">91.43</td><td align="left" valign="top">98.38</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.6</td><td align="left" valign="top">97.14</td><td align="left" valign="top">99.30</td><td align="left" valign="top">95.56</td><td align="left" valign="top">98.79<sup><xref ref-type="table-fn" rid="table6fn3">c</xref></sup></td><td align="left" valign="top">85.40</td><td align="left" valign="top">96.86<sup><xref ref-type="table-fn" rid="table6fn3">c</xref></sup></td><td align="left" valign="top">91.75</td><td align="left" valign="top">97.85</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.7</td><td align="left" valign="top">97.14</td><td align="left" valign="top">99.19</td><td align="left" valign="top">95.56</td><td align="left" valign="top">98.69</td><td align="left" valign="top">83.81</td><td align="left" valign="top">95.48</td><td align="left" valign="top">93.65</td><td align="left" valign="top">98.65</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.8</td><td align="left" valign="top">97.14</td><td align="left" valign="top">99.38</td><td align="left" valign="top">95.87</td><td align="left" valign="top">98.75</td><td align="left" valign="top">81.59</td><td align="left" valign="top">95.48</td><td align="left" valign="top">91.43</td><td align="left" valign="top">97.95</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.9</td><td align="left" valign="top">96.83</td><td align="left" valign="top">99.35</td><td align="left" valign="top">95.87</td><td align="left" valign="top">98.62</td><td align="left" valign="top">81.27</td><td align="left" valign="top">94.25</td><td align="left" valign="top">88.89</td><td align="left" valign="top">97.80</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>1.0</td><td align="left" valign="top">97.78<sup><xref ref-type="table-fn" rid="table6fn3">c</xref></sup></td><td align="left" valign="top">99.46<sup><xref ref-type="table-fn" rid="table6fn3">c</xref></sup></td><td align="left" valign="top">96.83<sup><xref ref-type="table-fn" rid="table6fn3">c</xref></sup></td><td align="left" valign="top">98.75</td><td align="left" valign="top">75.56</td><td align="left" valign="top">91.62</td><td align="left" valign="top">91.43</td><td align="left" valign="top">98.45</td></tr><tr><td align="left" valign="top" colspan="9">Multivalue</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.1&#x2010;0.2</td><td align="left" valign="top">97.14</td><td align="left" valign="top">99.27</td><td align="left" valign="top">94.92</td><td align="left" valign="top">98.59</td><td align="left" valign="top">75.56</td><td align="left" valign="top">92.85</td><td align="left" valign="top">93.65</td><td align="left" valign="top">98.51</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.1&#x2010;0.3</td><td align="left" valign="top">96.51</td><td align="left" valign="top">99.30</td><td align="left" valign="top">93.97</td><td align="left" valign="top">98.62</td><td align="left" valign="top">84.13</td><td align="left" valign="top">95.86</td><td align="left" valign="top">91.43</td><td align="left" valign="top">98.20</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.1&#x2010;0.4</td><td align="left" valign="top">96.51</td><td align="left" valign="top">99.00</td><td align="left" valign="top">94.60</td><td align="left" valign="top">98.55</td><td align="left" valign="top">76.83</td><td align="left" valign="top">92.77</td><td align="left" valign="top">93.65</td><td align="left" valign="top">98.73</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.1&#x2010;0.5</td><td align="left" valign="top">97.46</td><td align="left" valign="top">99.28</td><td align="left" valign="top">94.29</td><td align="left" valign="top">98.63</td><td align="left" valign="top">80.32</td><td align="left" valign="top">94.69</td><td align="left" valign="top">92.38</td><td align="left" valign="top">98.64</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.1&#x2010;0.6</td><td align="left" valign="top">96.51</td><td align="left" valign="top">99.37</td><td align="left" valign="top">95.24</td><td align="left" valign="top">98.59</td><td align="left" valign="top">82.54</td><td align="left" valign="top">95.63</td><td align="left" valign="top">96.19<sup><xref ref-type="table-fn" rid="table6fn3">c</xref></sup></td><td align="left" valign="top">98.90<sup><xref ref-type="table-fn" rid="table6fn3">c</xref></sup></td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.1&#x2010;0.7</td><td align="left" valign="top">97.78<sup><xref ref-type="table-fn" rid="table6fn3">c</xref></sup></td><td align="left" valign="top">99.39</td><td align="left" valign="top">94.92</td><td align="left" valign="top">98.65</td><td align="left" valign="top">86.03<sup><xref ref-type="table-fn" rid="table6fn3">c</xref></sup></td><td align="left" valign="top">96.84<sup><xref ref-type="table-fn" rid="table6fn3">c</xref></sup></td><td align="left" valign="top">94.29</td><td align="left" valign="top">98.90<sup><xref ref-type="table-fn" rid="table6fn3">c</xref></sup></td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.1&#x2010;0.8</td><td align="left" valign="top">97.46</td><td align="left" valign="top">99.25</td><td align="left" valign="top">93.97</td><td align="left" valign="top">98.62</td><td align="left" valign="top">81.90</td><td align="left" valign="top">94.88</td><td align="left" valign="top">93.02</td><td align="left" valign="top">98.83</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.1&#x2010;0.9</td><td align="left" valign="top">97.48</td><td align="left" valign="top">99.32</td><td align="left" valign="top">94.92</td><td align="left" valign="top">98.70</td><td align="left" valign="top">81.90</td><td align="left" valign="top">94.99</td><td align="left" valign="top">93.65</td><td align="left" valign="top">98.86</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0.1&#x2010;1.0</td><td align="left" valign="top">97.46</td><td align="left" valign="top">99.41<sup><xref ref-type="table-fn" rid="table6fn3">c</xref></sup></td><td align="left" valign="top">95.87<sup><xref ref-type="table-fn" rid="table6fn3">c</xref></sup></td><td align="left" valign="top">98.75<sup><xref ref-type="table-fn" rid="table6fn3">c</xref></sup></td><td align="left" valign="top">82.22</td><td align="left" valign="top">95.21</td><td align="left" valign="top">93.33</td><td align="left" valign="top">98.72</td></tr></tbody></table><table-wrap-foot><fn id="table6fn1"><p><sup>a</sup>CT: computed tomography.</p></fn><fn id="table6fn2"><p><sup>b</sup>AUROC: area under the receiver operating characteristic curve.</p></fn><fn id="table6fn3"><p><sup>c</sup>Highest value in the column.</p></fn></table-wrap-foot></table-wrap></sec><sec id="s3-5"><title>RPS With Lung Region of Interest Segmentation</title><p>Prior studies have demonstrated that segmenting lung regions of interest (ROIs) can significantly improve the diagnostic performance of DL models [<xref ref-type="bibr" rid="ref33">33</xref>,<xref ref-type="bibr" rid="ref56">56</xref>]. To evaluate the effectiveness of the RPS technique when applied to segmented images, we conducted experiments using the selected models (ResNet-34, MobileNetV3 [small], Vision Transformer [base-16], and Swin Transformer [tiny]). Our investigation used both the IQ-OTH/NCCD dataset and chest CT scan images dataset at 224&#x00D7;224 resolution.</p><p>The segmentation process involved multiple steps. We first applied a threshold algorithm to generate a lung mask, followed by dilation and hole-filling operations to ensure comprehensive coverage of pulmonary structures. The final lung ROI was extracted by cropping surrounding pixels along the mask boundaries. The complete procedure is illustrated in <xref ref-type="fig" rid="figure4">Figure 4</xref>. For comparative analysis, we evaluated model performance under three conditions: (1) training without augmentation, (2) training with RPS, and (3) training with established augmentation techniques (Cutout, Random Erasing, MixUp, and CutMix). This comprehensive evaluation framework allowed us to assess the relative benefits of RPS when applied to segmented lung images.</p><fig position="float" id="figure4"><label>Figure 4.</label><caption><p>Lung segmentation procedure.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="bioinform_v6i1e68848_fig04.png"/></fig><p>Our experiments with the IQ-OTH/NCCD dataset demonstrated that the RPS technique significantly improved performance across all evaluated models (<italic>P</italic>=.04) and most techniques (Cutout: <italic>P</italic>=.049; Random Erasing: <italic>P</italic>=.004; MixUp: <italic>P</italic>=.04; CutMix: <italic>P</italic>=.06). The most notable results were achieved by ResNet-34 with RPS, reaching 97.56% accuracy and 98.61% AUROC. While RPS outperformed all competing techniques for MobileNetV3 and Swin Transformer, CutMix showed superior performance for Vision Transformer, exceeding RPS by 1.52% in accuracy and 0.67% in AUROC. The overall performance ranking across techniques was as follows: (1) RPS (best technique), (2) CutMix, (3) Random Erasing, (4) Cutout, and (5) MixUp.</p><p>For the chest CT scan images dataset, the RPS technique consistently improved performance across models (<italic>P</italic>=.06) and most techniques (Cutout: <italic>P</italic>=.01; Random Erasing: <italic>P</italic>=.009; MixUp: <italic>P</italic>=.01; CutMix: <italic>P</italic>=.38). The highest performance was again achieved by ResNet-34 with RPS (95.51% accuracy and 98.86% AUROC). While RPS showed superior results for MobileNetV3 and Swin Transformer, CutMix performed better for Vision Transformer (3.21% higher accuracy and 0.6% higher AUROC). The comprehensive performance ranking was similar to that for the IQ-OTH/NCCD dataset and was as follows: (1) RPS, (2) CutMix, (3) Cutout, (4) Random Erasing, and (5) MixUp. The detailed results are presented in <xref ref-type="table" rid="table7">Table 7</xref>.</p><table-wrap id="t7" position="float"><label>Table 7.</label><caption><p>Classification results of the IQ-OTH/NCCD<sup><xref ref-type="table-fn" rid="table7fn1">a</xref></sup> and chest CT<sup><xref ref-type="table-fn" rid="table7fn2">b</xref></sup> scan images datasets using preactivated deep learning models with various data augmentation techniques and segmentation of the lung region of interest (224&#x00D7;224 image resolution).</p></caption><table id="table7" frame="hsides" rules="groups"><thead><tr><td align="left" valign="top">Data augmentation</td><td align="left" valign="top">Rank<sup><xref ref-type="table-fn" rid="table7fn3">c</xref></sup></td><td align="left" valign="top" colspan="2">ResNet-34</td><td align="left" valign="top" colspan="2">MobileNetV3 (small)</td><td align="left" valign="top" colspan="2">Vision Transformer (base-16)</td><td align="left" valign="top" colspan="2">Swin Transformer (tiny)</td></tr><tr><td align="left" valign="bottom"/><td align="left" valign="bottom"/><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC<sup><xref ref-type="table-fn" rid="table7fn4">d</xref></sup>, %</td><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC, %</td><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC, %</td><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC, %</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="10">IQ-OTH/NCCD dataset</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Base model<sup><xref ref-type="table-fn" rid="table7fn5">e</xref></sup></td><td align="left" valign="top">5</td><td align="left" valign="top">96.65</td><td align="left" valign="top">99.13</td><td align="left" valign="top">95.43</td><td align="left" valign="top">97.28</td><td align="left" valign="top">89.94</td><td align="left" valign="top">96.51</td><td align="left" valign="top">93.60</td><td align="left" valign="top">98.21</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Cutout<sup><xref ref-type="table-fn" rid="table7fn5">e</xref></sup></td><td align="left" valign="top">4</td><td align="left" valign="top">96.04</td><td align="left" valign="top">98.86</td><td align="left" valign="top">95.43</td><td align="left" valign="top">96.33</td><td align="left" valign="top">92.38</td><td align="left" valign="top">96.20</td><td align="left" valign="top">93.90</td><td align="left" valign="top">97.80</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Random Erasing<sup><xref ref-type="table-fn" rid="table7fn5">e</xref></sup></td><td align="left" valign="top">3</td><td align="left" valign="top">95.73</td><td align="left" valign="top">97.45</td><td align="left" valign="top">96.65<sup><xref ref-type="table-fn" rid="table7fn6">f</xref></sup></td><td align="left" valign="top">97.29</td><td align="left" valign="top">91.46</td><td align="left" valign="top">96.27</td><td align="left" valign="top">94.82<sup><xref ref-type="table-fn" rid="table7fn6">f</xref></sup></td><td align="left" valign="top">98.00</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>MixUp<sup><xref ref-type="table-fn" rid="table7fn5">e</xref></sup><sup>,</sup><sup><xref ref-type="table-fn" rid="table7fn7">g</xref></sup></td><td align="left" valign="top">6</td><td align="left" valign="top">95.87</td><td align="left" valign="top">99.19<sup><xref ref-type="table-fn" rid="table7fn6">f</xref></sup></td><td align="left" valign="top">91.77</td><td align="left" valign="top">97.11</td><td align="left" valign="top">91.77</td><td align="left" valign="top">96.27</td><td align="left" valign="top">93.29</td><td align="left" valign="top">97.52</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>CutMix</td><td align="left" valign="top">2</td><td align="left" valign="top">96.65</td><td align="left" valign="top">98.86</td><td align="left" valign="top">94.51</td><td align="left" valign="top">96.39</td><td align="left" valign="top">93.90<sup><xref ref-type="table-fn" rid="table7fn6">f</xref></sup></td><td align="left" valign="top">97.64<sup><xref ref-type="table-fn" rid="table7fn6">f</xref></sup></td><td align="left" valign="top">93.29</td><td align="left" valign="top">97.52</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Random Pixel Swap<sup><xref ref-type="table-fn" rid="table7fn7">g</xref></sup></td><td align="left" valign="top">1</td><td align="left" valign="top">97.56<sup><xref ref-type="table-fn" rid="table7fn6">f</xref></sup></td><td align="left" valign="top">98.61</td><td align="left" valign="top">96.65<sup><xref ref-type="table-fn" rid="table7fn6">f</xref></sup></td><td align="left" valign="top">98.00<sup><xref ref-type="table-fn" rid="table7fn6">f</xref></sup></td><td align="left" valign="top">92.38</td><td align="left" valign="top">96.97</td><td align="left" valign="top">94.82<sup><xref ref-type="table-fn" rid="table7fn6">f</xref></sup></td><td align="left" valign="top">98.12<sup><xref ref-type="table-fn" rid="table7fn6">f</xref></sup></td></tr><tr><td align="left" valign="top" colspan="10">Chest CT scan images dataset</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Base model</td><td align="left" valign="top">2</td><td align="left" valign="top">95.19</td><td align="left" valign="top">99.03</td><td align="left" valign="top">87.82</td><td align="left" valign="top">96.83<sup><xref ref-type="table-fn" rid="table7fn6">f</xref></sup></td><td align="left" valign="top">82.69</td><td align="left" valign="top">95.48</td><td align="left" valign="top">90.71</td><td align="left" valign="top">98.11</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Cutout<sup><xref ref-type="table-fn" rid="table7fn5">e</xref></sup></td><td align="left" valign="top">4</td><td align="left" valign="top">94.55</td><td align="left" valign="top">98.85</td><td align="left" valign="top">88.14</td><td align="left" valign="top">97.66</td><td align="left" valign="top">80.77</td><td align="left" valign="top">93.86</td><td align="left" valign="top">88.14</td><td align="left" valign="top">97.32</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Random Erasing<sup><xref ref-type="table-fn" rid="table7fn5">e</xref></sup><sup>,</sup><sup><xref ref-type="table-fn" rid="table7fn7">g</xref></sup></td><td align="left" valign="top">5</td><td align="left" valign="top">94.55</td><td align="left" valign="top">98.75</td><td align="left" valign="top">86.54</td><td align="left" valign="top">96.52</td><td align="left" valign="top">79.81</td><td align="left" valign="top">89.72</td><td align="left" valign="top">86.86</td><td align="left" valign="top">97.16</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>MixUp<sup><xref ref-type="table-fn" rid="table7fn5">e</xref></sup><sup>,</sup><sup><xref ref-type="table-fn" rid="table7fn7">g</xref></sup></td><td align="left" valign="top">6</td><td align="left" valign="top">94.55</td><td align="left" valign="top">98.77</td><td align="left" valign="top">82.05</td><td align="left" valign="top">95.33</td><td align="left" valign="top">78.85</td><td align="left" valign="top">93.29</td><td align="left" valign="top">85.90</td><td align="left" valign="top">97.10</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>CutMix</td><td align="left" valign="top">3</td><td align="left" valign="top">95.19</td><td align="left" valign="top">99.05<sup><xref ref-type="table-fn" rid="table7fn6">f</xref></sup></td><td align="left" valign="top">86.54</td><td align="left" valign="top">96.89</td><td align="left" valign="top">86.86<sup><xref ref-type="table-fn" rid="table7fn6">f</xref></sup></td><td align="left" valign="top">96.43<sup><xref ref-type="table-fn" rid="table7fn6">f</xref></sup></td><td align="left" valign="top">87.82</td><td align="left" valign="top">96.73</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Random Pixel Swap</td><td align="left" valign="top">1</td><td align="left" valign="top">95.51<sup><xref ref-type="table-fn" rid="table7fn6">f</xref></sup></td><td align="left" valign="top">98.86</td><td align="left" valign="top">90.71<sup><xref ref-type="table-fn" rid="table7fn6">f</xref></sup></td><td align="left" valign="top">97.51</td><td align="left" valign="top">83.65</td><td align="left" valign="top">95.83</td><td align="left" valign="top">91.35<sup><xref ref-type="table-fn" rid="table7fn6">f</xref></sup></td><td align="left" valign="top">98.36<sup><xref ref-type="table-fn" rid="table7fn6">f</xref></sup></td></tr></tbody></table><table-wrap-foot><fn id="table7fn1"><p><sup>a</sup>IQ-OTH/NCCD: Iraq-Oncology Teaching Hospital/National Center for Cancer Diseases.</p></fn><fn id="table7fn2"><p><sup>b</sup>CT: computed tomography.</p></fn><fn id="table7fn3"><p><sup>c</sup>Rank represents the overall rating for each technique, with &#x201C;1&#x201D; indicating the best technique across all models.</p></fn><fn id="table7fn4"><p><sup>d</sup>AUROC: area under the receiver operating characteristic curve.</p></fn><fn id="table7fn5"><p><sup>e</sup>Significant difference between an augmentation technique and the Random Pixel Swap technique across all models.</p></fn><fn id="table7fn6"><p><sup>f</sup>Highest value in the column.</p></fn><fn id="table7fn7"><p><sup>g</sup>Significant difference between training using an augmentation technique and the base model across all models.</p></fn></table-wrap-foot></table-wrap></sec><sec id="s3-6"><title>Performance Analysis of the Combination of RPS With Traditional DA Techniques for Lung Cancer Diagnosis</title><p>Traditionally, DA techniques, including image flipping and rotation, are widely employed in medical image analysis with DL [<xref ref-type="bibr" rid="ref44">44</xref>]. To evaluate the potential benefits of combining these methods with the RPS technique, we conducted a systematic comparison. First, we trained selected models (ResNet-34, MobileNetV3 [small], Vision Transformer [base-16], and Swin Transformer [tiny]) using individual traditional techniques: horizontal flipping, vertical flipping, and random rotation (&#x00B1;90&#x00B0;). Subsequently, we trained the models using combinations of each traditional technique with RPS.</p><p>Our experiments revealed that the combination of RPS with traditional techniques generally enhanced model performance compared to using traditional methods alone. However, when a traditional technique failed to improve baseline performance, its combination with RPS did not surpass RPS alone. For the IQ-OTH/NCCD dataset, using RPS alone surpassed the individual traditional techniques (horizontal flipping: <italic>P</italic>=.63; vertical flipping: <italic>P</italic>=.22; rotation: <italic>P</italic>=.93). RPS with rotation achieved peak performance for ResNet-34 and Vision Transformer (base-16), improving upon rotation alone by 2.14% and 2.75% in accuracy, respectively. RPS with vertical flipping performed the best for MobileNetV3 (small), exceeding vertical flipping alone by 0.61% in accuracy. However, RPS alone showed superior results for Swin Transformer (tiny).</p><p>Similarly, for the chest CT scan images dataset, using RPS alone surpassed the individual traditional techniques (horizontal flipping: <italic>P</italic>=.01; vertical flipping: <italic>P</italic>=.03; rotation: <italic>P</italic>=.04). RPS with rotation demonstrated the strongest overall performance, improving upon rotation by 0.95% in accuracy. RPS with horizontal flipping achieved optimal results for Vision Transformer (base-16), surpassing horizontal flipping alone by 5.71% in accuracy. However, RPS alone outperformed all combinations for MobileNetV3 (small) and Swin Transformer (tiny). The detailed results are presented in <xref ref-type="table" rid="table8">Table 8</xref>.</p><table-wrap id="t8" position="float"><label>Table 8.</label><caption><p>Classification results of the IQ-OTH/NCCD<sup><xref ref-type="table-fn" rid="table8fn1">a</xref></sup> and chest CT<sup><xref ref-type="table-fn" rid="table8fn2">b</xref></sup> scan images datasets using preactivated deep learning models when 3 traditional data augmentation techniques are combined with the Random Pixel Swap data augmentation technique (224&#x00D7;224 image resolution).</p></caption><table id="table8" frame="hsides" rules="groups"><thead><tr><td align="left" valign="top">Data augmentation</td><td align="left" valign="top">Rank<sup><xref ref-type="table-fn" rid="table8fn3">c</xref></sup></td><td align="left" valign="top" colspan="2">ResNet-34</td><td align="left" valign="top" colspan="2">MobileNetV3 (small)</td><td align="left" valign="top" colspan="2">Vision Transformer (base-16)</td><td align="left" valign="top" colspan="2">Swin Transformer (tiny)</td></tr><tr><td align="left" valign="bottom"/><td align="left" valign="bottom"/><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC<sup><xref ref-type="table-fn" rid="table8fn4">d</xref></sup>, %</td><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC, %</td><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC, %</td><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC, %</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="10">IQ-OTH/NCCD dataset</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Base model<sup><xref ref-type="table-fn" rid="table8fn5">e</xref></sup></td><td align="left" valign="top">8</td><td align="left" valign="top">85.98</td><td align="left" valign="top">83.39</td><td align="left" valign="top">86.59</td><td align="left" valign="top">93.16</td><td align="left" valign="top">57.62</td><td align="left" valign="top">64.88</td><td align="left" valign="top">85.67</td><td align="left" valign="top">89.06</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Horizontal flip</td><td align="left" valign="top">5</td><td align="left" valign="top">82.01</td><td align="left" valign="top">86.70</td><td align="left" valign="top">88.11</td><td align="left" valign="top">93.36</td><td align="left" valign="top">73.78</td><td align="left" valign="top">86.20</td><td align="left" valign="top">87.50</td><td align="left" valign="top">91.86</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Horizontal flip with Random Pixel Swap</td><td align="left" valign="top">2</td><td align="left" valign="top">87.20</td><td align="left" valign="top">90.43</td><td align="left" valign="top">88.41</td><td align="left" valign="top">91.21</td><td align="left" valign="top">78.36</td><td align="left" valign="top">89.31<sup><xref ref-type="table-fn" rid="table8fn6">f</xref></sup></td><td align="left" valign="top">88.72</td><td align="left" valign="top">92.12</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Vertical flip<sup><xref ref-type="table-fn" rid="table8fn7">g</xref></sup></td><td align="left" valign="top">7</td><td align="left" valign="top">87.50</td><td align="left" valign="top">89.82</td><td align="left" valign="top">90.55</td><td align="left" valign="top">93.89<sup><xref ref-type="table-fn" rid="table8fn6">f</xref></sup></td><td align="left" valign="top">62.50</td><td align="left" valign="top">76.54</td><td align="left" valign="top">89.02</td><td align="left" valign="top">92.41</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Vertical flip with Random Pixel Swap<sup><xref ref-type="table-fn" rid="table8fn5">e</xref></sup><sup>,</sup><sup><xref ref-type="table-fn" rid="table8fn7">g</xref></sup></td><td align="left" valign="top">6</td><td align="left" valign="top">88.11</td><td align="left" valign="top">89.54</td><td align="left" valign="top">91.16<sup><xref ref-type="table-fn" rid="table8fn6">f</xref></sup></td><td align="left" valign="top">93.28</td><td align="left" valign="top">68.29</td><td align="left" valign="top">73.94</td><td align="left" valign="top">88.72</td><td align="left" valign="top">93.10</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Rotation<sup><xref ref-type="table-fn" rid="table8fn7">g</xref></sup></td><td align="left" valign="top">4</td><td align="left" valign="top">87.80</td><td align="left" valign="top">90.30</td><td align="left" valign="top">89.63</td><td align="left" valign="top">91.53</td><td align="left" valign="top">75.91</td><td align="left" valign="top">80.70</td><td align="left" valign="top">88.41</td><td align="left" valign="top">92.67</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Rotation with Random Pixel Swap<sup><xref ref-type="table-fn" rid="table8fn7">g</xref></sup></td><td align="left" valign="top">1</td><td align="left" valign="top">89.94<sup><xref ref-type="table-fn" rid="table8fn6">f</xref></sup></td><td align="left" valign="top">90.28</td><td align="left" valign="top">89.02</td><td align="left" valign="top">91.75</td><td align="left" valign="top">78.66<sup><xref ref-type="table-fn" rid="table8fn6">f</xref></sup></td><td align="left" valign="top">87.16</td><td align="left" valign="top">89.02</td><td align="left" valign="top">93.10</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Random Pixel Swap<sup><xref ref-type="table-fn" rid="table8fn7">g</xref></sup></td><td align="left" valign="top">3</td><td align="left" valign="top">88.11</td><td align="left" valign="top">93.70<sup><xref ref-type="table-fn" rid="table8fn6">f</xref></sup></td><td align="left" valign="top">89.63</td><td align="left" valign="top">93.80</td><td align="left" valign="top">73.17</td><td align="left" valign="top">74.64</td><td align="left" valign="top">89.94<sup><xref ref-type="table-fn" rid="table8fn6">f</xref></sup></td><td align="left" valign="top">94.79<sup><xref ref-type="table-fn" rid="table8fn6">f</xref></sup></td></tr><tr><td align="left" valign="top" colspan="10">Chest CT scan images dataset</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Base model<sup><xref ref-type="table-fn" rid="table8fn5">e</xref></sup></td><td align="left" valign="top">8</td><td align="left" valign="top">93.33</td><td align="left" valign="top">99.00</td><td align="left" valign="top">87.30</td><td align="left" valign="top">97.09</td><td align="left" valign="top">82.86</td><td align="left" valign="top">95.84</td><td align="left" valign="top">84.76</td><td align="left" valign="top">96.92</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Horizontal flip<sup><xref ref-type="table-fn" rid="table8fn5">e</xref></sup></td><td align="left" valign="top">7</td><td align="left" valign="top">91.43</td><td align="left" valign="top">98.56</td><td align="left" valign="top">87.62</td><td align="left" valign="top">97.37</td><td align="left" valign="top">82.86</td><td align="left" valign="top">95.83</td><td align="left" valign="top">91.75</td><td align="left" valign="top">98.33</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Horizontal flip with Random Pixel Swap<sup><xref ref-type="table-fn" rid="table8fn7">g</xref></sup></td><td align="left" valign="top">4</td><td align="left" valign="top">93.97</td><td align="left" valign="top">98.96</td><td align="left" valign="top">89.21</td><td align="left" valign="top">97.48</td><td align="left" valign="top">88.57<sup><xref ref-type="table-fn" rid="table8fn6">f</xref></sup></td><td align="left" valign="top">97.63<sup><xref ref-type="table-fn" rid="table8fn6">f</xref></sup></td><td align="left" valign="top">92.70</td><td align="left" valign="top">98.09</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Vertical flip<sup><xref ref-type="table-fn" rid="table8fn5">e</xref></sup><sup>,</sup><sup><xref ref-type="table-fn" rid="table8fn7">g</xref></sup></td><td align="left" valign="top">6</td><td align="left" valign="top">93.33</td><td align="left" valign="top">98.86</td><td align="left" valign="top">83.81</td><td align="left" valign="top">96.83</td><td align="left" valign="top">84.72</td><td align="left" valign="top">96.10</td><td align="left" valign="top">92.06</td><td align="left" valign="top">98.58</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Vertical flip with Random Pixel Swap<sup><xref ref-type="table-fn" rid="table8fn5">e</xref></sup><sup>,</sup><sup><xref ref-type="table-fn" rid="table8fn7">g</xref></sup><break/></td><td align="left" valign="top">5</td><td align="left" valign="top">93.97</td><td align="left" valign="top">98.81</td><td align="left" valign="top">86.67</td><td align="left" valign="top">97.13</td><td align="left" valign="top">84.76</td><td align="left" valign="top">96.23</td><td align="left" valign="top">91.43</td><td align="left" valign="top">98.04</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Rotation<sup><xref ref-type="table-fn" rid="table8fn5">e</xref></sup><sup>,</sup><sup><xref ref-type="table-fn" rid="table8fn7">g</xref></sup><break/></td><td align="left" valign="top">3</td><td align="left" valign="top">95.24</td><td align="left" valign="top">99.22</td><td align="left" valign="top">90.16</td><td align="left" valign="top">97.58</td><td align="left" valign="top">84.57</td><td align="left" valign="top">94.95</td><td align="left" valign="top">95.87</td><td align="left" valign="top">99.03</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Rotation with Random Pixel Swap<sup><xref ref-type="table-fn" rid="table8fn7">g</xref></sup><break/></td><td align="left" valign="top">1</td><td align="left" valign="top">96.19<sup><xref ref-type="table-fn" rid="table8fn6">f</xref></sup></td><td align="left" valign="top">99.24<sup><xref ref-type="table-fn" rid="table8fn6">f</xref></sup></td><td align="left" valign="top">91.75</td><td align="left" valign="top">97.73</td><td align="left" valign="top">85.23</td><td align="left" valign="top">95.10</td><td align="left" valign="top">96.19</td><td align="left" valign="top">98.99</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Random Pixel Swap<sup><xref ref-type="table-fn" rid="table8fn7">g</xref></sup><break/></td><td align="left" valign="top">2</td><td align="left" valign="top">95.56</td><td align="left" valign="top">99.15</td><td align="left" valign="top">92.70<sup><xref ref-type="table-fn" rid="table8fn6">f</xref></sup></td><td align="left" valign="top">98.02<sup><xref ref-type="table-fn" rid="table8fn6">f</xref></sup></td><td align="left" valign="top">86.67</td><td align="left" valign="top">96.32</td><td align="left" valign="top">96.19<sup><xref ref-type="table-fn" rid="table8fn6">f</xref></sup></td><td align="left" valign="top">98.90<sup><xref ref-type="table-fn" rid="table8fn6">f</xref></sup></td></tr></tbody></table><table-wrap-foot><fn id="table8fn1"><p><sup>a</sup>IQ-OTH/NCCD: Iraq-Oncology Teaching Hospital/National Center for Cancer Diseases.</p></fn><fn id="table8fn2"><p><sup>b</sup>CT: computed tomography.</p></fn><fn id="table8fn3"><p><sup>c</sup>Rank represents the overall rating for each technique, with &#x201C;1&#x201D; indicating the best technique across all models.</p></fn><fn id="table8fn4"><p><sup>d</sup>AUROC: area under the receiver operating characteristic curve.</p></fn><fn id="table8fn5"><p><sup>e</sup>Significant difference between an augmentation technique and the Random Pixel Swap technique across all models.</p></fn><fn id="table8fn6"><p><sup>f</sup>Highest value in the column.</p></fn><fn id="table8fn7"><p><sup>g</sup>Significant difference between training using an augmentation technique and the base model across all models.</p></fn></table-wrap-foot></table-wrap></sec><sec id="s3-7"><title>Validation Results of the Generalization Capabilities of the RPS Technique</title><p>Enhancing the generalization ability of DL models to unseen data represents a critical objective of DA [<xref ref-type="bibr" rid="ref46">46</xref>]. To evaluate the RPS technique&#x2019;s capacity to improve model generalization, we conducted experiments using the selected models (ResNet-18, MobileNetV3 [small], Vision Transformer [base-16], and Swin Transformer [tiny]). Models were trained on the IQ-OTH/NCCD dataset and validated on the chest CT scan images dataset (distinct collections acquired using different imaging equipment, protocols, time periods, and geographical locations). All models performed binary classification (cancerous vs normal) of CT images.</p><p>Our comparative analysis included the base models, RPS implementation, and selected standard DA techniques. The results demonstrated RPS&#x2019;s superior performance across all architectures (Cutout: <italic>P</italic>=.05; Random Erasing: <italic>P</italic>=.054; MixUp: <italic>P</italic>=.04; CutMix: <italic>P</italic>=.03), with an exception for the Vision Transformer implementation. Random Erasing showed a marginal 0.8% accuracy advantage over RPS. However, RPS maintained a significant 9.28% improvement in AUROC over Random Erasing. Furthermore, the cumulative ranking was as follows: (1) RPS (best technique), (2) Cutout, (3) CutMix, (4) Random Erasing, and (5) MixUp. The detailed results are presented in <xref ref-type="table" rid="table9">Table 9</xref>.</p><table-wrap id="t9" position="float"><label>Table 9.</label><caption><p>Validation results of the generalization capabilities of different data augmentation techniques for lung cancer diagnosis using deep learning (224&#x00D7;224 image resolution).</p></caption><table id="table9" frame="hsides" rules="groups"><thead><tr><td align="left" valign="top">Data augmentation</td><td align="left" valign="top">Rank<sup><xref ref-type="table-fn" rid="table9fn1">a</xref></sup></td><td align="left" valign="top" colspan="2">ResNet-34</td><td align="left" valign="top" colspan="2">MobileNetV3 (small)</td><td align="left" valign="top" colspan="2">Vision Transformer (base-16)</td><td align="left" valign="top" colspan="2">Swin Transformer (tiny)</td></tr><tr><td align="left" valign="bottom"/><td align="left" valign="bottom"/><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC<sup><xref ref-type="table-fn" rid="table9fn2">b</xref></sup>, %</td><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC, %</td><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC, %</td><td align="left" valign="bottom">Accuracy, %</td><td align="left" valign="bottom">AUROC, %</td></tr></thead><tbody><tr><td align="left" valign="top">Base model<sup><xref ref-type="table-fn" rid="table9fn3">c</xref></sup></td><td align="left" valign="top">3</td><td align="left" valign="top">82.53</td><td align="left" valign="top">84.33</td><td align="left" valign="top">91.24</td><td align="left" valign="top">90.03</td><td align="left" valign="top">79.29</td><td align="left" valign="top">63.60</td><td align="left" valign="top">92.22</td><td align="left" valign="top">95.22<sup><xref ref-type="table-fn" rid="table9fn4">d</xref></sup></td></tr><tr><td align="left" valign="top">Cutout<sup><xref ref-type="table-fn" rid="table9fn3">c</xref></sup></td><td align="left" valign="top">2</td><td align="left" valign="top">82.65</td><td align="left" valign="top">97.29</td><td align="left" valign="top">92.09</td><td align="left" valign="top">90.70</td><td align="left" valign="top">81.80</td><td align="left" valign="top">63.49</td><td align="left" valign="top">92.22</td><td align="left" valign="top">85.77</td></tr><tr><td align="left" valign="top">Random Erasing<sup><xref ref-type="table-fn" rid="table9fn5">e</xref></sup></td><td align="left" valign="top">5</td><td align="left" valign="top">88.71</td><td align="left" valign="top">78.66</td><td align="left" valign="top">91.45</td><td align="left" valign="top">89.66</td><td align="left" valign="top">82.65<sup><xref ref-type="table-fn" rid="table9fn4">d</xref></sup></td><td align="left" valign="top">58.92</td><td align="left" valign="top">91.96</td><td align="left" valign="top">85.96</td></tr><tr><td align="left" valign="top">MixUp<sup><xref ref-type="table-fn" rid="table9fn3">c</xref></sup></td><td align="left" valign="top">6</td><td align="left" valign="top">83.80</td><td align="left" valign="top">95.17</td><td align="left" valign="top">91.58</td><td align="left" valign="top">90.36</td><td align="left" valign="top">80.74</td><td align="left" valign="top">52.24</td><td align="left" valign="top">91.58</td><td align="left" valign="top">79.73</td></tr><tr><td align="left" valign="top">CutMix<sup><xref ref-type="table-fn" rid="table9fn3">c</xref></sup></td><td align="left" valign="top">4</td><td align="left" valign="top">84.57</td><td align="left" valign="top">94.36</td><td align="left" valign="top">90.69</td><td align="left" valign="top">80.26</td><td align="left" valign="top">81.12</td><td align="left" valign="top">66.90</td><td align="left" valign="top">92.09</td><td align="left" valign="top">86.09</td></tr><tr><td align="left" valign="top">Random Pixel Swap<sup><xref ref-type="table-fn" rid="table9fn5">e</xref></sup></td><td align="left" valign="top">1</td><td align="left" valign="top">90.69<sup><xref ref-type="table-fn" rid="table9fn4">d</xref></sup></td><td align="left" valign="top">97.48<sup><xref ref-type="table-fn" rid="table9fn4">d</xref></sup></td><td align="left" valign="top">92.35<sup><xref ref-type="table-fn" rid="table9fn4">d</xref></sup></td><td align="left" valign="top">93.30<sup><xref ref-type="table-fn" rid="table9fn4">d</xref></sup></td><td align="left" valign="top">81.85</td><td align="left" valign="top">68.20<sup><xref ref-type="table-fn" rid="table9fn4">d</xref></sup></td><td align="left" valign="top">92.35<sup><xref ref-type="table-fn" rid="table9fn4">d</xref></sup></td><td align="left" valign="top">95.04</td></tr></tbody></table><table-wrap-foot><fn id="table9fn1"><p><sup>a</sup>Rank represents the overall rating for each technique, with &#x201C;1&#x201D; indicating the best technique across all models.</p></fn><fn id="table9fn2"><p><sup>b</sup>AUROC: area under the receiver operating characteristic curve.</p></fn><fn id="table9fn3"><p><sup>c</sup>Significant difference between an augmentation technique and the Random Pixel Swap technique across all models.</p></fn><fn id="table9fn4"><p><sup>d</sup>Highest value in the column.</p></fn><fn id="table9fn5"><p><sup>e</sup>Significant difference between training using an augmentation technique and the base model across all models.</p></fn></table-wrap-foot></table-wrap></sec><sec id="s3-8"><title>Comparison With Prior Work</title><p>Our experimental results demonstrated improvements over the results of previous studies using both the IQ-OTH/NCCD and chest CT scan images datasets. For the IQ-OTH/NCCD dataset, our approach achieved a 7.67% performance improvement over a machine learning technique in the study by Kareem et al [<xref ref-type="bibr" rid="ref57">57</xref>], a 4.76% improvement over an ensemble of VGG-16, ResNet-50, InceptionV3, and EfficientNetB7 models in the study by Solyman et al [<xref ref-type="bibr" rid="ref58">58</xref>], and a 2.13% enhancement over an ensemble of 3 custom CNNs in the study by Abe et al [<xref ref-type="bibr" rid="ref59">59</xref>]. Similarly, for the chest CT scan images dataset, our method showed a 5.78% improvement over a 3-layer custom CNN in the study by Mamun et al [<xref ref-type="bibr" rid="ref60">60</xref>] and a 2.22% improvement over a 5-layer CNN with a custom Mavage Pooling layer in the study by Abe et al [<xref ref-type="bibr" rid="ref47">47</xref>]. The comparative results are detailed in <xref ref-type="table" rid="table10">Table 10</xref>.</p><table-wrap id="t10" position="float"><label>Table 10.</label><caption><p>Comparison of our study results with the results of previous studies on the analysis of the IQ-OTH/NCCD<sup><xref ref-type="table-fn" rid="table10fn1">a</xref></sup> and chest CT<sup><xref ref-type="table-fn" rid="table10fn2">b</xref></sup> scan images datasets.</p></caption><table id="table10" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Dataset and study</td><td align="left" valign="top">Accuracy, %</td><td align="left" valign="bottom">Number of classes</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="3">IQ-OTH/NCCD</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Kareem et al [<xref ref-type="bibr" rid="ref57">57</xref>]</td><td align="left" valign="top">89.89</td><td align="left" valign="top">3</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Solyman et al [<xref ref-type="bibr" rid="ref58">58</xref>]</td><td align="left" valign="top">92.80</td><td align="left" valign="top">3</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Abe et al [<xref ref-type="bibr" rid="ref59">59</xref>]</td><td align="left" valign="top">95.43</td><td align="left" valign="top">3</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Our study</td><td align="left" valign="top">97.56</td><td align="left" valign="top">3</td></tr><tr><td align="left" valign="top" colspan="3">Chest CT scan images</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Mamun et al [<xref ref-type="bibr" rid="ref60">60</xref>]</td><td align="left" valign="top">92.00</td><td align="left" valign="top">4</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Abe et al [<xref ref-type="bibr" rid="ref47">47</xref>]</td><td align="left" valign="top">95.56</td><td align="left" valign="top">4</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Our study</td><td align="left" valign="top">97.78</td><td align="left" valign="top">4</td></tr></tbody></table><table-wrap-foot><fn id="table10fn1"><p><sup>a</sup>IQ-OTH/NCCD: Iraq-Oncology Teaching Hospital/National Center for Cancer Diseases.</p></fn><fn id="table10fn2"><p><sup>b</sup>CT: computed tomography.</p></fn></table-wrap-foot></table-wrap></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Findings</title><p>The experimental results of the study demonstrated that the RPS DA technique significantly enhanced the diagnostic performance of both CNN and transformer architectures for lung cancer diagnosis from CT scan images. Our comprehensive evaluation demonstrated that RPS consistently outperformed 4 established augmentation methods (CutMix, Random Erasing, MixUp, and Cutout) across multiple performance metrics and diverse experimental conditions. The superior efficacy of RPS stems from its unique capacity to preserve critical anatomical content while generating clinically meaningful variations through controlled intraimage pixel swapping. This characteristic makes RPS particularly valuable for medical imaging applications where maintaining content integrity is essential for an accurate diagnosis.</p><p>For CNN architectures, specifically ResNet-34, RPS yielded remarkable performance improvements. ResNet-34 achieved peak accuracies of 97.56% for the IQ-OTH/NCCD dataset and 97.78% for the chest CT scan images dataset, with corresponding AUROC scores of 98.61% and 99.46%, respectively, at 512&#x00D7;512 image resolution. The technique&#x2019;s effectiveness with MobileNetV3 (96.65% accuracy and 98.0% AUROC for the IQ-OTH/NCCD dataset; 96.83% accuracy and 98.75% AUROC for the chest CT scan images dataset) is particularly notable given this model&#x2019;s lightweight architecture, suggesting RPS&#x2019;s potential for deployment in resource-constrained clinical settings where efficient models are often preferred [<xref ref-type="bibr" rid="ref56">56</xref>]. The study results represent a substantial advancement over conventional augmentation approaches, as RPS effectively addresses the inherent limitation of CNNs in capturing global relationships by creating localized variations that enhance feature learning while preserving diagnostically relevant image features.</p><p>The transformer-based architectures (Vision Transformer and Swin Transformer) showed particularly notable improvements when augmented with RPS. While transformer models conventionally demand large-scale training datasets to achieve peak performance, RPS effectively compensated for data limitations by generating variations that preserved the overall image content for proper attention mechanism functioning. For the Vision Transformer, RPS augmentation significantly enhanced performance, reaching 92.38% accuracy and 96.93% AUROC on the IQ-OTH/NCCD dataset and 86.67% accuracy and 96.32% AUROC on the chest CT scan images dataset. The Swin Transformer demonstrated robust performance gains, achieving 94.82% accuracy and 98.12% AUROC on the IQ-OTH/NCCD dataset and 96.19% accuracy and 98.90% AUROC on the chest CT scan images dataset when enhanced with RPS. The study results showed that RPS enables transformer models to develop more robust and clinically relevant feature representations, even with limited training data.</p><p>Our comparative analysis revealed RPS&#x2019;s consistent dominance across evaluation metrics and experimental conditions. While CutMix showed marginal advantages in specific scenarios (notably a 0.31% accuracy improvement with ResNet-34 at 512&#x00D7;512 image resolution), RPS maintained substantially better AUROC scores (5.31% higher in the same comparison), indicating more reliable diagnostic discrimination capability. This performance pattern held true across both the IQ-OTH/NCCD and chest CT scan images datasets, with RPS consistently ranking the highest in our comprehensive evaluation framework. Importantly, while conventional augmentation techniques sometimes degraded model performance in certain scenarios [<xref ref-type="bibr" rid="ref38">38</xref>,<xref ref-type="bibr" rid="ref40">40</xref>], RPS demonstrated universal performance enhancement across all tested conditions. Three fundamental characteristics explain RPS&#x2019;s exceptional effectiveness. The first characteristic is anatomical content preservation. Unlike methods that erase or mix image regions, RPS maintains all original diagnostic information while creating realistic variations through a controlled, systematic, random patch-based pixel swap within carefully defined ROIs. This approach preserves the clinical relevance of training samples while providing valuable data diversity. The second characteristic is architecture agnostic adaptability. The technique&#x2019;s parameter-free implementation and tunable swap area factor enable optimal performance across diverse model architectures without requiring architecture-specific adjustments. This flexibility makes RPS particularly valuable for medical imaging research, where multiple architectures may be explored. The third characteristic is clinical pathological relevance. By restricting pixel swaps to anatomically plausible regions within lung tissue (especially when combined with ROI segmentation), RPS enhances the learning of pathological features that may appear anywhere in the pulmonary anatomy, a crucial capability given the unpredictable spatial distribution of malignant nodules in many cancer cases [<xref ref-type="bibr" rid="ref61">61</xref>].</p><p>Validation experiments using independently acquired datasets with different scanning protocols and equipment configurations demonstrated RPS&#x2019;s superior generalization capabilities. The technique achieved these results while adding minimal computational overhead (statistically insignificant increases in training time, <italic>P</italic>&#x003E;.05), making it practical for real-world clinical implementation. Furthermore, RPS showed excellent compatibility with conventional augmentation methods, providing additional performance gains when combined with rotation and flipping operations, which suggests easy integration into existing medical image processing pipelines.</p><p>These findings offer significant implications for the development of computer-aided diagnosis systems. RPS directly addresses two fundamental challenges in medical AI: (1) the scarcity of annotated medical imaging data and (2) the limited generalizability of many models across different clinical settings [<xref ref-type="bibr" rid="ref23">23</xref>]. By consistently outperforming current state-of-the-art techniques while maintaining computational efficiency, RPS emerges as a versatile solution suitable for both research investigations and clinical deployment. Additionally, the technique&#x2019;s effectiveness suggests promising applications in educational settings for training radiologists, where realistic image variations could enhance learning without requiring additional patient scans.</p></sec><sec id="s4-2"><title>Conclusions</title><p>The findings of this study demonstrate that RPS is a robust and versatile DA technique that significantly enhances the performance of both CNN and transformer architectures for lung cancer diagnosis from CT scan images. By preserving anatomical content while introducing meaningful variability, RPS outperforms existing augmentation methods across multiple metrics and datasets, achieving improved accuracy and AUROC scores. Its computational efficiency, adaptability to diverse architectures, and ability to improve generalization make it particularly valuable for medical imaging applications where data scarcity and model reliability are critical challenges. RPS not only advances the technical frontier of DA but also holds immediate promise for improving computer-aided diagnosis systems in clinical practice. Future work will explore its extension to other medical imaging modalities (magnetic resonance, ultrasound, and x-ray imaging) and extension to 3D applications.</p></sec></sec></body><back><ack><p>This research was funded by the Department of Science and Innovation&#x2013;Council for Scientific and Industrial Research (DSI-CSIR) Inter-Bursary Support (IBS) Programme.</p></ack><notes><sec><title>Data Availability</title><p>The data supporting the findings of this study are available upon reasonable request from the corresponding author.</p><p>The Iraq-Oncology Teaching Hospital/National Center for Cancer Diseases (IQ-OTH/NCCD) dataset is available on Mendeley Data [<xref ref-type="bibr" rid="ref35">35</xref>]. The chest computed tomography (CT) scan images dataset is available on Kaggle [<xref ref-type="bibr" rid="ref55">55</xref>]. The code is available on GitHub [<xref ref-type="bibr" rid="ref62">62</xref>].</p></sec></notes><fn-group><fn fn-type="con"><p>Conceptualization: AAA</p><p>Data curation: AAA</p><p>Formal analysis: AAA</p><p>Funding acquisition: AAA</p><p>Investigation: AAA</p><p>Methodology: AAA</p><p>Project administration: NM</p><p>Resources: AAA</p><p>Software: AAA</p><p>Supervision: NM</p><p>Validation: AAA</p><p>Visualization: AAA</p><p>Writing &#x2013; original draft: AAA</p><p>Writing &#x2013; review &#x0026; editing: NM</p><p>All authors have read and agreed to the published version of the manuscript.</p></fn><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AUROC</term><def><p>area under the receiver operating characteristic curve</p></def></def-item><def-item><term id="abb2">CNN</term><def><p>convolutional neural network</p></def></def-item><def-item><term id="abb3">CT</term><def><p>computed tomography</p></def></def-item><def-item><term id="abb4">DA</term><def><p>data augmentation</p></def></def-item><def-item><term id="abb5">DL</term><def><p>deep learning</p></def></def-item><def-item><term id="abb6">IQ-OTH/NCCD</term><def><p>Iraq-Oncology Teaching Hospital/National Center for Cancer Diseases</p></def></def-item><def-item><term id="abb7">MVSF</term><def><p>multivalue swap area factor</p></def></def-item><def-item><term id="abb8">ROI</term><def><p>region of interest</p></def></def-item><def-item><term id="abb9">RPS</term><def><p>Random Pixel Swap</p></def></def-item><def-item><term id="abb10">SVSF</term><def><p>single-value swap area factor</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yoon</surname><given-names>SM</given-names> </name><name name-style="western"><surname>Shaikh</surname><given-names>T</given-names> </name><name name-style="western"><surname>Hallman</surname><given-names>M</given-names> </name></person-group><article-title>Therapeutic management options for stage III non-small cell lung cancer</article-title><source>World J Clin Oncol</source><year>2017</year><month>02</month><day>10</day><volume>8</volume><issue>1</issue><fpage>1</fpage><lpage>20</lpage><pub-id pub-id-type="doi">10.5306/wjco.v8.i1.1</pub-id><pub-id pub-id-type="medline">28246582</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bray</surname><given-names>F</given-names> </name><name name-style="western"><surname>Ferlay</surname><given-names>J</given-names> </name><name name-style="western"><surname>Soerjomataram</surname><given-names>I</given-names> </name><name name-style="western"><surname>Siegel</surname><given-names>RL</given-names> </name><name name-style="western"><surname>Torre</surname><given-names>LA</given-names> </name><name name-style="western"><surname>Jemal</surname><given-names>A</given-names> </name></person-group><article-title>Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries</article-title><source>CA A Cancer J Clinicians</source><year>2018</year><month>11</month><volume>68</volume><issue>6</issue><fpage>394</fpage><lpage>424</lpage><pub-id pub-id-type="doi">10.3322/caac.21492</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Schabath</surname><given-names>MB</given-names> </name><name name-style="western"><surname>Cote</surname><given-names>ML</given-names> </name></person-group><article-title>Cancer progress and priorities: lung cancer</article-title><source>Cancer Epidemiol Biomarkers Prev</source><year>2019</year><month>10</month><volume>28</volume><issue>10</issue><fpage>1563</fpage><lpage>1579</lpage><pub-id pub-id-type="doi">10.1158/1055-9965.EPI-19-0221</pub-id><pub-id pub-id-type="medline">31575553</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Travis</surname><given-names>WD</given-names> </name><name name-style="western"><surname>Brambilla</surname><given-names>E</given-names> </name><name name-style="western"><surname>Noguchi</surname><given-names>M</given-names> </name><etal/></person-group><article-title>International association for the study of lung cancer/american thoracic society/european respiratory society international multidisciplinary classification of lung adenocarcinoma</article-title><source>J Thorac Oncol</source><year>2011</year><month>02</month><volume>6</volume><issue>2</issue><fpage>244</fpage><lpage>285</lpage><pub-id pub-id-type="doi">10.1097/JTO.0b013e318206a221</pub-id><pub-id pub-id-type="medline">21252716</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Shin</surname><given-names>HJ</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>MS</given-names> </name><name name-style="western"><surname>Kho</surname><given-names>BG</given-names> </name><etal/></person-group><article-title>Delayed diagnosis of lung cancer due to misdiagnosis as worsening of sarcoidosis: a case report</article-title><source>BMC Pulm Med</source><year>2020</year><month>03</month><day>21</day><volume>20</volume><issue>1</issue><fpage>71</fpage><pub-id pub-id-type="doi">10.1186/s12890-020-1105-2</pub-id><pub-id pub-id-type="medline">32199453</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ruano-Ravi&#x00F1;a</surname><given-names>A</given-names> </name><name name-style="western"><surname>Provencio</surname><given-names>M</given-names> </name><name name-style="western"><surname>Calvo de Juan</surname><given-names>V</given-names> </name><etal/></person-group><article-title>Lung cancer symptoms at diagnosis: results of a nationwide registry study</article-title><source>ESMO Open</source><year>2020</year><month>11</month><volume>5</volume><issue>6</issue><fpage>e001021</fpage><pub-id pub-id-type="doi">10.1136/esmoopen-2020-001021</pub-id><pub-id pub-id-type="medline">33214227</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Del Ciello</surname><given-names>A</given-names> </name><name name-style="western"><surname>Franchi</surname><given-names>P</given-names> </name><name name-style="western"><surname>Contegiacomo</surname><given-names>A</given-names> </name><name name-style="western"><surname>Cicchetti</surname><given-names>G</given-names> </name><name name-style="western"><surname>Bonomo</surname><given-names>L</given-names> </name><name name-style="western"><surname>Larici</surname><given-names>AR</given-names> </name></person-group><article-title>Missed lung cancer: when, where, and why?</article-title><source>Diagn Interv Radiol</source><year>2017</year><volume>23</volume><issue>2</issue><fpage>118</fpage><lpage>126</lpage><pub-id pub-id-type="doi">10.5152/dir.2016.16187</pub-id><pub-id pub-id-type="medline">28206951</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Blandin Knight</surname><given-names>S</given-names> </name><name name-style="western"><surname>Crosbie</surname><given-names>PA</given-names> </name><name name-style="western"><surname>Balata</surname><given-names>H</given-names> </name><name name-style="western"><surname>Chudziak</surname><given-names>J</given-names> </name><name name-style="western"><surname>Hussell</surname><given-names>T</given-names> </name><name name-style="western"><surname>Dive</surname><given-names>C</given-names> </name></person-group><article-title>Progress and prospects of early detection in lung cancer</article-title><source>Open Biol</source><year>2017</year><month>09</month><volume>7</volume><issue>9</issue><fpage>170070</fpage><pub-id pub-id-type="doi">10.1098/rsob.170070</pub-id><pub-id pub-id-type="medline">28878044</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Meza</surname><given-names>R</given-names> </name><name name-style="western"><surname>Jeon</surname><given-names>J</given-names> </name><name name-style="western"><surname>Toumazis</surname><given-names>I</given-names> </name><etal/></person-group><article-title>Evaluation of the benefits and harms of lung cancer screening with low-dose computed tomography: modeling study for the US preventive services task force</article-title><source>JAMA</source><year>2021</year><month>03</month><day>9</day><volume>325</volume><issue>10</issue><fpage>988</fpage><lpage>997</lpage><pub-id pub-id-type="doi">10.1001/jama.2021.1077</pub-id><pub-id pub-id-type="medline">33687469</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fitzgerald</surname><given-names>RC</given-names> </name><name name-style="western"><surname>Antoniou</surname><given-names>AC</given-names> </name><name name-style="western"><surname>Fruk</surname><given-names>L</given-names> </name><name name-style="western"><surname>Rosenfeld</surname><given-names>N</given-names> </name></person-group><article-title>The future of early cancer detection</article-title><source>Nat Med</source><year>2022</year><month>04</month><volume>28</volume><issue>4</issue><fpage>666</fpage><lpage>677</lpage><pub-id pub-id-type="doi">10.1038/s41591-022-01746-x</pub-id><pub-id pub-id-type="medline">35440720</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Abujudeh</surname><given-names>HH</given-names> </name><name name-style="western"><surname>Boland</surname><given-names>GW</given-names> </name><name name-style="western"><surname>Kaewlai</surname><given-names>R</given-names> </name><etal/></person-group><article-title>Abdominal and pelvic computed tomography (CT) interpretation: discrepancy rates among experienced radiologists</article-title><source>Eur Radiol</source><year>2010</year><month>08</month><volume>20</volume><issue>8</issue><fpage>1952</fpage><lpage>1957</lpage><pub-id pub-id-type="doi">10.1007/s00330-010-1763-1</pub-id><pub-id pub-id-type="medline">20336300</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hoffman</surname><given-names>RM</given-names> </name><name name-style="western"><surname>Atallah</surname><given-names>RP</given-names> </name><name name-style="western"><surname>Struble</surname><given-names>RD</given-names> </name><name name-style="western"><surname>Badgett</surname><given-names>RG</given-names> </name></person-group><article-title>Lung cancer screening with low-dose CT: a meta-analysis</article-title><source>J Gen Intern Med</source><year>2020</year><month>10</month><volume>35</volume><issue>10</issue><fpage>3015</fpage><lpage>3025</lpage><pub-id pub-id-type="doi">10.1007/s11606-020-05951-7</pub-id><pub-id pub-id-type="medline">32583338</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bonney</surname><given-names>A</given-names> </name><name name-style="western"><surname>Malouf</surname><given-names>R</given-names> </name><name name-style="western"><surname>Marchal</surname><given-names>C</given-names> </name><etal/></person-group><article-title>Impact of low-dose computed tomography (LDCT) screening on lung cancer-related mortality</article-title><source>Cochrane Database Syst Rev</source><year>2022</year><month>08</month><day>3</day><volume>8</volume><issue>8</issue><fpage>CD013829</fpage><pub-id pub-id-type="doi">10.1002/14651858.CD013829.pub2</pub-id><pub-id pub-id-type="medline">35921047</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Swensen</surname><given-names>SJ</given-names> </name><name name-style="western"><surname>Jett</surname><given-names>JR</given-names> </name><name name-style="western"><surname>Hartman</surname><given-names>TE</given-names> </name><etal/></person-group><article-title>Lung cancer screening with CT: Mayo Clinic experience</article-title><source>Radiology</source><year>2003</year><month>03</month><volume>226</volume><issue>3</issue><fpage>756</fpage><lpage>761</lpage><pub-id pub-id-type="doi">10.1148/radiol.2263020036</pub-id><pub-id pub-id-type="medline">12601181</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Silvestri</surname><given-names>GA</given-names> </name><name name-style="western"><surname>Goldman</surname><given-names>L</given-names> </name><name name-style="western"><surname>Tanner</surname><given-names>NT</given-names> </name><etal/></person-group><article-title>Outcomes from more than 1 million people screened for lung cancer with low-dose CT imaging</article-title><source>Chest</source><year>2023</year><month>07</month><volume>164</volume><issue>1</issue><fpage>241</fpage><lpage>251</lpage><pub-id pub-id-type="doi">10.1016/j.chest.2023.02.003</pub-id><pub-id pub-id-type="medline">36773935</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Krupinski</surname><given-names>EA</given-names> </name><name name-style="western"><surname>Berbaum</surname><given-names>KS</given-names> </name><name name-style="western"><surname>Caldwell</surname><given-names>RT</given-names> </name><name name-style="western"><surname>Schartz</surname><given-names>KM</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>J</given-names> </name></person-group><article-title>Long radiology workdays reduce detection and accommodation accuracy</article-title><source>J Am Coll Radiol</source><year>2010</year><month>09</month><volume>7</volume><issue>9</issue><fpage>698</fpage><lpage>704</lpage><pub-id pub-id-type="doi">10.1016/j.jacr.2010.03.004</pub-id><pub-id pub-id-type="medline">20816631</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Jacobsen</surname><given-names>MM</given-names> </name><name name-style="western"><surname>Silverstein</surname><given-names>SC</given-names> </name><name name-style="western"><surname>Quinn</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Timeliness of access to lung cancer diagnosis and treatment: a scoping literature review</article-title><source>Lung Cancer (Auckl)</source><year>2017</year><month>10</month><volume>112</volume><fpage>156</fpage><lpage>164</lpage><pub-id pub-id-type="doi">10.1016/j.lungcan.2017.08.011</pub-id><pub-id pub-id-type="medline">29191588</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sathyakumar</surname><given-names>K</given-names> </name><name name-style="western"><surname>Munoz</surname><given-names>M</given-names> </name><name name-style="western"><surname>Singh</surname><given-names>J</given-names> </name><name name-style="western"><surname>Hussain</surname><given-names>N</given-names> </name><name name-style="western"><surname>Babu</surname><given-names>BA</given-names> </name></person-group><article-title>Automated lung cancer detection using artificial intelligence (AI) deep convolutional neural networks: a narrative literature review</article-title><source>Cureus</source><year>2020</year><month>08</month><day>25</day><volume>12</volume><issue>8</issue><fpage>e10017</fpage><pub-id pub-id-type="doi">10.7759/cureus.10017</pub-id><pub-id pub-id-type="medline">32989411</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Huang</surname><given-names>S</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Shen</surname><given-names>N</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Zhao</surname><given-names>Q</given-names> </name></person-group><article-title>Artificial intelligence in lung cancer diagnosis and prognosis: current application and future perspective</article-title><source>Semin Cancer Biol</source><year>2023</year><month>02</month><volume>89</volume><fpage>30</fpage><lpage>37</lpage><pub-id pub-id-type="doi">10.1016/j.semcancer.2023.01.006</pub-id><pub-id pub-id-type="medline">36682439</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Marentakis</surname><given-names>P</given-names> </name><name name-style="western"><surname>Karaiskos</surname><given-names>P</given-names> </name><name name-style="western"><surname>Kouloulias</surname><given-names>V</given-names> </name><etal/></person-group><article-title>Lung cancer histology classification from CT images based on radiomics and deep learning models</article-title><source>Med Biol Eng Comput</source><year>2021</year><month>01</month><volume>59</volume><issue>1</issue><fpage>215</fpage><lpage>226</lpage><pub-id pub-id-type="doi">10.1007/s11517-020-02302-w</pub-id><pub-id pub-id-type="medline">33411267</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Buduma</surname><given-names>N</given-names> </name><name name-style="western"><surname>Buduma</surname><given-names>N</given-names> </name><name name-style="western"><surname>Papa</surname><given-names>J</given-names> </name></person-group><source>Fundamentals of Deep Learning</source><year>2022</year><edition>2</edition><publisher-name>O&#x2019;Reilly Media, Inc</publisher-name><pub-id pub-id-type="other">9781492082170</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Forte</surname><given-names>GC</given-names> </name><name name-style="western"><surname>Altmayer</surname><given-names>S</given-names> </name><name name-style="western"><surname>Silva</surname><given-names>RF</given-names> </name><etal/></person-group><article-title>Deep learning algorithms for diagnosis of lung cancer: a systematic review and meta-analysis</article-title><source>Cancers (Basel)</source><year>2022</year><month>08</month><day>9</day><volume>14</volume><issue>16</issue><fpage>3856</fpage><pub-id pub-id-type="doi">10.3390/cancers14163856</pub-id><pub-id pub-id-type="medline">36010850</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Dosovitskiy</surname><given-names>A</given-names> </name><name name-style="western"><surname>Beyer</surname><given-names>L</given-names> </name><name name-style="western"><surname>Kolesnikov</surname><given-names>A</given-names> </name><etal/></person-group><article-title>An image is worth 16x16 words: transformers for image recognition at scale</article-title><source>arXiv</source><year>2020</year><access-date>2025-08-22</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2010.11929">https://arxiv.org/abs/2010.11929</ext-link></comment></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lecun</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Bottou</surname><given-names>L</given-names> </name><name name-style="western"><surname>Bengio</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Haffner</surname><given-names>P</given-names> </name></person-group><article-title>Gradient-based learning applied to document recognition</article-title><source>Proc IEEE</source><year>1998</year><volume>86</volume><issue>11</issue><fpage>2278</fpage><lpage>2324</lpage><pub-id pub-id-type="doi">10.1109/5.726791</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Lu</surname><given-names>K</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>Y</given-names> </name></person-group><article-title>Comparison of the potential between transformer and CNN in image classification</article-title><access-date>2025-08-22</access-date><conf-name>2nd International Conference on Machine Learning and Computer Application</conf-name><conf-date>Dec 17-19, 2021</conf-date><conf-loc>Shenyang, China</conf-loc><comment><ext-link ext-link-type="uri" xlink:href="https://ieeexplore.ieee.org/document/9736894">https://ieeexplore.ieee.org/document/9736894</ext-link></comment></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>H</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Ai</surname><given-names>T</given-names> </name></person-group><article-title>Long-range dependencies learning based on non-local 1D-convolutional neural network for rolling bearing fault diagnosis</article-title><source>J Dyn Monit Diagn</source><year>2022</year><month>04</month><day>12</day><issue>3</issue><fpage>148</fpage><lpage>159</lpage><pub-id pub-id-type="doi">10.37965/jdmd.2022.53</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Zhai</surname><given-names>X</given-names> </name><name name-style="western"><surname>Kolesnikov</surname><given-names>A</given-names> </name><name name-style="western"><surname>Houlsby</surname><given-names>N</given-names> </name><name name-style="western"><surname>Beyer</surname><given-names>L</given-names> </name></person-group><article-title>Scaling vision transformers</article-title><conf-name>2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</conf-name><conf-date>Jun 18-24, 2022</conf-date><conf-loc>New Orleans, LA</conf-loc><pub-id pub-id-type="doi">10.1109/CVPR52688.2022.01179</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Matsoukas</surname><given-names>C</given-names> </name><name name-style="western"><surname>Haslum</surname><given-names>JF</given-names> </name><name name-style="western"><surname>S&#x00F6;derberg</surname><given-names>M</given-names> </name><name name-style="western"><surname>Smith</surname><given-names>K</given-names> </name></person-group><article-title>Is it time to replace CNNs with transformers for medical images?</article-title><source>arXiv</source><year>2021</year><access-date>2025-08-22</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2108.09038">https://arxiv.org/abs/2108.09038</ext-link></comment></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Lin</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Cao</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>Swin transformer: hierarchical vision transformer using shifted windows</article-title><conf-name>2021 IEEE/CVF International Conference on Computer Vision (ICCV)</conf-name><conf-date>Oct 10-17, 2021</conf-date><conf-loc>Montreal, QC, Canada</conf-loc><pub-id pub-id-type="doi">10.1109/ICCV48922.2021.00986</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>D</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>F</given-names> </name><name name-style="western"><surname>Tie</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Qi</surname><given-names>L</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>F</given-names> </name></person-group><article-title>Res-trans networks for lung nodule classification</article-title><source>Int J Comput Assist Radiol Surg</source><year>2022</year><month>06</month><volume>17</volume><issue>6</issue><fpage>1059</fpage><lpage>1068</lpage><pub-id pub-id-type="doi">10.1007/s11548-022-02576-5</pub-id><pub-id pub-id-type="medline">35290646</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Nejad</surname><given-names>RR</given-names> </name><name name-style="western"><surname>Hooshmand</surname><given-names>S</given-names> </name></person-group><article-title>HViT4Lung: hybrid vision transformers augmented by transfer learning to enhance lung cancer diagnosis</article-title><conf-name>5th International Conference on Bio-engineering for Smart Technologies (BioSMART)</conf-name><conf-date>Jun 7-9, 2023</conf-date><conf-loc>Paris, France</conf-loc><pub-id pub-id-type="doi">10.1109/BioSMART58455.2023.10162074</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Atmakuru</surname><given-names>A</given-names> </name><name name-style="western"><surname>Chakraborty</surname><given-names>S</given-names> </name><name name-style="western"><surname>Faust</surname><given-names>O</given-names> </name><etal/></person-group><article-title>Deep learning in radiology for lung cancer diagnostics: a systematic review of classification, segmentation, and predictive modeling techniques</article-title><source>Expert Syst Appl</source><year>2024</year><month>12</month><volume>255</volume><fpage>124665</fpage><pub-id pub-id-type="doi">10.1016/j.eswa.2024.124665</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Santos</surname><given-names>C</given-names> </name><name name-style="western"><surname>Papa</surname><given-names>JP</given-names> </name></person-group><article-title>Avoiding overfitting: a survey on regularization methods for convolutional neural networks</article-title><source>ACM Comput Surv</source><year>2022</year><month>01</month><day>31</day><volume>54</volume><issue>10s</issue><fpage>1</fpage><lpage>25</lpage><pub-id pub-id-type="doi">10.1145/3510413</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Razzak</surname><given-names>MI</given-names> </name><name name-style="western"><surname>Naz</surname><given-names>S</given-names> </name><name name-style="western"><surname>Zaib</surname><given-names>A</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Dey</surname><given-names>N</given-names> </name><name name-style="western"><surname>Ashour</surname><given-names>A</given-names> </name><name name-style="western"><surname>Borra</surname><given-names>S</given-names> </name></person-group><article-title>Deep learning for medical image processing: overview, challenges and the future</article-title><source>Classification in BioApps Lecture Notes in Computational Vision and Biomechanics</source><year>2018</year><publisher-name>Springer</publisher-name><fpage>323</fpage><lpage>350</lpage><pub-id pub-id-type="doi">10.1007/978-3-319-65981-7_12</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Alyasriy</surname><given-names>H</given-names> </name></person-group><article-title>The IQ-OTHNCCD lung cancer dataset</article-title><source>Mendeley Data</source><year>2020</year><access-date>2025-08-22</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://data.mendeley.com/datasets/bhmdr45bh2/1">https://data.mendeley.com/datasets/bhmdr45bh2/1</ext-link></comment></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Shorten</surname><given-names>C</given-names> </name><name name-style="western"><surname>Khoshgoftaar</surname><given-names>TM</given-names> </name></person-group><article-title>A survey on image data augmentation for deep learning</article-title><source>J Big Data</source><year>2019</year><month>12</month><volume>6</volume><issue>1</issue><fpage>1</fpage><lpage>48</lpage><pub-id pub-id-type="doi">10.1186/s40537-019-0197-0</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chlap</surname><given-names>P</given-names> </name><name name-style="western"><surname>Min</surname><given-names>H</given-names> </name><name name-style="western"><surname>Vandenberg</surname><given-names>N</given-names> </name><name name-style="western"><surname>Dowling</surname><given-names>J</given-names> </name><name name-style="western"><surname>Holloway</surname><given-names>L</given-names> </name><name name-style="western"><surname>Haworth</surname><given-names>A</given-names> </name></person-group><article-title>A review of medical image data augmentation techniques for deep learning applications</article-title><source>J Med Imaging Radiat Oncol</source><year>2021</year><month>08</month><volume>65</volume><issue>5</issue><fpage>545</fpage><lpage>563</lpage><pub-id pub-id-type="doi">10.1111/1754-9485.13261</pub-id><pub-id pub-id-type="medline">34145766</pub-id></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lin</surname><given-names>CH</given-names> </name><name name-style="western"><surname>Kaushik</surname><given-names>C</given-names> </name><name name-style="western"><surname>Dyer</surname><given-names>EL</given-names> </name><name name-style="western"><surname>Muthukumar</surname><given-names>V</given-names> </name></person-group><article-title>The good, the bad and the ugly sides of data augmentation: an implicit spectral regularization perspective</article-title><source>J Mach Learn Res</source><year>2024</year><access-date>2025-08-22</access-date><volume>25</volume><fpage>1</fpage><lpage>85</lpage><comment><ext-link ext-link-type="uri" xlink:href="https://jmlr.org/papers/volume25/22-1312/22-1312.pdf">https://jmlr.org/papers/volume25/22-1312/22-1312.pdf</ext-link></comment></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Maharana</surname><given-names>K</given-names> </name><name name-style="western"><surname>Mondal</surname><given-names>S</given-names> </name><name name-style="western"><surname>Nemade</surname><given-names>B</given-names> </name></person-group><article-title>A review: data pre-processing and data augmentation techniques</article-title><source>Global Transitions Proceedings</source><year>2022</year><month>06</month><volume>3</volume><issue>1</issue><fpage>91</fpage><lpage>99</lpage><pub-id pub-id-type="doi">10.1016/j.gltp.2022.04.020</pub-id></nlm-citation></ref><ref id="ref40"><label>40</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Balestriero</surname><given-names>R</given-names> </name><name name-style="western"><surname>Bottou</surname><given-names>L</given-names> </name><name name-style="western"><surname>LeCun</surname><given-names>Y</given-names> </name></person-group><article-title>The effects of regularization and data augmentation are class dependent</article-title><conf-name>36th International Conference on Neural Information Processing Systems</conf-name><conf-date>Nov 28 to Dec 9, 2022</conf-date><conf-loc>New Orleans, LA</conf-loc><pub-id pub-id-type="doi">10.5555/3600270.3603015</pub-id></nlm-citation></ref><ref id="ref41"><label>41</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Krizhevsky</surname><given-names>A</given-names> </name><name name-style="western"><surname>Sutskever</surname><given-names>I</given-names> </name><name name-style="western"><surname>Hinton</surname><given-names>GE</given-names> </name></person-group><article-title>ImageNet classification with deep convolutional neural networks</article-title><source>Commun ACM</source><year>2012</year><volume>60</volume><issue>6</issue><fpage>84</fpage><lpage>90</lpage><pub-id pub-id-type="doi">10.1145/306538</pub-id></nlm-citation></ref><ref id="ref42"><label>42</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Goodfellow</surname><given-names>IJ</given-names> </name><name name-style="western"><surname>Pouget-Abadie</surname><given-names>J</given-names> </name><name name-style="western"><surname>Mirza</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Generative adversarial nets</article-title><conf-name>28th International Conference on Neural Information Processing Systems</conf-name><conf-date>Dec 8-13, 2014</conf-date><conf-loc>Montreal, Canada</conf-loc><pub-id pub-id-type="doi">10.5555/2969033.2969125</pub-id></nlm-citation></ref><ref id="ref43"><label>43</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>DeVries</surname><given-names>T</given-names> </name><name name-style="western"><surname>Taylor</surname><given-names>GW</given-names> </name></person-group><article-title>Improved regularization of convolutional neural networks with cutout</article-title><source>arXiv</source><year>2017</year><access-date>2025-08-22</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1708.04552">https://arxiv.org/abs/1708.04552</ext-link></comment></nlm-citation></ref><ref id="ref44"><label>44</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhong</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Zheng</surname><given-names>L</given-names> </name><name name-style="western"><surname>Kang</surname><given-names>G</given-names> </name><name name-style="western"><surname>Li</surname><given-names>S</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>Y</given-names> </name></person-group><article-title>Random erasing data augmentation</article-title><source>Proc AAAI Conf Artif Intell</source><year>2020</year><month>04</month><day>3</day><volume>34</volume><issue>7</issue><fpage>13001</fpage><lpage>13008</lpage><pub-id pub-id-type="doi">10.1609/aaai.v34i07.7000</pub-id></nlm-citation></ref><ref id="ref45"><label>45</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>H</given-names> </name><name name-style="western"><surname>Cisse</surname><given-names>M</given-names> </name><name name-style="western"><surname>Dauphin</surname><given-names>YN</given-names> </name><name name-style="western"><surname>Lopez-Paz</surname><given-names>D</given-names> </name></person-group><article-title>Mixup: beyond empirical risk minimization</article-title><source>arXiv</source><year>2017</year><access-date>2025-08-22</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1710.09412">https://arxiv.org/abs/1710.09412</ext-link></comment></nlm-citation></ref><ref id="ref46"><label>46</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Yun</surname><given-names>S</given-names> </name><name name-style="western"><surname>Han</surname><given-names>D</given-names> </name><name name-style="western"><surname>Chun</surname><given-names>S</given-names> </name><name name-style="western"><surname>Oh</surname><given-names>SJ</given-names> </name><name name-style="western"><surname>Yoo</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Choe</surname><given-names>J</given-names> </name></person-group><article-title>CutMix: regularization strategy to train strong classifiers with localizable features</article-title><conf-name>2019 IEEE/CVF International Conference on Computer Vision (ICCV)</conf-name><conf-date>Oct 27 to Nov 2, 2019</conf-date><conf-loc>Seoul, Korea (South)</conf-loc><pub-id pub-id-type="doi">10.1109/ICCV.2019.00612</pub-id></nlm-citation></ref><ref id="ref47"><label>47</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Abe</surname><given-names>A</given-names> </name><name name-style="western"><surname>Nyathi</surname><given-names>M</given-names> </name><name name-style="western"><surname>Okunade</surname><given-names>A</given-names> </name></person-group><article-title>Lung cancer diagnosis from computed tomography scans using convolutional neural network architecture with Mavage pooling technique</article-title><source>AIMS Med Sci</source><year>2025</year><volume>12</volume><issue>1</issue><fpage>13</fpage><lpage>27</lpage><pub-id pub-id-type="doi">10.3934/medsci.2025002</pub-id></nlm-citation></ref><ref id="ref48"><label>48</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Jiang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Manem</surname><given-names>VSK</given-names> </name></person-group><article-title>Data augmented lung cancer prediction framework using the nested case control NLST cohort</article-title><source>Front Oncol</source><year>2025</year><month>02</month><day>25</day><volume>15</volume><fpage>1492758</fpage><pub-id pub-id-type="doi">10.3389/fonc.2025.1492758</pub-id></nlm-citation></ref><ref id="ref49"><label>49</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Jin</surname><given-names>X</given-names> </name><name name-style="western"><surname>Zhu</surname><given-names>H</given-names> </name><name name-style="western"><surname>Li</surname><given-names>S</given-names> </name><etal/></person-group><article-title>A survey on mixup augmentations and beyond</article-title><source>arXiv</source><year>2024</year><access-date>2025-08-22</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2409.05202">https://arxiv.org/abs/2409.05202</ext-link></comment></nlm-citation></ref><ref id="ref50"><label>50</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>He</surname><given-names>J</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>B</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>X</given-names> </name></person-group><article-title>Non-local patch mixup for unsupervised domain adaptation</article-title><conf-name>2022 IEEE International Conference on Data Mining (ICDM)</conf-name><conf-date>Nov 28 to Dec 1, 2022</conf-date><conf-loc>Orlando, FL</conf-loc><pub-id pub-id-type="doi">10.1109/ICDM54844.2022.00116</pub-id></nlm-citation></ref><ref id="ref51"><label>51</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Oh</surname><given-names>J</given-names> </name><name name-style="western"><surname>Yun</surname><given-names>C</given-names> </name></person-group><article-title>Provable benefit of Cutout and CutMix for feature learning</article-title><source>arXiv</source><year>2024</year><access-date>2025-08-22</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2410.23672">https://arxiv.org/abs/2410.23672</ext-link></comment></nlm-citation></ref><ref id="ref52"><label>52</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>He</surname><given-names>K</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>X</given-names> </name><name name-style="western"><surname>Ren</surname><given-names>S</given-names> </name><name name-style="western"><surname>Sun</surname><given-names>J</given-names> </name></person-group><article-title>Deep residual learning for image recognition</article-title><conf-name>2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</conf-name><conf-date>Jun 27-30, 2016</conf-date><conf-loc>Las Vegas, NV</conf-loc><pub-id pub-id-type="doi">10.1109/CVPR.2016.90</pub-id></nlm-citation></ref><ref id="ref53"><label>53</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Howard</surname><given-names>A</given-names> </name><name name-style="western"><surname>Sandler</surname><given-names>M</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>B</given-names> </name><etal/></person-group><article-title>Searching for mobilenetv3</article-title><conf-name>2019 IEEE/CVF International Conference on Computer Vision (ICCV)</conf-name><conf-date>Oct 27 to Nov 2, 2019</conf-date><conf-loc>Seoul, Korea (South)</conf-loc><pub-id pub-id-type="doi">10.1109/ICCV.2019.00140</pub-id></nlm-citation></ref><ref id="ref54"><label>54</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Paszke</surname><given-names>A</given-names> </name><name name-style="western"><surname>Gross</surname><given-names>S</given-names> </name><name name-style="western"><surname>Massa</surname><given-names>F</given-names> </name><etal/></person-group><article-title>PyTorch: an imperative style, high-performance deep learning library</article-title><conf-name>33rd International Conference on Neural Information Processing Systems</conf-name><conf-date>Dec 8-14, 2019</conf-date><conf-loc>Vancouver, BC, Canada</conf-loc><pub-id pub-id-type="doi">10.5555/3454287.3455008</pub-id></nlm-citation></ref><ref id="ref55"><label>55</label><nlm-citation citation-type="web"><article-title>Chest CT-scan images dataset</article-title><source>Kaggle</source><access-date>2025-08-22</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.kaggle.com/datasets/mohamedhanyyy/chest-ctscan-images">https://www.kaggle.com/datasets/mohamedhanyyy/chest-ctscan-images</ext-link></comment></nlm-citation></ref><ref id="ref56"><label>56</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Setio</surname><given-names>AAA</given-names> </name><name name-style="western"><surname>Traverso</surname><given-names>A</given-names> </name><name name-style="western"><surname>de Bel</surname><given-names>T</given-names> </name><etal/></person-group><article-title>Validation, comparison, and combination of algorithms for automatic detection of pulmonary nodules in computed tomography images: the LUNA16 challenge</article-title><source>Med Image Anal</source><year>2017</year><month>12</month><volume>42</volume><fpage>1</fpage><lpage>13</lpage><pub-id pub-id-type="doi">10.1016/j.media.2017.06.015</pub-id><pub-id pub-id-type="medline">28732268</pub-id></nlm-citation></ref><ref id="ref57"><label>57</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kareem</surname><given-names>HF</given-names> </name><name name-style="western"><surname>AL-Huseiny</surname><given-names>MS</given-names> </name><name name-style="western"><surname>Y. Mohsen</surname><given-names>F</given-names> </name><name name-style="western"><surname>A. Khalil</surname><given-names>E</given-names> </name><name name-style="western"><surname>S. Hassan</surname><given-names>Z</given-names> </name></person-group><article-title>Evaluation of SVM performance in the detection of lung cancer in marked CT scan dataset</article-title><source>Indones J Electr Eng Comput Sci</source><year>2021</year><month>03</month><volume>21</volume><issue>3</issue><fpage>1731</fpage><pub-id pub-id-type="doi">10.11591/ijeecs.v21.i3.pp1731-1738</pub-id></nlm-citation></ref><ref id="ref58"><label>58</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Solyman</surname><given-names>S</given-names> </name><name name-style="western"><surname>Schwenker</surname><given-names>F</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Girma Debelee</surname><given-names>T</given-names> </name><name name-style="western"><surname>Ibenthal</surname><given-names>A</given-names> </name><name name-style="western"><surname>Schwenker</surname><given-names>F</given-names> </name></person-group><article-title>Lung tumor detection and recognition using deep convolutional neural networks</article-title><source>Pan-African Conference on Artificial Intelligence</source><year>2023</year><fpage>79</fpage><lpage>91</lpage><pub-id pub-id-type="doi">10.1007/978-3-031-31327-1_5</pub-id></nlm-citation></ref><ref id="ref59"><label>59</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Abe</surname><given-names>AA</given-names> </name><name name-style="western"><surname>Nyathi</surname><given-names>M</given-names> </name><name name-style="western"><surname>Okunade</surname><given-names>AA</given-names> </name><name name-style="western"><surname>Pilloy</surname><given-names>W</given-names> </name><name name-style="western"><surname>Kgole</surname><given-names>B</given-names> </name><name name-style="western"><surname>Nyakale</surname><given-names>N</given-names> </name></person-group><article-title>A robust deep learning algorithm for lung cancer detection from computed tomography images</article-title><source>Intelligence-Based Medicine</source><year>2025</year><volume>11</volume><fpage>100203</fpage><pub-id pub-id-type="doi">10.1016/j.ibmed.2025.100203</pub-id></nlm-citation></ref><ref id="ref60"><label>60</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Mamun</surname><given-names>M</given-names> </name><name name-style="western"><surname>Mahmud</surname><given-names>MI</given-names> </name><name name-style="western"><surname>Meherin</surname><given-names>M</given-names> </name><name name-style="western"><surname>Abdelgawad</surname><given-names>A</given-names> </name></person-group><article-title>LCDctCNN: lung cancer diagnosis of CT scan images using CNN based model</article-title><conf-name>10th International Conference on Signal Processing and Integrated Networks (SPIN)</conf-name><conf-date>Mar 23-24, 2023</conf-date><conf-loc>Noida, India</conf-loc><pub-id pub-id-type="doi">10.1109/SPIN57001.2023.10116075</pub-id></nlm-citation></ref><ref id="ref61"><label>61</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ali</surname><given-names>H</given-names> </name><name name-style="western"><surname>Mohsen</surname><given-names>F</given-names> </name><name name-style="western"><surname>Shah</surname><given-names>Z</given-names> </name></person-group><article-title>Improving diagnosis and prognosis of lung cancer using vision transformers: a scoping review</article-title><source>BMC Med Imaging</source><year>2023</year><month>09</month><day>15</day><volume>23</volume><issue>1</issue><fpage>129</fpage><pub-id pub-id-type="doi">10.1186/s12880-023-01098-z</pub-id><pub-id pub-id-type="medline">37715137</pub-id></nlm-citation></ref><ref id="ref62"><label>62</label><nlm-citation citation-type="web"><article-title>Random pixel swap (RPS) data augmentation technique code</article-title><source>GitHub</source><access-date>2025-08-22</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://github.com/Saintcodded/Random-Pixel-Swap-RPS-Data-Augmentation-Technique">https://github.com/Saintcodded/Random-Pixel-Swap-RPS-Data-Augmentation-Technique</ext-link></comment></nlm-citation></ref></ref-list></back></article>