<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing with OASIS Tables v3.0 20080202//EN" "https://jats.nlm.nih.gov/nlm-dtd/publishing/3.0/journalpub-oasis3.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:oasis="http://docs.oasis-open.org/ns/oasis-exchange/table" xml:lang="en" dtd-version="3.0" article-type="research-article">
  <front>
    <journal-meta><journal-id journal-id-type="publisher">MS</journal-id><journal-title-group>
    <journal-title>Mechanical Sciences</journal-title>
    <abbrev-journal-title abbrev-type="publisher">MS</abbrev-journal-title><abbrev-journal-title abbrev-type="nlm-ta">Mech. Sci.</abbrev-journal-title>
  </journal-title-group><issn pub-type="epub">2191-916X</issn><publisher>
    <publisher-name>Copernicus Publications</publisher-name>
    <publisher-loc>Göttingen, Germany</publisher-loc>
  </publisher></journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.5194/ms-17-371-2026</article-id><title-group><article-title>RIL-YOLO: a lightweight real-time object detection model on mobile devices for kart racing</article-title><alt-title>RIL-YOLO: a lightweight real-time object detection model</alt-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author" corresp="no" rid="aff1">
          <name><surname>Sha</surname><given-names>Ang</given-names></name>
          
        </contrib>
        <contrib contrib-type="author" corresp="no" rid="aff2">
          <name><surname>Xue</surname><given-names>Fuen</given-names></name>
          
        </contrib>
        <contrib contrib-type="author" corresp="no" rid="aff2">
          <name><surname>Zhang</surname><given-names>Yong</given-names></name>
          
        </contrib>
        <contrib contrib-type="author" corresp="yes" rid="aff3">
          <name><surname>Zang</surname><given-names>Xiaolin</given-names></name>
          <email>zangxiaolin@qut.edu.cn</email>
        </contrib>
        <contrib contrib-type="author" corresp="no" rid="aff4">
          <name><surname>Zhao</surname><given-names>Jinying</given-names></name>
          
        </contrib>
        <aff id="aff1"><label>1</label><institution>Innovation Institute for Sustainable Maritime Architecture Research and Technology, Qingdao University of Technology, Qingdao, China</institution>
        </aff>
        <aff id="aff2"><label>2</label><institution>School of Information and Control Engineering, Qingdao University of Technology, Qingdao, China</institution>
        </aff>
        <aff id="aff3"><label>3</label><institution>School of Architecture and Urban Planning, Qingdao University of Technology, Qingdao, China</institution>
        </aff>
        <aff id="aff4"><label>4</label><institution>Qingdao Cyber Racing Technology Co., Ltd., Qingdao, China</institution>
        </aff>
      </contrib-group>
      <author-notes><corresp id="corr1">Xiaolin Zang (zangxiaolin@qut.edu.cn)</corresp></author-notes><pub-date><day>9</day><month>April</month><year>2026</year></pub-date>
      
      <volume>17</volume>
      <issue>1</issue>
      <fpage>371</fpage><lpage>379</lpage>
      <history>
        <date date-type="received"><day>12</day><month>January</month><year>2026</year></date>
           <date date-type="rev-recd"><day>12</day><month>March</month><year>2026</year></date>
           <date date-type="accepted"><day>25</day><month>March</month><year>2026</year></date>
      </history>
      <permissions>
        <copyright-statement>Copyright: © 2026 Ang Sha et al.</copyright-statement>
        <copyright-year>2026</copyright-year>
      <license license-type="open-access"><license-p>This work is licensed under the Creative Commons Attribution 4.0 International License. To view a copy of this licence, visit <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link></license-p></license></permissions><self-uri xlink:href="https://ms.copernicus.org/articles/17/371/2026/ms-17-371-2026.html">This article is available from https://ms.copernicus.org/articles/17/371/2026/ms-17-371-2026.html</self-uri><self-uri xlink:href="https://ms.copernicus.org/articles/17/371/2026/ms-17-371-2026.pdf">The full text article is available as a PDF file from https://ms.copernicus.org/articles/17/371/2026/ms-17-371-2026.pdf</self-uri>
      <abstract><title>Abstract</title>

      <p id="d2e137">The kart is a high-speed mechanical system that requires real-time and reliable visual perception, while motion blur, occlusion, and limited mobile computing resources pose significant challenges. To address these issues, we propose RIL-YOLO, a lightweight object detection framework based on YOLOv8 and optimized for mobile deployment. The method incorporates motion blur data augmentation, a re-parameterized shared convolutional detection head architecture, an inner-CIoU (complete intersection over union) loss, and LAMP (layer-adaptive sparsity for the magnitude-based pruning)-based pruning to improve robustness, localization accuracy, and inference efficiency.  Experimental results show that, compared with YOLOv8n, RIL-YOLO improves mAP@0.5 and mAP@<inline-formula><mml:math id="M1" display="inline"><mml:mrow><mml:mo>[</mml:mo><mml:mn mathvariant="normal">0.5</mml:mn><mml:mo>:</mml:mo><mml:mn mathvariant="normal">0.95</mml:mn><mml:mo>]</mml:mo></mml:mrow></mml:math></inline-formula> by 2.8 % and 2.2 %, respectively, while reducing parameters by 83 %, lowering FLOPs (floating point operations per second) by 53 %, and increasing inference speed by approximately 25 %. The proposed method achieves a favorable balance between accuracy and real-time performance on resource-constrained mobile platforms.</p>
  </abstract>
    
<funding-group>
<award-group id="gs1">
<funding-source>National Natural Science Foundation of China</funding-source>
<award-id>Grant NO. 42301266</award-id>
</award-group>
<award-group id="gs2">
<funding-source>Natural Science Foundation of Shandong Province</funding-source>
<award-id>Grant NO. ZR2023QD005</award-id>
</award-group>
</funding-group>
</article-meta>
  </front>
<body>
      

<sec id="Ch1.S1" sec-type="intro">
  <label>1</label><title>Introduction</title>
      <p id="d2e165">The kart is a high-speed mechanical system for racing. It imposes stringent requirements on real-time visual perception for applications including safety monitoring and race analysis (Anggrainy et al., 2024; Matsumura et al., 2011). In these scenarios, severe motion blur, rapid scale variations, frequent occlusion, and complex backgrounds significantly increase detection difficulty, while deployment on mobile or edge devices further constrains model complexity and computational cost.  Achieving an effective balance between accuracy, real-time performance, and lightweight deployment therefore remains challenging in high-speed dynamic environments.</p>
      <p id="d2e168">In recent years, visual perception has also become a core component of autonomous driving systems, where real-time object detection supports environment understanding, obstacle avoidance, and decision-making.  Autonomous vehicles must operate reliably under dynamic conditions such as high-speed motion, illumination changes, occlusion, and motion blur while increasingly relying on on-board or edge computing platforms with limited computational resources. Therefore, improving detection robustness and efficiency under resource constraints has become an important research direction in intelligent transportation systems.</p>
      <p id="d2e171">Among existing vision-based approaches, the YOLO (you only look once) family has been widely adopted due to its end-to-end design and high inference efficiency. Since its introduction by Redmon et al. (2016), the YOLO family has continuously evolved, achieving improved accuracy and speed. YOLOv8 further adopts an anchor-free paradigm and a decoupled detection head, enabling a favorable trade-off between accuracy and speed. However, YOLOv8 still exhibits degraded robustness under motion blur and limited real-time performance under strict resource constraints.</p>
      <p id="d2e174">Recent studies have explored lightweight optimization of YOLO-based detectors through detection head redesign, label assignment optimization, and model compression. Chen et al. (2025) introduced a hierarchical decoupled prediction head and dynamic label assignment to enable low-latency detection of dense and occluded pedestrians on edge devices, while Lu and Liu (2022) combined attention mechanisms, depth-wise separable convolutions, and channel pruning to achieve efficient real-time detection in industrial environments, and Ren et al. (2024) improved YOLOv8 for autonomous driving by incorporating attention mechanisms and an enhanced IoU-based loss to boost road object detection performance. Although effective, these methods are not specifically designed for high-speed motion scenarios and are rarely validated on real mobile devices.</p>
      <p id="d2e178">To address these issues, we propose RIL-YOLO, a lightweight object detection framework optimized for high-speed dynamic scenarios and mobile deployment.  The acronym “RIL” represents the three core improvements introduced in this work: the re-parameterized shared convolutional detection head (RSCD), the inner-CIoU (complete intersection over union) loss, and the LAMP (layer-adaptive sparsity for the magnitude-based pruning) pruning strategy. Experimental results based on a kart-racing dataset and real mobile devices demonstrate that RIL-YOLO effectively balances detection accuracy and real-time efficiency under resource-constrained conditions.</p>
</sec>
<sec id="Ch1.S2">
  <label>2</label><title>Materials and methods</title>
      <p id="d2e189">YOLOv8 is an object detection model framework released by Ultralytics in January 2023. Its architecture consists of three main components: a backbone network, a neck network, and a detection head, as illustrated in Fig. 1. The backbone is built upon a CSPDarknet53-based architecture and incorporates C2f and SPPF (spatial pyramid pooling–fast) modules to enhance feature extraction capability and multi-scale information aggregation. The neck network combines feature pyramid network (FPN) and path aggregation network (PANet) structures to achieve bidirectional fusion of multi-scale features, thereby improving detection performance for objects of varying sizes. The detection head adopts an anchor-free design with a decoupled structure, in which classification and regression tasks are handled separately, leading to improved detection accuracy and training stability.</p>

      <fig id="F1" specific-use="star"><label>Figure 1</label><caption><p id="d2e194">Structure of YOLOv8.</p></caption>
        <graphic xlink:href="https://ms.copernicus.org/articles/17/371/2026/ms-17-371-2026-f01.png"/>

      </fig>

<sec id="Ch1.S2.SS1">
  <label>2.1</label><title>Improved YOLOv8</title>
<sec id="Ch1.S2.SS1.SSS1">
  <label>2.1.1</label><title>Data augmentation</title>
      <p id="d2e217">We adopt a combination of online and offline data augmentation strategies to enhance the generalization capability of the model. The online augmentation techniques employed in this work include hue adjustment, saturation adjustment, brightness adjustment, image rotation, image translation, image scaling, and mosaic augmentation.</p>
      <p id="d2e220">Offline data augmentation is conducted prior to training by preprocessing the dataset and generating augmented images that are stored locally (Kaur et al., 2021). Its primary objective is to simulate motion blur caused by camera shake or high-speed target motion in real kart-racing scenarios. Such motion blur often leads to indistinct object boundaries and loss of fine-grained details, thereby increasing detection difficulty. To address this issue, motion-blur-augmented samples are incorporated into the training set, improving the robustness and recognition performance of the model on blurred images.</p>
</sec>
<sec id="Ch1.S2.SS1.SSS2">
  <label>2.1.2</label><title>Rep shared convolutional detection head</title>
      <p id="d2e231">In YOLOv8, the detection head processes multi-scale feature maps using three independent decoupled branches, which introduces considerable parameter redundancy and limits inference efficiency on resource-constrained devices.</p>
      <p id="d2e234">To address this issue, we propose a lightweight shared convolutional detection head (LSCD), which shares convolutional layers across different feature scales to reduce parameters and computational cost (Wang et al., 2024c). As illustrated in Fig. 2, feature maps from P3, P4, and P5 are first aligned by a <inline-formula><mml:math id="M2" display="inline"><mml:mrow><mml:mn mathvariant="normal">1</mml:mn><mml:mo>×</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow></mml:math></inline-formula> convolution, followed by two shared <inline-formula><mml:math id="M3" display="inline"><mml:mrow><mml:mn mathvariant="normal">3</mml:mn><mml:mo>×</mml:mo><mml:mn mathvariant="normal">3</mml:mn></mml:mrow></mml:math></inline-formula> convolutional layers for joint feature extraction. The shared features are then fed into separate classification and regression branches, where a scale layer preserves scale-specific adaptability.</p>

      <fig id="F2" specific-use="star"><label>Figure 2</label><caption><p id="d2e263">LSCD module.</p></caption>
            <graphic xlink:href="https://ms.copernicus.org/articles/17/371/2026/ms-17-371-2026-f02.png"/>

          </fig>

      <p id="d2e273">To further enhance representation capability without increasing inference overhead, a diverse branch block (DBB) is incorporated into the shared convolutional layers. DBB introduces multi-branch structures during training to enrich feature representation and is re-parameterized into a single convolution during inference (Ding et al., 2021). The resulting re-parameterized shared convolutional detection head (RSCD) achieves improved detection performance while maintaining high inference efficiency, as shown in Figs. 3 and 4.</p>

      <fig id="F3" specific-use="star"><label>Figure 3</label><caption><p id="d2e278">DBB module.</p></caption>
            <graphic xlink:href="https://ms.copernicus.org/articles/17/371/2026/ms-17-371-2026-f03.png"/>

          </fig>

      <fig id="F4" specific-use="star"><label>Figure 4</label><caption><p id="d2e289">RSCD module.</p></caption>
            <graphic xlink:href="https://ms.copernicus.org/articles/17/371/2026/ms-17-371-2026-f04.png"/>

          </fig>

</sec>
<sec id="Ch1.S2.SS1.SSS3">
  <label>2.1.3</label><title>Improved loss function</title>
      <p id="d2e306">Bounding-box regression plays a key role in detection accuracy. In YOLOv8, localization is optimized using CIoU loss, which combines overlap, center distance, and aspect ratio penalties. However, CIoU may converge slowly and shows limited adaptability because gradient updates for high-IoU and low-IoU samples are often unbalanced. The calculation formula is shown in Eq. (1):

              <disp-formula id="Ch1.E1" content-type="numbered"><label>1</label><mml:math id="M4" display="block"><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mtext>CIoU</mml:mtext></mml:msub><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn><mml:mo>-</mml:mo><mml:mtext>IoU</mml:mtext><mml:mo>+</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mrow><mml:msup><mml:mi mathvariant="italic">ρ</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup><mml:mo>(</mml:mo><mml:mi>b</mml:mi><mml:mo>,</mml:mo><mml:msup><mml:mi>b</mml:mi><mml:mi mathvariant="normal">g</mml:mi></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:msup><mml:mi>c</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:mfrac></mml:mstyle><mml:mo>+</mml:mo><mml:mi mathvariant="italic">α</mml:mi><mml:mi mathvariant="italic">υ</mml:mi><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>

            where IoU represents the intersection-over-union between the predicted box and the ground-truth box, measuring the degree of overlap between the two boxes; <inline-formula><mml:math id="M5" display="inline"><mml:mrow><mml:msup><mml:mi mathvariant="italic">ρ</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup><mml:mo>(</mml:mo><mml:mi>b</mml:mi><mml:mo>,</mml:mo><mml:msup><mml:mi>b</mml:mi><mml:mi mathvariant="normal">g</mml:mi></mml:msup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> is the squared Euclidean distance between the center point <inline-formula><mml:math id="M6" display="inline"><mml:mi>b</mml:mi></mml:math></inline-formula> of the predicted box and the center point <inline-formula><mml:math id="M7" display="inline"><mml:mrow><mml:msup><mml:mi>b</mml:mi><mml:mi mathvariant="normal">g</mml:mi></mml:msup></mml:mrow></mml:math></inline-formula> of the ground-truth box; <inline-formula><mml:math id="M8" display="inline"><mml:mi>c</mml:mi></mml:math></inline-formula> is the diagonal length of the smallest enclosing rectangle covering both the predicted box and the ground-truth box; <inline-formula><mml:math id="M9" display="inline"><mml:mi mathvariant="italic">α</mml:mi></mml:math></inline-formula> is a weighting factor used to balance the consistency between IoU and aspect ratio; and <inline-formula><mml:math id="M10" display="inline"><mml:mi mathvariant="italic">υ</mml:mi></mml:math></inline-formula> is a penalty term used to measure the consistency of width-to-height ratios.</p>
      <p id="d2e429">To improve convergence and generalization, we introduce inner-IoU, where auxiliary bounding boxes are generated by scaling the original predicted and ground-truth boxes with a factor ratio (Zhang et al., 2023). Specifically, smaller auxiliary boxes are used for high-IoU samples to strengthen gradients, while larger auxiliary boxes are used for low-IoU samples to enlarge the effective optimization region and avoid vanishing gradients. Intuitively, inner-IoU improves gradient stability by adaptively adjusting the effective overlap region used for regression. For high-IoU samples, shrinking the auxiliary bounding boxes increases the sensitivity of the loss to small localization errors, preventing gradient vanishing. For low-IoU samples, enlarging the auxiliary region smooths the optimization landscape and avoids excessively large or unstable gradient updates. As a result, gradient magnitudes become more balanced across samples of different qualities, leading to more stable convergence during training. Replacing IoU in CIoU with inner-IoU yields the proposed inner-CIoU loss, leading to more stable training and improved localization performance. The mechanism of inner-IoU is illustrated in Fig. 5, and the calculation formulas are shown as follows.

                  <disp-formula specific-use="align" content-type="numbered"><mml:math id="M11" display="block"><mml:mtable displaystyle="true"><mml:mlabeledtr id="Ch1.E2"><mml:mtd><mml:mtext>2</mml:mtext></mml:mtd><mml:mtd><mml:mrow><mml:mstyle displaystyle="true" class="stylechange"/><mml:msubsup><mml:mi>b</mml:mi><mml:mi>l</mml:mi><mml:mi>g</mml:mi></mml:msubsup></mml:mrow></mml:mtd><mml:mtd><mml:mrow><mml:mstyle displaystyle="true" class="stylechange"/><mml:mo>=</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mi>c</mml:mi><mml:mi>g</mml:mi></mml:msubsup><mml:mo>-</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mrow><mml:msup><mml:mi>w</mml:mi><mml:mi>g</mml:mi></mml:msup><mml:mo>×</mml:mo><mml:mtext>ratio</mml:mtext></mml:mrow><mml:mn mathvariant="normal">2</mml:mn></mml:mfrac></mml:mstyle><mml:mo>,</mml:mo><mml:mspace linebreak="nobreak" width="1em"/><mml:msubsup><mml:mi>b</mml:mi><mml:mi>r</mml:mi><mml:mi>g</mml:mi></mml:msubsup><mml:mo>=</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mi>c</mml:mi><mml:mi>g</mml:mi></mml:msubsup><mml:mo>+</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mrow><mml:msup><mml:mi>w</mml:mi><mml:mi>g</mml:mi></mml:msup><mml:mo>×</mml:mo><mml:mtext>ratio</mml:mtext></mml:mrow><mml:mn mathvariant="normal">2</mml:mn></mml:mfrac></mml:mstyle></mml:mrow></mml:mtd></mml:mlabeledtr><mml:mlabeledtr id="Ch1.E3"><mml:mtd><mml:mtext>3</mml:mtext></mml:mtd><mml:mtd><mml:mrow><mml:mstyle displaystyle="true" class="stylechange"/><mml:msubsup><mml:mi>b</mml:mi><mml:mi>t</mml:mi><mml:mi>g</mml:mi></mml:msubsup></mml:mrow></mml:mtd><mml:mtd><mml:mrow><mml:mstyle class="stylechange" displaystyle="true"/><mml:mo>=</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>g</mml:mi></mml:msubsup><mml:mo>-</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mrow><mml:msup><mml:mi>h</mml:mi><mml:mi>g</mml:mi></mml:msup><mml:mo>×</mml:mo><mml:mtext>ratio</mml:mtext></mml:mrow><mml:mn mathvariant="normal">2</mml:mn></mml:mfrac></mml:mstyle><mml:mo>,</mml:mo><mml:mspace width="1em" linebreak="nobreak"/><mml:msubsup><mml:mi>b</mml:mi><mml:mi>b</mml:mi><mml:mi mathvariant="normal">g</mml:mi></mml:msubsup><mml:mo>=</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mi>c</mml:mi><mml:mi>g</mml:mi></mml:msubsup><mml:mo>+</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mrow><mml:msup><mml:mi>h</mml:mi><mml:mi>g</mml:mi></mml:msup><mml:mo>×</mml:mo><mml:mtext>ratio</mml:mtext></mml:mrow><mml:mn mathvariant="normal">2</mml:mn></mml:mfrac></mml:mstyle></mml:mrow></mml:mtd></mml:mlabeledtr><mml:mlabeledtr id="Ch1.E4"><mml:mtd><mml:mtext>4</mml:mtext></mml:mtd><mml:mtd><mml:mrow><mml:mstyle class="stylechange" displaystyle="true"/><mml:msub><mml:mi>b</mml:mi><mml:mi>l</mml:mi></mml:msub></mml:mrow></mml:mtd><mml:mtd><mml:mrow><mml:mstyle displaystyle="true" class="stylechange"/><mml:mo>=</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>c</mml:mi></mml:msub><mml:mo>-</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mrow><mml:mi>w</mml:mi><mml:mo>×</mml:mo><mml:mtext>ratio</mml:mtext></mml:mrow><mml:mn mathvariant="normal">2</mml:mn></mml:mfrac></mml:mstyle><mml:mo>,</mml:mo><mml:mspace width="1em" linebreak="nobreak"/><mml:msub><mml:mi>b</mml:mi><mml:mi>r</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>c</mml:mi></mml:msub><mml:mo>+</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mrow><mml:mi>w</mml:mi><mml:mo>×</mml:mo><mml:mtext>ratio</mml:mtext></mml:mrow><mml:mn mathvariant="normal">2</mml:mn></mml:mfrac></mml:mstyle></mml:mrow></mml:mtd></mml:mlabeledtr><mml:mlabeledtr id="Ch1.E5"><mml:mtd><mml:mtext>5</mml:mtext></mml:mtd><mml:mtd><mml:mrow><mml:mstyle class="stylechange" displaystyle="true"/><mml:msub><mml:mi>b</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow></mml:mtd><mml:mtd><mml:mrow><mml:mstyle class="stylechange" displaystyle="true"/><mml:mo>=</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>c</mml:mi></mml:msub><mml:mo>-</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mrow><mml:mi>h</mml:mi><mml:mo>×</mml:mo><mml:mtext>ratio</mml:mtext></mml:mrow><mml:mn mathvariant="normal">2</mml:mn></mml:mfrac></mml:mstyle><mml:mo>,</mml:mo><mml:mspace width="1em" linebreak="nobreak"/><mml:msub><mml:mi>b</mml:mi><mml:mi>b</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>c</mml:mi></mml:msub><mml:mo>+</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mrow><mml:mi>h</mml:mi><mml:mo>×</mml:mo><mml:mtext>ratio</mml:mtext></mml:mrow><mml:mn mathvariant="normal">2</mml:mn></mml:mfrac></mml:mstyle></mml:mrow></mml:mtd></mml:mlabeledtr></mml:mtable></mml:math></disp-formula>

                  <disp-formula specific-use="align" content-type="numbered"><mml:math id="M12" display="block"><mml:mtable displaystyle="true"><mml:mlabeledtr id="Ch1.E6"><mml:mtd><mml:mtext>6</mml:mtext></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" class="stylechange"/></mml:mtd><mml:mtd><mml:mrow><mml:mstyle displaystyle="true" class="stylechange"/><mml:mtable columnspacing="1em" rowspacing="0.2ex" class="aligned" displaystyle="true" columnalign="right left"><mml:mtr><mml:mtd><mml:mrow><mml:mstyle displaystyle="true" class="stylechange"/><mml:mtext>inter</mml:mtext></mml:mrow></mml:mtd><mml:mtd><mml:mrow><mml:mstyle class="stylechange" displaystyle="true"/><mml:mo>=</mml:mo><mml:mo>(</mml:mo><mml:mo movablelimits="false">min⁡</mml:mo><mml:mo>(</mml:mo><mml:msubsup><mml:mi>b</mml:mi><mml:mi>r</mml:mi><mml:mi>g</mml:mi></mml:msubsup><mml:msub><mml:mi>b</mml:mi><mml:mi>r</mml:mi></mml:msub><mml:mo>)</mml:mo><mml:mo>-</mml:mo><mml:mo movablelimits="false">max⁡</mml:mo><mml:mo>(</mml:mo><mml:msubsup><mml:mi>b</mml:mi><mml:mi>l</mml:mi><mml:mi>g</mml:mi></mml:msubsup><mml:msub><mml:mi>b</mml:mi><mml:mi>r</mml:mi></mml:msub><mml:mo>)</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mstyle class="stylechange" displaystyle="true"/></mml:mtd><mml:mtd><mml:mrow><mml:mstyle displaystyle="true" class="stylechange"/><mml:mspace width="1em" linebreak="nobreak"/><mml:mo>×</mml:mo><mml:mo>(</mml:mo><mml:mo movablelimits="false">min⁡</mml:mo><mml:mo>(</mml:mo><mml:msubsup><mml:mi>b</mml:mi><mml:mi>b</mml:mi><mml:mi>g</mml:mi></mml:msubsup><mml:msub><mml:mi>b</mml:mi><mml:mi>b</mml:mi></mml:msub><mml:mo>)</mml:mo><mml:mo>-</mml:mo><mml:mo movablelimits="false">max⁡</mml:mo><mml:mo>(</mml:mo><mml:msubsup><mml:mi>b</mml:mi><mml:mi>t</mml:mi><mml:mi>g</mml:mi></mml:msubsup><mml:msub><mml:mi>b</mml:mi><mml:mi>t</mml:mi></mml:msub><mml:mo>)</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:mrow></mml:mtd></mml:mlabeledtr><mml:mlabeledtr id="Ch1.E7"><mml:mtd><mml:mtext>7</mml:mtext></mml:mtd><mml:mtd><mml:mstyle class="stylechange" displaystyle="true"/></mml:mtd><mml:mtd><mml:mrow><mml:mstyle class="stylechange" displaystyle="true"/><mml:mtable class="aligned" rowspacing="0.2ex" columnspacing="1em" displaystyle="true" columnalign="right left"><mml:mtr><mml:mtd><mml:mrow><mml:mstyle displaystyle="true" class="stylechange"/><mml:mtext>union</mml:mtext></mml:mrow></mml:mtd><mml:mtd><mml:mrow><mml:mstyle displaystyle="true" class="stylechange"/><mml:mo>=</mml:mo><mml:mo>(</mml:mo><mml:msup><mml:mi>w</mml:mi><mml:mi>g</mml:mi></mml:msup><mml:mo>×</mml:mo><mml:msup><mml:mi>h</mml:mi><mml:mi>g</mml:mi></mml:msup><mml:mo>)</mml:mo><mml:mo>×</mml:mo><mml:mo>(</mml:mo><mml:mtext>ratio</mml:mtext><mml:msup><mml:mo>)</mml:mo><mml:mn mathvariant="normal">2</mml:mn></mml:msup><mml:mo>+</mml:mo><mml:mo>(</mml:mo><mml:mi>w</mml:mi><mml:mo>×</mml:mo><mml:mi>h</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mstyle displaystyle="true" class="stylechange"/></mml:mtd><mml:mtd><mml:mrow><mml:mstyle displaystyle="true" class="stylechange"/><mml:mspace width="1em" linebreak="nobreak"/><mml:mo>×</mml:mo><mml:mo>(</mml:mo><mml:mtext>ratio</mml:mtext><mml:msup><mml:mo>)</mml:mo><mml:mn mathvariant="normal">2</mml:mn></mml:msup><mml:mo>-</mml:mo><mml:mtext>inter</mml:mtext></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:mrow></mml:mtd></mml:mlabeledtr><mml:mlabeledtr id="Ch1.E8"><mml:mtd><mml:mtext>8</mml:mtext></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" class="stylechange"/></mml:mtd><mml:mtd><mml:mrow><mml:mstyle displaystyle="true" class="stylechange"/><mml:msup><mml:mtext>IoU</mml:mtext><mml:mtext>inner</mml:mtext></mml:msup><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mtext>inter</mml:mtext><mml:mtext>union</mml:mtext></mml:mfrac></mml:mstyle></mml:mrow></mml:mtd></mml:mlabeledtr><mml:mlabeledtr id="Ch1.E9"><mml:mtd><mml:mtext>9</mml:mtext></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" class="stylechange"/></mml:mtd><mml:mtd><mml:mrow><mml:mstyle displaystyle="true" class="stylechange"/><mml:msub><mml:mi>L</mml:mi><mml:mtext>inner-CIoU</mml:mtext></mml:msub><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn><mml:mo>-</mml:mo><mml:msup><mml:mtext>IoU</mml:mtext><mml:mtext>inner</mml:mtext></mml:msup><mml:mo>+</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mrow><mml:msup><mml:mi mathvariant="italic">ρ</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup><mml:mo>(</mml:mo><mml:mi>b</mml:mi><mml:mo>,</mml:mo><mml:msup><mml:mi>b</mml:mi><mml:mi mathvariant="normal">g</mml:mi></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:msup><mml:mi>c</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:mfrac></mml:mstyle><mml:mo>+</mml:mo><mml:mi mathvariant="italic">α</mml:mi><mml:mi mathvariant="italic">ν</mml:mi></mml:mrow></mml:mtd></mml:mlabeledtr></mml:mtable></mml:math></disp-formula></p><fig id="F5" specific-use="star"><label>Figure 5</label><caption><p id="d2e1010">Inner-IoU schematic.</p></caption>
            <graphic xlink:href="https://ms.copernicus.org/articles/17/371/2026/ms-17-371-2026-f05.png"/>

          </fig>

</sec>
<sec id="Ch1.S2.SS1.SSS4">
  <label>2.1.4</label><title>Model pruning using the LAMP algorithm</title>
      <p id="d2e1027">To enhance the performance of YOLOv8n in the go-kart detection task while reducing computational cost, we introduce an adaptive pruning strategy known as layer-adaptive sparsity for the magnitude-based pruning (LAMP). LAMP dynamically prunes parameters across different layers based on layer-wise characteristics and importance instead of applying a uniform global pruning ratio (Lee et al., 2020). The pruning ratio of each layer is determined according to its contribution to overall model performance, enabling effective model compression while preserving detection accuracy.</p>
      <p id="d2e1030">LAMP proposes a novel global pruning-importance scoring method, which quantitatively measures the contribution of each layer to the network. In the LAMP, each layer of the model is first analyzed to evaluate the importance of its parameters and neurons. The formulation is presented as Eq. (10). Subsequently, the weights are ranked according to their LAMP scores, and pruning begins with those having the lowest scores. After each pruning iteration, the LAMP scores of the remaining weights are recalculated to ensure that parameters contributing less to model performance are progressively removed. This iterative process continues until the weights in each layer are compressed to a predefined pruning ratio.

              <disp-formula id="Ch1.E10" content-type="numbered"><label>10</label><mml:math id="M13" display="block"><mml:mrow><mml:mtext>Score</mml:mtext><mml:mo>(</mml:mo><mml:mi>u</mml:mi><mml:mo>;</mml:mo><mml:mi>W</mml:mi><mml:mo>)</mml:mo><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mrow><mml:mo>(</mml:mo><mml:mi>W</mml:mi><mml:mo>[</mml:mo><mml:mi>u</mml:mi><mml:mo>]</mml:mo><mml:msup><mml:mo>)</mml:mo><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow><mml:mrow><mml:msub><mml:mo>∑</mml:mo><mml:mrow><mml:mi>v</mml:mi><mml:mo>≥</mml:mo><mml:mi>u</mml:mi></mml:mrow></mml:msub><mml:mo>(</mml:mo><mml:mi>W</mml:mi><mml:mo>[</mml:mo><mml:mi>v</mml:mi><mml:mo>]</mml:mo><mml:msup><mml:mo>)</mml:mo><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:mfrac></mml:mstyle></mml:mrow></mml:math></disp-formula>

            In the above, <inline-formula><mml:math id="M14" display="inline"><mml:mrow><mml:mi>W</mml:mi><mml:mo>[</mml:mo><mml:mi>u</mml:mi><mml:mo>]</mml:mo></mml:mrow></mml:math></inline-formula> denotes the weight of the <inline-formula><mml:math id="M15" display="inline"><mml:mi>u</mml:mi></mml:math></inline-formula>th connection. The numerator represents the square of the connection weight, while the denominator corresponds to the sum of the squared weights of all connections with lower importance.</p>
</sec>
</sec>
<sec id="Ch1.S2.SS2">
  <label>2.2</label><title>Mobile deployment</title>
      <p id="d2e1129">To meet the application requirements of real-time monitoring and object detection on mobile platforms in high-speed dynamic scenarios, the optimized YOLOv8 model is deployed on a smartphone platform for validation.  Considering the limited computational resources of mobile devices, the model is optimized using the NCNN (neural computing neural network) framework in combination with model quantization techniques. NCNN is a high-performance open-source neural network inference framework specifically designed for low-power and resource-constrained devices, providing fast and low-latency inference across multiple platforms through optimized memory management and computational pipelines. Ultimately, the optimized YOLOv8 model is successfully deployed on mobile devices, enabling a real-time monitoring application tailored for kart-racing scenarios.</p>
</sec>
</sec>
<sec id="Ch1.S3">
  <label>3</label><title>Results and discussions</title>
<sec id="Ch1.S3.SS1">
  <label>3.1</label><title>Dataset</title>
      <p id="d2e1148">To evaluate the performance of the proposed method in high-speed dynamic scenarios, a kart-racing object detection dataset was constructed using both online image collection and on-site video recordings. Key frames were extracted from recorded videos, and the final dataset consists of 2337 images. The constructed kart-racing dataset will be made available upon reasonable request to the corresponding author for research purposes.</p>
      <p id="d2e1151">All images were manually annotated using the LabelImg tool, covering nine object categories relevant to kart-racing scenes: whole kart body (KT), front head of the ego kart (KHF), left apex lane marking (ApexL), finish line (FL), right apex lane marking (ApexR), approaching kart facing the camera (KHInv), half kart head on the left side (KHL), half kart head on the right side (KHR), pedestrian (P), and gantry structure (LMJ).</p>
      <p id="d2e1154">The distribution of samples across categories is moderately imbalanced. The whole kart body (KT) and ego front head (KHF) contain relatively more instances due to their frequent occurrence in racing scenarios, while other categories such as pedestrians and gantry structures appear less frequently but remain sufficiently represented for training and evaluation.</p>
      <p id="d2e1157">The dataset includes images captured under different lighting conditions, primarily during daytime and late afternoon, thereby introducing moderate illumination variation.</p>
      <p id="d2e1161">The dataset was divided into training, validation, and test sets with a ratio of <inline-formula><mml:math id="M16" display="inline"><mml:mrow><mml:mn mathvariant="normal">8</mml:mn><mml:mo>:</mml:mo><mml:mn mathvariant="normal">1</mml:mn><mml:mo>:</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow></mml:math></inline-formula>.</p>
</sec>
<sec id="Ch1.S3.SS2">
  <label>3.2</label><title>Experimental environment configuration</title>
      <p id="d2e1188">The experimental environment required is shown in Table 1.</p>

<table-wrap id="T1"><label>Table 1</label><caption><p id="d2e1194">Experimental environment configuration.</p></caption><oasis:table frame="topbot"><oasis:tgroup cols="3">
     <oasis:colspec colnum="1" colname="col1" align="left"/>
     <oasis:colspec colnum="2" colname="col2" align="justify" colwidth="23mm"/>
     <oasis:colspec colnum="3" colname="col3" align="left"/>
     <oasis:thead>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1">Device</oasis:entry>
         <oasis:entry colname="col2" align="left">Experimental environment</oasis:entry>
         <oasis:entry colname="col3">Version</oasis:entry>
       </oasis:row>
     </oasis:thead>
     <oasis:tbody>
       <oasis:row>
         <oasis:entry colname="col1">PC</oasis:entry>
         <oasis:entry colname="col2" align="left">Operating system</oasis:entry>
         <oasis:entry colname="col3">Windows11</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1"/>
         <oasis:entry colname="col2" align="left">CPU</oasis:entry>
         <oasis:entry colname="col3">Intel i7–14700K</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1"/>
         <oasis:entry colname="col2" align="left">GPU</oasis:entry>
         <oasis:entry colname="col3">Nvidia RTX 4060ti 16G</oasis:entry>
       </oasis:row>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1"/>
         <oasis:entry colname="col2" align="left">Memory</oasis:entry>
         <oasis:entry colname="col3">64G</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Smartphone</oasis:entry>
         <oasis:entry colname="col2" align="left">SOC</oasis:entry>
         <oasis:entry colname="col3">Snapdragon 8 Gen 3</oasis:entry>
       </oasis:row>
     </oasis:tbody>
   </oasis:tgroup></oasis:table></table-wrap>

</sec>
<sec id="Ch1.S3.SS3">
  <label>3.3</label><title>Experimental parameter</title>
      <p id="d2e1291">All models were trained using the SGD (stochastic gradient descent) optimizer with an initial learning rate of 0.01, a momentum of 0.937, and a weight decay of 0.01. A cosine annealing learning-rate schedule was adopted during training. Input images were resized to <inline-formula><mml:math id="M17" display="inline"><mml:mrow><mml:mn mathvariant="normal">640</mml:mn><mml:mo>×</mml:mo><mml:mn mathvariant="normal">640</mml:mn></mml:mrow></mml:math></inline-formula>, and training was conducted for 150 epochs with a batch size of 64. All models were trained from scratch without using pre-trained weights to ensure fair comparison across different configurations.</p>
</sec>
<sec id="Ch1.S3.SS4">
  <label>3.4</label><title>Evaluation metrics</title>
      <p id="d2e1314">To comprehensively evaluate the performance of the RIL-YOLO model, multiple evaluation metrics were adopted. These include precision (<inline-formula><mml:math id="M18" display="inline"><mml:mi>P</mml:mi></mml:math></inline-formula>), recall (<inline-formula><mml:math id="M19" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula>), F1 score, mAP, Params, FLOPs (floating point operations per second), and FPS. Precision, recall, and F1 score evaluate classification performance, while mAP measures overall detection accuracy by computing the area under the precision–recall curve. Params represent the total number of learnable parameters in the model, reflecting model size and memory requirements. FLOPs indicate the computational complexity required for a single forward pass. FPS evaluates inference efficiency and real-time performance. Their calculation formulas are given as follows:

                <disp-formula specific-use="gather" content-type="numbered"><mml:math id="M20" display="block"><mml:mtable displaystyle="true"><mml:mlabeledtr id="Ch1.E11"><mml:mtd><mml:mtext>11</mml:mtext></mml:mtd><mml:mtd><mml:mrow><mml:mstyle class="stylechange" displaystyle="true"/><mml:mi>P</mml:mi><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mtext>TP</mml:mtext><mml:mrow><mml:mtext>TP</mml:mtext><mml:mo>+</mml:mo><mml:mtext>FP</mml:mtext></mml:mrow></mml:mfrac></mml:mstyle><mml:mo>,</mml:mo><mml:mi>R</mml:mi><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mtext>TP</mml:mtext><mml:mrow><mml:mtext>TP</mml:mtext><mml:mo>+</mml:mo><mml:mtext>FN</mml:mtext></mml:mrow></mml:mfrac></mml:mstyle><mml:mo>,</mml:mo></mml:mrow></mml:mtd></mml:mlabeledtr><mml:mlabeledtr id="Ch1.E12"><mml:mtd><mml:mtext>12</mml:mtext></mml:mtd><mml:mtd><mml:mrow><mml:mstyle class="stylechange" displaystyle="true"/><mml:mtext>F1</mml:mtext><mml:mo>=</mml:mo><mml:mn mathvariant="normal">2</mml:mn><mml:mo>×</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mrow><mml:mi>P</mml:mi><mml:mo>×</mml:mo><mml:mi>R</mml:mi></mml:mrow><mml:mrow><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>R</mml:mi></mml:mrow></mml:mfrac></mml:mstyle><mml:mo>,</mml:mo><mml:mtext>mAP</mml:mtext><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mn mathvariant="normal">1</mml:mn><mml:mi>n</mml:mi></mml:mfrac></mml:mstyle><mml:msubsup><mml:mo>∑</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow><mml:mi>n</mml:mi></mml:msubsup><mml:msub><mml:mtext>AP</mml:mtext><mml:mi>i</mml:mi></mml:msub><mml:mo>,</mml:mo></mml:mrow></mml:mtd></mml:mlabeledtr><mml:mlabeledtr id="Ch1.E13"><mml:mtd><mml:mtext>13</mml:mtext></mml:mtd><mml:mtd><mml:mrow><mml:mstyle class="stylechange" displaystyle="true"/><mml:mtext>FPS</mml:mtext><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:mfrac style="display"><mml:mn mathvariant="normal">1</mml:mn><mml:mrow><mml:msub><mml:mi>T</mml:mi><mml:mtext>inf</mml:mtext></mml:msub></mml:mrow></mml:mfrac></mml:mstyle><mml:mo>,</mml:mo></mml:mrow></mml:mtd></mml:mlabeledtr></mml:mtable></mml:math></disp-formula>

          where TP, FP, and FN are the true positive, false positive, and false negative, respectively; AP is numerically equal to the area enclosed by the <inline-formula><mml:math id="M21" display="inline"><mml:mi>P</mml:mi></mml:math></inline-formula>–<inline-formula><mml:math id="M22" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula> curve and the coordinate axis; and <inline-formula><mml:math id="M23" display="inline"><mml:mrow><mml:msub><mml:mi>T</mml:mi><mml:mtext>inf</mml:mtext></mml:msub></mml:mrow></mml:math></inline-formula> is the average inference time per image. FPS values were measured on the mobile device by continuously processing 500 frames. The reported FPS corresponds to the mean <inline-formula><mml:math id="M24" display="inline"><mml:mo>±</mml:mo></mml:math></inline-formula> standard deviation across repeated measurements. Other metrics were obtained on the PC platform.</p>
</sec>
<sec id="Ch1.S3.SS5">
  <label>3.5</label><title>Ablation experiment</title>
      <p id="d2e1506">To systematically analyze the contribution of each improvement, all optimization modules were incrementally integrated into the experiments. The experimental results are presented in Table 2, which provides a detailed comparison of the performance gains contributed by each module. Here, <inline-formula><mml:math id="M25" display="inline"><mml:mi>A</mml:mi></mml:math></inline-formula>, <inline-formula><mml:math id="M26" display="inline"><mml:mi>B</mml:mi></mml:math></inline-formula>, <inline-formula><mml:math id="M27" display="inline"><mml:mi>C</mml:mi></mml:math></inline-formula>, and <inline-formula><mml:math id="M28" display="inline"><mml:mi>D</mml:mi></mml:math></inline-formula> denote motion blur augmentation, inner-CIoU loss, the RSCD head, and LAMP <inline-formula><mml:math id="M29" display="inline"><mml:mrow><mml:mtext>pruning</mml:mtext><mml:mo>(</mml:mo><mml:mtext>ratio</mml:mtext><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1.5</mml:mn><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>, respectively.</p>

<table-wrap id="T2" specific-use="star"><label>Table 2</label><caption><p id="d2e1558">Ablation experiment.</p></caption><oasis:table frame="topbot"><oasis:tgroup cols="9">
     <oasis:colspec colnum="1" colname="col1" align="left"/>
     <oasis:colspec colnum="2" colname="col2" align="right"/>
     <oasis:colspec colnum="3" colname="col3" align="right"/>
     <oasis:colspec colnum="4" colname="col4" align="right"/>
     <oasis:colspec colnum="5" colname="col5" align="right"/>
     <oasis:colspec colnum="6" colname="col6" align="right"/>
     <oasis:colspec colnum="7" colname="col7" align="right"/>
     <oasis:colspec colnum="8" colname="col8" align="right"/>
     <oasis:colspec colnum="9" colname="col9" align="right"/>
     <oasis:thead>
       <oasis:row>
         <oasis:entry colname="col1">Model</oasis:entry>
         <oasis:entry colname="col2"><inline-formula><mml:math id="M30" display="inline"><mml:mi>P</mml:mi></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col3"><inline-formula><mml:math id="M31" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col4">F1</oasis:entry>
         <oasis:entry colname="col5">mAP@0.5</oasis:entry>
         <oasis:entry colname="col6">mAP@<inline-formula><mml:math id="M32" display="inline"><mml:mrow><mml:mo>[</mml:mo><mml:mn mathvariant="normal">0.5</mml:mn><mml:mo>:</mml:mo><mml:mn mathvariant="normal">0.95</mml:mn><mml:mo>]</mml:mo></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col7">Params</oasis:entry>
         <oasis:entry colname="col8">FLOPs</oasis:entry>
         <oasis:entry colname="col9">FPS</oasis:entry>
       </oasis:row>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1"/>
         <oasis:entry colname="col2">(%)</oasis:entry>
         <oasis:entry colname="col3">(%)</oasis:entry>
         <oasis:entry colname="col4">(%)</oasis:entry>
         <oasis:entry colname="col5">(%)</oasis:entry>
         <oasis:entry colname="col6">(%)</oasis:entry>
         <oasis:entry colname="col7">(M)</oasis:entry>
         <oasis:entry colname="col8">(G)</oasis:entry>
         <oasis:entry colname="col9">(mean <inline-formula><mml:math id="M33" display="inline"><mml:mo>±</mml:mo></mml:math></inline-formula> SD)</oasis:entry>
       </oasis:row>
     </oasis:thead>
     <oasis:tbody>
       <oasis:row>
         <oasis:entry colname="col1">YOLOv8n</oasis:entry>
         <oasis:entry colname="col2">83.4</oasis:entry>
         <oasis:entry colname="col3">90.7</oasis:entry>
         <oasis:entry colname="col4">86.9</oasis:entry>
         <oasis:entry colname="col5">90.9</oasis:entry>
         <oasis:entry colname="col6">58.8</oasis:entry>
         <oasis:entry colname="col7">3.01</oasis:entry>
         <oasis:entry colname="col8">8.1</oasis:entry>
         <oasis:entry colname="col9"><inline-formula><mml:math id="M34" display="inline"><mml:mrow><mml:mn mathvariant="normal">19.6</mml:mn><mml:mo>±</mml:mo><mml:mn mathvariant="normal">1.7</mml:mn></mml:mrow></mml:math></inline-formula></oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1"><inline-formula><mml:math id="M35" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mi>A</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col2">90.3</oasis:entry>
         <oasis:entry colname="col3">85.1</oasis:entry>
         <oasis:entry colname="col4">87.6</oasis:entry>
         <oasis:entry colname="col5">91.4</oasis:entry>
         <oasis:entry colname="col6">59.9</oasis:entry>
         <oasis:entry colname="col7">3.01</oasis:entry>
         <oasis:entry colname="col8">8.1</oasis:entry>
         <oasis:entry colname="col9"><inline-formula><mml:math id="M36" display="inline"><mml:mrow><mml:mn mathvariant="normal">19.1</mml:mn><mml:mo>±</mml:mo><mml:mn mathvariant="normal">1.0</mml:mn></mml:mrow></mml:math></inline-formula>  (<inline-formula><mml:math id="M37" display="inline"><mml:mrow><mml:mo>-</mml:mo><mml:mn mathvariant="normal">2</mml:mn><mml:mspace linebreak="nobreak" width="0.125em"/><mml:mi mathvariant="italic">%</mml:mi></mml:mrow></mml:math></inline-formula>)</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1"><inline-formula><mml:math id="M38" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mi>B</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col2">88.9</oasis:entry>
         <oasis:entry colname="col3">86.4</oasis:entry>
         <oasis:entry colname="col4">87.6</oasis:entry>
         <oasis:entry colname="col5">93.8</oasis:entry>
         <oasis:entry colname="col6">58</oasis:entry>
         <oasis:entry colname="col7">3.01</oasis:entry>
         <oasis:entry colname="col8">8.1</oasis:entry>
         <oasis:entry colname="col9"><inline-formula><mml:math id="M39" display="inline"><mml:mrow><mml:mn mathvariant="normal">19.9</mml:mn><mml:mo>±</mml:mo><mml:mn mathvariant="normal">0.9</mml:mn></mml:mrow></mml:math></inline-formula>  (<inline-formula><mml:math id="M40" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mn mathvariant="normal">2</mml:mn><mml:mspace width="0.125em" linebreak="nobreak"/><mml:mi mathvariant="italic">%</mml:mi></mml:mrow></mml:math></inline-formula>)</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1"><inline-formula><mml:math id="M41" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mi>C</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col2">88.8</oasis:entry>
         <oasis:entry colname="col3">82.9</oasis:entry>
         <oasis:entry colname="col4">85.7</oasis:entry>
         <oasis:entry colname="col5">88.7</oasis:entry>
         <oasis:entry colname="col6">57.1</oasis:entry>
         <oasis:entry colname="col7">2.36</oasis:entry>
         <oasis:entry colname="col8">6.5</oasis:entry>
         <oasis:entry colname="col9"><inline-formula><mml:math id="M42" display="inline"><mml:mrow><mml:mn mathvariant="normal">21.8</mml:mn><mml:mo>±</mml:mo><mml:mn mathvariant="normal">0.6</mml:mn></mml:mrow></mml:math></inline-formula>  (<inline-formula><mml:math id="M43" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mn mathvariant="normal">12</mml:mn><mml:mspace linebreak="nobreak" width="0.125em"/><mml:mi mathvariant="italic">%</mml:mi></mml:mrow></mml:math></inline-formula>)</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1"><inline-formula><mml:math id="M44" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mi>A</mml:mi><mml:mo>+</mml:mo><mml:mi>B</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col2">86</oasis:entry>
         <oasis:entry colname="col3">90.4</oasis:entry>
         <oasis:entry colname="col4">88.1</oasis:entry>
         <oasis:entry colname="col5">92.4</oasis:entry>
         <oasis:entry colname="col6">60.9</oasis:entry>
         <oasis:entry colname="col7">3.01</oasis:entry>
         <oasis:entry colname="col8">8.1</oasis:entry>
         <oasis:entry colname="col9"><inline-formula><mml:math id="M45" display="inline"><mml:mrow><mml:mn mathvariant="normal">19.6</mml:mn><mml:mo>±</mml:mo><mml:mn mathvariant="normal">0.8</mml:mn></mml:mrow></mml:math></inline-formula>  (<inline-formula><mml:math id="M46" display="inline"><mml:mrow><mml:mn mathvariant="normal">0</mml:mn><mml:mspace linebreak="nobreak" width="0.125em"/><mml:mi mathvariant="italic">%</mml:mi></mml:mrow></mml:math></inline-formula>)</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1"><inline-formula><mml:math id="M47" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mi>A</mml:mi><mml:mo>+</mml:mo><mml:mi>C</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col2">83.4</oasis:entry>
         <oasis:entry colname="col3">89.7</oasis:entry>
         <oasis:entry colname="col4">86.4</oasis:entry>
         <oasis:entry colname="col5">90.1</oasis:entry>
         <oasis:entry colname="col6">60</oasis:entry>
         <oasis:entry colname="col7">2.36</oasis:entry>
         <oasis:entry colname="col8">6.5</oasis:entry>
         <oasis:entry colname="col9"><inline-formula><mml:math id="M48" display="inline"><mml:mrow><mml:mn mathvariant="normal">20.9</mml:mn><mml:mo>±</mml:mo><mml:mn mathvariant="normal">0.8</mml:mn></mml:mrow></mml:math></inline-formula>  (<inline-formula><mml:math id="M49" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mn mathvariant="normal">7</mml:mn><mml:mspace linebreak="nobreak" width="0.125em"/><mml:mi mathvariant="italic">%</mml:mi></mml:mrow></mml:math></inline-formula>)</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1"><inline-formula><mml:math id="M50" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mi>B</mml:mi><mml:mo>+</mml:mo><mml:mi>C</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col2">88</oasis:entry>
         <oasis:entry colname="col3">83.5</oasis:entry>
         <oasis:entry colname="col4">85.7</oasis:entry>
         <oasis:entry colname="col5">91.4</oasis:entry>
         <oasis:entry colname="col6">55.9</oasis:entry>
         <oasis:entry colname="col7">2.36</oasis:entry>
         <oasis:entry colname="col8">6.5</oasis:entry>
         <oasis:entry colname="col9"><inline-formula><mml:math id="M51" display="inline"><mml:mrow><mml:mn mathvariant="normal">21.1</mml:mn><mml:mo>±</mml:mo><mml:mn mathvariant="normal">1.2</mml:mn></mml:mrow></mml:math></inline-formula>  (<inline-formula><mml:math id="M52" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mn mathvariant="normal">8</mml:mn><mml:mspace linebreak="nobreak" width="0.125em"/><mml:mi mathvariant="italic">%</mml:mi></mml:mrow></mml:math></inline-formula>)</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1"><inline-formula><mml:math id="M53" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mi>A</mml:mi><mml:mo>+</mml:mo><mml:mi>B</mml:mi><mml:mo>+</mml:mo><mml:mi>C</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col2">85.1</oasis:entry>
         <oasis:entry colname="col3">88.1</oasis:entry>
         <oasis:entry colname="col4">86.6</oasis:entry>
         <oasis:entry colname="col5">90.4</oasis:entry>
         <oasis:entry colname="col6">58.1</oasis:entry>
         <oasis:entry colname="col7">2.36</oasis:entry>
         <oasis:entry colname="col8">6.5</oasis:entry>
         <oasis:entry colname="col9"><inline-formula><mml:math id="M54" display="inline"><mml:mrow><mml:mn mathvariant="normal">21.4</mml:mn><mml:mo>±</mml:mo><mml:mn mathvariant="normal">1.3</mml:mn></mml:mrow></mml:math></inline-formula>  (<inline-formula><mml:math id="M55" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mn mathvariant="normal">9</mml:mn><mml:mspace width="0.125em" linebreak="nobreak"/><mml:mi mathvariant="italic">%</mml:mi></mml:mrow></mml:math></inline-formula>)</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1"><inline-formula><mml:math id="M56" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mi>A</mml:mi><mml:mo>+</mml:mo><mml:mi>B</mml:mi><mml:mo>+</mml:mo><mml:mi>C</mml:mi><mml:mo>+</mml:mo><mml:mi>D</mml:mi></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col2">85.6</oasis:entry>
         <oasis:entry colname="col3">88.2</oasis:entry>
         <oasis:entry colname="col4">86.9</oasis:entry>
         <oasis:entry colname="col5">93.7</oasis:entry>
         <oasis:entry colname="col6">61</oasis:entry>
         <oasis:entry colname="col7">0.51</oasis:entry>
         <oasis:entry colname="col8">3.8</oasis:entry>
         <oasis:entry colname="col9"><inline-formula><mml:math id="M57" display="inline"><mml:mrow><mml:mn mathvariant="normal">24.4</mml:mn><mml:mo>±</mml:mo><mml:mn mathvariant="normal">0.5</mml:mn></mml:mrow></mml:math></inline-formula>  (<inline-formula><mml:math id="M58" display="inline"><mml:mrow><mml:mo>+</mml:mo><mml:mn mathvariant="normal">25</mml:mn><mml:mspace linebreak="nobreak" width="0.125em"/><mml:mi mathvariant="italic">%</mml:mi></mml:mrow></mml:math></inline-formula>)</oasis:entry>
       </oasis:row>
     </oasis:tbody>
   </oasis:tgroup></oasis:table></table-wrap>

      <p id="d2e2261">The experimental results indicate that motion blur augmentation enhances robustness in relation to dynamic scenes, achieving a recall of 90.7 % when used alone. When combined with inner-CIoU and RSCD, precision and recall become more balanced, leading to a notable improvement in mAP@0.5. Inner-CIoU alone increases precision to 90.3 % but reduces recall to 85.1 %; however, introducing RSCD restores recall to above 88 % while maintaining high mAP@0.5, highlighting the importance of coordination between localization optimization and detection head design. RSCD further reduces model parameters by 21.6 % with minimal performance degradation, providing a solid basis for lightweight deployment. After applying LAMP pruning, the parameter count is further reduced to 0.51 <inline-formula><mml:math id="M59" display="inline"><mml:mrow class="unit"><mml:mi mathvariant="normal">M</mml:mi></mml:mrow></mml:math></inline-formula>, resulting in a 25 % increase in inference speed.</p>
      <p id="d2e2273">The combination of RSCD and inner-CIoU exhibits a synergistic effect. RSCD focuses on structural optimization by sharing convolutional parameters across feature scales, which improves feature consistency and reduces overfitting under limited model capacity. In contrast, Inner-CIoU addresses optimization dynamics by balancing gradient magnitudes for high- and low-quality samples. When combined, RSCD provides stronger and more compact feature representations, while inner-CIoU ensures stable and balanced regression updates. This complementary interaction leads to more robust convergence and improved overall performance.</p>
</sec>
<sec id="Ch1.S3.SS6">
  <label>3.6</label><title>Comparative experiment</title>
      <p id="d2e2284">To validate the overall performance advantages of the proposed improved algorithm, several representative lightweight models were selected for comparison, including YOLOv8-fastnet, YOLOv8-mobilenetv4, YOLOv5n, YOLOv9t (Wang et al., 2024b), YOLOv10n (Wang et al., 2024a), and YOLOv11n. Among them, YOLOv8-fastnet and YOLOv8-mobilenetv4 replace the original backbone of YOLOv8 with FastNet (Chen et al., 2023) and MobileNetV4 (Qin et al., 2024), respectively. Since the proposed method is built upon the YOLOv8 framework and primarily targets lightweight real-time deployment on mobile devices, the comparative analysis focuses on representative lightweight one-stage detectors. Two-stage detectors (e.g., faster R-CNN) are generally more computationally intensive and less suitable for mobile real-time applications and therefore were not included in this comparison. For fairness, all comparison models, including YOLOv11n, were implemented using the official Ultralytics implementation (version 8.3.50). The comparative experimental results are summarized in Table 3.</p>

<table-wrap id="T3" specific-use="star"><label>Table 3</label><caption><p id="d2e2290">Comparative experiment.</p></caption><oasis:table frame="topbot"><oasis:tgroup cols="8">
     <oasis:colspec colnum="1" colname="col1" align="left"/>
     <oasis:colspec colnum="2" colname="col2" align="right"/>
     <oasis:colspec colnum="3" colname="col3" align="right"/>
     <oasis:colspec colnum="4" colname="col4" align="right"/>
     <oasis:colspec colnum="5" colname="col5" align="right"/>
     <oasis:colspec colnum="6" colname="col6" align="right"/>
     <oasis:colspec colnum="7" colname="col7" align="right"/>
     <oasis:colspec colnum="8" colname="col8" align="right"/>
     <oasis:thead>
       <oasis:row>
         <oasis:entry colname="col1">Model</oasis:entry>
         <oasis:entry colname="col2"><inline-formula><mml:math id="M60" display="inline"><mml:mi>P</mml:mi></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col3"><inline-formula><mml:math id="M61" display="inline"><mml:mi>R</mml:mi></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col4">F1</oasis:entry>
         <oasis:entry colname="col5">mAP@0.5</oasis:entry>
         <oasis:entry colname="col6">mAP@<inline-formula><mml:math id="M62" display="inline"><mml:mrow><mml:mo>[</mml:mo><mml:mn mathvariant="normal">0.5</mml:mn><mml:mo>:</mml:mo><mml:mn mathvariant="normal">0.95</mml:mn><mml:mo>]</mml:mo></mml:mrow></mml:math></inline-formula></oasis:entry>
         <oasis:entry colname="col7">Params</oasis:entry>
         <oasis:entry colname="col8">FLOPs</oasis:entry>
       </oasis:row>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1"/>
         <oasis:entry colname="col2">(%)</oasis:entry>
         <oasis:entry colname="col3">(%)</oasis:entry>
         <oasis:entry colname="col4">(%)</oasis:entry>
         <oasis:entry colname="col5">(%)</oasis:entry>
         <oasis:entry colname="col6">(%)</oasis:entry>
         <oasis:entry colname="col7">(M)</oasis:entry>
         <oasis:entry colname="col8">(G)</oasis:entry>
       </oasis:row>
     </oasis:thead>
     <oasis:tbody>
       <oasis:row>
         <oasis:entry colname="col1">YOLOv8-mobilenetv4</oasis:entry>
         <oasis:entry colname="col2">80.9</oasis:entry>
         <oasis:entry colname="col3">84.9</oasis:entry>
         <oasis:entry colname="col4">82.9</oasis:entry>
         <oasis:entry colname="col5">84</oasis:entry>
         <oasis:entry colname="col6">51.6</oasis:entry>
         <oasis:entry colname="col7">5.70</oasis:entry>
         <oasis:entry colname="col8">22.6</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">YOLOv8-fastnet</oasis:entry>
         <oasis:entry colname="col2">87.7</oasis:entry>
         <oasis:entry colname="col3">84.3</oasis:entry>
         <oasis:entry colname="col4">86.0</oasis:entry>
         <oasis:entry colname="col5">88.7</oasis:entry>
         <oasis:entry colname="col6">56.7</oasis:entry>
         <oasis:entry colname="col7">4.17</oasis:entry>
         <oasis:entry colname="col8">10.7</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">YOLOv5n</oasis:entry>
         <oasis:entry colname="col2">85.5</oasis:entry>
         <oasis:entry colname="col3">80.7</oasis:entry>
         <oasis:entry colname="col4">83.0</oasis:entry>
         <oasis:entry colname="col5">88.1</oasis:entry>
         <oasis:entry colname="col6">56.2</oasis:entry>
         <oasis:entry colname="col7">2.50</oasis:entry>
         <oasis:entry colname="col8">7.1</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">YOLOv9t</oasis:entry>
         <oasis:entry colname="col2">88.9</oasis:entry>
         <oasis:entry colname="col3">88</oasis:entry>
         <oasis:entry colname="col4">88.4</oasis:entry>
         <oasis:entry colname="col5">92.5</oasis:entry>
         <oasis:entry colname="col6">58.9</oasis:entry>
         <oasis:entry colname="col7">1.97</oasis:entry>
         <oasis:entry colname="col8">7.6</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">YOLOv10n</oasis:entry>
         <oasis:entry colname="col2">75.4</oasis:entry>
         <oasis:entry colname="col3">77.1</oasis:entry>
         <oasis:entry colname="col4">76.2</oasis:entry>
         <oasis:entry colname="col5">85.7</oasis:entry>
         <oasis:entry colname="col6">53.9</oasis:entry>
         <oasis:entry colname="col7">2.70</oasis:entry>
         <oasis:entry colname="col8">8.2</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">YOLOv11n</oasis:entry>
         <oasis:entry colname="col2">86.5</oasis:entry>
         <oasis:entry colname="col3">85.4</oasis:entry>
         <oasis:entry colname="col4">85.9</oasis:entry>
         <oasis:entry colname="col5">92.4</oasis:entry>
         <oasis:entry colname="col6">57.6</oasis:entry>
         <oasis:entry colname="col7">2.58</oasis:entry>
         <oasis:entry colname="col8">6.3</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">RIL-YOLO(ours)</oasis:entry>
         <oasis:entry colname="col2">85.6</oasis:entry>
         <oasis:entry colname="col3">88.2</oasis:entry>
         <oasis:entry colname="col4">86.9</oasis:entry>
         <oasis:entry colname="col5">93.7</oasis:entry>
         <oasis:entry colname="col6">61</oasis:entry>
         <oasis:entry colname="col7">0.51</oasis:entry>
         <oasis:entry colname="col8">3.8</oasis:entry>
       </oasis:row>
     </oasis:tbody>
   </oasis:tgroup></oasis:table></table-wrap>

      <p id="d2e2598">The experimental results demonstrate that our method outperforms all comparison models on both mAP@0.5 and mAP@<inline-formula><mml:math id="M63" display="inline"><mml:mrow><mml:mo>[</mml:mo><mml:mn mathvariant="normal">0.5</mml:mn><mml:mo>:</mml:mo><mml:mn mathvariant="normal">0.95</mml:mn><mml:mo>]</mml:mo></mml:mrow></mml:math></inline-formula>, achieving improvements of 1.2 % and 2.1 % over the second-best model, YOLOv9t, respectively, which verifies its strong adaptability to complex scenarios. With respect to model efficiency, the parameter count of our method is only 25.9 % of that of YOLOv9t, while the FLOPs are reduced by 39.7 % compared with YOLOv11n, demonstrating an effective balance between high accuracy and low computational cost.</p>
</sec>
</sec>
<sec id="Ch1.S4" sec-type="conclusions">
  <label>4</label><title>Conclusions</title>
      <p id="d2e2626">We propose RIL-YOLO, a lightweight object detection method for real-time visual perception in high-speed mechanical systems, using kart racing as a representative dynamic scenario. Based on the YOLOv8 framework, the model is systematically optimized through motion blur data augmentation, an RSCD head, an inner-CIoU loss function, and LAMP-based pruning. Experimental results show that, relative to the YOLOv8n baseline, the optimized model improves mAP@0.5 and mAP@<inline-formula><mml:math id="M64" display="inline"><mml:mrow><mml:mo>[</mml:mo><mml:mn mathvariant="normal">0.5</mml:mn><mml:mo>:</mml:mo><mml:mn mathvariant="normal">0.95</mml:mn><mml:mo>]</mml:mo></mml:mrow></mml:math></inline-formula> by 2.8 % and 2.2 %, respectively, while reducing parameters by approximately 83 %, lowering computational cost by about 53 %, and increasing mobile inference speed by around 25 %, achieving a favorable balance between accuracy and real-time performance on resource-constrained platforms.</p>
      <p id="d2e2645">Furthermore, in practical mobile deployment scenarios, reduced model complexity contributes not only to faster inference but also to lower power consumption and improved thermal stability during prolonged operation. By alleviating computational load, the proposed method helps maintain stable latency and mitigate performance degradation caused by thermal throttling on smartphones or embedded devices. During practical testing under continuous real-time monitoring, no obvious degradation in inference speed was observed, indicating stable runtime behavior under typical operating conditions. These characteristics enhance its suitability for continuous real-time monitoring applications in high-speed dynamic environments.</p>
      <p id="d2e2648">Despite the demonstrated improvements, several limitations remain. First, the dataset size is relatively limited compared to large-scale benchmarks, which may restrict generalization to more diverse environments. Second, although mobile inference stability was verified under typical operating conditions, comprehensive long-duration thermal and power consumption analysis was not systematically conducted. In addition, extreme lighting variations or highly crowded racing scenarios may still pose challenges to detection robustness. Future work will focus on expanding the dataset diversity and conducting more extensive real-world deployment evaluations.</p>
</sec>

      
      </body>
    <back><notes notes-type="codedataavailability"><title>Code and data availability</title>

      <p id="d2e2656">All the code and data used in this paper can be obtained upon request to the corresponding author.</p>
  </notes><notes notes-type="authorcontribution"><title>Author contributions</title>

      <p id="d2e2662">AS conceived the study; provided the key ideas; and performed the method design, the data collection, the paper revision, and the supervision.  FX designed the RIL-YOLO and constructed the paper. YZ provided the test program. XZ provided the key suggestions on the experiments. JZ provided the key suggestions on the research questions.</p>
  </notes><notes notes-type="competinginterests"><title>Competing interests</title>

      <p id="d2e2668">The contact author has declared that none of the authors has any competing interests.</p>
  </notes><notes notes-type="disclaimer"><title>Disclaimer</title>

      <p id="d2e2674">Publisher's note: Copernicus Publications remains neutral with regard to jurisdictional claims made in the text, published maps, institutional affiliations, or any other geographical representation in this paper. The authors bear the ultimate responsibility for providing appropriate place names. Views expressed in the text are those of the authors and do not necessarily reflect the views of the publisher.</p>
  </notes><notes notes-type="financialsupport"><title>Financial support</title>

      <p id="d2e2680">This research has been supported by the National Natural Science Foundation of China (grant no. 42301266) and the Natural Science Foundation of Shandong Province (grant no. ZR2023QD005).</p>
  </notes><notes notes-type="reviewstatement"><title>Review statement</title>

      <p id="d2e2686">This paper was edited by Benliang Zhu and reviewed by Marco Aiello and two anonymous referees.</p>
  </notes><ref-list>
    <title>References</title>

      <ref id="bib1.bib1"><label>1</label><mixed-citation>Anggrainy, R., Yoga, N. G., Wiyono, A., Putri, D. M., Wahyudi, Z. T., and Septiyan, Y. A.: Unveiling the Future of Safety: Cutting-Edge Simulation Testing of e-Kart Bumpers, J. Phys. Conf. Ser., 012096, <ext-link xlink:href="https://doi.org/10.1088/1742-6596/2866/1/012096" ext-link-type="DOI">10.1088/1742-6596/2866/1/012096</ext-link>, 2024.</mixed-citation></ref>
      <ref id="bib1.bib2"><label>2</label><mixed-citation>Chen, C., Li, J., Shuai, Z., Wang, Y., and Wang, Y.: A lightweight optimization framework for real-time pedestrian detection in dense and occluded scenes, Mech. Sci., 16, 877–886, <ext-link xlink:href="https://doi.org/10.5194/ms-16-877-2025" ext-link-type="DOI">10.5194/ms-16-877-2025</ext-link>, 2025.</mixed-citation></ref>
      <ref id="bib1.bib3"><label>3</label><mixed-citation>Chen, J., Kao, S.-h., He, H., Zhuo, W., Wen, S., Lee, C.-H., and Chan, S.-H. G.: Run, don't walk: chasing higher FLOPS for faster neural networks, arXiv [preprint], <ext-link xlink:href="https://doi.org/10.48550/arXiv.2303.03667" ext-link-type="DOI">10.48550/arXiv.2303.03667</ext-link>, 2023.</mixed-citation></ref>
      <ref id="bib1.bib4"><label>4</label><mixed-citation>Ding, X., Zhang, X., Han, J., and Ding, G.: Diverse branch block: Building a convolution as an inception-like unit, in: 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021, IEEE, 10881–10890, <ext-link xlink:href="https://doi.org/10.1109/CVPR46437.2021.01074" ext-link-type="DOI">10.1109/CVPR46437.2021.01074</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bib5"><label>5</label><mixed-citation>Kaur, P., Khehra, B. S., and Mavi, E. B. S.: Data augmentation for object detection: A review, in: 2021 IEEE International Midwest Symposium on Circuits and Systems (MWSCAS), 537–543, <ext-link xlink:href="https://doi.org/10.1109/MWSCAS47672.2021.9531849" ext-link-type="DOI">10.1109/MWSCAS47672.2021.9531849</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bib6"><label>6</label><mixed-citation>Lee, J., Park, S., Mo, S., Ahn, S., and Shin, J.: Layer-adaptive sparsity for the magnitude-based pruning, arXiv [preprint], <ext-link xlink:href="https://doi.org/10.48550/arXiv.2010.07611" ext-link-type="DOI">10.48550/arXiv.2010.07611</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bib7"><label>7</label><mixed-citation>Lu, J. and Liu, Y.: A real-time and accurate detection approach for bucket teeth falling off based on improved YOLOX, Mech. Sci., 13, 979–990, <ext-link xlink:href="https://doi.org/10.5194/ms-13-979-2022" ext-link-type="DOI">10.5194/ms-13-979-2022</ext-link>, 2022.</mixed-citation></ref>
      <ref id="bib1.bib8"><label>8</label><mixed-citation>Matsumura, K., Yamakoshi, T., Yamakoshi, Y., and Rolfe, P.: The effect of competition on heart rate during kart driving: A field study, BMC Research Notes, 4, 342, <ext-link xlink:href="https://doi.org/10.1186/1756-0500-4-342" ext-link-type="DOI">10.1186/1756-0500-4-342</ext-link>, 2011.</mixed-citation></ref>
      <ref id="bib1.bib9"><label>9</label><mixed-citation>Qin, D., Leichner, C., Delakis, M., Fornoni, M., Luo, S., Yang, F., Wang, W., Banbury, C., Ye, C., and Akin, B.: MobileNetV4: Universal models for the mobile ecosystem, arXiv [preprint], <ext-link xlink:href="https://doi.org/10.48550/arXiv.2404.10518" ext-link-type="DOI">10.48550/arXiv.2404.10518</ext-link>, 2024. </mixed-citation></ref>
      <ref id="bib1.bib10"><label>10</label><mixed-citation>Redmon, J., Divvala, S., Girshick, R., and Farhadi, A.: You only look once: Unified, real-time object detection, arXiv [preprint], <ext-link xlink:href="https://doi.org/10.48550/arXiv.1506.02640" ext-link-type="DOI">10.48550/arXiv.1506.02640</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bib11"><label>11</label><mixed-citation>Ren, H., Jing, F., and Li, S.: DCW-YOLO: road object detection algorithms for  autonomous driving, IEEE Access, 13, 125676–125688, <ext-link xlink:href="https://doi.org/10.1109/ACCESS.2024.3364681" ext-link-type="DOI">10.1109/ACCESS.2024.3364681</ext-link>, 2024.</mixed-citation></ref>
      <ref id="bib1.bib12"><label>12</label><mixed-citation>Wang, A., Chen, H., Liu, L., Chen, K., Lin, Z., and Han, J.: Yolov10: Real-time end-to-end object detection, arXiv [preprint], <ext-link xlink:href="https://doi.org/10.48550/arXiv.2405.14458" ext-link-type="DOI">10.48550/arXiv.2405.14458</ext-link>, 2024a.</mixed-citation></ref>
      <ref id="bib1.bib13"><label>13</label><mixed-citation>Wang, C.-Y., Yeh, I.-H., and Mark Liao, H.-Y.: Yolov9: Learning what you want to learn using programmable gradient information, arXiv [preprint], <ext-link xlink:href="https://doi.org/10.48550/arXiv.2402.13616" ext-link-type="DOI">10.48550/arXiv.2402.13616</ext-link>, 2024b.</mixed-citation></ref>
      <ref id="bib1.bib14"><label>14</label><mixed-citation>Wang, H., Liu, X., Song, L.,Zhang, Y., Rong, X., and Wang, Y.: Research on a train safety driving method based on fusion of an incremental clustering algorithm and lightweight shared convolution, Sensors, 24, 4951, <ext-link xlink:href="https://doi.org/10.3390/s24154951" ext-link-type="DOI">10.3390/s24154951</ext-link>, 2024c.</mixed-citation></ref>
      <ref id="bib1.bib15"><label>15</label><mixed-citation>Zhang, H., Xu, C., and Zhang, S.: Inner-IoU: more effective intersection over union loss with auxiliary bounding box, arXiv [preprint], <ext-link xlink:href="https://doi.org/10.48550/arXiv.2311.02877" ext-link-type="DOI">10.48550/arXiv.2311.02877</ext-link>, 2023.</mixed-citation></ref>

  </ref-list></back>
    <!--<article-title-html>RIL-YOLO: a lightweight real-time object detection model on mobile devices for kart racing</article-title-html>
<abstract-html/>
<ref-html id="bib1.bib1"><label>1</label><mixed-citation>
      
Anggrainy, R., Yoga, N. G., Wiyono, A., Putri, D. M., Wahyudi, Z. T., and Septiyan, Y. A.: Unveiling the Future of Safety: Cutting-Edge Simulation Testing of e-Kart Bumpers, J. Phys. Conf. Ser., 012096, <a href="https://doi.org/10.1088/1742-6596/2866/1/012096" target="_blank">https://doi.org/10.1088/1742-6596/2866/1/012096</a>, 2024.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib2"><label>2</label><mixed-citation>
      
Chen, C., Li, J., Shuai, Z., Wang, Y., and Wang, Y.: A lightweight optimization framework for real-time pedestrian detection in dense and occluded scenes, Mech. Sci., 16, 877–886, <a href="https://doi.org/10.5194/ms-16-877-2025" target="_blank">https://doi.org/10.5194/ms-16-877-2025</a>, 2025.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib3"><label>3</label><mixed-citation>
      
Chen, J., Kao, S.-h., He, H., Zhuo, W., Wen, S., Lee, C.-H., and Chan, S.-H. G.: Run, don't walk: chasing higher FLOPS for faster neural networks, arXiv [preprint], <a href="https://doi.org/10.48550/arXiv.2303.03667" target="_blank">https://doi.org/10.48550/arXiv.2303.03667</a>, 2023.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib4"><label>4</label><mixed-citation>
      
Ding, X., Zhang, X., Han, J., and Ding, G.: Diverse branch block: Building a convolution as an inception-like unit, in: 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021, IEEE, 10881–10890, <a href="https://doi.org/10.1109/CVPR46437.2021.01074" target="_blank">https://doi.org/10.1109/CVPR46437.2021.01074</a>, 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib5"><label>5</label><mixed-citation>
      
Kaur, P., Khehra, B. S., and Mavi, E. B. S.: Data augmentation for object detection: A review, in: 2021 IEEE International Midwest Symposium on Circuits and Systems (MWSCAS), 537–543, <a href="https://doi.org/10.1109/MWSCAS47672.2021.9531849" target="_blank">https://doi.org/10.1109/MWSCAS47672.2021.9531849</a>, 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib6"><label>6</label><mixed-citation>
      
Lee, J., Park, S., Mo, S., Ahn, S., and Shin, J.: Layer-adaptive sparsity for the magnitude-based pruning, arXiv [preprint], <a href="https://doi.org/10.48550/arXiv.2010.07611" target="_blank">https://doi.org/10.48550/arXiv.2010.07611</a>, 2020.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib7"><label>7</label><mixed-citation>
      
Lu, J. and Liu, Y.: A real-time and accurate detection approach for bucket teeth falling off based on improved YOLOX, Mech. Sci., 13, 979–990, <a href="https://doi.org/10.5194/ms-13-979-2022" target="_blank">https://doi.org/10.5194/ms-13-979-2022</a>, 2022.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib8"><label>8</label><mixed-citation>
      
Matsumura, K., Yamakoshi, T., Yamakoshi, Y., and Rolfe, P.: The effect of competition on heart rate during kart driving: A field study, BMC Research Notes, 4, 342, <a href="https://doi.org/10.1186/1756-0500-4-342" target="_blank">https://doi.org/10.1186/1756-0500-4-342</a>, 2011.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib9"><label>9</label><mixed-citation>
      
Qin, D., Leichner, C., Delakis, M., Fornoni, M., Luo, S., Yang, F., Wang, W., Banbury, C., Ye, C., and Akin, B.: MobileNetV4: Universal models for the mobile ecosystem, arXiv [preprint], <a href="https://doi.org/10.48550/arXiv.2404.10518" target="_blank">https://doi.org/10.48550/arXiv.2404.10518</a>, 2024.


    </mixed-citation></ref-html>
<ref-html id="bib1.bib10"><label>10</label><mixed-citation>
      
Redmon, J., Divvala, S., Girshick, R., and Farhadi, A.: You only look once: Unified, real-time object detection, arXiv [preprint], <a href="https://doi.org/10.48550/arXiv.1506.02640" target="_blank">https://doi.org/10.48550/arXiv.1506.02640</a>, 2016.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib11"><label>11</label><mixed-citation>
      
Ren, H., Jing, F., and Li, S.: DCW-YOLO: road object detection algorithms for  autonomous driving, IEEE Access, 13, 125676–125688, <a href="https://doi.org/10.1109/ACCESS.2024.3364681" target="_blank">https://doi.org/10.1109/ACCESS.2024.3364681</a>, 2024.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib12"><label>12</label><mixed-citation>
      
Wang, A., Chen, H., Liu, L., Chen, K., Lin, Z., and Han, J.: Yolov10: Real-time end-to-end object detection, arXiv [preprint], <a href="https://doi.org/10.48550/arXiv.2405.14458" target="_blank">https://doi.org/10.48550/arXiv.2405.14458</a>, 2024a.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib13"><label>13</label><mixed-citation>
      
Wang, C.-Y., Yeh, I.-H., and Mark Liao, H.-Y.: Yolov9: Learning what you want to learn using programmable gradient information, arXiv [preprint], <a href="https://doi.org/10.48550/arXiv.2402.13616" target="_blank">https://doi.org/10.48550/arXiv.2402.13616</a>, 2024b.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib14"><label>14</label><mixed-citation>
      
Wang, H., Liu, X., Song, L.,Zhang, Y., Rong, X., and Wang, Y.: Research on a train safety driving method based on fusion of an incremental clustering algorithm and lightweight shared convolution, Sensors, 24, 4951, <a href="https://doi.org/10.3390/s24154951" target="_blank">https://doi.org/10.3390/s24154951</a>, 2024c.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib15"><label>15</label><mixed-citation>
      
Zhang, H., Xu, C., and Zhang, S.: Inner-IoU: more effective intersection over union loss with auxiliary bounding box, arXiv [preprint], <a href="https://doi.org/10.48550/arXiv.2311.02877" target="_blank">https://doi.org/10.48550/arXiv.2311.02877</a>, 2023.

    </mixed-citation></ref-html>--></article>
