Modern farming systems face significant challenges due to an excessive reliance on chemical herbicides, which have led to biodiversity loss, soil degradation, the emergence of herbicide-resistant weed species and increased production costs (
Rao et al., 2020;
Nath et al., 2024). The European Union and other regions have set high standards to convert 30-40% of farmland to organic production by 2030 (
Pânzaru et al., 2023), as a result of global laws aimed at promoting sustainable practices. Rice farming, in particular, requires innovative weed management approaches, since it is one of the most herbicide-intensive crops (
Arockia et al., 2023;
Kokilam et al., 2023; Bandyopadhyay et al., 2024). Recent research, including our previous work, explored autonomous robotic weeders as potential alternatives to chemical control. Earlier studies primarily focused on machine vision for weed detection and eradication methods
(Mohanty et al., 2025). This research focuses on developing a robotic platform with artificial intelligence (AI) algorithms specifically designed for precision weeding in paddy and other crops grown in paddy-based cropping systems.
System architecture and operational framework
The robotic platform was designed primarily for mechanical weed suppression in high-density organic farming systems, where weed-to-crop ratios can exceed 12:1, posing a significant challenge for herbicide-free agriculture. (
Mesías-Ruiz et al., 2024).
To address this demand, new computational frameworks were created, including a topology-aware row detection algorithm and a multispectral plant discrimination model that were tuned for the platform’s AI-powered perception stack, which combines convolutional neural networks (CNNs) for spatial pattern recognition with adaptive decision-making algorithms that adjust weeding intensity in real time depending on Phyto-morphological inputs
(Zhu et al., 2025). This split-processing architecture separates structural row navigation from weed species detection, enabling both tasks to run concurrently
(Gao et al., 2025). This parallelism improves real-time response and precision in targeted weed suppression.
The AI-driven perception stack has a split architecture
(Ozer et al., 2025), as shown in Fig 1. CNN-based spatial pattern recognition examines crop row topology for navigation (left stream), while a separate multispectral analytic pipeline classifies weed species (right stream)
(Vignesh et al., 2025). Both outputs are fed into a central decision-making unit that regulates actuator intensity and path correction in real time.
Integrated methodological framework and software architecture
The methodological approach for autonomous weed management is designed to address the inherent challenges of rice cultivation, namely its row-based structure, which creates two distinct weeding spatial regimes: inter-row clearance (removing weeds between crop rows) and the more complex intra-row eradication (identifying and eliminating weeds within the crop line, requiring precise plant-level discrimination)
(Qu et al., (2024). To navigate this environment, the robotic platform utilises a modular propulsion system comprising an active front-wheel drive, powered by a single cage-wheel motor and stabilising rear casters. The system’s operational intelligence is governed by a sophisticated software architecture running on an NVIDIA Jetson Nano embedded computer, which synchronises the vision, navigation and actuation subsystems for continuous, real-time performance. The software workflow can be broken down into three core functions: perception and localisation, decision and control and actuation.
Technical implementation
In the software architecture shown in Fig 2, detection data from the YOLO model is routed through an intermediary buffer task to facilitate asynchronous communication between the application and control layers. This design ensures that time-sensitive control processes, such as navigation and weeding, operate on perception data without being delayed by ongoing inference computations, maintaining the control layer’s responsiveness
(Lan et al., 2024). The system is organised into a two-layer design to enhance modularity and maintainability. This structure, supported by asynchronous communication protocols and the NVIDIA Jetson Nano’s processing capabilities, enables non-blocking data flow through the buffer
(Dang et al., 2025). This ensures that updates to perception algorithms or the web-based dashboard for remote monitoring do not impact the core control logic, enabling robust, vision-guided autonomous operation in resource-constrained environments
(Moghadam et al., 2026).
Fig 3 compares the YOLO backbone architectures of v5, v7 and v8. The backbone of YOLOv5 has undergone significant changes compared to previous versions of the model. It mainly utilises the Focus structure for down-sampling, as well as the C3 (Cross Stage Partial) module, a modified version of the CSP constraint designed for enhanced gradient flow. The backbone also includes an SPP (Spatial Pyramid Pooling) layer to expand the receptive field.
YOLOv7 has a more complex backbone that incorporates the E-ELAN (Extended Efficient Layer Aggregation Network) computing block. This structure is designed to improve the network’s learning capabilities while preserving the original gradient route. It also uses SPP-CPC (Spatial Pyramid Pooling with Cross-Stage Partial Connections) for feature integration. The backbone of YOLOv8 replaces the C3 module with the C2f (Cross Stage Partial with 2 Convolutions) module, which provides improved gradient flow. It also employs an SPPF (Spatial Pyramid Pooling Fast) module, a faster version of SPP and transitions to an anchor-free detection head, simplifying the architecture.
Mechanical architecture framework and technical specifications of the robot
The robot system is designed to minimise soil compaction and maximise traction under saturated paddy field conditions. Fig 4 shows the mechanical flow diagram, which integrates several subsystems, including the chassis, drive system, vision system and cultivating tool assembly. The wheel assemblies with cage wheels are attached to the chassis via a chain-sprocket transmission mechanism and a motor-mounted shaft. This ensures effective torque transfer and improved grip on wet soil conditions. A Jetson Nano processing unit running the YOLO object detection model is connected to cameras as part of the vision system, enabling the real-time identification of crops and weeds
(Zhang et al., 2026; Ramos-Sanchez et al., 2026).
A knuckle joint (made of Grade 30C8 steel) connects the motor mounting body to the main robot chassis, as shown in Fig 5. For autonomous operation, actuation is provided by a NEMA 17 stepper motor, which is coupled to the steering system and governed by a Jetson Nano embedded controller. To ensure algorithmic stability and power efficiency during row-following, the software imposes a ±10-degree operational range around the neutral position. This soft limit prevents oversteer and mechanical shock while providing sufficient articulation for accurate path correction
(Raj et al., 2018).
Cage-wheel propulsion system
The robot utilises a front-driven cage-wheel propulsion system (Fig 6), designed for navigating soft, muddy terrain. A single, independently powered front cage wheel provides driven traction, while passive rear stabilisers ensure balance. The open-frame wheel design enhances grip on waterlogged soil, minimises debris accumulation and reduces slippage, which is critical for precise inter-row navigation and weeding tasks.
Motor mounting configuration with chain and sprocket assembly
The propulsion is powered by one high-torque 12V DC worm gear motor, each delivering a rated torque of 70 kg·cm (110 kg·cm stall torque) at approximately 10 RPM. The motor is mounted on the front chassis for optimal weight distribution and is connected to the drive shaft
via a slip-free torque transmission, as shown in Fig 7(a). The worm gear design provides inherent braking and high torque at low speeds, making it an ideal choice for agricultural applications. The mounting frame is built to withstand field vibrations and impacts, ensuring longevity. The motor is bolted to the wheel mount via four mounting bolts. The motor shaft is fitted with the driven sprocket, which engages with the driver sprocket through a drive chain to transmit power. A precision sprocket is securely affixed to a 16 mm inner diameter drive shaft Fig 7(b) using a set screw, ensuring reliable torque transmission. The gearbox is pre-lubricated with high-performance grease to reduce friction and ensure durable, low-maintenance operation under variable field loads. A roller chain and sprocket arrangement Fig 7(c) transmit power from the motor’s output shaft to the wheel axle. The system is designed to minimise backlash and slippage, with an adjustable chain tensioner that maintains wheel synchronisation and compensates for elongation over time. This ensures consistent and stable power delivery across uneven or hilly terrain.
Rotary blade as a weed cutter and the cultivator
A horizontally rotating blade assembly, as shown in Fig 8, positioned at an optimal height below the chassis, severs weeds between crop rows.
This mechanical subsystem features a dual-action mechanism that integrates a rotary weed cutter with a cultivator-incorporator.
Chassis with steering design
The chassis, as shown in Fig 9(a), is constructed from high-strength, lightweight materials (aluminium alloy) and designed to support all subsystems while withstanding field vibrations and environmental challenges. The steering module of the robot Fig 9(b) utilises simple, low-cost and no complex linkages. Self-aligning Casters automatically follow the path of motion with a maximum turn angle of 10°, driven by a NEMA 17 stepper motor. Fail-safe electromagnetic brakes are integrated to deploy automatically in the event of power failure, ensuring the robot’s position is secured. The steering system, controlled by the Jetson Nano board, is software-limited to an operational range of ±10° from neutral, preventing oversteer while ensuring precise path correction. The steering wheel is attached to a stepper motor, which is in turn mounted on the chassis.
Protection and safety systems
The chassis is elevated to provide sufficient ground clearance, protecting critical drive and electronic components from uneven terrain, crop residue and debris. Simultaneously, the cutting blades are positioned below the chassis to engage weeds at or below the soil surface. This purposeful vertical offset ensures effective weeding accuracy and crop safety while maintaining overall mobility and operational reliability in challenging field conditions. This spatial design, combined with synchronisation that reduces the machine vision processing lag to less than 200 ms, reduces the positional error between plant identification and mechanical action to less than 2 cm, resulting in accurate intervention. Power is supplied by modular Lithium-ion battery packs rated at 12 V and 12 Ah. Routine maintenance, including chain tension adjustment, wheel alignment verification and motor mounting inspection, is performed to ensure system reliability and optimal performance.
Use of appropriate herbicides through pneumatic control
The herbicide delivery system is shown in Fig 10. The system features a pneumatically controlled herbicide delivery mechanism that enables precise, targeted chemical treatment based on real-time weed classification. When a weed is identified using vision-based detection (
e.g., YOLO v5, v7 and v8 combined with stereo camera input), the appropriate herbicide is selected from several onboard containers, each labelled with a specific herbicide type. This modular design enables simple refilling, replacement and scaling, depending on the crop type and field size. It ensures that agrochemicals are deployed safely and automatically, aligning with precision agricultural goals.
Mechatronics architecture
The platform’s mechatronic architecture is built around a series of 12 V DC actuators equipped with worm-gear transmissions, enabling high torque and precise placement required for farming operations. These actuators drive a modular end-effector system that accommodates interchangeable tools, such as rotary blades and sliding-mount driven tines, to perform a diverse range of soil-intervention tasks. The vibration sensors are strategically mounted near the tool frame to capture oscillatory signatures generated during tool–soil interaction. Deviations from the nominal frequency response are interpreted as indicators of bearing wear, mechanical imbalance, or variations in soil compaction.