With the increase of population in the world, the demand for quality food is increasing too. In recent years, increasing demand and environmental factors have heavily influenced the agricultural production. Automation and robotics for fruit and vegetable production/monitoring have become the new standard. This paper discusses an autonomous Unmanned Aerial Vehicle (UAV) able to navigate through rows orchard rows. The UAV is comprised of a flight controller (AP stack), a microcontroller for analog reading of different sensors, and an On-Board Computer (OBC). Pictures are taken through a camera and streamed through WiFi to a Ground Control Computer (GCC) running a convolution neural network model. Based on prior training, the model outputs three directions: RIGHT, LEFT and STRAIGHT. A moving average of multiple frames per second is extracted and sent to a build-in Proportional-IntegralDerivative (PID) controller on the UAV. After error correction from this feedback, controller sends the direction to the flight controller using MAVLink protocol's radio channel overrides, thus performing autonomous navigation.
Breslla K., Bortolotti G., Boini A., Perulli G., Morandi B., Grappadelli L.C., et al. (2020). Sensor-fusion and deep neural networks for autonomous UAV navigation within orchards. Institute of Electrical and Electronics Engineers Inc. [10.1109/MetroAgriFor50201.2020.9277568].
Sensor-fusion and deep neural networks for autonomous UAV navigation within orchards
Bortolotti G.;Boini A.;Perulli G.;Morandi B.;Grappadelli L. C.;Manfrini L.
2020
Abstract
With the increase of population in the world, the demand for quality food is increasing too. In recent years, increasing demand and environmental factors have heavily influenced the agricultural production. Automation and robotics for fruit and vegetable production/monitoring have become the new standard. This paper discusses an autonomous Unmanned Aerial Vehicle (UAV) able to navigate through rows orchard rows. The UAV is comprised of a flight controller (AP stack), a microcontroller for analog reading of different sensors, and an On-Board Computer (OBC). Pictures are taken through a camera and streamed through WiFi to a Ground Control Computer (GCC) running a convolution neural network model. Based on prior training, the model outputs three directions: RIGHT, LEFT and STRAIGHT. A moving average of multiple frames per second is extracted and sent to a build-in Proportional-IntegralDerivative (PID) controller on the UAV. After error correction from this feedback, controller sends the direction to the flight controller using MAVLink protocol's radio channel overrides, thus performing autonomous navigation.I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.