AppleLeafNet: a lightweight and efficient deep learning framework for diagnosing apple leaf diseases

Front Plant Sci. 2024 Nov 27:15:1502314. doi: 10.3389/fpls.2024.1502314. eCollection 2024.

Abstract

Accurately identifying apple diseases is essential to control their spread and support the industry. Timely and precise detection is crucial for managing the spread of diseases, thereby improving the production and quality of apples. However, the development of algorithms for analyzing complex leaf images remains a significant challenge. Therefore, in this study, a lightweight deep learning model is designed from scratch to identify the apple leaf condition. The developed framework comprises two stages. First, the designed 37-layer model was employed to assess the condition of apple leaves (healthy or diseased). Second, transfer learning was used for further subclassification of the disease class (e.g., rust, complex, scab, and frogeye leaf spots). The trained lightweight model was reused because the model trained with correlated images facilitated transfer learning for further classification of the disease class. A dataset available online was used to validate the proposed two-stage framework, resulting in a classification rate of 98.25% for apple leaf condition identification and an accuracy of 98.60% for apple leaf disease diagnosis. Furthermore, the results confirm that the proposed model is lightweight and involves relatively fewer learnable parameters in comparison with other pre-trained deep learning models.

Keywords: apple leaf condition identification; apple leaf disease detection; crop monitoring; deep learning; lightweight model.

Grants and funding

The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This work was partially supported by the Ministry of Science and ICT (MSIT), Korea, under the ICT Challenge and Advanced Network of HRD (ICAN) program (IITP-2024-RS-2022-00156345) supervised by the Institute of Information & Communications Technology Planning & Evaluation (IITP). This work was partially supported by the National Research Foundation of Korea (NRF) under grant numbers RS-2023-00219051 and RS-2023-00209107 and the Unmanned Vehicles Core Technology Research and Development Program through the NRF and Unmanned Vehicle Advanced Research Center (UVARC), funded by the Ministry of Science and ICT, Republic of Korea (NRF-2023M3C1C1A01098408).