Enhancing Plant Leaf Classification with Deep Learning: Automating Feature Extraction for Accurate Species Identification

Authors

  • Chilukuri Ganesh Student, Department of AI and DS, Koneru Lakshmaiah Education Foundation Author
  • Gandikota Harshavardhan Student, Department of AI and DS, Koneru Lakshmaiah Education Foundation Author
  • Naishadham Radha Sri Keerthi Student, Department of AI and DS, Koneru Lakshmaiah Education Foundation Author
  • Raj Veer Yabaji Student, Department of AI and DS, Koneru Lakshmaiah Education Foundation Author
  • Rajveer Yabaji Meghana Sadhu Student, Department of AI and DS, Koneru Lakshmaiah Education Foundation Author

DOI:

https://doi.org/10.56294/piii2025513

Keywords:

Plant leaf classification, Deep learning, Evaluation metrics, CNN, Data augmentation, Optimizers, ANN, Regularization, VGG, RNN, LSTM, Autoencoders

Abstract

Plant leaf classification using deep learning provides an automated approach that surpasses traditional methods reliant on manual feature selection. Convolutional Neural Networks (CNNs) excel at learning intricate patterns from leaf images, extracting valuable features that contribute to accurate plant species identification. These models enhance classification precision by automating feature extraction, thereby improving efficiency and reliability. By leveraging deep learning, plant recognition systems can become more dependable, and their classification accuracy is significantly increased, minimizing human error and manual intervention.

References

1. Garcia, S., Luengo, J., & Herrera, F. (2016). Data Preprocessing in Data Mining. Springer International Publishing.

o This work discusses various techniques for preprocessing data in data mining applications, focusing on how to handle raw data before applying machine learning models.

2. Kingma, D. P., & Ba, J. (2014). Adam: A Method for Stochastic Optimization. arXiv preprint arXiv:1412.6980.

o This paper introduces the Adam optimization algorithm, a method that combines the advantages of two other extensions of stochastic gradient descent, providing a robust technique for training deep learning models.

3. Bottou, L. (2012). Stochastic Gradient Descent Tricks. In Neural Networks: Tricks of the Trade (pp. 421-436). Springer, Berlin, Heidelberg.

o This chapter explores various tricks and optimizations to improve the performance of stochastic gradient descent (SGD) in training neural networks.

4. Tieleman, T., & Hinton, G. (2012). Lecture 6.5 - RMSprop: Divide the Gradient by a Running Average of its Recent Magnitude. COURSERA: Neural Networks for Machine Learning, 4(2), 26-31.

o The authors present RMSprop, a gradient descent method that divides the gradient by a moving average of its recent magnitude to adjust the learning rate dynamically.

5. Bishop, C. M. (2006). Pattern Recognition and Machine Learning. Springer.

o This book offers comprehensive coverage of the theory and techniques involved in pattern recognition and machine learning, including the mathematical foundations and various models.

6. Bottou, L., & Bousquet, O. (2008). The Tradeoffs of Large Scale Learning. In Advances in Neural Information Processing Systems (pp. 161-168).

o This paper discusses the challenges and tradeoffs involved in training machine learning algorithms on large-scale datasets, focusing on computational efficiency and algorithmic scalability.

7. LeCun, Y., Bottou, L., Bengio, Y., & Haffner, P. (1998). Gradient-Based Learning Applied to Document Recognition. Proceedings of the IEEE, 86(11), 2278-2324.

o The authors discuss the application of gradient-based learning methods to document recognition, introducing the concept of convolutional neural networks for image classification tasks.

8. Shorten, C., & Khoshgoftaar, T. M. (2019). A Survey on Image Data Augmentation for Deep Learning. Journal of Big Data, 6(1), 60.

o This survey paper reviews different data augmentation techniques that can be applied to image datasets, emphasizing their impact on the performance of deep learning models.

9. Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., & Salakhutdinov, R. (2014). Dropout: A Simple Way to Prevent Neural Networks from Overfitting. The Journal of Machine Learning Research, 15(1), 1929-1958.

o The authors propose the dropout technique as a regularization method to prevent neural networks from overfitting by randomly setting a fraction of input units to zero during training.

10. Simonyan, K., & Zisserman, A. (2014). Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv preprint arXiv:1409.1556.

• This paper introduces a very deep convolutional network (VGG) for image classification, achieving significant performance improvements on large-scale benchmarks like ImageNet.

11. Hochreiter, S., & Schmidhuber, J. (1997). Long Short-Term Memory. Neural Computation, 9(8), 1735-1780.

• The authors present the Long Short-Term Memory (LSTM) network, a type of recurrent neural network (RNN) designed to overcome the vanishing gradient problem in training deep networks.

12. Graves, A., Mohamed, A. R., & Hinton, G. (2013). Speech Recognition with Deep Recurrent Neural Networks. In 2013 IEEE International Conference on Acoustics, Speech and Signal Processing (pp. 6645-6649). IEEE.

• This work applies deep recurrent neural networks (RNNs) to speech recognition, showing that RNNs, particularly LSTMs, outperform traditional methods for sequential data.

13. Hinton, G. E., & Salakhutdinov, R. R. (2006). Reducing the Dimensionality of Data with Neural Networks. Science, 313(5786), 504-507.

• The authors demonstrate how neural networks can be used to reduce the dimensionality of data, effectively capturing important patterns and structures in high-dimensional datasets.

14. Vincent, P., Larochelle, H., Bengio, Y., & Manzagol, P. A. (2008). Extracting and Composing Robust Features with Denoising Autoencoders. In Proceedings of the 25th International Conference on Machine Learning (pp. 1096-1103).

• This paper introduces denoising autoencoders, a technique for learning robust features by corrupting input data and training the model to reconstruct the original data.

15. He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 770-778).

• This work introduces the concept of residual learning, which helps to train deep networks by bypassing some layers to address the problem of vanishing gradients.

16. Kingma, D. P., & Welling, M. (2013). Auto-Encoding Variational Bayes. arXiv preprint arXiv:1312.6114.

Downloads

Published

2025-01-06

How to Cite

1.
Ganesh C, Harshavardhan G, Sri Keerthi NR, Yabaji RV, Rajveer Yabaji MS. Enhancing Plant Leaf Classification with Deep Learning: Automating Feature Extraction for Accurate Species Identification. SCT Proceedings in Interdisciplinary Insights and Innovations [Internet]. 2025 Jan. 6 [cited 2025 Feb. 14];3:513. Available from: https://proceedings.ageditor.ar/index.php/piii/article/view/513