Improved Tea Recognition Algorithm Based on RT-DETR
DOI:
https://doi.org/10.6911/WSRJ.202504_11(4).0017Keywords:
RT-DETR model, Tea identification, Object detection.Abstract
An improved RT-DETR model is proposed to solve the challenges of complex texture details, diverse target scales and high computational efficiency in tea recognition tasks. The convolutional gated linear unit (CGLU) is introduced to achieve dynamic weight adjustment, which enhances the robustness to complex background and variable target. OmniKernel module is designed to integrate multi-direction and multi-scale convolution kernel to improve the modeling ability of tea texture directivity and scale diversity. Combined with frequency-domain feature enhancement module (FSAM), the global context and local detail features are modeled jointly to suppress the interference of background noise. A feature segmentation and fusion module (SPDConv) is proposed to optimize the global consistency of features through space segmentation and channel fusion. Experiments show that the improved model, under the synergistic effect of enhanced input features and multi-scale feature pyramid network, significantly improves the accuracy of tea recognition and adaptability to complex scenes, while maintaining high computational efficiency, providing a high-precision and robust solution for tea detection tasks.
Downloads
References
[1] J.W. Zhao, Q.S. Chen, X.Y. Huang, C.H. Fang, Qualitative identification of tea categories by near infrared spectroscopy and support vector machine, J. Pharm. Biomed. Anal. 41(4) (2006) 1198-1204. https://doi.org/10.1016/j.jpba.2006.02.053.
[2] J. Zhou, H. Cheng, J.M. Zeng, L.Y. Wang, K. Wei, W. He, W.F. Wang, X. Liu, Study on Identification and Traceability of Tea Material Cultivar by Combined Analysis of Multi-Partial Least Squares Models Based on Near Infrared Spectroscopy, Spectrosc. Spectr. Anal. 30(10) (2010) 2650-2653. https://doi.org/10.3964/j.issn.1000-0593(2010)10-2650-04.
[3] W. Jian, Z. Xianyin, D. ShiPing, IDENTIFICATION AND GRADING OF TEA USING COMPUTER VISION, Appl. Eng. Agric. 26(4) (2010) 639-645.
[4] C. Zhang, J. Wang, G.D. Lu, S.M. Fei, T. Zheng, B.C. Huang, Automated tea quality identification based on deep convolutional neural networks and transfer learning, J. Food Process Eng. 46(4) (2023) 16. https://doi.org/10.1111/jfpe.14303.
[5] C.Y. Yan, Z.H. Chen, Z.L. Li, R.X. Liu, Y.X. Li, H. Xiao, P. Lu, B.L. Xie, Tea Sprout Picking Point Identification Based on Improved DeepLabV3+, Agriculture-Basel 12(10) (2022) 15. https://doi.org/10.3390/agriculture12101594.
[6] D.J.I. Shi, TransNeXt: Robust Foveal Visual Perception for Vision Transformers, (2023).
[7] Y.N. Cui, W.Q. Ren, A. Knoll, Omni-Kernel Modulation for Universal Image Restoration, IEEE Trans. Circuits Syst. Video Technol. 34(12) (2024) 12496-12509. https://doi.org/10.1109/tcsvt.2024.3429557.
[8] C.Y. Wang, A. Bochkovskiy, H.Y.M.J.a.e.-p. Liao, YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors, (2022).
Downloads
Published
Issue
Section
License
Copyright (c) 2025 World Scientific Research Journal

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.