[1] |
LOWE D G. Distinctive image features from scale-invariant keypoints[J]. International Journal of Computer Vision, 2004, 60(2): 91-110.
DOI
URL
|
[2] |
DALAL N, TRIGGS B. Histograms of oriented gradients for human detection[C]// 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2005: 886-893.
|
[3] |
郭艾侠, 彭明明, 邢仲璟. 机器视觉技术在荔枝识别与定位研究中的应用[J]. 计算机工程与应用, 2017, 53(17): 218-223, 259.
DOI
|
|
GUO A X, PENG M M, XING Z J. Study on recognition and positioning of litchi based on technology of machine vision[J]. Computer Engineering and Applications, 2017, 53(17): 218-223, 259. (in Chinese)
DOI
|
[4] |
王永皎, 张引, 张三元. 基于图像处理的植物叶面积测量方法[J]. 计算机工程, 2006, 32(8): 210-212.
|
|
WANG Y J, ZHANG Y, ZHANG S Y. Approach to measure plant leaf area based on image process[J]. Computer Engineering, 2006, 32(8): 210-212. (in Chinese)
|
[5] |
VASWANI A, SHAZEER N, PARMAR N, et al. Attention is all You need[C]// The 31st International Conference on Neural Information Processing Systems. New York: ACM, 2017: 6000-6010.
|
[6] |
YU W H, LUO M, ZHOU P, et al. MetaFormer is actually what You need for vision[EB/OL]. [2022-06-21]. https://arxiv.org/abs/2111.11418.
|
[7] |
PARK J, WOO S, LEE J Y, et al. BAM: bottleneck attention module[EB/OL]. [2022-05-14]. https://arxiv.org/abs/1807.06514.
|
[8] |
KRIZHEVSKY A, SUTSKEVER I, HINTON G E. ImageNet classification with deep convolutional neural networks[C]// The 25th International Conference on Neural Information Processing Systems-Volume 1. New York:ACM, 2012: 1097-1105.
|
[9] |
SIMONYAN K, ZISSERMAN A. Very deep convolutional networks for large-scale image recognition[EB/OL]. [2022-05- 16]. https://arxiv.org/abs/1409.1556.
|
[10] |
SZEGEDY C, LIU W, JIA Y Q, et al. Going deeper with convolutions[C]// 2015 IEEE Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2015: 1-9.
|
[11] |
HE K M, ZHANG X Y, REN S Q, et al. Deep residual learning for image recognition[C]// 2016 IEEE Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2016: 770-778.
|
[12] |
LEE S H, CHAN C S, WILKIN P, et al. Deep-plant: plant identification with convolutional neural networks[C]// 2015 IEEE International Conference on Image Processing. New York: IEEE Press, 2015: 452-456.
|
[13] |
XIA X L, XU C, NAN B. Inception-v3 for flower classification[C]// 2017 2nd International Conference on Image, Vision and Computing. New York: IEEE Press, 2017: 783-787.
|
[14] |
GAVAI N R, JAKHADE Y A, TRIBHUVAN S A, et al. MobileNets for flower classification using TensorFlow[C]// 2017 International Conference on Big Data, IoT and Data Science. New York: IEEE Press, 2018: 154-158.
|
[15] |
CAO S, SONG B. Visual attentional-driven deep learning method for flower recognition[J]. Mathematical Biosciences and Engineering: MBE, 2021, 18(3): 1981-1991.
DOI
URL
|
[16] |
DOSOVITSKIY A, BEYER L, KOLESNIKOV A, et al. An image is worth 16×16 words: transformers for image recognition at scale[EB/OL]. [2022-05-16]. https://arxiv.org/abs/2010.11929.
|
[17] |
TOUVRON H, CORD M, DOUZE M, et al. Training data-efficient image transformers & distillation through attention[EB/OL]. [2022-05-16]. https://arxiv.org/abs/2012.12877.
|
[18] |
YUAN L, CHEN Y P, WANG T, et al. Tokens-to-token ViT: training vision transformers from scratch on ImageNet[C]// 2021 IEEE/CVF International Conference on Computer Vision. New York: IEEE Press, 2022: 538-547.
|
[19] |
LEE-THORP J, AINSLIE J, ECKSTEIN I, et al. FNet: mixing tokens with Fourier transforms[EB/OL]. [2022-05-16]. https://arxiv.org/abs/2105.03824.
|
[20] |
TOLSTIKHIN I, HOULSBY N, KOLESNIKOV A, et al. MLP-mixer: an all-MLP architecture for vision[EB/OL]. [2022-05-16]. https://arxiv.org/abs/2105.01601.
|
[21] |
LIU Z, LIN Y T, CAO Y, et al. Swin transformer: hierarchical vision transformer using shifted windows[C]// 2021 IEEE/CVF International Conference on Computer Vision. New York: IEEE Press, 2022: 9992-10002.
|
[22] |
LIN T Y, DOLLÁR P, GIRSHICK R, et al. Feature pyramid networks for object detection[C]// 2017 IEEE Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2017: 936-944.
|
[23] |
RONNEBERGER O, FISCHER P, BROX T. U-net: convolutional networks for biomedical image segmentation[M]// Lecture Notes in Computer Science. Cham: Springer International Publishing, 2015: 234-241.
|
[24] |
HE K M, ZHANG X Y, REN S Q, et al. Deep residual learning for image recognition[C]// 2016 IEEE Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2016: 770-778.
|
[25] |
XIE S N, GIRSHICK R, DOLLÁR P, et al. Aggregated residual transformations for deep neural networks[C]// 2017 IEEE Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2017: 5987-5995.
|
[26] |
HENDRYCKS D, GIMPEL K. Gaussian error linear units (GELUs)[EB/OL]. [2022-05-16]. https://arxiv.org/abs/1606.08415.
|
[27] |
RADOSAVOVIC I, KOSARAJU R P, GIRSHICK R, et al. Designing network design spaces[C]// 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2020: 10425-10433.
|
[28] |
HUMPHREY E J, BELLO J P. Rethinking automatic chord recognition with convolutional neural networks[C]// The 11th International Conference on Machine Learning and Applications. New York: IEEE Press, 2013: 357-362.
|
[29] |
DONG X Y, BAO J M, CHEN D D, et al. CSWin transformer: a general vision transformer backbone with cross-shaped windows[EB/OL]. [2022-05-16]. https://arxiv.org/abs/2107.00652.
|