<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.4 20241031//EN"
        "https://jats.nlm.nih.gov/publishing/1.4/JATS-journalpublishing1-4.dtd">
<article  article-type="research-article"        dtd-version="1.4">
            <front>

                <journal-meta>
                                                                <journal-id>j. inst. sci. and tech.</journal-id>
            <journal-title-group>
                                                                                    <journal-title>Journal of the Institute of Science and Technology</journal-title>
            </journal-title-group>
                                        <issn pub-type="epub">2536-4618</issn>
                                                                                            <publisher>
                    <publisher-name>Igdir University</publisher-name>
                </publisher>
                    </journal-meta>
                <article-meta>
                                        <article-id pub-id-type="doi">10.21597/jist.1283491</article-id>
                                                                <article-categories>
                                            <subj-group  xml:lang="en">
                                                            <subject>Computer Software</subject>
                                                    </subj-group>
                                            <subj-group  xml:lang="tr">
                                                            <subject>Bilgisayar Yazılımı</subject>
                                                    </subj-group>
                                    </article-categories>
                                                                                                                                                        <title-group>
                                                                                                                                                            <article-title>Improving Plant Disease Recognition Through Gradient-Based Few-shot Learning with Attention Mechanisms</article-title>
                                                                                                    </title-group>
            
                                                    <contrib-group content-type="authors">
                                                                        <contrib contrib-type="author">
                                                                    <contrib-id contrib-id-type="orcid">
                                        https://orcid.org/0000-0003-3037-5586</contrib-id>
                                                                <name>
                                    <surname>Işık</surname>
                                    <given-names>Gültekin</given-names>
                                </name>
                                                                    <aff>IĞDIR ÜNİVERSİTESİ</aff>
                                                            </contrib>
                                                                                </contrib-group>
                        
                                        <pub-date pub-type="pub" iso-8601-date="20230901">
                    <day>09</day>
                    <month>01</month>
                    <year>2023</year>
                </pub-date>
                                        <volume>13</volume>
                                        <issue>3</issue>
                                        <fpage>1482</fpage>
                                        <lpage>1495</lpage>
                        
                        <history>
                                    <date date-type="received" iso-8601-date="20230414">
                        <day>04</day>
                        <month>14</month>
                        <year>2023</year>
                    </date>
                                                    <date date-type="accepted" iso-8601-date="20230505">
                        <day>05</day>
                        <month>05</month>
                        <year>2023</year>
                    </date>
                            </history>
                                        <permissions>
                    <copyright-statement>Copyright © 2011, Journal of the Institute of Science and Technology</copyright-statement>
                    <copyright-year>2011</copyright-year>
                    <copyright-holder>Journal of the Institute of Science and Technology</copyright-holder>
                </permissions>
            
                                                                                                                        <abstract><p>This study investigates the use of few-shot learning algorithms to improve classification performance in situations where traditional deep learning methods fail due to a lack of training data. Specifically, we propose a few-shot learning approach using the Almost No Inner Loop (ANIL) algorithm and attention modules to classify tomato diseases in the Plant Village dataset. The attended features obtained from the five separate attention modules are classified using a Multi Layer Perceptron (MLP) classifier, and the soft voting method is used to weigh the classification scores from each classifier. The results demonstrate that our proposed approach achieves state-of-the-art accuracy rates of 97.05% and 97.66% for 10-shot and 20-shot classification, respectively. Our approach demonstrates the potential for incorporating attention mechanisms in feature extraction processes and suggests new avenues for research in few-shot learning methods.</p></abstract>
                                                            
            
                                                                                        <kwd-group>
                                                    <kwd>Few-shot learning</kwd>
                                                    <kwd>  Meta-learning</kwd>
                                                    <kwd>  Attention mechanisms</kwd>
                                                    <kwd>  Plant diseases</kwd>
                                                    <kwd>  Deep learning</kwd>
                                            </kwd-group>
                            
                                                                                                                                                    </article-meta>
    </front>
    <back>
                            <ref-list>
                                    <ref id="ref1">
                        <label>1</label>
                        <mixed-citation publication-type="journal">Albattah, W., Nawaz, M., Javed, A., Masood, M., &amp; Albahli, S. (2022). A novel deep learning method for detection and classification of plant diseases. Complex &amp; Intelligent Systems, 1–18.</mixed-citation>
                    </ref>
                                    <ref id="ref2">
                        <label>2</label>
                        <mixed-citation publication-type="journal">Argüeso, D., Picon, A., Irusta, U., Medela, A., San-Emeterio, M. G., Bereciartua, A., &amp; Alvarez-Gila, A. (2020). Few-Shot Learning approach for plant disease classification using images taken in the field. Computers and Electronics in Agriculture, 175, 105542.</mixed-citation>
                    </ref>
                                    <ref id="ref3">
                        <label>3</label>
                        <mixed-citation publication-type="journal">Arnold, S. M. R., Mahajan, P., Datta, D., Bunner, I., &amp; Zarkias, K. S. (2020). learn2learn: A Library for Meta-Learning Research. http://arxiv.org/abs/2008.12284</mixed-citation>
                    </ref>
                                    <ref id="ref4">
                        <label>4</label>
                        <mixed-citation publication-type="journal">Bayat, S., &amp; Işık, G. (2022). Recognition of Aras Bird Species From Their Voices With Deep Learning Methods. Journal of the Institute of Science and Technology, 12(3), 1250–1263.</mixed-citation>
                    </ref>
                                    <ref id="ref5">
                        <label>5</label>
                        <mixed-citation publication-type="journal">Cao, Y., Xu, J., Lin, S., Wei, F., &amp; Hu, H. (2019). Gcnet: Non-local networks meet squeeze-excitation networks and beyond. Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, 0.</mixed-citation>
                    </ref>
                                    <ref id="ref6">
                        <label>6</label>
                        <mixed-citation publication-type="journal">Chen, L., Cui, X., &amp; Li, W. (2021). Meta-learning for few-shot plant disease detection. Foods, 10(10), 2441.</mixed-citation>
                    </ref>
                                    <ref id="ref7">
                        <label>7</label>
                        <mixed-citation publication-type="journal">Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., &amp; Fei-Fei, L. (2009). Imagenet: A large-scale hierarchical image database. 2009 IEEE Conference on Computer Vision and Pattern Recognition, 248–255.</mixed-citation>
                    </ref>
                                    <ref id="ref8">
                        <label>8</label>
                        <mixed-citation publication-type="journal">Dumoulin, V., Houlsby, N., Evci, U., Zhai, X., Goroshin, R., Gelly, S., &amp; Larochelle, H. (2021). Comparing transfer and meta learning approaches on a unified few-shot classification benchmark. ArXiv Preprint ArXiv:2104.02638.</mixed-citation>
                    </ref>
                                    <ref id="ref9">
                        <label>9</label>
                        <mixed-citation publication-type="journal">Finn, C., Abbeel, P., &amp; Levine, S. (2017). Model-agnostic meta-learning for fast adaptation of deep networks. International Conference on Machine Learning, 1126–1135.</mixed-citation>
                    </ref>
                                    <ref id="ref10">
                        <label>10</label>
                        <mixed-citation publication-type="journal">Gao, Z., Xie, J., Wang, Q., &amp; Li, P. (2019). Global second-order pooling convolutional networks. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 3024–3033.</mixed-citation>
                    </ref>
                                    <ref id="ref11">
                        <label>11</label>
                        <mixed-citation publication-type="journal">Gündüz, M. Ş., &amp; Işık, G. (2023). A new YOLO-based method for social distancing from real-time videos. Neural Computing and Applications, 1–11.</mixed-citation>
                    </ref>
                                    <ref id="ref12">
                        <label>12</label>
                        <mixed-citation publication-type="journal">Guo, M.-H., Xu, T.-X., Liu, J.-J., Liu, Z.-N., Jiang, P.-T., Mu, T.-J., Zhang, S.-H., Martin, R. R., Cheng, M.-M., &amp; Hu, S.-M. (2022). Attention mechanisms in computer vision: A survey. Computational Visual Media, 8(3), 331–368.</mixed-citation>
                    </ref>
                                    <ref id="ref13">
                        <label>13</label>
                        <mixed-citation publication-type="journal">He, K., Zhang, X., Ren, S., &amp; Sun, J. (2016). Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 770–778.</mixed-citation>
                    </ref>
                                    <ref id="ref14">
                        <label>14</label>
                        <mixed-citation publication-type="journal">Hu, J., Shen, L., &amp; Sun, G. (2018). Squeeze-and-excitation networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 7132–7141.</mixed-citation>
                    </ref>
                                    <ref id="ref15">
                        <label>15</label>
                        <mixed-citation publication-type="journal">Karaman, A., Pacal, I., Basturk, A., Akay, B., Nalbantoglu, U., Coskun, S., Sahin, O., &amp; Karaboga, D. (2023). Robust real-time polyp detection system design based on YOLO algorithms by optimizing activation functions and hyper-parameters with artificial bee colony (ABC). Expert Systems with Applications, 221, 119741.</mixed-citation>
                    </ref>
                                    <ref id="ref16">
                        <label>16</label>
                        <mixed-citation publication-type="journal">Kaya, A., Keceli, A. S., Catal, C., Yalic, H. Y., Temucin, H., &amp; Tekinerdogan, B. (2019). Analysis of transfer learning for deep neural network based plant classification models. Computers and Electronics in Agriculture, 158, 20–29.</mixed-citation>
                    </ref>
                                    <ref id="ref17">
                        <label>17</label>
                        <mixed-citation publication-type="journal">Keceli, A. S., Kaya, A., Catal, C., &amp; Tekinerdogan, B. (2022). Deep learning-based multi-task prediction system for plant disease and species detection. Ecological Informatics, 69, 101679.</mixed-citation>
                    </ref>
                                    <ref id="ref18">
                        <label>18</label>
                        <mixed-citation publication-type="journal">Li, Y., &amp; Chao, X. (2021). Semi-supervised few-shot learning approach for plant diseases recognition. Plant Methods, 17, 1–10.</mixed-citation>
                    </ref>
                                    <ref id="ref19">
                        <label>19</label>
                        <mixed-citation publication-type="journal">Lin, H., Tse, R., Tang, S.-K., Qiang, Z., &amp; Pau, G. (2022a). Few-shot learning approach with multi-scale feature fusion and attention for plant disease recognition. Frontiers in Plant Science, 13.</mixed-citation>
                    </ref>
                                    <ref id="ref20">
                        <label>20</label>
                        <mixed-citation publication-type="journal">Lin, H., Tse, R., Tang, S.-K., Qiang, Z., &amp; Pau, G. (2022b). Few-Shot Learning for Plant-Disease Recognition in the Frequency Domain. Plants, 11(21), 2814.</mixed-citation>
                    </ref>
                                    <ref id="ref21">
                        <label>21</label>
                        <mixed-citation publication-type="journal">Liu, X., Zhang, F., Hou, Z., Mian, L., Wang, Z., Zhang, J., &amp; Tang, J. (2021). Self-supervised learning: Generative or contrastive. IEEE Transactions on Knowledge and Data Engineering, 35(1), 857–876.</mixed-citation>
                    </ref>
                                    <ref id="ref22">
                        <label>22</label>
                        <mixed-citation publication-type="journal">Mohanty, S. P., Hughes, D. P., &amp; Salathé, M. (2016). Using deep learning for image-based plant disease detection. Frontiers in Plant Science, 7, 1419.</mixed-citation>
                    </ref>
                                    <ref id="ref23">
                        <label>23</label>
                        <mixed-citation publication-type="journal">Munkhdalai, T., &amp; Yu, H. (2017). Meta networks. International Conference on Machine Learning, 2554–2563.</mixed-citation>
                    </ref>
                                    <ref id="ref24">
                        <label>24</label>
                        <mixed-citation publication-type="journal">Nichol, A., Achiam, J., &amp; Schulman, J. (2018). On first-order meta-learning algorithms. ArXiv Preprint ArXiv:1803.02999.</mixed-citation>
                    </ref>
                                    <ref id="ref25">
                        <label>25</label>
                        <mixed-citation publication-type="journal">Pacal, I. (2022). Deep Learning Approaches for Classification of Breast Cancer in Ultrasound (US) Images. Journal of the Institute of Science and Technology, 12(4), 1917–1927.</mixed-citation>
                    </ref>
                                    <ref id="ref26">
                        <label>26</label>
                        <mixed-citation publication-type="journal">Patricio, D. I., &amp; Rieder, R. (2018). Computer vision and artificial intelligence in precision agriculture for grain crops: A systematic review. Computers and Electronics in Agriculture, 153, 69–81.</mixed-citation>
                    </ref>
                                    <ref id="ref27">
                        <label>27</label>
                        <mixed-citation publication-type="journal">Raghu, A., Raghu, M., Bengio, S., &amp; Vinyals, O. (2019). Rapid learning or feature reuse? towards understanding the effectiveness of maml. ArXiv Preprint ArXiv:1909.09157.</mixed-citation>
                    </ref>
                                    <ref id="ref28">
                        <label>28</label>
                        <mixed-citation publication-type="journal">Shen, T., Zhou, T., Long, G., Jiang, J., Wang, S., &amp; Zhang, C. (2018). Reinforced self-attention network: a hybrid of hard and soft attention for sequence modeling. ArXiv Preprint ArXiv:1801.10296.</mixed-citation>
                    </ref>
                                    <ref id="ref29">
                        <label>29</label>
                        <mixed-citation publication-type="journal">Snell, J., Swersky, K., &amp; Zemel, R. (2017). Prototypical networks for few-shot learning. Advances in Neural Information Processing Systems, 30.</mixed-citation>
                    </ref>
                                    <ref id="ref30">
                        <label>30</label>
                        <mixed-citation publication-type="journal">Sun, Q., Liu, Y., Chua, T.-S., &amp; Schiele, B. (2019). Meta-transfer learning for few-shot learning. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 403–412.</mixed-citation>
                    </ref>
                                    <ref id="ref31">
                        <label>31</label>
                        <mixed-citation publication-type="journal">Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., &amp; Wojna, Z. (2016). Rethinking the inception architecture for computer vision. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2818–2826.</mixed-citation>
                    </ref>
                                    <ref id="ref32">
                        <label>32</label>
                        <mixed-citation publication-type="journal">Vinyals, O., Blundell, C., Lillicrap, T., Wierstra, D., &amp; others. (2016). Matching networks for one shot learning. Advances in Neural Information Processing Systems, 29.</mixed-citation>
                    </ref>
                                    <ref id="ref33">
                        <label>33</label>
                        <mixed-citation publication-type="journal">Wang, Q., Wu, B., Zhu, P., Li, P., Zuo, W., &amp; Hu, Q. (2020). ECA-Net: Efficient channel attention for deep convolutional neural networks. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 11534–11542.</mixed-citation>
                    </ref>
                                    <ref id="ref34">
                        <label>34</label>
                        <mixed-citation publication-type="journal">Wang, S., Li, C., Wang, R., Liu, Z., Wang, M., Tan, H., Wu, Y., Liu, X., Sun, H., Yang, R., &amp; others. (2021). Annotation-efficient deep learning for automatic medical image segmentation. Nature Communications, 12(1), 5915.</mixed-citation>
                    </ref>
                                    <ref id="ref35">
                        <label>35</label>
                        <mixed-citation publication-type="journal">Wang, W., Song, H., Zhao, S., Shen, J., Zhao, S., Hoi, S. C. H., &amp; Ling, H. (2019). Learning unsupervised video object segmentation through visual attention. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 3064–3074.</mixed-citation>
                    </ref>
                                    <ref id="ref36">
                        <label>36</label>
                        <mixed-citation publication-type="journal">Woo, S., Park, J., Lee, J.-Y., &amp; Kweon, I. S. (2018). Cbam: Convolutional block attention module. Proceedings of the European Conference on Computer Vision (ECCV), 3–19.</mixed-citation>
                    </ref>
                                    <ref id="ref37">
                        <label>37</label>
                        <mixed-citation publication-type="journal">Yang, J., Guo, X., Li, Y., Marinello, F., Ercisli, S., &amp; Zhang, Z. (2022). A survey of few-shot learning in smart agriculture: developments, applications, and challenges. Plant Methods, 18(1), 1–12.</mixed-citation>
                    </ref>
                            </ref-list>
                    </back>
    </article>
