<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.4 20241031//EN"
        "https://jats.nlm.nih.gov/publishing/1.4/JATS-journalpublishing1-4.dtd">
<article  article-type="research-article"        dtd-version="1.4">
            <front>

                <journal-meta>
                                    <journal-id></journal-id>
            <journal-title-group>
                                                                                    <journal-title>Balkan Journal of Electrical and Computer Engineering</journal-title>
            </journal-title-group>
                            <issn pub-type="ppub">2147-284X</issn>
                                        <issn pub-type="epub">2147-284X</issn>
                                                                                            <publisher>
                    <publisher-name>MUSA YILMAZ</publisher-name>
                </publisher>
                    </journal-meta>
                <article-meta>
                                        <article-id pub-id-type="doi">10.17694/bajece.677326</article-id>
                                                                <article-categories>
                                            <subj-group  xml:lang="en">
                                                            <subject>Artificial Intelligence</subject>
                                                    </subj-group>
                                            <subj-group  xml:lang="tr">
                                                            <subject>Yapay Zeka</subject>
                                                    </subj-group>
                                    </article-categories>
                                                                                                                                                        <title-group>
                                                                                                                                                            <article-title>A Hybrid Framework for Matching Printing Design Files to Product Photos</article-title>
                                                                                                    </title-group>
            
                                                    <contrib-group content-type="authors">
                                                                        <contrib contrib-type="author">
                                                                <name>
                                    <surname>Kaplan</surname>
                                    <given-names>Alper</given-names>
                                </name>
                                                                    <aff>YEDITEPE UNIVERSITY</aff>
                                                            </contrib>
                                                    <contrib contrib-type="author">
                                                                    <contrib-id contrib-id-type="orcid">
                                        https://orcid.org/0000-0002-0792-7306</contrib-id>
                                                                <name>
                                    <surname>Akagunduz</surname>
                                    <given-names>Erdem</given-names>
                                </name>
                                                                    <aff>CANKAYA UNIVERSITY</aff>
                                                            </contrib>
                                                                                </contrib-group>
                        
                                        <pub-date pub-type="pub" iso-8601-date="20200430">
                    <day>04</day>
                    <month>30</month>
                    <year>2020</year>
                </pub-date>
                                        <volume>8</volume>
                                        <issue>2</issue>
                                        <fpage>170</fpage>
                                        <lpage>180</lpage>
                        
                        <history>
                                    <date date-type="received" iso-8601-date="20200120">
                        <day>01</day>
                        <month>20</month>
                        <year>2020</year>
                    </date>
                                                    <date date-type="accepted" iso-8601-date="20200416">
                        <day>04</day>
                        <month>16</month>
                        <year>2020</year>
                    </date>
                            </history>
                                        <permissions>
                    <copyright-statement>Copyright © 2013, Balkan Journal of Electrical and Computer Engineering</copyright-statement>
                    <copyright-year>2013</copyright-year>
                    <copyright-holder>Balkan Journal of Electrical and Computer Engineering</copyright-holder>
                </permissions>
            
                                                                                                                        <abstract><p>We propose a real-time image matching framework, which is hybrid in the sense that it uses both hand-crafted features and deep features obtained from a well-tuned deep convolutional network. The matching problem, which we concentrate on, is specific to a certain application, that is, printing design to product photo matching. Printing designs are any kind of template image files, created using a design tool, thus are perfect image signals. However, photographs of a printed product suffer many unwanted effects, such as uncontrolled shooting angle, uncontrolled illumination, occlusions, printing deficiencies in color, camera noise, optic blur, et cetera. For this purpose, we create an image set that includes printing design and corresponding product photo pairs with collaboration of an actual printing facility. Using this image set, we benchmark various hand-crafted and deep features for matching performance and propose a framework in which deep learning is utilized with highest contribution, but without disabling real-time operation using an ordinary desktop computer.</p></abstract>
                                                            
            
                                                                                        <kwd-group>
                                                    <kwd>image matching</kwd>
                                                    <kwd>  hand-crafted features</kwd>
                                                    <kwd>  deep features</kwd>
                                                    <kwd>  semantic segmentation</kwd>
                                                    <kwd>  product image processing</kwd>
                                            </kwd-group>
                            
                                                                                                                                                <funding-group specific-use="FundRef">
                    <award-group>
                                                    <funding-source>
                                <named-content content-type="funder_name">TUBITAK - TEYDEB</named-content>
                            </funding-source>
                                                                            <award-id>7170364</award-id>
                                            </award-group>
                </funding-group>
                                </article-meta>
    </front>
    <back>
                            <ref-list>
                                    <ref id="ref1">
                        <label>1</label>
                        <mixed-citation publication-type="journal">[1]	T. Dharani and I. L. Aroquiaraj, &quot;A survey on content based image retrieval,&quot; in 2013 International Conference on Pattern Recognition, Informatics and Mobile Engineering, pp. 485-490, Feb 2013.</mixed-citation>
                    </ref>
                                    <ref id="ref2">
                        <label>2</label>
                        <mixed-citation publication-type="journal">[2]	Y. Liu, D. Zhang, G. Lu, and W.-Y. Ma, &quot;A survey of content-based image retrieval with high-level semantics,&quot; Pattern Recognition, vol. 40, no. 1, pp. 262 - 282, 2007.</mixed-citation>
                    </ref>
                                    <ref id="ref3">
                        <label>3</label>
                        <mixed-citation publication-type="journal">[3]	Sivic and Zisserman, &quot;Video google: a text retrieval approach to object matching in videos,&quot; in Proceedings Ninth IEEE International Conference on Computer Vision, pp. 1470-1477 vol.2, Oct 2003.</mixed-citation>
                    </ref>
                                    <ref id="ref4">
                        <label>4</label>
                        <mixed-citation publication-type="journal">[4]	H. Wang, Y. Cai, Y. Zhang, H. Pan, W. Lv, and H. Han, &quot;Deep learning for image retrieval: What works and what doesn&#039;t,&quot; in 2015 IEEE International Conference on Data Mining Workshop (ICDMW), pp. 1576-1583, Nov 2015.</mixed-citation>
                    </ref>
                                    <ref id="ref5">
                        <label>5</label>
                        <mixed-citation publication-type="journal">[5]	J. Yosinski, J. Clune, A. Nguyen, T. Fuchs, and H. Lipson, &quot;Understanding neural networks through deep visualization,&quot; in Deep Learning Workshop, 31. International Conference on Machine Learning (ICML), 2015.</mixed-citation>
                    </ref>
                                    <ref id="ref6">
                        <label>6</label>
                        <mixed-citation publication-type="journal">[6]	K. Simonyan and A. Zisserman, &quot;Very deep convolutional networks for large-scale image recognition,&quot; in Proc. of Workshop at Int. Conf. on Learning Representations (ICLR) Workshops, 2015.</mixed-citation>
                    </ref>
                                    <ref id="ref7">
                        <label>7</label>
                        <mixed-citation publication-type="journal">[7]	C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabinovich, &quot;Going deeper with convolutions,&quot; in 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1-9, June 2015.</mixed-citation>
                    </ref>
                                    <ref id="ref8">
                        <label>8</label>
                        <mixed-citation publication-type="journal">[8]	A. Krizhevsky, I. Sutskever, and G. E. Hinton, &quot;Imagenet classification with deep convolutional neural networks,&quot; in Advances in Neural Information Processing Systems 25 (F. Pereira, C. J. C. Burges, L. Bottou, and K. Q. Weinberger, eds.), pp. 1097-1105, Curran Associates, Inc., 2012.</mixed-citation>
                    </ref>
                                    <ref id="ref9">
                        <label>9</label>
                        <mixed-citation publication-type="journal">[9]	P. Sermanet, D. Eigen, X. Zhang, M. Mathieu, R. Fergus, and Y. LeCun, &quot;Overfeat: Integrated recognition, localization and detection using convolutional networks,&quot; CoRR, vol. abs/1312.6229, 2013.</mixed-citation>
                    </ref>
                                    <ref id="ref10">
                        <label>10</label>
                        <mixed-citation publication-type="journal">[10]	A. Babenko, A. Slesarev, A. Chigorin, and V. Lempitsky, &quot;Neural codes for image retrieval,&quot; in Computer Vision - ECCV 2014 (D. Fleet, T. Pajdla, B. Schiele, and T. Tuytelaars, eds.), (Cham), pp. 584-599, Springer International Publishing, 2014.</mixed-citation>
                    </ref>
                                    <ref id="ref11">
                        <label>11</label>
                        <mixed-citation publication-type="journal">[11]	V. Chandrasekhar, J. Lin, O. Morère, H. Goh, and A. Veillard, &quot;A practical guide to cnns and fisher vectors for image instance retrieval,&quot; Signal Process., vol. 128, pp. 426-439, Nov. 2016.</mixed-citation>
                    </ref>
                                    <ref id="ref12">
                        <label>12</label>
                        <mixed-citation publication-type="journal">[12]	I. Melekhov, J. Kannala, and E. Rahtu, &quot;Siamese network features for image matching,&quot; in 2016 23rd International Conference on Pattern Recognition (ICPR), pp. 378-383, Dec 2016.</mixed-citation>
                    </ref>
                                    <ref id="ref13">
                        <label>13</label>
                        <mixed-citation publication-type="journal">[13]	Y. Taigman, M. Yang, M. Ranzato, and L. Wolf, &quot;Deepface: Closing the gap to human-level performance in face verification,&quot; in 2014 IEEE Conference on Computer Vision and Pattern Recognition, pp. 1701-1708, June 2014.</mixed-citation>
                    </ref>
                                    <ref id="ref14">
                        <label>14</label>
                        <mixed-citation publication-type="journal">[14]	T. Lin, Y. Cui, S. Belongie, and J. Hays, &quot;Learning deep representations for ground-to-aerial geolocalization,&quot; in 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5007-5015, June 2015.</mixed-citation>
                    </ref>
                                    <ref id="ref15">
                        <label>15</label>
                        <mixed-citation publication-type="journal">[15]	D. Cai, X. Gu, and C. Wang, &quot;A revisit on deep hashings for large-scale content based image retrieval,&quot; CoRR, vol. abs/1711.06016, 2017.</mixed-citation>
                    </ref>
                                    <ref id="ref16">
                        <label>16</label>
                        <mixed-citation publication-type="journal">[16]	R. Datta, J. Li, and J. Z. Wang, &quot;Content-based image retrieval: Approaches and trends of the new age,&quot; in Proceedings of the 7th ACM SIGMM International Workshop on Multimedia Information Retrieval, MIR &#039;05, (New York, NY, USA), pp. 253-262, ACM, 2005.</mixed-citation>
                    </ref>
                                    <ref id="ref17">
                        <label>17</label>
                        <mixed-citation publication-type="journal">[17]	P. Clough, H. Müller, T. Deselaers, M. Grubinger, T. Martin Lehmann, J. R. Jensen, and W. Hersh, &quot;The clef 2005 cross-language image retrieval track,&quot; CEUR Workshop Proceedings, vol. 1171, pp. 535-557, 09 2005.</mixed-citation>
                    </ref>
                                    <ref id="ref18">
                        <label>18</label>
                        <mixed-citation publication-type="journal">[18]	G. Schaefer, &quot;Ucid-raw - a colour image database in raw format,&quot; in VipIMAGE 2017, (Cham), pp. 179-184, Springer International Publishing, 2018.</mixed-citation>
                    </ref>
                                    <ref id="ref19">
                        <label>19</label>
                        <mixed-citation publication-type="journal">[19]	T. Ahmad, P. Campr, M. Cadik, and G. Bebis, &quot;Comparison of semantic segmentation approaches for horizon/sky line detection,&quot; 2017 International Joint Conference on Neural Networks (IJCNN), May 2017.</mixed-citation>
                    </ref>
                                    <ref id="ref20">
                        <label>20</label>
                        <mixed-citation publication-type="journal">[20]	F. Jiang, A. Grigorev, S. Rho, Z. Tian, Y. Fu, W. Jifara, A. Khan, and S. Liu, &quot;Medical image semantic segmentation based on deep learning,&quot; Neural Computing and Applications, 07 2017.</mixed-citation>
                    </ref>
                                    <ref id="ref21">
                        <label>21</label>
                        <mixed-citation publication-type="journal">[21]	M. Siam, S. Elkerdawy, M. Jägersand, and S. Yogamani, &quot;Deep semantic segmentation for automated driving: Taxonomy, roadmap and challenges,&quot; in 20th IEEE International Conference on Intelligent Transportation Systems, ITSC 2017, Yokohama, Japan, October 16-19, 2017, pp. 1-8, 2017.</mixed-citation>
                    </ref>
                                    <ref id="ref22">
                        <label>22</label>
                        <mixed-citation publication-type="journal">[22]	M. Thoma, &quot;A survey of semantic segmentation,&quot; arXiv preprint arXiv:1602.06541, 2016.</mixed-citation>
                    </ref>
                                    <ref id="ref23">
                        <label>23</label>
                        <mixed-citation publication-type="journal">[23]	M. H. Saffar, M. Fayyaz, M. Sabokrou, and M. Fathy, &quot;Semantic video segmentation: A review on recent approaches,&quot; 2018.</mixed-citation>
                    </ref>
                                    <ref id="ref24">
                        <label>24</label>
                        <mixed-citation publication-type="journal">[24]	H. Yu, Z. Yang, L. Tan, Y. Wang, W. Sun, M. Sun, and Y. Tang, &quot;Methods and datasets on semantic segmentation: A review,&quot; Neurocomputing, vol. 304, pp. 82 - 103, 2018.</mixed-citation>
                    </ref>
                                    <ref id="ref25">
                        <label>25</label>
                        <mixed-citation publication-type="journal">[25]	Y. Guo, Y. Liu, T. Georgiou, and M. S. Lew, &quot;A review of semantic segmentation using deep neural networks,&quot; International Journal of Multimedia Information Retrieval, vol. 7, pp. 87-93, Jun 2018.</mixed-citation>
                    </ref>
                                    <ref id="ref26">
                        <label>26</label>
                        <mixed-citation publication-type="journal">[26]	A. Garcia-Garcia, S. Orts-Escolano, S. Oprea, V. Villena-Martinez, and J. G. Rodríguez, &quot;A review on deep learning techniques applied to semantic segmentation,&quot; CoRR, vol. abs/1704.06857, 2017.</mixed-citation>
                    </ref>
                                    <ref id="ref27">
                        <label>27</label>
                        <mixed-citation publication-type="journal">[27]	E. Shelhamer, J. Long, and T. Darrell, &quot;Fully convolutional networks for semantic segmentation,&quot; IEEE Trans. Pattern Anal. Mach. Intell., vol. 39, pp. 640-651, Apr. 2017.</mixed-citation>
                    </ref>
                                    <ref id="ref28">
                        <label>28</label>
                        <mixed-citation publication-type="journal">[28]	A. Vedaldi and K. Lenc, &quot;Matconvnet: Convolutional neural networks for matlab,&quot; in Proceedings of the 23rd ACM International Conference on Multimedia, MM &#039;15, pp. 689-692, 2015.</mixed-citation>
                    </ref>
                                    <ref id="ref29">
                        <label>29</label>
                        <mixed-citation publication-type="journal">[29]	H. Zhang, J. E. Fritts, and S. A. Goldman, &quot;Image segmentation evaluation: A survey of unsupervised methods,&quot; Computer Vision and Image Understanding, vol. 110, no. 2, pp. 260 - 280, 2008.</mixed-citation>
                    </ref>
                                    <ref id="ref30">
                        <label>30</label>
                        <mixed-citation publication-type="journal">[30]	J. Harel, C. Koch, and P. Perona, &quot;Graph-based visual saliency,&quot; in Proc. Advances in Neural Information Processing Systems, (NIPS), pp. 545-552, 2006.</mixed-citation>
                    </ref>
                                    <ref id="ref31">
                        <label>31</label>
                        <mixed-citation publication-type="journal">[31]	Y. Xu, J. Li, J. Chen, G. Shen, and Y. Gao, &quot;A novel approach for visual saliency detection and segmentation based on objectness and top-down attention,&quot; in 2017 2nd International Conference on Image, Vision and Computing (ICIVC), pp. 361-365, June 2017.</mixed-citation>
                    </ref>
                                    <ref id="ref32">
                        <label>32</label>
                        <mixed-citation publication-type="journal">[32]	J. Lankinen, V. Kangas, and J. Kamarainen, &quot;A comparison of local feature detectors and descriptors for visual object categorization by intra-class repeatability and matching,&quot; in Proceedings of the 21st International Conference on Pattern Recognition (ICPR2012), pp. 780-783, Nov 2012.</mixed-citation>
                    </ref>
                                    <ref id="ref33">
                        <label>33</label>
                        <mixed-citation publication-type="journal">[33]	A. Oliva and A. Torralba, &quot;Modeling the shape of the scene: A holistic representation of the spatial envelope,&quot; Int. J. Comput. Vision, vol. 42, pp. 145-175, May 2001.</mixed-citation>
                    </ref>
                                    <ref id="ref34">
                        <label>34</label>
                        <mixed-citation publication-type="journal">[34]	N. Dalal and B. Triggs, &quot;Histograms of oriented gradients for human detection,&quot; in 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR&#039;05), vol. 1, pp. 886-893 vol. 1, June 2005.</mixed-citation>
                    </ref>
                                    <ref id="ref35">
                        <label>35</label>
                        <mixed-citation publication-type="journal">[35]	D. G. Lowe, &quot;Distinctive image features from scale-invariant keypoints,&quot; International Journal of Computer Vision, vol. 60, pp. 91-110, Nov 2004.</mixed-citation>
                    </ref>
                                    <ref id="ref36">
                        <label>36</label>
                        <mixed-citation publication-type="journal">[36]	H. Bay, T. Tuytelaars, and L. Van Gool, &quot;Surf: Speeded up robust features,&quot; in Computer Vision - ECCV 2006 (A. Leonardis, H. Bischof, and A. Pinz, eds.), (Berlin, Heidelberg), pp. 404-417, Springer Berlin Heidelberg, 2006.</mixed-citation>
                    </ref>
                                    <ref id="ref37">
                        <label>37</label>
                        <mixed-citation publication-type="journal">[37]	K. He, X. Zhang, S. Ren, and J. Sun, &quot;Spatial pyramid pooling in deep convolutional networks for visual recognition,&quot; CoRR, vol. abs/1406.4729, 2014.</mixed-citation>
                    </ref>
                                    <ref id="ref38">
                        <label>38</label>
                        <mixed-citation publication-type="journal">[38]	C. Hentschel and H. Sack, &quot;Does one size really fit all?: Evaluating classifiers in bag-of-visual-words classification,&quot; in Proceedings of the 14th International Conference on Knowledge Technologies and Data-driven Business, iKNOW &#039;14, (New York, NY, USA), pp. 7:1-7:8, ACM, 2014.</mixed-citation>
                    </ref>
                                    <ref id="ref39">
                        <label>39</label>
                        <mixed-citation publication-type="journal">[39]	L. I. Kuncheva, &quot;On the optimality of naïve bayes with dependent binary features,&quot; Pattern Recogn. Lett., vol. 27, pp. 830-837, May 2006.</mixed-citation>
                    </ref>
                                    <ref id="ref40">
                        <label>40</label>
                        <mixed-citation publication-type="journal">[40]	K. Lenc and A. Vedaldi, &quot;Understanding image representations by measuring their equivariance and equivalence,&quot; in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 991-999, 2015.</mixed-citation>
                    </ref>
                            </ref-list>
                    </back>
    </article>
