EN
TR
Context-aware CLIP for Enhanced Food Recognition
Abstract
Generalization of food image recognition frameworks is difficult due to the wide variety of food categories in cuisines across cultures. The performance of the deep neural network models highly depends on the training dataset. To overcome this problem, we propose to extract context information from images in order to increase the discrimination capacity of networks. In this work, we utilize the CLIP architecture with the automatically derived ingredient context from food images. A list of ingredients are associated with each food category, which is later modeled as text after a voting process and fed to a CLIP architecture together with input image. Experimental results on the Food101 dataset show that this approach significantly improves the model’s performance, achieving a 2% overall increase in accuracy. This improvement varies across food classes, with increases ranging from 0.5% to as much as 22%. The proposed framework, CLIP fed with ingredient text, outperforms Yolov8 (81.46%) with 81.80% top 1 overall accuracy over 101 classes.
Keywords
References
- Chen X, Kamavuako EN. “Vision-based methods for food and fluid intake monitoring: A literature review”, Sensors, (2023) 23(13), 2023.
- Ponte D et al. “Ontologydriven deep learning model for multitask visual food analysis”, VISIGRAPP (2024) 624-631.
- Zhang Y et al. “Deep learning in food category recognition”, Information Fusion, (2023) 98:101859.
- Zhao H et al. “Fusion learning using semantics and graph convolutional network for visual food recognition”, In WACV, (2021) 1710–1719.
- Liu C et al. “Deepfood: Deep learning-based food image recognition for computer-aided dietary assessment”, In International Conference on Smart Homes and Health Telematics”, Springer Intl. Publishing, (2016) 37–48.
- Shuqiang J et al. “Few-shot food recognition via multi-view representation learning”, ACM Transactions on Multi-media Computing, Communications and Applications, (2020). 1-4.
- Yang J et al. “Learning to classify new foods incrementally via compressed exemplars”, CVPRW, (2024) 3695-3704.
- Ergun OO, Ozturk B. “An ontology based semantic representation for turkish cuisine”, In 26th Signal Processing and Communications Applications Conference (SIU), (2018) 1–4.
Details
Primary Language
English
Subjects
Computer Vision, Artificial Intelligence (Other)
Journal Section
Research Article
Authors
Early Pub Date
June 16, 2025
Publication Date
June 16, 2025
Submission Date
May 28, 2025
Acceptance Date
June 9, 2025
Published in Issue
Year 2025 Volume: 5 Number: 1
APA
Öztürk Ergün, Ö. (2025). Context-aware CLIP for Enhanced Food Recognition. Advances in Artificial Intelligence Research, 5(1), 7-13. https://doi.org/10.54569/aair.1707867
AMA
1.Öztürk Ergün Ö. Context-aware CLIP for Enhanced Food Recognition. Adv. Artif. Intell. Res. 2025;5(1):7-13. doi:10.54569/aair.1707867
Chicago
Öztürk Ergün, Övgü. 2025. “Context-Aware CLIP for Enhanced Food Recognition”. Advances in Artificial Intelligence Research 5 (1): 7-13. https://doi.org/10.54569/aair.1707867.
EndNote
Öztürk Ergün Ö (June 1, 2025) Context-aware CLIP for Enhanced Food Recognition. Advances in Artificial Intelligence Research 5 1 7–13.
IEEE
[1]Ö. Öztürk Ergün, “Context-aware CLIP for Enhanced Food Recognition”, Adv. Artif. Intell. Res., vol. 5, no. 1, pp. 7–13, June 2025, doi: 10.54569/aair.1707867.
ISNAD
Öztürk Ergün, Övgü. “Context-Aware CLIP for Enhanced Food Recognition”. Advances in Artificial Intelligence Research 5/1 (June 1, 2025): 7-13. https://doi.org/10.54569/aair.1707867.
JAMA
1.Öztürk Ergün Ö. Context-aware CLIP for Enhanced Food Recognition. Adv. Artif. Intell. Res. 2025;5:7–13.
MLA
Öztürk Ergün, Övgü. “Context-Aware CLIP for Enhanced Food Recognition”. Advances in Artificial Intelligence Research, vol. 5, no. 1, June 2025, pp. 7-13, doi:10.54569/aair.1707867.
Vancouver
1.Övgü Öztürk Ergün. Context-aware CLIP for Enhanced Food Recognition. Adv. Artif. Intell. Res. 2025 Jun. 1;5(1):7-13. doi:10.54569/aair.1707867
