[{“box”:0,”content”:”[if 992 equals=”Open Access”]n
n
Open Access
nn
n
n[/if 992]n
n
n
n
n

n
Om Sanjay Awari, Shivaraj Mahendra Ghangale, Janmejay Aklesh Maurya, Aishwarya Ajit Ghangale, Keerti Kharatmol,
n
- n t
n
n
n[/foreach]
n
n[if 2099 not_equal=”Yes”]n
- [foreach 286] [if 1175 not_equal=””]n t
- Student, Student, Student, Student, Assistant Professor Department of Computer Science and Engineering, K.C. College of Engineering & Management Studies & Research, Thane, Department of Computer Science and Engineering, K.C. College of Engineering & Management Studies & Research, Thane, Department of Computer Science and Engineering, K.C. College of Engineering & Management Studies & Research, Thane, Department of Computer Science and Engineering, K.C. College of Engineering & Management Studies & Research, Thane, Department of Computer Science and Engineering, K.C. College of Engineering & Management Studies & Research, Thane Maharashtra, Maharashtra, Maharashtra, Maharashtra, Maharashtra India, India, India, India, India
n[/if 1175][/foreach]
n[/if 2099][if 2099 equals=”Yes”][/if 2099]n
Abstract
nMachine learning has become a crucial part of modern life, influencing various domains. Its applications range from enhancing data-driven business decisions to enabling autonomous vehicles. Advances in ML have brought about notable changes in how we interact with technology. In the culinary world, the idea of creating food recipes from images has gained increasing interest. This entails the development of innovative systems that seamlessly convert visual input, such as images of dishes, into comprehensive culinary instructions. This process involves a fusion of advanced computer vision techniques and natural language generation algorithms, enabling a bridge between visual data and textual information. Several cutting-edge systems have been proposed to tackle the challenge of generating recipes from images. These systems employ sophisticated methodologies that leverage object detection and recognition to accurately identify ingredients and quantities within the images. By consulting extensive recipe databases, these systems generate step-by-step cooking instructions. The applications of such systems are far-reaching. They serve a wide range of users, from seasoned chefs looking for new ideas to beginners needing cooking advice. This technology introduces a novel aspect to cooking by making culinary knowledge more accessible, encouraging creativity, and improving the cooking experience. The advantages of using ML for recipe generation are numerous. It acts as a digital sous-chef, offering personalized recommendations tailored to user preferences and dietary needs. Additionally, the integration of user feedback and continuous refinement through advanced ML techniques ensures the adaptability and accuracy of these systems. Moreover, the seamless integration of such systems with smart cooking devices promises to revolutionize the way individuals approach cooking, making it more accessible, interactive, and enjoyable for everyone.
n
Keywords: Deep Learning, recipe generation, cooking instructions, food images, culinary domain
n[if 424 equals=”Regular Issue”][This article belongs to Journal of Open Source Developments(joosd)]
n
n
n
n
n
nn[if 992 equals=”Open Access”] Full Text PDF Download[/if 992] n
nn[if 379 not_equal=””]n
Browse Figures
n
n
n[/if 379]n
References
n[if 1104 equals=””]n
- Chen JJ, Ngo CW, Chua TS. Cross-modal recipe retrieval with rich food attributes. InProceedings of the 25th ACM international conference on Multimedia 2017 Oct 23 (pp. 1771-1779).
- Salvador A, Drozdzal M, Giró-i-Nieto X, Romero A. Inverse cooking: Recipe generation from food images. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2019 (pp. 10453-10462).
- Han F, Guerrero R, Pavlovic V. CookGAN: Meal image synthesis from ingredients. InProceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision 2020 (pp. 1450-1458).
- Salvador A, Hynes N, Aytar Y, Marin J, Ofli F, Weber I, Torralba A. Learning cross-modal embeddings for cooking recipes and food images. InProceedings of the IEEE conference on computer vision and pattern recognition 2017 (pp. 3020-3028).
- Pan S, Dai L, Hou X, Li H, Sheng B. ChefGAN: Food image generation from recipes. InProceedings of the 28th ACM International Conference on Multimedia 2020 Oct 12 (pp. 4244-4252).
- Salvador A, Gundogdu E, Bazzani L, Donoser M. Revamping cross-modal recipe retrieval with hierarchical transformers and self-supervised learning. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2021 (pp. 15475-15484).
- Fujita J, Sato M, Nobuhara H. Model for cooking recipe generation using reinforcement learning. In2021 IEEE 37th International Conference on Data Engineering Workshops (ICDEW) 2021 Apr 19 (pp. 1-4). IEEE.
- Chen J, Pang L, Ngo CW. Cross-modal recipe retrieval: How to cook this dish?. InMultiMedia Modeling: 23rd International Conference, MMM 2017, Reykjavik, Iceland, January 4-6, 2017, Proceedings, Part I 23 2017 (pp. 588-600). Springer International Publishing.
- Zhu B, Ngo CW, Chen J, Hao Y. R2gan: Cross-modal recipe retrieval with generative adversarial network. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2019 (pp. 11477-11486).
- Bettadapura V, Thomaz E, Parnami A, Abowd GD, Essa I. Leveraging context to support automated food recognition in restaurants. In2015 IEEE Winter Conference on Applications of Computer Vision 2015 Jan 5 (pp. 580-587). IEEE.
- Zuo S, Xiao Y, Chang X, Wang X. Vision transformers for dense prediction: A survey. Knowledge-Based Systems. 2022 Oct 11;253:109552.
- Salvador A, Drozdzal M, Giró-i-Nieto X, Romero A. Inverse cooking: Recipe generation from food images. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2019 (pp. 10453-10462).
- Salvador A, Hynes N, Aytar Y, Marin J, Ofli F, Weber I, Torralba A. Learning cross-modal embeddings for cooking recipes and food images. InProceedings of the IEEE conference on computer vision and pattern recognition 2017 (pp. 3020-3028).
- Chhikara P, Chaurasia D, Jiang Y, Masur O, Ilievski F. Fire: Food image to recipe generation. InProceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision 2024 (pp. 8184-8194).
- Wang L, Li Q, Li N, Dong G, Yang Y. Substructure similarity measurement in chinese recipes. InProceedings of the 17th international conference on World Wide Web 2008 Apr 21 (pp. 979-988).
- Chen J, Ngo CW. Deep-based ingredient recognition for cooking recipe retrieval. InProceedings of the 24th ACM international conference on Multimedia 2016 Oct 1 (pp. 32-41).
- Wang X, Kumar D, Thome N, Cord M, Precioso F. Recipe recognition with large multimodal food dataset. In2015 IEEE International Conference on Multimedia & Expo Workshops (ICMEW) 2015 Jun 29 (pp. 1-6). IEEE.
- Wang S, Gao H, Zhu Y, Zhang W, Chen Y. A food dish image generation framework based on progressive growing GANs. InCollaborative Computing: Networking, Applications and Worksharing: 15th EAI International Conference, CollaborateCom 2019, London, UK, August 19-22, 2019, Proceedings 15 2019 (pp. 323-333). Springer International Publishing.
- Martinel N, Foresti GL, Micheloni C. Wide-slice residual networks for food recognition. In2018 IEEE Winter Conference on applications of computer vision (WACV) 2018 Mar 12 (pp. 567-576). IEEE.
- Teng CY, Lin YR, Adamic LA. Recipe recommendation using ingredient networks. InProceedings of the 4th annual ACM web science conference 2012 Jun 22 (pp. 298-307).
- Zhang XJ, Lu YF, Zhang SH. Multi-task learning for food identification and analysis with deep convolutional neural networks. Journal of Computer Science and Technology. 2016 May;31(3):489-500.
nn[/if 1104][if 1104 not_equal=””]n
- [foreach 1102]n t
- [if 1106 equals=””], [/if 1106][if 1106 not_equal=””],[/if 1106]
n[/foreach]
n[/if 1104]
nn
nn[if 1114 equals=”Yes”]n
n[/if 1114]
n
n
n
| Volume | 11 | |
| [if 424 equals=”Regular Issue”]Issue[/if 424][if 424 equals=”Special Issue”]Special Issue[/if 424] [if 424 equals=”Conference”][/if 424] | 02 | |
| Received | June 28, 2024 | |
| Accepted | July 31, 2024 | |
| Published | August 7, 2024 |
n
n
n
n
n
n nfunction myFunction2() {nvar x = document.getElementById(“browsefigure”);nif (x.style.display === “block”) {nx.style.display = “none”;n}nelse { x.style.display = “Block”; }n}ndocument.querySelector(“.prevBtn”).addEventListener(“click”, () => {nchangeSlides(-1);n});ndocument.querySelector(“.nextBtn”).addEventListener(“click”, () => {nchangeSlides(1);n});nvar slideIndex = 1;nshowSlides(slideIndex);nfunction changeSlides(n) {nshowSlides((slideIndex += n));n}nfunction currentSlide(n) {nshowSlides((slideIndex = n));n}nfunction showSlides(n) {nvar i;nvar slides = document.getElementsByClassName(“Slide”);nvar dots = document.getElementsByClassName(“Navdot”);nif (n > slides.length) { slideIndex = 1; }nif (n (item.style.display = “none”));nArray.from(dots).forEach(nitem => (item.className = item.className.replace(” selected”, “”))n);nslides[slideIndex – 1].style.display = “block”;ndots[slideIndex – 1].className += ” selected”;n}n”}]
