<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="review-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMIR</journal-id>
      <journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id>
      <journal-title>Journal of Medical Internet Research</journal-title>
      <issn pub-type="epub">1438-8871</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v26i1e51432</article-id>
      <article-id pub-id-type="pmid">39546777</article-id>
      <article-id pub-id-type="doi">10.2196/51432</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Review</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Review</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Advancements in Using AI for Dietary Assessment Based on Food Images: Scoping Review</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Coristine</surname>
            <given-names>Andrew</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Jia</surname>
            <given-names>Wenyan</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>DiFilippo</surname>
            <given-names>Kristen</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author">
          <name name-style="western">
            <surname>Chotwanvirat</surname>
            <given-names>Phawinpon</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-4714-790X</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author">
          <name name-style="western">
            <surname>Prachansuwan</surname>
            <given-names>Aree</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff3" ref-type="aff">3</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-3062-276X</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>Sridonpai</surname>
            <given-names>Pimnapanut</given-names>
          </name>
          <degrees>MSc</degrees>
          <xref rid="aff3" ref-type="aff">3</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-9669-4160</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author" corresp="yes">
          <name name-style="western">
            <surname>Kriengsinyos</surname>
            <given-names>Wantanee</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff3" ref-type="aff">3</xref>
          <address>
            <institution>Human Nutrition Unit, Food and Nutrition Academic and Research Cluster, Institute of Nutrition</institution>
            <institution>Mahidol University</institution>
            <addr-line>999 Phutthamonthon 4 Rd., Salaya</addr-line>
            <addr-line>Nakhon Pathom, 73170</addr-line>
            <country>Thailand</country>
            <phone>66 2 800 2380</phone>
            <fax>66 2 441 9344</fax>
            <email>wantanee.krieng@mahidol.ac.th</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-8262-5095</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>Theptarin Diabetes, Thyroid, and Endocrine Center</institution>
        <institution>Vimut-Theptarin Hospital</institution>
        <addr-line>Bangkok</addr-line>
        <country>Thailand</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>Diabetes and Metabolic Care Center</institution>
        <institution>Taksin Hospital</institution>
        <institution>Medical Service Department, Bangkok Metropolitan Administration</institution>
        <addr-line>Bangkok</addr-line>
        <country>Thailand</country>
      </aff>
      <aff id="aff3">
        <label>3</label>
        <institution>Human Nutrition Unit, Food and Nutrition Academic and Research Cluster, Institute of Nutrition</institution>
        <institution>Mahidol University</institution>
        <addr-line>Nakhon Pathom</addr-line>
        <country>Thailand</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Wantanee Kriengsinyos <email>wantanee.krieng@mahidol.ac.th</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <year>2024</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>15</day>
        <month>11</month>
        <year>2024</year>
      </pub-date>
      <volume>26</volume>
      <elocation-id>e51432</elocation-id>
      <history>
        <date date-type="received">
          <day>31</day>
          <month>7</month>
          <year>2023</year>
        </date>
        <date date-type="rev-request">
          <day>15</day>
          <month>2</month>
          <year>2024</year>
        </date>
        <date date-type="rev-recd">
          <day>13</day>
          <month>6</month>
          <year>2024</year>
        </date>
        <date date-type="accepted">
          <day>24</day>
          <month>9</month>
          <year>2024</year>
        </date>
      </history>
      <copyright-statement>©Phawinpon Chotwanvirat, Aree Prachansuwan, Pimnapanut Sridonpai, Wantanee Kriengsinyos. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 15.11.2024.</copyright-statement>
      <copyright-year>2024</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://www.jmir.org/2024/1/e51432" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>To accurately capture an individual’s food intake, dietitians are often required to ask clients about their food frequencies and portions, and they have to rely on the client’s memory, which can be burdensome. While taking food photos alongside food records can alleviate user burden and reduce errors in self-reporting, this method still requires trained staff to translate food photos into dietary intake data. Image-assisted dietary assessment (IADA) is an innovative approach that uses computer algorithms to mimic human performance in estimating dietary information from food images. This field has seen continuous improvement through advancements in computer science, particularly in artificial intelligence (AI). However, the technical nature of this field can make it challenging for those without a technical background to understand it completely.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>This review aims to fill the gap by providing a current overview of AI’s integration into dietary assessment using food images. The content is organized chronologically and presented in an accessible manner for those unfamiliar with AI terminology. In addition, we discuss the systems’ strengths and weaknesses and propose enhancements to improve IADA’s accuracy and adoption in the nutrition community.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>This scoping review used PubMed and Google Scholar databases to identify relevant studies. The review focused on computational techniques used in IADA, specifically AI models, devices, and sensors, or digital methods for food recognition and food volume estimation published between 2008 and 2021.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>A total of 522 articles were initially identified. On the basis of a rigorous selection process, 84 (16.1%) articles were ultimately included in this review. The selected articles reveal that early systems, developed before 2015, relied on handcrafted machine learning algorithms to manage traditional sequential processes, such as segmentation, food identification, portion estimation, and nutrient calculations. Since 2015, these handcrafted algorithms have been largely replaced by deep learning algorithms for handling the same tasks. More recently, the traditional sequential process has been superseded by advanced algorithms, including multitask convolutional neural networks and generative adversarial networks. Most of the systems were validated for macronutrient and energy estimation, while only a few were capable of estimating micronutrients, such as sodium. Notably, significant advancements have been made in the field of IADA, with efforts focused on replicating humanlike performance.</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>This review highlights the progress made by IADA, particularly in the areas of food identification and portion estimation. Advancements in AI techniques have shown great potential to improve the accuracy and efficiency of this field. However, it is crucial to involve dietitians and nutritionists in the development of these systems to ensure they meet the requirements and trust of professionals in the field.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>image-assisted dietary assessment</kwd>
        <kwd>artificial intelligence</kwd>
        <kwd>dietary assessment</kwd>
        <kwd>mobile phone</kwd>
        <kwd>food intake</kwd>
        <kwd>image recognition</kwd>
        <kwd>portion size</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <sec>
        <title>Background</title>
        <p>Dietary assessment is a technique for determining an individual’s intake, eating patterns, and food quality choices, as well as the nutritional values of consumed food. However, this technique’s procedures are costly, laborious, and time-consuming and rely on specially trained personnel (such as dietitians and nutritionists) to produce reliable results. Consequently, a strong need exists for novel methods having improved measurement capabilities that are accurate, convenient, less burdensome, and cost-effective [<xref ref-type="bibr" rid="ref1">1</xref>]. Instead of relying solely on client self-report, taking food photos before eating has been incorporated into traditional methods, such as a 3-day food record with food images, to reduce missing food records, incorrect food identification, and errors in portion size estimation. However, this technique still requires well-trained staff to translate food image information into reliable nutritional values and does not solve labor-intensive and time-consuming issues.</p>
        <p>The application of computer algorithms to translate food images into representative nutritional values has gained interest in both the nutrition and computer science communities. This combination has resulted in a new field called image-assisted dietary assessment (IADA), and various systems have been developed to address these limitations, ranging from simple estimation equations in early systems to more complex artificial intelligence (AI) models in recent years. By applying IADA alongside the increasing use of smartphones and devices with built-in digital cameras, real-time analysis of dietary intake data from food images has become possible with accurate results, reduced labor, and greater convenience, thus gaining attention among nutrition professionals. However, the technical nature of this field can make it difficult to understand for those without a background in computer science or engineering, leading to the low involvement of nutrition professionals in its development. This gap is the rationale for us to conduct this review.</p>
      </sec>
      <sec>
        <title>Objectives</title>
        <p>The objective of this review is to bridge that knowledge gap by providing an up-to-date overview of the gradual enhancement of AI integration in dietary assessment based on food images. The information is presented in chronological order and in a manner that is understandable and accessible to those who may not be familiar with the technical jargon and complexity of AI terminologies. In addition, the advantages and limitations of these systems are discussed. Finally, we proposed auxiliary systems to enhance the accuracy of IADA and its potential adoption within the nutrition community.</p>
      </sec>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Overview</title>
        <p>To conduct this scoping review, we followed the methodology suggested by Arksey and O’Malley [<xref ref-type="bibr" rid="ref2">2</xref>] and adhered to the PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews) guidelines [<xref ref-type="bibr" rid="ref3">3</xref>].</p>
      </sec>
      <sec>
        <title>Search Strategy</title>
        <p>We searched 2 web-based databases, PubMed and Google Scholar, between February 2023 and March 2023, using the following terms: ((“food image”[Title/Abstract]) AND (classification[Title/Abstract] OR recognition[Title/Abstract] OR (“computer vision”[Title/Abstract]))) and “artificial intelligence,” “dietary assessment,” “computer vision,” “food image” recognition, “portion size,” segmentation, and classification, respectively.</p>
      </sec>
      <sec>
        <title>Eligibility Criteria</title>
        <p>This review included studies that focused on AI techniques used for IADA, specifically AI models, systems, or digital methods for food recognition and food volume estimation. For mobile apps or systems, we considered only articles that explain algorithms beyond mobile apps, prototype testing, or conducting clinical research. Studies that used noncomputational techniques, such as using food images as a tool for training human portion estimation, are excluded. Eligible articles were published in peer-reviewed journals or conference papers and written in English.</p>
      </sec>
      <sec>
        <title>Selection Process</title>
        <p>We used Zotero (Corporation for Digital Scholarship) reference management software to collect search results using the add multiple results function. All automatic data retrieval functions were disabled to prevent data retrieval from exceeding Google Scholar’s traffic limitation. Zotero’s built-in duplicate merger was used to identify duplicated records, and unduplicated records were exported to Excel online (Microsoft Corp). In Excel, all authors independently screened article types, titles, and abstracts. The screening process removed all nonrelated titles or abstracts, review and editorial articles, non-English articles, or conference abstracts without full text. For thesis articles, the corresponding published articles were identified using keywords from the title, first author, or corresponding author whenever possible. Each article required 2 independent reviewers’ approval. In cases of conflict, a full-text review was necessary to resolve disagreements. After the initial screening process, the full texts of articles were obtained to assess eligibility. All full-text articles, whether they were excluded or not, and review articles were thoroughly read to identify interesting or related articles. These were classified as articles from other sources.</p>
      </sec>
      <sec>
        <title>Data Extraction</title>
        <p>A data extraction table was constructed, including the system name, classification algorithm, portion size estimation algorithm, accuracy of classification or portion estimated results, and the system’s noticeable advantages and drawbacks. Data were extracted from full texts.</p>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Literature Findings</title>
        <p>We retrieved 44 (8.4%) items from PubMed, while Google Scholar provided 478 (91.6%) results from the search terms, giving a total of 522 items retrieved. In total, 122 (23.4%) duplicate items were removed using Zotero’s built-in duplicate merger. The remaining 400 (76.6%) deduplicated items were screened based on their titles and abstracts, resulting in 104 (19.9%) records for full-text review. After the full-text review process, 72 (13.8%) articles were included in this study. In addition, we manually identified and included 12 (2.3%) additional articles from other sources. An overview of the literature identification method and results is shown in <xref rid="figure1" ref-type="fig">Figure 1</xref>, and the PRISMA-ScR checklist is available in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p>
        <fig id="figure1" position="float">
          <label>Figure 1</label>
          <caption>
            <p>PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews) flowchart of the structured literature search, screening, and selection methodology.</p>
          </caption>
          <graphic xlink:href="jmir_v26i1e51432_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Traditional Dietary Assessment Methods</title>
        <p>When measuring individual food intake, dietary assessment methods are typically divided into 2 sequential processes: methods to obtain dietary intake and methods to estimate the nutritional values of food. Principally, obtaining an individual’s intake can be done by recording all consumed foods, beverages, herbs, or supplements with their portion sizes on a day-to-day basis or within a specific time frame (eg, a week) based on variation in the nutrients of interest. These methods were developed early on and can be performed manually. Due to their simplicity, some methods are frequently used in nutrition professionals’ practices.</p>
        <p>The 24-hour dietary recall (24HR) method is the simplest way to measure dietary intake, but accurately obtaining dietary intake information can be very challenging. The participant or their caregiver are asked by a trained interviewer to recall the participant’s food intake within the last 24 hours. This method relies heavily on the client’s memory and estimation of food portion size [<xref ref-type="bibr" rid="ref4">4</xref>]. Unintentional misreporting of food intake is common, as clients often forget some foods. Underreporting of portion size is common because clients are not familiar with estimating food portion sizes [<xref ref-type="bibr" rid="ref5">5</xref>,<xref ref-type="bibr" rid="ref6">6</xref>]. In participants who are overweight or obese, intentional underreporting is also common [<xref ref-type="bibr" rid="ref7">7</xref>]. Although this method is the simplest for determining dietary intake, it takes approximately 1 hour to complete each interview. Moreover, a single 24HR result does not satisfactorily define an individual’s usual intake due to day-to-day variations in eating habits.</p>
        <p>Estimated food records (EFRs) are more reliable but time-consuming. Clients are asked to record all food and beverage intake during eating times for a specified period. Details of food are needed along with the portion sizes estimated by the client and rounded to household units (eg, half cup of soymilk with ground sesame and 4 tablespoons of kidney beans without syrup). To improve accuracy, training in estimating portion size using standard food models is required. The EFR places a burden on the clients, as they need to record all eating times. Moreover, some clients temporarily change their intake habits during recording to minimize this burden, while others may intentionally not report certain foods to cover up certain eating habits. Food portion size estimation errors are sometimes found, but taking food photographs before and after eating can lower these errors [<xref ref-type="bibr" rid="ref8">8</xref>-<xref ref-type="bibr" rid="ref12">12</xref>].</p>
        <p>A standardized weighing scale can be used to avoid errors caused by human estimation of portion sizes. This technique is known as weighed food records and is considered the gold standard for determining personal intake. However, it is impractical to weigh all eaten food in the long term because it becomes a burden for the client to measure the weight of food eaten throughout the day [<xref ref-type="bibr" rid="ref4">4</xref>]. This technique also only eliminates portion size estimation errors, while other issues with EFRs may still persist.</p>
        <p>After retrieving dietary intake information from sources, such as 24HR, EFR, or weighed food records, the next step is to estimate the representative nutritional value of the food using a food composition table. If the recorded foods match the food items and their description in an available food composition table, the nutritional values can be obtained by multiplying the consumed food weight directly. However, if the food items are not found, the food needs to be analyzed and broken down into its components. The nutritional values of each component can then be obtained from the food composition table (or its nutrition label) and multiplied by the actual weight of each consumed component. When the portion size is recorded instead of its actual weight, the estimated weight can be obtained using standardized portion sizes from the food composition table. Nutrient analysis software can easily accomplish this task.</p>
      </sec>
      <sec>
        <title>IADA Methods</title>
        <sec>
          <title>Overview</title>
          <p>Digital devices are often used for dietary assessment. The first well-documented attempt to develop such a digital device was called Wellnavi by Wang et al [<xref ref-type="bibr" rid="ref8">8</xref>]. Although the device yielded accurate results, its usability was limited by the technologies of the time, including short battery life, poor image quality, a bulky body, and a less sensitive touch screen [<xref ref-type="bibr" rid="ref10">10</xref>].</p>
          <p>Several attempts have been made to use generic devices, such as Palm (Palm Inc) PDAs [<xref ref-type="bibr" rid="ref13">13</xref>], compact digital cameras [<xref ref-type="bibr" rid="ref14">14</xref>], and smartphones [<xref ref-type="bibr" rid="ref15">15</xref>], instead of inventing a specific food recording device. In using these devices, users reported a decrease in the burden of completing food recording when compared with traditional methods [<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref17">17</xref>]. However, these devices still rely heavily on dietitians or nutritionists to analyze the nutritional values of food items.</p>
          <p>Recent advancements in mobile phone technologies, including high-performance processors and high-quality digital cameras, have created the opportunity to invent a food image analysis system on smartphones. While the exact origins of applying AI for IADA research are uncertain, one well-documented attempt to develop a simple system on smartphones was that of DiaWear [<xref ref-type="bibr" rid="ref18">18</xref>]. The system implemented an artificial neural network, which is a subset of deep learning, a recently advanced technique in the field of AI. Despite achieving an accuracy rate above 75%, which was considered incredible at that time, the system’s usefulness was limited because it could identify only 4 types of foods—hamburgers, fries, chicken nuggets, and apple pie. In addition, the system could not determine the portion size of the taken food image; thus, it gave a nutritional value based on a constant portion size directly.</p>
          <p>In this paper, the architecture of IADA is divided into multistage architectures, which were prevalent in the early stages of IADA development, and end-to-end architecture, which has emerged more recently with advancements in AI techniques and food image datasets. The multistage architectures, as implied by their name, include 4 individual processes: segmentation, food identification, portion estimation, and nutrient calculations using a food composition table. This sequential process is consistent across all early-stage IADA systems [<xref ref-type="bibr" rid="ref19">19</xref>-<xref ref-type="bibr" rid="ref23">23</xref>]. These subprocesses are trained independently because they require specific input variables, and optimization can only be done for each step individually, not for the entire process. By contrast, the end-to-end approach, which replaces a multistep pipeline with a single model, can be fine-tuned as a whole process, making it more advanced and increasingly the focus of researchers today.</p>
          <p>Nowadays, multistage architectures are becoming obsolete and are often referred to as traditional IADA. They played a significant role in the IADA timeline before the emergence of the end-to-end approach. Therefore, we delve into the multistage architectures, particularly focusing on food identification and portion estimation algorithms in their subsections, and provide details about the end-to-end approach in the Going Beyond the Traditional Approach With Deep Learning section. For better comparison, <xref rid="figure2" ref-type="fig">Figure 2</xref> illustrates traditional dietary assessment methods and the substitution processes of IADA, along with some notable systems that indicate combining certain processes of the multistage architecture into a single model through deep learning [<xref ref-type="bibr" rid="ref18">18</xref>,<xref ref-type="bibr" rid="ref23">23</xref>-<xref ref-type="bibr" rid="ref31">31</xref>].</p>
          <fig id="figure2" position="float">
            <label>Figure 2</label>
            <caption>
              <p>Comparison of traditional dietary assessment processes and the image-assisted dietary assessment (IADA) substitution processes for the same tasks, including systems that integrate multistage architecture into a single model using deep learning. Systems referenced include DiaWear from Shroff et al [<xref ref-type="bibr" rid="ref18">18</xref>], GoCARB from Anthimopoulos et al [<xref ref-type="bibr" rid="ref23">23</xref>], FIVR from Puri et al [<xref ref-type="bibr" rid="ref26">26</xref>], Im2Calories from Myers et al [<xref ref-type="bibr" rid="ref27">27</xref>], Diabetes60 from Christ et al [<xref ref-type="bibr" rid="ref28">28</xref>], Multitask CNN from Ege and Yanai [<xref ref-type="bibr" rid="ref29">29</xref>], Fang et al [<xref ref-type="bibr" rid="ref30">30</xref>], and technologies-assisted dietary assessment (TADA) from Zhu et al [<xref ref-type="bibr" rid="ref24">24</xref>, <xref ref-type="bibr" rid="ref25">25</xref>,<xref ref-type="bibr" rid="ref31">31</xref>]. 24HR: 24-hour dietary recall; CNN: convolutional neural network; EFR: estimated food record; GAN: generative adversarial network; ResNet50: residual network; SVM: support vector machine; VCG: visual geometry group; WFR: weighed food record.</p>
            </caption>
            <graphic xlink:href="jmir_v26i1e51432_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Food Identification System</title>
          <p>Image recognition systems are one of the milestones in the computer vision field. The goal is to detect and locate an interesting object in an image. Several researchers have applied this technique to food identification tasks that formerly relied on humans only. The early stages in the development of food identification systems were from 2009 to 2015. Most of the existing systems were powered by machine learning algorithms that required human-designed input information, or technical terms called features. Hence, all machine learning-based algorithms are classified as handcrafted algorithms.</p>
          <p>The era of handcrafted algorithms began in 2009 with the release of the Pittsburgh Fast-Food Image Dataset [<xref ref-type="bibr" rid="ref19">19</xref>], marking a significant historical landmark in promoting research into food identification algorithms. This dataset consisted of 4545 fast-food images, including 606 stereo image pairs of 101 different food items. In addition, researchers provided baseline detection accuracy results of 11% and 24% using only the image color histogram together with the support vector machines (SVMs)-based classifier and the bag-of-scale-invariant feature transform classifier, respectively. Although these classifiers were commonly used during that time, the results were not considered sufficient and demonstrated much room for improvement. Since then, various techniques have been proposed to improve the accuracy of food classification from images. In later studies, the same team used pairwise statistics to detect ingredient relations in food images, achieving an accuracy range of 19% to 28% on the Pittsburgh Fast-Food Image Dataset [<xref ref-type="bibr" rid="ref20">20</xref>]. Taichi and Keiji [<xref ref-type="bibr" rid="ref21">21</xref>], from the University of Electro-Communications (UEC) team, used multiple kernel learning, which integrates different image features such as color, texture, and scale-invariant feature transform. This method achieved 61% accuracy on a new dataset of 50 food images and 37.5% accuracy on real-world images captured using a mobile phone [<xref ref-type="bibr" rid="ref21">21</xref>]. In 2011, Bosch et al [<xref ref-type="bibr" rid="ref22">22</xref>] from the Technology Assisted Dietary Assessment (TADA) team achieved an accuracy of 86.1% for 39 food classes by using an SVM classifier. This approach incorporated 6 features derived from color and texture [<xref ref-type="bibr" rid="ref22">22</xref>]. These results suggest that including a larger number of features in the algorithms could potentially improve detection accuracy.</p>
          <p>After active research, the accuracy of handcrafted algorithms reached a saturation point for improvement during the 2014 period. The optimized bag-of-features model was applied to food image recognition by Anthimopoulos et al [<xref ref-type="bibr" rid="ref23">23</xref>]. It achieved an accuracy level of up to 77.8% for 11 classes of food on a food image dataset containing nearly 5000 images for the type 1 diabetes project called GoCARB. Pouladzadeh et al [<xref ref-type="bibr" rid="ref32">32</xref>] achieved a 90.41% accuracy for 15 food classes using an SVM classifier with 4 image features: color, texture, size, and shape. Kawano and Yanai [<xref ref-type="bibr" rid="ref33">33</xref>] (UEC) attained a 50.1% accuracy for a new dataset comprising 256 food classes, using a one-vs-rest classifier with a Fisher vector and a derived feature from a color histogram named RootHoG [<xref ref-type="bibr" rid="ref33">33</xref>]. While handcrafted algorithms yielded high-accuracy results for their specific test datasets with fewer food classes, they struggled to effectively handle larger class sets and real-world images. This difficulty arose due to factors, such as challenging lighting conditions, image noise, distorted food shapes, variations in food colors, and the presence of multiple items within the same image. Handcrafted algorithms may reach a limitation in their ability to improve further.</p>
          <p>In contrast, the novel approach called deep learning, which can automatically extract features from input data, appears to be more suitable for complex tasks such as food identification. The convolutional neural network (CNN), considered to be one of the approaches in deep learning, was developed for handling image analysis in 1998 [<xref ref-type="bibr" rid="ref34">34</xref>]. CNN reads a group of squared pixels of an input image, referred to as a receptive field, and then applies a mathematical function to the read data. The operation is performed repeatedly from the top-left corner until reaching the bottom-right corner of an input image. This operation is done in a similar manner to matrix multiplication or dot product in linear algebra. CNN and deep learning were applied to the food identification task in 2014 by the UEC team [<xref ref-type="bibr" rid="ref35">35</xref>]. This system achieved an accuracy of 72.3% on a dataset containing 100 classes of real-world Japanese food images, named UEC FOOD-100, surpassing their previous handcrafted system in 2012, which achieved 55.8% on the same dataset [<xref ref-type="bibr" rid="ref36">36</xref>]. This marked the beginning of the era of applying deep learning techniques for food identification. Later that year, the UEC team also released an international food image dataset called UEC FOOD-256 that contained 256 food classes to facilitate further research [<xref ref-type="bibr" rid="ref37">37</xref>]. Simultaneously, the FOOD-101 dataset was made available, comprising nearly 101,000 images of 101 different food items [<xref ref-type="bibr" rid="ref38">38</xref>]. They also presented baseline classification results from the random forest–based algorithm, one of the handcrafted algorithms, and compared it with CNN. They found that CNN achieved an accuracy of 56.4%, while random forest–based algorithm achieved 50.76% accuracy in this dataset. These food image datasets have become the favored benchmark for subsequent food identification systems.</p>
          <p>Another important technique is transfer learning, which is well-known for training many deep learning algorithms, including CNNs. It involves 2 stages: pretraining and fine-tuning. Initially, the model is trained with a large and diverse image dataset, and then it is further trained with a smaller, more specific dataset to enhance detection accuracy. This approach is similar to how humans are educated, where broad knowledge is learned in school followed by deeper knowledge in university. The UEC team applied this training approach to the food identification task in 2015 and successfully achieved an accuracy of 78.77% on the UEC FOOD-100 dataset [<xref ref-type="bibr" rid="ref39">39</xref>]. It has been reported that pretraining on large-scale datasets for both food and nonfood images could improve the classification system’s accuracy beyond 80% [<xref ref-type="bibr" rid="ref40">40</xref>-<xref ref-type="bibr" rid="ref45">45</xref>], which is considered to surpass all handcrafted algorithms and be sufficient for real-world applications.</p>
          <p>Currently, numerous state-of-the-art object detectors or classifier models, including the pretrain and fine-tune training paradigm, have been developed and are available, such as AlexNet (AlexNet is an object detection model that won the ImageNet Challenge in 2012; it is named after its inventors, Alex Krizhevsky) [<xref ref-type="bibr" rid="ref46">46</xref>], region-based CNN (R-CNN; an object detection model that significantly improved object detection performance by combining region proposals with CNNs) [<xref ref-type="bibr" rid="ref47">47</xref>], residual network (ResNet; a deep learning model that won the ImageNet Challenge in 2015, known for its innovative use of residual learning to train very deep networks) [<xref ref-type="bibr" rid="ref48">48</xref>], You Only Look Once (YOLO; it is an object detection model that introduced a novel approach by framing object detection as a single regression problem, predicting bounding boxes and class probabilities directly from full images in one step evaluation) [<xref ref-type="bibr" rid="ref49">49</xref>], Visual Geometry Group (VGG) [<xref ref-type="bibr" rid="ref50">50</xref>], and Inception (this is an object detection model that won the ImageNet Challenge in 2014, recognized for its use of a novel architecture that efficiently leverages computing resources inside the network) [<xref ref-type="bibr" rid="ref51">51</xref>]. These object detectors have been designed to automatically extract features from input images and learn distinct characteristics of each class during the training process. Deep learning-based object detection models have shown great promise in image recognition tasks, especially in complex tasks such as food identification. These models and their derivatives are commonly found in many of the food identification systems developed later. The use of these state-of-the-art models presents an exciting opportunity for nutrition researchers who may not have a background in computer engineering or data science. They can now create high-performance food identification systems for specific tasks by curating a food image dataset and training the model accordingly. With the various algorithms available, it is crucial to carefully consider their unique characteristics to select the most suitable one for a given application. The notable food identification systems are listed in <xref ref-type="table" rid="table1">Table 1</xref>.</p>
          <table-wrap position="float" id="table1">
            <label>Table 1</label>
            <caption>
              <p>Overview of notable food identification systems, classifier algorithms, selected features, number of classes, name of food dataset (if specified or noted as their own dataset if absent), and accuracy results<sup>a</sup>.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="160"/>
              <col width="150"/>
              <col width="170"/>
              <col width="150"/>
              <col width="230"/>
              <col width="140"/>
              <thead>
                <tr valign="top">
                  <td>Study, year</td>
                  <td>Projects or team</td>
                  <td>Classifier</td>
                  <td>Feature</td>
                  <td>Class (dataset)</td>
                  <td>Accuracy results percentages</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>Shroff et al [<xref ref-type="bibr" rid="ref18">18</xref>], 2008</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>DiaWear</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Neural network</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Color, size, shape, and texture</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>4</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>~75</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Chen et al [<xref ref-type="bibr" rid="ref19">19</xref>], 2009</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>PFID<sup>b</sup></p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>SVM<sup>c</sup></p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Color</p>
                      </list-item>
                      <list-item>
                        <p>BoSIFT<sup>d</sup></p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>61 (PFID)</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>~11</p>
                      </list-item>
                      <list-item>
                        <p>~24</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Taichi and Keiji [<xref ref-type="bibr" rid="ref21">21</xref>], 2009</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>UEC<sup>e</sup></p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>MKL<sup>f</sup></p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Color, texture, and SIFT<sup>g</sup></p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>50</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>61.34</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Hoashi et al [<xref ref-type="bibr" rid="ref52">52</xref>], 2010</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>UEC</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>MKL</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>BoF<sup>h</sup>, Gabor<sup>i</sup>, color, HOG<sup>j</sup>, and texture</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>85</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>62.53</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Yang et al [<xref ref-type="bibr" rid="ref20">20</xref>], 2010</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>PFID</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>SVM</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Pairwise local features</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>61 (PFID)</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>78.00</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Zhu et al [<xref ref-type="bibr" rid="ref31">31</xref>], 2010</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>TADA<sup>k</sup></p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>SVM with Gaussian radial basis kernel</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Color and texture</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>19</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>97.20</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Kong and Tan [<xref ref-type="bibr" rid="ref53">53</xref>], 2011</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>DietCam</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Multiclass SVM</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Nearest neighbor Gaussian region detector, and SIFT</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>61 (PFID)</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>84.00</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Bosch et al [<xref ref-type="bibr" rid="ref22">22</xref>], 2011</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>TADA</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>SVM</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Color, entropy, Gabor, Tamura<sup>l</sup>, SIFT, Haar wavelet<sup>m</sup>, steerable<sup>n</sup>, and DAISY<sup>o</sup></p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>39</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>86.10</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Matsuda et al [<xref ref-type="bibr" rid="ref36">36</xref>], 2012</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>UEC</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>MKL-SVM</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>HOG, SIFT, Gabor, color, and texture</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>100 (UEC-Food100)</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>55.80</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Anthimopoulos et al [<xref ref-type="bibr" rid="ref23">23</xref>], 2014</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>GoCARB</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>SVM</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>HSV<sup>p</sup>-SIFT, optimized BoF, and color moment invariant</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>11</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>78.00</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>He et al [<xref ref-type="bibr" rid="ref54">54</xref>], 2014</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>TADA</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>k-nearest neighbors</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>DCD<sup>q</sup>, SIFT, MDSIFT<sup>r</sup>, and SCD<sup>s</sup></p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>42</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>65.4</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Pouladzadeh et al [<xref ref-type="bibr" rid="ref32">32</xref>], 2014</td>
                  <td>—<sup>t</sup></td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>SVM</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Color, texture, size, and shape</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>15</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>90.41</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Kawano and Yanai [<xref ref-type="bibr" rid="ref35">35</xref>], 2014</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>UEC</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Pretrained CNN<sup>u</sup></p>
                      </list-item>
                    </list>
                  </td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>100 (UEC-Food100)</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>72.3</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Yanai and Kawano [<xref ref-type="bibr" rid="ref39">39</xref>], 2015</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>UEC</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Deep CNN</p>
                      </list-item>
                    </list>
                  </td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>100 (UEC-Food-100)</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>78.77</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Christodoulidis et al [<xref ref-type="bibr" rid="ref40">40</xref>], 2015</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>GoCARB</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Patch-wise CNN</p>
                      </list-item>
                    </list>
                  </td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>7</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>84.90</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Myers et al [<xref ref-type="bibr" rid="ref27">27</xref>], 2015</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Google</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>GoogLeNet</p>
                      </list-item>
                    </list>
                  </td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>101</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>79.00</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Liu et al [<xref ref-type="bibr" rid="ref41">41</xref>], 2016</td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>DeepFood</p>
                      </list-item>
                    </list>
                  </td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Food-101</p>
                      </list-item>
                      <list-item>
                        <p>UEC-256</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>77.40</p>
                      </list-item>
                      <list-item>
                        <p>54.70</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Singla et al [<xref ref-type="bibr" rid="ref42">42</xref>], 2016</td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>GoogLeNet</p>
                      </list-item>
                    </list>
                  </td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>11</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>83.60</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Hassannejad et al [<xref ref-type="bibr" rid="ref43">43</xref>], 2016</td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>InceptionV3<sup>v</sup></p>
                      </list-item>
                    </list>
                  </td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>101 (Food-101)</p>
                      </list-item>
                      <list-item>
                        <p>100 (UEC-Food100)</p>
                      </list-item>
                      <list-item>
                        <p>256 (UEC-Food256)</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>88.28</p>
                      </list-item>
                      <list-item>
                        <p>81.45</p>
                      </list-item>
                      <list-item>
                        <p>76.17</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Ciocca et al [<xref ref-type="bibr" rid="ref44">44</xref>], 2017</td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>VGG<sup>w</sup></p>
                      </list-item>
                    </list>
                  </td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>73 (UNIMINB2016)</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>78.30</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Mezgec and Koroušić Seljak [<xref ref-type="bibr" rid="ref45">45</xref>], 2017</td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>NutriNet (Modified AlexNet<sup>x</sup>)</p>
                      </list-item>
                    </list>
                  </td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>73 (UNIMINB2016)</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>86.72</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Pandey et al [<xref ref-type="bibr" rid="ref55">55</xref>], 2017</td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Ensemble net</p>
                      </list-item>
                    </list>
                  </td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>101 (Food-101)</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>72.10</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Martinel et al [<xref ref-type="bibr" rid="ref56">56</xref>], 2018</td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>WISeR<sup>y</sup></p>
                      </list-item>
                    </list>
                  </td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>101 (Food-101)</p>
                      </list-item>
                      <list-item>
                        <p>100 (UEC-Food100)</p>
                      </list-item>
                      <list-item>
                        <p>256 (UEC-Food256)</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>88.72</p>
                      </list-item>
                      <list-item>
                        <p>79.76</p>
                      </list-item>
                      <list-item>
                        <p>86.71</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Jiang et al [<xref ref-type="bibr" rid="ref57">57</xref>], 2020</td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>MSMVFA<sup>z</sup></p>
                      </list-item>
                    </list>
                  </td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>101 (Food-101)</p>
                      </list-item>
                      <list-item>
                        <p>172 (VireoFood-172)</p>
                      </list-item>
                      <list-item>
                        <p>208 (ChineseFoodNet)</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>~90.47</p>
                      </list-item>
                      <list-item>
                        <p>90.61</p>
                      </list-item>
                      <list-item>
                        <p>81.94</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Lu et al [<xref ref-type="bibr" rid="ref58">58</xref>], 2020</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>GoCARB</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Modified InceptionV3</p>
                      </list-item>
                    </list>
                  </td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>298 Generic food</p>
                      </list-item>
                      <list-item>
                        <p>Subgroups</p>
                      </list-item>
                      <list-item>
                        <p>Fine-grained</p>
                      </list-item>
                      <list-item>
                        <p>(MADiMA<sup>aa</sup>)</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>65.80</p>
                      </list-item>
                      <list-item>
                        <p>61.50</p>
                      </list-item>
                      <list-item>
                        <p>57.10</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>Wu et al [<xref ref-type="bibr" rid="ref59">59</xref>], 2021</td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Modified AlexNet</p>
                      </list-item>
                    </list>
                  </td>
                  <td>—</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>22 styles of Bento sets</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>96.30</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table1fn1">
                <p><sup>a</sup>Note that convolutional neural network–based classifiers do not require the number of features to be shown as they extract features autonomously.</p>
              </fn>
              <fn id="table1fn2">
                <p><sup>b</sup>PFID: Pittsburgh Fast-Food Image Dataset.</p>
              </fn>
              <fn id="table1fn3">
                <p><sup>c</sup>SVM: support vector machine.</p>
              </fn>
              <fn id="table1fn4">
                <p><sup>d</sup>BoSIFT: bag-of-scale-invariant feature transform.</p>
              </fn>
              <fn id="table1fn5">
                <p><sup>e</sup>UEC: University of Electro-Communications.</p>
              </fn>
              <fn id="table1fn6">
                <p><sup>f</sup>MKL: multiple kernel learning. This is a machine-learning technique that combines multiple kernels or similarity functions, to improve the performance and flexibility of kernel-based models such as support vector machines.</p>
              </fn>
              <fn id="table1fn7">
                <p><sup>g</sup>SIFT: scale-invariant feature transform.</p>
              </fn>
              <fn id="table1fn8">
                <p><sup>h</sup>BoF: bag-of-features.</p>
              </fn>
              <fn id="table1fn9">
                <p><sup>i</sup>Gabor is a texture feature extraction invented by Dennis Gabor.</p>
              </fn>
              <fn id="table1fn10">
                <p><sup>j</sup>HOG: histogram of orientated gradients—a feature descriptor based on color.</p>
              </fn>
              <fn id="table1fn11">
                <p><sup>k</sup>TADA: Technology Assisted Dietary Assessment.</p>
              </fn>
              <fn id="table1fn12">
                <p><sup>l</sup>Tamura is a 6-texture feature extraction invented by Hideyuki Tamura.</p>
              </fn>
              <fn id="table1fn13">
                <p><sup>m</sup>Haar wavelet is a mathematical analysis for wavelet sequence named after Alfréd Haar.</p>
              </fn>
              <fn id="table1fn14">
                <p><sup>n</sup>Steerable filter is an image filter introduced by Freeman and Adelson.</p>
              </fn>
              <fn id="table1fn15">
                <p><sup>o</sup>DAISY is a local image descriptor introduced by E Tola et al [<xref ref-type="bibr" rid="ref60">60</xref>], but they did not describe a true acronym of DAISY.</p>
              </fn>
              <fn id="table1fn16">
                <p><sup>p</sup>HSV is the name of a red-green-blue color model based on hue, saturation, and value.</p>
              </fn>
              <fn id="table1fn17">
                <p><sup>q</sup>DCD: dominant color descriptor.</p>
              </fn>
              <fn id="table1fn18">
                <p><sup>r</sup>MDSIFT: multiscale dense scale-invariant feature transform.</p>
              </fn>
              <fn id="table1fn19">
                <p><sup>s</sup>SCD: scalable color descriptor.</p>
              </fn>
              <fn id="table1fn20">
                <p><sup>t</sup>Not available.</p>
              </fn>
              <fn id="table1fn21">
                <p><sup>u</sup>CNN: convolutional neural network.</p>
              </fn>
              <fn id="table1fn22">
                <p><sup>v</sup>Inception is an object detection model that won the ImageNet Challenge in 2014, recognized for its use of a novel architecture that efficiently leverages computing resources inside the network.</p>
              </fn>
              <fn id="table1fn23">
                <p><sup>w</sup>VGG: visual geometry group—an object detection model named after a research group from the University of Oxford.</p>
              </fn>
              <fn id="table1fn24">
                <p><sup>x</sup>AlexNet is an object detection model that won the ImageNet Large-Scale Visual Recognition Challenge (also known as the ImageNet challenge) in 2012; it is named after its inventors, Alex Krizhevsky.</p>
              </fn>
              <fn id="table1fn25">
                <p><sup>y</sup>WISeR: wide-slice residual.</p>
              </fn>
              <fn id="table1fn26">
                <p><sup>z</sup>MSMVFA: multi-scale multi-view feature aggregation.</p>
              </fn>
              <fn id="table1fn27">
                <p><sup>aa</sup>MADiMA: Multimedia Assisted Dietary Management.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
        </sec>
        <sec>
          <title>Food Portion Size Estimation System</title>
          <sec>
            <title>Overview</title>
            <p>Food portion size estimation is a challenging task for researchers as it requires more accurate information on the amount of food, ingredients, or cooking methods that cannot be obtained from only a captured image without additional input, which makes it harder to create a food image dataset with portion size annotation. Furthermore, quantifying an object’s size from a single 2D image is faced with common image perspective distortion problems [<xref ref-type="bibr" rid="ref61">61</xref>,<xref ref-type="bibr" rid="ref62">62</xref>], as shown in <xref rid="figure3" ref-type="fig">Figure 3</xref>. First, the size of the object in the image can change due to the distance between the object (food) and the capturing device (smartphone or camera). The size of the white rice in <xref rid="figure3" ref-type="fig">Figure 3</xref>A is smaller compared with <xref rid="figure3" ref-type="fig">Figure 3</xref>B because the white rice in <xref rid="figure3" ref-type="fig">Figure 3</xref>B is closer to the camera. Second, the angle at which the photo is taken also alters the perceived object size. For example, flattened objects such as rice, that are spread out on a 23-cm (9-inch) circular plate appear in their full size in a bird’s-eye shot (90°), in <xref rid="figure3" ref-type="fig">Figure 3</xref>C, but they appear smaller when taken from approximately 30° from the tabletop as in <xref rid="figure3" ref-type="fig">Figure 3</xref>D. Thirdly, there is a loss of depth in a bird’s-eye view in <xref rid="figure3" ref-type="fig">Figures 3</xref>E and 3F, making it difficult to compare between food B and food C. The weights of foods A, B, C, and D are 48, 49, 62, and 149 grams, respectively. We use these images for teaching image-based portion estimation for dietetics students.</p>
            <p>While pretrain and fine-tune training for CNNs is a silver bullet for food image identification, currently there is no equivalent solution for portion estimation. Many researchers are actively finding ways to calibrate the object size within an image to mediate such an error, and several approaches have been discussed here. Basically, portion estimation can be broadly classified, based on complexity, into four progressive categories: (1) pixel density, (2) geometric modeling, (3) 3D reconstruction, and (4) depth camera. <xref ref-type="table" rid="table2">Table 2</xref> provides an overview of notable systems for volume estimation.</p>
            <fig id="figure3" position="float">
              <label>Figure 3</label>
              <caption>
                <p>There are common image perspective distortion problems. Firstly, position distortion: the size of the white rice in (A) is smaller compared to (B) because the white rice in (B) is closer to the camera. Secondly, angle distortion: the white rice in (C) is fully visible at 90 degrees, while it appears smaller when taken from 30 degrees, as in (D). Thirdly, there is a loss of depth information in the bird’s-eye view in (E) and (F), making it difficult to compare food B and food C.</p>
              </caption>
              <graphic xlink:href="jmir_v26i1e51432_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
            </fig>
            <table-wrap position="float" id="table2">
              <label>Table 2</label>
              <caption>
                <p>A comprehensive overview of notable publications for 4 volume estimation approaches, arranged chronologically.</p>
              </caption>
              <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
                <col width="30"/>
                <col width="180"/>
                <col width="170"/>
                <col width="210"/>
                <col width="160"/>
                <col width="250"/>
                <thead>
                  <tr valign="top">
                    <td colspan="2">Approach and study, year</td>
                    <td>Projects or team</td>
                    <td>Reference object</td>
                    <td>Item</td>
                    <td>Reported error</td>
                  </tr>
                </thead>
                <tbody>
                  <tr valign="top">
                    <td colspan="6">
                      <bold>Pixel density approach</bold>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Martin et al [<xref ref-type="bibr" rid="ref13">13</xref>], 2009</td>
                    <td>—<sup>a</sup></td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Physical card</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>N/A<sup>b</sup></p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>N/A</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Jia et al [<xref ref-type="bibr" rid="ref63">63</xref>], 2012</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>University of Pittsburgh</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Circular plate</p>
                        </list-item>
                        <list-item>
                          <p>Circular LED light</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>—</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>&#60;27.60</p>
                        </list-item>
                        <list-item>
                          <p>&#60;54.10</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Pouladzadeh et al [<xref ref-type="bibr" rid="ref32">32</xref>], 2014</td>
                    <td>—</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>User’s thumb</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>5</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>&#60;10</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Okamoto and Yanai [<xref ref-type="bibr" rid="ref64">64</xref>], 2016</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>UEC<sup>c</sup></p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Wallet</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>3</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Mean calorie error</p>
                          <list>
                            <list-item>
                              <p>
                    Beef rice bowl –242 (SD 55.1)
                  </p>
                            </list-item>
                            <list-item>
                              <p>
                    Croquette –47.08 (SD 52.5)
                  </p>
                            </list-item>
                            <list-item>
                              <p>
                    Salad 4.86 (SD 11.9)
                  </p>
                            </list-item>
                          </list>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Akpa et al [<xref ref-type="bibr" rid="ref65">65</xref>], 2017</td>
                    <td>—</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Chopstick</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>15</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>&#60;6.65</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Liang and Li [<xref ref-type="bibr" rid="ref66">66</xref>], 2017</td>
                    <td>—</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>1-yuan coin</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>19 fruits</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>15 items &#60;20%</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Yanai et al [<xref ref-type="bibr" rid="ref67">67</xref>], 2019 and Ege et al [<xref ref-type="bibr" rid="ref67">67</xref>], 2019</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>UEC</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Rice grain size</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>3</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>&#60;10%</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td colspan="6">
                      <bold>Geometric modeling approach</bold>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Zhu et al [<xref ref-type="bibr" rid="ref24">24</xref>], 2010 and Zhu et al [<xref ref-type="bibr" rid="ref25">25</xref>], 2008</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>TADA<sup>d</sup></p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Checkerboard</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>7</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Spherical 5.65%</p>
                        </list-item>
                        <list-item>
                          <p>Prismatic 28.85%</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Chae et al [<xref ref-type="bibr" rid="ref69">69</xref>], 2011</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>TADA</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Checkerboard</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>26</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Cylinders 11.1%</p>
                        </list-item>
                        <list-item>
                          <p>Flattop solid 11.7%</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Chen et al [<xref ref-type="bibr" rid="ref70">70</xref>], 2013</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>University of Pittsburgh</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Circular plate</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>17</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>3.69%</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Jia et al [<xref ref-type="bibr" rid="ref71">71</xref>], 2014</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>University of Pittsburgh</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Circular plate</p>
                        </list-item>
                        <list-item>
                          <p>Other container</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>100</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>&#60;30% from 85/100 of test items</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Tanno et al [<xref ref-type="bibr" rid="ref72">72</xref>], 2018</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>UEC</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Apple ARKit</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>3</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Mean calorie error</p>
                        </list-item>
                      </list>
                      <list list-type="bullet">
                        <list-item>
                          <p>Beef rice bowl –67.14 (SD 18.8)</p>
                        </list-item>
                        <list-item>
                          <p>Croquette–127.0 (SD 9.0)</p>
                        </list-item>
                        <list-item>
                          <p>Salad –0.95 (SD 0.16)</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Yang et al [<xref ref-type="bibr" rid="ref73">73</xref>], 2019</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>University of Pittsburgh</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Augmented reality</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>15</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Large objects 16.65%</p>
                        </list-item>
                        <list-item>
                          <p>Small objects 47.60%</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Smith et al [<xref ref-type="bibr" rid="ref74">74</xref>], 2022</td>
                    <td>—</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Checkerboard</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>26</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Single food items 32.4%-56.1%</p>
                        </list-item>
                        <list-item>
                          <p>Multiple food items 23.7%-32.6%</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td colspan="6">
                      <bold>3D reconstruction approach</bold>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Puri et al [<xref ref-type="bibr" rid="ref26">26</xref>], 2009</td>
                    <td>—</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>3 images</p>
                        </list-item>
                        <list-item>
                          <p>Checkerboard</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>26</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>2%-9.5%</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Kong and tan [<xref ref-type="bibr" rid="ref75">75</xref>], 2012</td>
                    <td>—</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>3 images</p>
                        </list-item>
                        <list-item>
                          <p>Checkerboard</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>7</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Volume estimation error 20%</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Rahman et al [<xref ref-type="bibr" rid="ref76">76</xref>], 2012</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>TADA</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>2 images</p>
                        </list-item>
                        <list-item>
                          <p>Checkerboard</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>6</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>7.70%</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Chang et al [<xref ref-type="bibr" rid="ref77">77</xref>], 2013</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>TADA</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Using food silhouettes to reconstruct a 3D object</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>4</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>10%</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Anthimopoulos et al [<xref ref-type="bibr" rid="ref78">78</xref>], 2015</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>GoCARB</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>2 images physical card Physical card</p>
                        </list-item>
                      </list>
                    </td>
                    <td>N/A</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Volume estimation error 9.4%</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Dehais et al [<xref ref-type="bibr" rid="ref79">79</xref>], 2017</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>GoCARB</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>2 images</p>
                        </list-item>
                        <list-item>
                          <p>Physical card</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>45 dishes</p>
                        </list-item>
                        <list-item>
                          <p>14 meals</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>8.2%-9.8%</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Gao et al [<xref ref-type="bibr" rid="ref80">80</xref>], 2018</td>
                    <td>—</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>SLAMe-based with Rubik cube</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>3</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>11.69%-19.20% for static measurement</p>
                        </list-item>
                        <list-item>
                          <p>16.32%-27.9% for continuous measurement</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Ando et al [<xref ref-type="bibr" rid="ref81">81</xref>], 2019</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>UEC</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Multiple cameras on iPhone X for depth estimation</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>3</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Calorie estimation error</p>
                          <list>
                            <list-item>
                              <p>
                    Sweet and sour pork &#60;1%
                  </p>
                            </list-item>
                            <list-item>
                              <p>
                    Fried chicken &#60;1%
                  </p>
                            </list-item>
                            <list-item>
                              <p>
                    Croquette &#60;15%
                  </p>
                            </list-item>
                          </list>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Lu et al [<xref ref-type="bibr" rid="ref58">58</xref>], 2020</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>GoCARB</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>2 images</p>
                        </list-item>
                        <list-item>
                          <p>Physical card and gravity information</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>234 items from MADiMA<sup>f</sup></p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>MARE<sup>g</sup> 19%, while their earlier system, GoCarb (2017), achieved 22.6% on the same task [<xref ref-type="bibr" rid="ref79">79</xref>].</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td colspan="6">
                      <bold>Depth camera approach</bold>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Shang et al [<xref ref-type="bibr" rid="ref82">82</xref>], 2011</td>
                    <td>—</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Specific food recording device</p>
                        </list-item>
                      </list>
                    </td>
                    <td>—</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>No performance report</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Chen et al [<xref ref-type="bibr" rid="ref83">83</xref>], 2012</td>
                    <td>—</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Depth camera</p>
                        </list-item>
                      </list>
                    </td>
                    <td>—</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>No performance report</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Fang et al [<xref ref-type="bibr" rid="ref84">84</xref>], 2016</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>TADA</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Camera from this study [<xref ref-type="bibr" rid="ref85">85</xref>]</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>10</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Depth method overestimates volume than geometric model</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Alfonsi et al [<xref ref-type="bibr" rid="ref86">86</xref>], 2020</td>
                    <td>—</td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>iPhone and Android devices</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>200</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Carbohydrate estimation error &#60;10 g</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                  <tr valign="top">
                    <td>
                      <break/>
                    </td>
                    <td>Herzig et al [<xref ref-type="bibr" rid="ref87">87</xref>], 2020</td>
                    <td>
                      <break/>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>iPhone X</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>128</p>
                        </list-item>
                      </list>
                    </td>
                    <td>
                      <list list-type="bullet">
                        <list-item>
                          <p>Relative error of weight estimation 14.0%</p>
                        </list-item>
                      </list>
                    </td>
                  </tr>
                </tbody>
              </table>
              <table-wrap-foot>
                <fn id="table2fn1">
                  <p><sup>a</sup>Not available.</p>
                </fn>
                <fn id="table2fn2">
                  <p><sup>b</sup>N/A: not applicable.</p>
                </fn>
                <fn id="table2fn3">
                  <p><sup>c</sup>UEC: University of Electro-Communications.</p>
                </fn>
                <fn id="table2fn4">
                  <p><sup>d</sup>TADA: Technology Assisted Dietary Assessment.</p>
                </fn>
                <fn id="table2fn5">
                  <p><sup>e</sup>SLAM: simultaneous localization and mapping.</p>
                </fn>
                <fn id="table2fn6">
                  <p><sup>f</sup>MADiMA: Multimedia Assisted Dietary Management.</p>
                </fn>
                <fn id="table2fn7">
                  <p><sup>g</sup>MARE: mean absolute relative error.</p>
                </fn>
              </table-wrap-foot>
            </table-wrap>
          </sec>
          <sec>
            <title>Revisiting the Classic Pixel Density Approach</title>
            <p>Pixel density is the simplest approach for providing good and effective estimation. After a food image is segmented, the number of pixels in each segmented section is determined. Mathematical equations or other transformations are then used to calculate the portion size of each section that is presented in the image.</p>
            <p>However, this approach suffers from image distortion problems, and several approaches have been implemented to combat this drawback. The simplest method is the use of a physical reference object or fiducial marker for calibrating the size of objects in an image. When the real size of the reference object is known, the real size of an object can be determined relative to the reference object. This method was chosen for food volume estimation during its early development stage [<xref ref-type="bibr" rid="ref13">13</xref>,<xref ref-type="bibr" rid="ref88">88</xref>,<xref ref-type="bibr" rid="ref89">89</xref>]. Various physical objects have been used as reference objects in the literature, including a special patterned card [<xref ref-type="bibr" rid="ref13">13</xref>,<xref ref-type="bibr" rid="ref89">89</xref>], a known-size circular plate [<xref ref-type="bibr" rid="ref63">63</xref>] or bowl [<xref ref-type="bibr" rid="ref90">90</xref>], chopsticks [<xref ref-type="bibr" rid="ref65">65</xref>], a 1-yuan coin [<xref ref-type="bibr" rid="ref66">66</xref>], a wallet [<xref ref-type="bibr" rid="ref64">64</xref>], a user’s thumb [<xref ref-type="bibr" rid="ref40">40</xref>,<xref ref-type="bibr" rid="ref91">91</xref>], or even rice grain size [<xref ref-type="bibr" rid="ref67">67</xref>].</p>
          </sec>
          <sec>
            <title>Geometric Modeling Approach</title>
            <p>Assuming that the food has a cylindrical shape, such as compressed steamed rice (<xref rid="figure4" ref-type="fig">Figure 4</xref>A), its volume can be calculated using the conventional formula 2πr<sup>2</sup> × h. The radius r and height h can be determined by counting the pixels in the image. While this approach is effective for geometric shapes, it is less reliable for irregular shapes that lack a specific equation. The demonstration of this approach is shown in <xref rid="figure4" ref-type="fig">Figure 4</xref>B, where the user selects a predefined shape and then manually fits (or registers) the geometric model with the image.</p>
            <p>The TADA team reported the use of several predefined shapes of foods, including cylindrical, flattop solid, spherical, and prismatic models [<xref ref-type="bibr" rid="ref24">24</xref>,<xref ref-type="bibr" rid="ref25">25</xref>,<xref ref-type="bibr" rid="ref68">68</xref>,<xref ref-type="bibr" rid="ref69">69</xref>]. Prismatic models were specifically used to estimate portion sizes of irregularly shaped foods. This approach allowed a more accurate estimation of portion sizes by considering the unique characteristics of each food item. The research team at the University of Pittsburgh proposed a similar technique known as wireframe modeling. This technique involves creating a skeletal representation of an object using lines and curves to define its structure to accurately capture the shape and dimensions of food items [<xref ref-type="bibr" rid="ref70">70</xref>,<xref ref-type="bibr" rid="ref71">71</xref>]. However, this approach is also affected by common image distortion problems. Initially, a physical reference object was used for calibration.</p>
            <p>Geometric modeling shares a fundamental principle with augmented reality (AR), a technology that transforms 2D environmental images into 3D coordinates in a computer system. As AR has become more widely available on smartphones, many researchers have explored the feasibility of using AR as a calibration method instead of using physical reference objects [<xref ref-type="bibr" rid="ref72">72</xref>,<xref ref-type="bibr" rid="ref73">73</xref>]. AR-based object length measurement is demonstrated in <xref rid="figure5" ref-type="fig">Figure 5</xref>.</p>
            <fig id="figure4" position="float">
              <label>Figure 4</label>
              <caption>
                <p>This figure demonstrates the various approaches to estimating food volume. (A) A cylindrical shape of 75 grams of brown rice taken from a 60° angle. (B) Geometric modeling with a predefined cylindrical shape, where the user needs to adjust each point manually to fit the object. (C) A predicted depth map from state-of-the-art dense prediction transformation. (D) A 3D reconstructed object using depth information from (C). These images have been adjusted in size for visual comparison purposes.</p>
              </caption>
              <graphic xlink:href="jmir_v26i1e51432_fig4.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
            </fig>
            <fig id="figure5" position="float">
              <label>Figure 5</label>
              <caption>
                <p>Measuring the size of the same banana can be done using different techniques, as shown in the figure. (A) A standard ruler is used as a ground truth measurement, (B) Samsung augmented reality Zone app, and (C) Apple iPhone Measure app. These apps use the gyroscope or accelerometer sensors in the mobile phone to accurately track the movement of the phone as the measurement line is drawn.</p>
              </caption>
              <graphic xlink:href="jmir_v26i1e51432_fig5.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
            </fig>
          </sec>
          <sec>
            <title>3D Reconstruction</title>
            <p>This technique involves using ≥2 images taken from different angles to create virtual 3D objects in 3D coordinates in a computer system. It shares the same principle as both AR and geometric modeling, where reconstructed objects are represented similarly to prismatic models in geometric modeling. Furthermore, this technique allows for the inclusion of shapes beyond traditional geometric shapes.</p>
            <p>While several researchers have explored the use of 3D reconstruction [<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref75">75</xref>,<xref ref-type="bibr" rid="ref76">76</xref>], 1 notable example is the GoCARB system [<xref ref-type="bibr" rid="ref78">78</xref>]. This system requires 2 images taken from different angles to construct a 3D model of the food, achieving an accuracy within 20 grams for carbohydrate content estimation. This level of accuracy is comparable to estimates made by dietitians when the food is completely visible on a single dish with an elliptical plate and flat base [<xref ref-type="bibr" rid="ref92">92</xref>].</p>
            <p><xref rid="figure4" ref-type="fig">Figures 4</xref>C and 4D demonstrate a similar 3D reconstruction approach but implemented using state-of-the-art dense prediction transformation models to predict depth maps from a single image (<xref rid="figure4" ref-type="fig">Figure 4</xref>A), followed by the reconstruction of the 3D object using the predicted depth map.</p>
          </sec>
          <sec>
            <title>Depth Camera Approach</title>
            <p>This method operates on the same principle as geometric modeling and 3D reconstruction, but it requires a special time-of-flight (ToF) sensor (also known as a depth camera) to measure an object’s size in 3D coordinates in a computer system. Initially, the application of depth cameras in food volume estimation was limited, primarily due to their high cost and limited availability [<xref ref-type="bibr" rid="ref82">82</xref>]. However, with the introduction of consumer-grade depth cameras, such as Kinect (Microsoft Corp), Intel RealSense, and smartphones equipped with depth sensors, their accessibility increased, leading to wider use in food volume estimation applications [<xref ref-type="bibr" rid="ref81">81</xref>,<xref ref-type="bibr" rid="ref83">83</xref>,<xref ref-type="bibr" rid="ref84">84</xref>,<xref ref-type="bibr" rid="ref86">86</xref>,<xref ref-type="bibr" rid="ref87">87</xref>].</p>
            <p>Nevertheless, the availability of depth sensors remains a significant challenge in implementing this system. Currently, only a limited number of mobile phone models are equipped with such sensors. In addition, some manufacturers integrate the sensor with the front camera for authentication purposes, such as Apple’s FaceID, making it impractical for capturing object photos. Moreover, certain mobile device manufacturers have omitted the ToF sensor in their recent models [<xref ref-type="bibr" rid="ref93">93</xref>], further reducing the availability of depth sensors and posing implementation challenges for the depth camera approach.</p>
            <p>An example of depth information captured by the Intel Realsense d435i depth camera displayed in RGB (red-green-blue; color model based on additive color primaries) with depth (RGB with depth; RGBD) format is shown in <xref rid="figure6" ref-type="fig">Figure 6</xref>B. Rendered objects from a captured polygon file are demonstrated as freely rotatable 3D objects in <xref rid="figure6" ref-type="fig">Figures 6</xref>C and 6D, with a regular RGB image shown for comparison in <xref rid="figure6" ref-type="fig">Figure 6</xref>A.</p>
            <fig id="figure6" position="float">
              <label>Figure 6</label>
              <caption>
                <p>(A) A typical red-green-blue image showing 3 Burmese grapes, each weighing approximately 20 grams. (B) A red-green-blue image with depth captured by Intel RealSense d435i from a bird’s-eye view. (C) and (D) 3D reconstructed objects from the polygon file, illustrating the height of each fruit from different angles.</p>
              </caption>
              <graphic xlink:href="jmir_v26i1e51432_fig6.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
            </fig>
          </sec>
        </sec>
        <sec>
          <title>Going Beyond the Traditional Approach With Deep Learning</title>
          <p>Advancements in deep learning are opening more possibilities to improve the IADA system by merging some steps (or even all steps) of the multistep pipeline into a single model, which can be fine-tuned as a whole process. Due to the rise in IADA research with the emergence of advanced algorithms, we can only highlight a few reports that demonstrate the gradual enhancements in IADA in this paper.</p>
          <p>In 2015, Myers et al [<xref ref-type="bibr" rid="ref27">27</xref>] from Google proposed the Im2Calories system, using deep learning for all stages of IADA. The classifiers are based on the GoogLeNet architecture, and the classification results are used to improve the semantic segmentation handled by the DeepLab network. For volume estimation, a new CNN architecture, trained with an RGBD dataset, estimates the depth map from a single RBG image and then converts the depth map to volume in the final step. Although the absolute error for some test foods could exceed 300 ml, the overall volume estimation results were deemed acceptable. The system still requires a food composition database to determine the nutritional values of the food in the final step.</p>
          <p>The idea of using deep learning to estimate food volume is gaining popularity, and several systems are transitioning to using deep learning algorithms to estimate food volume without the need for an actual ToF sensor. In 2017, carbohydrate counting algorithms named Diabetes60 were proposed by Christ et al [<xref ref-type="bibr" rid="ref28">28</xref>]. The system reported food-specific portions called “bread units,” which are defined to contain 12 to 15 grams of carbohydrates. This definition closely resembles the “carb unit” widely used in the diabetes field or the “exchange unit” in dietetic practice. The system was based on ResNet50 and trained using an RGBD image dataset that contained human-annotated bread unit information. It achieved a root mean square error of 1.53 (approximately 18.4-23 g of carbohydrate), while humans could achieve a root mean square error of 0.89 (approximately 10.7-13.4 g of carbohydrate) when compared with the ground truth. The modified ResNet was also used for fruit volume estimation, achieving an error of 2.04% to 14.3% for 5 types of fruit and 1 fruit model [<xref ref-type="bibr" rid="ref94">94</xref>]. Furthermore, Jiang et al [<xref ref-type="bibr" rid="ref95">95</xref>] introduced a system to classify liquid levels in bottles into 4 categories: 25%, 50%, 75%, and 100%. Using their own designed CNN architecture, they achieved a 92.4% classification accuracy when the system was trained with 3 methods of data augmentation. Furthermore, the system could achieve 100% classification accuracy when the bottle images had labels removed.</p>
          <p>One challenge in converting a single 2D image into a 3D object is the difficulty in capturing the back side of an object in single-view images due to factors such as view angle or occlusion. Therefore, the food volume may be underestimated. Point2Volume was introduced in 2020 by Lo et al [<xref ref-type="bibr" rid="ref96">96</xref>] to address the limitations. The system builds upon 2 of their previous works: a deep learning view synthesis [<xref ref-type="bibr" rid="ref97">97</xref>] and a point completion network [<xref ref-type="bibr" rid="ref98">98</xref>]. When a single-depth image is captured, a Mark region-based CNN—a combination of object detection and instance segmentation network—performs classification and segmentation, obtaining only partial point clouds due to occlusion. It then reconstructs the complete shapes and finally estimates the food volumes. This system demonstrated a volume estimation error of 7.7% for synthetic foods and 15.3% for real foods.</p>
          <p>While the estimation of exact food volume has improved recently, dietitians and nutritionists often use a different approach. They compare unknown food amounts with known reference volumes, such as a thumb, matchbox, tennis ball, deck of cards, or a series of known portion-size images. Yang et al [<xref ref-type="bibr" rid="ref99">99</xref>] introduced a system that mimics this mental estimation approach in 2021. The system classifies the unknown portion object to match the system’s set of reference volumes and then fine-tunes the predicted volume using the selected set. The system achieved a mean relative volumetric error of around 11.6% to 20.1% for their own real food image dataset. Interestingly, they noted that even when the system chose the wrong set of reference volumes—due to top-1 accuracy being &#60;50% in most cases—the mean relative volumetric error still remained acceptable, implying that fewer reference volume sets might be sufficient.</p>
          <p>Another crucial question is how many food classes should be included in the system to achieve usability in day-to-day situations. The goFood system [<xref ref-type="bibr" rid="ref58">58</xref>], successor to the previous carbohydrate estimation system GoCARB, takes a different approach to expand the coverage beyond their included food classes. Using a modified Inception V3 architecture to classify food into a 3-level hierarchical structure: 18 types of generic food (eg, meat, bread, and dairy), 40 types of subgroups (eg, white bread and red meat), and 319 types of specific foods. This strategy mirrors the concept of a food exchange list, allowing the handling of a large number of foods without the need for an extensive number of fine-grained classifications. This lowers the number of unidentified food objects and results in achieving at least a 3% higher accuracy for food identification than the single-level Inception V3 classifier. Their newer 3D reconstruction algorithm, incorporating gravity data from the smartphone’s inertial measurement unit (eg, accelerometer or gyroscope), achieved a mean absolute relative error of 19%, surpassing the algorithm in GoCARB, which had 22.6% error.</p>
          <p>Furthermore, CNN and deep learning could potentially estimate nutrients directly without relying on food composition tables, enabling an end-to-end approach for IADA. The originality of this method is unclear, but to the best of our knowledge, the first well-documented system was introduced by Miyazaki et al [<xref ref-type="bibr" rid="ref100">100</xref>] in 2011. This system extracts 4 features from food images and estimates calories from these features instead of relying on food identification, portion estimation, and food composition tables as in multistage IADA. The system achieved a relative error of approximately 20% for 35% of items and 40% for 79% of items, which is relatively high. This idea inspired subsequent works by Ege and Yanai [<xref ref-type="bibr" rid="ref29">29</xref>] from UEC in 2017. They applied a multitask CNN, a technique where a model is trained to perform multiple tasks simultaneously, using visual geometry group-16 for feature extraction and a calorie-annotated image dataset for training. The CNN system achieved an estimation error of 20% for 50% of items and 40% for 80% of items in their Japanese food image dataset. However, the system assumed that each food image contained only 1 food item; this limitation was addressed in their later works [<xref ref-type="bibr" rid="ref101">101</xref>,<xref ref-type="bibr" rid="ref102">102</xref>]. Multitask CNNs can be fine-tuned for the entire algorithm rather than for each stage as in a multistage architecture. This gives them the potential to surpass multistage architectures, similar to how deep learning and CNNs have outperformed handcrafted food identification algorithms. Therefore, they have gained significant attention from researchers [<xref ref-type="bibr" rid="ref103">103</xref>-<xref ref-type="bibr" rid="ref107">107</xref>].</p>
          <p>Not only multitask CNNs but also generative adversarial networks, which are the backbone of image generation AI, such as Dall-E (OpenAI), can be used to learn the energy distribution map and estimate food energy directly from a single RGB image. Fang et al [<xref ref-type="bibr" rid="ref30">30</xref>] from the TADA team applied this approach and achieved a mean energy estimation error of 209 kcal. Their subsequent work, which included adding food localization networks, improved accuracy by approximately 3.6% [<xref ref-type="bibr" rid="ref108">108</xref>]. While most system predictions focus on food portions (volume or weight), calories, or macronutrients such as carbohydrates, in 2019, Situju et al [<xref ref-type="bibr" rid="ref109">109</xref>] used a multitask CNN to predict the salt content of 14 types of food. This was achieved by training the multitask CNN with a dataset annotated for both calories and salt. The relative estimation error was 31.2% (89.6 kcal) for calories and 36.1% (0.74 g) for salt. These works provide evidence that advanced deep learning techniques yield promising results and offer room for improvement in IADA, garnering increasing attention from researchers today.</p>
        </sec>
      </sec>
      <sec>
        <title>Advancements and Challenges From the Dietitian’s Perspective</title>
        <sec>
          <title>Overview</title>
          <p>According to recently published information, both image classification and volume estimation techniques are comparable in accuracy to those of untrained humans or even trained professionals in some situations [<xref ref-type="bibr" rid="ref92">92</xref>,<xref ref-type="bibr" rid="ref110">110</xref>]. Some limitations exist, however, in relying on traditional methods, which indicates that another auxiliary system might be necessary to improve the overall accuracy and usefulness of a future developed system.</p>
        </sec>
        <sec>
          <title>Using Recipe-Specific Nutritional Values</title>
          <p>Currently, most existing systems rely on standard food composition tables to calculate the representative nutritional values of foods. While the United States Department of Agriculture National Nutrient Database is considered comprehensive, in practical dietetics, it is important to use recipe-specific nutritional values when available. For example, differentiating between a Subway sandwich (Subway IP LLC) and a Starbucks sandwich (Starbucks Corporation) using a food identification system may be feasible with a large image dataset of these specific sandwiches. However, it could be more straightforward to use location data to determine the brand of the sandwich.</p>
          <p>Furthermore, when a food product has a nutrition facts label, it is essential to obtain the representative values directly from the label instead of relying solely on food composition tables. This can be accomplished either through a system equipped with optical character recognition or by accessing a vast nutrition facts label database, such as Open Food Facts [<xref ref-type="bibr" rid="ref111">111</xref>]. By incorporating these recipe-specific and label-based nutritional values, the accuracy and relevance of food nutrient assessment systems can be significantly improved.</p>
        </sec>
        <sec>
          <title>Challenges With Density Determination</title>
          <p>The conversion of volume to weight in volume estimation approaches relies on food-specific density values, which can pose technical difficulties [<xref ref-type="bibr" rid="ref112">112</xref>]. Furthermore, food-specific density is not provided in all food compositions; therefore, it must be obtained through calculation. Most food composition tables provide nutrient content per 100 grams of edible food, as it is derived from direct chemical analysis procedures. By contrast, food portion sizes are often measured in household units, such as teaspoons, tablespoons, or measuring cups.</p>
          <p>The portion-specific weight must be divided by the standard volume of the household unit to calculate density. For example, according to the Thai food composition table, cooked mung bean sprouts weigh 78 and 34 grams for 1 serving (240 mL) and 1/3 serving (80 mL), respectively. This results in food-specific densities of 0.325 and 0.425 g/mL. However, relying on a single representative density value may not be appropriate, as it can contribute to overall system errors beyond just volume estimation. To address this challenge, a calibration curve-like method should be used instead of relying on a single density value. The accuracy and reliability of volume estimation systems can be improved, thus ensuring more precise and consistent results.</p>
        </sec>
        <sec>
          <title>Guessing Missing Information</title>
          <p>When assessing food intake, dietitians and nutritionists often encounter situations where certain food items are not readily available in food composition tables or nutrition databases. In such cases, a comprehensive analysis of the food needs to be conducted, breaking it down into its individual components. Using plain fried rice with egg as an example, the 2 cups of fried rice should be divided into at least 2 components: steamed white rice and chicken egg, which are visible in the image. However, additional components, such as seasonings and cooking oil, must be estimated. Seasonings, such as salt, soy sauce, and sugar, are typically added to enhance flavor, while cooking oil is often used to prevent food from sticking to the pan and to aid in the cooking process. Furthermore, the amount of seasoning and cooking oil may vary based on the personal experience or preference of the nutritionist who analyzes the food. Consequently, in nutrition research, it is recommended to have at least 2 or 3 analysts to reduce individual bias [<xref ref-type="bibr" rid="ref113">113</xref>]. Using algorithms, which are based on standardized criteria, the variation caused by personal experience and subjectivity can be reduced.</p>
        </sec>
        <sec>
          <title>Explainable System and Trust Issues</title>
          <p>Using AI in health care has attracted close attention from health care communities worldwide, raising concerns about how to trust unexplained systems [<xref ref-type="bibr" rid="ref114">114</xref>-<xref ref-type="bibr" rid="ref116">116</xref>]. This concern is also shared by nutrition professionals. The black-box nature of deep learning algorithms makes it difficult for users to identify incorrect outputs.</p>
          <p>When dietitians and nutritionists review a participant’s food photo and the estimated calorie intake is lower than expected, it could be due to underreporting or misreporting by the participant, selection of an inappropriate food item, forgetting to include certain amounts of oil in recipe analysis, or underestimating portion sizes. Dietitians and nutritionists can easily identify these errors. However, if the system only provides calorie outputs without additional information, it fails to establish trust with the users. Consequently, involving nutrition professionals in the development and evaluation of these systems is crucial to build trust and ensure that the technology meets their requirements.</p>
        </sec>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Principal Findings</title>
        <p>In this study, we investigated the AI techniques used for IADA and analyzed the available literature to identify the principal findings in this field. Our scoping review encompassed 522 articles, and after careful evaluation, we included 84 (16.1%) articles for analysis, spanning from 2008 to 2021. After 2015, the increase in the number of published articles in this field can be attributed to various factors, including the growing availability of large datasets, advancements in AI development frameworks, and improved accessibility of hardware resources for AI-related tasks.</p>
        <p>The principal findings were categorized into 2 main areas: food identification and food volume estimation. The chronological presentation of the articles allowed a better understanding of the algorithms’ complexity and the improvements achieved in accuracy. The transition from handcrafted food identification algorithms to deep learning-based algorithms occurred within a relatively short span of 5 years. This shift demonstrated the transformative power of deep learning in enhancing the accuracy and efficiency of food identification in image-based dietary assessment. Regarding food volume estimation, 4 different approaches were identified. However, all of these approaches share the common goal of translating 2D object views into 3D representations within a computer system and then converting these to weight to estimate representative nutritional values from a food composition table. While these approaches each have their strengths and limitations, the use of depth cameras is straightforward for measuring volume with fewer assumptions and might result in the lowest error rates compared with other methods. Nonetheless, the limited availability of depth cameras in some smartphones poses a significant challenge for implementing this approach. However, recent advancements in deep learning techniques offer promising alternatives to overcome the need for specific hardware to estimate volume and even directly estimate nutritional values without using a food composition table.</p>
      </sec>
      <sec>
        <title>Comparison With Prior Work</title>
        <p>During our search for relevant studies, we encountered several review articles published before ours. Gemming et al [<xref ref-type="bibr" rid="ref117">117</xref>] organized notable studies from the early stages of IADA development. Doulah et al [<xref ref-type="bibr" rid="ref118">118</xref>] primarily focused on computational methods for determining energy intake, including IADA techniques and wearable devices aimed at replacing traditional dietary assessment methods. Lo et al [<xref ref-type="bibr" rid="ref119">119</xref>] provided detailed explanations of techniques for both food recognition and volume estimation used in IADA studies. The survey from Subhi et al [<xref ref-type="bibr" rid="ref120">120</xref>] and the systematic review from Dalakleidi et al [<xref ref-type="bibr" rid="ref121">121</xref>] offer comprehensive comparisons of IADA systems, organized based on the subtasks of multistage architecture. Tay et al [<xref ref-type="bibr" rid="ref122">122</xref>] provided an exclusive report on computational food volume estimation. While these review articles provide extensive information, they may be difficult to comprehend for nontechnical individuals, such as dietitians and nutritionists. This review is tailored to serve as a starting point for those who may not be familiar with the technical terminology and complexity associated with this field, presenting information in clear chronological order for easy following and comparison.</p>
      </sec>
      <sec>
        <title>Strengths and Limitations</title>
        <p>While technology has advanced rapidly over the past 2 decades, it is important to acknowledge that some of the studies included in our review may have become outdated in terms of algorithm complexity, measurement techniques, and the accuracy of predicted results. Nonetheless, the findings from these earlier studies remain crucial from a dietitian’s perspective and provide valuable insights for future research and solution development. Although our search strategy was comprehensive and systematic, it is important to acknowledge that there may be studies that we were unable to identify or include in this study. Despite this limitation, our analysis provides a comprehensive overview of the principal findings in the field of IADA, shedding light on the potential and challenges of incorporating AI techniques into this domain.</p>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>The application of AI has demonstrated promising results in enhancing the accuracy and efficiency of IADA. Advanced technologies, such as deep learning, CNNs, multitask CNNs, and generative adversarial networks, have significantly improved digitization of dietary intake. However, despite their potential, there are still challenges to overcome when implementing these technologies in real-world settings. To achieve broader coverage and increased reliability, integrating various inputs, such as food barcodes, direct label readers through optical character recognition, and location-specific recipes, could enhance the capabilities of IADA systems.</p>
        <p>Additional research and development efforts are needed to address persistent issues, such as the limited availability of depth cameras, interassessor variation, missing information, and density estimation. While AI-based approaches offer valuable insights into dietary intake, it is essential to recognize that they were not designed to capture long-term usual intake entirely, which could be determined by aggregating self-reported and objective measures of dietary intake.</p>
        <p>Furthermore, combining usual intake with additional aspects of health, such as physical activity, sleep patterns, and body composition, is required for a comprehensive understanding of the relationship between lifestyle, health, and disease. By overcoming these challenges, AI-based approaches have the potential to revolutionize dietary assessment and contribute to a better understanding of an individual’s intake, eating patterns, and overall nutritional health.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group>
      <supplementary-material id="app1">
        <label>Multimedia Appendix 1</label>
        <p>PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews) checklist.</p>
        <media xlink:href="jmir_v26i1e51432_app1.pdf" xlink:title="PDF File  (Adobe PDF File), 101 KB"/>
      </supplementary-material>
    </app-group>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">24HR</term>
          <def>
            <p>24-hour dietary recall</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">AI</term>
          <def>
            <p>artificial intelligence</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">AR</term>
          <def>
            <p>augmented reality</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">CNN</term>
          <def>
            <p>convolutional neural network</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">EFR</term>
          <def>
            <p>estimated food record</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">IADA</term>
          <def>
            <p>image-assisted dietary assessment</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">PRISMA-ScR</term>
          <def>
            <p>Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">ResNet</term>
          <def>
            <p>residual network</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">RGB</term>
          <def>
            <p>red-green-blue (color model based on additive color primaries)</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb10">RGBD</term>
          <def>
            <p>red-green-blue with depth</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb11">SVM</term>
          <def>
            <p>support vector machine</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb12">TADA</term>
          <def>
            <p>Technology Assisted Dietary Assessment</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb13">ToF</term>
          <def>
            <p>time-of-flight</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb14">UEC</term>
          <def>
            <p>University of Electro-Communications</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>This work was funded by the Program Management Unit for Human Resources &#38; Institutional Development, Research, and Innovation agency under contract B04G640044. The authors would like to thank the Institute of Nutrition, Mahidol University, for the support and use of their facilities. The authors gratefully thank Sabri Bromage for his valuable suggestions and to George Attig for editing the manuscript.</p>
    </ack>
    <notes>
      <sec>
        <title>Data Availability</title>
        <p>All data generated or analyzed during this study are included in this published article.</p>
      </sec>
    </notes>
    <fn-group>
      <fn fn-type="con">
        <p>PC wrote the manuscript and provided data for tables and figures. PC and WK conceived and designed the conceptual framework. PC, AP, and PS discussed implications, limitations, and potential future directions. All authors reviewed, edited, and approved the final manuscript.</p>
      </fn>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Thompson</surname>
              <given-names>FE</given-names>
            </name>
            <name name-style="western">
              <surname>Subar</surname>
              <given-names>AF</given-names>
            </name>
            <name name-style="western">
              <surname>Loria</surname>
              <given-names>CM</given-names>
            </name>
            <name name-style="western">
              <surname>Reedy</surname>
              <given-names>JL</given-names>
            </name>
            <name name-style="western">
              <surname>Baranowski</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Need for technological innovation in dietary assessment</article-title>
          <source>J Am Diet Assoc</source>
          <year>2010</year>
          <month>01</month>
          <volume>110</volume>
          <issue>1</issue>
          <fpage>48</fpage>
          <lpage>51</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/20102826"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jada.2009.10.008</pub-id>
          <pub-id pub-id-type="medline">20102826</pub-id>
          <pub-id pub-id-type="pii">S0002-8223(09)01684-8</pub-id>
          <pub-id pub-id-type="pmcid">PMC2823476</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Arksey</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>O'Malley</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>Scoping studies: towards a methodological framework</article-title>
          <source>Int J Soc Res Methodol</source>
          <year>2005</year>
          <volume>8</volume>
          <issue>1</issue>
          <fpage>19</fpage>
          <lpage>32</lpage>
          <pub-id pub-id-type="doi">10.1080/1364557032000119616</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tricco</surname>
              <given-names>AC</given-names>
            </name>
            <name name-style="western">
              <surname>Lillie</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Zarin</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>O'Brien</surname>
              <given-names>KK</given-names>
            </name>
            <name name-style="western">
              <surname>Colquhoun</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Levac</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Moher</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Peters</surname>
              <given-names>MD</given-names>
            </name>
            <name name-style="western">
              <surname>Horsley</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Weeks</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Hempel</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Akl</surname>
              <given-names>EA</given-names>
            </name>
            <name name-style="western">
              <surname>Chang</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>McGowan</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Stewart</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Hartling</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Aldcroft</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Wilson</surname>
              <given-names>MG</given-names>
            </name>
            <name name-style="western">
              <surname>Garritty</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Lewin</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Godfrey</surname>
              <given-names>CM</given-names>
            </name>
            <name name-style="western">
              <surname>Macdonald</surname>
              <given-names>MT</given-names>
            </name>
            <name name-style="western">
              <surname>Langlois</surname>
              <given-names>EV</given-names>
            </name>
            <name name-style="western">
              <surname>Soares-Weiser</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Moriarty</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Clifford</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Tunçalp</surname>
              <given-names>Ö</given-names>
            </name>
            <name name-style="western">
              <surname>Straus</surname>
              <given-names>SE</given-names>
            </name>
          </person-group>
          <article-title>PRISMA extension for scoping reviews (PRISMA-ScR): checklist and explanation</article-title>
          <source>Ann Intern Med</source>
          <year>2018</year>
          <month>10</month>
          <day>02</day>
          <volume>169</volume>
          <issue>7</issue>
          <fpage>467</fpage>
          <lpage>73</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.acpjournals.org/doi/abs/10.7326/M18-0850?url_ver=Z39.88-2003&#38;rfr_id=ori:rid:crossref.org&#38;rfr_dat=cr_pub  0pubmed"/>
          </comment>
          <pub-id pub-id-type="doi">10.7326/M18-0850</pub-id>
          <pub-id pub-id-type="medline">30178033</pub-id>
          <pub-id pub-id-type="pii">2700389</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Gibson</surname>
              <given-names>RS</given-names>
            </name>
          </person-group>
          <source>Principles of Nutritional Assessment</source>
          <year>2005</year>
          <publisher-loc>Oxford, UK</publisher-loc>
          <publisher-name>Oxford University Press</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zegman</surname>
              <given-names>MA</given-names>
            </name>
          </person-group>
          <article-title>Errors in food recording and calorie estimation: clinical and theoretical implications for obesity</article-title>
          <source>Addict Behav</source>
          <year>1984</year>
          <volume>9</volume>
          <issue>4</issue>
          <fpage>347</fpage>
          <lpage>50</lpage>
          <pub-id pub-id-type="doi">10.1016/0306-4603(84)90033-9</pub-id>
          <pub-id pub-id-type="medline">6532141</pub-id>
          <pub-id pub-id-type="pii">0306-4603(84)90033-9</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Livingstone</surname>
              <given-names>MB</given-names>
            </name>
            <name name-style="western">
              <surname>Robson</surname>
              <given-names>PJ</given-names>
            </name>
            <name name-style="western">
              <surname>Wallace</surname>
              <given-names>JM</given-names>
            </name>
          </person-group>
          <article-title>Issues in dietary intake assessment of children and adolescents</article-title>
          <source>Br J Nutr</source>
          <year>2004</year>
          <month>10</month>
          <volume>92 Suppl 2</volume>
          <fpage>S213</fpage>
          <lpage>22</lpage>
          <pub-id pub-id-type="doi">10.1079/bjn20041169</pub-id>
          <pub-id pub-id-type="medline">15522159</pub-id>
          <pub-id pub-id-type="pii">S0007114504002326</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Goris</surname>
              <given-names>AH</given-names>
            </name>
            <name name-style="western">
              <surname>Westerterp-Plantenga</surname>
              <given-names>MS</given-names>
            </name>
            <name name-style="western">
              <surname>Westerterp</surname>
              <given-names>KR</given-names>
            </name>
          </person-group>
          <article-title>Undereating and underrecording of habitual food intake in obese men: selective underreporting of fat intake</article-title>
          <source>Am J Clin Nutr</source>
          <year>2000</year>
          <month>01</month>
          <volume>71</volume>
          <issue>1</issue>
          <fpage>130</fpage>
          <lpage>4</lpage>
          <pub-id pub-id-type="doi">10.1093/ajcn/71.1.130</pub-id>
          <pub-id pub-id-type="medline">10617957</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>DH</given-names>
            </name>
            <name name-style="western">
              <surname>Kogashiwa</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Ohta</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Kira</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Validity and reliability of a dietary assessment method: the application of a digital camera with a mobile phone card attachment</article-title>
          <source>J Nutr Sci Vitaminol (Tokyo)</source>
          <year>2002</year>
          <month>12</month>
          <volume>48</volume>
          <issue>6</issue>
          <fpage>498</fpage>
          <lpage>504</lpage>
          <pub-id pub-id-type="doi">10.3177/jnsv.48.498</pub-id>
          <pub-id pub-id-type="medline">12775117</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Nicklas</surname>
              <given-names>TA</given-names>
            </name>
            <name name-style="western">
              <surname>O'Neil</surname>
              <given-names>CE</given-names>
            </name>
            <name name-style="western">
              <surname>Stuff</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Goodell</surname>
              <given-names>LS</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Martin</surname>
              <given-names>CK</given-names>
            </name>
          </person-group>
          <article-title>Validity and feasibility of a digital diet estimation method for use with preschool children: a pilot study</article-title>
          <source>J Nutr Educ Behav</source>
          <year>2012</year>
          <volume>44</volume>
          <issue>6</issue>
          <fpage>618</fpage>
          <lpage>23</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/22727939"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jneb.2011.12.001</pub-id>
          <pub-id pub-id-type="medline">22727939</pub-id>
          <pub-id pub-id-type="pii">S1499-4046(11)00649-X</pub-id>
          <pub-id pub-id-type="pmcid">PMC3764479</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>DH</given-names>
            </name>
            <name name-style="western">
              <surname>Kogashiwa</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Kira</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Development of a new instrument for evaluating individuals' dietary intakes</article-title>
          <source>J Am Diet Assoc</source>
          <year>2006</year>
          <month>10</month>
          <volume>106</volume>
          <issue>10</issue>
          <fpage>1588</fpage>
          <lpage>93</lpage>
          <pub-id pub-id-type="doi">10.1016/j.jada.2006.07.004</pub-id>
          <pub-id pub-id-type="medline">17000191</pub-id>
          <pub-id pub-id-type="pii">S0002-8223(06)01691-9</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Gregory</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Walwyn</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Bloor</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Amin</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>A feasibility study of the use of photographic food diaries in the management of obesity</article-title>
          <source>Pract Diabetes Int</source>
          <year>2006</year>
          <month>04</month>
          <day>10</day>
          <volume>23</volume>
          <issue>2</issue>
          <fpage>66</fpage>
          <lpage>8</lpage>
          <pub-id pub-id-type="doi">10.1002/pdi.899</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Williamson</surname>
              <given-names>DA</given-names>
            </name>
            <name name-style="western">
              <surname>Allen</surname>
              <given-names>HR</given-names>
            </name>
            <name name-style="western">
              <surname>Martin</surname>
              <given-names>PD</given-names>
            </name>
            <name name-style="western">
              <surname>Alfonso</surname>
              <given-names>AJ</given-names>
            </name>
            <name name-style="western">
              <surname>Gerald</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Hunt</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Comparison of digital photography to weighed and visual estimation of portion sizes</article-title>
          <source>J Am Diet Assoc</source>
          <year>2003</year>
          <month>09</month>
          <volume>103</volume>
          <issue>9</issue>
          <fpage>1139</fpage>
          <lpage>45</lpage>
          <pub-id pub-id-type="doi">10.1016/s0002-8223(03)00974-x</pub-id>
          <pub-id pub-id-type="medline">12963941</pub-id>
          <pub-id pub-id-type="pii">S000282230300974X</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Martin</surname>
              <given-names>CK</given-names>
            </name>
            <name name-style="western">
              <surname>Han</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Coulon</surname>
              <given-names>SM</given-names>
            </name>
            <name name-style="western">
              <surname>Allen</surname>
              <given-names>HR</given-names>
            </name>
            <name name-style="western">
              <surname>Champagne</surname>
              <given-names>CM</given-names>
            </name>
            <name name-style="western">
              <surname>Anton</surname>
              <given-names>SD</given-names>
            </name>
          </person-group>
          <article-title>A novel method to remotely measure food intake of free-living individuals in real time: the remote food photography method</article-title>
          <source>Br J Nutr</source>
          <year>2009</year>
          <month>02</month>
          <volume>101</volume>
          <issue>3</issue>
          <fpage>446</fpage>
          <lpage>56</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/18616837"/>
          </comment>
          <pub-id pub-id-type="doi">10.1017/S0007114508027438</pub-id>
          <pub-id pub-id-type="medline">18616837</pub-id>
          <pub-id pub-id-type="pii">S0007114508027438</pub-id>
          <pub-id pub-id-type="pmcid">PMC2626133</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dahl Lassen</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Poulsen</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Ernst</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Kaae Andersen</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Biltoft-Jensen</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Tetens</surname>
              <given-names>I</given-names>
            </name>
          </person-group>
          <article-title>Evaluation of a digital method to assess evening meal intake in a free-living adult population</article-title>
          <source>Food Nutr Res</source>
          <year>2010</year>
          <month>11</month>
          <day>12</day>
          <volume>54</volume>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.tandfonline.com/doi/full/10.3402/fnr.v54i0.5311"/>
          </comment>
          <pub-id pub-id-type="doi">10.3402/fnr.v54i0.5311</pub-id>
          <pub-id pub-id-type="medline">21085516</pub-id>
          <pub-id pub-id-type="pii">5311</pub-id>
          <pub-id pub-id-type="pmcid">PMC2982786</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rollo</surname>
              <given-names>ME</given-names>
            </name>
            <name name-style="western">
              <surname>Ash</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Lyons-Wall</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Russell</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Trial of a mobile phone method for recording dietary intake in adults with type 2 diabetes: evaluation and implications for future applications</article-title>
          <source>J Telemed Telecare</source>
          <year>2011</year>
          <volume>17</volume>
          <issue>6</issue>
          <fpage>318</fpage>
          <lpage>23</lpage>
          <pub-id pub-id-type="doi">10.1258/jtt.2011.100906</pub-id>
          <pub-id pub-id-type="medline">21844173</pub-id>
          <pub-id pub-id-type="pii">jtt.2011.100906</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Daugherty</surname>
              <given-names>BL</given-names>
            </name>
            <name name-style="western">
              <surname>Schap</surname>
              <given-names>TE</given-names>
            </name>
            <name name-style="western">
              <surname>Ettienne-Gittens</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>FM</given-names>
            </name>
            <name name-style="western">
              <surname>Bosch</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Delp</surname>
              <given-names>EJ</given-names>
            </name>
            <name name-style="western">
              <surname>Ebert</surname>
              <given-names>DS</given-names>
            </name>
            <name name-style="western">
              <surname>Kerr</surname>
              <given-names>DA</given-names>
            </name>
            <name name-style="western">
              <surname>Boushey</surname>
              <given-names>CJ</given-names>
            </name>
          </person-group>
          <article-title>Novel technologies for assessing dietary intake: evaluating the usability of a mobile telephone food record among adults and adolescents</article-title>
          <source>J Med Internet Res</source>
          <year>2012</year>
          <month>04</month>
          <day>13</day>
          <volume>14</volume>
          <issue>2</issue>
          <fpage>e58</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://www.jmir.org/2012/2/e58/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/jmir.1967</pub-id>
          <pub-id pub-id-type="medline">22504018</pub-id>
          <pub-id pub-id-type="pii">v14i2e58</pub-id>
          <pub-id pub-id-type="pmcid">PMC3376510</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Six</surname>
              <given-names>BL</given-names>
            </name>
            <name name-style="western">
              <surname>Schap</surname>
              <given-names>TE</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>FM</given-names>
            </name>
            <name name-style="western">
              <surname>Mariappan</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Bosch</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Delp</surname>
              <given-names>EJ</given-names>
            </name>
            <name name-style="western">
              <surname>Ebert</surname>
              <given-names>DS</given-names>
            </name>
            <name name-style="western">
              <surname>Kerr</surname>
              <given-names>DA</given-names>
            </name>
            <name name-style="western">
              <surname>Boushey</surname>
              <given-names>CJ</given-names>
            </name>
          </person-group>
          <article-title>Evidence-based development of a mobile telephone food record</article-title>
          <source>J Am Diet Assoc</source>
          <year>2010</year>
          <month>01</month>
          <volume>110</volume>
          <issue>1</issue>
          <fpage>74</fpage>
          <lpage>9</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/20102830"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jada.2009.10.010</pub-id>
          <pub-id pub-id-type="medline">20102830</pub-id>
          <pub-id pub-id-type="pii">S0002-8223(09)01686-1</pub-id>
          <pub-id pub-id-type="pmcid">PMC3042797</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Shroff</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Smailagic</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Siewiorek</surname>
              <given-names>DP</given-names>
            </name>
          </person-group>
          <article-title>Wearable context-aware food recognition for calorie monitoring</article-title>
          <source>Proceedings of the 12th IEEE International Symposium on Wearable Computers</source>
          <year>2008</year>
          <conf-name>ISWC 2008</conf-name>
          <conf-date>September 28-October 1, 2008</conf-date>
          <conf-loc>Pittsburgh, PA</conf-loc>
          <pub-id pub-id-type="doi">10.1109/iswc.2008.4911602</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Dhingra</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Sukthankar</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>PFID: Pittsburgh fast-food image dataset</article-title>
          <source>Proceedings of the 16th IEEE International Conference on Image Processing</source>
          <year>2009</year>
          <conf-name>ICIP 2009</conf-name>
          <conf-date>November 7-10, 2009</conf-date>
          <conf-loc>Cairo, Egypt</conf-loc>
          <pub-id pub-id-type="doi">10.1109/icip.2009.5413511</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Pomerleau</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Sukthankar</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Food recognition using statistics of pairwise local features</article-title>
          <source>Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition</source>
          <year>2010</year>
          <conf-name>CVPR 2010</conf-name>
          <conf-date>June 13-18, 2010</conf-date>
          <conf-loc>San Francisco, CA</conf-loc>
          <pub-id pub-id-type="doi">10.1109/cvpr.2010.5539907</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Taichi</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Keiji</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>A food image recognition system with Multiple Kernel Learning</article-title>
          <source>Proceedings of the 16th IEEE International Conference on Image Processing</source>
          <year>2009</year>
          <conf-name>ICIP 2009</conf-name>
          <conf-date>November 7-10, 2009</conf-date>
          <conf-loc>Cairo, Egypt</conf-loc>
          <pub-id pub-id-type="doi">10.1109/icip.2009.5413400</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bosch</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Khanna</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Boushey</surname>
              <given-names>CJ</given-names>
            </name>
            <name name-style="western">
              <surname>Delp</surname>
              <given-names>EJ</given-names>
            </name>
          </person-group>
          <article-title>Combining global and local features for food identification in dietary assessment</article-title>
          <source>Proc Int Conf Image Proc</source>
          <year>2011</year>
          <month>09</month>
          <volume>2011</volume>
          <fpage>1789</fpage>
          <lpage>92</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/25110454"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/ICIP.2011.6115809</pub-id>
          <pub-id pub-id-type="medline">25110454</pub-id>
          <pub-id pub-id-type="pmcid">PMC4123454</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Anthimopoulos</surname>
              <given-names>MM</given-names>
            </name>
            <name name-style="western">
              <surname>Gianola</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Scarnato</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Diem</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Mougiakakou</surname>
              <given-names>SG</given-names>
            </name>
          </person-group>
          <article-title>A food recognition system for diabetic patients based on an optimized bag-of-features model</article-title>
          <source>IEEE J Biomed Health Inform</source>
          <year>2014</year>
          <month>07</month>
          <volume>18</volume>
          <issue>4</issue>
          <fpage>1261</fpage>
          <lpage>71</lpage>
          <pub-id pub-id-type="doi">10.1109/JBHI.2014.2308928</pub-id>
          <pub-id pub-id-type="medline">25014934</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Bosch</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Woo</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Boushey</surname>
              <given-names>CJ</given-names>
            </name>
            <name name-style="western">
              <surname>Ebert</surname>
              <given-names>DS</given-names>
            </name>
            <name name-style="western">
              <surname>Delp</surname>
              <given-names>EJ</given-names>
            </name>
          </person-group>
          <article-title>The use of mobile devices in aiding dietary assessment and evaluation</article-title>
          <source>IEEE J Sel Top Signal Process</source>
          <year>2010</year>
          <month>08</month>
          <volume>4</volume>
          <issue>4</issue>
          <fpage>756</fpage>
          <lpage>66</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/20862266"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/JSTSP.2010.2051471</pub-id>
          <pub-id pub-id-type="medline">20862266</pub-id>
          <pub-id pub-id-type="pmcid">PMC2941896</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Mariappan</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Boushey</surname>
              <given-names>CJ</given-names>
            </name>
            <name name-style="western">
              <surname>Kerr</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Lutes</surname>
              <given-names>KD</given-names>
            </name>
            <name name-style="western">
              <surname>Ebert</surname>
              <given-names>DS</given-names>
            </name>
            <name name-style="western">
              <surname>Delp</surname>
              <given-names>EJ</given-names>
            </name>
          </person-group>
          <article-title>Technology-assisted dietary assessment</article-title>
          <source>Proc SPIE Int Soc Opt Eng</source>
          <year>2008</year>
          <month>03</month>
          <day>20</day>
          <volume>6814</volume>
          <fpage>681411</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/22128303"/>
          </comment>
          <pub-id pub-id-type="doi">10.1117/12.778616</pub-id>
          <pub-id pub-id-type="medline">22128303</pub-id>
          <pub-id pub-id-type="pmcid">PMC3224859</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Puri</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Yu</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Divakaran</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Sawhney</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>Recognition and volume estimation of food intake using a mobile device</article-title>
          <source>Proceedings of the Workshop on Applications of Computer Vision</source>
          <year>2009</year>
          <conf-name>WACV 2009</conf-name>
          <conf-date>December 7-8, 2009</conf-date>
          <conf-loc>Snowbird, UT</conf-loc>
          <pub-id pub-id-type="doi">10.1109/wacv.2009.5403087</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Myers</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Johnston</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Rathod</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Korattikara</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Gorban</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Silberman</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Guadarrama</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Papandreou</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Murphy</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Im2Calories: towards an automated mobile vision food diary</article-title>
          <source>Proceedings of the  IEEE International Conference on Computer Vision</source>
          <year>2015</year>
          <conf-name>ICCV 2015</conf-name>
          <conf-date>December 7-13, 2015</conf-date>
          <conf-loc>Santiago, Chile</conf-loc>
          <pub-id pub-id-type="doi">10.1109/iccv.2015.146</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Christ</surname>
              <given-names>PF</given-names>
            </name>
            <name name-style="western">
              <surname>Schlecht</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Ettlinger</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Grün</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Heinle</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Tatavatry</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Ahmadi</surname>
              <given-names>SA</given-names>
            </name>
            <name name-style="western">
              <surname>Diepold</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Menze</surname>
              <given-names>BH</given-names>
            </name>
          </person-group>
          <article-title>Diabetes60 — inferring bread units from food images using fully convolutional neural networks</article-title>
          <source>Proceedings of the IEEE International Conference on Computer Vision Workshops</source>
          <year>2017</year>
          <conf-name>ICCVW 2017</conf-name>
          <conf-date>October 22-29, 2017</conf-date>
          <conf-loc>Venice, Italy</conf-loc>
          <pub-id pub-id-type="doi">10.1109/iccvw.2017.180</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ege</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Yanai</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Image-based food calorie estimation using knowledge on food categories, ingredients and cooking directions</article-title>
          <source>Proceedings of the on Thematic Workshops of ACM Multimedia 2017</source>
          <year>2017</year>
          <conf-name>Thematic Workshops '17</conf-name>
          <conf-date>October 23-27, 2017</conf-date>
          <conf-loc>Mountain View,CA</conf-loc>
          <pub-id pub-id-type="doi">10.1145/3126686.3126742</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Fang</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Shao</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Kerr</surname>
              <given-names>DA</given-names>
            </name>
            <name name-style="western">
              <surname>Boushey</surname>
              <given-names>CJ</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>F</given-names>
            </name>
          </person-group>
          <article-title>An end-to-end image-based automatic food energy estimation technique based on learned energy distribution images: protocol and methodology</article-title>
          <source>Nutrients</source>
          <year>2019</year>
          <month>04</month>
          <day>18</day>
          <volume>11</volume>
          <issue>4</issue>
          <fpage>877</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mdpi.com/resolver?pii=nu11040877"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/nu11040877</pub-id>
          <pub-id pub-id-type="medline">31003547</pub-id>
          <pub-id pub-id-type="pii">nu11040877</pub-id>
          <pub-id pub-id-type="pmcid">PMC6521161</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Bosch</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Boushey</surname>
              <given-names>CJ</given-names>
            </name>
            <name name-style="western">
              <surname>Delp</surname>
              <given-names>EJ</given-names>
            </name>
          </person-group>
          <article-title>An image analysis system for dietary assessment and evaluation</article-title>
          <source>Proc Int Conf Image Proc</source>
          <year>2010</year>
          <fpage>1853</fpage>
          <lpage>6</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/22025261"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/ICIP.2010.5650848</pub-id>
          <pub-id pub-id-type="medline">22025261</pub-id>
          <pub-id pub-id-type="pmcid">PMC3198857</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pouladzadeh</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Shirmohammadi</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Al-Maghrabi</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Measuring calorie and nutrition from food image</article-title>
          <source>IEEE Trans Instrum Meas</source>
          <year>2014</year>
          <month>8</month>
          <volume>63</volume>
          <issue>8</issue>
          <fpage>1947</fpage>
          <lpage>56</lpage>
          <pub-id pub-id-type="doi">10.1109/TIM.2014.2303533</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kawano</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Yanai</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>FoodCam-256: a large-scale real-time mobile food recognition system employing high-dimensional features and compression of classifier weights</article-title>
          <source>Proceedings of the 22nd ACM International Conference on Multimedia</source>
          <year>2014</year>
          <conf-name>MM '14</conf-name>
          <conf-date>November 3-7, 2014</conf-date>
          <conf-loc>Orlando, FL</conf-loc>
          <pub-id pub-id-type="doi">10.1145/2647868.2654869</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lecun</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Bottou</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Bengio</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Haffner</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Gradient-based learning applied to document recognition</article-title>
          <source>Proc IEEE</source>
          <year>1998</year>
          <month>11</month>
          <volume>86</volume>
          <issue>11</issue>
          <fpage>2278</fpage>
          <lpage>324</lpage>
          <pub-id pub-id-type="doi">10.1109/5.726791</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref35">
        <label>35</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kawano</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Yanai</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Food image recognition with deep convolutional features</article-title>
          <source>Proceedings of the 2014 ACM International Joint Conference on Pervasive and Ubiquitous Computing: Adjunct Publication</source>
          <year>2014</year>
          <conf-name>UbiComp '14 Adjunct</conf-name>
          <conf-date>September 13-17, 2014</conf-date>
          <conf-loc>Seattle, WA</conf-loc>
          <pub-id pub-id-type="doi">10.1145/2638728.2641339</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref36">
        <label>36</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Matsuda</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Hoashi</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Yanai</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Recognition of multiple-food images by detecting candidate regions</article-title>
          <source>Proceedings of the IEEE International Conference on Multimedia and Expo</source>
          <year>2012</year>
          <conf-name>ICME 2012</conf-name>
          <conf-date>July 9-13, 2012</conf-date>
          <conf-loc>Melbourne, Australia</conf-loc>
          <pub-id pub-id-type="doi">10.1109/icme.2012.157</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref37">
        <label>37</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kawano</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Yanai</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Automatic expansion of a food image dataset leveraging existing categories with domain adaptation</article-title>
          <source>Proceedings of the 13th European Conference on Computer Vision</source>
          <year>2014</year>
          <conf-name>ECCV 2014</conf-name>
          <conf-date>September 6-12, 2014</conf-date>
          <conf-loc>Zurich, Switzerland</conf-loc>
          <pub-id pub-id-type="doi">10.1007/978-3-319-16199-0_1</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref38">
        <label>38</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bossard</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Guillaumin</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Gool</surname>
              <given-names>LV</given-names>
            </name>
          </person-group>
          <article-title>Food-101 – mining discriminative components with random forests</article-title>
          <source>Proceedings of the 13th European Conference on Computer Vision</source>
          <year>2014</year>
          <conf-name>ECCV 2014</conf-name>
          <conf-date>September 6-12, 2014</conf-date>
          <conf-loc>Zurich, Switzerland</conf-loc>
          <pub-id pub-id-type="doi">10.1007/978-3-319-10599-4_29</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref39">
        <label>39</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yanai</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Kawano</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Food image recognition using deep convolutional network with pre-training and fine-tuning</article-title>
          <source>Proceedings of the IEEE International Conference on Multimedia &#38; Expo Workshops</source>
          <year>2015</year>
          <conf-name>ICMEW 2015</conf-name>
          <conf-date>June 29-July 3, 2015</conf-date>
          <conf-loc>Turin, Italy</conf-loc>
          <pub-id pub-id-type="doi">10.1109/icmew.2015.7169816</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref40">
        <label>40</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Christodoulidis</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Anthimopoulos</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Mougiakakou</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Food recognition for dietary assessment using deep convolutional neural networks</article-title>
          <source>Proceedings of the International Conference on Image Analysis and Processing</source>
          <year>2015</year>
          <conf-name>ICIAP 2015</conf-name>
          <conf-date>September 7-8, 2015</conf-date>
          <conf-loc>Genoa, Italy</conf-loc>
          <pub-id pub-id-type="doi">10.1007/978-3-319-23222-5_56</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref41">
        <label>41</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Cao</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Luo</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Vokkarane</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Ma</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>DeepFood: deep learning-based food image recognition for computer-aided dietary assessment</article-title>
          <source>Proceedings of the 14th International Conference on Smart Homes and Health Telematics</source>
          <year>2016</year>
          <conf-name>ICOST 2016</conf-name>
          <conf-date>May 25-27, 2016</conf-date>
          <conf-loc>Wuhan, China</conf-loc>
          <pub-id pub-id-type="doi">10.1007/978-3-319-39601-9_4</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref42">
        <label>42</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Singla</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Yuan</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Ebrahimi</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Food/non-food image classification and food categorization using pre-trained GoogLeNet model</article-title>
          <source>Proceedings of the 2nd International Workshop on Multimedia Assisted Dietary Management</source>
          <year>2016</year>
          <conf-name>MADiMa '16</conf-name>
          <conf-date>October 16, 2016</conf-date>
          <conf-loc>Amsterdam, The Netherlands</conf-loc>
          <pub-id pub-id-type="doi">10.1145/2986035.2986039</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref43">
        <label>43</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hassannejad</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Matrella</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Ciampolini</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>De Munari</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Mordonini</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Cagnoni</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Food image recognition using very deep convolutional networks</article-title>
          <source>Proceedings of the 2nd International Workshop on Multimedia Assisted Dietary Management</source>
          <year>2016</year>
          <conf-name>MADiMa '16</conf-name>
          <conf-date>October 16, 2016</conf-date>
          <conf-loc>Amsterdam, The Netherlands</conf-loc>
          <pub-id pub-id-type="doi">10.1145/2986035.2986042</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref44">
        <label>44</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ciocca</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Napoletano</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Schettini</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Food recognition: a new dataset, experiments, and results</article-title>
          <source>IEEE J Biomed Health Inform</source>
          <year>2017</year>
          <month>05</month>
          <volume>21</volume>
          <issue>3</issue>
          <fpage>588</fpage>
          <lpage>98</lpage>
          <pub-id pub-id-type="doi">10.1109/JBHI.2016.2636441</pub-id>
          <pub-id pub-id-type="medline">28114043</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref45">
        <label>45</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Mezgec</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Koroušić Seljak</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>NutriNet: a deep learning food and drink image recognition system for dietary assessment</article-title>
          <source>Nutrients</source>
          <year>2017</year>
          <month>06</month>
          <day>27</day>
          <volume>9</volume>
          <issue>7</issue>
          <fpage>657</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mdpi.com/resolver?pii=nu9070657"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/nu9070657</pub-id>
          <pub-id pub-id-type="medline">28653995</pub-id>
          <pub-id pub-id-type="pii">nu9070657</pub-id>
          <pub-id pub-id-type="pmcid">PMC5537777</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref46">
        <label>46</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Krizhevsky</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Sutskever</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Hinton</surname>
              <given-names>GE</given-names>
            </name>
          </person-group>
          <article-title>ImageNet classification with deep convolutional neural networks</article-title>
          <source>Commun ACM</source>
          <year>2017</year>
          <month>05</month>
          <day>24</day>
          <volume>60</volume>
          <issue>6</issue>
          <fpage>84</fpage>
          <lpage>90</lpage>
          <pub-id pub-id-type="doi">10.1145/3065386</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref47">
        <label>47</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Girshick</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Donahue</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Darrell</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Malik</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Rich feature hierarchies for accurate object detection and semantic segmentation</article-title>
          <source>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source>
          <year>2014</year>
          <conf-name>CVPR 2014</conf-name>
          <conf-date>June 23-28, 2014</conf-date>
          <conf-loc>Columbus, OH</conf-loc>
          <pub-id pub-id-type="doi">10.1109/cvpr.2014.81</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref48">
        <label>48</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>He</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Ren</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Deep residual learning for image recognition</article-title>
          <source>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source>
          <year>2016</year>
          <conf-name>CVPR 2016</conf-name>
          <conf-date>June 27-30, 2016</conf-date>
          <conf-loc>Las Vegas, NV</conf-loc>
          <pub-id pub-id-type="doi">10.1109/cvpr.2016.90</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref49">
        <label>49</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Redmon</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Divvala</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Girshick</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Farhadi</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>You only look once: unified, real-time object detection</article-title>
          <source>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source>
          <year>2016</year>
          <conf-name>CVPR 2016</conf-name>
          <conf-date>June 27-30, 2016</conf-date>
          <conf-loc>Las Vegas, NV</conf-loc>
          <pub-id pub-id-type="doi">10.1109/cvpr.2016.91</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref50">
        <label>50</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Simonyan</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Zisserman</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Very deep convolutional networks for large-scale image recognition</article-title>
          <source>arXiv. Preprint posted online on September 4, 2014</source>
          <year>2024</year>
          <pub-id pub-id-type="doi">10.48550/arXiv.1409.1556</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref51">
        <label>51</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Szegedy</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Jia</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Sermanet</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Reed</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Anguelov</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Erhan</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Vanhoucke</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Rabinovich</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Going deeper with convolutions</article-title>
          <source>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source>
          <year>2015</year>
          <conf-name>CVPR 2015</conf-name>
          <conf-date>June 7-12, 2015</conf-date>
          <conf-loc>Boston, MA</conf-loc>
          <pub-id pub-id-type="doi">10.1109/cvpr.2015.7298594</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref52">
        <label>52</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hoashi</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Joutou</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Yanai</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Image recognition of 85 food categories by feature fusion</article-title>
          <source>Proceedings of the IEEE International Symposium on Multimedia</source>
          <year>2010</year>
          <conf-name>ISM 2010</conf-name>
          <conf-date>December 13-15, 2010</conf-date>
          <conf-loc>Taichung, Taiwan</conf-loc>
          <pub-id pub-id-type="doi">10.1109/ism.2010.51</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref53">
        <label>53</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kong</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Tan</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>DietCam: regular shape food recognition with a camera phone</article-title>
          <source>Proceedings of the International Conference on Body Sensor Networks</source>
          <year>2011</year>
          <conf-name>BSN 2011</conf-name>
          <conf-date>May 23-25, 2011</conf-date>
          <conf-loc>Dallas, TX</conf-loc>
          <pub-id pub-id-type="doi">10.1109/bsn.2011.19</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref54">
        <label>54</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>He</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Khanna</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Boushey</surname>
              <given-names>CJ</given-names>
            </name>
            <name name-style="western">
              <surname>Delp</surname>
              <given-names>EJ</given-names>
            </name>
          </person-group>
          <article-title>Analysis of food images: features and classification</article-title>
          <source>Proc Int Conf Image Proc</source>
          <year>2014</year>
          <month>10</month>
          <volume>2014</volume>
          <fpage>2744</fpage>
          <lpage>8</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/28572748"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/ICIP.2014.7025555</pub-id>
          <pub-id pub-id-type="medline">28572748</pub-id>
          <pub-id pub-id-type="pmcid">PMC5448982</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref55">
        <label>55</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pandey</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Deepthi</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Mandal</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Puhan</surname>
              <given-names>NB</given-names>
            </name>
          </person-group>
          <article-title>FoodNet: recognizing foods using ensemble of deep networks</article-title>
          <source>IEEE Signal Process Lett</source>
          <year>2017</year>
          <month>12</month>
          <volume>24</volume>
          <issue>12</issue>
          <fpage>1758</fpage>
          <lpage>62</lpage>
          <pub-id pub-id-type="doi">10.1109/lsp.2017.2758862</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref56">
        <label>56</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Martinel</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Foresti</surname>
              <given-names>GL</given-names>
            </name>
            <name name-style="western">
              <surname>Micheloni</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Wide-slice residual networks for food recognition</article-title>
          <source>Proceedings of the IEEE Winter Conference on Applications of Computer Vision</source>
          <year>2018</year>
          <conf-name>WACV 2018</conf-name>
          <conf-date>March 12-15, 2018</conf-date>
          <conf-loc>Lake Tahoe, NV</conf-loc>
          <pub-id pub-id-type="doi">10.1109/wacv.2018.00068</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref57">
        <label>57</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jiang</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Min</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Luo</surname>
              <given-names>Z</given-names>
            </name>
          </person-group>
          <article-title>Multi-scale multi-view deep feature aggregation for food recognition</article-title>
          <source>IEEE Trans Image Process</source>
          <year>2020</year>
          <volume>29</volume>
          <fpage>265</fpage>
          <lpage>76</lpage>
          <pub-id pub-id-type="doi">10.1109/TIP.2019.2929447</pub-id>
          <pub-id pub-id-type="medline">31369375</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref58">
        <label>58</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Stathopoulou</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Vasiloglou</surname>
              <given-names>MF</given-names>
            </name>
            <name name-style="western">
              <surname>Pinault</surname>
              <given-names>LF</given-names>
            </name>
            <name name-style="western">
              <surname>Kiley</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Spanakis</surname>
              <given-names>EK</given-names>
            </name>
            <name name-style="western">
              <surname>Mougiakakou</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>goFOOD: an artificial intelligence system for dietary assessment</article-title>
          <source>Sensors (Basel)</source>
          <year>2020</year>
          <month>07</month>
          <day>31</day>
          <volume>20</volume>
          <issue>15</issue>
          <fpage>4283</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mdpi.com/resolver?pii=s20154283"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/s20154283</pub-id>
          <pub-id pub-id-type="medline">32752007</pub-id>
          <pub-id pub-id-type="pii">s20154283</pub-id>
          <pub-id pub-id-type="pmcid">PMC7436102</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref59">
        <label>59</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>MY</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>JH</given-names>
            </name>
            <name name-style="western">
              <surname>Hsueh</surname>
              <given-names>CY</given-names>
            </name>
          </person-group>
          <article-title>A framework of visual checkout system using convolutional neural networks for Bento buffet</article-title>
          <source>Sensors (Basel)</source>
          <year>2021</year>
          <month>04</month>
          <day>08</day>
          <volume>21</volume>
          <issue>8</issue>
          <fpage>2627</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mdpi.com/resolver?pii=s21082627"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/s21082627</pub-id>
          <pub-id pub-id-type="medline">33918027</pub-id>
          <pub-id pub-id-type="pii">s21082627</pub-id>
          <pub-id pub-id-type="pmcid">PMC8069312</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref60">
        <label>60</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tola</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Lepetit</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Fua</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>DAISY: an efficient dense descriptor applied to wide-baseline stereo</article-title>
          <source>IEEE Trans Pattern Anal Mach Intell</source>
          <year>2010</year>
          <month>05</month>
          <volume>32</volume>
          <issue>5</issue>
          <fpage>815</fpage>
          <lpage>30</lpage>
          <pub-id pub-id-type="doi">10.1109/TPAMI.2009.77</pub-id>
          <pub-id pub-id-type="medline">20299707</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref61">
        <label>61</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Galer</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <source>Photography: Foundations for Art &#38; Design: The Creative Photography</source>
          <year>2004</year>
          <publisher-loc>Waltham, MA</publisher-loc>
          <publisher-name>Focal Press</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref62">
        <label>62</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Katz</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <source>Introduction to Geometrical Optics</source>
          <year>2002</year>
          <publisher-loc>Singapore, Singapore</publisher-loc>
          <publisher-name>World Scientific</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref63">
        <label>63</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jia</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Yue</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Fernstrom</surname>
              <given-names>JD</given-names>
            </name>
            <name name-style="western">
              <surname>Yao</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Sclabassi</surname>
              <given-names>RJ</given-names>
            </name>
            <name name-style="western">
              <surname>Fernstrom</surname>
              <given-names>MH</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Imaged based estimation of food volume using circular referents in dietary assessment</article-title>
          <source>J Food Eng</source>
          <year>2012</year>
          <month>03</month>
          <volume>109</volume>
          <issue>1</issue>
          <fpage>76</fpage>
          <lpage>86</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/22523440"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jfoodeng.2011.09.031</pub-id>
          <pub-id pub-id-type="medline">22523440</pub-id>
          <pub-id pub-id-type="pmcid">PMC3328298</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref64">
        <label>64</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Okamoto</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Yanai</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>An automatic calorie estimation system of food images on a smartphone</article-title>
          <source>Proceedings of the 2nd International Workshop on Multimedia Assisted Dietary Management</source>
          <year>2016</year>
          <conf-name>MADiMa '16</conf-name>
          <conf-date>October 16, 2016</conf-date>
          <conf-loc>Amsterdam, The Netherlands</conf-loc>
          <pub-id pub-id-type="doi">10.1145/2986035.2986040</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref65">
        <label>65</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Akpa</surname>
              <given-names>EA</given-names>
            </name>
            <name name-style="western">
              <surname>Suwa</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Arakawa</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Yasumoto</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Smartphone-based food weight and calorie estimation method for effective food journaling</article-title>
          <source>SICE J Control Meas Syst Integr</source>
          <year>2021</year>
          <month>01</month>
          <day>18</day>
          <volume>10</volume>
          <issue>5</issue>
          <fpage>360</fpage>
          <lpage>9</lpage>
          <pub-id pub-id-type="doi">10.9746/jcmsi.10.360</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref66">
        <label>66</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Liang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Deep learning-based food calorie estimation method in dietary assessment</article-title>
          <source>arXiv. Preprint posted online on June 10, 2017</source>
          <year>2024</year>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/1706.04062"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref67">
        <label>67</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ege</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Shimoda</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Yanai</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>A new large-scale food image segmentation dataset and its application to food calorie estimation based on grains of rice</article-title>
          <source>Proceedings of the 5th International Workshop on Multimedia Assisted Dietary Management</source>
          <year>2019</year>
          <conf-name>MADiMa '19</conf-name>
          <conf-date>October 21, 2019</conf-date>
          <conf-loc>Nice, France</conf-loc>
          <pub-id pub-id-type="doi">10.1145/3347448.3357162</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref68">
        <label>68</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Woo</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Otsmo</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Ebert</surname>
              <given-names>DS</given-names>
            </name>
            <name name-style="western">
              <surname>Delp</surname>
              <given-names>EJ</given-names>
            </name>
            <name name-style="western">
              <surname>Boushey</surname>
              <given-names>CJ</given-names>
            </name>
          </person-group>
          <article-title>Automatic portion estimation and visual refinement in mobile dietary assessment</article-title>
          <source>Proc SPIE Int Soc Opt Eng</source>
          <year>2010</year>
          <month>01</month>
          <day>01</day>
          <volume>7533</volume>
          <fpage>75330O</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/22242198"/>
          </comment>
          <pub-id pub-id-type="doi">10.1117/12.849051</pub-id>
          <pub-id pub-id-type="medline">22242198</pub-id>
          <pub-id pub-id-type="pii">75330O</pub-id>
          <pub-id pub-id-type="pmcid">PMC3254118</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref69">
        <label>69</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chae</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Woo</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Maciejewski</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Delp</surname>
              <given-names>EJ</given-names>
            </name>
            <name name-style="western">
              <surname>Boushey</surname>
              <given-names>CJ</given-names>
            </name>
            <name name-style="western">
              <surname>Ebert</surname>
              <given-names>DS</given-names>
            </name>
          </person-group>
          <article-title>Volume estimation using food specific shape templates in mobile image-based dietary assessment</article-title>
          <source>Proc SPIE Int Soc Opt Eng</source>
          <year>2011</year>
          <month>02</month>
          <day>07</day>
          <volume>7873</volume>
          <fpage>78730K</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/22025936"/>
          </comment>
          <pub-id pub-id-type="doi">10.1117/12.876669</pub-id>
          <pub-id pub-id-type="medline">22025936</pub-id>
          <pub-id pub-id-type="pmcid">PMC3198859</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref70">
        <label>70</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>HC</given-names>
            </name>
            <name name-style="western">
              <surname>Jia</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Yue</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>YN</given-names>
            </name>
            <name name-style="western">
              <surname>Fernstrom</surname>
              <given-names>JD</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Model-based measurement of food portion size for image-based dietary assessment using 3D/2D registration</article-title>
          <source>Meas Sci Technol</source>
          <year>2013</year>
          <month>10</month>
          <volume>24</volume>
          <issue>10</issue>
          <fpage>10.1088/0957-0233/24/10/105701</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/24223474"/>
          </comment>
          <pub-id pub-id-type="doi">10.1088/0957-0233/24/10/105701</pub-id>
          <pub-id pub-id-type="medline">24223474</pub-id>
          <pub-id pub-id-type="pmcid">PMC3819104</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref71">
        <label>71</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jia</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>HC</given-names>
            </name>
            <name name-style="western">
              <surname>Yue</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Fernstrom</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Bai</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Accuracy of food portion size estimation from digital pictures acquired by a chest-worn camera</article-title>
          <source>Public Health Nutr</source>
          <year>2014</year>
          <month>08</month>
          <volume>17</volume>
          <issue>8</issue>
          <fpage>1671</fpage>
          <lpage>81</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/24476848"/>
          </comment>
          <pub-id pub-id-type="doi">10.1017/S1368980013003236</pub-id>
          <pub-id pub-id-type="medline">24476848</pub-id>
          <pub-id pub-id-type="pii">S1368980013003236</pub-id>
          <pub-id pub-id-type="pmcid">PMC4152011</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref72">
        <label>72</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tanno</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Ege</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Yanai</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>AR DeepCalorieCam V2: food calorie estimation with CNN and AR-based actual size estimation</article-title>
          <source>Proceedings of the 24th ACM Symposium on Virtual Reality Software and Technology</source>
          <year>2018</year>
          <conf-name>VRST '18</conf-name>
          <conf-date>November 28-December 1, 2018</conf-date>
          <conf-loc>Tokyo Japan</conf-loc>
          <pub-id pub-id-type="doi">10.1145/3281505.3281580</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref73">
        <label>73</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Jia</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Bucher</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Image-based food portion size estimation using a smartphone without a fiducial marker</article-title>
          <source>Public Health Nutr</source>
          <year>2019</year>
          <month>05</month>
          <volume>22</volume>
          <issue>7</issue>
          <fpage>1180</fpage>
          <lpage>92</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/29623867"/>
          </comment>
          <pub-id pub-id-type="doi">10.1017/S136898001800054X</pub-id>
          <pub-id pub-id-type="medline">29623867</pub-id>
          <pub-id pub-id-type="pii">S136898001800054X</pub-id>
          <pub-id pub-id-type="pmcid">PMC8115205</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref74">
        <label>74</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Smith</surname>
              <given-names>SP</given-names>
            </name>
            <name name-style="western">
              <surname>Adam</surname>
              <given-names>MT</given-names>
            </name>
            <name name-style="western">
              <surname>Manning</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Burrows</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Collins</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Rollo</surname>
              <given-names>ME</given-names>
            </name>
          </person-group>
          <article-title>Food volume estimation by integrating 3D image projection and manual wire mesh transformations</article-title>
          <source>IEEE Access</source>
          <year>2022</year>
          <month>05</month>
          <day>02</day>
          <volume>10</volume>
          <fpage>48367</fpage>
          <lpage>78</lpage>
          <pub-id pub-id-type="doi">10.1109/ACCESS.2022.3171584</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref75">
        <label>75</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kong</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Tan</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>DietCam: automatic dietary assessment with mobile camera phones</article-title>
          <source>Pervasive Mob Comput</source>
          <year>2012</year>
          <month>2</month>
          <volume>8</volume>
          <issue>1</issue>
          <fpage>147</fpage>
          <lpage>63</lpage>
          <pub-id pub-id-type="doi">10.1016/j.pmcj.2011.07.003</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref76">
        <label>76</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rahman</surname>
              <given-names>MH</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Pickering</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Frater</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Kerr</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Bouchey</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Delp</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Food volume estimation in a mobile phone based dietary assessment system</article-title>
          <source>Proceedings of the Eighth International Conference on Signal Image Technology and Internet Based Systems</source>
          <year>2012</year>
          <conf-name>SITIS 2012</conf-name>
          <conf-date>November 25-29, 2012</conf-date>
          <conf-loc>Sorrento, Italy</conf-loc>
          <pub-id pub-id-type="doi">10.1109/sitis.2012.146</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref77">
        <label>77</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Ye</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Albert</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Edward</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Nitin</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Carol</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>Image-based food volume estimation</article-title>
          <source>CEA13 (2013)</source>
          <year>2013</year>
          <month>10</month>
          <volume>2013</volume>
          <fpage>75</fpage>
          <lpage>80</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/28573255"/>
          </comment>
          <pub-id pub-id-type="doi">10.1145/2506023.2506037</pub-id>
          <pub-id pub-id-type="medline">28573255</pub-id>
          <pub-id pub-id-type="pmcid">PMC5448987</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref78">
        <label>78</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Anthimopoulos</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Dehais</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Shevchik</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Ransford</surname>
              <given-names>BH</given-names>
            </name>
            <name name-style="western">
              <surname>Duke</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Diem</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Mougiakakou</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Computer vision-based carbohydrate estimation for type 1 patients with diabetes using smartphones</article-title>
          <source>J Diabetes Sci Technol</source>
          <year>2015</year>
          <month>05</month>
          <volume>9</volume>
          <issue>3</issue>
          <fpage>507</fpage>
          <lpage>15</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/25883163"/>
          </comment>
          <pub-id pub-id-type="doi">10.1177/1932296815580159</pub-id>
          <pub-id pub-id-type="medline">25883163</pub-id>
          <pub-id pub-id-type="pii">1932296815580159</pub-id>
          <pub-id pub-id-type="pmcid">PMC4604531</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref79">
        <label>79</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dehais</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Anthimopoulos</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Shevchik</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Mougiakakou</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Two-view 3D reconstruction for food volume estimation</article-title>
          <source>IEEE Trans Multimedia</source>
          <year>2017</year>
          <month>5</month>
          <volume>19</volume>
          <issue>5</issue>
          <fpage>1090</fpage>
          <lpage>9</lpage>
          <pub-id pub-id-type="doi">10.1109/tmm.2016.2642792</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref80">
        <label>80</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Gao</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Lo</surname>
              <given-names>FP</given-names>
            </name>
            <name name-style="western">
              <surname>Lo</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>Food volume estimation for quantifying dietary intake with a wearable camera</article-title>
          <source>Proceedings of the IEEE 15th International Conference on Wearable and Implantable Body Sensor Networks</source>
          <year>2018</year>
          <conf-name>BSN 2018</conf-name>
          <conf-date>March 4-7, 2018</conf-date>
          <conf-loc>Las Vegas, NV</conf-loc>
          <pub-id pub-id-type="doi">10.1109/bsn.2018.8329671</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref81">
        <label>81</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ando</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Ege</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Cho</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Yanai</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>DepthCalorieCam: a mobile application for volume-based food calorie estimation using depth cameras</article-title>
          <source>Proceedings of the 5th International Workshop on Multimedia Assisted Dietary Management</source>
          <year>2019</year>
          <conf-name>MADiMa '19</conf-name>
          <conf-date>October 21, 2019</conf-date>
          <conf-loc>Nice, France</conf-loc>
          <pub-id pub-id-type="doi">10.1145/3347448.3357172</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref82">
        <label>82</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Shang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Duong</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Pepin</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Sandara-Rajan</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Mamishev</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Kristal</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>A mobile structured light system for food volume estimation</article-title>
          <source>Proceedings of the IEEE International Conference on Computer Vision Workshops</source>
          <year>2011</year>
          <conf-name>ICCVW 2011</conf-name>
          <conf-date>November 6-13, 2011</conf-date>
          <conf-loc>Barcelona, Spain</conf-loc>
          <pub-id pub-id-type="doi">10.1109/iccvw.2011.6130229</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref83">
        <label>83</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>MY</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>YH</given-names>
            </name>
            <name name-style="western">
              <surname>Ho</surname>
              <given-names>CJ</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>SH</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>SM</given-names>
            </name>
            <name name-style="western">
              <surname>Chang</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Yeh</surname>
              <given-names>CH</given-names>
            </name>
            <name name-style="western">
              <surname>Ouhyoung</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Automatic Chinese food identification and quantity estimation</article-title>
          <source>Proceedings of the SIGGRAPH Asia 2012 Technical Briefs</source>
          <year>2012</year>
          <conf-name>SA '12</conf-name>
          <conf-date>November 28-December 1, 2012</conf-date>
          <conf-loc>Singapore, Singapore</conf-loc>
          <pub-id pub-id-type="doi">10.1145/2407746.2407775</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref84">
        <label>84</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Fang</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Jiang</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Boushey</surname>
              <given-names>CJ</given-names>
            </name>
            <name name-style="western">
              <surname>Delp</surname>
              <given-names>EJ</given-names>
            </name>
          </person-group>
          <article-title>A comparison of food portion size estimation using geometric models and depth images</article-title>
          <source>Proceedings of the IEEE International Conference on Image Processing</source>
          <year>2016</year>
          <conf-name>ICIP 2016</conf-name>
          <conf-date>September 25-28, 2016</conf-date>
          <conf-loc>Phoenix, AZ</conf-loc>
          <pub-id pub-id-type="doi">10.1109/icip.2016.7532312</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref85">
        <label>85</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Flexible 3D shape measurement using projector defocusing: extended measurement range</article-title>
          <source>Opt Lett</source>
          <year>2010</year>
          <month>04</month>
          <day>01</day>
          <volume>35</volume>
          <issue>7</issue>
          <fpage>934</fpage>
          <lpage>6</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://core.ac.uk/reader/38936099?utm_source=linkout"/>
          </comment>
          <pub-id pub-id-type="doi">10.1364/OL.35.000934</pub-id>
          <pub-id pub-id-type="medline">20364174</pub-id>
          <pub-id pub-id-type="pii">196694</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref86">
        <label>86</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Alfonsi</surname>
              <given-names>JE</given-names>
            </name>
            <name name-style="western">
              <surname>Choi</surname>
              <given-names>EE</given-names>
            </name>
            <name name-style="western">
              <surname>Arshad</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Sammott</surname>
              <given-names>SA</given-names>
            </name>
            <name name-style="western">
              <surname>Pais</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Nguyen</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Maguire</surname>
              <given-names>BR</given-names>
            </name>
            <name name-style="western">
              <surname>Stinson</surname>
              <given-names>JN</given-names>
            </name>
            <name name-style="western">
              <surname>Palmert</surname>
              <given-names>MR</given-names>
            </name>
          </person-group>
          <article-title>Carbohydrate counting app using image recognition for youth with type 1 diabetes: pilot randomized control trial</article-title>
          <source>JMIR Mhealth Uhealth</source>
          <year>2020</year>
          <month>10</month>
          <day>28</day>
          <volume>8</volume>
          <issue>10</issue>
          <fpage>e22074</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://mhealth.jmir.org/2020/10/e22074/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/22074</pub-id>
          <pub-id pub-id-type="medline">33112249</pub-id>
          <pub-id pub-id-type="pii">v8i10e22074</pub-id>
          <pub-id pub-id-type="pmcid">PMC7657721</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref87">
        <label>87</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Herzig</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Nakas</surname>
              <given-names>CT</given-names>
            </name>
            <name name-style="western">
              <surname>Stalder</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Kosinski</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Laesser</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Dehais</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Jaeggi</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Leichtle</surname>
              <given-names>AB</given-names>
            </name>
            <name name-style="western">
              <surname>Dahlweid</surname>
              <given-names>FM</given-names>
            </name>
            <name name-style="western">
              <surname>Stettler</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Bally</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>Volumetric food quantification using computer vision on a depth-sensing smartphone: preclinical study</article-title>
          <source>JMIR Mhealth Uhealth</source>
          <year>2020</year>
          <month>03</month>
          <day>25</day>
          <volume>8</volume>
          <issue>3</issue>
          <fpage>e15294</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://boris.unibe.ch/id/eprint/147780"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/15294</pub-id>
          <pub-id pub-id-type="medline">32209531</pub-id>
          <pub-id pub-id-type="pii">v8i3e15294</pub-id>
          <pub-id pub-id-type="pmcid">PMC7142738</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref88">
        <label>88</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Yu</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Siddiquie</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Divakaran</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Sawhney</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>"Snap-n-eat": food recognition and nutrition estimation on a smartphone</article-title>
          <source>J Diabetes Sci Technol</source>
          <year>2015</year>
          <month>05</month>
          <volume>9</volume>
          <issue>3</issue>
          <fpage>525</fpage>
          <lpage>33</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/25901024"/>
          </comment>
          <pub-id pub-id-type="doi">10.1177/1932296815582222</pub-id>
          <pub-id pub-id-type="medline">25901024</pub-id>
          <pub-id pub-id-type="pii">1932296815582222</pub-id>
          <pub-id pub-id-type="pmcid">PMC4604540</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref89">
        <label>89</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>He</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Khanna</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Boushey</surname>
              <given-names>CJ</given-names>
            </name>
            <name name-style="western">
              <surname>Delp</surname>
              <given-names>EJ</given-names>
            </name>
          </person-group>
          <article-title>Food image analysis: segmentation, identification and weight estimation</article-title>
          <source>Proceedings of the IEEE International Conference on Multimedia and Expo</source>
          <year>2013</year>
          <conf-name>ICME 2013</conf-name>
          <conf-date>July 15-19, 2013</conf-date>
          <conf-loc>San Jose, CA</conf-loc>
          <pub-id pub-id-type="doi">10.1109/icme.2013.6607548</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref90">
        <label>90</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yue</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Jia</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Measurement of food volume based on single 2-D image without conventional camera calibration</article-title>
          <source>Annu Int Conf IEEE Eng Med Biol Soc</source>
          <year>2012</year>
          <volume>2012</volume>
          <fpage>2166</fpage>
          <lpage>9</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/23366351"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/EMBC.2012.6346390</pub-id>
          <pub-id pub-id-type="medline">23366351</pub-id>
          <pub-id pub-id-type="pmcid">PMC3739717</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref91">
        <label>91</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pouladzadeh</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Villalobos</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Almaghrabi</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Shirmohammadi</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>A novel SVM based food recognition method for calorie measurement applications</article-title>
          <source>Proceedings of the IEEE International Conference on Multimedia and Expo Workshops</source>
          <year>2012</year>
          <conf-name>ICMEW 2012</conf-name>
          <conf-date>July 9-13, 2012</conf-date>
          <conf-loc>Melbourne, Australia</conf-loc>
          <pub-id pub-id-type="doi">10.1109/icmew.2012.92</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref92">
        <label>92</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Vasiloglou</surname>
              <given-names>MF</given-names>
            </name>
            <name name-style="western">
              <surname>Mougiakakou</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Aubry</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Bokelmann</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Fricker</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Gomes</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Guntermann</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Meyer</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Studerus</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Stanga</surname>
              <given-names>Z</given-names>
            </name>
          </person-group>
          <article-title>A comparative study on carbohydrate estimation: GoCARB vs. dietitians</article-title>
          <source>Nutrients</source>
          <year>2018</year>
          <month>06</month>
          <day>07</day>
          <volume>10</volume>
          <issue>6</issue>
          <fpage>741</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://boris.unibe.ch/id/eprint/118556"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/nu10060741</pub-id>
          <pub-id pub-id-type="medline">29880772</pub-id>
          <pub-id pub-id-type="pii">nu10060741</pub-id>
          <pub-id pub-id-type="pmcid">PMC6024682</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref93">
        <label>93</label>
        <nlm-citation citation-type="web">
          <article-title>Samsung Galaxy A24, A34, and A54 to launch without depth sensing cameras</article-title>
          <source>GSMArena</source>
          <year>2022</year>
          <month>7</month>
          <day>21</day>
          <access-date>2023-04-09</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.gsmarena.com/samsung_is_going_to_remove_useless_depth_sensing_cameras_from_its_a_series_devices_next_year-news-55135.php">https://www.gsmarena.com/samsung_is_going_to_remove_useless_depth_sensing_cameras_from_its_a_series_devices_next_year-news-55135.php</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref94">
        <label>94</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Li</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Han</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>DeepVol: deep fruit volume estimation</article-title>
          <source>Proceedings of the 27th International Conference on Artificial Neural Networks</source>
          <year>2018</year>
          <conf-name>ICANN 2018</conf-name>
          <conf-date>October 4-7, 2018</conf-date>
          <conf-loc>Rhodes, Greece</conf-loc>
          <pub-id pub-id-type="doi">10.1007/978-3-030-01424-7_33</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref95">
        <label>95</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jiang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Schenck</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Kranz</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Banerjee</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Banerjee</surname>
              <given-names>NK</given-names>
            </name>
          </person-group>
          <article-title>CNN-based non-contact detection of food level in bottles from RGB images</article-title>
          <source>Proceedings of the 25th International Conference on MultiMedia Modeling</source>
          <year>2019</year>
          <conf-name>MMM 2019</conf-name>
          <conf-date>January 8-11, 2019</conf-date>
          <conf-loc>Thessaloniki, Greece</conf-loc>
          <pub-id pub-id-type="doi">10.1007/978-3-030-05710-7_17</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref96">
        <label>96</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lo</surname>
              <given-names>FP</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Qiu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Lo</surname>
              <given-names>BP</given-names>
            </name>
          </person-group>
          <article-title>Point2Volume: a vision-based dietary assessment approach using view synthesis</article-title>
          <source>IEEE Trans Ind Inform</source>
          <year>2020</year>
          <month>1</month>
          <volume>16</volume>
          <issue>1</issue>
          <fpage>577</fpage>
          <lpage>86</lpage>
          <pub-id pub-id-type="doi">10.1109/TII.2019.2942831</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref97">
        <label>97</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lo</surname>
              <given-names>FP</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Qiu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Lo</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>Food volume estimation based on deep learning view synthesis from a single depth map</article-title>
          <source>Nutrients</source>
          <year>2018</year>
          <month>12</month>
          <day>18</day>
          <volume>10</volume>
          <issue>12</issue>
          <fpage>2005</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mdpi.com/resolver?pii=nu10122005"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/nu10122005</pub-id>
          <pub-id pub-id-type="medline">30567362</pub-id>
          <pub-id pub-id-type="pii">nu10122005</pub-id>
          <pub-id pub-id-type="pmcid">PMC6316017</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref98">
        <label>98</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lo</surname>
              <given-names>FP</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Qiu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Lo</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>A novel vision-based approach for dietary assessment using deep learning view synthesis</article-title>
          <source>Proceedings of the IEEE 16th International Conference on Wearable and Implantable Body Sensor Networks</source>
          <year>2019</year>
          <conf-name>BSN 2019</conf-name>
          <conf-date>May 19-22, 2019</conf-date>
          <conf-loc>Chicago, IL</conf-loc>
          <pub-id pub-id-type="doi">10.1109/bsn.2019.8771089</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref99">
        <label>99</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Yu</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Cao</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Yuan</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Jia</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Mao</surname>
              <given-names>ZH</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Human-mimetic estimation of food volume from a single-view RGB image using an AI system</article-title>
          <source>Electronics (Basel)</source>
          <year>2021</year>
          <month>07</month>
          <day>28</day>
          <volume>10</volume>
          <issue>13</issue>
          <fpage>1556</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/34552763"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/electronics10131556</pub-id>
          <pub-id pub-id-type="medline">34552763</pub-id>
          <pub-id pub-id-type="pii">1556</pub-id>
          <pub-id pub-id-type="pmcid">PMC8455030</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref100">
        <label>100</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Miyazaki</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>de Silva</surname>
              <given-names>GC</given-names>
            </name>
            <name name-style="western">
              <surname>Aizawa</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Image-based calorie content estimation for dietary assessment</article-title>
          <source>Proceedings of the IEEE International Symposium on Multimedia</source>
          <year>2011</year>
          <conf-name>ISM 2011</conf-name>
          <conf-date>December 5-7, 2011</conf-date>
          <conf-loc>Dana Point, CA</conf-loc>
          <pub-id pub-id-type="doi">10.1109/ism.2011.66</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref101">
        <label>101</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ege</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Yanai</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Multi-task learning of dish detection and calorie estimation</article-title>
          <source>Proceedings of the Joint Workshop on Multimedia for Cooking and Eating Activities and Multimedia Assisted Dietary Management</source>
          <year>2018</year>
          <conf-name>CEA/MADiMa '18</conf-name>
          <conf-date>July 15, 2018</conf-date>
          <conf-loc>Stockholm, Sweden</conf-loc>
          <pub-id pub-id-type="doi">10.1145/3230519.3230594</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref102">
        <label>102</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ege</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Yanai</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Simultaneous estimation of dish locations and calories with multi-task learning</article-title>
          <source>IEICE Trans Inf Syst</source>
          <year>2019</year>
          <volume>E102.D</volume>
          <issue>7</issue>
          <fpage>1240</fpage>
          <lpage>6</lpage>
          <pub-id pub-id-type="doi">10.1587/transinf.2018cep0004</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref103">
        <label>103</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Allegra</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Anthimopoulos</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Stanco</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Farinella</surname>
              <given-names>GM</given-names>
            </name>
            <name name-style="western">
              <surname>Mougiakakou</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>A multi-task learning approach for meal assessment</article-title>
          <source>Proceedings of the Joint Workshop on Multimedia for Cooking and Eating Activities and Multimedia Assisted Dietary Management</source>
          <year>2018</year>
          <conf-name>CEA/MADiMa '18</conf-name>
          <conf-date>July 15, 2018</conf-date>
          <conf-loc>Stockholm, Sweden</conf-loc>
          <pub-id pub-id-type="doi">10.1145/3230519.3230593</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref104">
        <label>104</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>He</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Shao</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Wright</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Kerr</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Boushey</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>F</given-names>
            </name>
          </person-group>
          <article-title>Multi-task image-based dietary assessment for food recognition and portion size estimation</article-title>
          <source>Proceedings of the IEEE Conference on Multimedia Information Processing and Retrieval</source>
          <year>2020</year>
          <conf-name>MIPR 2020</conf-name>
          <conf-date>August 6-8, 2020</conf-date>
          <conf-loc>Shenzhen, China</conf-loc>
          <pub-id pub-id-type="doi">10.1109/mipr49039.2020.00018</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref105">
        <label>105</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Thames</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Karpur</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Norris</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Xia</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Panait</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Weyand</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Sim</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Nutrition5k: towards automatic nutritional understanding of generic food</article-title>
          <source>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source>
          <year>2021</year>
          <conf-name>CVPR 2021</conf-name>
          <conf-date>June 20-25, 2021</conf-date>
          <conf-loc>Nashville, TN</conf-loc>
          <pub-id pub-id-type="doi">10.1109/cvpr46437.2021.00879</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref106">
        <label>106</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ruede</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Heusser</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Frank</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Roitberg</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Haurilet</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Stiefelhagen</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Multi-task learning for calorie prediction on a novel large-scale recipe dataset enriched with nutritional information</article-title>
          <source>Proceedings of the 25th International Conference on Pattern Recognition</source>
          <year>2020</year>
          <conf-name>ICPR 2020</conf-name>
          <conf-date>January 10-15, 2021</conf-date>
          <conf-loc>Virtual Event</conf-loc>
          <pub-id pub-id-type="doi">10.1109/icpr48806.2021.9412839</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref107">
        <label>107</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Stathopoulou</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Mougiakakou</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Partially supervised multi-task network for single-view dietary assessment</article-title>
          <source>Proceedings of the 25th International Conference on Pattern Recognition</source>
          <year>2020</year>
          <conf-name>ICPR 2020</conf-name>
          <conf-date>July 10-15, 2021</conf-date>
          <conf-loc>Virtual Event</conf-loc>
          <pub-id pub-id-type="doi">10.1109/icpr48806.2021.9412339</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref108">
        <label>108</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>He</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Mao</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Shao</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Wright</surname>
              <given-names>JL</given-names>
            </name>
            <name name-style="western">
              <surname>Kerr</surname>
              <given-names>DA</given-names>
            </name>
            <name name-style="western">
              <surname>Boushey</surname>
              <given-names>CJ</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>F</given-names>
            </name>
          </person-group>
          <article-title>An end-to-end food image analysis system</article-title>
          <source>Electron Imaging</source>
          <year>2021</year>
          <month>1</month>
          <volume>33</volume>
          <fpage>285-1</fpage>
          <lpage>-7</lpage>
          <pub-id pub-id-type="doi">10.2352/issn.2470-1173.2021.8.imawm-285</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref109">
        <label>109</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Situju</surname>
              <given-names>SF</given-names>
            </name>
            <name name-style="western">
              <surname>Takimoto</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Sato</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Yamauchi</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Kanagawa</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Lawi</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Food constituent estimation for lifestyle disease prevention by multi-task CNN</article-title>
          <source>Appl Artif Intell</source>
          <year>2019</year>
          <month>04</month>
          <day>23</day>
          <volume>33</volume>
          <issue>8</issue>
          <fpage>732</fpage>
          <lpage>46</lpage>
          <pub-id pub-id-type="doi">10.1080/08839514.2019.1602318</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref110">
        <label>110</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Boushey</surname>
              <given-names>CJ</given-names>
            </name>
            <name name-style="western">
              <surname>Spoden</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Delp</surname>
              <given-names>EJ</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Bosch</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Ahmad</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Shvetsov</surname>
              <given-names>YB</given-names>
            </name>
            <name name-style="western">
              <surname>DeLany</surname>
              <given-names>JP</given-names>
            </name>
            <name name-style="western">
              <surname>Kerr</surname>
              <given-names>DA</given-names>
            </name>
          </person-group>
          <article-title>Reported energy intake accuracy compared to doubly labeled water and usability of the mobile food record among community dwelling adults</article-title>
          <source>Nutrients</source>
          <year>2017</year>
          <month>03</month>
          <day>22</day>
          <volume>9</volume>
          <issue>3</issue>
          <fpage>312</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mdpi.com/resolver?pii=nu9030312"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/nu9030312</pub-id>
          <pub-id pub-id-type="medline">28327502</pub-id>
          <pub-id pub-id-type="pii">nu9030312</pub-id>
          <pub-id pub-id-type="pmcid">PMC5372975</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref111">
        <label>111</label>
        <nlm-citation citation-type="web">
          <source>Open Food Facts</source>
          <access-date>2024-06-13</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://world.openfoodfacts.org">https://world.openfoodfacts.org</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref112">
        <label>112</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Partridge</surname>
              <given-names>EK</given-names>
            </name>
            <name name-style="western">
              <surname>Neuhouser</surname>
              <given-names>ML</given-names>
            </name>
            <name name-style="western">
              <surname>Breymeyer</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Schenk</surname>
              <given-names>JM</given-names>
            </name>
          </person-group>
          <article-title>Comparison of nutrient estimates based on food volume versus weight: implications for dietary assessment methods</article-title>
          <source>Nutrients</source>
          <year>2018</year>
          <month>07</month>
          <day>27</day>
          <volume>10</volume>
          <issue>8</issue>
          <fpage>973</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mdpi.com/resolver?pii=nu10080973"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/nu10080973</pub-id>
          <pub-id pub-id-type="medline">30060455</pub-id>
          <pub-id pub-id-type="pii">nu10080973</pub-id>
          <pub-id pub-id-type="pmcid">PMC6115952</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref113">
        <label>113</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Conway</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Robertson</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Dennis</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Stamler</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Elliott</surname>
              <given-names>P</given-names>
            </name>
            <collab>INTERMAP Research Group</collab>
          </person-group>
          <article-title>Standardised coding of diet records: experiences from INTERMAP UK</article-title>
          <source>Br J Nutr</source>
          <year>2004</year>
          <month>05</month>
          <volume>91</volume>
          <issue>5</issue>
          <fpage>765</fpage>
          <lpage>71</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/15152639"/>
          </comment>
          <pub-id pub-id-type="doi">10.1079/BJN20041095</pub-id>
          <pub-id pub-id-type="medline">15152639</pub-id>
          <pub-id pub-id-type="pii">S0007114504000959</pub-id>
          <pub-id pub-id-type="pmcid">PMC6660142</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref114">
        <label>114</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Abdul</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Vermeulen</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Lim</surname>
              <given-names>BY</given-names>
            </name>
            <name name-style="western">
              <surname>Kankanhalli</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Trends and trajectories for explainable, accountable and intelligible systems: an HCI research agenda</article-title>
          <source>Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems</source>
          <year>2018</year>
          <conf-name>CHI '18</conf-name>
          <conf-date>April 21-26, 2018</conf-date>
          <conf-loc>Montreal, QC</conf-loc>
          <pub-id pub-id-type="doi">10.1145/3173574.3174156</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref115">
        <label>115</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Amann</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Blasimme</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Vayena</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Frey</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Madai</surname>
              <given-names>VI</given-names>
            </name>
            <collab>Precise4Q consortium</collab>
          </person-group>
          <article-title>Explainability for artificial intelligence in healthcare: a multidisciplinary perspective</article-title>
          <source>BMC Med Inform Decis Mak</source>
          <year>2020</year>
          <month>11</month>
          <day>30</day>
          <volume>20</volume>
          <issue>1</issue>
          <fpage>310</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://bmcmedinformdecismak.biomedcentral.com/articles/10.1186/s12911-020-01332-6"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s12911-020-01332-6</pub-id>
          <pub-id pub-id-type="medline">33256715</pub-id>
          <pub-id pub-id-type="pii">10.1186/s12911-020-01332-6</pub-id>
          <pub-id pub-id-type="pmcid">PMC7706019</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref116">
        <label>116</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chang</surname>
              <given-names>EM</given-names>
            </name>
            <name name-style="western">
              <surname>Saigal</surname>
              <given-names>CS</given-names>
            </name>
            <name name-style="western">
              <surname>Raldow</surname>
              <given-names>AC</given-names>
            </name>
          </person-group>
          <article-title>Explaining health state utility assessment</article-title>
          <source>JAMA</source>
          <year>2020</year>
          <month>03</month>
          <day>17</day>
          <volume>323</volume>
          <issue>11</issue>
          <fpage>1085</fpage>
          <lpage>6</lpage>
          <pub-id pub-id-type="doi">10.1001/jama.2020.0656</pub-id>
          <pub-id pub-id-type="medline">32091541</pub-id>
          <pub-id pub-id-type="pii">2762126</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref117">
        <label>117</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Gemming</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Utter</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Ni Mhurchu</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Image-assisted dietary assessment: a systematic review of the evidence</article-title>
          <source>J Acad Nutr Diet</source>
          <year>2015</year>
          <month>01</month>
          <volume>115</volume>
          <issue>1</issue>
          <fpage>64</fpage>
          <lpage>77</lpage>
          <pub-id pub-id-type="doi">10.1016/j.jand.2014.09.015</pub-id>
          <pub-id pub-id-type="medline">25441955</pub-id>
          <pub-id pub-id-type="pii">S2212-2672(14)01469-5</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref118">
        <label>118</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Doulah</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>McCrory</surname>
              <given-names>MA</given-names>
            </name>
            <name name-style="western">
              <surname>Higgins</surname>
              <given-names>JA</given-names>
            </name>
            <name name-style="western">
              <surname>Sazonov</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>A systematic review of technology-driven methodologies for estimation of energy intake</article-title>
          <source>IEEE Access</source>
          <year>2019</year>
          <volume>7</volume>
          <fpage>49653</fpage>
          <lpage>68</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/32489752"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/access.2019.2910308</pub-id>
          <pub-id pub-id-type="medline">32489752</pub-id>
          <pub-id pub-id-type="pmcid">PMC7266287</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref119">
        <label>119</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lo</surname>
              <given-names>FP</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Qiu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Lo</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>Image-based food classification and volume estimation for dietary assessment: a review</article-title>
          <source>IEEE J Biomed Health Inform</source>
          <year>2020</year>
          <month>7</month>
          <volume>24</volume>
          <issue>7</issue>
          <fpage>1926</fpage>
          <lpage>39</lpage>
          <pub-id pub-id-type="doi">10.1109/jbhi.2020.2987943</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref120">
        <label>120</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Subhi</surname>
              <given-names>MA</given-names>
            </name>
            <name name-style="western">
              <surname>Ali</surname>
              <given-names>SH</given-names>
            </name>
            <name name-style="western">
              <surname>Mohammed</surname>
              <given-names>MA</given-names>
            </name>
          </person-group>
          <article-title>Vision-based approaches for automatic food recognition and dietary assessment: a survey</article-title>
          <source>IEEE Access</source>
          <year>2019</year>
          <month>03</month>
          <day>13</day>
          <volume>7</volume>
          <fpage>35370</fpage>
          <lpage>81</lpage>
          <pub-id pub-id-type="doi">10.1109/access.2019.2904519</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref121">
        <label>121</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dalakleidi</surname>
              <given-names>KV</given-names>
            </name>
            <name name-style="western">
              <surname>Papadelli</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Kapolos</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Papadimitriou</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Applying image-based food-recognition systems on dietary assessment: a systematic review</article-title>
          <source>Adv Nutr</source>
          <year>2022</year>
          <month>12</month>
          <day>22</day>
          <volume>13</volume>
          <issue>6</issue>
          <fpage>2590</fpage>
          <lpage>619</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S2161-8313(23)00093-5"/>
          </comment>
          <pub-id pub-id-type="doi">10.1093/advances/nmac078</pub-id>
          <pub-id pub-id-type="medline">35803496</pub-id>
          <pub-id pub-id-type="pii">S2161-8313(23)00093-5</pub-id>
          <pub-id pub-id-type="pmcid">PMC9776640</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref122">
        <label>122</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tay</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Kaur</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Quek</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Lim</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Henry</surname>
              <given-names>CJ</given-names>
            </name>
          </person-group>
          <article-title>Current developments in digital quantitative volume estimation for the optimisation of dietary assessment</article-title>
          <source>Nutrients</source>
          <year>2020</year>
          <month>04</month>
          <day>22</day>
          <volume>12</volume>
          <issue>4</issue>
          <fpage>1167</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mdpi.com/resolver?pii=nu12041167"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/nu12041167</pub-id>
          <pub-id pub-id-type="medline">32331262</pub-id>
          <pub-id pub-id-type="pii">nu12041167</pub-id>
          <pub-id pub-id-type="pmcid">PMC7231293</pub-id>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
