<?xml version="1.0"?>
<metadata xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:dc="http://purl.org/dc/elements/1.1/"><dc:title>Large language models in food and nutrition science</dc:title><dc:creator>Gjorgjevikj,	Ana	(Avtor)
	</dc:creator><dc:creator>Martinc,	Matej	(Avtor)
	</dc:creator><dc:creator>Cenikj,	Gjorgjina	(Avtor)
	</dc:creator><dc:creator>Drole,	Jan	(Avtor)
	</dc:creator><dc:creator>Ogrinc,	Nives	(Avtor)
	</dc:creator><dc:creator>Džeroski,	Sašo	(Avtor)
	</dc:creator><dc:creator>Koroušić-Seljak,	Barbara	(Avtor)
	</dc:creator><dc:creator>Eftimov,	Tome	(Avtor)
	</dc:creator><dc:subject>FoodyLLM</dc:subject><dc:subject>nutrient estimation</dc:subject><dc:subject>data interoperability</dc:subject><dc:description>Background Reliable nutrient profiling and semantic interoperability are essential for scalable dietary assessment, food labeling (e.g., traffic-light schemes), and FAIR integration of food composition and consumption data. However, general-purpose large language models (LLMs) are not systematically exposed to structured recipe–nutrition mappings and food ontologies, limiting their accuracy and trustworthiness in food and nutrition tasks. Scope and approach We review recent LLM advances in life sciences and healthcare and analyze the gap in food and nutrition applications. To address this gap, we introduce FoodyLLM, a domain-specialized LLM fine-tuned on 225k task-aligned QA pairs for (i) recipe nutrient estimation, (ii) traffic-light classification, and (iii) ontology-based entity linking to support FAIR food data interoperability. We benchmark FoodyLLM against strong general-purpose baselines (e.g., Llama 3 8B, Gemini 2.0) under zero-/few-shot prompting across five evaluation folds. Key findings Across all tasks, FoodyLLM substantially outperforms general-purpose LLMs for nutrient estimation across all macronutrients (fat, protein, salt, saturates, sugar), accuracy increases from 0.43 to 0.63 to 0.91–0.97; for traffic-light classification across all nutrients and color categories, macro F1 improves from 0.46 to 0.80 to 0.86–0.97; and for ontology-based food entity linking across FoodOn, SNOMED-CT, and Hansard, macro F1 increases from 0.33 to 0.44 (best general-purpose baseline) to 0.93–0.98 on artificial NEL data, and from 0.24 to 0.51 to 0.67–0.84 on real corpora (CafeteriaSA and CafeteriaFCD). Overall, our results demonstrate the practical value of domain-specialized LLMs in food and nutrition research. They enable automated dietary assessment, large-scale nutritional monitoring, and FAIR data integration, while opening new pathways toward sustainable and personalized nutrition.</dc:description><dc:publisher>Elsevier</dc:publisher><dc:date>2026</dc:date><dc:date>2026-03-04 12:39:51</dc:date><dc:type>Neznano</dc:type><dc:identifier>27977</dc:identifier><dc:identifier>UDK: 004.8</dc:identifier><dc:identifier>ISSN pri članku: 2665-9271</dc:identifier><dc:identifier>DOI: 10.1016/j.crfs.2026.101351</dc:identifier><dc:identifier>COBISS_ID: 270414595</dc:identifier><dc:source>Nizozemska</dc:source><dc:language>sl</dc:language><dc:rights>© 2026 The Authors. </dc:rights></metadata>
