|国家预印本平台
首页|Multimodal Fusion of Glucose Monitoring and Food Imagery for Caloric Content Prediction

Multimodal Fusion of Glucose Monitoring and Food Imagery for Caloric Content Prediction

Multimodal Fusion of Glucose Monitoring and Food Imagery for Caloric Content Prediction

来源:Arxiv_logoArxiv
英文摘要

Effective dietary monitoring is critical for managing Type 2 diabetes, yet accurately estimating caloric intake remains a major challenge. While continuous glucose monitors (CGMs) offer valuable physiological data, they often fall short in capturing the full nutritional profile of meals due to inter-individual and meal-specific variability. In this work, we introduce a multimodal deep learning framework that jointly leverages CGM time-series data, Demographic/Microbiome, and pre-meal food images to enhance caloric estimation. Our model utilizes attention based encoding and a convolutional feature extraction for meal imagery, multi-layer perceptrons for CGM and Microbiome data followed by a late fusion strategy for joint reasoning. We evaluate our approach on a curated dataset of over 40 participants, incorporating synchronized CGM, Demographic and Microbiome data and meal photographs with standardized caloric labels. Our model achieves a Root Mean Squared Relative Error (RMSRE) of 0.2544, outperforming the baselines models by over 50%. These findings demonstrate the potential of multimodal sensing to improve automated dietary assessment tools for chronic disease management.

Adarsh Kumar

医学研究方法计算技术、计算机技术

Adarsh Kumar.Multimodal Fusion of Glucose Monitoring and Food Imagery for Caloric Content Prediction[EB/OL].(2025-05-13)[2025-06-14].https://arxiv.org/abs/2505.09018.点此复制

评论