Section 01
[Introduction] FoodSense Project: A Breakthrough in Enabling AI to Perceive Multisensory Experiences from Food Images
The FoodSense project aims to address the gap in AI cross-sensory reasoning. It has built a dataset with annotations from 66,842 participants, covering 2,987 food images, supporting the prediction of taste, smell, texture, and sound from visuals, and generating explainable reasoning. The trained FoodSense-VL model advances food image understanding from surface-level recognition to multisensory perception, bridging cognitive science and AI.