Section 01
Introduction to Clinical Application of Multimodal Deep Learning in Speech Assessment for Cleft Lip and Palate Patients
This study integrates audio, facial video, fluoroscopic imaging, and clinical variables to construct a multimodal deep learning model, enabling automated detection of compensatory articulation and hypernasality in cleft lip and palate patients. It aims to provide an objective auxiliary tool for clinical speech assessment and address the limitations of traditional subjective evaluation.