Integrated Vocal Deviation Index (IVDI): a Machine Learning Model to Classifier of the General Grade of Vocal Deviation
OBJECTIVE: To develop a multiparametric index based on machine learning (ML) to predict and classify the overall degree of vocal deviation (GG). METHOD: The sample consisted of 300 dysphonic and non-dysphonic participants of both sexes. Two speech tasks were: sustained vowel [a] and connected speech (counting numbers from 1 to 10). Five speech-language pathologists (SLP) performed the perceptual-auditory judgment (APJ) of the GG and the degrees of roughness (GR), breathiness (GB), instability (GI), and strain (GS). We extracted 47 acoustic measurements from these tasks. The APJ result and the acoustic measurements were used to develop the multiparametric index. We used mean absolute error (MAE), root mean square error (RMSE), and coefficient of determination (R_) to select the best model of ML to predict GG, and feature importance to select the best set of variables for the index. After classifying the GG between non-dysphonic, mild, moderate and severe, the final model was validated using accuracy, sensitivity, specificity, predictive values, likelihood ratios, F1-Score, and weighted kappa. RESULTS: The gradient boost model showed the best performance among the ML models. Eight features were selected in the model, including four acoustic measures (jitterLoc, CPPS, HNRmean, and Correlation) and four APJ measures (GR, GB, GS, GI). The final model correctly classified 93.75% of participants and obtained a weighted kappa index of 0.9374, demonstrating the model's excellent performance. CONCLUSION: The Integrated Vocal Deviation Index (IVDI) includes four acoustic measures and four perceptual-auditory measures, and showed excellent performance in classifying voices according to GG.