Author: Yip, S. S. F.; Klanecek, Z.; Naganawa, S.; Kim, J.; Studen, A.; Rivetti, L.; Jeraj, R.
Title: Performance and Robustness of Machine Learning-based Radiomic COVID-19 Severity Prediction Cord-id: fr96tbef Document date: 2020_9_9
ID: fr96tbef
Snippet: Objectives: This study investigated the performance and robustness of radiomics in predicting COVID-19 severity in a large public cohort. Methods: A public dataset of 1110 COVID-19 patients (1 CT/patient) was used. Using CTs and clinical data, each patient was classified into mild, moderate, and severe by two observers: (1) dataset provider and (2) a board-certified radiologist. For each CT, 107 radiomic features were extracted. The dataset was randomly divided into a training (60%) and holdout
Document: Objectives: This study investigated the performance and robustness of radiomics in predicting COVID-19 severity in a large public cohort. Methods: A public dataset of 1110 COVID-19 patients (1 CT/patient) was used. Using CTs and clinical data, each patient was classified into mild, moderate, and severe by two observers: (1) dataset provider and (2) a board-certified radiologist. For each CT, 107 radiomic features were extracted. The dataset was randomly divided into a training (60%) and holdout validation (40%) set. During training, features were selected and combined into a logistic regression model for predicting severe cases from mild and moderate cases. The models were trained and validated on the classifications by both observers. AUC quantified the predictive power of models. To determine model robustness, the trained models was cross-validated on the inter-observer classifications. Results: A single feature alone was sufficient to predict mild from severe COVID-19 with AUC_valid^provider=0.85 and AUC_valid^radiologist=0.74 (p<<0.01). The most predictive features were the distribution of small size-zones (GLSZM-SmallAreaEmphasis) for provider classification and linear dependency of neighboring voxels (GLCM-Correlation) for radiologist classification. Cross-validation showed that both AUC_valid^ {approx}0.80 (p<<0.01). In predicting moderate from severe COVID-19, first-order-Median alone had sufficient predictive power of AUC_valid^provider=0.65 (p=0.01). For radiologist classification, the predictive power of the model increased to AUC_valid^radiologist=0.66 (p<<0.01) as the number of features grew from 1 to 5. Cross-validation yielded AUC_valid^radiologist=0.63 (p=0.002) and AUC_valid^provider=0.60 (p=0.09). Conclusions: Radiomics significantly predicted different levels of COVID-19 severity. The prediction was moderately sensitive to inter-observer classifications, and thus need to be used with caution.
Search related documents:
Co phrase search for related documents- logistic regression and lung radiomic: 1, 2, 3
- logistic regression and lung segmentation: 1, 2, 3, 4
- logistic regression and lung tissue: 1, 2, 3, 4, 5, 6, 7
- logistic regression and lung window: 1, 2
- logistic regression and machine learning: 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74
- logistic regression and machine learning model: 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48
- logistic regression model and lung involvement: 1, 2, 3, 4, 5, 6
- logistic regression model and lung radiomic: 1
- logistic regression model and lung segmentation: 1
- logistic regression model and lung window: 1
- logistic regression model and machine learning: 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55
- logistic regression model and machine learning model: 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18
- lung appearance and machine learning: 1, 2
- lung involvement and machine learning: 1, 2, 3, 4, 5, 6
- lung involvement and machine learning model: 1, 2, 3
- lung radiomic and machine learning: 1, 2
- lung radiomic and machine learning model: 1
- lung tissue and machine learning: 1, 2, 3, 4, 5, 6
- lung tissue and machine learning model: 1
Co phrase search for related documents, hyperlinks ordered by date