Automated Medical Coding on MIMIC-III and MIMIC-IV: A Critical Review and Replicability Study

Medical coding is the task of assigning medical codes to clinical free-text documentation. Healthcare professionals manually assign such codes to track patient diagnoses and treatments. Automated medical coding can considerably alleviate this administrative burden. In this paper, we reproduce, compare, and analyze state-of-the-art automated medical coding machine learning models. We show that several models underperform due to weak configurations, poorly sampled train-test splits, and insufficient evaluation. In previous work, the macro F1 score has been calculated sub-optimally, and our correction doubles it. We contribute a revised model comparison using stratified sampling and identical experimental setups, including hyperparameters and decision boundary tuning. We analyze prediction errors to validate and falsify assumptions of previous works. The analysis confirms that all models struggle with rare codes, while long documents only have a negligible impact. Finally, we present the first comprehensive results on the newly released MIMIC-IV dataset using the reproduced models. We release our code, model parameters, and new MIMIC-III and MIMIC-IV training and evaluation pipelines to accommodate fair future comparisons.

PDF Abstract

Datasets


Introduced in the Paper:

MIMIC-IV ICD-10 MIMIC-IV ICD-9

Used in the Paper:

MIMIC-III
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Medical Code Prediction MIMIC-IV ICD-10 PLM-ICD F1 Micro 58.5 # 1
F1 Macro 21.1 # 1
Exact Match Ratio 0.4 # 1
Precision@8 69.9 # 1
Precision@15 55.0 # 1
R-Prec 57.9 # 1
mAP 61.9 # 1
AUC Micro 99.2 # 1
AUC Macro 96.6 # 1
Medical Code Prediction MIMIC-IV ICD-10 LAAT F1 Micro 57.9 # 2
F1 Macro 20.3 # 3
Exact Match Ratio 0.4 # 1
Precision@8 68.9 # 2
Precision@15 54.3 # 2
R-Prec 57.2 # 2
mAP 60.6 # 2
AUC Micro 99.0 # 2
AUC Macro 95.4 # 2
Medical Code Prediction MIMIC-IV ICD-10 MultiResCNN F1 Micro 56.9 # 3
F1 Macro 21.1 # 1
Exact Match Ratio 0.4 # 1
Precision@8 67.8 # 3
Precision@15 53.5 # 3
R-Prec 56.1 # 3
mAP 59.3 # 3
AUC Micro 99.0 # 2
AUC Macro 94.5 # 3
Medical Code Prediction MIMIC-IV ICD-10 CAML F1 Micro 55.4 # 4
F1 Macro 16.0 # 4
Exact Match Ratio 0.3 # 4
Precision@8 66.8 # 4
Precision@15 52.2 # 4
R-Prec 54.5 # 4
mAP 57.4 # 4
AUC Micro 98.5 # 4
AUC Macro 91.1 # 5
Medical Code Prediction MIMIC-IV ICD-10 Bi-GRU F1 Micro 50.1 # 5
F1 Macro 10.6 # 5
Exact Match Ratio 0.3 # 4
Precision@8 62.6 # 5
Precision@15 47.7 # 5
R-Prec 49.6 # 5
mAP 51.1 # 5
AUC Micro 98.3 # 5
AUC Macro 92.4 # 4
Medical Code Prediction MIMIC-IV ICD-10 CNN F1 Micro 47.2 # 6
F1 Macro 8.0 # 6
Exact Match Ratio 0.3 # 4
Precision@8 60.3 # 6
Precision@15 45.7 # 6
R-Prec 47.3 # 6
mAP 48.2 # 6
AUC Micro 97.5 # 6
AUC Macro 87.9 # 6
Medical Code Prediction MIMIC-IV ICD-9 PLM-ICD F1 Micro 62.6 # 1
F1 Macro 29.8 # 1
Exact Match Ratio 1.0 # 1
Precision@8 70.0 # 1
Precision@15 53.5 # 1
R-Prec 62.7 # 1
mAP 68.0 # 1
AUC Micro 99.4 # 1
AUC Macro 97.2 # 1
Medical Code Prediction MIMIC-IV ICD-9 LAAT F1 Micro 61.7 # 2
F1 Macro 26.4 # 3
Exact Match Ratio 0.9 # 2
Precision@8 68.9 # 2
Precision@15 52.7 # 2
R-Prec 61.7 # 2
mAP 66.3 # 2
AUC Micro 99.3 # 2
AUC Macro 96.0 # 2
Medical Code Prediction MIMIC-IV ICD-9 MultiResCNN F1 Micro 60.4 # 3
F1 Macro 27.7 # 2
Exact Match Ratio 0.8 # 3
Precision@8 67.6 # 3
Precision@15 51.8 # 3
R-Prec 60.4 # 3
mAP 64.7 # 3
AUC Micro 99.2 # 3
AUC Macro 95.1 # 3
Medical Code Prediction MIMIC-IV ICD-9 CAML F1 Micro 58.6 # 4
F1 Macro 19.3 # 4
Exact Match Ratio 0.6 # 5
Precision@8 66.3 # 4
Precision@15 50.3 # 4
R-Prec 58.5 # 4
mAP 62.4 # 4
AUC Micro 98.8 # 4
AUC Macro 90.7 # 5
Medical Code Prediction MIMIC-IV ICD-9 Bi-GRU F1 Micro 55.5 # 5
F1 Macro 16.6 # 5
Exact Match Ratio 0.7 # 4
Precision@8 64.1 # 5
Precision@15 47.8 # 5
R-Prec 55.8 # 5
mAP 58.9 # 5
AUC Micro 98.8 # 4
AUC Macro 93.8 # 4
Medical Code Prediction MIMIC-IV ICD-9 CNN F1 Micro 52.4 # 6
F1 Macro 12.6 # 6
Exact Match Ratio 0.6 # 5
Precision@8 61.3 # 6
Precision@15 45.6 # 6
R-Prec 52.9 # 6
mAP 55.2 # 6
AUC Micro 98.1 # 6
AUC Macro 89.4 # 6

Methods


No methods listed for this paper. Add relevant methods here