Multi-Label Classification
-
Medical Code Prediction from Discharge Summary: Document to Sequence BERT using Sequence Attention
Tak-Sung Heo
,Yongmin Yoo
,Yeongjoon Park
,Byeong-Cheol Jo
,Kyounguk Lee
,Kyungsun Kim
-
-
Text - noteevents
-
Label - diagnoses_icd, procedure_icd
-
-
Document to Sequence Preprocessing
-
Collecting the CLS_token extracted through Document-to-Sequence BERT (D2SBERT)
- We used BioBERT
-
Sequence Attention
-
Classifier
Model | F1-Macro | F1-Micro |
---|---|---|
CAML | 0.56924 | 0.64993 |
SWAM | 0.58025 | 0.65994 |
EnCAML | 0.59653 | 0.66594 |
BERT-head | 0.49376 | 0.56627 |
BERT-tail | 0.45453 | 0.54011 |
BERT-head-tail | 0.49362 | 0.56566 |
Proposed model | 0.62898 | 0.68555 |