Abstract
We target the automatic classification of fractures from clinical X-Ray images following the Arbeitsgemeinschaft Osteosynthese (AO) classification standard. We decompose the problem into the localization of the region-of-interest (ROI) and the classification of the localized region. Our solution relies on current advances in multi-task end-to-end deep learning. More specifically, we adapt an attention model known as Spatial Transformer (ST) to learn an image-dependent localization of the ROI trained only from image classification labels. As a case study, we focus here on the classification of proximal femur fractures. We provide a detailed quantitative and qualitative validation on a dataset of 1000 images and report high accuracy with regard to inter-expert correlation values reported in the literature.
| Original language | English |
|---|---|
| Title of host publication | MLMI 2017: Machine Learning in Medical Imaging |
| Volume | 10541 |
| Publication date | 7 Sept 2017 |
| ISBN (Print) | 978-3-319-67388-2 |
| ISBN (Electronic) | 978-3-319-67389-9 |
| DOIs | |
| Publication status | Published - 7 Sept 2017 |
| Externally published | Yes |
Keywords
- label image classification
- spatial transformer network
- classification model parameters
- trauma surgery department
- adaptive network architecture
Fingerprint
Dive into the research topics of 'Automatic Classification of Proximal Femur Fractures Based on Attention Models'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver