Optimizing Body Region Classification with Deep Convolutional Activation Features
Abstract
The goal of this work is to automatically apply generated image keywords as text representations, to optimize medical image classification accuracies of body regions. To create a keyword generative model, a Long Short-Term Memory (LSTM) based Recurrent Neural Network (RNN) is adopted, which is trained with preprocessed biomedical image captions as text representation and visual features extracted using Convolutional Neural Networks (CNN). For image representation, deep convolutional activation features and Bag-of-Keypoints (BoK) features are extracted for each radiograph and combined with the automatically generated keywords. Random Forest models and Support Vector Machines are trained with these multimodal image representations, as well as just visual representation, to predict body regions. Adopting multimodal image features proves to be the better approach, as the prediction accuracy for body regions is increased.
Cite
Text
Pelka et al. "Optimizing Body Region Classification with Deep Convolutional Activation Features." European Conference on Computer Vision Workshops, 2018. doi:10.1007/978-3-030-11018-5_60Markdown
[Pelka et al. "Optimizing Body Region Classification with Deep Convolutional Activation Features." European Conference on Computer Vision Workshops, 2018.](https://mlanthology.org/eccvw/2018/pelka2018eccvw-optimizing/) doi:10.1007/978-3-030-11018-5_60BibTeX
@inproceedings{pelka2018eccvw-optimizing,
title = {{Optimizing Body Region Classification with Deep Convolutional Activation Features}},
author = {Pelka, Obioma and Nensa, Felix and Friedrich, Christoph M.},
booktitle = {European Conference on Computer Vision Workshops},
year = {2018},
pages = {699-704},
doi = {10.1007/978-3-030-11018-5_60},
url = {https://mlanthology.org/eccvw/2018/pelka2018eccvw-optimizing/}
}