Deep learning models have been proved to be promising and efficient lately on image parsing tasks. However, deep learning models are not fully capable of incorporating visual and contextual information simultaneously. We propose a new three-layer context-based deep architecture to integrate context explicitly with visual information. The novel idea here is to have a visual layer to learn visual characteristics from binary class-based learners, a contextual layer to learn context, and then an integration layer to learn from both via genetic algorithm-based optimal fusion to produce a final decision. The experimental outcomes when evaluated on benchmark datasets show our approach outperforms existing baseline approaches. Further analysis shows that optimized network weights can improve performance and make stable predictions.
CITATION STYLE
Mandal, R., Azam, B., & Verma, B. (2021). Context-Based Deep Learning Architecture with Optimal Integration Layer for Image Parsing. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 13109 LNCS, pp. 285–296). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-030-92270-2_25
Mendeley helps you to discover research relevant for your work.