TY - JOUR
T1 - Multi-modal convolutional parameterisation network for guided image inverse problems
AU - Czerkawski, Mikolaj
AU - Upadhyay, Priti
AU - Davison, Christopher
AU - Atkinson, Robert
AU - Michie, Craig
AU - Andonovic, Ivan
AU - Macdonald, Malcolm
AU - Cardona, Javier
AU - Tachtatzis, Christos
PY - 2024/3/12
Y1 - 2024/3/12
N2 - There are several image inverse tasks, such as inpainting or super-resolution, which can be solved using deep internal learning, a paradigm that involves employing deep neural networks to find a solution by learning from the sample itself rather than a dataset. For example, Deep Image Prior is a technique based on fitting a convolutional neural network to output the known parts of the image (such as non-inpainted regions or a low-resolution version of the image). However, this approach is not well adjusted for samples composed of multiple modalities. In some domains, such as satellite image processing, accommodating multi-modal representations could be beneficial or even essential. In this work, Multi-Modal Convolutional Parameterisation Network (MCPN) is proposed, where a convolutional neural network approximates shared information between multiple modes by combining a core shared network with modality-specific head networks. The results demonstrate that these approaches can significantly outperform the single-mode adoption of a convolutional parameterisation network on guided image inverse problems of inpainting and super-resolution.
AB - There are several image inverse tasks, such as inpainting or super-resolution, which can be solved using deep internal learning, a paradigm that involves employing deep neural networks to find a solution by learning from the sample itself rather than a dataset. For example, Deep Image Prior is a technique based on fitting a convolutional neural network to output the known parts of the image (such as non-inpainted regions or a low-resolution version of the image). However, this approach is not well adjusted for samples composed of multiple modalities. In some domains, such as satellite image processing, accommodating multi-modal representations could be beneficial or even essential. In this work, Multi-Modal Convolutional Parameterisation Network (MCPN) is proposed, where a convolutional neural network approximates shared information between multiple modes by combining a core shared network with modality-specific head networks. The results demonstrate that these approaches can significantly outperform the single-mode adoption of a convolutional parameterisation network on guided image inverse problems of inpainting and super-resolution.
KW - image synthesis
KW - internal learning
KW - image inpainting
KW - image super-resolution
KW - multi-modal learning
U2 - 10.3390/jimaging10030069
DO - 10.3390/jimaging10030069
M3 - Article
SN - 2313-433X
VL - 10
JO - Journal of Imaging
JF - Journal of Imaging
IS - 3
M1 - 69
ER -