TY - UNPB
T1 - Neural knitworks
T2 - patched neural implicit representation networks
AU - Czerkawski, Mikolaj
AU - Cardona, Javier
AU - Atkinson, Robert
AU - Michie, Craig
AU - Andonovic, Ivan
AU - Clemente, Carmine
AU - Tachtatzis, Christos
PY - 2021/9/29
Y1 - 2021/9/29
N2 - Coordinate-based Multilayer Perceptron (MLP) networks, despite being capable of learning neural implicit representations, are not performant for internal image synthesis applications. Convolutional Neural Networks (CNNs) are typically used instead for a variety of internal generative tasks, at the cost of a larger model. We propose Neural Knitwork, an architecture for neural implicit representation learning of natural images that achieves image synthesis by optimizing the distribution of image patches in an adversarial manner and by enforcing consistency between the patch predictions. To the best of our knowledge, this is the first implementation of a coordinate-based MLP tailored for synthesis tasks such as image inpainting, super-resolution, and denoising. We demonstrate the utility of the proposed technique by training on these three tasks. The results show that modeling natural images using patches, rather than pixels, produces results of higher fidelity. The resulting model requires 80% fewer parameters than alternative CNN-based solutions while achieving comparable performance and training time.
AB - Coordinate-based Multilayer Perceptron (MLP) networks, despite being capable of learning neural implicit representations, are not performant for internal image synthesis applications. Convolutional Neural Networks (CNNs) are typically used instead for a variety of internal generative tasks, at the cost of a larger model. We propose Neural Knitwork, an architecture for neural implicit representation learning of natural images that achieves image synthesis by optimizing the distribution of image patches in an adversarial manner and by enforcing consistency between the patch predictions. To the best of our knowledge, this is the first implementation of a coordinate-based MLP tailored for synthesis tasks such as image inpainting, super-resolution, and denoising. We demonstrate the utility of the proposed technique by training on these three tasks. The results show that modeling natural images using patches, rather than pixels, produces results of higher fidelity. The resulting model requires 80% fewer parameters than alternative CNN-based solutions while achieving comparable performance and training time.
KW - multilayer perceptron (MLP)
KW - convolutional neural network (CNN)
KW - neural knitwork
U2 - 10.48550/arXiv.2109.14406
DO - 10.48550/arXiv.2109.14406
M3 - Working Paper/Preprint
BT - Neural knitworks
CY - Ithaca, N.Y.
ER -