Position-dependent partial convolutions for supervised spatial interpolation

Hirotaka Hachiya, Kotaro Nagayoshi, Asako Iwaki, Takahiro Maeda, Naonori Ueda, Hiroyuki Fujiwara
Proceedings of The 14th Asian Conference on Machine Learning, PMLR 189:420-435, 2023.

Abstract

Acquiring continuous spatial data, e.g., spatial ground motion is essential to assess the damaged area and appropriately assign rescue and medical teams. To this purpose, spatial interpolation methods have been developed to estimate the value of unobserved points linearly from neighbor observed values, i.e., inverse distance weighting and Kriging. Recently, realistic spatial continuous environmental data with various scenarios can be generated by 3D finite difference methods with a high-resolution structure model. It enables us to collect supervised data even for unobserved points. Along this line, we propose a framework of supervised spatial interpolation and apply highly advanced deep inpainting methods where we treat spatially distributed observed points as a masked image and non-linearly expand them through convolutional encoder-decoder networks. However, the property of translation invariance would avoid locally fine-grained interpolation since the relation between the target and surrounding observation points varies over regions due to its topography and subsurface structure. To overcome this problem, we propose introducing position-dependent convolution where kernel weights are adjusted depending on their position on an image based on the trainable position-feature map. We show the effectiveness of our proposed method, called, PoDIM (Position-dependent Deep Inpainting Method), through experiments using simulated ground-motion data.

Cite this Paper


BibTeX
@InProceedings{pmlr-v189-hachiya23a, title = {Position-dependent partial convolutions for supervised spatial interpolation}, author = {Hachiya, Hirotaka and Nagayoshi, Kotaro and Iwaki, Asako and Maeda, Takahiro and Ueda, Naonori and Fujiwara, Hiroyuki}, booktitle = {Proceedings of The 14th Asian Conference on Machine Learning}, pages = {420--435}, year = {2023}, editor = {Khan, Emtiyaz and Gonen, Mehmet}, volume = {189}, series = {Proceedings of Machine Learning Research}, month = {12--14 Dec}, publisher = {PMLR}, pdf = {https://proceedings.mlr.press/v189/hachiya23a/hachiya23a.pdf}, url = {https://proceedings.mlr.press/v189/hachiya23a.html}, abstract = {Acquiring continuous spatial data, e.g., spatial ground motion is essential to assess the damaged area and appropriately assign rescue and medical teams. To this purpose, spatial interpolation methods have been developed to estimate the value of unobserved points linearly from neighbor observed values, i.e., inverse distance weighting and Kriging. Recently, realistic spatial continuous environmental data with various scenarios can be generated by 3D finite difference methods with a high-resolution structure model. It enables us to collect supervised data even for unobserved points. Along this line, we propose a framework of supervised spatial interpolation and apply highly advanced deep inpainting methods where we treat spatially distributed observed points as a masked image and non-linearly expand them through convolutional encoder-decoder networks. However, the property of translation invariance would avoid locally fine-grained interpolation since the relation between the target and surrounding observation points varies over regions due to its topography and subsurface structure. To overcome this problem, we propose introducing position-dependent convolution where kernel weights are adjusted depending on their position on an image based on the trainable position-feature map. We show the effectiveness of our proposed method, called, PoDIM (Position-dependent Deep Inpainting Method), through experiments using simulated ground-motion data.} }
Endnote
%0 Conference Paper %T Position-dependent partial convolutions for supervised spatial interpolation %A Hirotaka Hachiya %A Kotaro Nagayoshi %A Asako Iwaki %A Takahiro Maeda %A Naonori Ueda %A Hiroyuki Fujiwara %B Proceedings of The 14th Asian Conference on Machine Learning %C Proceedings of Machine Learning Research %D 2023 %E Emtiyaz Khan %E Mehmet Gonen %F pmlr-v189-hachiya23a %I PMLR %P 420--435 %U https://proceedings.mlr.press/v189/hachiya23a.html %V 189 %X Acquiring continuous spatial data, e.g., spatial ground motion is essential to assess the damaged area and appropriately assign rescue and medical teams. To this purpose, spatial interpolation methods have been developed to estimate the value of unobserved points linearly from neighbor observed values, i.e., inverse distance weighting and Kriging. Recently, realistic spatial continuous environmental data with various scenarios can be generated by 3D finite difference methods with a high-resolution structure model. It enables us to collect supervised data even for unobserved points. Along this line, we propose a framework of supervised spatial interpolation and apply highly advanced deep inpainting methods where we treat spatially distributed observed points as a masked image and non-linearly expand them through convolutional encoder-decoder networks. However, the property of translation invariance would avoid locally fine-grained interpolation since the relation between the target and surrounding observation points varies over regions due to its topography and subsurface structure. To overcome this problem, we propose introducing position-dependent convolution where kernel weights are adjusted depending on their position on an image based on the trainable position-feature map. We show the effectiveness of our proposed method, called, PoDIM (Position-dependent Deep Inpainting Method), through experiments using simulated ground-motion data.
APA
Hachiya, H., Nagayoshi, K., Iwaki, A., Maeda, T., Ueda, N. & Fujiwara, H.. (2023). Position-dependent partial convolutions for supervised spatial interpolation. Proceedings of The 14th Asian Conference on Machine Learning, in Proceedings of Machine Learning Research 189:420-435 Available from https://proceedings.mlr.press/v189/hachiya23a.html.

Related Material