This repository is the official implementation of the paper:
A Hierarchical Spatial Transformer for Massive Point Samples in Continuous Space
Wenchong He1, Zhe Jiang1, Tingsong Xiao1, Zelin Xu1, Shigang Chen1, Ronald Fick2, Miles Medina2, Christine Angelini2
1 Dept. of Comp. & Info. Sci. & Eng., University of Florida
2 Center for Coastal Solutions, University of Florida
[Paper] [Code]
Abstract: Given a set of point samples in continuous space with explanatory features and target response variables, the problem is to learn the spatial structured representation of the continuous space and to infer the target variable at any new point location. The problem is fundamental with broad applications, such as coastal water quality monitoring, air quality monitoring, and operator learning on physical simulations. However, the problem is challenging due to the implicit dependency structure on irregular point locations in continuous space, the potential high computational costs of modeling long-range interactions between a large number of points, and the risks of over-confident predictions due to local point sparsity. Existing works either assume a given spatial neighbor structure or infer unknown spatial dependency under a rigid distribution assumption. In recent years, transformer architecture has been widely used to learn spatial representation, but existing methods are primarily designed for regular grids or graphs and thus cannot be directly applied to irregular point samples in continuous space. There are also works related to operator learning on numerical simulations in the continuous space, but these methods often do not address the hierarchical spatial representation of irregular points. To fill this gap, this paper proposes a new hierarchical spatial transformer model for a large number of irregular point samples in continuous space. Our key idea is to learn the multi-scale spatial representation through a quad-tree hierarchy and to conduct efficient attention operations by using a coarser representation for distant points. We also design an uncertainty quantification component to measure spatial prediction confidence. Evaluations of several datasets confirm that our method outperforms multiple baselines in existing works.
TODO
TODO
TODO
TODO
We provide the results on prediction performance, uncertainty quantification performance, and computation and memory cost analysis.
- prediction performance
- sensitivity analysis
TODO
TODO: Figure 7
- activation memory costs of HST encoder
Please cite our paper if you find this code useful for your work:
@inproceedings{he2023hierarchical,
title={A Hierarchical Spatial Transformer for Massive Point Samples in Continuous Space},
author={He, Wenchong and Jiang, Zhe and Xiao, Tingsong and Xu, Zelin and Chen, Shigang and Fick, Ronald and Medina, Miles and Angelini, Christine},
booktitle={Proceedings of the 2023 Conference on Neural Information Processing Systems (NeurIPS)},
year={2023},
}
TODO