NIERT: Accurate Numerical Interpolation Through Unifying Scattered Data Representations Using Transformer Encoder

被引:0
|
作者
Ding, Shizhe [1 ,2 ]
Xia, Boyang [1 ,2 ]
Ren, Milong [1 ,2 ]
Bu, Dongbo [1 ,2 ,3 ]
机构
[1] Chinese Acad Sci, Inst Comp Technol, Key Lab Intelligent Informat Proc, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Beijing 100049, Peoples R China
[3] Henan Acad Sci, Cent China Artificial Intelligence Res Inst, Zhengzhou 450046, Peoples R China
基金
中国国家自然科学基金;
关键词
Interpolation algorithm; pre-trained models; scattered data; transformer encoder;
D O I
10.1109/TKDE.2024.3402444
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Interpolation for scattered data is a classical problem in numerical analysis, with a long history of theoretical and practical contributions. Recent advances have utilized deep neural networks to construct interpolators, exhibiting excellent and generalizable performance. However, they still fall short in two aspects: 1) inadequate representation learning, resulting from separate embeddings of observed and target points in popular encoder-decoder frameworks and 2) limited generalization power, caused by overlooking prior interpolation knowledge shared across different domains. To overcome these limitations, we present a Numerical Interpolation approach using Encoder Representation of Transformers (called NIERT). On one hand, NIERT utilizes an encoder-only framework rather than the encoder-decoder structure. This way, NIERT can embed observed and target points into a unified encoder representation space, thus effectively exploiting the correlations among them and obtaining more precise representations. On the other hand, we propose to pre-train NIERT on large-scale synthetic mathematical functions to acquire prior interpolation knowledge, and transfer it to multiple interpolation domains with consistent performance gain. On both synthetic and real-world datasets, NIERT outperforms the existing approaches by a large margin, i.e., 4.3 similar to 14.3x lower MAE on TFRD subsets, and 1.7/1.8/8.7x lower MSE on Mathit/PhysioNet/PTV datasets.
引用
收藏
页码:6731 / 6744
页数:14
相关论文
共 50 条