...
首页> 外文期刊>Multiscale modeling & simulation >MULTILEVEL FINE-TUNING: CLOSING GENERALIZATION GAPS IN APPROXIMATION OF SOLUTION MAPS UNDER A LIMITED BUDGET FOR TRAINING DATA
【24h】

MULTILEVEL FINE-TUNING: CLOSING GENERALIZATION GAPS IN APPROXIMATION OF SOLUTION MAPS UNDER A LIMITED BUDGET FOR TRAINING DATA

机译:多级微调:在培训数据的有限预算下的解决方案地图近似的概括间隙

获取原文
获取原文并翻译 | 示例
           

摘要

In scientific machine learning, regression networks have been recently applied to approximate solution maps (e.g., the potential-ground state map of the Schrodinger equation). In this paper, we aim to reduce the generalization error without spending more time on generating training samples. However, to reduce the generalization error, the regression network needs to be fit on a large number of training samples (e.g., a collection of potential-ground state pairs). The training samples can be produced by running numerical solvers, which takes significant time in many applications. In this paper, we aim to reduce the generalization error without spending more time on generating training samples. Inspired by few-shot learning techniques, we develop the multilevel fine-tuning algorithm by introducing levels of training: we first train the regression network on samples generated at the coarsest grid and then successively fine-tune the network on samples generated at finer grids. Within the same amount of time, numerical solvers generate more samples on coarse grids than on fine grids. We demonstrate a significant reduction of generalization error in numerical experiments on challenging problems with oscillations, discontinuities, or rough coefficients. Further analysis can be conducted in the neural tangent kernel regime, and we provide practical estimators to the generalization error. The number of training samples at different levels can be optimized for the smallest estimated generalization error under the constraint of budget for training data. The optimized distribution of budget over levels provides practical guidance with theoretical insight as in the celebrated multilevel Monte Carlo algorithm.
机译:None

著录项

相似文献

  • 外文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号