Nonlocal vector calculus, which is based on the nonlocal forms of gradient, divergence, and Laplace operators in multiple dimensions, has shown promising applications in fields such as hydrology, mechanics, and image processing. In this work, we study the analytical underpinnings of these operators. We rigorously treat compositions of nonlocal operators, prove nonlocal vector calculus identities, and connect weighted and unweighted variational frameworks. We combine these results to obtain a weighted fractional Helmholtz decomposition which is valid for sufficiently smooth vector fields. Our approach identifies the function spaces in which the stated identities and decompositions hold, providing a rigorous foundation to the nonlocal vector calculus identities that can serve as tools for nonlocal modeling in higher dimensions.
Nonlocal models provide a much-needed predictive capability for important Sandia mission applications, ranging from fracture mechanics for nuclear components to subsurface flow for nuclear waste disposal, where traditional partial differential equations (PDEs) models fail to capture effects due to long-range forces at the microscale and mesoscale. However, utilization of this capability is seriously compromised by the lack of a rigorous nonlocal interface theory, required for both application and efficient solution of nonlocal models. To unlock the full potential of nonlocal modeling we developed a mathematically rigorous and physically consistent interface theory and demonstrate its scope in mission-relevant exemplar problems.
Modeling of phenomena such as anomalous transport via fractional-order differential equations has been established as an effective alternative to partial differential equations, due to the inherent ability to describe large-scale behavior with greater efficiency than fully resolved classical models. In this review article, we first provide a broad overview of fractional-order derivatives with a clear emphasis on the stochastic processes that underlie their use. We then survey three exemplary application areas — subsurface transport, turbulence, and anomalous materials — in which fractional-order differential equations provide accurate and predictive models. For each area, we report on the evidence of anomalous behavior that justifies the use of fractional-order models, and survey both foundational models as well as more expressive state-of-the-art models. We also propose avenues for future research, including more advanced and physically sound models, as well as tools for calibration and discovery of fractional-order models.
We develop a framework for Gaussian processes regression constrained by boundary value problems. The framework may be applied to infer the solution of a well-posed boundary value problem with a known second-order differential operator and boundary conditions, but for which only scattered observations of the source term are available. Scattered observations of the solution may also be used in the regression. The framework combines co-kriging with the linear transformation of a Gaussian process together with the use of kernels given by spectral expansions in eigenfunctions of the boundary value problem. Thus, it benefits from a reduced-rank property of covariance matrices. We demonstrate that the resulting framework yields more accurate and stable solution inference as compared to physics-informed Gaussian process regression without boundary condition constraints.
A key challenge to nonlocal models is the analytical complexity of deriving them from first principles, and frequently their use is justified a posteriori. In this work we extract nonlocal models from data, circumventing these challenges and providing data-driven justification for the resulting model form. Extracting data-driven surrogates is a major challenge for machine learning (ML) approaches, due to nonlinearities and lack of convexity — it is particularly challenging to extract surrogates which are provably well-posed and numerically stable. Our scheme not only yields a convex optimization problem, but also allows extraction of nonlocal models whose kernels may be partially negative while maintaining well-posedness even in small-data regimes. To achieve this, based on established nonlocal theory, we embed in our algorithm sufficient conditions on the non-positive part of the kernel that guarantee well-posedness of the learnt operator. These conditions are imposed as inequality constraints to meet the requisite conditions of the nonlocal theory. We demonstrate this workflow for a range of applications, including reproduction of manufactured nonlocal kernels; numerical homogenization of Darcy flow associated with a heterogeneous periodic microstructure; nonlocal approximation to high-order local transport phenomena; and approximation of globally supported fractional diffusion operators by truncated kernels.
We analyze the well-posedness of an anisotropic, nonlocal diffusion equation. Establishing an equivalence between weighted and unweighted anisotropic nonlocal diffusion operators in the vein of unified nonlocal vector calculus, we apply our analysis to a class of fractional-order operators and present rigorous estimates for the solution of the corresponding anisotropic anomalous diffusion equation. Furthermore, we extend our analysis to the anisotropic diffusion-advection equation and prove well-posedness for fractional orders s ∊ [0.5, 1). We also present an application of the advection-diffusion equation to anomalous transport of solutes.
Second-order optimizers hold intriguing potential for deep learning, but suffer from increased cost and sensitivity to the non-convexity of the loss surface as compared to gradient-based approaches. We introduce a coordinate descent method to train deep neural networks for classification tasks that exploits global convexity of the cross-entropy loss in the weights of the linear layer. Our hybrid Newton/Gradient Descent (NGD) method is consistent with the interpretation of hidden layers as providing an adaptive basis and the linear layer as providing an optimal fit of the basis to data. By alternating between a second-order method to find globally optimal parameters for the linear layer and gradient descent to train the hidden layers, we ensure an optimal fit of the adaptive basis to data throughout training. The size of the Hessian in the second-order step scales only with the number weights in the linear layer and not the depth and width of the hidden layers; furthermore, the approach is applicable to arbitrary hidden layer architecture. Previous work applying this adaptive basis perspective to regression problems demonstrated significant improvements in accuracy at reduced training cost, and this work can be viewed as an extension of this approach to classification problems. We first prove that the resulting Hessian matrix is symmetric semi-definite, and that the Newton step realizes a global minimizer. By studying classification of manufactured two-dimensional point cloud data, we demonstrate both an improvement in validation error and a striking qualitative difference in the basis functions encoded in the hidden layer when trained using NGD. Application to image classification benchmarks for both dense and convolutional architectures reveals improved training accuracy, suggesting gains of second-order methods over gradient descent. A Tensorflow implementation of the algorithm is available at github.com/rgp62/.