论文标题

PDE对由神经网络计算的平滑分层功能的约束

PDE constraints on smooth hierarchical functions computed by neural networks

论文作者

Filom, Khashayar, Kording, Konrad Paul, Farhoodi, Roozbeh

论文摘要

神经网络是用于计算的多功能工具,具有近似广泛功能的能力。深度神经网络理论中的一个重要问题是表现力。也就是说,我们想了解给定网络可计算的功能。我们研究了馈电神经网络实现的真实无限(平滑)层次函数,通过在两种情况下组成更简单的功能: 1)组成的每个组成函数的输入少于所得函数; 2)构成函数以非线性单变量函数(例如TANH)的更特定但普遍的形式应用于线性多元函数。 我们确定在每个制度中,都存在非平凡的代数偏微分方程(PDE),这些方程是由计算函数满足的。这些PDE纯粹是基于部分导数,仅取决于网络的拓扑。对于多项式函数的组成,代数PDE在环境多项式空间中产生非平凡的方程(仅取决于体系结构的程度),这些方程在相关的功能品种上满足。相反,我们猜想这样的PDE约束曾经伴随着适当的非象征性条件以及可能涉及部分导数的某些不平等现象,请保证网络可以考虑正在考虑的光滑函数。在许多示例中验证了猜想,包括具有神经科学意义的树木体系结构的情况。我们的方法是制定与特定神经网络相关的功能空间的代数描述的一步,并可能为构建神经网络提供新的有用工具。

Neural networks are versatile tools for computation, having the ability to approximate a broad range of functions. An important problem in the theory of deep neural networks is expressivity; that is, we want to understand the functions that are computable by a given network. We study real infinitely differentiable (smooth) hierarchical functions implemented by feedforward neural networks via composing simpler functions in two cases: 1) each constituent function of the composition has fewer inputs than the resulting function; 2) constituent functions are in the more specific yet prevalent form of a non-linear univariate function (e.g. tanh) applied to a linear multivariate function. We establish that in each of these regimes there exist non-trivial algebraic partial differential equations (PDEs), which are satisfied by the computed functions. These PDEs are purely in terms of the partial derivatives and are dependent only on the topology of the network. For compositions of polynomial functions, the algebraic PDEs yield non-trivial equations (of degrees dependent only on the architecture) in the ambient polynomial space that are satisfied on the associated functional varieties. Conversely, we conjecture that such PDE constraints, once accompanied by appropriate non-singularity conditions and perhaps certain inequalities involving partial derivatives, guarantee that the smooth function under consideration can be represented by the network. The conjecture is verified in numerous examples including the case of tree architectures which are of neuroscientific interest. Our approach is a step toward formulating an algebraic description of functional spaces associated with specific neural networks, and may provide new, useful tools for constructing neural networks.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源