pith. machine review for the scientific record. sign in

arxiv: 1705.02302 · v5 · submitted 2017-05-05 · 💻 cs.LG · cs.NE

Recognition: unknown

Analysis and Design of Convolutional Networks via Hierarchical Tensor Decompositions

Authors on Pith no claims yet
classification 💻 cs.LG cs.NE
keywords convolutionalarchitecturebiasefficiencyexpressiveinductivenetworknetworks
0
0 comments X
read the original abstract

The driving force behind convolutional networks - the most successful deep learning architecture to date, is their expressive power. Despite its wide acceptance and vast empirical evidence, formal analyses supporting this belief are scarce. The primary notions for formally reasoning about expressiveness are efficiency and inductive bias. Expressive efficiency refers to the ability of a network architecture to realize functions that require an alternative architecture to be much larger. Inductive bias refers to the prioritization of some functions over others given prior knowledge regarding a task at hand. In this paper we overview a series of works written by the authors, that through an equivalence to hierarchical tensor decompositions, analyze the expressive efficiency and inductive bias of various convolutional network architectural features (depth, width, strides and more). The results presented shed light on the demonstrated effectiveness of convolutional networks, and in addition, provide new tools for network design.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 1 Pith paper

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Neural Operator: Graph Kernel Network for Partial Differential Equations

    cs.LG 2020-03 unverdicted novelty 7.0

    Graph Kernel Networks learn PDE solution operators that generalize across discretization methods and grid resolutions using graph-based kernel integration.