Deep Declarative Networks

Stephen Gould*, Richard Hartley, Dylan Campbell

*Corresponding author for this work

    Research output: Contribution to journalArticlepeer-review

    32 Citations (Scopus)

    Abstract

    We explore a class of end-to-end learnable models wherein data processing nodes (or network layers) are defined in terms of desired behavior rather than an explicit forward function. Specifically, the forward function is implicitly defined as the solution to a mathematical optimization problem. Consistent with nomenclature in the programming languages community, we name these models deep declarative networks. Importantly, it can be shown that the class of deep declarative networks subsumes current deep learning models. Moreover, invoking the implicit function theorem, we show how gradients can be back-propagated through many declaratively defined data processing nodes thereby enabling end-to-end learning. We discuss how these declarative processing nodes can be implemented in the popular PyTorch deep learning software library allowing declarative and imperative nodes to co-exist within the same network. We also provide numerous insights and illustrative examples of declarative nodes and demonstrate their application for image and point cloud classification tasks.

    Original languageEnglish
    Pages (from-to)3988-4004
    Number of pages17
    JournalIEEE Transactions on Pattern Analysis and Machine Intelligence
    Volume44
    Issue number8
    DOIs
    Publication statusPublished - 1 Aug 2022

    Fingerprint

    Dive into the research topics of 'Deep Declarative Networks'. Together they form a unique fingerprint.

    Cite this