{ "id": "1611.05742", "version": "v1", "published": "2016-11-17T15:37:23.000Z", "updated": "2016-11-17T15:37:23.000Z", "title": "Building Deep Networks on Grassmann Manifolds", "authors": [ "Zhiwu Huang", "Jiqing Wu", "Luc Van Gool" ], "categories": [ "cs.CV" ], "abstract": "Representing the data on Grassmann manifolds is popular in quite a few image and video recognition tasks. In order to enable deep learning on Grassmann manifolds, this paper proposes a deep network architecture which generalizes the Euclidean network paradigm to Grassmann manifolds. In particular, we design full rank mapping layers to transform input Grassmannian data into more desirable ones, exploit orthogonal re-normalization layers to normalize the resulting matrices, study projection pooling layers to reduce the model complexity in the Grassmannian context, and devise projection mapping layers to turn the resulting Grassmannian data into Euclidean forms for regular output layers. To train the deep network, we exploit a stochastic gradient descent setting on manifolds where the connection weights reside on, and study a matrix generalization of backpropagation to update the structured data. We experimentally evaluate the proposed network for three computer vision tasks, and show that it has clear advantages over existing Grassmann learning methods, and achieves results comparable with state-of-the-art approaches.", "revisions": [ { "version": "v1", "updated": "2016-11-17T15:37:23.000Z" } ], "analyses": { "keywords": [ "grassmann manifolds", "building deep networks", "design full rank mapping layers", "transform input grassmannian data", "exploit orthogonal re-normalization layers" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }