{ "id": "1602.02644", "version": "v1", "published": "2016-02-08T16:50:28.000Z", "updated": "2016-02-08T16:50:28.000Z", "title": "Generating Images with Perceptual Similarity Metrics based on Deep Networks", "authors": [ "Alexey Dosovitskiy", "Thomas Brox" ], "categories": [ "cs.LG", "cs.CV", "cs.NE" ], "abstract": "Image-generating machine learning models are typically trained with loss functions based on distance in the image space. This often leads to over-smoothed results. We propose a class of loss functions, which we call deep perceptual similarity metrics (DeePSiM), that mitigate this problem. Instead of computing distances in the image space, we compute distances between image features extracted by deep neural networks. This metric better reflects perceptually similarity of images and thus leads to better results. We show three applications: autoencoder training, a modification of a variational autoencoder, and inversion of deep convolutional networks. In all cases, the generated images look sharp and resemble natural images.", "revisions": [ { "version": "v1", "updated": "2016-02-08T16:50:28.000Z" } ], "analyses": { "keywords": [ "deep networks", "generating images", "deep perceptual similarity metrics", "metric better reflects perceptually similarity", "loss functions" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable", "adsabs": "2016arXiv160202644D" } } }