{ "id": "1905.13686", "version": "v1", "published": "2019-05-31T15:51:29.000Z", "updated": "2019-05-31T15:51:29.000Z", "title": "Explainability Techniques for Graph Convolutional Networks", "authors": [ "Federico Baldassarre", "Hossein Azizpour" ], "comment": "Accepted at the ICML 2019 Workshop \"Learning and Reasoning with Graph-Structured Representations\" (poster + spotlight talk)", "categories": [ "cs.LG", "cs.AI", "stat.ML" ], "abstract": "Graph Networks are used to make decisions in potentially complex scenarios but it is usually not obvious how or why they made them. In this work, we study the explainability of Graph Network decisions using two main classes of techniques, gradient-based and decomposition-based, on a toy dataset and a chemistry task. Our study sets the ground for future development as well as application to real-world problems.", "revisions": [ { "version": "v1", "updated": "2019-05-31T15:51:29.000Z" } ], "analyses": { "keywords": [ "graph convolutional networks", "explainability techniques", "graph network decisions", "real-world problems", "main classes" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }