Finding the best not the most: regularized loss minimization subgraph selection for graph classification

Shirui Pan*, Jia Wu, Xingquan Zhu, Guodong Long, Chengqi Zhang

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

34 Citations (Scopus)

Abstract

Classification on structure data, such as graphs, has drawn wide interest in recent years. Due to the lack of explicit features to represent graphs for training classification models, extensive studies have been focused on extracting the most discriminative subgraphs features from the training graph dataset to transfer graphs into vector data. However, such filter-based methods suffer from two major disadvantages: (1) the subgraph feature selection is separated from the model learning process, so the selected most discriminative subgraphs may not best fit the subsequent learning model, resulting in deteriorated classification results; (2) all these methods rely on users to specify the number of subgraph features K, and suboptimally specified K values often result in significantly reduced classification accuracy. In this paper, we propose a new graph classification paradigm which overcomes the above disadvantages by formulating subgraph feature selection as learning a K-dimensional feature space from an implicit and large subgraph space, with the optimal K value being automatically determined. To achieve the goal, we propose a regularized loss minimization-driven (RLMD) feature selection method for graph classification. RLMD integrates subgraph selection and model learning into a unified framework to find discriminative subgraphs with guaranteed minimum loss w.r.t. the objective function. To automatically determine the optimal number of subgraphs K from the exponentially large subgraph space, an effective elastic net and a subgradient method are proposed to derive the stopping criterion, so that K can be automatically obtained once RLMD converges. The proposed RLMD method enjoys gratifying property including proved convergence and applicability to various loss functions. Experimental results on real-life graph datasets demonstrate significant performance gain.

Original languageEnglish
Pages (from-to)3783-3796
Number of pages14
JournalPattern Recognition
Volume48
Issue number11
DOIs
Publication statusPublished - 1 Nov 2015
Externally publishedYes

Keywords

  • Classification
  • Feature selection
  • Graph classification
  • Sparse learning

Fingerprint

Dive into the research topics of 'Finding the best not the most: regularized loss minimization subgraph selection for graph classification'. Together they form a unique fingerprint.

Cite this