- Proceedings
- Open Access

# Protein functional properties prediction in sparsely-label PPI networks through regularized non-negative matrix factorization

- Qingyao Wu
^{1}, - Zhenyu Wang
^{1}, - Chunshan Li
^{2}, - Yunming Ye
^{3}, - Yueping Li
^{4}and - Ning Sun
^{5}Email author

**9 (Suppl 1)**:S9

https://doi.org/10.1186/1752-0509-9-S1-S9

© Wu et al.; licensee BioMed Central Ltd. 2015

**Published:**21 January 2015

## Abstract

### Background

Predicting functional properties of proteins in protein-protein interaction (PPI) networks presents a challenging problem and has important implication in computational biology. Collective classification (CC) that utilizes both attribute features and relational information to jointly classify related proteins in PPI networks has been shown to be a powerful computational method for this problem setting. Enabling CC usually increases accuracy when given a fully-labeled PPI network with a large amount of labeled data. However, such labels can be difficult to obtain in many real-world PPI networks in which there are usually only a limited number of labeled proteins and there are a large amount of unlabeled proteins. In this case, most of the unlabeled proteins may not connected to the labeled ones, the supervision knowledge cannot be obtained effectively from local network connections. As a consequence, learning a CC model in sparsely-labeled PPI networks can lead to poor performance.

### Results

We investigate a latent graph approach for finding an integration latent graph by exploiting various latent linkages and judiciously integrate the investigated linkages to link (separate) the proteins with similar (different) functions. We develop a regularized non-negative matrix factorization (RNMF) algorithm for CC to make protein functional properties prediction by utilizing various data sources that are available in this problem setting, including attribute features, latent graph, and unlabeled data information. In RNMF, a label matrix factorization term and a network regularization term are incorporated into the non-negative matrix factorization (NMF) objective function to seek a matrix factorization that respects the network structure and label information for classification prediction.

### Conclusion

Experimental results on KDD Cup tasks predicting the localization and functions of proteins to yeast genes demonstrate the effectiveness of the proposed RNMF method for predicting the protein properties. In the comparison, we find that the performance of the new method is better than those of the other compared CC algorithms especially in paucity of labeled proteins.

## Keywords

- Label Data
- Unlabeled Data
- Nonnegative Matrix Factorization
- Nonnegative Matrix Factorization
- Label Information

## Background

Advances in experimental methods in sequencing technologies results in the rapid growth of genome sequences and gene expression profiles in last decade. A critical problem in making use of these sequenced and associated experimental data is the assignment of functional information. Although the knowledge of protein functions can be acquired by conducting various biochemical experiments, it is both expensive and time-consuming by relying doing experiments alone to identify the functional properties of newly sequenced proteins which can no longer catch up with their rapid growth. Therefore, various computational methods have been developed for automated prediction in the biological literature.

The task of protein functional properties prediction has been explored widely (e.g., see an extensive review on this task for overviews [1]). The conventional prediction methods usually concentrated on protein sequence homology through finding homologies of a protein based on their similarity. Typically, each protein is represented as a feature vector (e.g., textual features from MEDLINE), and the attribute features are taken as input to machine learning algorithms, such as SVM [2], neural networks [3], and random forest [4], to infer annotation rules for predicting the functional properties of unlabeled proteins [5]. However, these kinds of methods do not consider the function diversification when a protein produces interactions with other ones.

Protein-protein interaction (PPI) networks are becoming increasing rich and useful in delineating the biological processes, pathways and complexes that proteins take part in. As a consequence, many works have considered using protein interactions to make prediction. The network-based methods study the task of protein functional properties prediction in the context of PPI networks based on the assumption that the interaction partners of a protein are likely to share similar functions with it. Sharan et al. [6] summarize the methods into two groups: direct annotation schemes, which infer the function of a protein based on its connections in the network [7–9]; and module-assisted schemes which first identify modules of related proteins and then annotate each module based on the known functions of its members [10, 11]. However, these types of methods using only interaction partners limit predictions to proteins that have at least one interaction partner with known annotation.

In recent years, there is an increasing concern about using collective classification (CC) that utilizes both attribute features and protein interactions to jointly classifying related proteins in PPI networks [12–14]. CC methods, such as the iterative classification algorithm (ICA), usually explore dependencies between proteins based on the analysis of attributes and functions of neighboring partners. To do so, the attribute features of each protein, together with the additional relational features derived from the linked neighbors are combined for prediction. The additional relational features can potentially increase classification accuracy. But as some of the neighboring proteins' functions may initially unknown, and thus this inferring process may decrease accuracy as well when there are only a limited number of labeled neighboring proteins.

To tackle this challenge, we investigate a latent graph approach for finding an integration latent graph by exploiting various latent linkages among protein nodes to link (separate) the proteins with similar (different) functional properties (see Figure 1(c)). Via the latent graph constructed, the supervision knowledge may be able to propagate more effectively from labeled proteins to unlabeled proteins. Then, we develop a regularized non-negative matrix factorization (RNMF) algorithm to make prediction on the latent graph. Conventional non-negative matrix factorization (NMF) method is specifically designed for unsupervised learning and cannot be directly used for network data classification. In RNMF, we extend the NMF objective function by adding a label matrix factorization term and an additional network regularization term to encode the network structure and label information of proteins, and we seek a matrix factorization which gives a new data representation that provides a good approximation of the original data matrix to make prediction for the unlabeled proteins. In this way, the prediction has local smoothness on the network from labeled proteins to unlabeled proteins. As a result, RNMF can have more discriminating power than the ordinary NMF approach which only considers the Euclidean structure of the data.

We study the KDD Cup 2001 tasks of predicting properties (protein localization and their functions) of the protein corresponding to a given yeast gene. Experimental results show that the proposed RNMF algorithm is able to deliver better performance than other compared CC algorithms in paucity of labeled proteins. In summary, the main contributions of this paper are listed as follows:

1 This article studies the protein functional properties prediction problem on sparsely-labeled PPI networks with only a limited number of labeled proteins, which is a very common situation in functional genomics but traditional prediction approaches rarely consider such label deficiency problem.

2 It is the first one to propose a NMF based algorithm that utilizes various additional data sources, including attribute features, latent graph, and unlabeled data, to improve the performance of protein functional properties prediction.

3 The proposed RNMF algorithm extends NMF method by incorporating two additional terms into the model to encode the network structure and labeled information to obtain a local smoothness of predictions on the PPI network. This leads to better prediction performance against the other CC methods especially in paucity of labeled data.

## Methods

### Protein functional properties prediction task

Conventional supervised learning methods assume that the instances to be classified are independent of each other while collective classification (CC) considers to jointly classify interrelated instances in a network by exploiting their interrelations [15]. From this viewpoint, the task of protein functional properties prediction can be cast into the collective classification problem of learning a predictive model from PPI networks. Generally, a PPI network can be represented by a graph where nodes (proteins) interconnected with each other by edges reflecting the interactions between the proteins. Information on each protein node is represented as an attribute feature vector. We are given a set of labeled proteins of known functional classes, and the task is to predict the functions of the remaining nodes of unlabeled proteins. Nevertheless, the functional class membership of one protein may influence the class membership of a related protein.

Formally, the protein functional properties prediction task is described as follows: let **G** = (**V**, **E**, **X**, **Y**) be a protein network dataset. **V** is a set of protein nodes {*v*_{1},...,*v*_{
N
}}. $E=\left[{E}_{ij}\right]\in {\mathbb{R}}^{N\times N}$ is the weighting matrix whereas *E*_{i,j}indicates the weights on the edge between node *v*_{
i
} and node *v*_{
j
} . $X=\left[{x}_{\mathsf{\text{1}}},...,{x}_{N}\right]\in {\mathbb{R}}^{M\times N}$ denotes a data matrix consists of *N* protein attribute feature vectors of dimensionality *M*, where each *x*_{
i
} ∈ **X** is an attribute vector for a node *v*_{
i
} ∈ **V**. {*c*_{1}, *c*_{2}, ..., *c*_{
q
}} is the set of *q* possible labels. $Y=\left[{Y}_{\mathsf{\text{1}}},...,\phantom{\rule{2.77695pt}{0ex}}{Y}_{N}\right]\in {\mathbb{R}}^{q\times N}$ denotes the set of class labels where *Y*_{
i
} is the class labels of protein *v*_{
i
}. Each ${Y}_{i}={\left[{{Y}_{i,}}_{\mathsf{\text{1}}},...,{Y}_{i,q}\right]}^{T}\in {\left\{0,\mathsf{\text{1}}\right\}}_{q}$ such that *Y*_{i,j}= 1 means that protein *v*_{
i
} is associated with class *c*_{
j
} and *Y*_{i,j}= 0 otherwise. Assume that we have *n'* labeled proteins ${\left\{\left({x}_{i},{Y}_{i}\right)\right\}}_{i=\mathsf{\text{1}}}^{n\prime}$ and *n"* unlabeled data ${\left\{\left({x}_{i}\right)\right\}}_{i=n\prime +\mathsf{\text{1}}}^{n\prime +n\prime \prime}$ with *N* = *n'* + *n"*. The task is to predict the functional classes of unlabeled proteins. When there are only a limited number of labeled proteins in the network, i.e. *n'* ≪ *n"*, most of the proteins may not connect to labeled ones, which makes the task very challenging. As such, it is natural to consider semi-supervised learning and network exploration techniques to utilize different data sources that are available in this problem setting, including attribute features, protein interactions, and unlabeled data, to improve the prediction performance.

### Nonnegative matrix factorization

Nonnegative Matrix Factorization (NMF) is a matrix factorization technique for discovering low dimensional representations of data [16, 17]. In many applications, the input data matrix is of very high dimension, NMF seeks to find two lower dimensional matrices (nonnegative) whose product provides a good approximation to the original data matrix. NMF has received much attention because the learned bases can be interpreted as a natural parts-based representation of data and this interpretation is consistent with the psychological intuition of combining parts to form a whole, like face images and text documents [16, 18]. That is, we can explain each data instance by additive linear combination of nonnegative basis vectors because NMF allows only additive combinations. For this reason, NMF has been widely used in various real world applications, such as face recognition [19], document clustering [20] and gene expression analysis [21].

*N*nonnegative column vectors (each is an input instance vector of dimensionality

*M*), NMF seeks to find two nonnegative matrices $U=\mathsf{\text{[}}{u}_{ik}\mathsf{\text{]}}\in {\mathbb{R}}^{M\times K}$ and $V=\mathsf{\text{[}}{v}_{jk}\mathsf{\text{]}}\in {\mathbb{R}}^{N\times K}$ whose product provides a good approximate to the original data matrix

**X**, typically

*K*≪

*M*and

*K*≪

*N*, with the following form

where **U**, **V** ≥ 0,**U** is called a basis matrix and **V** is called a coefficient matrix.

The algorithm minimizing the objective function in Eq.(2) using the above multiplicative updates. These updates are guaranteed to decrease the approximation cost at each iteration, and converge to a local minimum of the objective function.

### Latent graphs for protein function prediction

In the protein functional properties prediction task, we are given a PPI network data represented as a graph **G**(**V**, **E**, **X**, **Y**). Our objective is to make prediction for unlabeled proteins on the graph. Recent studies [22, 23] have shown that learning performance can be significantly enhanced when the network structure is exploited and the local invariance is considered. The power of these approaches lies in the fact that the exploited network topology generally exhibits the predictable relationships between the input instances and the output class labels.

For protein functional properties prediction, there are two ways of looking at this problem by considering local invariance: i) two neighboring proteins *v*_{
i
} and *v*_{
j
} with large linkage weight *E*_{i,j}are likely to share similar functional classes; ii) if *v*_{
i
} and *v*_{
j
} have small *E*_{i,j}, they tend to have different functional classes. Suppose we have a network structure that can well respect the predictable relationships between the proteins and functional classes, we should be able to have a good performance of predicting the functional properties of unlabeled proteins.

However, the above scheme may not work well on sparsely-labeled PPI networks where there are plenty of links among the proteins but only few of these neighboring proteins are labeled or there are only few links shared between labeled proteins and unlabeled proteins. Recently, researchers [24, 25] have considered exploiting various latent linkages among the nodes to find latent graphs with more desirable form of network structures for prediction. Specifically, a weight matrix **E** = [*E*_{
ij
}] is defined for one constructed graph, one latent edge is created for each pair nodes and a weight is assigned on the edge based on the proximity of the nodes. For protein function prediction, we can define the weight matrix **E** for latent graph generation using different data sources that are available in this problem setting. Three of the most commonly used methods are as follows:

*PPI latent graph:*The original PPI network can be considered as a latent graph. We define the weight matrix

*E*

^{(1)}of the PPI latent graph as follows

where *E*(*i*, *j*) = 1 if node *vi* and node *v*_{
j
} are connected in the PPI network, and *E*(*i*, *j*) = 0 otherwise.

*Random walk latent graph:*It is observed that proteins that interact with level-2 neighbors (indirect neighbors in the PPI network) also have a great likelihood of sharing similar characteristics [7]. Thus, we also use the idea of

*even-step*random walk with restart (ERWR) [25] to construct the random walk latent graph. Given the weight matrix

**E**of the original PPI network, we compute

**P**=

**EE**and normalize its entries with respect to each column to obtain a normalized transition probability matrix

**P**. The ERWR uses a random walker to iteratively visit the neighbourhood nodes with transition probability given in

**P**. Also at each step, it has probability

*α*(e.g.,

*α*= 0.1) to return to the start node. We define the weight matrix

**E**

^{(2)}of the random walk latent graph as follows

where $R={\mathrm{\Sigma}}_{t=\mathsf{\text{1}}}^{T}\phantom{\rule{2.77695pt}{0ex}}\alpha {\left(\mathsf{\text{1}}-\alpha \right)}^{t}{P}^{t}$ is the steady-state probability matrix after *T* steps, and *R*(*i*, *j*) is the (*i*, *j*)th entry in **R**.

*Prediction similarity latent graph:*We consider the values of class labels of the labeled proteins as input features to build a classifier, and give prediction to the remaining proteins. Specifically, we use SVM classifier with probability outputs implemented in the LIBSVM library [26] to compute the classification confidence ${Y}_{j}^{\prime}$ of a protein

*x*

_{ i }to different classes, where ${Y}_{i}^{\prime}=\left[P\left({c}_{\mathsf{\text{1}}}\phantom{\rule{0.3em}{0ex}}|\phantom{\rule{0.3em}{0ex}}{x}_{i}\right),\phantom{\rule{2.77695pt}{0ex}}\cdot \phantom{\rule{2.77695pt}{0ex}}\cdot \phantom{\rule{2.77695pt}{0ex}}\cdot \phantom{\rule{2.77695pt}{0ex}},\phantom{\rule{2.77695pt}{0ex}}P\left({c}_{q}\phantom{\rule{0.3em}{0ex}}|\phantom{\rule{0.3em}{0ex}}{x}_{i}\right)\right]$ where

*P*(

*c*

_{ j }|

*x*

_{ i }) is the probability of the protein

*xi*belongs to the class

*c*

_{ j }. The weight matrix

**E**

^{(3)}of latent graph is based on the cosine similarity of prediction confidences of two proteins, and it is defined as follow

where $cosin\left({Y}_{i}^{\prime},{Y}_{j}^{\prime}\right)=\frac{{{Y}^{\prime}}_{i}\cdot {{Y}^{\prime}}_{j}}{\left|\right|{{Y}^{\prime}}_{i}\left|\right|\cdot \left|\right|{{Y}^{\prime}}_{j}\left|\right|}$.

### Latent graphs integration

**E**

^{(1)},

**E**

^{(2)}, ...,

**E**

^{(t)}} to integrate the latent linkages to infer a integration latent graph

**Ê**such that the weight on the edge of two protein nodes linked together is large (small) if they have similar (different) functions. Here,

*t*is the number of different latent graphs. Formally, we define the integration latent graph

**Ê**as follows

where *w*_{
i
} is the combination weight for the *i*th latent graph.

*q*-by-

*N*matrix

**Y**as the label matrix of all the data, we define another

*q*-by-

*n'*label matrix

**Ῡ**which respects the label information of the labeled examples

where *q* is the number of class labels, *n'* is the number of labeled examples, and *N* is the number of all examples.

Similar, denote a *N*-by-*N* matrix **E**^{(i) }as the weight matrix of *i*th latent graph, we define another *n'*-by-*n'* weight matrix **Ē**^{(i) }which only respects the linkage weights among the labeled examples of the *i*th latent graph.

**Ῡ**and

**Ē**that respects the labeled proteins' label information and linkage information to learn the weights. The idea is to ensure that proteins have similar (different) class labels have large (small) linkage weights. To this end, the square of the Euclidean distance between the matrices

**Ῡ**

^{ T }

**Ῡ**and $\sum _{i=1}^{q}{w}_{i}{\stackrel{\u0304}{E}}^{\left(i\right)}$ is used as the objective function to quantify the quality of the combination weights because the optimal similarity matrix for the labeled examples should be

**Ῡ**

^{ T }

**Ῡ**. Formally, the objective function can be written as

where **Ē** and **Ῡ**^{
T
}**Ῡ** are *n'*-by-*n'* matrices, **w** = [*w*_{1}, ...,*w*_{
t
}]^{
T
} is the combination weight vector, *λ* is a smoothing parameter (e.g., *λ* = 0.01). The objective is solved to learn the weights **w** for latent graph integration. Then, collective classification methods can be applied on the learnt latent graph to make prediction.

### Regularized NMF with latent graph for protein function prediction

Various methods have been developed for protein functional properties prediction. Previous works have shown that NMF is a general method for robust pattern discovery in complex biological systems [21]. NMF appears to have advantages over other methods such as hierarchical clustering to recover meaningful biological information based on the protein attribute feature matrix. Concretely, NMF aims to find two non-negative matrices whose product provides a good approximation to the original matrix. The nonnegative constraints lead to a parts-based representation. For instance, when applied to face image, NMF yielded a decomposition of faces into parts reminiscent of features such as eyes, nose, etc. The protein patterns in terms of attribute features of the proteins are summarized while applying NMF to the problem of protein function prediction, i.e., NMF is able to provide an interesting decomposition of proteins analogous to facial features in Lee and Seung's work [16] on images.

The main aim of this paper is to study the effectiveness of the NMF and latent graph learning approaches for the problem of protein functional properties prediction. The ordinary NMF method as well as most of its variants cannot be directly applied for network data classification task because the approaches are developed for analysis of unlabeled examples in the context of Euclidean structure of the data. The updates in Eq. (3) and (4) derived from the objective function of NMF in Eq. (2) simply ignore the label information and network structure which play a crucial role for functional genomics problems.

To leverage the power of both NMF and latent graph learning, in this paper, we propose a novel regularized nonnegative matrix factorization (RNMF) algorithm, which seeks a matrix factorization that respects the label information and network structure on the constructed latent graph for protein functional properties prediction. To achieve this, a label matrix factorization term and an additional network regularization term are incorporated into the NMF objective function, and an optimization scheme is developed to solve the objective function of the new NMF method.

*x*

_{ i }∈

**X**, and $Y=\left[{Y}_{\mathsf{\text{1}}},\phantom{\rule{2.77695pt}{0ex}}...,\phantom{\rule{2.77695pt}{0ex}}{Y}_{N}\right]\in {\mathbb{R}}^{q\times N}$denotes the label matrix encoding the label information of all the data. For labeled data,

*Y*

_{ ij }= 1 if

*x*

_{ i }is labeled with

*c*

_{ j }, and

*Y*

_{ ij }= 0 otherwise. For unlabeled data,

*Y*

_{ ij }= 0. With the protein attribute feature matrix

**X**and label matrix

**Y**, the objective function of NMF is extended as follows

*A*⊙

*B*]

_{ ij }= [

*A*]

_{ ij }

*·*[

*B*]

_{ ij }, and

*α*is a tradeoff parameter to determine the importance of the label matrix term. Here $B\in {\mathbb{R}}^{q\times K}$ is a basis matrix for the second term, and $W\in {\mathbb{R}}^{q\times N}$is a weight matrix such that elements of

**W**are with nonzero values if the labels of corresponding proteins are known, otherwise elements of

**W**are 0. Specifically, we have

where *Y*_{
ij
} is either 1 or 0 depending on the class membership of the instance.

In the objective function in Eq. (8), the approximation matrix [**BV**^{
T
}]_{
ij
} (with nonnegative value) does not need to be exactly equal to 1 when *Y*_{
ij
} = 1. On the other hand, we hope [**BV**^{
T
}]_{
ij
} to be close to 0 when *Y*_{
ij
} = 0. Thus, in Eq. (9), the weights with respect to *Y*_{
ij
} = 0 is set to be larger than those of *Y*_{
ij
} = 1 for the labeled data.

By using the new NMF method, the supervised knowledge can be effectively preserved, and we seek a matrix factorization which gives a good approximation for both of the data matrix and label matrix. On the other hand, with the integration latent graph **Ê** constructed, one might further hope that the intrinsic network structure can be considered while applying the NMF method to make prediction. In the following, we incorporate a network regularizer into the NMF objective function to seek a matrix factorization that also respects the intrinsic network structure.

*Ê*

_{ jl }of two proteins

*x*

_{ j }and

*x*

_{ l }on the constructed latent graph

**Ê**is large, these two nodes also should be close to each other in terms of the new representations of the matrix factorization. To achieve this, we denote that the new representations of two neighboring nodes

*x*

_{ j }and

*x*

_{ l }with respect to the new basis matrices are

**z**

_{ j }= [

*v*

_{j1}, ...,

*v*

_{ jK }]

^{ T }and

**z**

_{ l }= [

*v*

_{l1},...,

*v*

_{ lK }]

^{ T }, respectively. Again, we use the square of the Euclidean distance between these two vectors to measure their distance

**Ê**and the distribution distance measure

*d*(

**z**

_{ j },

**z**

_{ l }), we can compute the smoothness of the proteins on the latent graph as follows

where Tr(·) denotes the trace of a matrix and **D** is a diagonal matrix whose entries are column sum of **Ê**, **D** = ∑_{
l
} **Ê**_{
jl
}. **L** = **D** - **Ê** is the graph Laplacian.

where *β* is the regularization parameter controlling the importance of the network regularization term.

**U**,

**B**and V for minimizing the objective function. In the following, we introduce an iterative algorithm which can achieve a local minimum for the objective function

*O*in Eq. (11). Using the matrix properties Tr(

**AB**) = Tr(

**BA**) and Tr(

**A**) = Tr(

**A**

^{ T }), the objective function can be rewritten as follows

*ψ*

_{ ik },

*γ*

_{ ik }and

*ϕ*

_{ jk }be the lagrange multiplier for constraint

*u*

_{ ik }≥ 0,

*b*

_{ ik }≥ 0 and

*v*

_{ jk }≥ 0, respectively. We need to minimize $\mathcal{O}$ with respect to

**U**,

**B**and V subject to the lagrange multiplier constraints. Then we have the Lagrange function $\mathcal{L}$ as follows

where Ψ = [*ψ*_{
ik
}], ϒ = [*γ*_{
ik
}] and Φ = [*ϕ*_{
jk
}].

*L*with respect to

**U**,

**B**and

**V**are

*ψ*

_{ ik }

*u*

_{ ik }= 0,

*γ*

_{ ik }

*b*

_{ ik }= 0 and

*ϕ*

_{ jk }

*v*

_{ jk }= 0, we have

when *α* = 0 and *β* = 0 the above updating rules reduce to the updating rules of the original NMF.

**Algorithm 1** RNMF

**Input:** data matrix **X**, label matrix **Y**, linkage weighting matrix **Ê**, label weighting matrix **W**

**Output:** new label matrix *Y '*

1: Initialize **V** using Eq.(23) and Eq.(25).

2: **repeat**

3: Update **U** using Eq.(20)

4: Update **B** using Eq.(21)

5: Update **V** using Eq.(22)

6: Reset **V** for labeled data using Eq.(25)

7: **until** stopping criteria is met

8: **for** each unlabeled protein *x*_{
i
} **do**

9: $\widehat{k}\leftarrow \mathsf{\text{arg}}\underset{k}{\mathsf{\text{max}}}\left({v}_{ik}\right)$

10: ${Y}^{\prime}\left(i,\widehat{k}\right)\leftarrow 1$

11: **end for**

**V**for the updates in Eq. (20-22) based on the class priors (using the labeled data). Specifically, for the labeled data we have

*v*

_{ jk }are initialized as

where *n*(*c*_{
k
}, *x*_{
i
}) = 1 if *x*_{
i
} is labeled as *c*_{
k
} and 0 otherwise.

The matrices **U**, **B** and **V** are then updated alternately until the objective value of Eq. (11) does not change or the maximum number of iterations is met (line 2-7). In this procedure, the values of *v*_{
jk
} of the labeled data are reset at each iteration to preserve the label information (line 6). In practice, only a small portion of entries of **V** will be reset when we have limited number of labeled data and do not affect the convergence of the algorithm as we see in the experiment section.

By alternatively updating the nonnegative matrices, we obtain a local optimum solution of the coefficient matrix **V**. In the following, we describe how to use **V** for protein function prediction. We specify the column dimension of the new representation $V=\left[{v}_{jk}\right]\in {\mathbb{R}}^{N\times K}$ of the original data with respect to the new basis as the same as the number of possible class labels *q*, i.e., we set *K* equals to *q*, each dimension of the new representation corresponds to one class label.

*x*

_{ j }function prediction, it is then assigned with the class with the largest

*v*

_{ jk }value, i.e.,

For multi-label protein *x*_{
j
} function prediction, we are primarily interested in learning a model that generate a ranking of possible labels for the given instance such that its correct labels receive higher ranking than the other irrelevant labels. The *v*_{
jk
} value give ranking of labels to indicate the importance of a set of labels associated with the instance. That is, the class label is ordered according to value of *v*_{
jk
} for each instance. A large value of *v*_{
jk
} has a high rank of the corresponding class label. If *v*_{
jk
}*' > v*_{
jk
}*"*, the label *k'* is considered to be ranked higher than the label *k"*. The model is then evaluated in terms of its ability to predict a good approximation of ranks for labels associated with the unlabeled instances.

Our proposed RNMF model is different from the other variants of NMF methods. Recently, various researchers have considered manifold learning in matrix factorization. For instance, Cai et al. [22] showed that adding manifold learning in matrix factorization will improve clustering performance substantially. But these NMF methods only deal with unsupervised modeling so far. They cannot directly used for supervised protein functional properties prediction problems where PPI interaction networks are involved. One hopes then to find a matrix factorization which uncovers the network structure and simultaneously respects the label information of the labeled data. In our RNMF model, a label matrix factorization term and a network regularization term are incorporated into the NMF model for this purpose.

## Experiments

In this section, we conduct extensive experiments to compare the performance of our proposed RNMF method with the other compared baselines: SVM, wvRN+RL, ICA, semi-ICA and ICML, and show that the proposed RNMF method is able to achieve better performance against these algorithms.

### Yeast dataset and baselines

We conduct experiments to predict properties of the proteins corresponding to a given yeast gene from KDD Cup 2001 [27] (available at http://www.kdd.org/kdd-cup-2001-molecular-bioactivity-plus-protein-locale-prediction). These properties are (1) the localization of the proteins encoded by the genes (2) one (or several) of categories of protein function(s). A protein can have more than one function, but only one localization. Problem (1) is a binary problem, i.e., proteins are localized (or not localized) to the corresponding organelle. Problem (2) is a multi-label problem with 14 functional classes, and we are primarily interested in learning a ranking of possible functions for the proteins.

The dataset for these two problems includes 1,243 protein instances and 1,806 interactions among the pair of proteins interact with one another. The protein features include the attributes refer to the chromosome on which the genes appears, to whether the gene is essential for survival, observable characteristics of the phenotype, structural category of the protein, the existence of characteristic motifs in the amino acid sequence of the protein, and whether the protein forms larger proteins with others [27, 5].

and problem (2) by two multi-label learning evaluation metrics *Coverage* and *RankingLoss* [28].

*Coverage*evaluates how far we need, on the average, to go down the list of labels in order to cover all the true labels of an instance:

where *ranks*(*x*_{
i
}, *c*_{
k
}) denotes the ranks of class label *c*_{
k
} derived from a confidence function *s*(*x*_{
i
}, *c*_{
k
}) which indicates the confidence for the class label *c*_{
k
} to be a proper label of *x*_{
i
}.

*Ranking loss*evaluates the average fraction of label pairs that are reversely ordered for the instance:

where *R*_{
i
} = {(*c*_{1}, *c*_{2})*|h*(*x*_{
i
}, *c*_{1}) ≤ *h*(*x*_{
i
}, *c*_{2}), (*c*_{1}, *c*_{2}) ∈ *Y*_{
i
} × *Ῡ*_{
i
}|}, and *Ῡ*_{
i
} denotes the complementary set of *Y*_{
i
}.

1 **SVM** [26]. This baseline is a feature-based method only using the attribute features of the proteins for learning without considering to use any network sources.

2 **wvRN+RL** [29]. This algorithm is a relational-only method only using the PPI network for prediction. wvRN+RL computes a new label distribution for an unlabeled node by averaging the current estimated distributions of its linked neighbors. This process is repeated until reaching the maximum iteration number.

3 **ICA** [15]. This denotes a collective classification algorithm which uses both attribute features and relational features to train a base classifier iteratively for prediction. The relational features are constructed based on the labels of neighbors. We use logistic regression (LR) as base classifier because prior works have found that LR to be superior to other classifiers such as naive bayes and *k*NN, as base classifier for ICA.

4 **semi-ICA** [30]. This method extends ICA to leverage the unlabeled data using semi-supervised learning. There are four semi-ICA variants (KNOWN-EM, ALL-EM, KNOWN-ONEPASS, ALL-ONEPASS) for semi-ICA, we run all four variants and choose the best one as the result of semi-ICA.

5 ICML [31]. This method extends ICA to handle multi-label learning by constructing additional label correlation features to exploit the dependencies among the labels as additional input features to learn base classifier.

It is generally more difficult to determine the classifier parameter values when the number of labeled data available is smaller. Learning from limited number of labeled data is the focus of this study. Thus, we do not tune the algorithm parameters using cross validation. In the experiments, we use default parameter values for the compared methods as recommended by previous works. In particular, we use the LibSVM (available at http://www.csie.ntu.edu.tw/ cjlin/libsvm/) library [26] with linear kernel as base classifier for the SVM algorithm, and set the penalty parameter *C* = 1.0 for the SVM as default. The maximum number of iterations for ICA, semi-ICA are set to 10 as in [30, 31]. While the wvRN+RL uses 1000 iterations. The parameters *α* and *β* for our proposed method are set to 10 and 5. The parameter selection will be discussed in the later section.

### Results on protein localization prediction

We first consider problem (1) of KDD Cup 2001, i.e., the protein localization prediction problem. We compare RNMF with the learning algorithms: SVM, wvRN+RN, ICA and semi-ICA. The performance is measured in classification accuracy.

Accuracy (mean ± standard deviation) of the compared algorithms against different label ratios on problem (1) of KDD Cup 2001.

label ratio | RNMF | SVM | wvRN+RL | ICA | semiICA |
---|---|---|---|---|---|

2% |
| 0.700 ± 0.044 | 0.633 ± 0.012 | 0.700 ± 0.058 | 0.725 ± 0.052 |

3% |
| 0.736 ± 0.004 | 0.624 ± 0.013 | 0.731 ± 0.063 | 0.755 ± 0.004 |

4% |
| 0.774 ± 0.005 | 0.650 ± 0.004 | 0.760 ± 0.052 | 0.774 ± 0.055 |

5% |
| 0.770 ± 0.003 | 0.675 ± 0.023 | 0.771 ± 0.058 | 0.792 ± 0.001 |

Avg. |
| 0.745 ± 0.014 | 0.645 ± 0.013 | 0.740 ± 0.057 | 0.762 ± 0.028 |

We further analyze the performance difference between RNMF and the compared methods and count the results of the win-tie-loss with pairwise t-tests at 0.10 significance level. The label ratios used are 2%, 3%, 4% and 5%. For each label ratio, a win (or loss) is counted when RNMF is significantly better (or worse) than the compared algorithm over 10 runs. Otherwise, a tie is recorded. We find that the win/tie/lose counts with pairwise t-test for RNMF against other algorithms are 4/0/0 over all comparisons. This result reveals that the RNMF method is statistically superior to other methods at 0.10 significance level when there is limited number of labeled data. This is consistent with our earlier assertions that our approach can work well in the paucity of labeled proteins.

*X*axis and true positive (tp) rate is plotted on the

*Y*axis. In classification evaluation, the classifier model produces a continuous output (i.e., an estimate of an instance's class membership probability) to which different thresholds are applied to predict class membership. If the classifier output is above the threshold, the classifier predicts the instance as class

*c*, else _

*c*. In this way, each threshold value produces a different prediction result to compute the results of tp and fp. Each of the thresholds corresponds to a different point in ROC space. The area under the ROC curve (the larger the better) is used to evaluate the strength of a classifier across various thresholds. Figure 2 shows the ROC curves of the RNMF method and the baselines (SVM and wvRN+RL). We see from the figure that the area under the ROC curve of our RNMF (the red curve) is larger than those of the SVM method (the blue curve) and the wvRN+RL method (the green curve), which implies that the RNMF method is able to deliver better performance against the baselines for protein localization prediction.

### Convergence study

*x*-axis is the number of iteration number in the process of optimizing the objective value $\mathcal{O}$ and the

*y*axis is the value of successive computed objective value $\left|\right|\phantom{\rule{0.3em}{0ex}}\mathcal{O}\left(t+\mathsf{\text{1}}\right)-\mathcal{O}\left(t\right)\phantom{\rule{0.3em}{0ex}}\left|\right|/\left|\right|\phantom{\rule{0.3em}{0ex}}\mathcal{O}\left(t\right)\phantom{\rule{0.3em}{0ex}}\left|\right|$. We observe that the algorithm converge after about 10 iterations.

### Parameter sensitivity

*α*and

*β*which quantify the importance of the label matrix factorization term and the network regularization term of the objective function in Eq. (11). In this experiment, we investigate how different values of the parameters

*α*and

*β*affect the classification accuracy of the proposed method. We examine the sensitivity of RNMF with respect to different

*α*and

*β*. (i) We fix

*β*= 5 and vary

*α*. Figure 4 shows the classification accuracy of RNMF against different values of

*α*on problem (1) of the KDD Cup 2001 dataset. From the figure, we observe that when

*α*is small the accuracy is poor because the RNMF algorithm boils down to an unsupervised NMF approach in this situation. The accuracy of the proposed RNMF method increases as the value of

*α*increases, and the accurate for

*α*between 5 to 60 does not change significantly. (ii) We fix

*α*= 10 and vary

*β*. Figure 5 shows the classification accuracy of RNMF against different values of

*β*. One observes that when

*β*is small, the classification accuracy is degraded, because no smoothness is used in this situation. As the parameter

*β*increases, the accuracy reaches a plateau between 5 to 60, and does not change significantly. In summary, the experimental results show that one can use the method in a robust way across a wide range of parameters. The best performance is achieved at

*α*= 10 and

*β*= 5. Therefore, we set

*α*= 10 and

*β*= 5 as default values in the experiments.

### Interaction relations

**V**learnt by the proposed RNMF method can be used to estimate the interaction effects among the proteins. Given two protein instances

*xi*and

*x*

_{ j }, their interaction can be estimated by the cosine similarity between their coefficient vectors

**v**

_{ i }= [

*v*

_{i,1},...,

*v*

_{i,K}] and

**v**

_{ j }= [

*v*

_{j,1}, ...,

*v*

_{j,K}]. The resulting similarity ranges from 0 to 1, with 0 indicating the instances are independence, and 1 indicating the instances are highly interrelated. We apply the cosine similarity measure to evaluate the interaction relations of 5 randomly selected genes (G238510, G238510, G234935, G235158, G237021, G234980) to their interrelated genes in the KDD Cup 2001 dataset. Table 2 shows these interrelated proteins (discovered by previous studies) and their similarity values (computed by using the matrix

**V**). In general, we can see that these interrelated genes tend to have large similarity values. This provides evidence of the advantages of using our proposed method to detect the interactions.

Selected interrelated genes and their similarity computed by the proposed method.

GeneID | GeneID | Similarity |
---|---|---|

G238510 | G239467 | 0.9984 |

G238510 | G239178 | 0.9987 |

G238510 | G235250 | 0.9983 |

G234935 | G234445 | 0.9094 |

G234935 | G239966 | 0.9388 |

G234935 | G235763 | 0.9589 |

G234935 | G235329 | 0.9700 |

G235158 | G234735 | 0.9776 |

G235158 | G234074 | 0.9808 |

G235158 | G234177 | 0.9837 |

G235158 | G235216 | 0.9554 |

G237021 | G234486 | 0.8831 |

G237021 | G234065 | 0.9222 |

G237021 | G239804 | 0.9285 |

G237021 | G239266 | 0.8751 |

G234980 | G235439 | 0.9865 |

G234980 | G235231 | 0.9843 |

G234980 | G234914 | 0.9939 |

G234980 | G235780 | 0.9305 |

### Results on protein function prediction

We also conduct experiments for problem (2) of KDD Cup 2001, i.e., the multi-label protein function prediction problem. We compare the proposed RNMF algorithms with baseline classifiers: SVM, wvRN+RN, ICA, semi-ICA and ICML. For SVM, wvRN+RN, ICA and semi-ICA, we use the binary relevance (BR) method [33] to decompose the multi-label problem into a set of *q* binary classification problems using one-against-all strategy, and train independent classifier for each single-label problem. The predictions for all *q* binary classification problems are combined to make the final prediction.

*Coverage*and

*RankingLoss*, respectively. For these two evaluation metrics, the smaller the value of the metrics, the better the performance of the algorithms. From the experimental results, we see that the RNMF method (the black line) has the best performance (lies under the other curves) across different percentages of labeled data from 2% to 10%. This provides evidence of the advantage of the proposed RNMF method for multi-label protein function prediction.

## Conclusion

In this paper, we utilize a latent graph approach for finding an integration latent graph by exploiting various latent linkages and judiciously integrate the linkages to generate a latent graph to effectively propagate the label information from labeled data to unlabeled data. For protein function prediction, we developed a novel method, called regularized non-negative matrix factorization (RNMF), to seek a matrix factorization which respect the attribute features, latent graph, and unlabeled data for classification prediction. In RNMF, a label matrix factorization term and a network regularization term are incorporated into the NMF objective function to encode the network structure and label information. As such, the learnt RNMF has more discriminating power than the other compared baseline methods.

Several questions remain to be investigated in our future work: 1) A challenge problem is the lack of large benchmark datasets for evaluation the scalability of the proposed method. Future work includes collecting and generating large datasets for more extensive empirical study. 2) Using suitable parameters are critical to the RNMF model and the compared methods. We will further investigate how to select the parameters effectively and efficiently with limited number of labeled data. 3) The convergence proofs of the RNMF can follow the idea in the proofs of Lee and Seung's paper [17] for the original NMF. It is interesting to apply Lee and Seung's idea [17] to theoretically proof the convergence of the RNMF model. 4) Advances in biotechnology have generated a wide variety of heterogeneous biology networks. This suggests investigating the performance of the RNMF model on different networks.

## Declarations

### Acknowledgements

This research was supported in part by National Key Technology R&D Program of MOST China under Grant No. 2012BAK17B08, and NSFC under Grant No. 61272538 and No. 61100190, and the Shenzhen Strategic Emerging Industries Program under Grant No. GJHS20120627112429515, and the Science and Technology Program of Shenzhen Polytechnic under Grant No. 2212K3190005 and 2213K3190016, and the cooperation Project of Industry, Education and Academy of Guangdong under Grant No. 2012B091120420, and the Key Laboratory of Guangdong under Grant No. 2013WSYS0002, and the Major Project on the General Administration of Press and Publications under Grant No. 2rjD9132490.

**Declarations**

Publication of this article was funded by the corresponding author.

This article has been published as part of *BMC Systems Biology* Volume 9 Supplement 1, 2015: Selected articles from the Thirteenth Asia Pacific Bioinformatics Conference (APBC 2015): Systems Biology. The full contents of the supplement are available online at http://www.biomedcentral.com/bmcsystbiol/supplements/9/S1

## Authors’ Affiliations

## References

- Pandey G, Kumar V, Steinbach M: Computational approaches for protein function prediction: A survey. Twin Cities: Department of Computer Science and Engineering, University of Minnesota. 2006Google Scholar
- Cai C, Han L, Ji ZL, Chen X, Chen YZ: Svm-prot: web-based support vector machine software for functional classification of a protein from its primary sequence. Nucleic acids research. 2003, 31 (13): 3692-3697. 10.1093/nar/gkg600.PubMed CentralView ArticlePubMedGoogle Scholar
- Mooney C, Pollastri G, et al: Sclpred: protein subcellular localization prediction by n-to-1 neural networks. Bioinformatics. 2011, 27 (20): 2812-2819. 10.1093/bioinformatics/btr494.View ArticlePubMedGoogle Scholar
- Wu Q, Ye Y, Liu Y, Ng MK: Snp selection and classification of genome-wide snp data using stratified sampling random forests. IEEE Transactions on NanoBioscience. 2012, 11 (3): 216-227.View ArticlePubMedGoogle Scholar
- Krogel M-A, Scheffer T: Multi-relational learning, text mining, and semi-supervised learning for functional genomics. Machine Learning. 2004, 57 (1-2): 61-81.View ArticleGoogle Scholar
- Sharan R, Ulitsky I, Shamir R: Network-based prediction of protein function. Molecular systems biology. 2007, 3 (1):Google Scholar
- Chua HN, Sung W-K, Wong L: Exploiting indirect neighbours and topological weight to predict protein function from protein-protein interactions. Bioinformatics. 2006, 22 (13): 1623-1630. 10.1093/bioinformatics/btl145.View ArticlePubMedGoogle Scholar
- Nabieva E, Jim K, Agarwal A, Chazelle B, Singh M: Whole-proteome prediction of protein function via graph-theoretic analysis of interaction maps. Bioinformatics. 2005, 21 (suppl 1): 302-310. 10.1093/bioinformatics/bti1054.View ArticleGoogle Scholar
- Deng M, Tu Z, Sun F, Chen T: Mapping gene ontology to proteins based on protein-protein interaction data. Bioinformatics. 2004, 20 (6): 895-902. 10.1093/bioinformatics/btg500.View ArticlePubMedGoogle Scholar
- Arnau V, Mars S, Marín I: Iterative cluster analysis of protein interaction data. Bioinformatics. 2005, 21 (3): 364-378. 10.1093/bioinformatics/bti021.View ArticlePubMedGoogle Scholar
- Adamcsek B, Palla G, Farkas IJ, Derényi I, Vicsek T: Cfinder: locating cliques and overlapping modules in biological networks. Bioinformatics. 2006, 22 (8): 1021-1023. 10.1093/bioinformatics/btl039.View ArticlePubMedGoogle Scholar
- Wu Q, Ye Y, Ng MK, Ho S-S, Shi R: Collective prediction of protein functions from protein-protein interaction networks. BMC bioinformatics. 2014, 15 (Suppl 2): 9-10.1186/1471-2105-15-S2-S9.View ArticleGoogle Scholar
- Wu Q, Ye Y, Ho S-S, Zhou S: Semi-supervised multi-label collective classification ensemble for functional genomics. BMC Genomics. (to appear)Google Scholar
- Shi R, Wu Q, Ye Y, Ho S-S: A generative model with network regularization for semi-supervised collective classification. SIAM International Conference on Data Mining (SDM14). (to appear)Google Scholar
- Neville J, Jensen D: Iterative classification in relational data. Proc. AAAI-2000 Workshop on Learning Statistical Models from Relational Data. 2000, 13-20.Google Scholar
- Lee DD, Seung HS: Learning the parts of objects by non-negative matrix factorization. Nature. 1999, 401 (6755): 788-791. 10.1038/44565.View ArticlePubMedGoogle Scholar
- Lee DD, Seung HS: Algorithms for non-negative matrix factorization. Advances in Neural Information Processing Systems. 2001, 556-562.Google Scholar
- Berry MW, Browne M, Langville AN, Pauca VP, Plemmons RJ: Algorithms and applications for approximate nonnegative matrix factorization. Computational Statistics & Data Analysis. 2007, 52 (1): 155-173. 10.1016/j.csda.2006.11.006.View ArticleGoogle Scholar
- Li SZ, Hou X, Zhang H, Cheng Q: Learning spatially localized, parts-based representation. Computer Vision and Pattern Recognition, 2001. CVPR 2001. Proceedings of the 2001 IEEE Computer Society Conference. 2001, 1: 207-Google Scholar
- Xu W, Liu X, Gong Y: Document clustering based on non-negative matrix factorization. Proceedings of the 26th Annual International ACM SIGIR Conference on Research and Development in Informaion Retrieval. 2003, 267-273.View ArticleGoogle Scholar
- Brunet J-P, Tamayo P, Golub TR, Mesirov JP: Metagenes and molecular pattern discovery using matrix factorization. Proceedings of the national academy of sciences. 2004, 101 (12): 4164-4169. 10.1073/pnas.0308531101.View ArticleGoogle Scholar
- Cai D, He X, Han J, Huang TS: Graph regularized nonnegative matrix factorization for data representation. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2011, 33 (8): 1548-1560.View ArticlePubMedGoogle Scholar
- Gu Q, Zhou J: Co-clustering on manifolds. Proceedings of the 15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 2009, 359-368.View ArticleGoogle Scholar
- Shi X, Li Y, Yu P: Collective prediction with latent graphs. Proceedings of the 20th ACM International Conference on Information and Knowledge Management. 2011, 1127-1136.Google Scholar
- Gallagher B, Tong H, Eliassi-Rad T, Faloutsos C: Using ghost edges for classification in sparsely labeled networks. Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 2008, 256-264.View ArticleGoogle Scholar
- Chang C-C, Lin C-J: Libsvm: a library for support vector machines. ACM Transactions on Intelligent Systems and Technology (TIST). 2011, 2 (3): 27-Google Scholar
- Cheng J, Hatzis C, Hayashi H, Krogel M-A, Morishita S, Page D, Sese J: Kdd cup 2001 report. ACM SIGKDD Explorations Newsletter. 2002, 3 (2): 47-64. 10.1145/507515.507523.View ArticleGoogle Scholar
- Madjarov G, Kocev D, Gjorgjevikj D, Dẑeroski S: An extensive experimental comparison of methods for multi-label learning. Pattern Recognition. 2012, 45 (9): 3084-3104. 10.1016/j.patcog.2012.03.004.View ArticleGoogle Scholar
- Macskassy SA, Provost F: Classification in networked data: A toolkit and a univariate case study. The Journal of Machine Learning Research. 2007, 8: 935-983.Google Scholar
- McDowell L, Aha D: Semi-supervised collective classification via hybrid label regularization. Proc. of the 29th International Conference on Machine Learning. 2012, 975-982.Google Scholar
- Kong X, Shi X, Yu PS: Multi-label collective classification. SIAM International Conference on Data Mining (SDM). 2011, 618-629.Google Scholar
- Huang J, Ling CX: Using auc and accuracy in evaluating learning algorithms. IEEE Transactions on Knowledge and Data Engineering. 2005, 17 (3): 299-310.View ArticleGoogle Scholar
- Zhang M-L, Zhou Z-H: A review on multi-label learning algorithms. IEEE Transactions on Knowledge and Data Engineering. 2013, 99 (PrePrints): 1-View ArticleGoogle Scholar

## Copyright

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.