Open Access

Affine Transformation Based Ontology Sparse Vector Learning Algorithm


Cite

Introduction

As a kind of information representation and shared model, ontology is introduced in nearly all fields of computer science. Acting as a concept semantic framework, ontology works high effectiveness and is widely employed in other engineering applications such as biology science, medical science, pharmaceutical science, material science, mechanical science and chemical science (for instance, see Coronnello et al. [2], Vishnu et al. [3], Roantree et al. [4], Kim and Park [5], Hinkelmann et al. [6], Pesaranghader et al. [7], Daly et al. [8], Agapito et al. [9], Umadevi et al. [10] and Cohen [11]).

The model of ontology can be regarded as a graph G = (V, E), in which each vertex v expresses a concept and each edge e = vivj represents a directly contact between two concepts vi and vj. The aim of ontology similarity calculating is to learn a similarity function Sim : V × V → ℝ+ ∪ {0} which maps each pair of vertices to a score real number. Moreover, the purpose of ontology mapping is to bridge the link between two or more different ontologies based on the similarity between their concepts. Using two graphs G1 and G2 to express two ontologies O1 and O2, respectively. The target is to determine a set SvV(G2) for each vG1 where the vertices in Sv are semantically high similarity to the concept corresponding to v. Hence, it may compute the similarity S(v, vj) for each vjV (G2) and select a parameter 0 < M < 1 for each vG1. Sv is set for vertex v and its element meets S(v, vj) ≥ M. From this perspective, the essence of ontology mapping is to yield a similarity function S and to determine a suitable parameter M according to detailed applications.

There are several effective learning tricks in ontology similarity measure and ontology mapping. Gao and Zhu [12] studied the gradient learning algorithms for ontology similarity computing and ontology mapping. Gao and Xu [13] obtained the stability analysis for ontology learning algorithms. Gao et al. [14] manifested an ontology sparse vector learning approach for ontology similarity measuring and ontology mapping based on ADAL trick. Gao et al. [15] researched an ontology optimization tactics according to distance calculating techniques. More theoretical analysis of ontology learning algorithm can be referred to Gao et al. [16].

In this paper, we propose a new ontology learning trick based on affine transformation. Furthermore, we present the efficiency of the algorithm in the biology and chemical applications via experiments.

Setting

Let V be an instance space. We use p dimension vector to express the semantics information of each vertex in ontology graph. Specifically, let v = {v1, ···, vp} be a vector corresponding to a vertex v. To facilitate the representation, we slightly confused the notations and v is used to represent both the ontology vertex and its corresponding vector.In the learning setting, the aim of ontology algorithms is to yield an vertices can be determined according to the difference between their corresponding real numbers. Obviously, the ontology function can be regarded as a dimensionality reduction operator f : ℝp → ℝ.

In recent years, the application of ontology algorithm faces many challenges. When it comes to the field of chemical and biology, the situation may become very complex since we need to deal with high dimensional data or big data. Under this background, sparse vector learning algorithms are introduced in biology and chemical ontology computation (see Afzali et al. [17], Khormuji, and Bazrafkan [18], Ciaramella and Borzi [19], Lorincz et al. [20], Saadat et al. [21], Yamamoto et al. [22], Lorintiu et al. [23], Mesnil and Ruzzene [24], Gopi et al. [25], and Dowell and Pinson [26] for more details). For example, if we aim to find what kind of genes causes a certain genetic disease, there are millions of genes in human’s bodies and the computation task is complex and tough. However, in fact, only a few classes of genes cause this kind of genetic disease. The sparse vector learning algorithm can effectively help scientists pinpoint genes in the mass disease genes.

One computational method of ontology function via sparse vector is expressed by

fw(v)=i=1pviwi+δ,$$\begin{array}{} \displaystyle {f_{\bf{w}}}(v) = \mathop \sum \limits_{i = 1}^p {v_i}{w_i} + \delta , \end{array}$$

where w = {w1, ···, wp} is a sparse vector which is used to shrink irrelevant component to zero and δ is a noise term. Using this model, the key to determine the ontology function f is to learn the optimal sparse vector w.

For example, the standard framework with the penalize term via the l1 -norm of the unknown sparse vector w ∈ ℝp can be stated as:

Yw=l(w)+λw1,$$\begin{array}{} \displaystyle {Y_{\bf{w}}} = l({\bf{w}}) + \lambda {\left\| {\bf{w}} \right\|_1}, \end{array}$$

where λ > 0 is a balance parameter and l is the principal function to measure the error of w. The balance term λ ||w||1 is used to measure the sparsity of sparse vector w.

Algorithm description

Let D be a (p − 1) × p matrix which is denoted as

D=(11000011000011000011)$$\begin{array}{} \displaystyle {\bf{D}} = \left( {\begin{array}{*{20}{c}} 1&{ - 1}&0& \cdots &0&0\\ 0&1&{ - 1}& \cdots &0&0\\ \vdots & \vdots & \ddots & \ddots & \vdots & \vdots \\ 0&0& \cdots &1&{ - 1}&0\\ 0&0& \cdots &0&1&{ - 1} \end{array}} \right) \end{array}$$

One general ontology sparse vector learning framework can be stated as

minwRp12yVw+ρ2w2+λDw1,$$\begin{array}{} \displaystyle \mathop {\min }\limits_{\bf{w} \in {\mathbb{R}^p}} \frac{1}{2}\left\| {{\bf{y}} - {\bf{Vw}}} \right\| + \frac{\rho }{2}{\left\| {\bf{w}} \right\|^2} + \lambda {\left\| {{\bf{Dw}}} \right\|_1}, \end{array}$$

where ρ ≥ 0 is also a balance parameter. Let V˜=(VρI)$\begin{array}{} \displaystyle {\bf{\tilde V}} = (\begin{array}{*{20}{c}} {\bf{V}} \\ {\sqrt {\rho {\bf{I}}} } \\ \end{array}) \end{array}$ and y˜=(y0)$\begin{array}{} \displaystyle {\bf{\tilde y}} = (\begin{array}{*{20}{c}} {\bf{y}} \\ 0 \\ \end{array}) \end{array}$. Then ontology sparse vector learning problem (4) can be expressed as

minwp12y˜V˜w22+λDw1.$$\begin{array}{} \displaystyle \mathop {\min }\limits_{{\bf{w}} \in {\mathbb{R}^p}} \frac{1}{2}\left\| {{\bf{\tilde y}} - {\bf{\tilde Vw}}} \right\|_2^2 + \lambda {\left\| {{\bf{Dw}}} \right\|_1}. \end{array}$$

An effective method to get the solution is to set variables β = Dw, then it becomes

minwRp,βRp1=12||y˜V˜w||22+λ||β||1,β=Dw.$$\begin{array}{} \displaystyle \begin{array}{*{20}{c}} {\begin{array}{*{20}{c}} {{\rm{min}}} \\ {{\bf{w}} \in {\mathbb{R}^p},\beta \in {\mathbb{R}^{p - 1}}} \\ \end{array} = \frac{1}{2}\left\| {{\bf{\tilde y}} - {\bf{\tilde Vw}}} \right\|_2^2 + \lambda {{\left\| \beta \right\|}_1},} \\ {\beta = {\bf{Dw}}.} \\ \end{array} \end{array}$$

By derivation, the Lagrange dual problem of above ontology problem is formulated as

minup112(V˜Ty˜DTu)T(V˜TV˜)(V˜Ty˜DTu),uλ,DTuspan(V˜T).$$\begin{array}{c} \min\limits_{u\in\mathbb{R}^{p-1}}\frac{1}{2}(\tilde{{\bf V}}^{T}\tilde{{\bf y}}-{\bf D}^{T}{\bf u})^{T}(\tilde{{\bf V}}^{T}\tilde{{\bf V}})^{\dagger}(\tilde{{\bf V}}^{T}\tilde{{\bf y}}-{\bf D}^{T}{\bf u}),\\ \|u\|_{\infty}\le\lambda,{\bf D}^{T}{\bf u}\in {\rm span}(\tilde{{\bf V}}^{T}). \end{array}$$

The other version of ontology framework can be simply expressed as

minw12||yVw||+λ||w||1.$$\begin{array}{} \displaystyle {\min\limits _{\bf{w}}}\frac{1}{2}{\bf{||y}} - {\bf{Vw}}|| + \lambda {\bf{||w||}}{_1}. \end{array}$$

And, the ontology dual problem of (6) can be written as

supθ12y22λ22yλθ2,s.t.|viTθ|1$$\begin{array}{} \displaystyle \begin{array}{*{20}{c}} {\mathop {\sup }\limits_\theta \frac{1}{2}\left\| {\bf{y}} \right\|_2^2 - \frac{{{\lambda ^2}}}{2}{{\left\| {\frac{{\bf{y}}}{\lambda } - \theta } \right\|}^2},} \\ {s.t.\quad |\upsilon _i^T\theta | \le 1} \\ \end{array} \end{array}$$

for any i ∈ {1,2, · · · , p}. Moreover, ontology problem (7) is equal to the following problem:

minθ12yλθ2,s.t.|υiTθ|1$$\begin{array}{} \displaystyle \begin{array}{*{20}{l}} {{\min\limits_\theta }\frac{1}{2}{{\left\| {\frac{{\bf{y}}}{\lambda } - \theta } \right\|}^2},} \\ {{\rm{s}}{\rm{.t}}.\quad |\upsilon _i^T\theta | \le 1} \\ \end{array} \end{array}$$

for any i ∈ {1,2, · · · ,p}.

Set D={θ:|viTθ|1,i{1,2,,p}}$\begin{array}{} \displaystyle {\scr D} = \left\{ {\theta :\left| {v_i^T\theta } \right| \le 1,i \in \left\{ {1,2, \cdots ,p} \right\}} \right\} \end{array}$ as the feasible set of ontology problem (8). Obviously, 𝒟 can be regarded as the intersection of a collection of closed half spaces which is a closed convex set, and 𝒟 ≠ ∅ since 0 ∈ 𝒟 .By means of (8), the projection of yλ$\begin{array}{} \displaystyle \frac{{\rm{y}}}{\lambda } \end{array}$ onto 𝒟 is the dual optimal solution θ* which is stated as θ*=PDyλ$\begin{array}{} \displaystyle {\theta ^*} = {\mathbb{P}_{\scr D}}\frac{{\bf{y}}}{\lambda } \end{array}$.

Next, we present our dual framework of ontology problem which can be formulated as a problem of projection. Set

β=y˜V˜w,$$\begin{array}{} \displaystyle \beta = {\bf{\tilde y}} - {\bf{\tilde Vw}}, \end{array}$$

then the ontology problem (5) becomes

minwp,βn12β2+λDw1,s.t.β=y˜V˜w.$$\begin{array}{} \displaystyle \begin{array}{*{20}{c}} {\mathop {\min }\limits_{\bf{w} \in {\mathbb{R}^p},\beta \in {\mathbb{R}^n}} \frac{1}{2}{{\left\| \beta \right\|}^2} + \lambda {{\left\| {{\bf{Dw}}} \right\|}_1},} \\ {s.t.\quad \beta = {\bf{\tilde y}} - {\bf{\tilde Vw}}.} \\ \end{array} \end{array}$$

Set λ θN as the Lagrangian multipliers. Hence, the Lagrangian of ontology problem (5) can be expressed as

L(w,β;θ)=12β2+λDw1+λ<θ,y˜V˜wβ>=12β2λ<θ,β>+λDw1λ<θ,V˜w>+λ<θ,y˜>.$$\begin{array}{} \displaystyle \begin{array}{*{20}{c}} {L({\bf{w}},\beta ;\theta )} \hfill & { = \frac{1}{2}{{\left\| \beta \right\|}^2} + \lambda {{\left\| {{\bf{Dw}}} \right\|}_1} + \lambda < \theta ,{\bf{\tilde y}} - {\bf{\tilde Vw}} - \beta > } \hfill \\ {} \hfill & { = \frac{1}{2}{{\left\| \beta \right\|}^2} - \lambda < \theta ,\beta > + \lambda {{\left\| {{\bf{Dw}}} \right\|}_1} - \lambda < \theta ,{\bf{\tilde Vw}} > + \lambda < \theta ,{\bf{\tilde y}} > .} \hfill \\ \end{array} \end{array}$$

Set g1(β)=12β||2λ<θ,β>$\begin{array}{} \displaystyle {g_1}(\beta ) = \frac{1}{2}{\left\| \beta \right\|^2} - \lambda < \theta ,\beta > \end{array}$ and g2(w)=λDw1λ<θ,V˜w>$\begin{array}{} \displaystyle {g_2}({\bf{w}}) = \lambda {\left\| {{\bf{Dw}}} \right\|_1} - \lambda < \theta ,{\bf{\tilde Vw}} > \end{array}$. Since g1(β) is a quadratic, we deduce

Δg1(β)=0λθ=βminβg1(β)=λ22θ2.$$\begin{array}{} \displaystyle \Delta {g_1}(\beta ) = 0 \to \lambda \theta = \beta \to \mathop {\min }\limits_\beta {g_1}(\beta ) = - \frac{{{\lambda ^2}}}{2}{\left\| \theta \right\|^2}. \end{array}$$

Furthermore, we infer

0g2(w)minwg2(w)=0.$$\begin{array}{} \displaystyle 0 \in \partial {g_2}({\bf{w}}) \to \mathop {\min }\limits_{\bf{w}} {g_2}({\bf{w}}) = 0. \end{array}$$

In terms of (10) and (11), we get the following equal ontology optimization version:

minwp,βnL(w,β;θ)=λ22θ2+λ<θ,y˜>=12y˜2λ22y˜λθ2.$$\begin{array}{} \displaystyle \mathop {\min }\limits_{\bf{w} \in {\mathbb{R}^p},\beta \in {\mathbb{R}^n}} L({\bf{w}},\beta ;\theta ) = - \frac{{{\lambda ^2}}}{2}{\left\| \theta \right\|^2} + \lambda < \theta ,{\bf{\tilde y}} > = \frac{1}{2}{\left\| {{\bf{\tilde y}}} \right\|^2} - \frac{{{\lambda ^2}}}{2}{\left\| {\frac{{{\bf{\tilde y}}}}{\lambda } - \theta } \right\|^2}. \end{array}$$

Now, we discuss the deep ontology optimization model in view of affine transformation. Set Θ1 ∈ ℝp × p as

Θ1=(1111011100110001)$$\begin{array}{} \displaystyle {\Theta _1} = \left( {\begin{array}{*{20}{c}} 1&1&1& \cdots &1\\ 0&1&1& \cdots &1\\ 0&0&1& \cdots &1\\ \vdots & \vdots & \ddots & \ddots & \vdots \\ 0&0&0& \cdots &1 \end{array}} \right) \end{array}$$

and V¯=(v¯1,,v¯p)=V˜Θ1$\begin{array}{} \displaystyle {\bf{\bar V}} = ({{\bf{\bar V}}_1}, \cdots ,{{\bf{\bar V}}_p}) = {\bf{\tilde V}}{\Theta _1} \end{array}$. Then, we obtain v¯i=Σk=1iv˜k$\begin{array}{} \displaystyle {{\bf{\bar v}}_i} = \Sigma _{k = 1}^i{\tilde v_k} \end{array}$ for any i ’ {1, · · · , p}. Thus, the dual problem of the ontology sparse vector problem (5) becomes

maxθn=12y˜2λ22y˜λθ2,s.t.|v¯iTθ|1,,i{1,,p1},v¯pTθ=0.$$\begin{array}{} \displaystyle \begin{array}{*{20}{c}} {\mathop {\max }\limits_{\theta \in {\mathbb{R}^n}} = \frac{1}{2}{{\left\| {{\bf{\tilde y}}} \right\|}^2} - \frac{{{\lambda ^2}}}{2}{{\left\| {\frac{{{\bf{\tilde y}}}}{\lambda } - \theta } \right\|}^2},} \\ {{\rm{s}}{\rm{.t}}.\quad |{\bf{\bar v}}_i^T\theta | \le 1,\quad ,i \in \{ 1, \cdots ,p - 1\} ,} \\ {{\bf{\bar v}}_p^T\theta = 0.} \\ \end{array} \end{array}$$

In addition, the ontology problem (14) has the equal solution with the following ontology optimization problem

minθn=12y˜λθ2s.t.|v¯iTθ|1,,i{1,,p1},v¯pTθ=0.$$\begin{array}{} \begin{array}{*{20}{c}} {\mathop {\min }\limits_{\theta \in {\mathbb{R}^n}} = \frac{1}{2}{{\left\| {\frac{{{\bf{\tilde y}}}}{\lambda } - \theta } \right\|}^2}} \\ {{\rm{s}}{\rm{.t}}{\rm{.}}\quad |{\bf{\bar v}}_i^T\theta | \le 1,\quad ,i \in \{ 1, \cdots ,p - 1\} ,} \\ {{\bf{\bar v}}_p^T\theta = 0.} \\ \end{array} \end{array}$$

Set H¯$\begin{array}{} \bar {\scr H} \end{array}$ as the feasible set of ontology problem (15), we yield

H¯={θ:|v¯iTθ|1,,i{1,,p1},v¯pTθ=0}.$$\begin{array}{} \bar {\scr H} = \left\{ {\theta :\left| {\overline {\bf{v}} _i^T\theta } \right| \le 1,\quad ,i \in \left\{ {1, \cdots ,p - 1} \right\},\overline {\bf{v}} _p^T\theta = 0} \right\}. \end{array}$$

It is not hard to check that the dual optimal solution of ontology problem is the projection of y˜λ$\begin{array}{} \frac{{{\bf{\tilde y}}}}{\lambda } \end{array}$ onto H¯$\begin{array}{} \bar {\scr H} \end{array}$.

In the following contexts, we show the equivalent optimization model of our ontology sparse vector problem. Our discussion can be divided into two cases according to whether the value of v¯p$\begin{array}{} {{\bf{\bar v}}_p} \end{array}$ equals to zero or not.

If v¯p=0$\begin{array}{} \displaystyle {{\bf{\bar v}}_p} = 0 \end{array}$, then we can skip the condition v¯pTθ=0$\begin{array}{} \displaystyle {\rm{\bar v}}_p^T\theta = 0 \end{array}$ and the ontology framework can be reduced to

minθn=12y˜λθ2s.t.|v¯iTθ|1,,i{1,,p1}.$$\begin{array}{} \displaystyle \begin{array}{*{20}{c}} {\mathop {\min }\limits_{\theta \in {\mathbb{R}^n}} = \frac{1}{2}{{\left\| {\frac{{{\bf{\tilde y}}}}{\lambda } - \theta } \right\|}^2}} \\ {{\rm{s}}{\rm{.t}}.\quad |{\bf{\bar v}}_i^T\theta | \le 1,\quad ,i \in \{ 1, \cdots ,p - 1\} .} \\ \end{array} \end{array}$$

If v¯p0$\begin{array}{} \displaystyle {\overline {\bf{v}} _p} \ne 0 \end{array}$, we set U={αv¯p:αR},$\begin{array}{} \displaystyle {\scr U} = \left\{ {\alpha {{\overline {\bf{v}} }_p}:\alpha \in } \right\}, \end{array}$, V={x:xTx¯p=0},$\begin{array}{} \displaystyle {\scr V} = \left\{ {{\bf{x}}:{{\bf{x}}^T}{{\overline {\bf{x}} }_p} = 0} \right\}, \end{array}$, Θ2=Iv¯pv¯pT||v¯p||2$\begin{array}{} \displaystyle {\Theta _2} = {\bf{I}} - \frac{{{{{\rm{\bar v}}}_p}{\rm{\bar v}}_p^T}}{{|\left| {{{{\rm{\bar v}}}_p}} \right|{|^2}}} \end{array}$, y˜=Θ2y˜=y˜v¯pTy˜||v¯p||2$\begin{array}{} \displaystyle {{\bf{\tilde y}}^ \bot } = {{\rm{\Theta }}_2}{\bf{\tilde y}} = {\bf{\tilde y}} - \frac{{{\bf{\bar v}}_p^T{\bf{\tilde y}}}}{{||{{{\bf{\bar v}}}_p}|{|^2}}} \end{array}$, v¯i=Θ2vi˜=vi˜v¯pTvi¯||v¯p||2$\begin{array}{} \displaystyle {\rm{\bar v}}_i^ \bot = {{\rm{\Theta }}_2}\mathop {{{\rm{v}}_i}}\limits^ = \mathop {{{\rm{v}}_i}}\limits^ - \frac{{{\rm{\bar v}}_p^T\overline {{{\rm{v}}_i}} }}{{|\left| {{{{\rm{\bar v}}}_p}} \right|{|^2}}} \end{array}$, for i ∈ {1, · · · , p — 1}. We get <y˜,v¯p>=0$\begin{array}{} \displaystyle < {{\bf{\tilde y}}^ \bot },{{\bf{\bar v}}_p} > = 0 \end{array}$ and <θ,v¯p>=0$\begin{array}{} \displaystyle \lt {\theta ,{{\overline {\bf{v}} }_p}} \gt = 0 \end{array}$ for any θ ∈ ℋ according to the fact that Θ2 is a projection operator onto the linear subspace 𝒱 which is the orthogonal complement subspace of 𝒰. Therefore, for any θ ∈ ℋ, we yield <y˜λθ,v¯p>=0$\begin{array}{} \displaystyle < \frac{{{{{\bf{\tilde y}}}^ \bot }}}{\lambda } - \theta ,{{\bf{\bar v}}_p} > = 0 \end{array}$ and further y˜λθ2=y˜λθ2+(v¯pTy˜)2λ2v¯p2$\begin{array}{} \displaystyle {\left\| {\frac{{{\bf{\tilde y}}}}{\lambda } - \theta } \right\|^2} = {\left\| {\frac{{{{{\bf{\tilde y}}}^ \bot }}}{\lambda } - \theta } \right\|^2} + \frac{{{{({\bf{\bar v}}_p^T{\bf{\tilde y}})}^2}}}{{{\lambda ^2}{{\left\| {{{{\bf{\bar v}}}_p}} \right\|}^2}}} \end{array}$.

Therefore, our ontology problem (15) can be expressed as

minθRn=12y~λθ2+12(v¯pTy~)2λ2v¯p2s.t.|v¯iTθ|1,,i{1,,p1},v¯pTθ=0.$$\begin{array}{} \displaystyle \begin{array}{*{20}{c}} {{\min\limits_{\theta \in {\mathbb{R}^n}}} = \frac{1}{2}{{\left\| {\frac{{{{{\bf{\tilde y}}}^ \bot }}}{\lambda } - \theta } \right\|}^2} + \frac{1}{2}\frac{{{{({\bf{\bar v}}_p^T{\bf{\tilde y}})}^2}}}{{{\lambda ^2}{{\left\| {{{{\bf{\bar v}}}_p}} \right\|}^2}}}} \\ {{\rm{s}}{\rm{.t}}.\quad |{\bf{\bar v}}_i^T\theta | \le 1,\quad ,i \in \{ 1, \cdots ,p - 1\} ,} \\ {{\bf{\bar v}}_p^T\theta = 0.} \\ \end{array} \end{array}$$

Note that the second term in (18) doesn’t rely on θ, the ontology problem (18) can be determined by

minθRn=12y~λθ2s.t.|v¯iTθ|1,,i{1,,p1},v¯pTθ=0.$$\begin{array}{} \displaystyle \begin{array}{*{20}{c}} {{\min\limits_{\theta \in {\mathbb{R}^n}}} = \frac{1}{2}{{\left\| {\frac{{{{{\bf{\tilde y}}}^ \bot }}}{\lambda } - \theta } \right\|}^2}} \\ {{\rm{s}}{\rm{.t}}{\rm{.}}\quad |{\bf{\bar v}}_i^T\theta | \le 1,\quad ,i \in \{ 1, \cdots ,p - 1\} ,} \\ {{\bf{\bar v}}_p^T\theta = 0.} \\ \end{array} \end{array}$$

Let H={θ:|<v¯i,θ>|1,i{1,,p1},v¯iTθ=0}$\begin{array}{} \displaystyle {{\scr H}^ \bot } = \left\{ {\theta :\left| {\lt {{\rm{\bar v}}_i^ \bot ,\theta } \gt } \right| \le 1,i \in \left\{ {1, \cdots ,p - 1} \right\},{\rm{\bar v}}_i^T\theta = 0} \right\} \end{array}$. We can derive that ℋ = ℋ. Thus, the ontology problem (19) can be further stated as

minθRn=12y~λθ2,s.t.|<v¯i,θ>|1,,i{1,,p1},v¯pTθ=0.$$\begin{array}{} \displaystyle \begin{array}{*{20}{c}} {\min\limits_{\theta \in {\mathbb{R}^n}} = \frac{1}{2} \in {{\left\| {\frac{{{{{\bf{\tilde y}}}^ \bot }}}{\lambda } - \theta } \right\|}^2},} \\ {{\rm{s}}{\rm{.t}}.\quad | < {\bf{\bar v}}_i^ \bot ,\theta \gt | \le 1,\quad ,i \in \{ 1, \cdots ,p - 1\} ,} \\ {{\bf{\bar v}}_p^T\theta = 0.} \\ \end{array} \end{array}$$

It implies that the dual optimal solution of ontology problem is the projection of y˜λ$\begin{array}{} \displaystyle \frac{{{{{\bf{\tilde y}}}^ \bot }}}{\lambda } \end{array}$ onto the feasible set . Finally, we have the final version of ontology sparse vector learning problem which has the same optimal solution with ontology problem (20):

minθRn=12y~λθR2,s.t.|<v¯i,θ>|1,,i{1,,p1}.$$\begin{array}{} \displaystyle \begin{array}{*{20}{c}} {\min\limits_{\theta \in {\mathbb{R}^n}} = \frac{1}{2} \|\frac{{{{{\bf{\tilde y}}}^ \bot }}}{\lambda } - \theta \|{\mathbb{R}^2},} \\ {s.t.\quad |\lt {\bf{\bar v}}_i^ \bot ,\theta \gt | \le 1,\quad ,i \in \{ 1, \cdots ,p - 1\} .} \\ \end{array} \end{array}$$

The feasible set of ontology problem (21) is stated as

H^={θ:|v¯i,θ>|1,   ,i{1,....,p1}}.$$\begin{array}{} \displaystyle \hat {\scr H} = \{ \theta :|{\bf{\bar v}}_i^ \bot ,\theta \gt | \le 1,{\rm{ }},i \in \{ 1,....,p - 1\} \} . \end{array}$$

Experiments

In this section, we test the feasibility of our new algorithm via the following four simulation experiments related to ontology similarity measure and ontology mapping below. After obtaining the sparse vector w, the ontology function is given by fw(v)=i=1pviwi$\begin{array}{} \displaystyle {f_{\bf{w}}}(v) = \sum\nolimits_{i = 1}^p {} {v_i}{w_i} \end{array}$ in which we ignore the noise term.

Ontology similarity measure experiment on biology data

In biology science, “GO” ontology (denoted by O1 which was constructed in http: //www. geneontology. org, and Fig. 1 presents the basic structure of O1) is a widely used database for gene researchers. Now, we apply this data set for our first experiment. We use P@N (Precision Ratio, see Craswell and Hawking [27] for more details) to measure the effectiveness of the experiment. In the first step, the closest N concepts (have highest similarity) for each vertex was deduced by the expert. Then, in the second step, the first N concepts for each vertex on ontology graph are determined by the algorithm and the precision ratios are obtained. In addition to our ontology learning algorithm, programming from Huang et al. [29], Gao and Liang [30] and Gao et al. [16] are employed to “GO” ontology, and the precision ratios which we inferred from these tricks are compared. Partial experiment results can be referred to Tab. 1.

Fig. 1

The Structure of “GO” Ontology

The experiment results of ontology similarity measure

P@3 average precision ratioP@5 average precision ratioP@10 average precision ratioP@20 average precision ratio
Algorithm in our paper0.47620.55040.67310.7918
Algorithm in Huang et al. [29]0.46380.53480.62340.7459
Algorithm in Gao and Liang [30]0.43560.49380.56470.7194
Algorithm in Gao et al. [16]0.42130.51830.60190.7239

From Fig. 1, take N = 3,5,10 or 20, the precision ratio in terms of our sparse vector ontology learning algorithm is higher than the precision ratio computed by algorithms by Huang et al. [29], Gao and Liang [30] and Gao et al. [16]. Specially, such precision ratios apparently increase as N increases. Thus, one result can be concluded that the ontology learning algorithm described in our paper is superior to that proposed by Huang et al. [29], Gao and Liang [30] and Gao et al. [16].

Ontology mapping experiment on physical data

Physical ontologies O2 and O3 (the structures of O2 and O3 can refer to Fig. 2 and Fig. 3, respectively) are used for our second experiment which aims to test the utility of ontology mapping. The ontology mapping between O2 and O3 are determined by means of our new ontology learning algorithm and P@N criterion is applied as well to test the equality of the experiment. Huang et al. [29], Gao and Liang [30] and Gao et al. [31] also employed ontology algorithms to “Physical” ontology, and we made a comparison among the precision ratios which we get from four methods. Several experiment results can be referred to Tab. 2.

Fig. 2

“Physical” ontology O2

Fig. 3

“Physical” ontology O3

The experiment results of ontology mapping

P@1 average precision ratioP@3 average precision ratioP@5 average precision ratio
Algorithm in our paper0.69130.77420.9161
Algorithm in Huang et al. [29]0.61290.73120.7935
Algorithm in Gao and Liang [30]0.69130.75560.8452
Algorithm in Gao et al. [31]0.67740.77420.8968

It can be seen that our algorithm is more efficient than ontology learning algorithms raised in Huang et al. [29], Gao and Liang [30] and Gao et al. [31] in particular when N is sufficiently large.

Ontology similarity measure experiment on plant data

In this part, “PO” ontology O4 (which was constructed in http: //www.plantontology.org. Fig. 4 shows the basic structure of O4) is used to test the efficiency of our new ontology learning algorithm for ontology similarity calculating. This ontology is famous in plant science which can be used as a dictionary for scientists to learn and search concepts and botanical features. P@N standard is used again for this experiment. Furthermore, ontology learning approaches in Wang et al. [28], Huang et al. [29] and Gao and Liang [30] are borrowed to the “PO” ontology in our experiment for comparison requirements. The accuracy by these ontology learning algorithms are computed and parts of the results are compared and presented in Tab. 3.

The experiment results of ontology similarity measure

P@3 average precision ratioP@5 average precision ratioP@10 average precision ratio
Algorithm in our paper0.48650.60520.7393
Algorithm in Wang et al. [28]0.45490.51170.5859
Algorithm in Huang et al. [29]0.42820.48490.5632
Algorithm in Gao and Liang [30]0.48310.56350.6871

It’s revealed in the Tab. 3 that the precision ratio in view of our ontology sparse vector learning algorithm is higher than the precision ratio proposed by ontology learning algorithms that Wang et al. [28], Huang et al. [29] and Gao and Liang [30] when N=3, 5 or 10. Furthermore, such precision ratios are increasing apparently as N increases. Therefore, we can conclude that the ontology sparse vector learning algorithm described in our paper is superior to the trick recommended in Wang et al. [28], Huang et al. [29] and Gao and Liang [30].

Fig. 4

The Structure of “PO” Ontology.

Ontology mapping experiment on humanoid robotics data

Humanoid robotics ontologies (denoted by O5 and O6, constructed by Gao and Zhu [12], and the structures of O5 and O6 can refer to in Fig. 5 and Fig. 6 respectively) are employed for our last experiment. Humanoid robotics ontologies are used to orderly and clearly express the humanoid robotics, and this experiment aims to determine ontology mapping between O5 and O6. Again, we use P@N criterion to measure the accuracy of data gotten in the experiment. Beside our ontology learning algorithm, ontology algorithms raised in Gao and Lan [32], Gao and Liang [30] and Gao et al. [31] are also applied on humanoid robotics ontologies, and the precision ratios which are obtained from four ontology learning algorithms are compared. Partial experiment results can refer to Tab. 4.

Fig. 5

“Humanoid Robotics”ontology O5

Fig. 6

“Humanoid Robotics “ ontology O6

The experiment results of ontology mapping

P@1 average precision ratioP@3 average precision ratioP@5 average precision ratio
Algorithm in our paper0.27780.50000.6556
Algorithm in Gao and Lan [32]0.27780.48150.5444
Algorithm in Gao and Liang [30]0.22220.40740.4889
Algorithm in Gao et al. [31]0.27780.46300.5333

The experiment results presented in Table 4 imply that our ontology sparse vector learning algorithm works with more efficiency than other ontology learning algorithms obtained in Gao and Lan [32], Gao and Liang [30] and Gao et al. [31] especially when N is sufficiently large.

Conclusion

In our paper, an affine transformation based computation technology is considered and presented to the readers. This ontology technology is suitable for biological and chemical ontology engineering applications because of its similarity measure and ontology mapping. The main approach is based on affine transformation and its theoretical derivation. At last, simulation data show that our ontology scheming has high efficiency in biology, physics, plant and humanoid robotics fields. The ontology sparse vector learning algorithm raised in our paper illustrates the promising application prospects in multiple disciplines.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

eISSN:
2444-8656
Language:
English
Publication timeframe:
2 times per year
Journal Subjects:
Life Sciences, other, Mathematics, Applied Mathematics, General Mathematics, Physics