Article Preview
TopIntroduction
Ontologies are the cornerstone of the semantic web. They provide representing, sharing and reuse of knowledge as a communication tool for applications developed in different ways. An ontology is an explicit description of the concepts, properties, relationships and individuals that may exist for a particular domain. It reflects knowledge from a certain domain of discourse (Zamazal, 2020). However, most applications require access to multiple ontologies. In addition, the construction of ontologies by various experts leads to heterogeneity at different levels, due to the rapid development of the semantic web.
Therefore, it is very interesting to identify correspondences between semantically related entities of heterogeneous ontologies. That allows agents using different ontologies to inter-operate. These correspondences, called alignment or mapping, are the backbone of the ontology matching task, which is the promising solution to this ontology semantic heterogeneity problem. In addition, automatic and semi-automatic ontology matching techniques should be developed in order to reduce the burden of its manual creation and maintenance (Khiat & Benaissa, 2015).
Likewise, ontologies of most applications (like in medicine and astronomy) are of big size. And, large ontologies include a high conceptual heterogeneity. That could decrease the efficiency of ontology matching systems facing other challenges as shortage of memory and long-time processing. Such issue makes scaling up the ontology matching process a very interesting problem.
Deep learning techniques have been recently used to address important problems in many research axes, such as image processing, natural language processing, information retrieval, signal processing and many others as in (Gupta et al., 2019; Sedik et al., 2021; Al-Smadi et al., 2018). Sophisticated artificial intelligence systems use deep learning to solve computational tasks and complex problems quickly (Fiorini, 2020). These techniques are very appropriate for dealing with large datasets. They have the ability to analyse and interpret massive amounts of data, that require efficient and effective computational tools.
Although deep learning techniques are very appropriate for dealing with large datasets, they have limited use in ontology matching. Even the few approaches that employ these computational models aim at enhancing the performance of the ontology matching task, and not at handling the large-scale heterogeneity problem (Portisch & Paulheim, 2018; Chang et al., 2019; Hertling & Paulheim, 2018; Monych et al., 2020; Roussille et al., 2018). Then, they tested their methods on ontologies of small sizes. The commonly promising solution for dealing with the large-scale ontology matching issue is partitioning (Tran et al., 2012; Laadhar et al., 2019; Laadhar et al., 2018; Jiménez-Ruiz et al., 2018; Balachandran et al., 2019). It consists on dividing the input ontologies into several sub-ontologies. The overall result of matching is obtained after combining the individual results of matching sub-ontologies. However, the partitioning phase is also challenging; The method of dividing input ontologies, the number of resulted partitions, sizes of partitions, and all parameters of the partitioning process are delicate to define as well. Moreover, several semantic links inside ontologies are expected to be lost when dividing input ontologies. That affects the matching quality.