PreprintPDF Available
Preprints and early-stage research may not have been peer reviewed yet.

Abstract

This paper presents the learning strategy and the environment for Ontology Learning (OL) relations discovery task for the scientific publications domain by adjusting Deep Belief Network (DBN). The adjusted DBN is called Relu Dropout DBN (Re-DDBN). This paper elaborates on the adjusted Re-DDBN configuration, its structure, hyper-parameters, and functions. In addition, the adjusted Re-DDBN was compared with traditional DBN and other comparative models (e.g., Support vector machine (SVM) and Naïve Bayes (NB)) for ontology semantic relations discovery. The outcomes revealed that the adjusted Re-DDBN displayed the best performance when compared to the other models. The SemEval-2018 task 7 dataset was applied in this study.
RELU DROPOUT DEEP BELIEF NETWORK FOR
ONTOLOGY SEMANTIC RELATION DISCOVERY
FATIMA N. AL-ASWADI, HUAH YONG CHAN, & KENG HOON GAN
Universiti Sains Malaysia, Pulau Pinang, Malaysia.
Hodeidah University, Hodeidah , Yemen.
PRESENTED BY: Fatima N. AL-Aswadi
OUTLINES
Background
Research Problem
Configuring Learning Strategy
Deep Belief Network (DBN)
Adjusted DBN (Relu Dropout DBN)
Experiments & Results
Conclusion
BACKGROUND
Identifying Concepts & Relations
Mapping
Concepts &
Relations
R
R
R
Relations Discovery is considered the Backbone task of any Ontology Learning
System
ONTOLOGY & ONTOLOGY LEARNING
Many applications are based on ontology
Semantic Searching
Decision-Support Systems
Automated Fraud Detection
Question-Answering Systems
IMPORTANCE OF ONTOLOGY
RELATIONS DISCOVERY
Relations are the taxonomic or non-taxonomic (semantic) connections between
extracted concepts that form the facts and rules of specific domain.
Relations discovery task aims to extract the relationships among selected concepts.
Relations discovery techniques are combination of Natural Language Processing
(NLP) and Machine Learning (ML).
RESEARCH PROBLEM
RESEARCH PROBLEM
Extracting only taxonomic relation (or very limited relations), or based on large
predefined patterns, which have very low recall result such as in (A. El-Kilany et
al.,2017;Gillani Andleeb, 2015; L. Bergelid, 2018;Sureshkumar & Zayaraz, 2015; J.
Zhang et al., 2016; J. Wang et al., 2018;Zhong et al., 2016).
Proposed solution:
Develop efficient automatic method includes dual stages :
adjusting and configuring the appropriate learning
strategy to can automatically detect or learn relations
developing relations extraction and classification
technique based on the established learning strategy
This paper contribution
Upcoming contribution
CONFIGURING LEARNING STRATEGY
DEEP BELIEF NETWORK (DBN)
A stack of Restricted
Boltzmann machines
(RBMs)
The learning algorithm of
DBN has dual stages:
pre-training and fine-
tuning
RBM is a symmetrical graph
(each visible node is
connected with each hidden
node) and consists of two
layers (visible and hidden)
h0h1hm
v0v1v2vn
h
v
w
...
...
b
0
b
1
b
m
a
0
a
1
a
2
a
n
w
00
w
nm
Each layer in the DBNs has a
double role, it serves as the
hidden layer to the nodes that
come before and as the visible
layer to the nodes that come
after.
Input Layer
Output Layer
h1
h2
h3
RBM1
RBM2
RBM3
before
after
DEEP BELIEF NETWORK (DBN) CONT.
DBN
Input Layer
Output Layer
h1
h2
h3
Input Layer
Output Layer
h1
h2
h3
Without Dropout With Dropout
Inactive unit
Active unit
()
=

+
=1


()=+
=1 
=
+
=1 , if = 1
0 , other else = 0
Traditional Adjusted (Re-DDBN)
Sigmoid
=
+
Relu (Rectified Linear)
=,=,>
,
Re-DDBN HYPER-PARAMETERS
Hyper-parameter Value/Type
Hidden Layers 3
Neurons/nodes 800,600,500
Hidden Layers Activation Relu
Output Layers Activation Softmax
learning rate 0.2,0.3
Batch size Small dataset: 100~180
Large dataset: 180~270
epochs 50~80
Dropout 0.4
Optimization algorithm SGD
Hundreds of scenarios and
experiments have been
conducted to determine the
hyper-parameters’ values
EXPERIMENTS & RESULTS
DATASET
SemEval-2018 task 7 is a collection of two corpora, ACLRelAcS and ACL RD-TEC
2.0, within the computational linguistics domain. They are based on the ACL
Anthology Reference Corpus (a digital archive of published papers in both journals
and conferences for NLP and computational linguistics domain).
It is divided into two sizes; the small fraction had 800 samples and the large part
contained 1800 samples.
https://lipn.univ-paris13.fr/~gabor/semeval2018task7/
DROPOUT STRATEGY EXPERIMENTS
Different percentages of
dropout applied for Re-DDBN
and the outcomes revealed
that 40% dropout (=0.4) had the
best results for both small
and large sampling.
Accuracy of Re-DDBN for Different Dropout Percentages
COMPARATIVE MODELS
Traditional DBN: such as used in (H. Wang, 2015; Zhong et al., 2016)(Sigmoid +DBN)
Re-DBN: such as used in (Dai et al., 2017)(Relu +DBN)
Sig-DDBN: such as used in (J. Huang & Guan, 2021) (Sigmoid +Dropout +DBN)
Multinomial NB: same as used in (Sureshkumar & Zayaraz, 2015)
Linear SVM: same as used in (Bergelid, 2018)
Re-DDBN: Adjusted model (Relu +Dropout +DBN)
Hundreds of scenarios and experiments respectively were conducted to select the hyper-parameters for
these variations of DBN.
COMPARISON EXPERIMENT
The comparative experiments was conducted 10 times with 3-
fold cross-validation (=30) for each model to obtain the results.
The results were calculated using the macro function of k-fold
cross-validation.
COMPARISON EXPERIMENT
DL/ ML Model Hidden Layers
Activation With/without
Dropout Samples size Accuracy
DL
Re-DBN Relu without Small 0.4310±0. 0701
Re-DDBN Relu with Small 0.4620±0.0805
Traditional DBN Sigmoid without Small 0.4093±0.0690
Sig-DDBN Sigmoid with Small 0.4227±0.0498
Re-DBN Relu without Large 0.4575±0.1632
Re-DDBN Relu with Large 0.5244±0.1407
Traditional DBN Sigmoid without Large 0.4070±0.1305
Sig-DDBN Sigmoid with Large 0.4106±0.1026
ML
SVM Linear -- Small 0.4428±0.1379
SVM Linear -- Large 0.5058±0.0839
NB Multinomial -- Small 0.4563±0.0690
NB Multinomial -- Large 0.4868±0.0839
Comparison of the Developed Re-DDBN, Traditional DBN, Re-DBN, Sig-DDBN, SVM, and NB for Small
and Large Samples
RESULTS & DISCUSSION
Re-DDBN obtained the best accuracy result among the comparative models in small
and large samples.
The second highest accuracy was obtained by linear SVM for large samples and by
multinomial NB for small samples.
The sigmoid function showed weaker accuracy result than other models due to the
vanishing problem.
As for the imbalanced, noisy dataset and low characteristic dimensions of samples,
the adjusted model (Re-DDBN) performs better than other comparative models.
CONCLUSION
A Re-DDBN model, which uses a dropout strategy and Relu activation function, is proposed in this
study to establish the learning strategy for relations discovery.
The adjusted model, Re-DDBN, outperformed the traditional model DBN and other comparative
models.
The upcoming paper shall present the development of relations extraction and classification
techniques for OL based on the Re-DDBN model.
In future work, we will conduct the comparison of using Relu functions with and without batch
normalization with applying the dropout strategy or applying without it.
REFERENCES
1Franco, W., Avila, C.V.S., Oliveira, A., Maia, G., Brayner, A., Vidal, V.M.P., et al.: Ontology-based Question Answering Systems over Knowledge Bases: A Survey. 22nd International Conference on
Enterprise Information Systems (ICEIS) p. 532-9: SCITEPRESS Digital Library; (2020).
2. Al-Aswadi, F.N., Chan, H.Y., Gan, K.H.:Extracting Semantic Concepts and Relations from Scientific Publications by Using Deep Learning. In: Saeed F, Mohammed F, Al-Nahari A, editors. Innovative
Systems for Intelligent Health Informatics. IRICT2020 (p. 374-83). Cham: Springer International Publishing; (2021).
3. Ahmed, I.A., Al-Aswadi, F.N., Noaman, K.M.G., Alma'aitah, W.Z.a.: Arabic Knowledge Graph Construction: A close look in the present and into the future. Journal of King Saud University -
Computer and Information Sciences. (2022).
4. Tiwari, S., Al-Aswadi, F.N., Gaurav, D.: Recent trends in knowledge graphs: theory and practice. Soft Computing. (2021).
5. Wong, W., Liu, W., Bennamoun, M.: Ontology learning from text: A look back and into the future. ACM Computing Surveys (CSUR).44(4):20. (2012).
6. Al-Aswadi, F.N., Chan, H.Y., Gan, K.H.: Automatic ontology construction from text: a review from shallow to deep learning trend. Artificial Intelligence Review.53(6):3901-28. (2020).
7. Zhao, Z., Han, S.-K., So, I.-M.: Architecture of knowledge graph construction techniques. International Journal of Pure and Applied Mathematics.118(19):1869-83. (2018).
8. Alma’aitah, W.Z.a., Talib, A.Z., Osman, M.A.: Opportunities and challenges in enhancing access to metadata of cultural heritage collections: a survey. Artificial Intelligence Review.53(5):3621-46.
(2020).
9. Saber, Y.M., Abdel-Galil, H., El-Fatah Belal, M.A.: Arabic ontology extraction model from unstructured text. Journal of King Saud University -Computer and Information Sciences. (2022)
10. Sathiya, B., Geetha, T.V.: Automatic Ontology Learning from Multiple Knowledge Sources of Text. Int J Intell Inf Technol.14(2):1–21. (2018).
11. Arefyev, N., Sheludko, B., Davletov, A., Kharchev, D., Nevidomsky, A., Panchenko, A.: Neural GRANNy at SemEval-2019 task 2: A combined approach for better modeling of semantic relationships in
semantic frame induction. Proceedings of the 13th International Workshop on Semantic Evaluation. p. 31-8(2019).
12. Gillani Andleeb, S.: From text mining to knowledge mining: An integrated framework of concept extraction and categorization for domain ontology. Department of Information Systems. p. 146.
Budapest: Budapesti Corvinus Egyetem; (2015).
13. Sombatsrisomboon, R., Matsuo, Y., Ishizuka, M.: Acquisition of hypernyms and hyponyms from the WWW. Proceedings of the 2nd International Workshop on Active Mining(2003).
14. Specia, L., Motta, E.: A hybrid approach for relation extraction aimed at the semantic web. International Conference on Flexible Query Answering Systems. p. 564-76: Springer; (2006).
15. Sánchez, D., Moreno, A.: Learning non-taxonomic relationships from web documents for domain ontology construction. Data & Knowledge Engineering.64(3):600-23. (2008).
REFERENCES
16. El-Kilany, A., Tazi, N.E., Ezzat, E.:Building Relation Extraction Templates via Unsupervised Learning.In:Proceedings of the 21st International Database Engineering & Applications Symposium.228-34.
Bristol, United Kingdom ACM, (2017).
17. Minard, A.-L., Ligozat, A.-L., Grau, B.: Multi-class SVM for relation extraction from clinical reports. Recent Advances in Natural Language Processing. Hissar, Bulgaria(2011).
18. Bergelid, L.: Classification of explicit music content using lyrics and music metadata. TRITA-EECS-EX. STOCKHOLM, SWEDEN: KTH ROYAL INSTITUTE OF TECHNOLOGY; (2018).
19. Sureshkumar, G., Zayaraz, G.: Automatic relation extraction using naïve Bayes classifier for concept relational ontology development. International Journal of Computer Aided Engineering and
Technology.7(4):421-35. (2015).
20. Zhang, J., Liu, J., Wang, X.: Simultaneous Entities and Relationship Extraction from Unstructured Text. International Journal of Database Theory and Application.9(6):151-60. (2016).
21. Etzioni, O., Banko, M., Soderland, S., Weld, D.S.: Open information extraction from the web. Commun ACM.51(12):68-74. (2008).
22. Zhong, B., Liu, J., Du, Y., Liaozheng, Y., Pu, J.: Extracting Attributes of Named Entity from Unstructured Text with Deep Belief Network. International Journal of Database Theory and
Application.9(5):187-96. (2016).
23. Chen, Y., Li, W., Liu, Y., Zheng, D., Zhao, T.: Exploring deep belief network for chinese relation extraction. Proceedings of the Joint Conference on Chinese Language Processing (CLP’10). p. 28-
9(2010).
24. Feng, J., Lu, S.: Performance analysis of various activation functions in artificial neural networks. Journal of physics: conference series. p. 022030: IOP Publishing; (2019).
25. Dai, J., Song, H., Sheng, G., Jiang, X.: Dissolved gas analysis of insulating oil for power transformer fault diagnosis with deep belief network. IEEE Transactions on Dielectrics and Electrical
Insulation.24(5):2828-35. (2017).
26. Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a simple way to prevent neural networks from overfitting. The journal of machine learning research.15(1):1929-58.
(2014).
27. Nematzadeh, Z., Ibrahim, R., Selamat, A.: Comparative studies on breast cancer classifications with k-fold cross validations using machine learning techniques. 2015 10th Asian Control Conference
(ASCC). p. 1-6(2015).
28. Wang, H.: Semantic Deep Learning. University of Oregon. (2015).
29. Huang, J., Guan, Y.: Dropout Deep Belief Network Based Chinese Ancient Ceramic Non-Destructive Identification. Sensors.21(4):1318. (2021).
Fatima N.
AL-Aswadi
Huah
Yo n g
Chan
Keng
Hoon
Gan
Ph.D. candidate in School of
Computer Sciences at Universiti
Sains Malaysia (USM).Also,
research assistant and instructor
in the Computer Sciences
Department, at Hodeidah
University,Yemen.
Associate Professor in the
School of Computer
Sciences, Universiti Sains
Malaysia (USM).
Senior lecturer in the School
of Computer Sciences,
Universiti Sains Malaysia
(USM).
THANK YOU
ResearchGate has not been able to resolve any citations for this publication.
Article
Full-text available
With the widespread growth of data on the Web, the need for efficient methods to get and arrange valuable information from these big noisy data is increased. The knowledge graph (KG) is a way to represent and organize the data in a more efficient and easy way to modify, use, and understand. Recently, KG has become a new hotspot topic in academic and business research; it is used in many applications such as intelligent question-answering (QA), recommender systems, map navigation, etc. There has been a tendency to construct KG in different languages such as English, Chinese, Persian, or Arabic. Constructing KG faces many challenges and obstacles, especially constructing Arabic Knowledge Graph (AKG) due to the sparse Arabic data in online encyclopedias and academic research, as well as the lack of tools that can process the proprietary nature of the Arabic language effectively, besides other challenges. This research aims to review and discuss KG construction best practices (systems, phases, problems, and challenges) with highlighting the Arabic perspective. Besides, it elaborates a classification of the AKG challenges and investigates the potential solutions and opportunities that might define the key future research directions of constructing AKG.
Chapter
Full-text available
With the large volume of unstructured data that increases constantly on the web, the motivation of representing the knowledge in this data in the machine-understandable form is increased. Ontology is one of the major cornerstones of representing the information in a more meaningful way on the semantic Web. The current ontology repositories are quite limited either for their scope or for currentness. In addition, the current ontology extraction systems have many shortcomings and drawbacks, such as using a small dataset, depending on a large amount predefined patterns to extract semantic relations, and extracting a very few types of relations. The aim of this paper is to introduce a proposal of automatically extracting semantic concepts and relations from scientific publications. This paper introduces a novel relevance measurement for concepts, and it suggests new types of semantic relations. Also, it points out of using deep learning (DL) models for semantic relation extraction.
Article
Full-text available
With the extensive growth of data that has been joined with the thriving development of the Internet in this century, finding or getting valuable information and knowledge from these huge noisy data became harder. The Concept of Knowledge Graph (KG) is one of the concepts that has come into the public view as a result of this development. In addition, with that thriving development especially in the last two decades, the need to process and extract valuable information in a more efficient way is increased. KG presents a common framework for knowledge representation, based on the analysis and extraction of entities and relationships. Techniques for KG construction can extract information from either structured, unstructured or even semistructured data sources, and finally organize the information into knowledge, represented in a graph. This paper presents a characterization of different types of KGs along with their construction approaches. It reviews the existing academia, industry and expert KG systems and discusses in detail about the features of it. A systematic review methodology has been followed to conduct the review. Several databases (Scopus, GS, WoS) and journals (SWJ, Applied Ontology, JWS) are analysed to collect the relevant study and filtered by using inclusion and exclusion criteria. This review includes the state-of-the-art, literature review, characterization of KGs, and the knowledge extraction techniques of KGs. In addition, this paper overviews the current KG applications, problems, and challenges as well as discuss the perspective of future research. The main aim of this paper is to analyse all existing KGs with their features, techniques, applications, problems, and challenges. To the best of our knowledge, such a characterization table among these most commonly used KGs has not been presented earlier.
Article
Full-text available
A non-destructive identification method was developed here based on dropout deep belief network in multi-spectral data of ancient ceramic. A fractional differential algorithm was proposed to enhance the spectral details by making use of the difference between the first and second-order differential pre-process spectral data. An unsupervised multi-layer restricted Boltzmann machine (RBM) was employed to extract some high-level features during pre-training. Some weight and bias values trained by RBM were used to initialize a back propagation (BP) neural network. The RBM deep belief network was fine-tuned by the BP neural network to promote the initiative performance of network training, which helped to overcome local optimal limitation of the network due to the random initializing weight parameter. The dropout strategy has been put forward into the RBM network to solve the over-fitting of small sample spectral data. The experimental results show that the proposed method has excellent recognition performance of the ceramics by comparisons with some other ones.
Article
Full-text available
The explosive growth of textual data on the web coupled with the increase on demand for ontologies to promote the semantic web, have made the automatic ontology construction from the text a very promising research area. Ontology learning (OL) from text is a process that aims to (semi-) automatically extract and represent the knowledge from text in machine-readable form. Ontology is considered one of the main cornerstones of representing the knowledge in a more meaningful way on the semantic web. Usage of ontologies has proven to be beneficial and efficient in different applications (e.g. information retrieval, information extraction, and question answering). Nevertheless, manually construction of ontologies is time-consuming as well extremely laborious and costly process. In recent years, many approaches and systems that try to automate the construction of ontologies have been developed. This paper reviews various approaches, systems, and challenges of automatic ontology construction from the text. In addition, it also discusses ways the ontology construction process could be enhanced in the future by presenting techniques from shallow learning to deep learning (DL).
Article
Full-text available
Machine processable data that narrate digital/non-digital resources are termed as metadata. Different metadata standards exist for describing various types of digital objects. Several researches have reported on how to address issues related to accessing of metadata resources. Most studies on metadata involve cultural heritage domain, and this is an indication of the importance of this domain in metadata research and development. Research on metadata in cultural heritage mainly revolves around three fundamental issues: (1) lack of quality in metadata contents in most of the cases, (2) difficulty in accessing metadata contents due largely to limited user’s knowledge on the content of the metadata, and (3) heterogeneity of the data at the level of schemas which makes the access even more difficult. The lack of quality in metadata makes it difficult for the users to retrieve and explore information that satisfies their needs. So, in order to make its contents more accessible, enhancing the metadata content is required, especially for cultural heritage collections which consist of digital objects (structured documents) described by a variety of metadata schemas. This paper presents issues and challenges in enhancing access to metadata by reviewing the existing approaches in metadata environment with a particular emphasis on cultural heritage collections. In this paper, firstly, we look at the classification of metadata which is divided into two categories namely data retrieval and information retrieval. Then, we present the analysis, findings and suggestions on how to address issues in enhancing access to metadata contents especially in cultural heritage collections. A detailed comparison is given between information retrieval and data retrieval, and it focuses on the applicability of one approach over the other. A framework that aims to improve the effectiveness of retrieval when searching metadata is also proposed and tested. The proposed framework consists of approaches and methods that are expected to enhance access to metadata especially in cultural heritage collections and be useful for those with limited knowledge on cultural heritage. The experiments were conducted on CHiC2013 which is a collection on cultural heritage. The results show a considerable enhancement over other IR approaches that use the expansion methods.
Article
Full-text available
The prime textual sources used for ontology learning are a domain corpus and dynamic large text from web pages. The first source is limited and possibly outdated, while the second is uncertain. To overcome these shortcomings, a novel ontology learning methodology is proposed to utilize the different sources of text such as a corpus, web pages and the massive probabilistic knowledge base, Probase, for an effective automated construction of ontology. Specifically, to discover taxonomical relations among the concept of the ontology, a new web page based two-level semantic query formation methodology using the lexical syntactic patterns (LSP) and a novel scoring measure: Fitness built on Probase are proposed. Also, a syntactic and statistical measure called COS (Co-occurrence Strength) scoring, and Domain and Range-NTRD (Non-Taxonomical Relation Discovery) algorithms are proposed to accurately identify non-taxonomical relations(NTR) among concepts, using evidence from the corpus and web pages.
Article
Ontologies enable the sharing and reusing of knowledge, allowing the interoperation and combination of information systems. Automatic ontological relation extraction from text is an important factor for representing documents and their contents in a useful computerized manner. This paper proposes a model for automatic ontology extraction from Arabic text by parsing sentences and extracting part of speech (POS). Then, the proposed rule-based model is applied to extract the triple attributes of a sentence (subject, predicate, and object) from the parsing tree. Finally, the semantic relations can be used to extract new triples inferred from the initially extracted triples. The results were evaluated using two methods: the first method was based on a comparison with manual extraction because there is no standard method for measuring Arabic triple extraction; the second method is based on translating the dataset into English and comparing the output results using the Stanford dependencies extraction web tool. The proposed model achieved an accuracy of 73.6% for Arabic triple extraction and a 35% increase in overall triples owing to new inferred triples. However, when the dataset was translated into English and tested by the Stanford dependencies extraction web tool, the accuracy was 71.8% without inferred triples.
Article
Dissolved gas analysis (DGA) of insulating oil can provide an important basis for transformer fault diagnosis. To improve diagnosis accuracy, this paper presents a new transformer fault diagnosis method based on deep belief networks (DBN). By analyzing the relationship between the gases dissolved in transformer oil and fault types, the Non­code ratios of the gases are determined as the characterizing parameter of the DBN model. DBN adopts multi-layer and multi-dimension mapping to extract more detailed differences of fault types. In this process, the diagnosis parameters are pre-trained. A back-propagation algorithm adjusts them with the labels of the samples and optimizes the parameters. To verify the effect of the proposed method, the diagnostic DBN model is constructed and tested using various oil chromatographic datasets collected from the State Grid Corporation of China and previous publications. The performances of the DBN diagnosis model are analyzed by different characterizing parameters, different training datasets and sample datasets. In addition, the influence of discharge and overheating multiple faults on the diagnosis model is studied. The performance of the proposed approach is compared with that derived from support vector machine (SVM), back-propagation neural network (BPNN) and ratio methods respectively. The results show that the proposed method significantly improves the accuracy of power transformer fault diagnosis.