Open access peer-reviewed chapter

Finding New Connections between Concepts from Medline Database Incorporating Domain Knowledge

Written By

Yang Weikang, Chowdhury S.M. Mazharul Hoque and Jin Wei

Reviewed: 31 August 2023 Published: 01 November 2023

DOI: 10.5772/intechopen.113081

From the Edited Volume

Research Advances in Data Mining Techniques and Applications

Edited by Yves Rybarczyk

Chapter metrics overview

266 Chapter Downloads

View Full Metrics

Abstract

In this digital world, data is everything and significantly impacts our everyday lives. Interestingly, in this small world, everything is part of an ecosystem, where everything is connected, directly or indirectly. The same thing happens to data as well. In most cases, it may seem like a particular topic does not have any connection with another one, but in reality, they are connected through a mutually related topic. Therefore, in this research, we will discuss an adaptive model modified from the ABC model by Don R. Swanson, a Literature-Based Discovery (LBD) Model, to find the hidden connections between Concepts of Interest. The model demonstrates that two topics, “A” and “C,” are different and have no relationship. But they have a common topic, “B,” that can be used to connect topics “A” and “C.” This famous model will be used in this discussion to connect Medical Concepts.

Keywords

  • medical data analysis
  • NLP
  • Medline database
  • text extraction
  • web mining

1. Introduction

In this era of technology, the use of data is increasing rapidly, and with the growth of AI, data is growing even faster, as it requires a lot of data to come to a conclusion about a given problem. Throughout the world, the volume of scientific literature is expanding immensely. Because of that, it has become very challenging to keep track of the recent innovations, even for the specialized researchers with their slender domains. From all those literature-based contents, extracting useful information has become a challenge to everyone, and this can be solved using Literature-Based Discovery. LBD minimizes the difficulty among the researchers by reducing the complexity regarding finding useful information.

In this paper, we have worked on an extended version of the existing ABC model, and we applied the model to a medical research article database, the MEDLINE database. The reason behind this research was that every day, the medical sector is growing, and there are new problems that are coming up. There are a lot of existing problems that need to be solved, and researchers are working on them. As a result, every year, a huge number of articles are generated, and for anyone, keeping track of the data or the most recent research or research outcome would be a challenge. In the year 1999, a simple logic-based model was proposed by Swanson, known as Complementary Structures in Disjoint Literature (CSD). In this model, he wanted to present that, in a research article, there can be many theories and findings that can be interconnected. For example, it is possible that in an article, it is mentioned that a certain medicine is used to treat the “Migraine Headache,” and in another article, it says that that particular medicine commonly uses “Magnesium” as the main component [1]. Though “Migraine Headache” and “Magnesium” do not have a direct contact or relationship, and they exist in two different articles, it is still possible to make a relationship between those two.

This logic was explained by Swanson as follows – ““If concept A influences concept B, and concept B influences concept C, then concept A may influence concept C”, and this model is also a Literature Based Discovery model called Swanson’s ABC model [2]. Our research worked on an extended version of the ABC model. The model was applied to the Medline database, which is a database for worldwide medical-related research articles submitted to Medline and contains basic information, such as Title, Abstract, and Keywords. It is published by the National Library of Medicine (NLM). The proposed model uses the Pipe and Filter Software Architectural Design principles. The raw data was preprocessed to extract information that is relevant to the work before sending it to the final processing.

Moreover, the model was developed in such a way that it is not dependent only on two documents. Instead of that, it can go for cross-document discovery and produce a much better outcome. At the same time, the model was tested with a variety of cores to measure the impact on the result and processing time. It was found that the result on impact from the number of the cores was surprising. This model was tested and compared against other existing models and presented notable improvements. Data collected from the Medline database was processed using the MetaMap tool. It is a very popular tool that is used worldwide to process medical data, and it is also developed by NLM.

Advertisement

2. Swanson’s ABC model

Don R. Swanson was the pioneer of Literature-Based Discovery, which is a form of knowledge extraction that can automatically generate hypotheses from a given dataset (basically a large amount of text data, such as abstracts of different scientific publications) to present new interdependence between existing concepts. The aim of LBD is to present indirect associations that are not totally direct. According to Don R. Swanson’s hypothesis, two completely different studies that exhibit the A-B and B-C relationship may seem unchained, but an A-C relationship may emerge that was unexplored [3]. This process is also known as Swanson’s Linking.

Don Swanson’s original ABC model can be elaborated or explained through the “One Node Search” approach [4]. In this process, an initial topic or problem needs to be identified from an article that must have enough ground on the selected problem (Literature A). A second problem or topic needs to be selected from a second article with enough ground on the second topic and contains important information that would lead to the solution of problem A (Literature C). Here both articles have their own explanation and do not have any kind of interconnection. Now, from Literature A and C filtering important words and phrases needs to be carried out (B terms). B terms are the implicit information from both Literature A and C. Later, using those B terms (B1, B2, B3,………), searches need to be carried out to create B1, B2, B3….. literature from Literature A and C. By scoring each Bi literature from the title words and phrases from A literature it needs to figure out how many B literatures they are in. To define A literature, each Ai term needs to be searched and analyzed. From the analysis, common Bi terms from literature A and literature C would hypothetically contain common information related to the solution of the given problem.

Through this process, it has become much easier and faster to analyze and inspect indirect connections to gain new knowledge that has not been explicitly stated. This technique showed significant progress, and similar models were used to make new breakthroughs.

Advertisement

3. Literature review

The use of semantic predictions for LBD was presented and advocated by Hristovski, who was among the first few people who understand its importance [5]. To recover existing knowledge and find new relationships, he used the pattern-based approach that leveraged both term co-occurrence from the BioMedLEE and semantics from the SemRep [6, 7, 8, 9]. To find possible connections, he used a specified priori-based discovery pattern. However, this model’s major drawback was it could not be easily extended to find complex patterns. Therefore, interesting and more sophisticated patterns may remain untouched.

Delory Cameron et al. worked on Swanson’s hypothesis using a graph-based recovery and decomposition model using semantic predictions [5]. As biomedical literature-based semantic predictions allow the development of labeled directed graphs and from the literature, various concepts can be associated, a methodology for the semi-automatic discovery model was developed. Their model was a new approach to the field and was able to find associations at multiple levels.

Another study was done by Erwan Moreau et al. on the literature-based discovery of the ABC model to find the limitations of the ABC model [10]. Their work presents that the ABC model can find a relationship among the contexts, but how far it is effective or not effective is unknown to everyone, and the ABC model does not have the ability to judge it from a technical point of view. Therefore, they have used a data-driven task to work as a middle ground between the quantitative and qualitative learning-based discoveries. However, the model has an issue with the relatively small predefined targets.

Sam Henry et al. worked on the biomedical domain with an extensive literature review of literature-based discovery, where a unifying framework was selected, and different models and methods were differentiated and elaborated [11]. The purpose of the model was to give ideas to the readers about different LBD publications, models, methodologies, and challenges. Finally, with that knowledge, users will be able to develop an idea to create their own model.

Another work on the multilevel context-specific relationship discovery model of biological data was developed by Sejoon Lee et al. Their research goal was to find the relationship between drugs and diseases [12]. Their model proposed a three-step procedure to find the expected outcome, and the steps were multi-level entity identification, interaction extraction from the literature, and context vector-based similarity score calculation. The outcome from their model presented that context-based relation inference can perform better than the traditional ABC model.

Researcher Yong Hwan Kim focused on the limitations of the ABC model as it only gives priority to the B entity only and on the way to the B entity it may grab some relationships that are not entirely valid [13]. To counter the limitations, they presented a context-based approach to connect entities.

Advertisement

4. Method

As this model uses Pipe and Filter Architecture Design, multiple layers of filters work together to produce the final outcome from the raw data. Each filter works independently by taking input data from the previous filter in a processable format and producing the desired output format for the next filter. The whole analysis can be divided into three parts- preprocessing, preparation phase, and output phase. The whole process maintains a data flow that is given in Figure 1. The tool used to process data is the MetaMap, which is a popular tool that has great implementation in the text extraction and web mining field [14, 15].

Figure 1.

Data flow diagram.

4.1 Data collection and preprocessing

As mentioned earlier, all the data was collected from the Medline database through an API. This database is maintained by the NLM, and researchers across the world can get free access to research articles. At this moment, more than 23 million articles are stored in the Medline database. However, the basic type of data available in the database is the title of the article and abstract. Data were downloaded in XML format and from those XML files based on specific tags, information such as Publication ID, Publication Date, Article Title, and Article Abstract was extracted and stored. A total of 746 Medline documents were extracted from the website, which is - ftp://ftp.ncbi.nlm.nih.gov/pubmed/baseline. Each of the Medline documents has 30,000 Medline references, and each reference points towards a research document. Therefore, in total, around 22,380,000 references were analyzed by the system. All the documents that were collected were in English, and if any of the titles were in another language, they were translated into English first. Figure 2 shows a sample of the collected raw data.

Figure 2.

Sample XML file containing publication date, article ID, title and abstract.

4.2 Data processing through multiple filter

In this proposed model, filtering was done through multiple modules, and each module works independently. Here, the first module is the MedMeta module and it initially monitors the preprocessing. Later, it forwards the preprocessed data to the MetaMap API, where the MetaMap breaks down all the sentences in the document into phrases. In this case, those sentences are the title and abstract of the research articles. It normalizes those phrases to find mapping concepts through its database search in order to assign a mapping score to the concepts. Based on those concepts that were generated by the MedMeta server, indexes of concept occurrence relationships are produced and passed to the next filter.

The processes from the beginning to the end are monitored by one of the modules that can be called the MedMeta class. It observes the creation of the objects and connects them to the other objects. This model uses a parser to control event occurrence depending on a certain threshold or situation and the main reason behind it is that the amount of data to be processed can reduce the effectiveness of the model. For example, the meline documents were around 97 gigabytes in size, and an individual document reached up to 200 megabytes. The use of a Java-based purser reduced that pressure and ensured maximum effectiveness.

Another module named MedlineCite was responsible for storing data that was produced after parsing from the previous module. It divided the output data from the previous module into four categories, such as ID, Title, Abstract, and Date.

The next part of the module is the most important one, as it is responsible for communication with the MetaMap server in order to build indexes. A code sample for this part of the module is given below in Figure 3.

Figure 3.

Code snippet for connecting MetaMap server and index creation.

Titles and Abstracts will be passed through the server as input data to be processed, and the server will use Word Sense Disambiguation to separate the phrases and ignore any stop phrases. After processing the data, the MetaMap server will return the output. The generated output will be forwarded to the next filter to continue the analysis.

After that, the next filter will take the output from the previous filter as input and will gather important information. A three-level data structure (semantic class, concept class, and occurrence class) will hold the collected information. The task of the semantic class will be holding one or many concept objects. The concept class will hold one or many occurrence objects. Finally, the occurrence class is responsible for representing the occurrence of a certain word in a Medline document. It will hold information such as the genuine word that appears in the text, the original word preferable name, word location in the citation, Title, Abstract, ID of the citation and publication date where the word appears, and the exact location of the word that appaired on the title abstract. A sample of the MetaMapped document is given below in Figure 4.

Figure 4.

Sample MetaMapped document.

Now, this three-level data structure produces XLM type of output document to hold those generated data that is shown in the picture.

At this end of the processing MedMeta module uses a multithread based parallel processing in order to improve the performance. As mentioned earlier, each Medline document holds 30,000 citations and those citations are divided into three parts, each having 10,000 citations. Each part will be processed by a single thread, therefore, total three threads will be working together at the same time to access those data. Each thread will be able to communicate with the MetaMap server individually and will produce separate results. When all three of the threads are done with the analysis, all three threads will merge, and all the results will be forwarded to the main thread to start producing the output. Due to the use of multiple threads, it was possible to present the impact of using multithread against a single thread in the evaluation.

Here, another module was used named Semantic Type to Concept (S2C Module), which is responsible for generating a simplified version of the MetaMapped document or S2C document. Previously a three-level data structure was used to process data, and now, frequently, two-level S2C relationships will be used. There is a reason behind it. If every time a request is made and against that every time a three-level relationship is transformed into a two-level relationship, it will take a lot of time to process the data.

Therefore, by using a two-level S2C relationship, document processing time can be reduced. The S2C generator module starts its job by looking for the file path to pass them through the module as input, and a parser builds a two-level relationship. The parser basically uses an event-driven technique to go through all the MetaMapped documents to find concepts and include them in semantic-type classes. The semantic type class maintains a Hash set and only unique values will be added to the hash set. In this research, there were 133 semantic types, and individually they had a Hash set of concepts. Outcomes were stored in an XML file to be processed by the next module. A sample S2C document is presented in Figure 5.

Figure 5.

Sample S2C document.

A special note should be mentioned that, as only the title and abstract are playing the main role here, it is not necessary to keep anything other than those two in the main processing part. However, other relevant data can be used to know more about the source of the concept. Therefore, the Medline document will be more simplified and ready for concept generation. A sample of the simplified Medline document is shown in Figure 6.

Figure 6.

A simplified Medline document.

The final module in this model is called the ClosedDiscovery module, and it uses the MetaMapped Document, Simplified Medline Document, S2C Document, and the user-provided keyword as input. Those inputs will be used to generate Semantic Type to Concept and Weight (S2CW), which is the final output of the model, and it will be displayed through a Graphical User Interface (GUI).

The graphical user interface has two input boxes to provide input to the system, basically, those inputs are the keywords- Topic A and Topic C, whom the user wants to connect. To start the process, the user needs to click the start button on the screen (given in Figure 7). In the next stage, users will be able to select the S2C document that they want to be analyzed. Other S2C documents will not be considered for processing. As soon as the processing begins, it will start looking for the sentences relevant to the given keywords and will be stored in two different XML files based on their keywords (Topic A and C) (Figures 810).

Figure 7.

GUI for the model (showing input boxes for topics A and B, and their concepts, including the outcome C).

Figure 8.

Page to selecting S2C documents for processing.

Figure 9.

GUI showing all the relevant sentences.

Figure 10.

GUI presents sentences with intermediate linking concepts.

A concept chain will be created from those data by calculating the Term Frequency (TF) and the Inverse Document Frequency (IDF) in the sentences. The appearance of a concept in all sentences is considered as its Term Frequency. Multiple appearances of a concept in the same sentence will be considered as one appearance.

The Inverse Document Frequency measures the concept’s importance. For concept C it is measured by –.

IDF(c)=loge(Total Number of SentencesNumber of Sentences with conceptCinit)E1

In theory the minimal appearance of a concept in a sentence leads to high IDF value, making it important to the context. Later, the concept weight will be calculated by multiplying the TF and IDF, which is shown in Eq. (2).

WeightC=TFCIDFCE2

As the results from the MetaMap server may contain multiple concepts, therefore, by iterating them it is possible to build the Concept to Weight (C2W) and Concept to Sentence (C2Sents) relationship. Now the S2C relationship will be joined with the C2W relationship to find the S2CW relationship, and the weight of each concept is normalized by

Normalized Weight(c)=the weight of the conceptCmaximum weight in this Semantic TypeE3

Now to get the intermediate level information S2CW for topic B (S2CWB), S2CW for topic A (S2CWA) and S2CW for topic C (S2CWC) will be merged together. Potential connecting terms between topic A and B will be represented by S2CWB by crisscrossing their concepts. That means, all the concepts in S2CWB will be present in both S2CWA and S2CWC. The weight of S2CWA and S2CWB will determine the weight of S2CWB. The resulting S2CWB will be the relationship between topic A and topic C. The data was presented in the GUI, the left side represents the concepts and references for Topic A, and the right side represents the concepts and references for Topic C. In the center, topic B, or the relationship between input topics are presented.

4.3 System evaluation

There were quite a few modifications during the development of the MedMeta module. One of the major modifications was changing the model from a single-thread to a multi-thread model. Therefore, each of those threads can independently communicate with the MetaMap server, and the performance significantly improved and processing time reduced, though CPU use increased. An experiment was done to know how far performance changes due to changes in the number of threads used to communicate with the MetaMap server. The performance comparison is shown in Table 1 below. However, in this research, a computer was used that has a Core i7 4th generation CPU with eight cores, and 16 GB of RAM with Windows 10 64-bit OS. The document used for the test had 180 Medline citations. Testing was performed on 1 thread, 2 threads and 3 threads.

Number of threadsProcessing timeAverage memory usageAverage CPU usagePerformance compared to single thread% of theoretical performance
1145 s98 MB25%1
287 s110 MB40%1.6785%
362 s128 MB73%2.3478%

Table 1.

Performance comparison on test data.

According to Table 1, performance improved greatly based on the increment of the number of threads. First of all, processing time was reduced to 145 s (1 thread) to 87 s when 2 threads were used and again dropped to 62 s when 3 threads were used. On the other hand, performance increased 1.67 times while it was 85% of the theoretical performance when 2 threads were used, and 2.34 times while it was 78% of the theoretical performance when 3 threads were used. However, due to hardware limitations and as running multiple MetaMap servers costs a lot, the number of threads could not be increased, but with a computer with better configuration, it can be improved more.

4.4 Result analysis

From the given data set, the model found some meaningful connecting concepts based on the input topics A and C. All the relevant information was stored in the S2CWB XML files. But to find the accuracy of the result, the model was compared with another model by Gopalakrishnan and Kishlay [16]. They have used their own model on Mesh terms related to discovery from given topics, and connecting mesh terms were discovered that explain the relationship of input topics and found to be meaningful. The same topics, such as Fish-oil and Raynaud’s Disease, were given to this Medline Module with the same dataset to test the accuracy.

4.4.1 Test set 1 - fish oil and Raynaud disease

In the first test set by Gopalakrishnan and Kishlay was on ‘Fish oil and Raynaud Disease’. In their experiment, they found keywords such as Arthritis Rheumatoid, Platelet Aggregation, Prostaglandin, Blood Viscosity, and Vascular Reactivity (Table 2).

Connecting conceptFind?WeightRank in the semantic type
Platelet aggregationTrue0.872
Vascular reactivityTrue1.271
Blood viscosityFalse, but found something about blood pressure00
ProstaglandinTrue0.3096
Arthritis rheumatoidTrue, it appears as “Rheumatoid Arthritis”0.432

Table 2.

Analysis of fish oil and Raynaud disease.

The table presents that out of five connecting words, four were found by the model. On the other hand, three of the connecting words gained higher rank in their semantic type. This model used article title and abstract to find the linking contexts, but model by Gopalakrishnan and Kishlay used Mesh terms as potential connecting terms, that were assigned by biomedical experts. However, this model also found some linking terms that were undetected by their model, such as “Hemodynamic” and “Atherosclerosis”.

4.4.2 Test set 2 - migraine and magnesium

In the second study, two input topics were ‘Migraine and Magnesium’. In the experiment of Gopalakrishnan and Kishlay, they found connecting words like, Serotonin, Norepinephrine, Propranolol, Calcium, Ergotamine, Adenine Nucleotides, Adenosine Triphosphate, and Epinephrine (Table 3).

Connecting conceptFind?WeightRank in the semantic type
PropranololTrue0.418
Adenosine triphosphateTrue0.5210
CalciumTrue1.501
ErgotamineTrue, the system found “Ergot Alkaloids”0.345
SerotoninTrue1.491
NorepinephrineTrue0.796
Adenine nucleotidesTrue0.3214
EpinephrineTrue0.4913

Table 3.

Analysis on migraine and magnesium.

For Migraine and Magnesium this model was able to find all eight connecting words from the Gopalakrishnan and Kishlay’s model. Among those words, two of them got a very high semantic rank, and six of them were among the top ten semantic rank. Moreover, the model was able to find some connecting words, such as ‘Insulin’ that were not found previously.

4.4.3 Test set 3 - schizophrenia and phospholipase A2

In Gopalakrishnan and Kishlay’s model ‘Schizophrenia and Phospholipase A2’ was used as the third test data and here the connecting words their model found are Trifluoperazine, Chlorpromazine, Prolactin, Choline, Norepinephrine, Arachidonic Acid, Receptors Dopamine and Phenothiazines (Table 4).

Connecting conceptFind?WeightRank in the semantic type
ChlorpromazineTrue0.2211
TrifluoperazineTrue0.0732
Receptors dopamineTrue0.335
ProlactinTrue1.172
CholineTrue0.1221
Arachidonic acidTrue1.181
PhenothiazinesTrue0.0556
NorepinephrineTrue0.311

Table 4.

Analysis of schizophrenia and phospholipase A2.

In this case study the result presented that all the connecting words from Gopalakrishnan and Kishlay’s model were found again. Among eight of those words, three gained very high semantic score. Moreover, again, this model was able to find connecting words like, “PGE2”, that were new to the model by Gopalakrishnan and Kishlay.

Advertisement

5. Future work

This model is currently discovering only one level of intermediate terms (chain length 1). In the future, our goal is to upgrade the model into multiple levels, so that when there is only a small amount of information available, it becomes easier to find possible connections. Moreover, MetaMapped files will be combined together for easier access to the files. Finally, other MetaMap options will be explored to find the one that can provide the best result.

Advertisement

6. Conclusion

This research discussed the Learning Based Discovery (LBD) models and, more specifically, one of the LBD models, which is the ABC model. An extended version of the ABC model was presented with a complete application. This era of technology is growing very fast, and with its growth, the use of data is also increasing. Therefore, sometimes, it is difficult to find relevant data about a particular topic. This complexity can be reduced by the use of LBD models. There is no LBD model that can be considered as the best model, because all the systems have their problems and benefits. A model may not provide the best result for a particular problem, but can perform amazingly for another. Moreover, interaction with the user interface makes any system easier. Because of that, the popularity of UI-based systems is increasing among researchers as well. However, this modified Swanson’s ABC model can be improved more to solve problems faster and with a lot more data. This type of model will be able to make things easier for researchers by finding relevant data based on its user’s input data from a larger data set.

Advertisement

Acknowledgments

This research is supported by the National Science Foundation award IIS-1739095.

References

  1. 1. Andreou AP, Edvinsson L. Mechanisms of migraine as a chronic evolutive condition. The Journal of Headache and Pain. 2019;20:117. DOI: 10.1186/s10194-019-1066-0
  2. 2. Swason DR. Complementary structures in disjoint science literatures. In: Proceedings of the 14th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval. Chicago, IL: ACM Press; 1991. pp. 280-289
  3. 3. Swanson DR. Fish oil, Raynaud’s syndrome, and undiscovered public knowledge. Perspectives in Biology and Medicine. 1986;30(1):7-18. DOI: 10.1353/pbm.1986.0087
  4. 4. Smalheiser NR. Rediscovering Don Swanson: The past, present and future of literature-based discovery. Journal of Data and Information Science (Warsaw, Poland). 2017;2(4):43. DOI: 10.1515/jdis-2017-0019
  5. 5. Cameron D, Bodenreider O, Hima Yalamanchili T, Danh SV, Thirunarayan K, Sheth AP, et al. A graph-based recovery and decomposition of Swanson’s hypothesis using semantic predications. Journal of Biomedical Informatics. 2013;46(2):238-251, ISSN 1532-0464. DOI: 10.1016/j.jbi.2012.09.004
  6. 6. Hristovski D, Friedman C, Rindflesch TC, Peterlin B. Exploiting semantic relations for literature-based discovery. In: AMIA Annual Symposium Proceedings. AMIA Symposium. National Library of Medicine; 2006. p. 349-53
  7. 7. Hristovski D, Friedman C, Rindflesch TC, Peterlin B. Literature-based knowledge discovery using natural language processing. In: Bruza P, Weeber M, editors. Literature-Based Discovery, Information Science and Knowledge Management. Vol. 15. Berlin, Heidelberg: Springer; 2008. pp. 133-152
  8. 8. Rindflesch TC, Fiszman M. The interaction of domain knowledge and linguistic structure in natural language processing: Interpreting hypernymic propositions in biomedical text. Journal of Biomedical Informatics. 2003;36(6):462-477
  9. 9. Lussier Y, Borlawsky T, Rappaport D, Liu Y, Friedman C. Phenogo: Assigning phenotypic context to gene ontology annotations with natural language processing. In: Pacific Symposium on Biocomputing. National Library of Medicine; 2006. p. 64-75
  10. 10. Moreau E, Hardiman O, Heverin M, O’Sullivan D. Literature-based discovery beyond the ABC paradigm: A contrastive approach. BioRxiv. 2021;09(22):461375. DOI: 10.1101/2021.09.22.461375
  11. 11. Henry S, McInnes BT. Literature based discovery: Models, methods, and trends. Journal of Biomedical Informatics. 2017;74:20-32, ISSN 1532-0464,. DOI: 10.1016/j.jbi.2017.08.011
  12. 12. Lee S, Choi J, Park K, et al. Discovering context-specific relationships from biological literature by using multi-level context terms. BMC Medical Informatics and Decision Making. 2012;12(Suppl 1):S1. DOI: 10.1186/1472-6947-12-S1-S1
  13. 13. Kim YH, Song M. A context-based ABC model for literature-based discovery. PLoS One. 2019;14(4):e0215313. DOI: 10.1371/journal.pone.0215313
  14. 14. Aronson AR. Effective mapping of biomedical text to the UMLS Metathesaurus: The MetaMap program. In: AMIA Annual Symposium Proceedings. AMIA Symposium. National Library of Medicine; 2001. p. 17-21. DOI: D010001275
  15. 15. U.S. National Library of Medicine. MetaMap - A Tool for Recognizing UMLS Concepts in Text. 2017. Available from: https://metamap.nlm.nih.gov/
  16. 16. Gopalakrishnan V, Jha K, Zhang A, Jin W. Generating hypothesis: Using global and local features in graph to discover new knowledge from medical literature. In: Proceedings of the 8th International Conference on Bioinformatics and Computational Biology, BICOB. USA; 2016. pp. 23-30

Written By

Yang Weikang, Chowdhury S.M. Mazharul Hoque and Jin Wei

Reviewed: 31 August 2023 Published: 01 November 2023