Mode assessments for teamwork, checkmarks indicate the mode assessment of the expert.
A major challenge in constructing a Bayesian network (BN) is defining the node probability tables (NPT), which can be learned from data or elicited from domain experts. In practice, it is common not to have enough data for learning, and elicitation from experts is the only option. However, the complexity of defining NPT grows exponentially, making their elicitation process costly and error-prone. In this research, we conducted an exploratory study through a literature review that identified the main issues related to the task of probability elicitation and solutions to construct large-scale NPT while reducing the exposure to these issues. In this chapter, we present in detail three semiautomatic methods that reduce the burden for experts. We discuss the benefits and drawbacks of these methods, and present directions on how to improve them.
- Bayesian networks
- probability elicitation
- node probability table
- expert systems
- artificial intelligence
Bayesian network (BN) is a mathematical model that graphically and numerically represents the probabilistic relationships between random variables through the Bayes theorem. This technique is becoming popular to aid in decision-making in several domains due to the evolution of the computational capacity that makes possible the calculation of complex BN . Some examples of BN application areas are: software development project management [2, 3]; large-scale engineering projects ; and the prediction of success in innovation projects .
On the other hand, there are open challenges related to the construction of BN. One of these challenges is to build the node probability tables (NPT). In cases where there are databases with enough information for the problem in question, it is possible to automate the process of constructing NPT through
Furthermore, experts can often understand and identify key relationships that data alone may fail to discover . Therefore, the concept of smart data is defined by : a method that supports data engineering and knowledge engineering approaches with emphasis on applying causal knowledge and real-world facts to develop models.
In this context, it is necessary to manually elicit data from experts to define the NPT. However, given that the complexity of defining NPT increases exponentially, for large-scale BN, it becomes impracticable to manually define all the probability functions that compose each NPT . In addition, experts often have time constraints and are rarely interested in manually defining NPT, partially because it is necessary to work with many probabilistic distributions for long periods .
In addition, other factors may compromise the process of probability elicitation to construct the NPT, such as commonly used heuristics. Some well know heuristics used to reduce the cognitive effort in probability assessment task may lead the expert towards biased judgment of probability, leading to systematic errors. Moreover, the experts are hardly able to keep mutually consistent distributions during the NPT definition . In addition, factors such as boredom and fatigue are enough to make the criteria deviate during probability assessment , when in fact, it should be uniformly applied throughout the whole elicitation process.
A solution to solve this problem has been proposed by , which will be referenced herein as the ranked nodes method (RNM). Its goal is to define the NPT of the parent nodes and then generate the NPT of the child nodes. Ref.  introduces the concept of ranked nodes, ordinal random variables represented on a monotonically ordered continuous scale. A fundamental feature of this method is that mathematical expressions generate the child node’s NPT. These expressions define the central tendency of the child node for each combination of states of the parent nodes and have as input a set of weights of the parent nodes, which quantifies the relative strengths of their influence on the child node, and a variance parameter.
Another approach was proposed by , which will be referenced here as the weighted sum algorithm (WSA). This method uses well know heuristics in its favor, more precisely, the availability  heuristic and the simulation  heuristic. The main focus of this method is to assemble part of the NPT from experts by asking questions that comprehend cases that are easy to recall by experts, which is likely to be associated to more realistic probabilities. In the WSA, the remainder of the NPT is generated using interpolation techniques.
A systematic approach to generate NPT of nodes with multiple parents is proposed in . This approach is an adaptation of the analytic hierarchy process (AHP) method for the task of probability elicitation and semiautomatic generation of NPT, in which the expert needs only to make the assessment of probabilities conditioned on single parents. In this approach, the probability assessment is indirect by means of paired state judgments and the NPT is generated through the calculation of the product of the probabilities of the child node conditioned on single parents.
The three methods stated above reduce the burden for experts and allow the construction of complex BN in which manual elicitation of the NPT is unfeasible and, generally, there is not enough data to use batch learning. The reduced number of parameters to generate the NPT and consequently, reduced number of questions to ask the experts, makes it easier for the facilitator (e.g., BN expert) to deal with heuristics and possible biases during the NPT construction process. These methods can yet be extended with elaborate probability elicitation techniques (i.e., to improve its input).
Therefore, the objective of this research is to assess in detail three semiautomatic methods to generate NPT. We identified these methods in an exploratory study through a literature review. Additionally, we present heuristics that must be acknowledged during probability assessment for NPT construction and discuss extensions to these methods. It is our understanding that these methods can yet benefit from elaborate probability elicitation techniques. Such techniques can add additional overhead when manually defining the NPT, but this overhead is hugely reduced with semiautomatic methods (i.e., given the reduced number of questions to ask the experts) making them a viable choice to improve the method’s input.
This chapter is organized as follows. Section 2 presents an introduction to BN. Section 3 presents common heuristics which should be acknowledged and considered during the probability elicitation process. Section 4 presents a probability elicitation technique which can extend some of the semiautomatic methods. Section 5 presents three semiautomatic methods to generate NPT. Section 6 presents our conclusions and future works.
Bayesian networks are graph models used to represent knowledge about an uncertain domain . The Bayesian network, , is a directed acyclic graph that represents a joint probability distribution over a set of random variables . The network is defined by the pair , where is a directed acyclic graph with nodes representing random variables and edges representing the direct dependencies between these variables. is the set of probability functions (i.e., node probability table) which contains the parameter for each in conditioned by , the set of parameters of in . Eq. (1) portraits the joint probability distribution defined by over . An example of a BN is depicted in Figure 1.
In the above example, the probability of a person having cancer is calculated according to two variables: “Relatives had cancer” () and “Smoke” (). The ellipses represent the nodes and the arrows represent the arcs. Even though the arcs represent the causal connection’s direction between the variables, information can propagate in any direction . Hence, the direction of the arrows indicates the dependency to define the probability functions. In this example, it is assumed that all the variables are Booleans. Since the node “Cancer” is pointed out by and , the probability function is composed of probabilities for all possible combinations of states of and .
2.1. NPT’s complexity
A challenge in constructing a BN is defining the NPT, which can be learned from data or elicited from domain experts. In practice, it is common not to have enough data for learning and elicitation from experts is the only option. However, the complexity of defining NPT grows exponentially, which makes the elicitation process costly and error-prone.
Let us consider the following example shown in Figure 2. In this BN, we want to assess
To elicit all the probabilities needed to construct the NPT of the child node
Methods to address this problem were proposed. Noisy-OR and Noisy-MAX are two popular ones. However, the disadvantage of Noisy-OR is that it only applies to Boolean nodes. According to , the disadvantage of Noisy-MAX is that it does not model the extent of relationships required for large-scale BN. In this chapter, we present methods found in the literature that are applicable to a larger range of BN.
3. Heuristics in probability
The quantification process of a BN consists in converting expert knowledge, acquired through personal experiences, into probabilistic knowledge by eliciting a large number of subjective probabilities that reflect the expert’s belief at a given moment about something. Probability assessment can be described as the task of quantifying the chances of an event occur, using percentages. However, as the degree of complexity increases, it becomes increasingly difficult to size the probability of occurrence of each of the possible events in a given scenario.
For instance, we may have a hunch as to who will be the winner of a particular tournament at a particular time, but we will never know for sure the exact probability since the number of factors that can influence the event goes beyond our reach. Apart from that, epistemic uncertainties (e.g., lack of knowledge about all the participants in the tournament) and aleatory uncertainties (e.g., possibility of a team losing a player) play an important role in probability assessment. Nonetheless, if asked, one is capable of making an evaluation and give a quick answer. How do people manage to judge the probability of highly uncertain events?
According to , people make use of a limited number of heuristics, mental shortcuts, to reduce the complexity of judging the probability of an uncertain event. These mental shortcuts reduce the cognitive effort required to judge the probability of such events. However, they can lead to biases that result in systematic errors. In , three commonly used heuristics are presented: representativeness; availability; and anchoring.
The representative heuristic  describes the process by which people use the similarity of two events to estimate the degree to which one event is representative of another. It is used to answer questions such as: What is the probability that an event A originates from a process B? What is the probability of a process B generating event A? That is, if A is highly representative of B, the probability of A generating B is considered high. Conversely, if A is not representative of B, the probability of A originating from B is low.
Consider the following example adapted from : “
The availability heuristic [9, 16] is related to the judgment of probability of events occurring based on the ease with which we retrieve instances of these events in our mind. For example, to evaluate the likelihood that a person under the age of 30 years will suffer a heart attack, people usually do a quick search in their memory for cases they know of young people who have suffered a heart attack. This heuristic is useful because instances of larger classes are easier to remember than instances of smaller classes. However, the availability is affected by factors other than the frequency of events or probability. One may overestimate the probability of a young person getting cancer based on how recent an instance of such an event has occurred in his life, for example.
Anchoring and adjustment heuristic  occur when people judge probabilities based on an initial value, which is adjusted until the final response is reached. The problem with this heuristic is that the adjustments are usually insufficient. In other words, the expert assessment is likely to fluctuate around the initial anchor provided. It is important noting that, an anchor may be embedded in the formulation of a question to the domain expert (i.e., when a starting point is given), but it can also be the result of an incomplete computation.
In short, heuristics are mental shortcuts that reduce the cognitive effort in the task of reasoning about the probability of events with uncertainty. Although useful, it has its disadvantages that must be considered in the knowledge elicitation process. Therefore, it is imperative to acknowledge the possible biases derived from heuristics during the process of probability assessment, explicitly informing the experts of their existence and adopting appropriate methods to reduce their effects.
The number of probabilities to be elicited to construct an NPT may inevitably fall under some bias considering the effort needed from the experts. The semiautomatic methods reduce the number of questions to be asked to the expert or entirely removes the need of direct evaluation of probabilities during the construction of the NPT, which makes it easier for the facilitator and the expert to deal with these heuristics during the elicitation process, seizing the benefits of the heuristics and reducing their possible negative effects.
4. Probability elicitation methods
The process of probability elicitation can be supported by a variety of techniques designed to aid experts when they find it hard to express their degrees of belief with numbers. These methods are based on setting-controlled situations in which probability assessments can be inferred from the expert’s behaviors . In this section, we describe the use of probability scales with visual aids to make probability assessment easier for experts. However, it is worth noting, visual aids like probability scales (i.e., which uses numbers) still tend to be biased.
It is our understanding that the use of visual elements such as probability scales can improve the input quality of semiautomatic methods (i.e., the ones which needs probability distributions as input), but indirect methods, which we do not discuss here, may improve the input quality as well. Several methods for indirect elicitation of probabilities have been developed. Some well know methods are: the odds method; the bid method; the lottery method; the probability-wheel method; among others [17, 18], these methods allow the extraction of probabilities without have to explicitly mention probabilities, so to speak.
Both direct and indirect methods can be incorporated at some degree into semiautomatic methods. The purpose of this section is to show one of these techniques which can extend semiautomatic methods, as an example. Also, different techniques may produce different results, so we encourage readers to check a comprehensive review of issues related to the probability elicitation task which has a section dedicated for direct and indirect methods .
4.1. Probability scale
A probability scale is composed of a line that can be arranged vertically or horizontally with discrete numerical anchors which denotes the probabilities. It is a direct probability assessment method. To assess a probability, the experts mark a position on the scale. The probability value is given by the marking distance to the zero point of the scale. An example of a numerical probability scale can be seen in Figure 3.
There is no standard scale. For instance, anchors may vary in distance and values according to the domain, and lines can be arranged in different positions. Moreover, during probability assessment, one can use both numerical and verbal anchors. In  it is proposed a double scale that combines numbers and textual descriptions of probability to aid in the communication of probabilities. According to , verbal descriptions commonly used by people to express probabilities are directly related to the numerical values of the probabilities itself. In Figure 4, we can see an example of a double scale arranged in the vertical position with numerical and verbal anchors.
The advantage of using a scale is that it allows for the domain experts to think in terms of visual proportion rather than in terms of precise numbers. However, it is important to consider bias that may be introduced using probability scales. For example, let us say an expert is requested to indicate several assessments on a single line. In such a case, he is likely to introduce bias towards esthetically distributed marks. This bias is known as the spacing effect  and can be easily avoided by using a separate scale for each probability. Another bias that may be introduced by the use of probability scales is the tendency of people to use the middle of the scale. This bias is known as the centering effect .
Furthermore, scales can be used in combination with other components that may help in the task of probability assessment. In , a method is presented for elicitation of a large number of conditional probabilities in short time. This method was used to build a real-world BN for the diagnosis of esophageal cancer with more than 4000 conditional probabilities. This BN predicted the correct cancer stage for 85% of the patients . The main idea of this method is to present to the expert a figure with a double scale and a text fragment for each conditional probability. An example of combining probability scales with other components can be seen in Figure 5.
On the left side is a text fragment describing the conditional probability to be assessed. On the right side, we have the double scale proposed in . The text fragment is stated in terms of likelihood rather than frequency which circumvents the need for mathematical notation of the conditional probability. According to , the frequency format has been reported to be less liable to lead to biases and experts may experience considerable difficulty understanding conditional probabilities in mathematical notation. Conversely, such an approach may be less intuitive for domains in which it is difficult to imagine 100 occurrences of a rare event.
Nonetheless, in , the fragments of text and associated scales are grouped up accordingly to the conditional probability distribution. In so doing, domain experts can assess probabilities from the same conditional probability distribution simultaneously. In other words, the centering effect is avoided by presenting all the related probabilities (i.e., from the same probability distribution) at once for the expert to assess. This approach considerable reduces the number of mental changes during the probability elicitation process. In regards to the spacing effect, the proposed method avoids it by using a separated scale for each probability.
5. Semiautomatic methods
In this section, we present three methods to generate NPT that ease the burden for experts during the quantification process of a BN. These methods allow the construction of large-scale BN. The first is the RNM, which completely eliminates the need for direct probability assessment. The second is the WSA, which is based on two heuristics and needs only part of the NPT to be elicited from the expert. The third is an adaptation of the analytic hierarchy process (AHP) which reduces the cognitive effort, biases and inaccuracies from estimating probabilities to all combinations of states of multiple parents at a time. From now on, we will reference the latter as simply AHP. These three methods attack the magnitude problem of building NPT.
In , the ranked nodes method (RNM) is presented. In this chapter, it is introduced the concept of ranked nodes, ordinal random variables represented on a continuous scale ordered monotonically in the interval [0, 1]. For example, for the ordinal scale [“Low”, “Medium”, “High”], “Low” is represented by the interval [0, 1/3], “Medium”, by the interval [1/3, 2/3], and “High”, by the interval [2/3, 1]. This concept is based on the doubly truncated Normal (TNormal) distribution.
A normal distribution is made of four parameters: , mean (i.e., central tendency); , variance (i.e., uncertainty about the central tendency); a, the lower bound (i.e., 0); and, b, upper bound (i.e., 1). With a normal distribution, it is possible to model a variety of curves (i.e., relationships) as a uniform distribution, achieved when , and highly skewed distributions, achieved when . In Figure 6, we show an example of TNormal with same but different .
In this method, is defined by a weighted function of the parent nodes. There are four function types: weighted mean () Eq. (2), weighted minimum () Eq. (3), weighted maximum () Eq. (4) and a mix of both MIN and MAX functions () Eq. (5). In practice, these functions define the central tendency of the child node for the combination of parent nodes states. The weight of each parent node, which quantifies the relative strengths of the influences of the parents on the child node, must be defined by a constant in which .
Fenton et al.  do not present the details to, in practice, implement the solution. Despite presenting the mixture functions, there is no information regarding the algorithms used to generate and mix TNormal, define samples size and define a conventional NPT given the calculated TNormals. The latter enables the integration of ranked nodes with other types of nodes such as Boolean and continuous, which brings more modeling flexibility.
In , it is proposed a probabilistic algorithm for this purpose, composed of two main steps: (i) generate samples for the parent nodes and (ii) construct the NPT. In step (ii), for each possible combination of values for the parent nodes (i.e., each column of the NPT), the samples defined in the previous step are mixed given a function selected by the user and a TNormal is generated using the resulting mix and a variance defined by the user. An overview of the algorithm is shown in Figure 7.
As already mentioned, a ranked node is conceptually represented by an ordinal scale, which is mapped to the continuous interval [0, 1]. Thus, it is represented as a set of uniform distributions. For an ordinal scale with three values (e.g., “Bad”, “Moderate” and “Good”):, where
For the example shown in Figure 8, the set of uniform distributions is composed of the union of three uniform distributions. Numerically, this union is calculated using samples. Considering a sample size of 10,000, to represent the NPT of the example shown in Figure 8, it is necessary to collect 5700 random samples from , 3650 random samples from and 880 random samples from .
Figure 7 shows that the algorithm is composed of four collections: , a vector to store the samples of base states for the parent nodes; , a vector to store references to the parent nodes of each child node, in which is the number of parents; , a vector to store the states of each node, in which is the number of possible values for the child node given the combination states of its parents; and , a vector to store the resulting distribution for each possible combination of states of the parent nodes.
The repository strategy is used for optimization purposes. First, it is registered in memory (i.e., in ) distributions that represent the base states, which are states with hard evidence (i.e., a node has 100% of chance for a given state). For instance, for a node composed of the states [“Bad”, “Moderate”, “Good”], are registered samples for: 100% “Bad”, 100% “Moderate” and 100% “Good”, which respectively has , and . For this purpose, samples from a uniform distribution with the limits defined given the thresholds of the scale are collected.
For instance, for 100% “Good”, it is collected samples of a uniform distribution limited in the interval [2/3, 1]. In  it is empirically defined that using a sample size of 10,000 is enough to guarantee a margin of error less than 0.1%. Each sample is registered with meta-data regarding its configuration (i.e., number of states and ). The data in is used to generate samples for a node. Therefore, the samples for a base state are only generated once and reused later. The next step consists of, for each combination of the parent nodes, mix the TNormal using equidistant samples, randomly selected for each parent node. The samples are mixed using one of the given functions (e.g., , , or ) and the defined variance.
To mix the distributions, a random element from each sample of the parents is removed and used to calculate a resulting element using a given function. For instance, consider node A with two parents B and C. If we are calculating the probabilities of A for the combination “Low”-“High” and the selected function is with equal weights, if the values removed in an iteration were 0.1 and 0.7, the resulting value would be 0.4. This step must be repeated until the collections of samples are empty.
Afterwards, the set of calculated elements and the given are used as input to generate a TNormal. The resulting distribution is converted to an ordinal scale and represents a column in the NPT of the child node (i.e., in the given example, the column for the combination “Low”-“High”). At the end of this step, all the possible combinations of states of the parent nodes are evaluated and the NPT for the child node is completed.
Accordingly, the inputs to generate the NPT of a child node are: a weighted expression capable of generating curves equivalent to distributions expected by the experts; a set of weights of the parent nodes; and a value for . To determine the weighted expression one can ask the experts to assess the mode of the child node for different combinations of the extreme states of the parent nodes . For instance, let us consider the Bayesian network shown in Figure 8 along with the mode assessments of the experts in Table 1.
First, let us consider the rows 1 and 4, where all the parent nodes are in the highest and lowest states respectively. As can be seen in Table 1, when o all the parent nodes are in the lowest or highest states, the mode of the child node is also the lowest or highest state. Such a probability distribution can be obtained by any of the weighted expressions.
Now, let us consider the row 1 as the initial state, rows 2, 6 and 8 indicate that when the state of a single parent node shifts from lowest to highest state the mode of the child node shifts towards the highest state. Similarly, consider row 4 as the initial state, rows 3, 5 and 7 indicate that when the state of a single parent node shifts from highest to lowest state, the mode of the child node also shifts towards the lowest state.
However, it is quite clear that the shift effect is stronger when it occurs from the lowest to highest state. Hence, Table 1 reveals that the mode of the child node is inclined to go more towards the highest than lowest states which makes the function more suitable to express the distribution expected by the experts.
The process to determine the weights of the parent nodes and the variance parameter is not as straightforward as to determine the weighted expression. There is no guideline in the literature, as far as we know, to aid in this task. Nonetheless, one can use the mode assessments in Table 1 as a starting point to define the weight of the parent nodes. For instance, considering is the most suitable function to express the probability distribution, let us examine the rows in which the states shift from lowest to highest in Table 1.
Finally, let us consider row 1 as the initial state, rows 2, 6 and 8 indicate that the parent nodes have different strengths of influence on the child node. That is, when the parent node
This method solves the magnitude problem of constructing NPT in complex Bayesian networks. On the other hand, a drawback to this method is that the domain context needs to fit a pattern that can be modeled by one of the weighted expressions. This solution has been validated through case studies in different real-world domains, such as human resources management in software projects , software quality forecasting , air traffic control  and operational management .
In  the WSA method is proposed. This work introduced the concept of compatible parental configuration. The availability heuristic and the simulation heuristic are the base for this concept. As previously stated, the availability heuristic operates under the assumption that is easier to remember events that are more likely to occur. The simulation heuristic, in turn, operates according to which people determine the probability of an event based on how easy it is to simulate it mentally.
To formally define the concept of compatible parental configuration, we take as a basis the work of . Superscript is used to represent the states of a node and subscript to differentiate the parent-nodes. Therefore, consider that for is assigned an arbitrary state , that is, , since is another parent node, such that is considered compatible with only when is in the state which is most likely, according to the expert’s knowledge, to coexist with . Hence, we use the notation to represent the set of states that are compatible with for all parent nodes.
The compatible parental configurations are captured during the elicitation process by asking the domain experts to choose off the top of their head a plausible combination of states for each , which, theoretically, are easier to simulate and therefore, prone to more realistic probabilities. Hence, it is elicited from experts the probability distributions for all compatible parental configuration and relative weights. The NPT is calculated using a weighted sum algorithm  which takes these probability distributions and weights as input. The input data of the algorithm is obtained from the experts’ knowledge, as follows:
relative weight (between zero and one) for each parent node, denoting its degree of influence on the child node ;
probability distributions of for compatible parental configurations.
where is the relative weight of the parent node, and . A constraint must be observed: the sum of all the relative weights (i.e., of all parent nodes) must be exactly one. A weight equal to zero indicates that the parent node has no influence on the child node and therefore can be omitted from the relation. Conversely, a relative weight equal to one indicates that the parent node is the only determinant of the conditional probabilities on the child node.
For instance, let us consider the Bayesian network shown in Figure 2 where we wish to assess teamwork. For the sake of simplicity let us say that all the parents have the states “Low”, “Medium” and “High” instead of the five states from the original example. With WSA 3 × 3 distributions are needed to construct a complete NPT against in case of manual elicitation. Starting with the parent , let us say that the domain expert subjectively interprets the compatible parental configurations as an equivalence relation as follows:
When the domain expert provides 3 probability distributions over the node then all 3 × 3 distributions for compatible parental configurations are obtained. To generate the NPT, the expert must assign relative weights to the parents to quantify the relative strengths of their influence on the child node. Let us say that the expert interprets
With the weights and 3 probability distributions over the node as inputs, the weighted sum algorithm calculates all the distributions required to populate the NPT. On the other hand, let us say that Eq. (8) is not satisfied, then all the 3 × 3 probability distributions must be elicited.
In such a case, the probability of
This summarizes the WSA method, for an in-depth description please check . Unfortunately,  do not describe how to deal with situations where the expert cannot select a single compatible parental configuration. Hence, an extension to this method is proposed by  to deal with such situations by averaging the probabilities of valid compatible parental configurations that experts might select.
Although the direct assessment of probabilities in the construction of NPT is feasible for small Bayesian networks and relatively simple domains, for medium to large networks the complexity and burden for experts grows substantially. As the number of parents and states increase, the more difficult it becomes for experts to reason about conditional probabilities with multiple parents and multiple combinations of states at once, and the more susceptible it becomes to biases and inaccuracies .
In  it is proposed a systematic approach for generating conditional probabilities of nodes with multiple parents. It is an adaptation of the AHP method for the task of probability elicitation and semiautomatic generation of NPT where the expert only needs to provide probability assessments (i.e., indirect) conditioned on single parents. In this approach, the probability assessments are extracted from pairwise judgments of the states. The NPT is generated through the product of the probabilities of the child node conditioned on single parents.
Before using the proposed method  it is required to define an agreed upon scale to perform the pairwise judgments over the states of the node. Saaty’s scale  can be used for this purpose or a custom one can be created. A good example of how to obtain a scale can be consulted in  in which four successive experiments were performed to generate a scale with numbers and words. The Saaty’s scale has nine values as seen in Table 2.
|1||Equal likely||Event A and evet B are equal likely|
|2||Weak or slight|
|3||Moderate more likely||Event A is moderate more likely than event B|
|5||Strong more likely||Event A is Strong more likely than event B|
|7||Very strong more likely||Event A is very strong more likely than event B|
|8||Very, very strong|
|9||Extremely more likely||Event A is extremely more likely than event B|
For a better understanding of the method, we substitute the original terminology used in the AHP for terms more appropriate to the probability context. Thus, the term
In the above matrix, is specified by the question “comparing the state with , which is more likely and how more likely?”. Once we have filled the values for we can find the values of by calculating the inverse of , i.e., . The final result is a reciprocal matrix with all elements in the diagonal equal to 1, that is, for all .
The relative priority of is obtained from the maximum eigenvector of the matrix and the consistency of the matrix is the consistency ratio , where is the consistency index, defined by (where is the maximum eigenvalue corresponding to , and is the random index given by Table 3. A comparison matrix with less than 0.10 is considered acceptable . Although  has observed that this threshold may be inappropriate for the purpose of evaluating probabilities. Since the sum of all elements in is 1 and the ith element represents the relative importance of the state , is now interpreted as the prior probability of the state , that is, .
Similarly, to obtain the probabilities of a node with a single parent we estimate . In Figure 10 we see the resulting matrix when node :
In the above matrix is specified by questions such as “if the node is in the state, comparing the states and of the child node , which one is more likely and how more likely? “. After obtaining we have . The number of matrices needed to obtain is equal to the number of states of . The obtained results compose the NPT of a child node conditioned to the states of a parent , as shown in Figure 11.
The approach to generate the conditional probabilities for multi-parent nodes is based on , which states that when a node A in a Bayesian network has two parents B and C, its conditional probability in B and C can be approximated by where is a normalizing factor that ensures that . Hence, to generate the complete NPT Eq. (10) is applied:
This approach focuses on easing the burden for experts by automatically generating probabilistic distributions of nodes with multiple parents, and consequently, the complete NPT through the calculation of the product of the probabilities conditioned on single parents. Thus, the expert assesses the probabilities of a particular child node conditioned to each of its parents, one at a time, and these probabilities are combined to get the node’s conditional probability conditional on all its parents.
In  a similar method is proposed, also based on the AHP, which allow the quantitative evaluation of the inconsistency of experts in the task of probability assessment. The difference of the proposed methods is that in  the magnitude problem to construct NPT is reduced with a semiautomatic approach for the generation of the NPT and the cognitive effort is reduced because the experts only need to evaluate, indirectly, probabilistic distributions conditioned on a single parent at a time, whereas in  the effort is even greater than the direct elicitation of probabilities. Nonetheless, it is our understanding that the method proposed in  can somewhat extend other methods such as the WSA, without causing too much overhead. However, further studies are needed to confirm this.
Despite recent popularity, the construction of BN is still a challenging task. One of the main obstacles refers to defining the NPT for large-scale BN. It is possible to automate this process using batch learning, but it requires a database with enough information. In practice, this is not common. The other option is to elicit data from experts, which is unfeasible in most cases due to the number of probabilities required. A third option is to use semiautomatic methods that given an input (i.e., elicited from experts) generates the NPT.
In this chapter, we present three semiautomatic methods, found in an exploratory study through a literature review. These methods help, to a certain extent, to minimize the effects of human biases by reducing the parameters that are required to construct complete NPT. However, these methods are highly reliable on the input data elicited from experts. Therefore, flawed input necessarily produces nonsense output. For this reason, we present one of many probability elicitation techniques as an example, which can improve the input data needed by the semiautomatic methods and reduce the garbage in/garbage out effect.
The biggest problem with elaborated probability elicitation techniques is undoubtedly its cost, which is often greater than the direct elicitation of probabilities. Thus, these methods are not well suited for the construction of large-scale BN, despite been useful to deal with well know biases. However, it is our understanding that the cost to use elaborated probability elicitation techniques is drastically reduced when only is needed to elicit a small fraction of data of what would be necessary for manual definition of NPT. Therefore, the combination of semiautomatic methods and elaborated probability elicitation techniques might help building more reliable BN.
For example, let us consider the WSA method that uses a partial elicited NPT to generate a complete one using the concept of compatible parental configurations, weights of the parents and a weighted sum algorithm. Once the compatible parental configurations have been chosen, its probabilities can be elicited using a sophisticated probability elicitation technique with a rather small overhead. In one way, the probability elicitation technique becomes feasible and, theoretically, the input of the semiautomatic method is improved.
Nonetheless, it is evident that some methods may benefit more from elaborated probability elicitation techniques than others. However, it is still possible to use these techniques even in a method such as RNM. For example, the expert can inform the probabilities rather than the mode of each probabilistic distribution of the combination of extreme states (see Table 1). We believe that studies must be carried out to check if combining elaborated probability elicitation techniques with semiautomatic method can indeed improve the construction of large-scale BN.