Open access peer-reviewed chapter

# Bayesian Inference Application

By Wiyada Kumam, Plern Saipara and Poom Kumam

Submitted: December 6th 2016Reviewed: August 7th 2017Published: November 2nd 2017

DOI: 10.5772/intechopen.70530

## Abstract

In this chapter, we were introduced the concept of Bayesian inference and application to the real world problems such as game theory (Bayesian Game) etc. This chapter was organized as follows. In Sections 2 and 3, we present Model-based Bayesian inference and the components of Bayesian inference, respectively. The last section contains some applications of Bayesian inference.

### Keywords

• statistical inference
• Frequentist inference
• Bayesian inference

## 1. Introduction

In statistical inference, there are two ways for interpretations of probability include Frequentist (or Classical) inference and Bayesian inference. It usually is unlike with each other in the classical nature of probability. Classical inference defines probability as the limit of an event’s relative frequency for a large number of experiments and only in the sense of random experiments which are well defined. Other side, Bayesian inference can to impose probabilities to each statement when a random process is not associated. In the sense of Bayesian, probability is a way to show an individual’s degree of believes in a statement. Bayesian inferences are different interpretations of probability, and also different approaches depend on those interpretations. Bayes’ theorem presents the relativity about two conditional probabilities that are the reverse of anything other. The initials of the term Bayes’ theorem is in honor of Reverend Thomas Bayes, and is referred to as Bayes’ law (see [1]). This theorem shows the conditional probability or posterior probabilityof an event A after B is observed in terms of the prior probabilityof A, prior probability of B and the conditional probability of B given A. It is valid in all interpretations of probability. Bayes’ formula is how to revise probability statements using data. The Bayes’ law (or Bayes’ rule) is

PAB=PBAPAPB.E1

The conditional probability definition is defined as follows

PAB=PABPB=PBAPA.E2

For example, let a dice is thrown under a dice-box. From the standard model, all of outcomes have probability equal to 1/6. Now, the dice is lifted a bit and a random corner of the upper side is able to visible which it contains a dot. The new probability distribution of the outcomes shows as follows. Let Aiis the outcome of the throw, for i = 1 , 2 , 3 , 4 , 5 , 6 and Bis the randomly chosen corner contains a dot. So, we get P(Ai) = 1/6 and P(B) = 2/3. We get the following table:

AiP(Ai)P(B| Ai)P(Ai ∩ B)P(Ai| B)
A11/6000
A21/61/21/121/8
A31/61/21/121/8
A41/611/61/4
A51/611/61/4
A61/611/61/4

The simplest way to construct the fourth column is to multiply. For any Ai, P(Ai| B) and P(B| Ai), to sum these values and divide by this sum. This final term is said to be scaling and corresponds to the formula as

i=16PBAiPAi=i=16PAiB=PB.

An simpler argument is that P(Ai| B) has to be a probability distribution, thus sum to unity. As the scaling operation is trivial, Bayes’ rule is also shown as

PABPAPBA
where P(A) the prior (distribution), P(B|A) is the likelihood and P(A|B) is the posterior (distribution).

The main result of Bayesians statistics is that statistical inference may depend on the simple device posterior ∝ prior ∗ likelihood. By dice-throwing example is not of controversial. The disputations about the possibility of using Bay’s rule as

PTruthData=PDataTruthPTruthPData.E3

So, we get

PTruth=the prior.E4

The second ingredient we need is data, plus a how the data associate to the truth which is nothing but the classical concept of specifying a random relationship

PDataTruth=the likelihoodE5
for all associated values of Truth. Note that P(Data| Truth) is not applied as probability distribution for different data, but as the probability of the given data for different values of Truth. Various authors do apply P(Data| Truth) for likelihood to sheer this misconstrue.

Now, noting that (replace Truthwith T), probability of Data (P(Data)) can be written as

PData=PTPDataTdTE6
that is as a function of P(T) and P(Data| T), it is obvious that the prior and likelihood enable, using 1 to construct a new probability statement about Tgiven the data as follows
PTruthData=the posterior.E7

The purpose of this chapter was to introduce the concept of Bayesian inference and application to the real world problem such as game theory (Bayesian Game). In this chapter was organized as follows. In Sections 2 and 3, we present Model-based Bayesian inference and the components of Bayesian inference, respectively. The last section contains some applications of Bayesian inference.

## 2. Model-based Bayesian inference

The basic of Bayesian inference is continued by Bayes’ theorem. From (1), replacement Bwith observations y, Awith the set of parameter Θ, and probabilities Pwith densities p, results as the following

pΘy=pyΘpΘpyE8
which p(y) is the marginal likelihood of y, p(Θ) is the set prior distributions of the set of parameter Θ before yis observed, p(y|Θ) is the likelihood of yunderneath a model and p(Θ|y) is the joint posterior distribution of Θ that expresses uncertainty about parameter set Θ after taking both the prior and data into system. Because there are often multiple parameters, Θ presents a set of jparameters, we have
Θ=θ1,θ2,,θj.

The term

py=pyΘpΘdΘE9
determines the marginal likelihood(or the prior predictive distribution) of ywhich it was introduced by Jeffreys [2], and may be set to cwhere cis an unknown constant. This distribution shows what yshould be similar to given the model, before yhas been observed. Only the prior probabilities and the model’s likelihood function are applied for p(y). The presence of p(y) normalizes the joint posterior distribution, p(Θ|y) guarantee it is a proper distribution and integrates to 1. From replacement p(y) with a constant of proportionality c, the Bayes’ theorem becomes to
pΘy=pyΘpΘc.E10

We get

pΘypyΘpΘE11
when ∝ is proportional to.

This formulation (11) be shown as the unnormalized joint posterior being proportional to the likelihood multiply with the prior. Howsoever, the aim of this model is often not to concluding the non-normalized joint posterior distribution, however to concluding the marginal distributions of the parameters. The set of all Θ can partitioned as

Θ=ΦΛE12
when the interest sub-vector denote by Φand the complementary sub-vector of Θ denoted by Λ, usually called to as a vector of nuisance parameters. For a Bayesian scope, the presence of nuisance parameters does not pose any formal, theoretical problems. A nuisance parameter is a parameter that exists in the joint posterior distribution of a model, though it is not a interest parameter. The marginal posterior distribution of φ, the interest parameter, can be shown as
pφy=pφΛydΛ.E13

In model-based Bayesian inference, Bayes’ theorem is applied to approximate the non-normalized joint posterior distribution, and lastly the user can evaluate and make inferences by the marginal posterior distributions.

## 3. The components of Bayesian inference

In this section, we presents about the components of Bayesian inference which contains the prior distributions, the likelihood or likelihood function and the joint posterior distribution as follows.

1. p(Θ) is the prior distributions for set of Θ, and uses probability as a methods of quantifying uncertainty about Θ before taking the data into system.

2. p(y|Θ) is the function of likelihood which all variables are associated in a full probability model.

3. p(Θ|y) is the joint posterior distribution that shows uncertainty about Θ after taking both the prior and the data into system. If Θ is partitioned into a single parameter of interest φand the remaining parameters are considered nuisance parameters, then the marginal posterior distribution of φdenote by p(φ|y).

### 3.1. Prior distribution

The prior distribution is a main concept of Bayesian and shows the information about an uncertain Θ that is merged with the probability distribution of new data to yield the posterior distribution which in turn is applied for future inferences and decisions about Θ. The existence of a prior distribution for any problem can justified by some axioms of decision theory; which we now focus for how to set up a prior distribution for every given application. Generally, Θ will be a vector, but for easiness we will point as on p(Θ).

By well-identified and large sample sizes, suitable alternatives of p(Θ) will have minor effects on posterior inferences. This definition might look like to be circular, but in practice one can check the dependence on p(Θ) by a sensitivity analysis: comparing posterior inferences under different suitable alternatives of p(Θ).

If the sample size is small, or available data provide only indirect information about the parameters of interest, then p(Θ) becomes more important. In various cases, nevertheless, models can be set up hierarchically, such that clusters of parameters have shared p(Θ), which can themselves be approximated from data. Prior probability distributions have belonged to one of two kinds as informative and uninformative priors. In this section, four kinds of priors which include informative, weakly informative, least informative, and uninformative, are shown according to information and the aim in the use of the prior.

#### 3.1.1. Informative prior

If prior information is obtainable about Θ, it should be included in p(Θ). If the current model is homologous to a previous model, and the current model is goal to be an adjusted version dependent on more current data, then the posterior distribution of Θ from the previous model maybe used as p(Θ) for the current model.

Now, every version of a model is not start from scratch, based only on the current data, but the cumulative effects of all data, past and current, can be taken into system. To sure the current data do not dominate the prior, in 2000, Ibrahim and Chen [3] presented the power prior which it is a class of informative prior distribution that takes early data and results into system. If the current data is very homologous to the previous data, then the precision of the posterior distribution increases when including more information from previous models. If the current data differs tremendously, then the posterior distribution of Θ maybe in the tails of the prior distribution for Θ, therefore p(Θ) contributes less density in its tails.

Sometimes informative prior is not ready to be applied, for example when it resides in other person, as in an expert. For this way, their human personal beliefs of the probability for the event must be elicited into the form of a suitable probability density function which this process is said to be prior elicitation.

#### 3.1.2. Weakly informative prior

Weakly informative prior (in the short term: WIP) use prior information for regularization and stabilization, providing sufficient prior information to prevent results that contradict our knowledge for example an algorithmic failure to explore the state space. Other aim is for WIPs to use less prior information than is really available. WIPs should provide some of the useful of prior information while avoiding some of the risk from using information which does not exist. WIPs are the most common priors in practice and are liked by subjective Bayesians.

Selecting WIPs may be cumbersome. WIPs distributions should change with the sample size, since a model should have sufficient prior information to learn from the data, but the prior information must also be weak sufficient to learn from the data.

In practice, this is an example of WIPs. It is favor, for well reasons, to center and scale all continuous predictors [4]. Though centering and scaling predictors is not talked about here, but it should be clear that the potential range of the posterior distribution of θfor a centered and scaled predictor should be small. A favor WIPs for a centered and scaled predictor may be θN010000where θis normal distribution agreeable to a mean of 0 and a variance of 10,000. Here, the density for θis nearly flat. Nonetheless, the fact that it is not perfectly at yields well properties for numerical estimation algorithms. In both Bayesian and Frequentist inference, it is possible for numerical estimation algorithms to become stuck in regions of at density which become more common as sample size decreases or model complexity increases. Numerical estimation algorithms in Frequentist inference function as though a at prior were used, thus numerical estimation algorithms in Frequentist inference become stuck more frequently than numerical estimation algorithms in Bayesian inference. Prior distributions that are not completely at allow sufficient information for the numerical estimation algorithm to continue to diagnose the goal density, the posterior distribution.

After updating a model in which WIPs exist, the user should be investigating the posterior. If the posterior contradicts knowledge, then the WIPs must be revised by including information that will make the posterior consistent with knowledge [4]. A favor purpose Bayesian criticism against WIPs is that there is no precise mathematical form to derive the optimal WIPs for a given model and data.

#### 3.1.2.1. Vague priors

A vague prior, is said to be a diffuse prior which difficult to define, after considering WIPs. In 2005, Lambert, Sutton, Burton, Abrams and Jones introduce the first formal move from vague to WIPs. After conjugate priors were introduced by Raiffa and Schlaifer in 1961 which most applied Bayesian has applied vague priors, parameterized to estimate the idea of uninformative priors.

Normally, a vague prior is a conjugate prior together with a large size parameter. Howsoever, if the sample size is small then vague priors may be problems. All most problems about vague priors and small sample size are implicated with scale rather than location. The problem can be particularly acute in random-effects models which it is used rather loosely in this here to imply exchangeable, hierarchical and multilevel structures. A vague prior is defined as commonly being a conjugate prior that is intent to estimate an uninformative prior and without two goals of regularization and stabilization.

#### 3.1.3. Least informative prior

Least informative priors(for short term LIP) is applied here to describe a class of prior in which the aim is to minimize the amount of subjective information content, and to apply a prior that is determined only by the model and observed data. The rationale for using LIPs is often called to let the data speak for themselves. LIPs are preferred by objective Bayesians. LIPs are contains Flat Priors [12], Hierarchical Prior [4], Jeffreys Prior [2], MAXENT [5] and Reference Priors [68] etc.

#### 3.1.4. Uninformative prior

Traditionally, most of the above descriptions of prior distributions were classified as uninformative priors. However, uninformative priors do not really exist (see in [9]) and all priors are informative in some ways. Moreover, there have been various names associated with uninformative priors including diffuse, minimal, non-informative, objective, reference, uniform, vague, and perhaps weakly informative etc.

#### 3.1.5. Proper and improper priors

It is important for the prior distribution to be proper. A prior distribution, p(θ), is improper when ∫p(θ) =  ∞ .

Before, an unbounded uniform prior distribution is an inappropriate prior distribution since p(θ) ∝ 1, for θ ∈ (−∞, ∞). An inappropriate prior distribution may be cause an inappropriate posterior distribution. If the posterior distribution is inappropriate, then inferences are invalid.

To determine the propriety of a joint posterior distribution, the marginal likelihood should be finite for any y. Again, the marginal likelihood is p(y) = ∫ p(y|Θ)p(Θ). Although inappropriate prior distributions can be applied, it is good practice to avoid them.

### 3.2. Likelihood

To completely for the definition of a Bayesian, both the prior distributions and the likelihood must be estimated or completely specified. The likelihood or p(y|Θ), contains the available information provided by the sample. The likelihood is pyΘ=i=1npyiΘ.

The data yeffect to the posterior distribution p(Θ| y) only through the likelihood p(Θ| y). In this way, Bayesian inference believes the likelihood principle which states that for a given sample of data, any two probability models p(Θ| y) that have the same likelihood yield the same inference for Θ.

### 3.3. Posterior distribution

Recent theoretical and applied overviews of Bayesian statistics, including many examples and uses of posterior distributions, see [1012]. The posterior distributions for decision-making about home radon exposure are discussed in [13].

The posterior distribution summarizes the current state of knowledge about all the uncertain quantities in a Bayesian analysis. Analytically, the posterior density is the product of the prior density and the likelihood. In a complicated analysis, the joint posterior distribution can be summarized by a set of Lsimulation draws of the vector of uncertain quantities w1 , w2 ,  …  , wJ, as illustrated in the following matrix:

lw1w2wJ
1...
2..
.
L...

The marginal posterior distribution for any unknown quantity wlcan be summarized by its column of Lsimulation draws. In many examples it is not necessary to construct the entire table ahead of time; rather, one creates the Lvectors of posterior simulations for the parameters of the model and then uses these to construct posterior simulations for other unknown quantities of interest, as necessary.

## 4. Application to games theory

In this section, we present the application of Bayesian inference to the real world problems such as Bayesian Game as follows.

### 4.1. The classical games

The basic contents of the n-person game was presented by John Forbes Nash [14] in 1950. Also, he first shows the existence of equilibrium for this model when the player’s preferences are representable by continuous quasi-concave utilities and the sets of strategy are simplex. The definition of an n-person game can be written as below.

Definition 4.1

The normal form of an n− person game is Xirii=1n, where for each i ∈ {1, 2,  … , n}, the set of individual strategies of player idenoted by Xiwhich Xiis a non-empty set and riis the preference relation on X ≔ ∏i ∈ IXiof player i.

The individual preferences riare usually represented by utility functions, i.e. for each i ∈ {1, 2,  … , n} there exist a real valued function ui : X ≔ ∏i ∈ IXiRsuch that:

xriyui(x)ui(y),x,yX.

Then the normal form of n− person game is transformed to Xiuii=1n.

The solution of this game is called Nash equilibrium as below.

Definition 4.2

The Nash equilibrium for the game Xiuii=1nis a point x ∈ Xwhich satisfies for each i ∈ {1, 2,… , n} : ui(x) ≥ ui(x, xi) for each xi ∈ Xi.

The following theorem offers sufficient conditions for the existence of Nash equilibrium.

Theorem 4.3

LetΓ=Xiuii=1nbe a n-person game and denoted by f the real-valued function on X × X defined byfxy=Σi=1nuixiyi. Let us assume that

1. for each i ∈ {1, 2,  … , n}, Xi is a non-empty compact convex subset of a Hausdorff linear topological space;

2. for each i ∈ {1, 2,  … , n}, ui(⋅, xi) is continuous on Xi = ∏i ≠ jXj for each fixed xi ∈ Xi;

3. Σi=1nuiis continuous on X;

4. f(x, ⋅) is quasi-concave on X, for each x ∈ X.

Then, Γ has an equilibrium.

Proof.See in [34].

Next, we present some examples of Nash equilibrium for two persons game as follows.

Example 4.4

The battle of the sexes game has two Nash equilibrium (MT, FT), (MS, FS) with (3, 2) and (2, 3), where “Male like playing tennis” denoted by MT, “Male like shopping” denoted by MS, “Female like playing tennis” denoted by FTand “Female like shopping” denoted by FS, see in Figure 1.

Example 4.5

The oligopoly behavior game is a unique Nash equilibrium (Aa, Ba) where “A coffee shop use a strategy for don’t advertising” denoted by Ad, “A coffee shop use a strategy for advertising” denoted by Aa, “A coffee shop use a strategy for do not advertising” denoted by Bd, and “A coffee shop use a strategy for advertising” denoted by Ba, see in Figure 2.

### 4.2. The Bayesian games

For a long time, we have been supposed that everything in the game was normal knowledge for everyone playing. However, real players may have private information about their own payoffs, their type or preferences, etc. The way to modeling this situation of asymmetrical information is by recurring to the concept was defined by Harsanyi in 1967. The key is to introduce a move by the nature, which changes the uncertainty by converting an asymmetrical information problem into an imperfect information problem. The concept is the nature moves determining players’ types, a concept that collects all the private information relevant them (i.e. payoffs, preferences, beliefs of another players, etc.).

Definition 4.6

The normal form of Bayesian games with incomplete information include:

1. the players i ∈ {1, 2,  … , I};

2. the set of finite action for each player ai ∈ Ai;

3. the finite type set for each player θi ∈ Θi;

4. a probability distribution on types p(θ)

5. ui : A1 × A2 ×…× AI × Θ1 × Θ2 ×…× ΘIR, where uiis utilities function.

It is important to discuss some parts of the definition. Players’ types comprise all relevant information about some player’s private characteristics. The type of θiis only observed by player iwho uses this information both to make decisions and to update itself beliefs about the likelihood of opponents’ types.

Combining actions and types for each player it is possible to create the strategies. Strategies will be given by si : ΘiAi, with elements si(θi) where Θiis the type space and Aiis the action space. A strategy may determine different actions to different types. Lastly, utilities are computed by each player by taking expectations over types using itself own conditional beliefs about opponents’ types. Hence, if player iuses the pure strategy si, other players use the strategies siand player i’s type is θi, the expected utility can be presented as follows

Euisisiθi=θiΘiuisisiθiθiθipθiθi.

A Bayesian Nash Equilibrium (for short term: BNE) is basically the same concept than a Nash Equilibrium with the addition that players need to take expectations over opponents’ types as follows.

Definition 4.7

A Bayesian Nash Equilibrium is a Nash Equilibrium of a Bayesian Game, i.e. Euisis1θiEuisisiθifor all siSiand for all types θioccurring with positive probability.

The following theorem for the existence of Bayesian Nash Equilibrium.

Theorem 4.8

Every finite Bayesian Games has a Bayesian Nash Equilibrium.

Example 4.9

Consider the Bayesian games as follows:

1. Nature decides that the payoffs are as in matrix I or II, with probabilities;

2. ROW is informed of the choice of nature but COL is not;

3. The choices of ROW include Uor Dand the choices of COL include Lor Rwhere these choices are made simultaneously;

4. Payoffs are as in the matrix chosen from nature.

For any of the Bayesian games, we will find all BNE. All equilibrium in mixed behavioral strategies can be written as.

Matrix I:

LR
U(1, 1)(0, 0)
D(0, 0)(0, 0)

Matrix II:

LR
U(0, 0)(0, 0)
D(0, 0)(2, 2)

#### 4.2.1. Pure strategy BNE

First, we will deflate the case of incomplete information problem as a static extended game with all of possible strategies: Γ̂. It can be presented follow Harsanyi, that the Nash Equilibrium of Γ̂is the same equilibrium of the imperfect game presented. The idea is to deflate a game such that all the ways the game can follow is considered in the extended game Γ̂.

The first step is to define the strategies for all player.

Since he does not know in which matrix the game is played, so, COL has only two strategies which contain Land R.

ROW knows in which Matrix the game occurs, and the strategies are UU, UD, DUand DDwhere UDis played Uin Matrix I and Din Matrix II.

The probability knowledge, the nature locates the game in any matrix. The new extended game Γ̂can be shown as:

LR
UU1212(0, 0)
UD1212(1, 1)
DU(0, 0)(0, 0)
DD(0, 0)(1, 1)

Remember that DUis a dominated strategy for ROW. After displacement that possibility, the game has 3 pure Nash Equilibrium as follows {(UU; L); (UD; R); (DD; R)}.

#### 4.2.2. Mixed strategy BNE

Sequent to obtain the mixed strategies we will make another kind of analysis and try to repeat the three pure BNE obtained before.

Suppose the probabilities of playing each action are as displayed in the matrices as below, where yis the probability COL plays L, if the game is in Matrix I then xis the probability ROW plays Uand if the game is in Matrix II then zis the probability ROW plays U.

#### 4.2.3. Player’s best respones

• In Matrix I: we get ROW’s best response as follows

ROW would play U, x = 1, if 1y + 0(1 − y) > 0, then y > 0, which can be concluded as:

1. a.    if y > 0, then x = 1;

2. b.    if y = 0, then x ∈ [0, 1].

• In Matrix II: we get ROW’s best response as follows

ROW would play D, z = 0 if 0 < 2(1 − y) then y < 1 which can be concluded as:

1. c.     if y < 1, then z = 0;

2. d.    if y = 1 then z ∈ [0, 1].

• In Matrix I and II: we get COL’s best response as follows

COL would play L, y = 1 if

121x+01x+120z+01z>120x+01x+120z+21z

then x > 2(1 − z) which can be summarized as:

1. e.    if x = 2(1 − z), then y ∈ [0, 1];

2. f.    if x > 2(1 − z), then y = 1;

3. g.    if x < 2(1 − z), then y = 0.

Next, we can check each the possibilities in order to find the Nash Equilibrium, such as those strategies stable for any players. Let us start by checking COL’s strategies since there are less combinations.

#### 4.2.4. Mixed equilibrium

Case 1:

If y = 0, we have b. x ∈ [0, 1] and c. z = 0. Here, we want to check this is a equilibrium from COL’s point of view. By g., we can see that if z = 0, then x < 2 which always hold and that y = 0.

This Nash equilibrium supports two of the three pure BNE found before: (DD, R), which is the same as y = 0, x = 0 and z = 0 and (UD, R) which is the same as y = 0, x = 1 and z = 0.

Thus, we get Nash equilibrium of y = 0, x ∈ [0, 1] and z = 0.

There are many BNE in which column plays Rand row plays xU + (1 − x)D, when x ∈ [0, 1] if Matrix I occurs and Dif Matrix II occurs.

Case 2:

If y = 0, we have d. z ∈ [0, 1] and from a. x = 1.

From f., we can see that when x = 1, then it should be the case that z12in order to be true that y = 1. Hence, these BNE are restricted to y = 1, z121and x = 1.

This BNE supports the third pure Nash Equilibrium found before: (UU, L), which is the same as y = 1, x = 1 and z = 1.

There are many BNE in which column plays Land row plays Uif Matrix I occurs and zU + (1 − z)D, where z121if Matrix II occurs.

Case 3:

If y ∈ (0, 1), we have a. x = 1 and c. z = 0. By e., we can see that in order ybelongs to [0, 1] it should be the case that x = (1 − z). However it is impossible this equality to hold if both z = 0 and x = 1.

Therefore, the case if y ∈ (0, 1) is not a Bayesian Nash equilibrium.

### 4.3. Abstract economy model

Later, the existence of social equilibrium was proved Debreu [15]. Also Arrow and Debreu [16] proved the existence of Walrasian equilibrium. The classical abstract economy game introduced by Shafer and Sonnenschein [17] or Borglin and Keiding [18] consists of a finite set of agents, each characterized by certain constraints and preferences, explained by correspondences. Following many previous authors ideas, they studied the existence of equilibrium for generalized games (see, for example, [1927] and the references therein). Now, we show some definitions of an abstract economy model and equilibrium of this model as follows. Let the set of agents be the finite set {1, 2,  … , n}. For each i ∈ {1, 2,  … , n} let Xibe a non-empty set.

Definition 4.10

An abstract economy Γ=XiAiPii=1nis defined as a family of nordered triplets (Xi, Ai, Pi), where for each i ∈ I:

1. Ai : ∏i ∈ IXi→2Xiis constraint correspondence and

2. Pi : ∏i ∈ IXi→2Xiis preference correspondence.

Definition 4.11

An equilibrium for Γis a point x ∈ ∏i ∈ IXiwhich satisfies for each i ∈ {1, 2,  … , n}:

1. x ∈ Ai(x);

2. Ai(x) ∩ Pi(x) = ∅.

Theorem 4.12

LetΓ=XiAiPii=1nbe an abstract economy which satisfies, for each i ∈ {1, 2,  … , n}:

1. Xi is a non-empty compact convex subset inRl;

2. Ai is a continuous correspondence;

3. for each x ∈ X, Ai(x) is non-empty compact and convex;

4. Pi has an open graph in X × Xi and for each x ∈ X, Pi(x) is convex;

5. for each x ∈ X, xi ∉ Pi(x).

Then,Γ has an equilibrium.

Proof.See in [34].

### 4.4. Fuzzy games

The first concept of a fuzzy set was introduced by Zadeh [28] in 1965. Fuzzy set theory has been shown to be a gainful tool to describe situations in which the data are imprecise or vague. The theory of fuzzy sets has become a well framework for studying results concerning fuzzy equilibrium existence for abstract fuzzy economies. The first study of a fuzzy abstract economy (or a fuzzy game) has been studied by Kim and Lee in [29], they shown the existence of the equilibrium for 1-person fuzzy game. Also Kim and Lee [29] shown the existence of equilibrium for generalized games when the constraints or preferences are vague due to the agent’s behavior. In 2009, Patriche [30] studied the Bayesian abstract economy game and proved the existence of equilibrium for an abstract economy game with differential information and a measure space of agents. However, the existence of random fuzzy equilibrium for fuzzy game has not been studied so far. In 2013, Patriche [31] defined the Bayesian abstract economy game and proved the existence of the Bayesian fuzzy equilibrium for this game. Also, Patriche [32] defined the new Bayesian abstract fuzzy economy game and proved the existence of the Bayesian fuzzy equilibrium for this game which it is characterized by a private information set, an action fuzzy mapping, a random fuzzy constraint one and a random fuzzy preference mapping. Recently, Patriche [33] defined the fuzzy games and applications to systems of generalized quasi-variational inequalities problem. The Bayesian fuzzy equilibrium concept is an extension of the deterministic equilibrium. She also generalized and extended the former deterministic models introduced by Debreu [15], Shafer and Sonnenschein [17] and Patriche [34]. Very recently, Saipara and Kumam [35] introduced the model of general Bayesian abstract fuzzy economy for product measurable spaces, and proved the existence for Bayesian fuzzy equilibrium of this model as follows.

For each i ∈ I, let ΩiZibe a measurable space, ΩZbe the product measurable space where ΩiIΩi,ZiIZiand μis a probability measure on ΩZ. Let Ydenote the strategy or commodity space, where Yis a separable Banach space.

Let Ibe a non-empty finite set (the set of agents). For each i ∈ I, let Xi : ΩiF(Y) be a fuzzy mapping, and let zi ∈ (0, 1].

Let LXi = {xi ∈ S(Xi(⋅))zi : xiis Σi-measurable}. Denote by LX = ∏i ∈ ILXiand by the set ∏i ≠ jLXj. An element xiof LXiis called a strategy for agent i. The typical element of LXiis denoted by xiand that of (Xi(ωi))ziby xi(ωi) (or xi). We can define a general Bayesian abstract fuzzy economy model of product measurable spaces as follow.

Definition 4.13

A general Bayesian abstract fuzzy economy model of product measurable spaces is defined as follows:

Γ=ΩiZiiIμXiΣiAiaiPipiziiI,

where Iis non-empty finite set (the set of agents) and:

1. Xi : Ωi → F(Y) is a action (strategy) fuzzy mappingof agent i;

2. Σiis a sub σ-algebra of Z=iIZi, which denotes the private informationof agent i;

3. for each ωi ∈ Ωi , Ai(ωi, ⋅) : LX → F(Y) is the random fuzzy constraint mappingof agent i;

4. for each ωi ∈ Ωi , Pi(ωi, ⋅) : LX → F(Y) is the random fuzzy preference mappingof agent i;

5. ai : LX → (0, 1] is a random fuzzy constraint function, and pi : LX → (0, 1] is a random fuzzy preference functionof agent i;

6. zi ∈ (0, 1] is such that for all ωixΩi×LX,Aiωix˜aix˜Xiωiziand Piωix˜pix˜Xiωizi.

The Bayesian fuzzy equilibrium for a general Bayesian abstract fuzzy economy model of product measurable spaces is defined as follows.

Definition 4.14

A Bayesian fuzzy equilibrium for Γis a strategy profile x˜LXsuch that for all i ∈ I,

1. x˜ωiclAiωix˜aix˜μa.e.;

2. Aiωix˜aix˜Piωix˜pix˜=μa.e..

Theorem 4.15

Let I be a non-empty finite set. Let the family

Γ=ΩiZiiIμXiΣiAiaiPipiziiIbe a general Bayesian abstract economy model of product spaces satisfy (a)-(j). Then, there exists a Bayesian fuzzy equilibrium for Γ.

For each i ∈ I, the following conditions are sastisfied:

1. Xi : Ωi → F(Y) is such that ωiXi(ωi)zi : Ωi → 2Y is a non-empty convex weakly compact-valued and integrably bounded correspondence;

2. Xi : Ωi → F(Y) is such that ωiXi(ωi)zi : Ωi → 2Y isilower measurable;

3. For eachωix˜Ωi×LX,Aiωix˜aix˜is convex and has a non-empty interior in the relative norm topology of(Xi(ωi))zi ;

4. the correspondenceωix˜Aiωix˜aix˜:Ωi×LX2Yhas a measurable graph, i.e.,ωix˜yΩi×LX×Y:yAiωix˜aix˜FiBLXBY,whereBωiLXis the Borel σalgebra for the weak topology on LX andBYis the Borel σalgebra for the norm topology on Y;

5. the correspondenceωix˜Aiωix˜aix˜has weakly open lower sections, i.e., for each ωi ∈ Ωi and for each y ∈ Y, the setAiωix˜aix˜1ωix˜=x˜LX:yAiωix˜aix˜is weakly open in LX;

6. For eachωiΩi,x˜clAiωix˜aix˜:LX2Yis upper semicontinuous in the sense that the setx˜LX:clAiωix˜aix˜is weakly open in LX for every norm open subset V of Y;

7. the correspondenceωix˜Piωix˜pix˜:Ωi×LX2Yhas open convex values such thatPiωix˜pix˜Xωizifor eachωix˜Ωi×LX;

8. the correspondenceωix˜Piωix˜pix˜:Ωi×LX2Yhas a measurable graph;

9. the correspondenceωix˜Piωix˜pix˜:Ωi×LX2Yhas weakly open lower sections, i.e. for each ωi ∈ Ωi and for each y ∈ Y, the set(Piωix˜pix˜)1ωiy=x˜LX:yPiωix˜pix˜is weakly open in LX;

10. For eachx˜iLXi,for eachωiΩi,x˜iAiωix˜aix˜Piωix˜pix˜.

Proof.See in [35].

Moreover, in 1960, Fichera and Stampacchia first introduced the variational inequalities problem, this issue has been widely studied. Next, the basic concept of variational inequalities for fuzzy mappings was first introduced by Chang and Zhu [36] in 1989. In the topic of variational inequalities problem, there are many mathematicians who studied this topic (see, for example, [37, 38]). In 1993, the concept of a random variational inequality was introduced by Noor and Elsanousi [39]. Recently, Patriche [31] used the model of the Bayesian abstract fuzzy economy to prove the existence of solution for the two types of random quasi-variational inequalities with random fuzzy mappings.

## 5. Conclusion

The main objectives of this chapter was introduced the concept of Bayesian inference and application to some real world problems. In this chapter, we were presented about the basic concept of Bayesian inference which it can be application to the Bayesian game and a general Bayesian abstract fuzzy economy game or a fuzzy game. For application to Bayesian game, we were shown the solution of Bayesian Nash Equilibrium (BNE) for a Bayesian game with examples. Finally, we were shown the existence of Bayesian fuzzy equilibrium for a fuzzy game.

## Acknowledgments

This project was supported by the Theoretical and Computation Science (TaCS) Center under Computational and Applied Science for Smart Innovation Cluster (CLASSIC), Faculty of Science, KMUTT. Moreover, Poom Kumam was supported by the Thailand Research Fund (TRF) and the King Mongkut’s University of Technology Thonburi (KMUTT) under the TRF Research Scholar Award (Grant No. RSA6080047).

chapter PDF
Citations in RIS format
Citations in bibtex format

## More

© 2017 The Author(s). Licensee IntechOpen. This chapter is distributed under the terms of the Creative Commons Attribution 3.0 License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

## How to cite and reference

### Cite this chapter Copy to clipboard

Wiyada Kumam, Plern Saipara and Poom Kumam (November 2nd 2017). Bayesian Inference Application, Bayesian Inference, Javier Prieto Tejedor, IntechOpen, DOI: 10.5772/intechopen.70530. Available from:

### Related Content

Next chapter

#### Node-Level Conflict Measures in Bayesian Hierarchical Models Based on Directed Acyclic Graphs

By Jørund I. Gåsemyr and Bent Natvig

First chapter

#### Making a Predictive Diagnostic Model for Rangeland Management by Implementing a State and Transition Model Within a Bayesian Belief Network (Case Study: Ghom- Iran)

By Hossein Bashari

We are IntechOpen, the world's leading publisher of Open Access books. Built by scientists, for scientists. Our readership spans scientists, professors, researchers, librarians, and students, as well as business professionals. We share our knowledge and peer-reveiwed research papers with libraries, scientific and engineering societies, and also work with corporate R&D departments and government entities.