Next Article in Journal
Some Remarks on Odd Edge Colorings of Digraphs
Next Article in Special Issue
Combined Games with Randomly Delayed Beginnings
Previous Article in Journal
On a Metric Affine Manifold with Several Orthogonal Complementary Distributions
Previous Article in Special Issue
Mixed Mechanisms for Auctioning Ranked Items
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cooperative Stochastic Games with Mean-Variance Preferences

by
Elena Parilina
1,2,3,*,† and
Stepan Akimochkin
4,†
1
Department of Mathematical Game Theory and Statistical Decisions, Saint Petersburg State University, 7/9 Universitetskaya nab., 199034 Saint Petersburg, Russia
2
School of Mathematics and Statistics, Qingdao University, Qingdao 266071, China
3
Institute of Applied Mathematics of Shandong, Qingdao 266071, China
4
Australia and New Zealand Banking Group Limited, 242 Pitt Street, Sydney, NSW 2000, Australia
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2021, 9(3), 230; https://doi.org/10.3390/math9030230
Submission received: 30 November 2020 / Revised: 6 January 2021 / Accepted: 20 January 2021 / Published: 25 January 2021
(This article belongs to the Special Issue Statistical and Probabilistic Methods in the Game Theory)

Abstract

:
In stochastic games, the player’s payoff is a stochastic variable. In most papers, expected payoff is considered as a payoff, which means the risk neutrality of the players. However, there may exist risk-sensitive players who would take into account “risk” measuring their stochastic payoffs. In the paper, we propose a model of stochastic games with mean-variance payoff functions, which is the sum of expectation and standard deviation multiplied by a coefficient characterizing a player’s attention to risk. We construct a cooperative version of a stochastic game with mean-variance preferences by defining characteristic function using a maxmin approach. The imputation in a cooperative stochastic game with mean-variance preferences is supposed to be a random vector. We construct the core of a cooperative stochastic game with mean-variance preferences. The paper extends existing models of discrete-time stochastic games and approaches to find cooperative solutions in these games.

1. Introduction

The class of stochastic games was initially introduced by L. Shapley [1]. He considered two-player zero-sum stochastic games with finite state space and finite strategy spaces and proved the existence of optimal stationary strategies when players’ payoff function is discounted mathematical expected payoff. This result has been extended to the case of n-player games by Fink [2], Takahashi [3], Sobel [4]. The results in stochastic games of the recent 60 years, mostly for non-cooperative case, are discussed by Neyman and Sorin [5], Solan [6], and Solan and Vieille [7].
Players’ payoffs in stochastic games are stochastic variables. In most papers, the expected discounted payoff is considered as a measure of a player’s payoff in a stochastic game. In this case, it is supposed that the players are risk neutral. Hence, the risk averse and risk loving attitudes cannot be modeled considering only expectation as a criterion. We propose a model of a stochastic game in which the players’ attentions to risk are taken into account. We use a mean-variance preference popular in decision making theory. In particular, this type of preference for random payoffs is used in financial risk management, in a portfolio decision theory proposed by Markowitz [8]. The risk allocation problem in portfolio management based on cooperative game theory approach is also examined by Csóka et al. [9].
We consider n-person stochastic games with finite sets of actions in any state of the game. A mean-variance stochastic game in discrete time with the payoff function, which is a linear combination of expectation and standard deviation of the payoff, is proposed. The goal of the paper is to construct the cooperative version of a mean-variance stochastic game and define the core as a cooperative solution of such a cooperative game. In the paper, we first define the players’ payoffs according to mean-variance preferences in a non-cooperative setting, and second, we define a cooperative game in the form of characteristic function. Then, we determine a solution of a cooperative stochastic game using the method proposed by Suijs and Borm [10], and Suijs et al. in [11]. The contribution of the paper is in proposing the theory of cooperative stochastic games with mean-variances preferences. The research question about the stability of the cooperative solution over time is open and briefly discussed in Section 5.
First, we consider a non-cooperative stochastic game with mean-variance utility function. Second, we define the cooperative version of the game constructing characteristic function using maxmin approach (see Neumann and Morgenstern [12]). This approach supposes that the value of characteristic function for any coalition is the maxmin payoff in zero-sum stochastic game between the coalition and anti-coalition. As the players may have different coefficients before standard deviations in the utility functions, the summation of the utilities does not make sense. When a coalition is formed, the maximal utility of the coalition payoff over the coalition members is used to measure the utility of the coalition payoff. There are other approaches of defining characteristic function in dynamic games proposed by Gromova and Petrosyan [13], Petrosjan and Zaccour [14] and examined by Reddy and Zaccour [15]. However, applications of these approaches need corrections, considering the stochastic nature of payoffs and mean-variance utility functions. Cooperative stochastic games with expected payoff as a utility function are introduced by Petrosjan [16] and Petrosjan and Baranova [17], and then extended by Parilina [18] and Parilina and Tampieri [19] by examining the stability of cooperative solutions in stochastic dynamic settings. As proposed by Suijs et al. [11], we consider the core as a solution of a cooperative game. The conditions of node-consistency of the core in dynamic games played over event trees with stochastic nature are obtained by Parilina and Zaccour [20]. Recently, the sufficient conditions of strong subgame consistency of the core are obtained by Parilina and Petrosyan [21].
Then, we define a solution of a cooperative stochastic game. The idea is briefly described below. Suijs et al. introduce a new class of cooperative games arising from cooperative decision making problems in the stochastic environment [10,11]. They introduce some types of preferences to order stochastic payoffs and model the games with risk neutral, as well as risk averse and risk loving players. Suijs et al. in [11] define the core of the cooperative game with stochastic payoffs and obtain the necessary and sufficient conditions for the non-emptiness of the core. The deterministic equivalent of a stochastic payoff with some properties is introduced by Suijs and Borm [10]. The nucleolus for this type of games is introduced by Suijs et al. [11]. The Shapley-like solutions for the cooperative game with stochastic payoffs are proposed by Timmer et al. [22]. The model of transferable utility game with uncertainty is introduced by Habis and Herings [23]. They propose a solution concept of a weak sequential core for this class of games and show that it is non-empty if all ex post TU-games are convex.
The goal of the paper is to propose a scheme of construction of a cooperative version of a stochastic game, in which not only expectation but also standard deviation of players’ payoffs is taken into account. The idea of considering risk-sensitive players recently appears in different variations in some papers on dynamic and differential games. The risk sensitive nonzero-sum continuous time stochastic games are considered by Wei [24], where the existence of Nash equilibrium is proved. The risk-sensitive average payoff criterion for zero-sum (see Bäuerle and Rieder [25]) and for the nonzero-sum discrete-time stochastic games with a denumerable state space (see Wei and Chen [26]), in which an existence of the Nash equilibrium is examined. The risk sensitive Nash equilibria in stochastic games are discussed by Nowak [27]. The class of risk sensitive stochastic games with overlapping generations is considered by Jaśkiewicz and Nowak [28]. Recently, the Nash equilibria for stochastic games of resource extraction in which risk coefficient is included into utility function are found by Asienkiewicz and Balbus [29]. In mean-field games, the model with both expectation and variance in payoff criteria is proposed by Saldi et al. [30].
The paper is organized as follows. The class of mean-variance stochastic games in discrete time is introduced in Section 2. The propositions providing formulae to calculate a variance and covariance of players’ payoffs in stochastic games are given in Section 3. In Section 4, we define a cooperative mean-variance stochastic game and an imputation as a cooperative solution of the game. We briefly conclude in Section 5. The proofs of all propositions are given in the appendix.

2. Stochastic Games with Mean-Variance Preferences

2.1. Stochastic Game

Define a finite stochastic game G using the following notations. The finite set of players is N = { 1 , , n } , and the finite set of states is S = { 1 , 2 , , t } . In any state s S , an n-person normal-form game N , { A i s } i N , { K i s } i N is defined, where A i s is the finite set of Player i’s actions in state s, K i s a 1 s , , a n s = K i s ( a s ) R is a payoff function of Player i in state s, i N , s S . The action profile a s = ( a 1 s , , a n s ) belongs to the set A s = i N A i s .
The transition map π is T Δ ( S ) function, where T = { ( s , a s ) | s S ; a s A s } and Δ ( S ) is a probability distribution on S. For each pair ( s , a s ) , mapping π identifies the vector of probabilities ( π ( 1 | s ; a s ) , …, π ( s | s ; a s ) , …, π ( t | s ; a s ) ) , where π ( s | s ; a s ) is the probability that state s will be realized at the next stage if at the present stage (state s) action-profile a s has been realized, π ( s | s ; a s ) 0 and s S π ( s | s ; a s ) = 1 for any a s A s and s S .
Let vector p 0 = ( p 1 0 , , p t 0 ) be a vector of initial distribution on S, and p s 0 is the probability that state s will be realized at the first stage of the game, p s 0 0 for any s S , and s S p s 0 = 1 .
Let all players have the same discount factor δ ( 0 , 1 ) for their payoffs.
Definition 1.
The stochastic game G is defined by
N , S , π , p 0 , δ .
Definition 2.
Stochastic subgame G s is a finite stochastic game ( 1 ) with initial distribution vector p 0 , such that p s 0 = 1 . Subgame G s is a stochastic game starting from state s S .
The implementation of a stochastic game can be briefly described in the following way. First, the chance moves and chooses the initial state of the game according to vector p 0 . Let the initial state be s S . Second, the first stage of game G is realised, i.e., the normal-form game N , { A i s } i N , { K i s } i N is played. Players simultaneously choose their actions from the action sets { A i s } i N . Let Player i’s action in state s be a i s A i s , s S . Therefore, an action profile a s = ( a 1 s , , a n s ) i N A i s is implemented. Player i obtains her payoff K i s ( a s ) , i N . Then, the transition to the next stage is realized. The probability that state s will appear at the next stage is equal to π ( s | s ; a s ) , s S . After that, the game proceeds in the described way.
Assume that players know all the parameters of the game. We consider a class of stationary strategies. Let Player i’s strategy in game G be a mapping η i ( · ) , such that η i ( s ) = a i s A i s , s S . Denote the set of stationary strategies of Player i as H i . The stationary strategy defines an action at any stage depending only on which state is realized at the stage, i.e., η i : s a i s A i s , s S , and not depending on the stage number and the history of the stage. A vector of stationary strategies is a strategy profile η = ( η 1 , , η n ) H 1 × × H n in stochastic game G. Obviously, η is a strategy profile in any subgame G s .

2.2. Mean-Variance Payoff Functions

Players’ payoffs in stochastic games are random variables and they are functions of stationary strategy profile η . Denote Player i’s payoff in stochastic game G as ξ ¯ i = ξ ¯ i ( η ) and Player i’s payoff in stochastic subgame G s as ξ i s = ξ i s ( η ) for any s S . Random variables ξ ¯ i and ξ i = ( ξ i 1 , , ξ i t ) satisfy the following equation:
ξ ¯ i ( η ) = p 0 ξ i ( η ) .
To find a player’s “optimal” strategy, it is necessary to determine how to compare random payoffs. The most common way is to measure a random payoff by its mathematical expectation, assuming the risk neutrality of the players. Strictly speaking, ξ i is preferred to ψ i or ξ i E ψ i , if and only if E ( ξ i ) > E ( ψ i ) , where E ( ξ ) is the mathematical expectation of random variable ξ . This preference is complete and transitive but it does not allow one to model different attentions to risk.
Example 1.
One may compare random payoffs using the principle of stochastic domination. Let ξ i , ψ i L 1 ( R ) be random payoffs of Player i and F ξ i ( · ) and F ψ i ( · ) be cumulative distribution functions of ξ i and ψ i , respectively. It is said that ξ i stochastically dominates ψ i and write ξ i F ψ i , if and only if, for all x R inequality F ξ i ( x ) F ψ i ( x ) holds. This preference is incomplete, therefore, random payoffs can be incomparable with this preference. Another disadvantage of this preference is its computational complexity for a class of stochastic games. Thus, its application to the theory of stochastic games is questionable.
We use the method of ordering random payoffs proposed in [10,11]. Let every player i have a utility function u i of a random payoff such that u i : L 1 ( R ) R . For random payoffs ξ , ψ L 1 ( R ) , payoff ξ is preferred or equivalent to payoff ψ by Player i if and only if u i ( ξ ) u i ( ψ ) . In [10] the preferences ( i ) i N are considered for the cooperative games with stochastic payoffs and they are such that for any player i N there exists a utility function u i : L 1 ( R ) R satisfying properties:
  • For all ξ L 1 ( R ) : ξ i u i ( ξ ) .
  • For all ξ , ψ L 1 ( R ) : ξ i ψ if and only if u i ( ξ ) u i ( ψ ) .
  • For all d R : u i ( d ) = d .
  • For all ξ L 1 ( R ) : u i ( ξ u i ( ξ ) ) = 0 .
  • For all ξ L 1 ( R ) and all d, d R with d < d : u i ( d + ξ ) < u i ( d + ξ ) .
Expression ξ i u i ( ξ ) means that the random payoff ξ is equivalent to u i ( ξ ) , and ξ i ψ means that random payoff ξ is preferred or equivalent to random payoff ψ for Player i. In [10], the function u i ( · ) is called a deterministic equivalent of the preference. In our work, we call function u i ( ξ ) the utility function of a random payoff ξ . Obviously, the utility function u i ( ξ i ) = E ( ξ i ) satisfies properties 1–5 given above.
Example 2.
One may consider a preference α such that for any ξ, ψ L 1 ( R ) the relation ξ α ψ holds if α-quantile of c.d.f. F ξ of a random variable ξ is not less than the α-quantile of c.d.f. F ψ , i.e., u ( ξ ) = u α ξ = s u p { t | F ξ ( t ) < α } . This utility function also satisfies properties 1–5 given above.
Let Player i’s utility function of random payoff ξ i be
u i ( ξ i ) = E ( ξ i ) + b i V a r ( ξ i ) 1 2 ,
where V a r ( ξ i ) is the variance of ξ i , parameter b i R shows player’s attention to risk. Note that if b i = 0 , it is understood as Player i is risk neutral. If b i > 0 , then Player i is risk loving, and if b i < 0 , then Player i is a risk averse person. Denote the preference corresponding to utility function (2) as b i . The utility function (2) satisfies properties 1–5.
Preference b i is complete but not implied by F . The preference b i is attractive for applications because of its computation simplicity and possibility of modeling attention to risk. If b i = 0 for any i N , utility function (2) becomes u i ( ξ i ) = E ( ξ i ) , which coincides with preference E .

2.3. Stochastic Games with Mean-Variance Preferences

We define a stochastic game in which any player i N has a utility function u i : L 1 ( R ) R of form (2).
Definition 3.
Stochastic game G b with mean-variance preferences is defined by tuple
G b = N , S , π , p 0 , δ , b ,
where b = ( b 1 , , b n ) R n is a profile of parameters determining players’ utility functions (2) for any i N .
By G b s , s S , we define the subgame of stochastic game (3) with p 0 = ( p 1 0 , , p t 0 ) such that p s 0 = 1 .
In the next section, we present the formulae to calculate the utilities of random payoffs in stochastic game (3) and any subgame G b s . We also consider the case when players form a coalition and calculate its summarized stochastic payoff.

3. Expectation and Variance of Random Payoffs in Stochastic Games

Consider a strategy profile η = ( η 1 , , η n ) , where η i is a stationary strategy of Player i, such that η i ( s ) = a i s A i s for all s S , i N . First, find expectation E ( ξ i s ) of Player i’s payoff in subgame G s :
E ( ξ i s ( η ) ) = K i s ( a s ) + δ s S π ( s | s ; a s ) E ( ξ i s ( η ) )
or in a vector form:
E ( ξ i ( η ) ) = K i ( a ) + δ Π ( η ) E ( ξ i ( η ) ) ,
where E ( ξ i ( η ) ) = ( E ( ξ i 1 ( η ) ) , , E ( ξ i t ( η ) ) ) , K i ( a ) = ( K i 1 ( a 1 ) , , K i t ( a t ) ) . Equation ( 4 ) is equivalent to the following one:
E ( ξ i ( η ) ) = ( I δ Π ( η ) ) 1 K i ( a ) ,
where I is t × t identity matrix and Π ( η ) is a matrix of transition probabilities in stochastic game (1), which is a function of strategy profile η taking a form of
Π ( η ) = π ( 1 | 1 ; a 1 ) π ( s | 1 ; a 1 ) π ( t | 1 ; a 1 ) π ( 1 | t ; a t ) π ( s | t ; a t ) π ( t | t ; a t ) .
Matrix ( I δ Π ( η ) ) 1 is denoted as Π ^ ( η ) and it always exists for δ ( 0 , 1 ) .
Remark 1.
All the eigenvalues of stochastic matrix Π ( η ) are from the interval [ 1 , 1 ] . For the existence of matrix Π ^ ( η ) , it is necessary and sufficient that the determinant of matrix ( Π ( η ) I / δ ) be not equal to zero. Therefore, matrix ( Π ( η ) I / δ ) must not have an eigenvalue equal to 1 / δ . The last condition is satisfied because 1 / δ > 1 . Therefore, it cannot be an eigenvalue of stochastic matrix Π ( η ) .
An expectation of Player i’s payoff in stochastic game G equals
E ( ξ ¯ i ( η ) ) = p 0 E ( ξ i ( η ) ) = p 0 Π ^ ( η ) K i ( a ) .
The variance V a r ( ξ ¯ i ( η ) ) of Player i’s payoff ξ ¯ i ( η ) in game G can be found using the following proposition.
Proposition 1.
Let in stochastic game G or G b , Player i’s strategy be η i H i . The variance of Player i’s payoff is
V a r ( ξ ¯ i ( η ) ) = ( p 0 Π ˜ ( η ) K i ( a ) ) ( 2 Π ^ ( η ) I ) K i ( p 0 Π ^ ( η ) K i ( a ) ) 2 ,
where operation A B is the Hadamard product (Let A and B be m × n matrices with entries in R . The Hadamard product of A and B is an m × n matrix defined by [ A B ] i , j = [ A ] i , j [ B ] i , j for all 1 i m , 1 j n .) of matrices A and B of the same size, and Π ˜ ( η ) = ( I δ 2 Π ( η ) ) 1 .
Proof. 
See Appendix A.1. □
Corollary 1.
The variance V a r ( ξ i s ( η ) ) of Player i’s payoff in subgame G s , s S , can be found by formula ( 7 ) substituting p 0 = ( p 1 0 , , p t 0 ) s. t. p s 0 = 1 and p s 0 = 0 for any s s , s S .
For further calculations, in particular, for the construction of a cooperative form of stochastic game G b and subgames G b s , it is required to determine the random payoff of any coalition C N , which is denoted as ξ ¯ C and equals
ξ ¯ C ( η ) = i C ξ ¯ i ( η ) ,
and for subgame G s :
ξ C s ( η ) = i C ξ i s ( η ) .
We should notice that players’ payoffs ξ i s and ξ j s are not independent variables for i j , i , j C . The following proposition gives the formulae to calculate an expectation and variance of random payoff ξ ¯ C ( η ) of any coalition C.
Proposition 2.
The expectation of payoff ξ ¯ C ( η ) of coalition C in stochastic game G or G b is
E ( ξ ¯ C ( η ) ) = p 0 Π ^ ( η ) i C K i ( a ) .
The variance of payoff ξ ¯ C ( η ) of coalition C in stochastic game G or G b is
V a r ( ξ ¯ C ( η ) ) = i C V a r ( ξ ¯ i ( η ) ) + 2 l < m , l , m C C o v ( ξ ¯ l ( η ) , ξ ¯ m ( η ) ) ,
where variance V a r ( ξ ¯ i ( η ) ) of Player i’s payoff is defined by formula (7), and C o v ( ξ ¯ l ( η ) , ξ ¯ m ( η ) ) is a covariance of Players l and m’ payoffs, which is equal to
C o v ( ξ ¯ l ( η ) , ξ ¯ m ( η ) ) = ( p 0 Π ˜ ( η ) K l ( a ) ) ( Π ^ ( η ) 0.5 I ) K m ( a ) + ( p 0 Π ˜ ( η ) K m ( a ) ) ( Π ^ ( η ) 0.5 I ) K l ( a ) ( p 0 Π ^ ( η ) K l ( a ) ) ( p 0 Π ^ ( η ) K m ( a ) ) .
Proof. 
See Appendix A.2. □
Corollary 2.
The variance of payoff ξ C s ( η ) of coalition C in subgame G s or G b s , s S , can be found by formula ( 9 ) substituting p 0 = ( p 1 0 , , p t 0 ) s. t. p s 0 = 1 and p s 0 = 0 for any s S , s s .
Equation (6) and Proposition 1 provide formulae to calculate the utilities of players’ payoffs in stochastic games with mean-variance preferences for any strategy profile η . Proposition 2 provides formulae to calculate the utilities of coalition payoffs in stochastic games with mean-variance preferences.

4. Cooperative Stochastic Games with Mean-Variance Preferences

In this section, we construct a cooperative version of a stochastic game G b and subgame G b s , starting from state s S . Therefore, we determine characteristic function v ( · ) : 2 N R satisfying two properties:
  • v ( ) = 0 ,
  • (superadditivity) For any disjoint coalitions C , C N , C C = , inequality v ( C ) + v ( C ) v ( C C ) holds.
The value v ( C ) can be interpreted as a “worth” or “power” of coalition C when its members play together as a unit. We use “maxmin” approach to determine superadditive characteristic function [12].
The characteristic function for subgame G b s is denoted as v s ( · ) and defined as
v s ( C ) = max η C min η N C max i C u i ( ξ C s ( η C , η N C ) ) ,
for any coalition C N . Here, η C i C H i is a stationary strategy of coalition C and η N C i N C H i is a stationary strategy of coalition N C . Strategies ( η C , η N C ) form a strategy profile in any subgame G b s and game G b . In Equation (11) and further calculations, the utility function of any player i is of form (2), as required in the definition of stochastic game G b . An expression in the right-handed side of (11); we find maxmin of the maximal value of utility among players from C. Maxmin value is such that coalition C can guarantee by strategy η C when coalition N C plays against coalition C using strategy η N C .
Characteristic function for game G b is denoted by v ¯ ( · ) and it can be defined for coalition C using the following expression:
v ¯ ( C ) = max η C min η N C max i C u i ( ξ ¯ C ( η C , η N C ) ) .
This function is equal to the maximal value of utility function of random payoff ξ ¯ C ( η C , η N C ) (among the players from coalition C), which coalition C can guarantee by itself using stationary strategy η C i C H i if coalition N C plays against C in zero-sum game using strategy η N C i N C H i .
Cooperation in stochastic game G b assumes that players form the grand coalition N and play together to maximize the maximal utility of their total payoff ξ ¯ N . We call the stationary strategy profile η * in game G b a cooperative one if
η * = arg max η i N H i max i N u i ( ξ ¯ N ( η ) ) .
Therefore, the value of characteristic function v ¯ ( N ) can be rewritten as
v ¯ ( N ) = max i N u i ( ξ ¯ N ( η * ) ) .
The value v s ( N ) of the characteristic function v s ( N ) in cooperative subgame G b s is
v s ( N ) = max η i N H i max i N u i ( ξ N s ( η ) ) .
Definition 4.
Cooperative stochastic game with mean-variance preferences is a tuple N , v ¯ with characteristic function v ¯ : 2 N R defined by ( 12 ) and ( 14 ) , and v ¯ ( ) = 0 .
Definition 5.
Cooperative stochastic subgame with mean-variance preferences is a tuple N , v s with characteristic function v s : 2 N R defined by ( 11 ) and ( 15 ) , and v ( ) = 0 .
In cooperation, players with mean-variance preferences realize a cooperative stationary strategy profile η * given by (13).
The main problem of cooperative game theory is to find a ”fair” imputation of coalition N’s total payoff. In a classical cooperative game theory, values of characteristic function are non-random variables and there is a variety of solution concepts to determine, in some sense, ”fair” imputation.
The concept of a random payoff imputation is proposed in [11], in which it is supposed that the components of the imputation are random variables. We give a definition of an imputation of a random payoff ξ ¯ N ( η * ) .
Definition 6
(see [11]). An imputation of random payoff ξ ¯ N ( η * ) is a vector φ = ( φ 1 , , φ n ) , such that φ i = d i + r i ξ ¯ N ( η * ) E ( ξ ¯ N ( η * ) ) and d R n , r R n satisfy the following conditions:
1. 
i N d i = E ( ξ ¯ N ( η * ) ) ,
2. 
i N r i = 1 and r i 0 for any i N .
According to Definition 6, the payoff ξ ¯ N ( η * ) of coalition N in game G b is a random vector characterized by an allocation of expected payoff E ( ξ ¯ N ( η * ) ) and allocation of random variable ( ξ ¯ N ( η * ) E ( ξ ¯ N ( η * ) ) , which can be interpreted as a risk of random payoff ξ ¯ N ( η * ) . Denote the set of imputations in cooperative game G b as Φ ( N ) .
We use the statement from [11] about necessary and sufficient conditions that an imputation belongs to the core of cooperative stochastic game G b with mean-variance preferences as a definition of the core.
Definition 7.
An imputation φ belongs to the core of the cooperative game G b , if and only if the inequality
i C d i + r i u i ( ξ ¯ N ( η * ) ) E ( ξ ¯ N ( η * ) ) v ¯ ( C )
holds for any coalition C N .
To illustrate the idea of defining the core in cooperative stochastic game with mean-variance preferences, we provide a numerical example demonstrating calculations step by step.
Example 3.
Consider two-person stochastic game with two states N , A 1 s , A 2 s , K 1 s , K 2 s , s = 1 , 2 , with players’ payoffs:
s = 1 : ( 3 , 5 ) ( 3 , 10 ) ( 10 , 4 ) ( 8 , 6 ) , s = 2 : ( 2 , 4 ) ( 2 , 10 ) ( 7 , 3 ) ( 5 , 4 )
Transition probabilities are defined in matrices:
s = 1 : ( 0.7 , 0.3 ) ( 0.4 , 0.6 ) ( 0.4 , 0.6 ) ( 0.3 , 0.7 ) , s = 2 : ( 0.9 , 0.1 ) ( 0.4 , 0.6 ) ( 0.2 , 0.8 ) ( 0.3 , 0.7 ) ,
where element ( i , j ) of the first (second) matrix consists of transition probabilities from states 1 (2) to the states 1, 2 respectively under condition that Player 1 chooses strategy i and Player 2 chooses strategy j in state 1 (2).
Let discount factor δ be 0.9 and vector of initial distribution p 0 on state set { 1 , 2 } be ( 0.5 , 0.5 ) . Suppose that both players are risk averse and b 1 = 2 , b 2 = 1 .
Using condition ( 13 ) , we obtain η ¯ = ( η ¯ 1 , η ¯ 2 ) , where η ¯ 1 = ( 0 , 0 ) , η ¯ 2 = ( 1 , 1 ) , i.e., player 1 chooses action 2 in both states, and player 2 chooses action 1 in both states. The maximal value of the maximal utility function is equal to the following one:
v ¯ ( N ) = max i N u i ( ξ ¯ N ( η ¯ ) ) = 94.1709 .
Therefore, imputation defined by ( d , r ) is in the core of game G b according to Definition 7 if and only if the following inequality takes place for any C N :
i C ( d i + r i ( u i ( ξ ¯ N ( η ¯ ) ) E ( ξ ¯ N ( η ¯ ) ) ) v ¯ ( C ) .
If C = { 1 } , the inequality is
d 1 + r 1 ( u 1 ( ξ ¯ N ( η ¯ ) ) E ( ξ ¯ N ( η ¯ ) ) ) v ¯ ( { 1 } ) ,
and for C = { 2 } :
d 2 + r 2 ( u 2 ( ξ ¯ N ( η ¯ ) ) E ( ξ ¯ N ( η ¯ ) ) ) v ¯ ( { 2 } ) .
Taking into account Definition 6 of the imputation, the following conditions hold:
d 1 + d 2 = E ( ξ ¯ N ( η ¯ ) ) = 128.2 , r 1 + r 2 = 1 .
The following system defines the core of the cooperative stochastic game with mean-variance preferences:
d 1 + r 1 ( 60.1419 128.2000 ) 10.9955 , d 2 + r 2 ( 94.1709 128.2000 ) 25.4971 , d 1 + d 2 = 128.2000 , r 1 + r 2 = 1 ,
or in equivalent form:
d 1 [ 10.9955 + 68.0581 r 1 , 68.6739 + 34.0291 r 1 ] , d 2 = 128.2000 d 1 , r 1 [ 0 , 1 ] , r 2 = 1 r 1 .
The region of ( r 1 , d 1 ) , r 1 [ 0 , 1 ] , satisfying the first inequality of the latter system, is depicted in Figure 1.
We can easily notice that the core is non-empty and contains an infinite number of imputations. The imputation of the core is a vector
d 1 + r 1 ( ξ ¯ N ( η * ) 128.2 ) , d 2 + r 2 ( ξ ¯ N ( η * ) 128.2 ) ,
where d 1 , d 2 , r 1 , r 2 R and satisfy (16).

5. Conclusions

We have constructed a model of cooperative stochastic games with mean-variance preferences. The utility function for mean-variance preference takes into account players’ attention to risk. The risk sensitivity of a player is modeled by a linear combination of mathematical expectation and standard deviation of the payoff. We have provided formulas to calculate player’s and coalition utilities for any stationary strategy profile. We have defined a cooperative stochastic game with mean-variance preferences in the form of characteristic function determined by a maxmin approach. A particular cooperative solution proposed for the game is the core. The natural research question arising in the area of dynamic games is the sustainability and time-consistency of the cooperative solution over time. The problem of stability of a cooperative agreement in stochastic games with mean preferences is discussed by Parilina and Tampieri [19], Parilina [18], Parilina and Petrosyan [21], where conditions of stable cooperation are examined. The other research question is how to determine another cooperative solution like the Shapley value, nucleolus, etc., for stochastic games with mean-variance preferences and how to sustain them over time.

Author Contributions

Conceptualization, E.P. and S.A.; methodology, E.P. and S.A.; software, E.P. and S.A.; validation, E.P. and S.A.; writing—original draft preparation, E.P. and S.A.; writing—review and editing, E.P. and S.A.; project administration, E.P. and S.A. All authors have read and agreed to the published version of the manuscript.

Funding

The work of the first author was supported by the Shandong Province “Double-Hundred Talent Plan” (No. WST2017009).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Appendix A.1. Proof of Proposition 1

In the proof, we omit the symbol of strategy-profile η and action-profile a to simplify the description of calculations. We use the following notations: K i s = K i s ( a ) , E ( ξ ¯ i ) = E ( ξ ¯ i ( η ) ) , Π = Π ( η ) , Π ^ = Π ^ ( η ) , π ( l | m ; a m ) = π ( l | m ) . To prove Equation (7), we use the definition of the variance:
V a r ( ξ ¯ i ) = E ( ξ ¯ i 2 ) E 2 ( ξ ¯ i ) .
Note some useful properties of Hadamard product d h of vectors d , h R t used to prove Proposition 1:
  • d h = h d for all d , h R t ;
  • ( d h ) g = ( d g ) h for all d , h , g R t ;
  • d ( h + g ) = d h + d g for all d , h , g R t ;
  • ( d L h ) = ( d L h ) for all d , h R t and for any matrix L = [ L i j ] i = 1 , , t j = 1 , , t , L i j R .
Determine an expectation of the squared payoff to Player i in stochastic game G as follows:
E ( ξ ¯ i 2 ) = W = lim k W k ,
where W k is the expectation of the squared payoff of the ith player in stochastic game G under condition that the game ends at stage k. The value of W k can be obtained by the following equation:
W k = j 1 = 1 t j 2 = 1 t j k = 1 t p j 1 0 π ( j 2 | j 1 ) π ( j k | j k 1 ) l = 1 k δ l 1 K i j l 2
or another form of the same equation:
W k = j 1 = 1 t p j 1 0 j 2 = 1 t π ( j 2 | j 1 ) j k = 1 t π ( j k | j k 1 ) l = 1 k δ l 1 K i j l 2 .
We calculate W k recurrently and establish the general form of W k using the method of mathematical induction. It is obvious that
W 1 = p 0 ( K i K i ) .
The following computations are needed to determine W 2 :
W 2 = j 1 = 1 t p j 1 0 j 2 = 1 t π ( j 2 | j 1 ) ( K i j 1 + δ K i j 2 ) 2 = j 1 = 1 t p j 1 0 j 2 = 1 t π ( j 2 | j 1 ) ( K i j 1 ) 2 + + δ 2 j 1 = 1 t p j 1 0 j 2 = 1 t π ( j 2 | j 1 ) ( K i j 2 ) 2 + 2 δ j 1 = 1 t p j 1 0 j 2 = 1 t π ( j 2 | j 1 ) K i j 1 K i j 2 .
We obtain
W 2 = j 1 = 1 t p j 1 0 ( K i j 1 ) 2 j 2 = 1 t π ( j 2 | j 1 ) + δ 2 j 1 = 1 t p j 1 0 Π j 1 ( K i K i ) + 2 δ j 1 = 1 t p j 1 0 K i j 1 j 2 = 1 t π ( j 2 | j 1 ) K i j 2 = = j 1 = 1 t p j 1 0 ( K i j 1 ) 2 + δ 2 p 0 Π ( K i K i ) + 2 δ j 1 = 1 t p j 1 0 K i j 1 Π j 1 K i = = p 0 ( K i K i ) + δ 2 p 0 Π ( K i K i ) + 2 δ ( p 0 K i ) Π K i ,
where Π j is the jth row of matrix Π . We should notice that
W 2 = W 1 + δ { δ p 0 Π ( K i K i ) + 2 ( p 0 K i ) Π K i } .
Making similar calculations, we obtain an expression for the value of W 3 :
W 3 = W 2 + δ 2 { δ 2 p 0 Π 2 ( K i K i ) + 2 ( p 0 K i ) Π 2 K i + 2 δ ( p 0 Π K i ) Π K i } .
Suppose the following formula is true for any k = 1 , 2 , , n 1 :
W k + 1 = W k + δ k { δ k p 0 Π k ( K i K i ) + 2 j = 1 k δ k j ( p 0 Π k j K i ) Π j K i } ,
and we prove it using the method of mathematical induction for k = n . Let Equation ( A 2 ) hold for k = n 1 . The value of W n + 1 is equal to the following one:
W n + 1 = j 1 = 1 t p j 1 0 j 2 = 1 t π ( j 2 | j 1 ) j n + 1 = 1 t π ( j n + 1 | j n ) l = 1 n + 1 δ l 1 K i j l 2 .
Taking into account that
l = 1 n + 1 δ l 1 K i j l 2 = l = 1 n δ l 1 K i j l 2 + ( δ n K i j n + 1 ) 2 + 2 δ n K i j n + 1 l = 1 n δ l 1 K i j l ,
we can write an expression:
W n + 1 = W n + j 1 = 1 t p j 1 0 j 2 = 1 t j n + 1 = 1 t π ( j n + 1 | j n ) ( δ n K i j n + 1 ) 2 + + j 1 = 1 t p j 1 0 j 2 = 1 t j n + 1 = 1 t π ( j n + 1 | j n ) δ 2 n ( K i j n + 1 ) 2 + + 2 δ n j 1 = 1 t p j 1 0 j 2 = 1 t j n + 1 = 1 t π ( j n + 1 | j n ) K i j n + 1 l = 1 n δ l 1 K i j l = = W n + δ n δ n p 0 Π n ( K i K i ) + 2 δ n l = 1 n j 1 = 1 t j n + 1 = 1 t p j 1 0 π ( j 2 | j 1 ) π ( j l | j l 1 ) δ l 1 K i j l × π ( j l + 1 | j l ) π ( j n + 1 | j n ) K i j n + 1
Taking into account Equation (A2) for k = n , which is true by an assumption of mathematical induction, we obtain an expression for W n + 1 :
W n + 1 = W n + δ n δ n p 0 Π n ( K i K i ) + 2 l = 1 n δ n l ( p 0 Π l 1 K i ) Π n l + 1 K i = W n + δ n δ n p 0 Π n ( K i K i ) + 2 l = 1 n δ n l ( p 0 Π k l K i ) Π l K i .
Taking into account Equation (A1) and expression for W k , we obtain
W = lim k ( W k 1 + Ω k ) = lim k ( W k 2 + Ω k 1 + Ω k ) = = lim k ( W 1 + j = 1 k Ω j ) = W 1 + lim k j = 1 k Ω j ,
where
Ω j = δ j { δ j p 0 Π j ( K i K i ) + 2 l = 1 j δ j l ( p 0 Π j l K i ) Π l K i } = p 0 ( δ 2 Π ) j ( K i K i ) + 2 δ j l = 1 j ( p 0 ( δ Π ) j l K i ) Π l K i .
To find the sum j = 1 k Ω j , we first calculate
j = 1 k p 0 ( δ 2 Π ) j ( K i K i ) = p 0 ( Π ˜ I ) ( K i K i ) ,
where Π ˜ = ( I δ 2 Π ) 1 .
Then, we calculate the second sum:
2 j = 1 k δ j l = 1 j ( p 0 ( δ Π ) j l K i ) Π l K i = 2 n = 0 s = 1 δ n ( p 0 ( δ Π ) n K i ) ( δ Π ) s K i = 2 ( p 0 Π ˜ K i ) ( Π ^ I ) K i .
Taking into account the last two expressions, we obtain
W = p 0 ( K i K i ) + p 0 ( Π ˜ I ) ( K i K i ) + 2 ( p 0 Π ˜ K i ) ( Π ^ I ) K i = p 0 Π ˜ ( K i K i ) + 2 ( p 0 Π ˜ K i ) ( Π ^ I ) K i .
Therefore, we obtain the following expression for E ( ξ ¯ i 2 ) :
E ( ξ ¯ i 2 ) = p 0 Π ˜ ( K i K i ) + 2 ( p 0 Π ˜ K i ) ( Π ^ I ) K i .
Since p 0 Π ˜ ( K i K i ) is equal to ( p 0 Π ˜ K i ) K i by the property of Hadamard product, Equation (A3) can be rewritten as follows
E ( ξ ¯ i 2 ) = ( p 0 Π ˜ K i ) ( 2 Π ^ I ) K i .
There is a final formula for the variance of Player i’s payoff:
V a r ( ξ ¯ i ) = ( p 0 Π ˜ K i ) ( 2 Π ^ I ) K i ( p 0 Π ^ K i ) 2 .

Appendix A.2. Proof of Proposition 2

As in Appendix A.1, we omit the symbols of strategy profile in notations. So, we have K i s = K i s ( a s ) , E ( ξ ¯ i ) = E ( ξ ¯ i ( η ) ) , Π = Π ( η ) , Π ^ = Π ^ ( η ) , ξ = ξ ( η ) , π ( l | m ; a m ) = π ( l | m ) .
Equation (8) takes place because the expectation of payoff ξ C is equal to the sum of expectations of the payoffs ξ i of the players from coalition C and it follows from (5).
To prove Equation (9), we take into account that ξ ¯ C = i C ξ ¯ i and use the following formula for computation of the variance of coalition C’s payoff in stochastic game G:
V a r ( ξ ¯ C ) = i C V a r ( ξ ¯ i ) + 2 l < m , l , m C c o v ( ξ ¯ l , ξ ¯ m ) .
The variance of ξ ¯ i can be calculated by Equation (7), so we should obtain the formula to calculate covariance of the lth and mth players’ payoffs. Covariance c o v ( ξ ¯ l , ξ ¯ m ) can be derived as follows:
c o v ( ξ ¯ l , ξ ¯ m ) = E ( ξ ¯ l ξ ¯ m ) E ( ξ ¯ l ) E ( ξ ¯ m ) .
The way of determining covariance c o v ( ξ ¯ l , ξ ¯ m ) is similar to the way of determining E ( ξ ¯ i 2 ) in Appendix A.1. That is why we will not perform all computations in detail.
The expectation of multiplication of Player l and m’s payoffs is
E ( ξ ¯ l ξ ¯ m ) = W = lim k W k ,
where W k is the expectation of payoffs multiplication under condition that the game ends at stage k. The value of W k can be expressed in the following form:
W k = j 1 = 1 t p j 1 0 j 2 = 1 t π ( j 2 | j 1 ) j k = 1 t π ( j k | j k 1 ) r = 1 k δ r 1 K l j r r = 1 k δ r 1 K m j r .
It is obvious that
W 1 = p 0 ( K l K m ) .
Expression for W 2 is
W 2 = W 1 + δ δ p 0 Π ( K l K m ) + ( p 0 K l ) Π K m + ( p 0 K m ) Π K l .
Making similar computation as in the previous proof, we can write the recurrent equation for W k :
W k + 1 = W k + δ k δ k p 0 Π k ( K l K m ) + j = 1 k δ k j ( p 0 Π k j K l ) Π j K m + ( p 0 Π k j K m ) Π j K l .
With respect to (A5), we obtain
E ( ξ ¯ l ξ ¯ m ) = k = 0 p 0 ( δ 2 Π ) k ( K l K m ) + k = 1 δ k j = 1 k ( p 0 ( δ Π ) k j K l ) Π j K m + j = 1 k ( p 0 ( δ Π ) k j K m ) Π j K l .
Simplifying the summands of the last expression of Equation (A6), we can rewrite it as follows
E ( ξ ¯ l ξ ¯ m ) = p 0 Π ˜ ( K l K m ) + ( p 0 Π ˜ K l ) ( Π ^ I ) K m + ( p 0 Π ˜ K m ) ( Π ^ I ) K l .
Since p 0 Π ^ ( K l K m ) is equal to ( p 0 Π ^ K m ) K l and to ( p 0 Π ^ K l ) K m by the property of Hadamard product, Equation (A7) can be rewritten as follows
E ( ξ ¯ l ξ ¯ m ) = ( p 0 Π ˜ K l ) ( Π ^ 0.5 I ) K m + ( p 0 Π ˜ K m ) ( Π ^ 0.5 I ) K l .
Following Equations (7), (A4) and (A8), the covariance of payoffs of Players l and m in stochastic game G can be found by formula:
c o v ( ξ ¯ l , ξ ¯ m ) = ( p 0 Π ˜ K l ) ( Π ^ 0.5 I ) K m + ( p 0 Π ˜ K m ) ( Π ^ 0.5 I ) K l ( p 0 Π ^ K l ) ( p 0 Π ^ K m ) .
So, we prove that formula (9) is true, where c o v ( ξ ¯ l , ξ ¯ m ) can be determined by (10).

References

  1. Shapley, L.S. Stochastic games. Proc. Natl. Acad. Sci. USA 1953, 39, 1095–1100. [Google Scholar] [CrossRef]
  2. Fink, A.M. Equilibrium in a stochastic n-person game. J. Sci. Hiroshima Univ. 1964, 28, 89–93. [Google Scholar] [CrossRef]
  3. Takahashi, M. Equilibrium points of stochastic, noncooperative n-person games. J. Sci. Hiroshima Univ. 1964, 28, 95–99. [Google Scholar] [CrossRef]
  4. Sobel, M.J. Non-cooperative stochastic games. Ann. Math. Statist. 1971, 42, 1930–1935. [Google Scholar] [CrossRef]
  5. Neyman, A.; Sorin, S. Stochastic Games and Applications; Nato Science Series C; Kluwer Academic Publishers: Dordrecht, The Netherlands, 2003; Volume 570. [Google Scholar]
  6. Solan, E. Stochastic games. In Encyclopedia of Database Systems; Springer: Berlin, Germany, 2009. [Google Scholar]
  7. Solan, E.; Vieille, N. Stochastic games. Proc. Natl. Acad. Sci. USA 2015, 112, 13743–13746. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Markowitz, H.M. Portfolio Selection. J. Financ. 1952, 7, 77–91. [Google Scholar]
  9. Csóka, P.; Herings, P.J.-J.; Kóczy, L.A. Stable allocations of risk. Games Econ. Behav. 2009, 67, 266–276. [Google Scholar] [CrossRef] [Green Version]
  10. Suijs, J.; Borm, P. Stochastic cooperative games: Superadditivity, convexity, and certainty equivalents. Games Econ. Behav. 1999, 27, 331–345. [Google Scholar] [CrossRef] [Green Version]
  11. Suijs, J.; Borm, P.; De Waegenaere, A.; Tijs, S. Cooperative Games with Stochastic Payoffs. Eur. J. Oper. Res. 1999, 113, 193–205. [Google Scholar] [CrossRef] [Green Version]
  12. von Neumann, J.; Morgenstern, O. Theory of Games and Economic Behavior; Princeton University Press: Princeton, NJ, USA, 1944. [Google Scholar]
  13. Gromova, E.V.; Petrosyan, L.A. On an approach to constructing a characteristic function in cooperative differential games. Automat. Remote Control 2017, 78, 1680–1692. [Google Scholar] [CrossRef]
  14. Petrosjan, L.; Zaccour, G. Time-consistent Shapley value allocation of pollution cost reduction. J. Econ. Dyn. Control 2003, 27, 381–398. [Google Scholar] [CrossRef]
  15. Reddy, P.V.; Zaccour, G. A friendly computable characteristic function. Math. Soc. Sci. 2016, 82, 18–25. [Google Scholar] [CrossRef]
  16. Petrosjan, L.A. Cooperative Stochastic Games. In Advances in Dynamic Games: Application to Economics, Engineering and Environmental Management; Annals of the International Society of Dynamic Games; Haurie, A., Muto, S., Petrosjan, L.A., Raghavan, T.E.S., Eds.; Birkhäuser: Basel, Switzerland, 2006; pp. 139–146. [Google Scholar]
  17. Petrosjan, L.A.; Baranova, E.M. Cooperative Stochastic Games in Stationary Strategies. Game Theory Appl. 2006, 11, 7–17. [Google Scholar]
  18. Parilina, E.M. Stable cooperation in stochastic games. Automat. Remote Control 2015, 76, 1111–1122. [Google Scholar] [CrossRef]
  19. Parilina, E.M.; Tampieri, A. Stability and cooperative solution in stochastic games. Theory Decision 2018, 80, 601–625. [Google Scholar] [CrossRef] [Green Version]
  20. Parilina, E.; Zaccour, G. Node-consistent core for games played over event trees. Automatica 2015, 53, 304–311. [Google Scholar] [CrossRef]
  21. Parilina, E.; Petrosyan, L. On a Simplified Method of Defining Characteristic Function in Stochastic Games. Mathematics 2020, 8, 1135. [Google Scholar] [CrossRef]
  22. Timmer, J.; Borm, P.; Tijs, S. On three Shapley-like solutions for cooperative games with random payoffs. Int. J. Game Theory 2003, 32, 595–613. [Google Scholar]
  23. Habis, H.; Herings, P.J.-J. Transferable utility games with uncertainty. J. Econ. Theory 2011, 146, 2126–2139. [Google Scholar] [CrossRef] [Green Version]
  24. Wei, Q. Nonzero-sum risk-sensitive finite-horizon continuous-time stochastic games. Stat. Probabil. Lett. 2019, 147, 96–104. [Google Scholar] [CrossRef]
  25. Bäuerle, N.; Rieder, U. Zero-sum risk-sensitive stochastic games. Stoch. Process. Appl. 2017, 127, 622–642. [Google Scholar] [CrossRef] [Green Version]
  26. Wei, Q.; Chen, X. Risk-Sensitive Average Equilibria for Discrete-Time Stochastic Games. Dyn. Games Appl. 2019, 9, 521–549. [Google Scholar] [CrossRef]
  27. Nowak, A.S. Notes on Risk-Sensitive Nash Equilibria. In Advances in Dynamic Games. Annals of the International Society of Dynamic Games; Nowak, A.S., Szajowski, K., Eds.; Birkhäuser: Boston, MA, USA, 2005; Volume 7. [Google Scholar]
  28. Jaśkiewicz, A.; Nowak, A.S. Stationary Markov perfect equilibria in risk sensitive stochastic overlapping generations models. J. Econ. Theory 2014, 151, 411–447. [Google Scholar] [CrossRef]
  29. Asienkiewicz, H.; Balbus, Ł. Existence of Nash equilibria in stochastic games of resource extraction with risk-sensitive players. TOP 2019, 27, 502–518. [Google Scholar] [CrossRef] [Green Version]
  30. Saldi, N.; Basar, T.; Raginsky, M. Approximate markov-nash equilibria for discrete-time risk-sensitive mean-field games. Math. Oper. Res. 2020, 45, 1596–1620. [Google Scholar] [CrossRef]
Figure 1. The region of ( r 1 , d 1 ) defining the core in Example 3.
Figure 1. The region of ( r 1 , d 1 ) defining the core in Example 3.
Mathematics 09 00230 g001
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Parilina, E.; Akimochkin, S. Cooperative Stochastic Games with Mean-Variance Preferences. Mathematics 2021, 9, 230. https://doi.org/10.3390/math9030230

AMA Style

Parilina E, Akimochkin S. Cooperative Stochastic Games with Mean-Variance Preferences. Mathematics. 2021; 9(3):230. https://doi.org/10.3390/math9030230

Chicago/Turabian Style

Parilina, Elena, and Stepan Akimochkin. 2021. "Cooperative Stochastic Games with Mean-Variance Preferences" Mathematics 9, no. 3: 230. https://doi.org/10.3390/math9030230

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop