Next Article in Journal
Do Political Economy Factors Influence Funding Allocations for Disaster Risk Reduction?
Next Article in Special Issue
Credit Risk Model Based on Central Bank Credit Registry Data
Previous Article in Journal
Natural Resources Volatility and Economic Growth: Evidence from the Resource-Rich Region
Previous Article in Special Issue
Market Graph Clustering via QUBO and Digital Annealing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Time-Consistent Investment and Consumption Strategies under a General Discount Function

1
Laboratory of Applied Mathematics, University Mohamed Khider, P.O. Box 145, Biskra 07000, Algeria
2
Departament de Matemàtiques i Informàtica, Universitat de Barcelona, Gran Via 585, 08007 Barcelona, Spain
*
Author to whom correspondence should be addressed.
J. Risk Financial Manag. 2021, 14(2), 86; https://0-doi-org.brum.beds.ac.uk/10.3390/jrfm14020086
Submission received: 11 January 2021 / Revised: 9 February 2021 / Accepted: 11 February 2021 / Published: 20 February 2021
(This article belongs to the Special Issue Financial Optimization and Risk Management)

Abstract

:
In the present paper, we investigate the Merton portfolio management problem in the context of non-exponential discounting, a context that gives rise to time-inconsistency of the decision-maker. We consider equilibrium policies within the class of open-loop controls that are characterized, in our context, by means of a variational method which leads to a stochastic system that consists of a flow of forward-backward stochastic differential equations and an equilibrium condition. An explicit representation of the equilibrium policies is provided for the special cases of power, logarithmic and exponential utility functions.

1. Introduction

In recent years, there has been a renewed attention in time inconsistency for optimality problems, as well as financial models. Generally, time inconsistency arises in several optimality problems when the optimal strategy selected at some time s is no longer optimal at time t > s . In other words, a strategy is time-inconsistent when the decision-maker at future time t > s is tempted to deviate from the strategy determined at time s. It is well known that an optimization problem gives raise to time-inconsistent strategies when the dynamic programming principle cannot be applied, and Bellman’s principle does not hold.
In practice, an important illustration of time-inconsistent problem is the mean-variance selection problem, where the time inconsistency is due to the fact that there is a nonlinear function of the expectation of the final wealth in the objective criterion. Another important problem which produces a time-inconsistent behavior is the investment-consumption problem with non-exponential discounting. This was the case studied by Strotz (1955–1956), where the time inconsistency arises by the fact that the initial point in time enters in a crucial manner the objective criterion.
The common assumption in classical investment-consumption problems under discounted utility is that the discount rate is assumed to be constant over time which leads to the discount function be exponential. This assumption provides the possibility to compare outcomes occurring at different times by discounting future utility by some constant factor. But, on the other hand, results from experimental studies contradict this assumption, indicating that discount rates for the near future are much lower than discount rates for the time further away in future. Ainslie (1995) established experimental studies on human and animal behavior and found that discount functions are almost hyperbolic; that is, they decrease like a negative power of time rather than an exponential. Loewenstein and Prelec (1992) showed that economic decision-makers are impatient about choices in the short term but are more patient when choosing between long-term alternatives; therefore, a hyperbolic type discount function would be more realistic.
Unfortunately, as soon as a discount function is non-exponential, discounted utility models become time-inconsistent in the sense that they do not admit the Bellman’s optimality principle. Consequently, the classical dynamic programming approach may not be applicable to solve these problems. According to Strotz (1955–1956), there are two basic ways of handling time inconsistency in non-exponential discounted utility models. In the first one, under the notion of naive agents, every decision is taken without taking into account that their preferences will change in the near future. The agent at time t 0 , T will solve the problem as a standard optimal control problem with initial condition X ( t ) = x t . If we suppose that the naive agent at time 0 solves the problem, his or her solution corresponds to the so-called pre-commitment solution, in the sense that it is optimal as long as the agent can pre-commit his or her future behavior at time t = 0 .
Kydland and Prescott (1997) indeed argue that a pre-committed strategy may be economically meaningful in certain circumstances. The second approach consists in the formulation of a time-inconsistent decision problem as a non-cooperative game between incarnations of the decision-maker at different instants of time. Nash equilibrium of these strategies are then considered to define the new concept of solution to the original problem. Strotz (1955–1956) was the first who proposed a game theoretic formulation to handle the dynamic time inconsistent optimal decision problem on the deterministic Ramsey problem; see Ramsey (1928). Then, by capturing the idea of non-commitment, and letting the commitment period being infinitesimally small, he provided a primitive notion of Nash equilibrium strategy.

1.1. Related Works

Further work along this line in continuous and discrete time has been done by Pollak (1968); Phelps and Pollak (1968); Goldman (1980); Barro (1990); and Krusell and Smith (2003). Keeping the same game theoretic approach, Ekeland and Lazrak (2008) and Marín-Solano and Navas (2010) treated the optimal consumption problem where the utility involves a non-exponential discount function in the deterministic framework. They characterized the equilibrium strategies by a value function that has to satisfy a certain “extended Hamilton–Jacobi–Bellman (HJB) equation”, which is a non-linear differential equation displaying a non-local term that depends on the global behavior of the solution. In this situation, every decision at time t is taken by a t agent which represents the incarnation of the controller at time t and is referred in Marín-Solano and Navas (2010) as a “sophisticated t agent”.
Björk and Murgoci (2010) extended the idea to the stochastic setting where the controlled dynamic is driven by a quite general class of Markov process and a fairly general objective function. Yong in Yong (2011), by a discretization of time, studied a class of time inconsistent deterministic linear quadratic models and derive equilibrium controls via some class of Riccati-Voltera equations. Yong (2012), also by a discretization of time, investigated a general discounting time inconsistent stochastic optimal control problem and characterizes a feedback time-consistent Nash equilibrium control via the so-called “equilibrium HJB equation”.
The Nash equilibrium solution to the mean-variance problem was established first by Basak and Chabakauri (2010) and then extended to a further general class of time-inconsistent problems by Björk and Murgoci (2010). Other papers on the consistent planning approach for the mean-variance problem are Hu et al. (2012); Czichowsky (2013); and Björk et al. (2014).
Concerning equilibrium strategies for an optimal consumption-investment problem with a general discount function, Ekeland and Pirvu (2008) were the first to investigate Nash equilibrium strategies where the price process of the risky asset is driven by geometric Brownian motion. They characterized the equilibrium strategies through the solutions of a flow of BSDEs, and they show, for an special form of the discount function, that the BSDEs reduce to a system of two ODEs, which has a solution. Ekeland et al. (1995) added life insurance to the investor’s portfolio and they characterize the equilibrium strategy by an integral equation. In Yong (2012), the case of time-inconsistent consumption-investment problem under a power utility function is discussed. Following Yong’s approach, Zhao et al. (2014) studied the consumption-investment problem with a general discount function and a logarithmic utility function. Recently, Zou et al. (2014) investigated equilibrium consumption-investment decisions for Merton’s portfolio problem with stochastic hyperbolic discounting.
As for the comparison between different methods to time inconsistency, Wang and Forsyth (2012) evaluate time-consistent against pre-commitment strategies and compare their related efficient frontiers for a mean-variance optimization problem. The evaluation among the naive and the sophisticated approaches is given by Chen et al. (2014), who study the optimal dividend model of an insurance company in the existence of time inconsistency created by non-exponential discount factor. Cong and Oosterlee (2016) found a relation linking the time-consistent and the pre-commitment investment strategies in a defined contribution pension scheme. Cui et al. (2017) emphasize the shortcomings of pre-commitment and game theoretical strategies, and examine a self-coordination strategy that aims at corresponding global concern and local interests of the decision-maker. Van Staden et al. (2018) consider the pre-commitment and the time-consistent policies in the attendance of realistic investment constraints. Bensoussan et al. (2019) evaluate the produce of constraints on the value function of both pre-commitment and game theoretical approaches, and discover the unexpected result that for the game theoretical approach the occurrence of constraints can improve the payoff, whereas, for the pre-commitment approach, this paradox does not arise. Menoncin and Vigna (2020) evaluate the contribution pension scheme and prove that the dynamically optimal policy reacts better to extreme scenarios of market returns.
Time-inconsistent consumption-investment problem with a non-exponential discount function and a general utility function. We use the game theoretic approach to handle the time inconsistency in the same perspective as Björk and Murgoci (2010). Noting that, the game perspective that we will consider is as follows: We first consider a game with one player at each point t in time. This player represents the incarnation of the decision maker at time t and can be referred to as “player t”. This t t h player can control the system only at time t by taking his/her strategies. A control process is then viewed as a complete description of the chosen strategies of all players in the game. The reward to player t is given by utility function of an investment-consumption optimization problem. From this description, we introduce the concept of a “perfect Nash equilibrium strategy” of the game. This is an admissible control process satisfying some admissibility conditions.
We focus on a variational technique approach leading to a version of a necessary and sufficient condition for equilibrium, which involves a flow of forward-backward stochastic differential equations (FBSDEs) along with a certain equilibrium condition. We also present a verification theorem that covers some possible examples of utility functions. Then, by decoupling the flow of the FBSDEs, we derive a closed-loop representation of the equilibrium strategies via a parabolic non-linear partial differential equation (PDE). We show that within a special form of the utility function (logarithmic, power, and exponential) the PDE reduces to a system of ODEs which has an explicit solution.

1.2. Novelty and Contribution

Different from Marín-Solano and Navas (2010) and Ekeland and Pirvu (2008), where the authors derived explicit solutions for special forms of the discount factor, in our model, the non-exponential discount function is in a fairly general form. Moreover, we consider equilibrium strategies in the open-loop sense, as defined in Hu et al. (2012) and Hu et al. (2015), which is different from most of the existing literature on this topic. Note also that the time-inconsistency, in our paper, arises from a non-exponential discounting in the objective function, while the works Hu et al. (2012) and Hu et al. (2015) are concerned with a quite different kind of time-inconsistency which is caused by the presence of non-linear terms of expectations in the terminal cost. On other hand, the objective functional, in our paper, is not reduced to the quadratic form as in Hu et al. (2012) and Hu et al. (2015).
We accentuate that, different from most of the existing literature on this topic, where some feedback equilibrium strategies are derived via several very complicated highly non-linear integro-differential equations, an explicit representation of the equilibrium strategies are obtained in our work via simple ODEs. In addition, this method can provide the necessary and sufficient conditions to characterize the equilibrium strategies, while the extended HJB techniques can create, in general, only the sufficient condition in the form of a verification theorem that characterizes the equilibrium strategies.

1.3. Structure of the Paper

The rest of the paper is organized as follows. In Section 2, we formulate the problem and give the necessary notations and preliminaries. In Section 3, we present the main results of the paper, Theorem 1 and Theorem 2, that characterize the equilibrium decisions by some necessary and sufficient conditions. In Section 4, we derive an explicit representation of the equilibrium consumption-investment strategy. Section 5 is devoted to some comparisons with existing results in the literature. The paper ends with an Appendix containing some proofs.

2. Problem Formulation

In what follows, we assume that W · = W 1 · , , W d · is a d-dimensional standard Brownian motion defined on a filtered probability space ( Ω , F , F , P ) , such that F : = F t t 0 , T is a natural filtration that satisfies the usual conditions, in particular, F 0 contains all P -null sets and F T = F for an arbitrarily fixed finite time horizon T > 0 . Recall that F t stands for the information available up to time t, and any decision made at time t is based on this information.

2.1. Notations

Throughout this paper, we use the following notations: M : the transpose of the vector (or matrix) M, χ , ζ : the inner product of χ and ζ , that is, χ , ζ : = t r ( χ T ζ ) . For a function f , we denote by f x (resp. f x x ) the first (resp. the second) derivative of f with respect to the variable x .
For any Euclidean space E with Frobenius norm · , we let, for any t 0 , T ,
  • L p Ω , F t , P ; E : for any p 1 , the set of E valued F t measurable random variables X , such that E X p < .
  • L F 2 t , T ; E : the space of E valued, F s s t , T adapted continuous processes Y · , with
    Y · L F 2 t , T ; E = E sup s t , T Y s 2 < .
  • M F p t , T ; E : for any p 1 , the space of E valued, F s s t , T adapted processes Z · , with
    Z · M F p t , T ; E = E 0 T Z s p d s 1 p < .

2.2. Financial Market

Consider an individual facing the inter-temporal consumption and portfolio problem where the market environment consists of one riskless and d risky securities. The risky securities are stocks and their prices are modeled as Itô processes. Namely, for i = 1 , 2 , . . , d , the price S i s , for s 0 , T , of the i-th risky asset, satisfies
d S i s = S i s μ i s d s + j = 1 d σ i j s d W j s ,
with S i 0 > 0 , for i = 1 , 2 , . . . , d , and the coefficients μ i · and σ i · = σ i 1 · , , σ i d · , for i = 1 , . . , d , are F progressively measurable processes with values in R and R d , respectively. For brevity, we use μ · = μ 1 · , μ 2 · , , μ d · to denote the drift rate vector and σ · = σ i j · 1 i , j d to denote the random volatility matrix.
The riskless asset, or the savings account, has the price process S 0 s , for s 0 , T , governed by
d S 0 s = r 0 s S 0 s d s , S 0 0 = 1 ,
where r 0 · is a process with values in 0 , that represents the interest rate. We assume that E μ i t > r 0 t 0 , d t a . e . , for i = 1 , 2 , . . , d . This is a very natural assumption since, otherwise, nobody is willing to invest in the risky stocks.

2.3. Investment-Consumption Policies and Wealth Process

Starting from an initial capital x 0 > 0 at time 0, during the time horizon 0 , T , the decision-maker is allowed to dynamically invest in the stocks, as well as in the bond and consuming. A consumption-investment strategy is described by a d + 1 -dimensional stochastic process u · = c · , u 1 · , , u d · , where c s represents the consumption rate at time s 0 , T and u i s , for i = 1 , 2 , . . , d , represents the amount invested in the i- t h risky stock at time s 0 , T . The process u I · = u 1 · , , u d · is called an investment strategy. The amount invested in the bond at time s is
X x 0 , u s i = 1 d u i s ,
where X x 0 , u · is the wealth process associated with the strategy u · and the initial capital x 0 . The evolution of X x 0 , u · can be described as
d X x 0 , u s = X x 0 , u s i = 1 d u i s d S 0 s S 0 s + i = 1 d u i s d S i s S i s c s d s , for s 0 , T , X x 0 , u 0 = x 0 .
Accordingly, the wealth process solves the SDE:
d X x 0 , u s = r 0 s X x 0 , u s + u I s r s c s d s + u I s σ s d W s , for s 0 , T , X x 0 , u 0 = x 0 .
where r · = μ 1 · r 0 · , , μ d · r 0 · .
As time evolves, it is natural to consider the controlled stochastic differential equation parametrized by t , ξ 0 , T × L 2 Ω , F t , P ; R and satisfied by X · = X t , ξ · ; u · ,
d X s = r 0 s X s + u I s r s c s d s + u I s σ s d W s , for s t , T , X t = ξ .
Definition 1
(Admissible Strategy). A strategy u · = c · , u I · is said to be admissible over t , T if u · M F 1 t , T ; R × M F 2 t , T ; R d and for any t , ξ 0 , T × L 2 Ω , F t , P ; R , the equation 4 has a unique solution X · = X t , ξ · ; u · .
We impose the following assumption about the coefficients.
(H1) 
Processes r 0 · , r · and σ · are uniformly bounded and moreover we assume the following uniform ellipticity condition:
σ s σ s ϵ I d , d s a . e , d P a . s .
for some ϵ > 0 , where I d denotes the identity matrix on R d × d .
Under (H1), for any t , ξ , u · 0 , T × L 2 Ω , F t , P ; R × M F 1 t , T ; R × M F 2 t , T ; R d , the state equation 4 has a unique solution X · L F 2 t , T ; R . Moreover, we have the estimate
E sup t s T X s 2 K 1 + E ξ 2 ,
for some positive constant K. In particular, for t = 0 , x 0 > 0 and u · = c · , u I · M F 1 0 , T ; R × M F 2 0 , T ; R d , the state equation 3 has a unique solution X x 0 , u · L F 2 0 , T ; R , and the following estimate holds:
E sup 0 s T X x 0 , u s 2 K 1 + x 0 2 .

2.4. General Discounted Utility Function

Most of financial-economics works have considered that the rate of time preference is constant (exponential discounting). However, there is growing evidence to suggest that this may not be the case. In this section, we discuss the general discounting preferences. We also introduce the basic modeling framework of Merton’s consumption and portfolio problem. We refer the reader to Ainslie et al. (1991); Karatzas et al. (1987); Merton (1969, 1971), and Pliska (1986) for more detail about the classical Merton model.

2.4.1. Discount Function

As soon as discounting is non-exponential, most papers work with special form of the non-exponential discount factor. Differently from these works, we consider a general form of the discount factor.
Definition 2.
A discount function λ · : 0 , T R is a continuous and deterministic function satisfying λ 0 = 1 , λ s > 0 d s a . e . and 0 T λ s d s < .
Remark 1.
Some examples of discount functions are given in many papers, such as exponential discount functions (see Merton (1969) and Merton (1971)), mixture of exponential functions (see Ekeland and Pirvu (2008)), and hyperbolic discount functions (see Zhao et al. (2014)).

2.4.2. Utility Functions and Objective

In order to evaluate the performance of a consumption-investment strategy, the decision-maker derives utility from inter-temporal consumption and final wealth. Let φ · be the utility of inter-temporal consumption and h · the utility of the terminal wealth at some non-random horizon T (which is a primitive of the model). Then, for any t , ξ 0 , T × L 2 Ω , F t , P ; R , the investment-consumption optimization problem is reduced to maximize the utility function J t , ξ , . given by
J t , ξ , u · = E t t T λ s t φ c s d s + λ T t h X T ,
over u · M F 1 t , T ; R × M F 2 t , T ; R d , subject to 4 , where E t · = E · F t . We restrict ourselves to utility functions which satisfy the following conditions.
(H2) 
The maps φ · , h · : R R are strictly increasing, strictly concave and satisfy the integrability condition
E 0 T φ c s d s + h X T < .
Noting that, most literature on the necessary and/or sufficient optimality conditions in stochastic control problem that considers more strong conditions about the coefficients, in which the derivatives φ x x · and h x x · are bounded or have linear or quadratic growth; see, e.g., Yong and Zhou (1999). It is also worth mentioning that , several papers impose an L p bounds on the control process for p > 2 . Those restrictions make it impossible to apply the stochastic maximum principle approach directly to study the consumption–investment problem. Certainly, in that important problem, the optimal control is not necessarily L p -integrable, and the derivatives of the running cost and the terminal cost do not necessarily follow the global polynomial growth conditions.
In this paper, we overcome the technical difficulties mentioned above and treat some limiting procedures. To do so, let us introduce further technical integrability conditions of the utilities, which will be used in the proof of the main result.
(H3) 
The maps φ · , h · are twice continuously differentiable functions, so all the derivatives φ x · , h x · , φ x x · and h x x · are continuous.
(H4) 
For all admissible strategy pairs, there exists a constant p > 1 such that
E 0 T φ x c s p d s + h x X T p < , E 0 T sup η R , η M φ x x c s + η p d s < , for M 0 .
If we write W s = 0 , W s , and we denote B s = 1 , r s , Γ = 1 , 0 R d and
D s = 0 0 R d 0 R d σ s ,
then the optimal control problem associated with 4 and 7 is equivalent to maximize
J t , ξ , u · = E t t T λ s t φ Γ u · d s + λ T t h X T ,
subject to
d X s = r 0 s X s + u s B s d s + u s D s d W s , for s t , T , X t = ξ .
over u · M F 1 t , T ; R × M F 2 t , T ; R d .

2.4.3. Time Inconsistency

Let us first note that the optimal policies, although they exist, will not be time-consistent in general. First of all, as an illustration, let us consider the model in (8)–(9) with logarithmic utility functions. We suppose that the financial market consists of one riskless asset and d risky assets. Arguing as in Ekeland and Pirvu (2008), we can prove that, if the agent is naive and starts with a given positive wealth x, at some instant t , then, by the standard dynamic programming approach, the value function associated with this stochastic control problem solves the following Hamilton–Jacobi–Bellman equation.
V s t s , x + sup c , u I R d + 1 r 0 s X s + u I r s c V x t s , x + 1 2 u I σ s σ s u I V x x t s , x + λ s t λ s t V t s , x + φ c = 0 , for s t , T , V t T , x = h x .
The HJB equation contains the term λ s t λ s t , which depends not only on the current time s but also on initial time t, so the optimal policy will depend on t, as well. Indeed, the first order necessary conditions yield the t optimal policy
u ¯ I t s = r s σ s σ s 1 V x t s , x V x x t s , x , c ¯ t s = φ 1 V x t s , x .
Let us consider the following example: φ x = h x = log x . The naive agent for the initial pair 0 , x 0 solves the problem, assuming that the discount rate of time preference will be λ s , for s 0 , T , and the optimal consumption strategy will be
c ¯ 0 , x 0 s = 1 + s T exp λ r s + log λ r λ s d r 1 , for s 0 , T .
This solution corresponds to the so-called pre-commitment solution, in the sense that it is optimal as long as the agent can pre-commit (by signing a contract, for example) his or her future behavior at time t = 0 . If there is no commitment, the 0-agent will take the action c ¯ 0 , x 0 s , but, in the near future, the ϵ -agent will change his decision rule (time-inconsistency) to the solution of the HJB equation 10 with t = ϵ . In this case, the optimal control trajectory for s > ϵ will be changed to c ¯ ϵ , x ϵ s , given by
c ¯ ϵ , x ϵ s = c ¯ ϵ , X ¯ ϵ s = 1 + s T exp λ r s + log λ r ϵ λ s ϵ d r 1 , for s ϵ , T .
If λ t = e δ t , where δ > 0 is the constant discount rate, then
c ¯ | ϵ , T 0 , x 0 s = c ¯ ϵ , x ϵ s , for s ϵ , T ;
hence, the optimal consumption plan is time consistent. As soon as discount function is non-exponential
c ¯ | ϵ , T 0 , x 0 s c ¯ ϵ , x ϵ s , for s ϵ , T .
Then, the optimal consumption plan is not time consistent. In general, the solution for the naive agent will be constructed by solving the family of HJB equations 10 for t 0 , T , and patching together the “optimal” solutions c ¯ t , x t t . If the agent is sophisticated, things become more complicated. The standard HJB equation cannot be used to construct the solution, and a new method is required in what follows.

3. Equilibrium Strategies

It is well known that the problem described above by (8)–(9) turns out to be time inconsistent in the sense that it does not satisfy the Bellman optimality principle, since a restriction of an optimal control for a specific initial pair on a later time interval might not be optimal for that corresponding initial pair. For a more detailed discussion see Ekeland and Pirvu (2008) and Yong (2012). Due to the lack of time consistency, we consider open-loop Nash equilibrium controls instead of optimal controls. As in Hu et al. (2012), we first consider an equilibrium by local spike variation, given, for t 0 , T , an admissible consumption-investment strategy u ^ · M F 1 t , T ; R × M F 2 t , T ; R d . For any R d + 1 valued, F t measurable and bounded random variable v and for any ε > 0 , define
u ε s : = u ^ s + v , for s t , t + ε , u ^ s , for s t + ε , T .
We have the following definition.
Definition 3
(Open-loop Nash equilibrium). An admissible strategy u ^ · M F 1 t , T ; R × M F 2 t , T ; R d is an open-loop Nash equilibrium strategy if
lim ε 0 inf 1 ε J t , X ^ t , u ε · J t , X ^ t , u ^ · 0 ,
for any t 0 , T , where X ^ is the equilibrium wealth process that solves the SDE
d X ^ s = r 0 s X ^ s + u ^ s B s d s + u ^ s D s d W s , for s t , T , X ^ t = ξ .

3.1. A Necessary and Sufficient Condition for Equilibrium Controls

In this paper, we follow an alternative approach, which is essentially a necessary and sufficient condition for equilibrium. In the same spirit of proving the stochastic Pontryagin’s maximum principle for equilibrium in Hu et al. (2012) for the case of linear quadratic models, we derive this condition by a second-order expansion in the spike variation.
Now, we introduce the adjoint equations involved in the characterization of open-loop Nash equilibrium controls.

3.1.1. Adjoint Processes

Let u ^ · = c ^ · , u ^ I · M F 1 0 , T ; R × M F 2 0 , T ; R d an admissible strategy and denote by X ^ · L F 2 0 , T ; R the corresponding wealth process. For each t 0 , T , we introduce the first order adjoint equation defined on the time interval t , T , and satisfied by the pair of processes p · ; t , q · ; t as follows
d p s ; t = r 0 s p s ; t d s + q s ; t d W s , for s t , T , p T ; t = λ T t h x X ^ T ,
where q · ; t = q 1 · ; t , , q d · ; t . Under the assumption (H1), the BSDE 14 is uniquely solvable in L F 2 t , T ; R × M F 2 t , T ; R d . Moreover, there exists a constant K > 0 such that, for any t 0 , T , we have the following estimate
p · ; t L F 2 t , T ; R 2 + q · ; t M 2 t , T ; R d 2 K 1 + ξ 2 .
The second order adjoint equation is defined on the time interval t , T and satisfied by the pair of processes P · ; t , Q · ; t L F 2 t , T ; R × M F 2 t , T ; R d as follows:
d P s ; t = 2 r 0 s P s ; t d s + Q s ; t d W s , for s t , T , P T ; t = λ T t h x x X ^ T ,
where Q · ; t = Q 1 · ; t , , Q d · ; t . Under (H1), the above BSDE has a unique solution P · ; t , Q · ; t L F 2 t , T ; R × M F 2 t , T ; R d . Moreover, we have the following representation for P · ; t :
P s ; t = E s λ T t e s T 2 r 0 τ d τ h x x X ^ T , for s t , T .
Indeed, if we define the function Θ · , t , for each t 0 , T , as the fundamental solution of the linear ODE
d Θ τ , t = r 0 τ Θ τ , t d τ , for τ t , T , Θ t , t = 1 ,
and we apply the Itô’s formula to τ P τ ; t Θ τ , t 2 on t , T , by taking conditional expectations, we obtain 17 . Note that, since h x x X ^ T 0 , then P s ; t 0 , d s a . e .

3.1.2. A Characterization of Equilibrium Strategies

The following theorem is the first main result of this work, and it provides a necessary and sufficient condition for equilibrium. As we have said before, the proof is inspired by Hu et al. (2012) and Hu et al. (2015).
First, we define the process q ˜ s ; t = 0 , q s ; t , and we introduce the following notations:
H s ; t p s ; t B s + D s q ˜ s ; t + λ s t φ x Γ u ^ s Γ ,
and, for a certain θ [ 0 , 1 ] ,
A ε s ; t λ s t φ x x Γ u ^ s + θ v 1 t , t + ε ( s ) 0 R d 0 R d σ s σ s P s ; t .
Theorem 1.
Let (H1)–(H4) hold. Given an admissible strategy u ^ · M F 1 0 , T ; R × M F 2 0 , T ; R d , let for any t 0 , T , the process
p · ; t , q · ; t L F 2 t , T ; R × M F 2 t , T ; R d
be the unique solution to the BSDE 14 . Then, u ^ · is an equilibrium consumption-investment strategy, if and only if, the following condition holds
H t ; t = 0 , d P a . s . , d t a . e .
In order to derive the proof of this theorem, let us, first of all, derive some technical results. First, denote by X ^ ε · the solution of the state equation corresponding to u ε · . Since the coefficients of the controlled state equation are linear, using the standard perturbation approach (see, e.g., Yong and Zhou (1999)), we have
X ^ ε s X ^ s = y ε , v s + z ε , v s , for s t , T ,
where, for any R d + 1 valued, F t measurable, and bounded random variable v, and for any ε 0 , T t , y ε , v · and z ε , v · , solve, respectively, the following linear stochastic differential equations:
d y ε , v s = r 0 s y ε , v s d s + v D s 1 t , t + ε s d W s , for s t , T , y ε , v t = 0 ,
and
d z ε , v s = r 0 s z ε , v s + v B s 1 t , t + ε s d s , for s t , T , z ε , v t = 0 .
Proposition 1.
Let (H1)–(H4) hold. For any t 0 , T , the following estimates hold for any k 1 :
E t sup s t , T y ε , v s 2 k = O ε k ,
E t sup s t , T z ε , v s 2 k = O ε 2 k ,
E t sup s t , T y ε , v s + z ε , v s 2 k = O ε k .
In addition, we have the following equality:
J t , X ^ t , u ε · J t , X ^ t , u ^ · = t t + ε E t H s ; t , v + 1 2 A ε s ; t v , v d s + o ε .
Proof. 
See Appendix A. □
Now, we present the following technical lemma needed later. The proof follows an argument adapted from Hamaguchi (2019).
Lemma 1.
Under assumptions (H1)–(H4), there exists a sequence ε n t n N ( 0 , T t ) satisfying ε n t 0 as n , such that
(1) 
lim n 1 ε n t t t + ε n t E t H s ; t d s = H t ; t , d P a . s , d t a . e .
(2) 
lim n 1 ε n t t t + ε n t E t A ε n t s ; t d s = A 0 t ; t , d P a . s , d t a . e .
Proof. 
See Appendix A. □
Proof 
(Proof of Theorem 1.). Given an admissible strategy
u ^ · M F 1 0 , T ; R × M F 2 0 , T ; R d ,
for which 21 holds, according to Lemma 1, we have from 28 that, for any t 0 , T , and for any R d + 1 valued, F t measurable, and bounded random variable v , there exists a sequence ε n t n N ( 0 , T t ) satisfying ε n t 0 as n , such that
lim n 0 1 ε n t J t , X ^ t , u ε · J t , X ^ t , u ^ · = H t ; t , v + 1 2 A 0 t ; t v , v , = 1 2 A 0 t ; t v , v , 0 ,
where we have used in the last inequality the fact that, under the concavity condition of φ · and h · ,it follows A 0 t ; t v , v 0 . Hence, u ^ · is an equilibrium strategy.
Conversely, assume that u ^ · is an equilibrium strategy. Then, by 12 , together with 28 and Lemma 1, for any t , u 0 , T × R d + 1 , the following inequality holds:
H t ; t , u + 1 2 A 0 t ; t u , u 0 .
Now, we define t , u 0 , T × R d + 1 ,
Φ t , u = H t ; t , u + 1 2 A 0 t ; t u , u .
Clearly, Φ · , · is well defined. In fact, it is a second order polynomial in terms of the components of vector u . Easy manipulations show that the inequality 29 is equivalent to
Φ t , 0 = max u R d + 1 Φ t , u , d P a . s , t 0 , T .
So, it is easy to see that the maximum condition 30 leads to the following condition: t 0 , T ,
Φ u t , 0 = H t ; t = 0 , d P a . s .
According to Lemma 1, the expression 21 follows immediately. This completes the proof. □

3.2. A Characterization of Equilibrium Strategies by Verification Argument

In classical (time-consistent) stochastic control theory, the sufficient condition of optimality is of significant importance for computing optimal controls. It says that, if an admissible control satisfies the maximum condition of the Hamiltonian function, then the control is indeed optimal for the stochastic control problem. This allows one to solve examples of optimal control problems, where one can find a smooth solution to the associated adjoint equation.
The aim of the following theorem is to characterize the open-loop equilibrium pair only by a sufficient condition of equilibrium. Let us introduce an alternative to (H3) hypothesis:
(H3′) 
The maps φ · , h · are continuously differentiable, and the first order derivatives φ x · , h x · are continuous.
Then, we have the following theorem:
Theorem 2.
Let (H1), (H2) and (H3’) hold. Given an admissible strategy u ^ · M F 1 0 , T ; R × M F 2 0 , T ; R d , let, for any t 0 , T , the process
p · ; t , q · ; t L F 2 t , T ; R × M F 2 t , T ; R d
be the unique solution to the BSDE 14 . Then, u ^ · is an equilibrium consumption-investment strategy, if the following condition holds:
H t ; t = 0 , d P a . s . , d t a . e .
Proof. 
Suppose that u ^ ( · ) is an admissible control for which the condition (32) holds. In addition, for any t [ 0 , T ] and ε [ 0 , T t ) , we consider u ε · by 11 . Then, we have the following difference.
J t , X ^ t , u ^ · J t , X ^ t , u ε · = E t t T λ s t φ Γ u ^ s φ Γ u ε s d s + λ T t h X ^ T h X ^ ε T .
Noting that, by the concavity of h · , we have
E t λ T t h X ^ T h X ^ ε T E t λ T t X ^ T X ^ ε T T h x X ^ T .
Accordingly, by the terminal condition in the BSDE (14), we obtain that
J t , X ^ t , u ^ · J t , X ^ t , u ε · E t t T λ s t φ Γ u ^ s φ Γ u ε s d s + X ^ T X ^ ε T T p T ; t .
By applying Ito’s formula to s X ^ s X ^ ε s T p s ; t on t , T , we get
E t X ^ T X ^ ε T T p T ; t = E t t T u ^ s u ε s T B s p s ; t + D s q ˜ s ; t d s .
By the concavity of φ · , we find
E t t T λ s t φ Γ u ^ s φ Γ u ε s d s E t t T λ s t φ x Γ u ^ s Γ , u ^ s u ε s d s .
By taking 34 and 35 in 33 , it follows that
J t , X ^ t , u ε · J t , X ^ t , u ^ · E t t T B s p s ; t + D s q ˜ s ; t + λ s t φ x Γ u ^ s Γ , u ε s u ^ s d s = E t t t + ε H s ; t , v d s .
Now, dividing both sides by ε and taking the limit when ε vanishes, by Lemma 1, we conclude that u ^ ( ·) is an equilibrium control. □
Remark 2.
The purpose of the sufficient condition of optimality is to find an optimal control by computing the difference J u ^ · J u · in terms of the Hamiltonian function, where u · is an arbitrary admissible control. Here, the spike variation perturbation 11 plays a key role in deriving the sufficient condition for equilibrium strategies, which reduces to the computation of the difference J t , X ^ t , u ^ · J t , X ^ t , u ε · , without the necessity to achieving the second order expansion in the spike variation.

4. Equilibrium When the Coefficients Are Deterministic

Theorems 1 and 2 show that one can obtain equilibrium consumption-investment strategies by solving a system of FBSDEs which is not standard since the “flow” of the unknown process p · ; t , q · ; t t 0 , T is involved. Moreover, there is an additional constraint that act on the “diagonal” (i.e., when s = t ) of the flow. As far as we know, the explicit solvability of this type of equations remains an open problem, except for some particular form of the utility function. However, we are able to solve quite thoroughly this problem when the parameters r 0 · , μ · and σ · are deterministic functions. In this section, we define what we mean by an equilibrium rule, and then we derive a parabolic backward PDE. Our PDE is comparable with the one obtained in Marín-Solano and Navas (2010) and Ekeland and Pirvu (2008), for some particular discount functions in a finite horizon with different utility functions.
In this section, let us look at the Merton’s portfolio problem with general discounting and deterministic parameters. At first, we consider the following parabolic backward partial differential equation:
θ t t , x + θ x t , x r 0 t x r t Σ t r t θ t , x θ x t , x I λ T t θ t , x + 1 2 θ x x t , x r t Σ t r t θ t , x θ x t , x 2 + θ t , x r 0 t = 0 , t , x 0 , T × R , θ T , x = h x x ,
where we denote by I · the inverse function of the strictly decreasing marginal derivative utility φ x · and Σ s σ s σ s 1 .
We have the following verification theorem.
Theorem 3.
Let (H1)–(H4) hold. If there exists a classical solution
θ · , · C 1 , 2 ( 0 , T ) × R , R C 0 , T × R , R
of the PDE 36 such that the stochastic differential equation,
d X ^ s = r 0 s X ^ s r s Σ s r s θ s , X ^ s θ x s , X ^ s I λ T s θ s , X ^ s d s θ s , X ^ s θ x s , X ^ s r s Σ s σ s d W s , s 0 , T , X ^ 0 = x 0 ,
has a unique solution X ^ · , in which the following estimate holds
E sup 0 t T X t 2 K 1 + x 0 2 ,
then, the equilibrium consumption-investment strategy u ^ · = c ^ · , u ^ I · is given by
c ^ t = I λ T t θ t , X ^ t , d t a . e . ,
u ^ I t = Σ t r t θ t , X ^ t θ x t , X ^ t , d t a . e .
Proof. 
Suppose that u ^ · = c ^ · , u ^ I · is an equilibrium control and denote by X ^ · the corresponding wealth process. Then, in view of Theorem 1, there exists an adapted process, X ^ · , p · ; t , q · ; t t 0 , T , solution of the following flow of forward-backward SDEs, parametrized by t 0 , T :
d X s = r 0 s X ^ s + u ^ I s r s c ^ s d s + u ^ I s σ s d W s , s t , T , d p s ; t = r 0 s p s ; t d s + q s , t d W s , 0 t s T , X ^ 0 = x 0 , p T ; t = λ T t h x X ^ T , t 0 , T ,
with conditions
p t ; t + φ x c ^ t = 0 , d t a . e . ,
p t ; t r t + σ t q t ; t = 0 , d t a . e .
From the terminal condition in the first order adjoint process, we consider the following Ansatz
p s ; t = λ T t V s , X ^ s , 0 t s T ,
for some deterministic function V · , · C 1 , 2 0 , T × R , R such that V T , · = h x · .
Applying Itô’s formula to 43 , it yields
d p s ; t = λ T t V s s , X ^ s + V x s , X ^ s X ^ s r 0 s + u ^ I s r s c ^ s + 1 2 V x x s , X ^ s u ^ I s σ s σ s u ^ I s d s + λ T t V x s , X ^ s u ^ I s σ s d W s .
Next, comparing the d s term in 44 by the ones in the second equation in 40 , we deduce that
V s s , X ^ s + V x s , X ^ s X ^ s r 0 s + u ^ I s r s c ^ s + 1 2 V x x s , X ^ s u ^ I s σ s σ s u ^ I s = r 0 s V s , X ^ s ,
and by comparing the d W s terms we also get
q s , t = λ T t V x s , X ^ s σ s u ^ I s .
We put the above expressions of p s ; t and q s ; t at s = t into 41 and 42 , and then
λ T t V t , X ^ t φ x c ^ t = 0 ,
and
V x t , X ^ t σ t σ t u ^ I t = r t V t , X ^ t ,
which leads to the following representation
c ^ t = I λ T t V t , X ^ t , d t a . e . ,
u ^ I t = Σ t r t V t , X ^ t V x t , X ^ t , d t a . e .
Then, by taking expressions 49 and 50 into 45 , this suggests that V · , · coincides with the solution of the PDE 36 , evaluated along the trajectory X ^ · , solution of the state equation. □
Remark 3.
Equation 36 is comparable with the one in Marín-Solano and Navas (2010) and Ekeland and Pirvu (2008), in which the equilibrium is defined within the class of feedback controls.
Remark 4.
Theorem 3 enables us to derive a suitable equilibrium strategy u ^ I t , as well as c ^ t , at each t [ 0 , T ] , and this permits us to derive directly an explicit expression of equilibrium control in the cases of power, logarithmic, and exponential utility functions. While the duality approach used in Hamaguchi (2019) permits to characterize a stochastic equilibrium solution in terms of a complicated FBSDE system of a closed form, it does not provide an explicit representation.

5. Special Utility Functions

Equilibrium investment-consumption strategies for Merton’s portfolio problem with general discounting and deterministic parameters have been studied in Marín-Solano and Navas (2010); Ekeland and Pirvu (2008); and Yong (2012), among others, in different frameworks. In this section, we discuss some special cases in which the function θ · , · may be separated into functions of time and state variables. Then, one needs only to solve a system of ODEs in order to completely determine the equilibrium strategies. We will compare our results with some existing ones in the literature.

5.1. Power Utility Function

To make the problem (8)–(9) explicitly solvable, we consider power utility functions for the running and terminal costs. That is, φ c = c γ γ and h x = a x γ γ , with a > 0 and γ 0 , 1 . In this case, the PDE 36 reduces to
θ t t , x + θ x t , x r 0 t x r t Σ t r t θ t , x θ x t , x λ T t 1 γ θ t , x γ 1 + 1 2 θ x x t , x r t Σ t r t θ t , x θ x t , x 2 + r 0 t θ t , x = 0 , t , x 0 , T × R , θ T , x = a x γ 1 .
From the terminal condition, we consider the following trial solution
θ s , x = a Π s x γ 1 ,
for some deterministic function Π · C 1 0 , T , R with the terminal condition Π T = 1 . Then, by substituting in 36 , we obtain
Π t t + K t + Q t Π t 1 γ 1 Π t = 0 , for t 0 , T , Π T = 1 .
where
K t γ r 0 t + 1 2 γ 1 γ r t Σ t r t ,
and
Q t 1 γ a λ T t 1 γ 1 .
It remains to determine the function Π · . First, by the change of variable
Π t = y t 1 γ , for t 0 , T ,
we find that y · should solve the following ODE
y t t K t γ 1 y t Q t γ 1 = 0 , for t 0 , T , y T = 1 .
A variation of constant formula yields to
y t = 1 t T Q τ γ 1 e τ T K l γ 1 d l d τ exp t T K τ γ 1 d τ , for t 0 , T ,
and, subsequently, we obtain
Π t = 1 t T Q τ γ 1 e τ T K l γ 1 d l d τ 1 γ exp t T K τ d τ , for t 0 , T .
In view of Theorem 3, the representation of the Nash equilibrium strategies 38 39 gives
c ^ t = a λ T t Π t 1 γ 1 X ^ t , d t a . e . ,
u ^ I t = Σ t r t X ^ t 1 γ , d t a . e .
This consumption–investment strategy determines a wealth process given by
X t = x 0 + 0 t r 0 s + 1 1 γ r s Σ s r s a λ T s Π s 1 γ 1 X ^ s d s + 0 t X ^ s 1 γ r s Σ s σ s d W s , t 0 , T .
The above solution is comparable with the one obtained by Marín-Solano and Navas (2010); Ekeland and Pirvu (2008); and Yong (2012).

5.2. Logarithmic Utility Function

Now, let us analyze the case where φ c = ln c and h x = a ln x , with a > 0 . In this case, the PDE 36 reduces to
θ t t , x + θ x t , x r 0 t x r t Σ t r t θ t , x θ x t , x λ T t θ t , x 1 + 1 2 θ x x t , x r t Σ t r t θ t , x θ x t , x 2 + r 0 t θ t , x = 0 , t , x 0 , T × R , θ T , x = a x .
Once again, we know that the solution of 57 will be of the form
θ t , x = Π t a x , for t 0 , T ,
where Π · C 1 0 , T , R . By substituting in 57 , we get
Π t t + 1 a λ T t = 0 , for t 0 , T , Π T = 1 ,
which is explicitly solved by
Π t = 1 + t T 1 a λ T r d r , for t 0 , T .
In view of Theorem 3, the representation of the Nash equilibrium strategies 38 39 gives
c ^ t = a λ T t + t T λ T t λ T r d r 1 X ^ t , d t a . e . ,
u ^ I t = Σ t r t X ^ t , d t a . e .
This consumption–investment strategy determines a wealth process given by
X t = x 0 + 0 t r 0 s + r s Σ s r s a λ T s + s T λ T s λ T r d r 1 X ^ s d s + 0 t r s Σ s σ s X ^ s d W s .

5.3. Exponential Utility Function

Next, we consider the case where φ c = e γ c γ and h x = a e γ x γ , with a , γ > 0 . The terminal condition PDE 36 becomes
θ t t , x + θ x t , x r 0 t x r t Σ t r t θ t , x θ x t , x 1 γ ln λ T t θ t , x + 1 2 θ x x t , x r t Σ t r t θ t , x θ x t , x 2 + r 0 t θ t , x = 0 , t , x 0 , T × R , θ T , x = a e γ x .
We try a solution of the form
θ t , x = a e γ ϕ t x + ψ t , for t 0 , T ,
where ϕ · , ψ · C 1 0 , T , R such that ϕ T = 1 and ψ T = 0 . By substituting in 62 , we get
γ ϕ t t + γ ϕ t 2 γ ϕ t r 0 t x 1 2 r t Σ t r t γ ψ t t ϕ t ln a λ T t + γ ϕ t ψ t + r 0 t = 0 .
This suggests that functions ϕ · and ψ · should solve the following system of equations:
ϕ t t = r 0 t ϕ t + ϕ t 2 , t 0 , T , ψ t t = 1 γ ϕ t ln a λ T t + ϕ t ψ t 1 2 γ r t Σ t r t + 1 γ r 0 t , t 0 , T , ϕ T = 1 , ψ T = 0 ,
which is explicitly solvable for t 0 , T , by
ϕ t = e t T r 0 τ d τ 1 + t T e l T r 0 τ d τ d l ,
and
ψ t = e t T ϕ τ d τ t T e l T ϕ τ d τ 1 γ ϕ l ln λ T l a + 1 2 γ r t Σ t r t r 0 l γ d t .
The representation of the Nash equilibrium strategies 38 39 gives
c ^ t = 1 γ ln a λ T t + ϕ t X ^ t + ψ t , d t a . e .
u ^ I t = 1 γ Σ t r t ϕ t 1 , d t a . e .
This consumption–investment strategy determines a wealth process given by
X t = x 0 + 0 t r 0 s ϕ s X ^ s + 1 γ r s Σ s r s ϕ s 1 ln a λ T s ψ s d s + 0 t 1 γ ϕ s 1 r s Σ s σ s d W s , t 0 , T .
The above solution is comparable with the ones obtained in Marín-Solano and Navas (2010) by solving an extended Hamilton–Jacobi–Bellman (HJB) equations.

6. Special Discount Function

As well documented in Marín-Solano and Navas (2010), an agent making a decision at time t is usually called the t-agent, and can act in two different ways: naive and sophisticated. Naive agents make decisions without taking into account that their preferences will change in the near future, and then any t-agent will solve the problem as a standard optimal control problem with initial condition X ( t ) = x t , and his decision will be, in general, time-inconsistent. In order to obtain a time consistent strategy, the t-agent should be sophisticated, in the sense of taking into account the preferences of all the s-agents, for s t , T . Therefore, the approach to handle the time inconsistency in dynamic decision-making problems is by considering time-inconsistent problems as non-cooperative games with a continuous number of players, in which decisions at every instant of time are selected. The solution to the problem of the agent with non-constant discounting should be constructed by looking for the sub-game perfect equilibrium of the associated game with an infinite number of t-agents. In Marín-Solano and Navas (2010), the authors looked for a solution of a sophisticated agent to the modified HJB (which is not a partial differential equation due to the presence of a non-local term). Then, they need to define the Markov equilibrium strategies, while, in our work, and different from Marín-Solano and Navas (2010), we use the open-loop equilibrium strategies. This is a significant difference which leads to obtain an important change in the results.

6.1. Exponential Discounting with Constant Discount Rate (Classical Model)

At first, we consider the standard exponential discount function λ t = e δ 0 t , t 0 , T , where δ 0 > 0 is a constant representing the discount rate. In this case, our equilibrium solution for the three cases become:
(1) Logarithmic utility
c ^ ( t ) = 1 a e T t δ 0 + t T e l t δ 0 d l X ^ t , d t a . e . , u ^ I ( t ) = Σ t r t X ^ t , d t a . e .
(2) Power utility
c ^ t = a e T t δ 0 1 γ 1 e t T K τ γ 1 d τ 1 + t T a e T τ δ 0 1 γ 1 e τ T K l γ 1 d l d τ X ^ t , d t a . e . , u ^ I ( t ) = Σ t r t X ^ t 1 γ , d t a . e .
(3) Exponential utility
c ^ t = 1 γ ln a e T t δ 0 + ϕ t X ^ t + ψ t , d t a . e . , u ^ I t = Σ t r t 1 γ ϕ t , d t a . e .
where K · , ϕ · are given by 52 and 65 , respectively, and
ψ t = 1 γ e t T ϕ τ d τ t T e l T ϕ τ d τ ϕ l ln e T l δ 0 a + 1 2 r l Σ l r l r 0 l d l .
Notice that our solutions given above coincide with the optimal solutions of classical Merton portfolio problem (see, e.g., Marín-Solano and Navas (2010) in the case with constant discount rate). This confirms the well-known fact that the time-consistent equilibrium strategy for an exponential discount function is nothing but the optimal strategy. A relevant remark is that the portfolio rule is independent of the discount factor, and it is the same for a non-exponential discount function.

6.2. Exponential Discounting with Non-Constant Discount Rate (Karp’s Model)

Now, following Karp (2007), let us assume that the instantaneous discount rate is non-constant, but a continuous and positive function of time δ l , for l 0 , T . Impatient agents will be characterized by a non-increasing discount rate δ · . The discount factor used to evaluate a payoff at times τ ≥ 0, is given by
λ τ = e 0 τ δ l d l .
In this case, the objective is exactly the same as Marín-Solano and Navas (2010), in which the equilibrium is, however, defined within the class of feedback controls. In Marín-Solano and Navas (2010), the (feedback) equilibrium consumption-investment solutions (also called the sophisticated consumption-investment strategies) are summarized as:
(1) Logarithmic utility
c ^ t = 1 a e 0 T t δ τ d τ + t T e 0 l t δ τ d τ d l X ^ t , d t a . e . ,
u ^ I ( t ) = Σ t r t X ^ t , d t a . e .
(2) Power utility
c ^ t = α t 1 γ 1 X ^ t , d t a . e . ,
u ^ I ( t ) = Σ t r t X ^ t 1 γ , d t a . e . ,
where α · is the solution of the integro-differential equation,
α t t δ T t K t α t + 1 γ α t γ 1 γ t T e 0 s t δ l d l δ s t δ T t α s γ 1 γ e γ t s Δ τ d τ d s = 0 , α T = a .
with K t given by 52 and
Δ τ = r 0 τ + 1 1 γ r τ Σ τ r τ α τ 1 1 γ .
(3) Exponential utility
c ^ t = ϕ t X ^ t + C t ln γ a ϕ t γ , d t a . e . ,
u ^ I ( t ) = Σ t r t 1 γ ϕ t , d t a . e . ,
where ϕ · is given by 65 and C · satisfies the following very complicated integro-differential equation,
C t t C t ϕ t + 1 γ ϕ t ln a γ ϕ t + 1 2 γ r t Σ t r t + 1 γ δ T t ϕ t K C t , t = 0 , C T = 0 ,
where
K C t , t = E t T e 0 s t δ l d l δ s t δ T t ϕ t × e γ C s C t + t s ϕ τ Z τ d τ + t s 1 γ r τ Σ τ σ τ d W τ d s ,
with
Z τ = 1 γ ϕ τ r τ Σ τ r τ C τ + 1 γ ln γ a ϕ τ .
Our (open-loop) equilibrium solutions reduce to
(1) Logarithmic utility
c ^ t = 1 a e 0 T t δ τ d τ + t T e T l T t δ τ d τ d l X ^ t , d t a . e . ,
u ^ I ( t ) = Σ t r t X ^ t , d t a . e .
(2) Power utility
c ^ t = a e 0 T t δ τ d τ 1 γ 1 e t T K τ γ 1 d τ 1 + t T a e 0 T τ δ τ d τ 1 γ 1 e τ T K l γ 1 d l d τ X ^ t , d t a . e . ,
u ^ I ( t ) = Σ t r t X ^ t 1 γ , d t a . e .
(3) Exponential utility
c ^ t = 1 γ ln a e 0 T t δ τ d τ + ϕ t X ^ t + ψ t , d t a . e . ,
u ^ I ( t ) = Σ t r t 1 γ X ^ t ϕ t , d t a . e . ,
where K · , ϕ · are given by 52 and 65 , respectively, and
ψ t = e t T ϕ τ d τ t T e l T ϕ τ d τ 1 γ ϕ l ln e 0 T t δ τ d τ a + 1 2 γ r l Σ l r l r 0 l γ d l .
Remark 5.
Comparing the results of this special case with our solutions, we find the following facts: The equilibrium proportion investment strategies coincide in the three cases. The consumption strategies are different in the three cases. Moreover, our equilibrium consumption strategies are well defined and explicitly given, while, in Marín-Solano and Navas (2010), equilibrium consumption strategies in the case of Power utility, as well as in the case of Exponential utility, are obtained via a very complicated integro-differential equations, in which unique solvability are not established.

Author Contributions

All authors have contributed equally. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

The authors are particularly grateful for the editor and anonymous reviewers for their valuable and constructive suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

We derive the proof of Proposition 1 by means of the duality analysis and some limiting procedures. Moreover, since our objective function is not in quadratic form, we need to adapt some results obtained in Hu et al. (2012) and Hu et al. (2015) according to our control problem which concerns a general and non-necessary quadratic utility maximization.
Proof 
(Proof of Proposition 1.). The estimates ( 25 ) ( 27 ) follow from Theorem 4.4 in Yong and Zhou (1999). Moreover, the following representation holds for the objective functional:
J t , X ^ t , u ε · J t , X ^ t , u ^ · = E t t T λ s t φ Γ u ε s φ Γ u ^ s d s + λ T t h X ^ ε T h X ^ T .
Now, from 22 and by applying the second order Taylor-Young expansion, we find that
h X ^ ε T h X ^ T = h x X ^ T y ε , v T + z ε , v T + 1 2 h x x X ^ T y ε , v T + z ε , v T 2 + o y ε , v T + z ε , v T 2 .
By applying the second order Taylor-Lagrange expansion, we get
φ Γ u ε s φ Γ u ^ s = φ x Γ u ^ s Γ , v + 1 2 φ x x Γ u ^ s + θ v 1 t , t + ε Γ Γ v , v .
From 27 , it holds that
J t , X ^ t , u ε · J t , X ^ t , u ^ · = E t t T λ s t φ x Γ u ^ s Γ , v + 1 2 φ x x Γ u ^ s + θ v 1 t , t + ε Γ Γ v , v 1 t , t + ε d s + λ T t h x X ^ T y ε , v T + z ε , v T + 1 2 h x x X ^ T y ε , v T + z ε , v T 2 + o ε .
Notice that
λ T t h x X ^ T y ε , v T + z ε , v T + 1 2 h x x X ^ T y ε , v T + z ε , v T 2 = p T ; t y ε , v T + z ε , v T + 1 2 P T ; t y ε , v T + z ε , v T 2 .
Now, by applying Itô’s formula to s p s ; t y ε , v s + z ε , v s on t , T , we get
E t p T ; t y ε , v T + z ε , v T = E t t t + ε v B s p s ; t + v D s q ˜ s ; t d s .
Again, by applying Itô’s formula to s P s ; t y ε , v s + z ε , v s 2 on t , T , we get
E t P T ; t y ε , v T + z ε , v T 2 = E t t t + ε 2 v y ε , v s + z ε , v s B s P s , t + D s Q ˜ s , t + v D s D s v P s , t d s ,
where Q ˜ s ; t = 0 , Q s ; t . On the other hand, we conclude from (H1), together with 27 , that
E t t t + ε y ε , v s + z ε , v s B s P s , t + D s Q ˜ s , t d s = o ε .
By taking A . 1.4 , A . 1.5 , and A . 1.6 in A . 1.3 , it follows that
J t , X ^ t , u ε · J t , X ^ t , u ^ · = E t t t + ε B s p s ; t + D s q ˜ s ; t + λ s t φ x Γ u ^ s 1 t , t + ε Γ , v + 1 2 λ ( s t ) φ x x Γ , u ^ s + θ v 1 t , t + ε Γ Γ + P s , t D s D s v , v d s + o ε ,
which is equivalent to ( 28 ) .
Now, we derive the proof of Lemma 1 by using some limiting procedures. First, let us recall the following lemma which was proved by Wang (2020) (Lemma 3.3). □
Lemma A1.
If ϕ · = ( ϕ 1 · , , ϕ m · ) M F p ( 0 , T ; R m ) with m N and p > 1 , then, for a.e. t [ 0 , T ) , there exists a sequence { ε n t } n N ( 0 , T t ) depending on t such that l i m n ε n t = 0 and
lim n 1 ε n t E t t t + ε n t ϕ i s ϕ i t p d s = 0 , for i = 1 , . . . , m , d P a . s .
Proof 
(Proof of Lemma 1). We define, for t 0 , T and s t , T ,
p ¯ s ; t , q ¯ s ; t : = 1 λ T t e s T r 0 τ d τ p s ; t , q s ; t .
Then, for any t 0 , T , in the interval t , T , the pair p ¯ · ; t , q ¯ · ; t satisfies
d p ¯ s ; t = q ¯ s ; t d W s , s t , T , p ¯ T ; t = h x X ^ T .
Moreover, it is clear that, from the uniqueness of solutions to A . 2.1 , we have the equality p ¯ s ; t 1 , q ¯ s ; t 1 = p ¯ s ; t 2 , q ¯ s ; t 2 , for any t 1 , t 2 , s 0 , T such that 0 < t 1 < t 2 < s < T . Hence, the solution p ¯ · ; t , q ¯ · ; t does not depend on the variable t, and this allows us to denote the solution of A . 2.1 by p ¯ · , q ¯ · .
We have then, for any t 0 , T , and s t , T ,
p s ; t , q s ; t = λ T t e s T r 0 τ d τ p ¯ s , q ¯ s .
Now, using A . 2.2 we have, under (H2), for any t 0 , T and s t , T ,
p s ; t p s ; s sup t s t + ε λ T t λ T s e s T r 0 τ d τ p ¯ s ,
and
q s ; t q s ; s sup t s t + ε λ T t λ T s e s T r 0 τ d τ q ¯ s .
From which, we have, for any a > 0 , t 0 , T , and ε 0 , T t ,
P 1 ε E t t t + ε H s ; t d s 1 ε E t t t + ε H s ; s d s a , 1 a E 1 ε E t t t + ε H s ; t d s 1 ε E t t t + ε H s ; s d s , C sup t s t + ε λ T t λ T s 1 ε E t t + ε p ¯ s + q ¯ s d s + sup t s t + ε λ s t 1 1 ε t t + ε E φ x Γ u ^ s d s .
Noting that, since λ · is continuous, we get
lim ε 0 sup t s t + ε λ T t λ T s = 0
for t 0 , T . Moreover, since p ¯ · , q ¯ · L F 2 0 , T ; R × M F 2 0 , T ; R d , we get
lim ε 0 sup t s t + ε λ T t λ T s 1 ε E t t + ε p ¯ s + q ¯ s d s = 0 .
Noting that λ 0 = 1 , then lim ε 0 sup t s t + ε λ s t 1 = 0 . According to (H3), by using the dominated convergence theorem,
lim ε 0 1 ε t t + ε E φ x Γ u ^ s d s = E φ x Γ u ^ t < , d t a . e .
Therefore,
lim ε 0 E 1 ε E t t t + ε H s ; t d s 1 ε E t t t + ε H s ; s d s = 0 .
Hence, for each t, there exists a sequence ε n t n 0 0 , T t such that lim n ε n t = 0 and
lim n 1 ε n t E t t t + ε n t H s ; t d s 1 ε n t E t t t + ε n t H s ; s d s = 0 , d P a . s .
Moreover, since φ x Γ u ^ · M F p 0 , T ; R and
p ¯ · , q ¯ · L F 2 0 , T ; R × M F 2 0 , T ; R d ,
we get, from Lemma A1, that there exists a sub-sequence of ε n t n 0 , which is also denoted by ε n t n 0 such that
lim n 1 ε n t E t t t + ε n t H s ; s d s = H t ; t , d t a . e , d P a . s .
To derive the statement 2) in the Lemma 1, it is sufficient to prove the following, for each t there exists a sequence ε n t n 0 0 , T t such that lim n ε n t = 0 and
lim n 1 ε n t E t t t + ε n t λ s t φ x x Γ u ^ s + θ v 1 t , t + ε d s = φ x x Γ u ^ t , lim n 1 ε n t E t t t + ε n t λ s t σ s σ s P s ; t d s = σ t σ t P t ; t .
Let us prove the first limit. We have
1 ε E t t t + ε λ s t φ x x Γ u ^ s + θ v 1 t , t + ε d s 1 ε E t t t + ε φ x x Γ u ^ s d s sup t s t + ε λ s t 1 1 ε E t t t + ε sup η M φ x x Γ u ^ s + η d s .
Applying the same arguments used in the first limit, we obtain, according to Lemma A1,
lim n 1 ε n t E t t t + ε n t φ x x Γ u ^ s d s = φ x x Γ u ^ t ,
at least for a sub-sequence. □

Appendix B. Conclusions

It is well known that the inconsistent optimal control problems are complicated to solve in general. In this paper, we have studied optimal investment and consumption problem, since the objective functional may depend on the non-exponential discount function. So, inspired from the previous literature, we made some restrictive conditions on the coefficients of the utility function, which allows to treat some limiting procedures in the original problem.
We have used the game theoretic approach to handle the time inconsistency. Specifically, feedback Nash equilibrium controls are explicitly constructed as an alternative of optimal controls. This has been accomplished through the necessary and sufficient condition for equilibrium during stochastic system that includes a flow of forward-backward stochastic differential equations. We derive the closed-from expression of the equilibrium investment-consumption strategy. Moreover, some particular cases of our model are discussed and compared with the previous literature from the extended HJB equations with a verification theorem.
The work can be extended in several ways. For example, this approach can be extended to a general continuous-time stochastic control problem with delay under a fairly general time-inconsistent objective functional. Another challenging problem, is the study of some statistical testing method for validation of the smoothness assumptions about the coefficients; see, e.g., Pešta and Wendler (2020). The research on these topics is in progress and will appear in our forthcoming papers.

References

  1. Ainslie, George. 1995. Specious reward: A behavioral theory of impulsiveness and impulse control. Psychological Bulletin 82: 463–96. [Google Scholar] [CrossRef] [Green Version]
  2. Barro, Robert J. 1999. Ramsey meets Laibson in the neoclassical growth model. Quarterly Journal of Economics 114: 1125–52. [Google Scholar] [CrossRef]
  3. Basak, Suleyman, and Georgy Chabakauri. 2010. Dynamic mean-variance asset allocation. Review of Financial Studies 23: 2970–3016. [Google Scholar] [CrossRef]
  4. Bensoussan, Alain, Kwok C. Wong, and Sheung C. P. Yam. 2019. A paradox in time-consistency in the mean-variance problem? Finance and Stochastics 23: 173–207. [Google Scholar] [CrossRef]
  5. Björk, Tomas, and Agatha Murgoci. 2010. A general theory of Markovian time-inconsistent stochastic control problems. SSRN Electronic Journal 55: 1694759. [Google Scholar] [CrossRef] [Green Version]
  6. Björk, Tomas, Agatha Murgoci, and Xun Y. Zhou. 2014. Mean-variance portfolio optimization with state-dependent risk aversion. Mathematical Finance 24: 1–24. [Google Scholar] [CrossRef]
  7. Chen, Shumin, Zhongfei Li, and Yan Zeng. 2014. Optimal dividend strategies with time-inconsistent preferences. Journal of Economic Dynamics and Control 46: 150–72. [Google Scholar] [CrossRef]
  8. Cong, Fengyu, and Cornelis W. Oosterlee. 2016. On pre-commitment aspects of a time-consistent strategy for a mean-variance investor. Journal of Economic Dynamics and Control 70: 178–93. [Google Scholar] [CrossRef] [Green Version]
  9. Cui, Xiangyu, Duan Li, and Yun Shi. 2017. Self-coordination in time inconsistent stochastic decision problems: A planner-doer game framework. Journal of Economic Dynamics and Control 75: 91–113. [Google Scholar] [CrossRef]
  10. Czichowsky, Christoph. 2013. Time-mean-variance portfolio selection in discrete and continuous time. Finance and Stochastics 17: 227–71. [Google Scholar] [CrossRef] [Green Version]
  11. Ekeland, Ivar, and Ali Lazrak. 2008. Equilibrium policies when preferences are time inconsistent. arXiv arXiv:0808.3790v1. [Google Scholar]
  12. Ekeland, Ivar, and Traian A. Pirvu. 2008. Investment and consumption without commitment. Mathematics and Financial Economics 2: 57–86. [Google Scholar] [CrossRef] [Green Version]
  13. Ekeland, Ivar, Oumar Mbodji, and Traian A. Pirvu. 2012. Time-consistent portfolio management. SIAM Journal of Financial Mathematics 3: 1–32. [Google Scholar] [CrossRef] [Green Version]
  14. Fleming, Wendell H., and annd Thaleia Zariphopoulou. 1991. An optimal investment/consumption model with borrowing constraints. Mathematics of Operations Research 16: 802–22. [Google Scholar] [CrossRef]
  15. Goldman, Steven M. 1980. Consistent plans. Review of Financial Studies 47: 533–37. [Google Scholar] [CrossRef]
  16. Hamaguchi, Yushi. 2019. Time-inconsistent consumption-investment problems in incomplete markets under general discount functions. arXiv arXiv:1912.01281v2. [Google Scholar]
  17. Hu, Y., H. Jin, and X. Y. Zhou. 2012. Time-inconsistent stochastic linear quadratic control. SIAM Journal on Control and Optimization 50: 1548–72. [Google Scholar] [CrossRef] [Green Version]
  18. Hu, Ying, Hanqing Jin, and Xun Y. Zhou. 2012. Time-Inconsistent Stochastic Linear-Quadratic Control: Characterization and Uniqueness of Equilibrium. SIAM Journal on Control and Optimization 55: 1261–79. [Google Scholar] [CrossRef]
  19. Karatzas, Ioannis, John P. Lehoczky, and Steven E. Shreve. 1987. Optimal portfolio and consumption decisions for a “small investor” on a finite horizon. SIAM Journal on Control and Optimization 25: 1157–86. [Google Scholar] [CrossRef] [Green Version]
  20. Karp, Larry. 2007. Non-constant discounting in continuous time. Journal of Economic Theory 132: 557–68. [Google Scholar] [CrossRef] [Green Version]
  21. Krusell, Per, and Anthony A. Smith. 2003. Consumption and savings decisions with quasi-geometric discounting. Econometrica 71: 366–75. [Google Scholar] [CrossRef] [Green Version]
  22. Kydland, Finn E., and Edward E. Prescott. 1997. Rules rather than discretion: The inconsistency of optimal plans. Journal of Political Economy 85: 473–92. [Google Scholar] [CrossRef] [Green Version]
  23. Loewenstein, George, and Drezen Prelec. 1992. Anomalies in inter-temporal choice: Evidence and an interpretation. The Quarterly Journal of Economics 107: 573–97. [Google Scholar] [CrossRef]
  24. Marín-Solano, Jesús, and Jorge Navas. 2010. Consumption and portfolio rules for time-inconsistent investors. European Journal of Operational Research 201: 860–72. [Google Scholar] [CrossRef] [Green Version]
  25. Menoncin, Francesco, and Elena Vigna. 2020. Mean-variance dynamic optimality for DC pension schemes. European Actuarial Journal 10: 125–48. [Google Scholar] [CrossRef] [Green Version]
  26. Merton, Robert C. 1969. Lifetime portfolio selection under uncertainty: The continuous-time case. Review of Econometrics Statistics 51: 247–57. [Google Scholar] [CrossRef] [Green Version]
  27. Merton, Robert C. 1971. Optimum consumption and portfolio rules in a continuous-time model. Journal of Economic Theory 3: 373–413. [Google Scholar] [CrossRef] [Green Version]
  28. Phelps, Edmund S., and Robert A. Pollak. 1968. On second-best national saving and game-equilibrium growth. The Review of Economic Studies 35: 185–99. [Google Scholar] [CrossRef]
  29. Pešta, Michal, and Martin Wendler. 2020. Nuisance-parameter-free changepoint detection in non-stationary series. TEST 29: 379–408. [Google Scholar] [CrossRef]
  30. Pliska, Stanley R. 1986. A stochastic calculus model of continuous trading: Optimal portfolios. Mathematics of Operations Research 11: 371–82. [Google Scholar] [CrossRef]
  31. Pollak, Robert A. 1968. Consistent planning. The Review of Economic Studies 35: 201–8. [Google Scholar] [CrossRef]
  32. Ramsey, Frank P. 1928. A Mathematical Theory of Saving. The Economic Journal 38: 543–59. [Google Scholar] [CrossRef]
  33. Strotz, Robert H. 1955–1956. Myopia and inconsistency in dynamic utility maximization. The Review of Economic Studies 23: 165–80. [Google Scholar] [CrossRef]
  34. Van Staden, Pieter M., Duy-Mihn Dang, and Peter A. Forsyth. 2018. Time-consistent mean-variance portfolio optimization: A numerical impulse control approach. Insurance: Mathematics and Economics 83: 9–28. [Google Scholar] [CrossRef] [Green Version]
  35. Wang, Tianxiao. 2020. Uniqueness of equilibrium strategies in dynamic mean-variance problems with random coefficients. Journal of Mathematical Analysis and Applications 490: 124199. [Google Scholar] [CrossRef]
  36. Wang, Jin, and Peter A. Forsyth. 2012. Comparison of mean variance like strategies for optimal asset allocation problems. International Journal of Theoretical and Applied Finance 15: 1250014. [Google Scholar] [CrossRef] [Green Version]
  37. Yong, Jiongmin. 2011. A deterministic linear quadratic time-inconsistent optimal control problem. Mathematical Control and Related Fields 1: 83–118. [Google Scholar] [CrossRef] [Green Version]
  38. Yong, Jiongmin. 2012. Time-inconsistent optimal control problems and the equilibrium HJB equation. Mathematical Control and Related Fields 2: 271–329. [Google Scholar] [CrossRef]
  39. Yong, Jiongmin, and Xun Y. Zhou. 1999. Stochastic Controls: Hamiltonian Systems and HJB Equations. New York: Springer. [Google Scholar]
  40. Zhao, Qian, Yang Shen, and Jiaqin Wei. 2014. Consumption-investment strategies with non-exponential discounting and logarithmic utility. European Journal of Operational Research 283: 824–35. [Google Scholar] [CrossRef]
  41. Zou, Ziran, Shou Chen, and Lei Wedge. 2014. Finite horizon consumption and portfolio decisions with stochastic hyperbolic discounting. Journal of Mathematical Economics 52: 70–80. [Google Scholar] [CrossRef] [Green Version]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Alia, I.; Chighoub, F.; Khelfallah, N.; Vives, J. Time-Consistent Investment and Consumption Strategies under a General Discount Function. J. Risk Financial Manag. 2021, 14, 86. https://0-doi-org.brum.beds.ac.uk/10.3390/jrfm14020086

AMA Style

Alia I, Chighoub F, Khelfallah N, Vives J. Time-Consistent Investment and Consumption Strategies under a General Discount Function. Journal of Risk and Financial Management. 2021; 14(2):86. https://0-doi-org.brum.beds.ac.uk/10.3390/jrfm14020086

Chicago/Turabian Style

Alia, Ishak, Farid Chighoub, Nabil Khelfallah, and Josep Vives. 2021. "Time-Consistent Investment and Consumption Strategies under a General Discount Function" Journal of Risk and Financial Management 14, no. 2: 86. https://0-doi-org.brum.beds.ac.uk/10.3390/jrfm14020086

Article Metrics

Back to TopTop