1. Introduction
Theoretical and empirical finance research involves the evaluation of conditional expectations, which, in a continuous time jump-diffusion setting, can be related to second order partial integral differential equations of parabolic type (PIDEs) by the Feynman-Kac theorem, and other types of equations such as backward stochastic differential equations with jumps (BSDEJs) or quasi-linear PIDEs in more complicated settings. In theoretical continuous-time finance, many problems, such as asset pricing with market frictions, dynamic hedging or dynamic portfolio-consumption choice problems, can be related to Hamilton-Jacobi-Bellman (HJB) equations via dynamic programming techniques. The HJB equations, from another perspective, are equivalent to BSDEs derived from a probabilistic approach. The nonlinear BSDEs, studied in [2], can be decomposed into a sequence of linear equations, which can be solved by taking conditional expectations, via Picard iteration. For empirical studies, the focus of the literature has been the evaluation of the cross sectional conditional risk-adjusted expected returns and the explanation of them using factors. See [3] [4] and [5] as good illustrations. It is easily seen that, regardless of the fact whether the underlying models are continuous-time or discrete-time, evaluating conditional expectations is inevitable in finance literature. Moreover, in order to perform XVA computations for the measurement of counterparty credit risk, we need to evaluate the conditional expectations, i.e., the derivative prices, on a future simulation grid, as outlined in [6]. These facts call for efficient methods to compute the quantities aforementioned.
In this paper, we extend the basis function expansion approach proposed in [1] with machine learning techniques. Specifically, we propose new efficient methods to evaluate conditional expectations, regardless of the dynamics of the underlying stochastic process, as long as they can be simulated. Rigorous convergence proofs are given using Hilbert space theory. The methodologies can be applied to time zero pricing as well as pricing on a future simulation grid, with the advantage of ANN approximation most prominent in high dimensional problems. In the sequel, we show applications of our methodologies on the pricing of European derivatives and extension to contracts with optimal stopping feature is straightforward through either [1] approach or reflected-BSDEs.
Compared to the literature on traditional stochastic analysis, our methodologies are able to handle large data sets and high-dimensional problems, therefore suffering much less from the curse of dimensionality due to the nature of ANN methods. Moreover, our methodologies are very efficient when evaluating solutions of BSDEJs and PIDEs on a future simulation grid, where none of the traditional methodologies applies. With respect to recent machine learning literature on numerical solutions to BSDEs and PDEs, our methodologies enjoy the theoretical advantage of being able to handle equations with jump-diffusion and convergence results are provided. When applied to the solutions of BSDEJs and PIDEs, our methodologies require much less number of parameters, as compared to the current machine learning based methods to be mentioned below. At any step in the solution process, only one ANN is needed and we do not require nested optimization. In terms of application, not all the prices of OTC derivatives can be easily translated into BSDEJs and PIDEs, for example, a range accrual with both American and barrier (knock-out, for example) feature. However, our methodologies are naturally suitable in those situations. To conclude, our methods enjoy many theoretical and empirical advantages, which makes them attractive and novel.
There has been a huge literature on applications of machine learning techniques to financial research. Classical applications focus on the prediction of market variables such as equity indexes or FX rates and the detection of market anomalies, for example, [7] and [8]. Option pricing via a brute-force curving fitting by ANNs dates back to [9]. More applications of machine learning in finance, especially option pricing prediction, are surveyed in [10]. See references therein. Pricing of American options in high dimensions can be found in [11], which is closest to our method 1. However, there are several improvements of our methods compared to this reference. First of all, we enable deep neural network (DNN) approximation and show convergence. Second, we can incorporate constraints in DNN approximation estimation and prove the mathematical validity of this approach. Third, we propose two more efficient methods to complement the first method of ours. Our treatment of constraints in the estimation of DNNs extends the work of [12] in that we can deal with a larger class of constraints by specifying a general Hilbert subspace as the constrained set. Risk measure computation using machine learning can be found in [13]. Applications of machine learning function approximation on financial econometrics can be found in [14], [15], [16] and [17]. Recent applications include empirical and theoretical asset pricing, reinforcement learning and Q-learning in solving dynamic programming problems such as optimal investment-consumption choice, option pricing and optimal trading strategies construction, e.g., [18], [19], [20], [21], [22], [23], [24], [25], [26], [27], [28] and references therein. Numerical methods to solve PDEs and BSDEs or the related inverse problems can be found in [29], [30], [31], [32], [33], [34], [35], [36], [37], [38] and [39]. Machine learning based methods enjoy the advantage of being fast, able to handle large data sets and high dimensional problems.
Our methodologies are combinations of traditional statistical learning theory and stochastic analysis with advanced machine learning techniques, introducing powerful function approximation method via the universal approximation theorem and artificial neural networks (ANNs), while preserving the regression-type analysis documented in [1]. The methods are very easy to use, effective, accurate as illustrated by numerical experiments and time efficient. They are different from the convergent expansion method, e.g., [40], simulation methods such as [41], [42], [43] and [44] or the asymptotic expansion method proposed by [45], [46], [47], [48] [49] [50] [51] [52], in that we no longer resort to polynomial basis function expansion or small-diffusion type analysis. Our methods are also different from the pure machine learning based ones documented in [29], [30], [31], [32], [33], [34] and [35], in that we utilize the lead-lag regression formula to evaluate the conditional expectations, preserving the time dependent structure and our methods are able to handle jump-diffusion processes easily.
The organization of this paper is as follows. Section 2 documents the methodologies. Section 3 illustrates the usefulness of our methods by considering European and American derivatives pricing. Section 4 considers numerical experiments and Section 5 concludes. An outline of the proofs and other applications can be found in the appendices.
2. The Methodology
Mathematical Setup
We use a Markov process modeled by a jump-diffusion as illustration. Suppose that we have a stochastic differential equation with jumps
(1)
where
,
is a standard d-dimensional Brownian motion and
is a q-dimensional compensated Poisson random measure, with the compensator
. Information filtration
is generated by
. We hope to evaluate the conditional expectation
for any
, e.g., see [53]. Assumptions on
and X are stated below.
Assumption 1 (On Growth Condition of ψ).
has polynomial growth in its argument x, i.e., there exists a positive integer P, independent of x, such that for all
, we have, for constant C independent of x
(2)
The following assumption is w.r.t. X.
Assumption 2 (On X). There exists a unique strong solution to Equation (1) and X has finite polynomial moments of all orders.
The General Approximation Theory
First, we need the following assumptions, definitions and results. Please note that, some of the spaces we introduce are actually conditional ones. The discussions of conditional Hilbert spaces can be found in [54], e.g.,
is a conditional Hilbert space for all
.
Definition 3 (Projection Operator). For Hilbert spaces
and
, where
. Define
as the projection of
onto
.
Definition 4 (Orthogonal Space). For Hilbert spaces
and
, where
. Define
as the orthogonal space of
in
.
Definition 5 (Spanning the Hilbert Space). Assume that
is a set of elements in Hilbert space
and
is an index set. Define
as the intersection of all Hilbert subspaces of
containing
.
Assumption 6 (On Joint Continuity).
and
are two Hilbert spaces and
. Moreover,
is a sequence of Hilbert sub-spaces of
satisfying
for any
and
. We have
for any
and
.
The next two theorems are well-known in the literature.
Theorem 7 (Hilbert Projection Theorem). Let
be two Hilbert spaces and let
. Then,
exists and is unique. Moreover, it is characterized uniquely by
.
Theorem 8 (Repeated Projection Theorem). Let
be three Hilbert spaces. Then, for any
,
.
Remark 9 The conditions of Theorems 7 and 8 on
and
can be relaxed to convexity and completeness instead of Hilbert sub-spaces.
Finally, we have the result below.
Theorem 10. Suppose
is a Hilbert space,
and
are Hilbert subspaces of
satisfying
and
.
, define
and
. Then we have
w.r.t. the norm topology in
, if Assumption 6 is satisfied.
Sometimes we need to add constraints on the calibrated ANN, e.g., the shape constraints. The following assumption and theorem deal with this situation.
Assumption 11 (On Constrained Sub-space). Suppose that
such that
is a sequence of non-empty convex and complete subspaces of
satisfying Assumption 6, where
and
are described.
The following theorem handles the constrained approximation and its convergence.
Theorem 12 (On Constrained Approximation). Under Assumptions 6 and 11, for
, if
, then, we have
.
Remark 13 (On ψ). In Theorem 12, the set
represents prior knowledge on constraints that h satisfies. It can be represented by a set of non-linear inequalities or equalities on functionals of h. Common constraints for option pricing include non-negativity constraint and the positiveness constraint on the second order derivatives. The verification of
satisfying Assumption 6 should be based on a case-by-case manner.
To proceed further, we need the following assumptions.
Assumption 14 (On Some Spaces).
is an increasing sequence of Hilbert sub-spaces of
,
,
. Moreover,
.
Assumption 15 (On Structure of
).
is a set of elements of
, such that
, where
for any
and
, satisfies Assumption 141.
Then, we have the following results.
Lemma 1. For any adapted stochastic process
such that
, if
, we have
(3)
The following proposition is a natural extension of Lemma 1.
Proposition 16. For any measurable function
and stochastic process X such that
and
, we have
(4)
Here
and the above minimization problem has a unique solution. In particular, if X is a Markov process, then
, i.e.,
is a function of time t and
.
We then have the following theorem.
Theorem 17. Under Assumptions 1, 2, 6, 14 and 15, for any adapted stochastic process
such that
and
, we have
(5)
Further, for any measurable function
and stochastic process X such that
and
, we have the following equality
(6)
If X is Markov, then we have
, i.e.,
is a function of time t and
.
The following theorem justifies the Monte Carlo approximation of expectation in the above optimization problems.
Theorem 18 (On Sequential Convergence). Under Assumptions 1, 2, 6, 14 and 15, suppose that
for all
,
and
are M i.i.d. copies of
and
. Then we have
(7)
The following results justify the universal approximation and ANN approximation approaches proposed in this paper.
Proposition 19 (On Universal Approximation Theory). Let
denote the function in the universal approximation theorem mentioned in [55], [56] and [57]. Define
, where X satisfies Equation (1) and Assumption 2,
and
have at most n significant digits in total, where
, i.e., n belongs to the set of natural numbers, j runs from 1 to
and
is the number of all related
, i.e.,
. Then,
satisfies Assumptions 6, 14 and 15. Therefore, Theorems 17 and 18 apply.
Proposition 20 (On Deep Neural Network Approximation). For the DNN defined in ( [58], Definition 1.1], observe that
. Define
(8)
where
satisfies that
,
have at most n total significant digits and
. Then,
, where 1 means
function
for all x, satisfies Assumptions 6, 14 and 15. Therefore, Theorems 17 and 18 apply after a localization argument on
and X on a compact sub-domain in
.
Remark 21 (On DNN). Please note that, in Proposition 20, we do not intend to prove the convergence when the number of layers goes to infinity. Instead, we show convergence when the number of connections goes to infinity, which can be achieved via enlarging the number of neurons in each layer with the total number of layers remaining fixed.
Remark 22 (On Euler Time Discretization). [59] proposes an exact simulation method for multi-dimensional stochastic differential equations. The discussion of discretization error, of the regression approach proposed in this paper, with Euler method is not hard if
satisfies Assumption 1, in which case the dominated convergence theorem and
convergence of Euler method can be applied to show the convergence.
The proofs of the above results can be found in Appendix A. In what follows, we will propose three methods to compute, approximately, the function
in Proposition 16.
Method 1
In general,
, defined in Proposition 16 and Theorem 17, can not be found in closed-form. A natural thought would be to resort to function expansion representations, i.e., to find the solution to the following problem
(9)
where
is an appropriate space for coefficients
and
is a set of functions, with
2 dense in an appropriate function space
3. To further proceed, we seek a truncation of the function representation formula as follows
(10)
for J sufficiently large, where
is a compact set in the Euclidean space where
take values. The last step would be to use Monte Carlo simulation to approximate the unconditional expectation appearing in Equations (9) and (10). Therefore turning the conditional expectation computation problem, into a least-square function regression problem, similar to [1]. An obvious choice of
is polynomial basis, for example, the set of Fourier-Hermite basis functions. For expansion using Fourier-Hermite basis functions in high dimensions, see [60].
In fact, Artificial Neural Networks (ANNs) prove to be an efficient and convergent function approximation tool that we can utilize in the above expressions. Write
(11)
where
denotes an ANN with parameters
.
Note that, via proper time discretization and fixed point iteration, solving a BSDE with jumps can be decomposed into a series of evaluations of conditional expectations. The machine learning based method outlined above can be applied there. We will write down the algorithm to solve a general Coupled Forward-Backward Stochastic Differential Equation with Jumps (CFBSDEJs) in the appendix. Extensions to other types of BSDEJs are possible.
Here we assume that X is a Markov process. To handle path dependency or non-Markov processes, we can apply the backward induction method outlined in [1]. With the machine learning approach, it is easy to see that this method enables us to get the values of conditional expectations on a future simulation grid.
Method 2
Another method to utilize the idea of [1] is inspired by the boosting random tree method (BRT), see, [61], for example. Partition the domain space
4, where
is a set of disjoint sets in
and consider
(12)
The choice of
is important and we can use the machine learning classification techniques (or any classification rule), such as kmeans function in R programming language, in Monte Carlo simulation and related computations. Denote
. It is possible to show that as long as
, we only need finite number of functions, for example,
, to approximate each
and obtain convergence. In practice, although the domain of
is
, it might be centered at a small subspace
, therefore facilitating the partition process. Note also that this method might require us to mollify the function
, if it is not smooth. We adopt finite order Taylor expansion as the function expansion representation approach. The following theorems provide convergence analysis for this method.
Theorem 23. For an appropriate function space
, we have
(13)
Theorem 24. Let
be as described previously and
. Then, we have
(14)
with J large enough, fixed, finite and
is an approximation to
, which satisfies
(15)
for any
,
,
and
is independent of k when K is sufficiently large.
Method 3
Next, we propose an algorithm combining the ANN and universal approximation theorem (UAT). Suppose that
is the space where we are performing the approximation. Also assume that
, i.e., the information filtration is equivalently generated by X. Define an ANN with connection N by
, where x is the state variables that the ANN depends on,
is the vector of parameters and j is its label. We define the following nested regression approximation
(16)
(17)
(18)
(19)
(20)
(21)
where
is the approximate sequence of
.
In this paper, we will test and compare the performance of all of the proposed methods. A general discussion and rigorous proofs can be found in Appendix A5.
3. Applications in Derivatives Pricing
3.1. European Option Pricing
Suppose that the payoff of a European claim can be written as, similar to [62] and [63],
, where
is a stream of cash flows materialized at each time instance t and
is a one-time terminal payoff at time T. Therefore, under no-arbitrage condition, the price of this European payoff can be written as, under risk neutral measure
(22)
where
is the stochastic discount factor. If we assume a Markov structure
and
, then
, i.e.,
is a function of time t and state vector
. This problem is a canonical application of the evaluation of conditional expectations and we can apply the methodologies outlined in Section 2 to solve it. European claims with barrier features can be incorporated and priced in a similar way. For example, the price of a knock-in European claim can be written as
(23)
where
, where
. In our setting, the dynamics of X can be arbitrary, possibly stochastic differential equations with jumps, Markov chains, or even non-Markov processes. Previously, Monte Carlo based method for option pricing can be found in [64] and [65], among others.
3.2. American Option Pricing
Still use
to denote the payoff structure of an American claim, whose price can be obtained via formula
(24)
Here
is the space of all the stopping times in
. We refer the interested readers to [62] and [66] for general derivation and explanation of Equation (24). It is also possible to derive the general BSDE that an American claim price satisfies, for example [67]. Moreover, in [27] and [1], the authors utilize a backward induction approach to solve optimal stopping problems. The idea can be carried out using the methodologies documented in Section 2. American claims with barrier features can be incorporated and priced in a similar way. It is also known that American option prices can be related to reflected BSDEs (RBSDEs), a rigorous discussion of existence and uniqueness of such equations can be found in [68] and references therein.
4. Numerical Experiments
4.1. European Option Pricing
In this section, we consider a Heston model
(25)
(26)
where
is a two dimensional standard Brownian motion. The parameter values are chosen as
,
,
,
,
,
,
and
. Time to maturity is set to be
,
with time discretization step
and
. The number of
simulation paths is
. We price a plain vanilla European call option
as an illustration. The QQ-plots are displayed in Figures 1-10. The first three correspond to a recursive evaluation, i.e., regressing the values at
on state variables at time t. The rest of the plots correspond to direct regression, i.e., regressing the discounted payoffs at time T on state variables at time t. Figures 10-12 are for the prices of a digital call option under Black-Scholes setting and Figures 13-15 are QQ-plots for Delta values. Figure 16 and Figure 17 show the QQ-plots for method 3 under Heston model with 3 nested ANN approximations of size 4 and one ANN approximation of size 12 using R routine nnet. The absolute RMSE for the former is 0.1938% and latter 0.2581%, with the running time 10.36 seconds compared to 52.31 for ANN approximation with size 12.
Figure 1. QQ-plot for Method 1,
and relative pricing error is 1.20%.
Figure 2. QQ-plot for Method 1,
and relative pricing error is 1.50%.
Figure 3. QQ-plot for Method 1,
and relative pricing error is 1.20%.
Figure 4. QQ-plot for Method 1,
and relative pricing error is 1.66%.
Figure 5. QQ-plot for Method 1,
and relative pricing error is 1.75%.
Figure 6. QQ-plot for Method 1,
and relative pricing error is 3.00%.
Figure 7. QQ-plot for Method 2,
and relative pricing error is 1.80%.
Figure 8. QQ-plot for Method 2,
and relative pricing error is 3.50%.
Figure 9. QQ-plot for Method 2,
and relative pricing error is 3.53%.
Figure 10. QQ-plot for Method 1,
and relative pricing error is 0.40%.
Figure 11. QQ-plot for Method 1,
and relative pricing error is 0.80%.
Figure 12. QQ-plot for Method 1,
and relative pricing error is 0.60%.
Figure 13. Delta QQ-plot for Method 1,
.
Figure 14. Delta QQ-plot for Method 1,
.
Figure 15. Delta QQ-plot for Method 1,
.
Figure 16. Price QQ-plot for Method 3,
.
Figure 17. Price QQ-plot for Method 3,
.
4.2. American Option Pricing
Here we refer the readers to [67] for the BSDE satisfied by a plain vanilla American option. For
,
,
,
,
,
and
, the benchmark American option price at
is 9.0660 and the relative difference of our Monte-Carlo price is 0.27%. The running time is less than 30 seconds.
5. Conclusion and Future Research
In this paper, we show how machine learning techniques, specifically, ANN function approximation methods, can be applied to derivatives pricing. We relate pricing problems to the evaluation of conditional expectations via BSDEJs and PIDEs. Future research topics can, potentially, be the development of reinforcement learning methodologies to solve dynamic programming problems and apply them in the context of empirical asset pricing literature. Moreover, the evaluation of energy derivatives calls for SDEJs defined in a Hilbert space. The same theoretical constructions can also be found in the evaluation of fixed income derivatives, such as the random field models proposed and studied in [69]. One can, of course, apply Karhunen-Loéve expansion for a dimension reduction to reduce the problem to the evaluation of conditional expectations of regular SDEJs. However, the development of machine learning based methods to solve directly the conditional expectations on the stochastic processes defined in a Hilbert space is important. In addition, stochastic differential games, that arise in the context of American game options, equity swaps, and the related Mckean-Vlasov type FBSDEJs (mean-field FBSDEJ, see [70] ) are important topics in mathematical finance. They are also related to the theoretical analysis of high-frequency trading. Finding machine-learning based numerical methods to solve these equations is of great interest to us. Last, but not least, machine learning methods in asset pricing and portfolio optimization, which can be found in [71], [72], [73], [28], [74] and [75], admit an elegant way to price financial derivatives under
-measure. For example, we can use the method in [72] to calibrate the SDF process and use [75] to generate market scenarios. These methodologies, combined with the methods documented in this paper and [1], have the potential to solve for any derivative price. We leave all the development to future research.
Acknowledgements
We thank the Editor and the referee for their comments. Moreover, we are grateful to Professor Jérôme Detemple, Professor Marcel Rindisbacher and Professor Weidong Tian for their useful suggestions.
Appendix
A. Convergence of the Proposed Methodologies
Proof of Theorem 10. It is known from the projection theorem of Hilbert space that
and h actually exist and are unique. Moreover,
as indicated by the repeated projection theorem. It is also known that
. As we ask that Assumption 6 hold, we know that
as
.
Proof of Theorem 12. The proof follows from Assumption 6 and Theorem 8. We have
(27)
(28)
(29)
(30)
(31)
This concludes the proof.
Proof of Lemma 1. For any
, we have
(32)
(33)
(34)
(35)
(36)
Therefore we have the claim announced.
Proof of Theorem 17. The proof of this theorem follows from Assumptions 1, 2, 6, 14, 15 and Theorem 10, by choosing
.
Proof of Theorem 18. Essentially, Equation (7) is the result of Gauss-Markov Theorem and the consistency property of OLS estimator.
Proof of Proposition 19. This is a direct consequence of the discussion in ( [57], Section 3) (see Equation (5)) and Theorem 10. To elaborate, consider
,
, its projections h and
on
and
defined in this proposition. Suppose that
and
,
where
and
is a set of orthonormal basis in
. From the repeated projection theorem, we know that
for any
6 and
. From the
property of h, we know that
. Therefore,
as
.
Proof of Proposition 20. This is a direct consequence of the discussion in ( [58], Theorem 2.2), localization arguments, Theorem 10 and the proof of Proposition 19.
Proof of Theorem 23. The first, second and third equality are obvious given an appropriate choice of
depending on the Markov property of X and its moment conditions in Assumption 2. Actually, because of the existence and uniqueness of
such that the RHS of the first equality achieves minimum, we know that
(37)
(38)
From another perspective, we know that
is a piecewise minimization. Therefore
(39)
(40)
The last equality in Equation (13) holds.
Proof of Theorem 24. The proof of this theorem is a direct consequence of Equations (13), (15) and triangle inequality.
B. Other Applications
In this section, we document other applications of our methodologies in finance.
B.1. Joint Valuation and Calibration
Suppose that there are N derivatives contracts whose prices at time
can be expressed as
. Their payoffs are
, where X is an
r-dimensional vector of state variables. Sometimes we write
to explicitly state dependence of X on its vector of parameters
. Here suppose
satisfies a system of stochastic differential equations with jumps
(41)
The main idea is that
might contain derivatives contracts from different asset classes or hybrid ones. Therefore, we need to model X as a joint high dimensional cross-asset system. One potential problem is that
is in general a high-dimensional vector, which will be hard to estimate using usual optimization routines in R or MATLAB software system. However, we can apply ADAM method, studied in [76] for the parameter estimation. It is based on a stochastic iteration method via the gradient of the MSE function. The key to evaluate the gradient of the MSE function is to evaluate the dynamics of
. It satisfies the following system of SDEJ
(42)
The existence and uniqueness of the solution to the SDEJ system (42) can be obtained with necessary regularity conditions on the coefficients.
B.2. Option Surface Fitting
There is a strand of literature that strives to fit option panels using different dynamics for the underlying assets, for example, [77] on stochastic volatility models, [78] on local volatility models and [79] on local-stochastic volatility models. Models that incorporate jumps can be found in [80], [81] and references therein.
Consider the following stochastic differential equation
(43)
Here we model
by a DNN. The advantage of doing so is that it might fully capture the market volatility surface meantime ensuring a good dynamic fit, while still preserving the existence and uniqueness result for the related stochastic differential equation system (43).
B.3. Credit Risk Management: Evaluation on a Future Simulation Grid
We refer the problem definition to [6]. It is easy to illustrate that the problem is equivalent to the evaluation of conditional expectations on a future simulation grid and our methods are suitable for this type of problems. Note that, some XVA quantities, such as KVA, require the evaluation of CVA on a future simulation grid. Our methodologies, such as the ones proposed in Sections 2 and B.7, can be applied on the evaluation of KVA, once we obtain future present values of financial claims.
B.4. Dynamic Hedging
There are references that utilize machine learning (mainly Reinforcement Learning, or RL) to solve dynamic hedging problems, e.g., [82], [83] and [84]. However, here in this paper we will not follow this route. Instead, we use the BSDE formulation of the problem in [2] and try to solve the BSDE that characterizes the hedging problem. The methodology is outlined in Appendix B.11.
B.5. Dynamic Portfolio-Consumption Choice
We use [85] as an example and try to solve the related coupled FBSDE with jumps. The methodology is outlined in Appendix B.11. Other examples of dynamic portfolio optimization can be found in [53], [86], [87], [88], [89], [90], [91], [92], [93] and [94]. Essentially, dynamic portfolio-consumption choice problems are stochastic programming in nature and can be related to HJB equations or BSDEs. An example of using HJB representation of the problem can be found in [95]. The equations can be solved using the methodologies outlined in Section 0 and Appendix B.11.
B.6. Transition Density Approximation
We can generalize the theory in [96] and [97] to approximate the transition density of a multivariate time-inhomogeneous stochastic differential equation with jumps. According to [96] and [97], the transition density of a multivariate time-inhomogeneous stochastic differential equation with or without jumps can be approximated by polynomials in a weighted-Hilbert space. See ( [97], Equation (2.1)), for example. The key is to evaluate the coefficients
, which is, again, the evaluation of conditional expectations. The resulted transition density can be used in option pricing, MLE estimation for MSDEJs and prediction, filtering and smoothing problems for hidden Markov models, see [98].
B.7. Evaluating Conditional Expectations via a Measure Change
Consider the following equation
(44)
(45)
where
is the transition density of a stochastic differential equation with jumps, which can be simulated for arbitrary
without using time discretization7 and
is the transition density function of X.
can be approximated by the method outlined in Appendix B.6. It is immediately obvious that we can generate random numbers from
and reuse them for the evaluation of the conditional expectation on the left hand side of Equation (44) for different
.
B.8. Empirical Asset Pricing with Factor Models: Evaluating Expected Returns
In this section, we propose to use machine learning, mainly, ANN techniques, to construct factor models and evaluate the conditional expected asset returns and risk-premium cross-sectionally. Related references are [28] and [74], among others. [3] provide a good example with basis function expansion to capture the non-linearity in asset returns. Specifically, consider the following lead-lag regression
(46)
Here
and X is a set of risk factors. Then,
. Linear factor models assume that
. f can also be approximated by basis function expansion, using universal approximation theorem, or via ANNs. The fitted conditional expected asset returns can be fed into the mean-variance optimizer, i.e., [99] and construct long-short portfolios or other trading strategies.
B.9. Recovery and Representation Theorem
In [100], the authors propose a model-free recovery theorem, based on a series expansion of higher order conditional moments of asset returns. Their work inspires us to exploit the ANN-factor models to represent the higher order conditional moments of the asset returns and therefore validating the recovery theorem proposed there-in. Moreover, similar to [57], our machine learning approximation to the conditional expectations of financial payoffs amounts to a compound option representation of arbitrary
-claims in the financial economic system. Also, the second numerical method means that any financial claim, can be locally approximated by a linear combination of power derivatives, following the same idea.
B.10. Theoretical Asset Pricing via Dynamic Stochastic General Equilibrium
Note that, the equation systems proposed in [101], [102] and [103] can be transformed into BSDEs and we can use time discretization and apply the techniques proposed in Section 2 and Appendix B.11 to solve them. In this paper, however, we will not test our methods on this strand of literature.
B.11. Solving High-Dimensional CFBSDEJs
A coupled forward-backward stochastic differential equation with jumps (CFBSDEJ) can be written as
(47)
where
is a compensated Poisson random measure. We take the following steps to solve Equation (47) numerically.
Time Discretization
Discretize time interval
into n-equal distance sub-intervals
with
,
and
. Consider the following Euler discretized equation.
(48)
where
and
. Denote the solution to the time-discretized CFBSDEJ as
. We need the following assumption.
Assumption 25. Under the norm
introduced in [104], we have
(49)
as
.
Mollification
Define a sequence of functions
, which are bounded and have bounded derivatives of all orders and
(50)
in a point-wise sense. Also denote the solution to the CFBSDEJ with coefficients
as
. Then, we have the following theorem.
Theorem 26. Under Assumption 25
(51)
as
for arbitrary
. g is a function with at most polynomial growth in its arguments.
Picard Iteration
After the time discretization and mollification are done, we will resort to Picard fixed point iteration technique to decompose the solution
to a sequence of uncoupled FBSDEJs whose solutions are denoted by
, where k denotes the index of Picard iteration. For zeroth order, consider
(52)
For
, define
(53)
Evaluation of Conditional Expectations
For Equation system (53), we can start from the last time interval and work backwards. The problem is transformed into the evaluation of
, where u is the intermediate solution and satisfies
.
B.12. Pricing Kernel Approximation
A pricing kernel
is an
stochastic process, adapted to the information filtration
, such that
(54)
where
is an
payoff,
and
. It is obvious that
, i.e.,
is a
-martingale. Represent
where
is a set of orthonormal basis in
space and
is the vector of coefficients of
. Suppose that we have K derivative contracts, denoted by
, with basis representation
. Therefore
(55)
Equation (55), if truncated after J terms, formulates a linear equation system and the unknowns
and
can be recovered from ordinary least square optimization. After we obtain
,
can be recovered by
, via the methodology outlined in Section 2.
Remark 27. If
is not orthonormal, Equation (55) becomes nonlinear in
. The evaluations remain the same, with only more complicated numerical computations. The basis can also be represented by ANNs.
Remark 28. For a specific representation via universal approximation theorem, see [55].
Remark 29. It is possible to allow shape constraints in the estimation (55) and formulate a constrained optimization problem, see [105], for example.
We can also directly utilize the method proposed in Section 2, when used with time discretization and Monte Carlo simulation. Denote M as the number of
sample paths and
as M simulated final payoffs for each of the K derivatives. Define
as M real numbers. Let
be K derivative prices at time
. Find the solution to the following optimization problem
(56)
After obtaining
, we try to find function relation g such that
where
is a set of simulated state variables at time T. When fitting g, we can add some shape or no-arbitrage constraints, or other regularization conditions, to the optimization problem and formulate a constrained ANN
(ACNN). We always assume that the matrix
is a
invertible matrix, where
is the matrix transpose operator.
C. Intuition of Convergence Proof for Appendix B.11
In Appendix B.11, we propose a method to solve numerically a CFBSDEJ. As long as the time discretization step is convergent, we can argue that the methodology converges, in some sense, to the true one, as outlined above in Appendix B.11. Potentially, we need an a priori estimate formula, similar to the one in [2], for coupled BSDEs, to justify Picard iteration at every time discretization step.
NOTES
1It is obvious that
can be the basis or frame of
. However, we do not assume so in this paper.
2It is the linear space spanned by the set
.
3We should understand that distance can be defined in function space
.
4K can be positive infinity, i.e.,
.
5We will only show convergence of Methods 1 and 2.
6Here we only consider the case where
for any
. The case with
is analogous.
7For example, a Lévy process.