Theoretical and empirical finance research involves the evaluation of conditional expectations, which, in a continuous time jump-diffusion setting, can be related to second order partial integral differential equations of parabolic type (PIDEs) by the Feynman-Kac theorem, and other types of equations such as backward stochastic differential equations with jumps (BSDEJs) or quasi-linear PIDEs in more complicated settings. In theoretical continuous-time finance, many problems, such as asset pricing with market frictions, dynamic hedging or dynamic portfolio-consumption choice problems, can be related to Hamilton-Jacobi-Bellman (HJB) equations via dynamic programming techniques. The HJB equations, from another perspective, are equivalent to BSDEs derived from a probabilistic approach. The nonlinear BSDEs, studied in , can be decomposed into a sequence of linear equations, which can be solved by taking conditional expectations, via Picard iteration. For empirical studies, the focus of the literature has been the evaluation of the cross sectional conditional risk-adjusted expected returns and the explanation of them using factors. See   and  as good illustrations. It is easily seen that, regardless of the fact whether the underlying models are continuous-time or discrete-time, evaluating conditional expectations is inevitable in finance literature. Moreover, in order to perform XVA computations for the measurement of counterparty credit risk, we need to evaluate the conditional expectations, i.e., the derivative prices, on a future simulation grid, as outlined in . These facts call for efficient methods to compute the quantities aforementioned.
In this paper, we extend the basis function expansion approach proposed in  with machine learning techniques. Specifically, we propose new efficient methods to evaluate conditional expectations, regardless of the dynamics of the underlying stochastic process, as long as they can be simulated. Rigorous convergence proofs are given using Hilbert space theory. The methodologies can be applied to time zero pricing as well as pricing on a future simulation grid, with the advantage of ANN approximation most prominent in high dimensional problems. In the sequel, we show applications of our methodologies on the pricing of European derivatives and extension to contracts with optimal stopping feature is straightforward through either  approach or reflected-BSDEs.
Compared to the literature on traditional stochastic analysis, our methodologies are able to handle large data sets and high-dimensional problems, therefore suffering much less from the curse of dimensionality due to the nature of ANN methods. Moreover, our methodologies are very efficient when evaluating solutions of BSDEJs and PIDEs on a future simulation grid, where none of the traditional methodologies applies. With respect to recent machine learning literature on numerical solutions to BSDEs and PDEs, our methodologies enjoy the theoretical advantage of being able to handle equations with jump-diffusion and convergence results are provided. When applied to the solutions of BSDEJs and PIDEs, our methodologies require much less number of parameters, as compared to the current machine learning based methods to be mentioned below. At any step in the solution process, only one ANN is needed and we do not require nested optimization. In terms of application, not all the prices of OTC derivatives can be easily translated into BSDEJs and PIDEs, for example, a range accrual with both American and barrier (knock-out, for example) feature. However, our methodologies are naturally suitable in those situations. To conclude, our methods enjoy many theoretical and empirical advantages, which makes them attractive and novel.
There has been a huge literature on applications of machine learning techniques to financial research. Classical applications focus on the prediction of market variables such as equity indexes or FX rates and the detection of market anomalies, for example,  and . Option pricing via a brute-force curving fitting by ANNs dates back to . More applications of machine learning in finance, especially option pricing prediction, are surveyed in . See references therein. Pricing of American options in high dimensions can be found in , which is closest to our method 1. However, there are several improvements of our methods compared to this reference. First of all, we enable deep neural network (DNN) approximation and show convergence. Second, we can incorporate constraints in DNN approximation estimation and prove the mathematical validity of this approach. Third, we propose two more efficient methods to complement the first method of ours. Our treatment of constraints in the estimation of DNNs extends the work of  in that we can deal with a larger class of constraints by specifying a general Hilbert subspace as the constrained set. Risk measure computation using machine learning can be found in . Applications of machine learning function approximation on financial econometrics can be found in , ,  and . Recent applications include empirical and theoretical asset pricing, reinforcement learning and Q-learning in solving dynamic programming problems such as optimal investment-consumption choice, option pricing and optimal trading strategies construction, e.g., , , , , , , , , , ,  and references therein. Numerical methods to solve PDEs and BSDEs or the related inverse problems can be found in , , , , , , , , ,  and . Machine learning based methods enjoy the advantage of being fast, able to handle large data sets and high dimensional problems.
Our methodologies are combinations of traditional statistical learning theory and stochastic analysis with advanced machine learning techniques, introducing powerful function approximation method via the universal approximation theorem and artificial neural networks (ANNs), while preserving the regression-type analysis documented in . The methods are very easy to use, effective, accurate as illustrated by numerical experiments and time efficient. They are different from the convergent expansion method, e.g., , simulation methods such as , ,  and  or the asymptotic expansion method proposed by , , ,     , in that we no longer resort to polynomial basis function expansion or small-diffusion type analysis. Our methods are also different from the pure machine learning based ones documented in , , , , ,  and , in that we utilize the lead-lag regression formula to evaluate the conditional expectations, preserving the time dependent structure and our methods are able to handle jump-diffusion processes easily.
The organization of this paper is as follows. Section 2 documents the methodologies. Section 3 illustrates the usefulness of our methods by considering European and American derivatives pricing. Section 4 considers numerical experiments and Section 5 concludes. An outline of the proofs and other applications can be found in the appendices.
2. The Methodology
We use a Markov process modeled by a jump-diffusion as illustration. Suppose that we have a stochastic differential equation with jumps
where , is a standard d-dimensional Brownian motion and is a q-dimensional compensated Poisson random measure, with the compensator . Information filtration is generated by . We hope to evaluate the conditional expectation for any , e.g., see . Assumptions on and X are stated below.
Assumption 1 (On Growth Condition of ψ). has polynomial growth in its argument x, i.e., there exists a positive integer P, independent of x, such that for all , we have, for constant C independent of x
The following assumption is w.r.t. X.
Assumption 2 (On X). There exists a unique strong solution to Equation (1) and X has finite polynomial moments of all orders.
The General Approximation Theory
First, we need the following assumptions, definitions and results. Please note that, some of the spaces we introduce are actually conditional ones. The discussions of conditional Hilbert spaces can be found in , e.g., is a conditional Hilbert space for all .
Definition 3 (Projection Operator). For Hilbert spaces and , where . Define as the projection of onto .
Definition 4 (Orthogonal Space). For Hilbert spaces and , where . Define as the orthogonal space of in .
Definition 5 (Spanning the Hilbert Space). Assume that is a set of elements in Hilbert space and is an index set. Define as the intersection of all Hilbert subspaces of containing .
Assumption 6 (On Joint Continuity). and are two Hilbert spaces and . Moreover, is a sequence of Hilbert sub-spaces of satisfying for any and . We have for any and .
The next two theorems are well-known in the literature.
Theorem 7 (Hilbert Projection Theorem). Let be two Hilbert spaces and let . Then, exists and is unique. Moreover, it is characterized uniquely by .
Theorem 8 (Repeated Projection Theorem). Let be three Hilbert spaces. Then, for any , .
Remark 9 The conditions of Theorems 7 and 8 on and can be relaxed to convexity and completeness instead of Hilbert sub-spaces.
Finally, we have the result below.
Theorem 10. Suppose is a Hilbert space, and are Hilbert subspaces of satisfying and . , define and . Then we have w.r.t. the norm topology in , if Assumption 6 is satisfied.
Sometimes we need to add constraints on the calibrated ANN, e.g., the shape constraints. The following assumption and theorem deal with this situation.
Assumption 11 (On Constrained Sub-space). Suppose that such that is a sequence of non-empty convex and complete subspaces of satisfying Assumption 6, where and are described.
The following theorem handles the constrained approximation and its convergence.
Theorem 12 (On Constrained Approximation). Under Assumptions 6 and 11, for , if , then, we have .
Remark 13 (On ψ). In Theorem 12, the set represents prior knowledge on constraints that h satisfies. It can be represented by a set of non-linear inequalities or equalities on functionals of h. Common constraints for option pricing include non-negativity constraint and the positiveness constraint on the second order derivatives. The verification of satisfying Assumption 6 should be based on a case-by-case manner.
To proceed further, we need the following assumptions.
Assumption 14 (On Some Spaces). is an increasing sequence of Hilbert sub-spaces of , , . Moreover, .
Assumption 15 (On Structure of ). is a set of elements of , such that , where for any and , satisfies Assumption 141.
Then, we have the following results.
Lemma 1. For any adapted stochastic process such that , if , we have
The following proposition is a natural extension of Lemma 1.
Proposition 16. For any measurable function and stochastic process X such that and , we have
Here and the above minimization problem has a unique solution. In particular, if X is a Markov process, then , i.e., is a function of time t and .
We then have the following theorem.
Theorem 17. Under Assumptions 1, 2, 6, 14 and 15, for any adapted stochastic process such that and , we have
Further, for any measurable function and stochastic process X such that and , we have the following equality
If X is Markov, then we have , i.e., is a function of time t and .
The following theorem justifies the Monte Carlo approximation of expectation in the above optimization problems.
Theorem 18 (On Sequential Convergence). Under Assumptions 1, 2, 6, 14 and 15, suppose that for all , and are M i.i.d. copies of and . Then we have
The following results justify the universal approximation and ANN approximation approaches proposed in this paper.
Proposition 19 (On Universal Approximation Theory). Let denote the function in the universal approximation theorem mentioned in ,  and . Define , where X satisfies Equation (1) and Assumption 2, and have at most n significant digits in total, where , i.e., n belongs to the set of natural numbers, j runs from 1 to and
is the number of all related , i.e., . Then, satisfies Assumptions 6, 14 and 15. Therefore, Theorems 17 and 18 apply.
Proposition 20 (On Deep Neural Network Approximation). For the DNN defined in ( , Definition 1.1], observe that . Define
where satisfies that , have at most n total significant digits and . Then, , where 1 means
function for all x, satisfies Assumptions 6, 14 and 15. Therefore, Theorems 17 and 18 apply after a localization argument on and X on a compact sub-domain in .
Remark 21 (On DNN). Please note that, in Proposition 20, we do not intend to prove the convergence when the number of layers goes to infinity. Instead, we show convergence when the number of connections goes to infinity, which can be achieved via enlarging the number of neurons in each layer with the total number of layers remaining fixed.
Remark 22 (On Euler Time Discretization).  proposes an exact simulation method for multi-dimensional stochastic differential equations. The discussion of discretization error, of the regression approach proposed in this paper, with Euler method is not hard if satisfies Assumption 1, in which case the dominated convergence theorem and convergence of Euler method can be applied to show the convergence.
The proofs of the above results can be found in Appendix A. In what follows, we will propose three methods to compute, approximately, the function in Proposition 16.
In general, , defined in Proposition 16 and Theorem 17, can not be found in closed-form. A natural thought would be to resort to function expansion representations, i.e., to find the solution to the following problem
where is an appropriate space for coefficients and is a set of functions, with 2 dense in an appropriate function space 3. To further proceed, we seek a truncation of the function representation formula as follows
for J sufficiently large, where is a compact set in the Euclidean space where take values. The last step would be to use Monte Carlo simulation to approximate the unconditional expectation appearing in Equations (9) and (10). Therefore turning the conditional expectation computation problem, into a least-square function regression problem, similar to . An obvious choice of is polynomial basis, for example, the set of Fourier-Hermite basis functions. For expansion using Fourier-Hermite basis functions in high dimensions, see .
In fact, Artificial Neural Networks (ANNs) prove to be an efficient and convergent function approximation tool that we can utilize in the above expressions. Write
where denotes an ANN with parameters .
Note that, via proper time discretization and fixed point iteration, solving a BSDE with jumps can be decomposed into a series of evaluations of conditional expectations. The machine learning based method outlined above can be applied there. We will write down the algorithm to solve a general Coupled Forward-Backward Stochastic Differential Equation with Jumps (CFBSDEJs) in the appendix. Extensions to other types of BSDEJs are possible.
Here we assume that X is a Markov process. To handle path dependency or non-Markov processes, we can apply the backward induction method outlined in . With the machine learning approach, it is easy to see that this method enables us to get the values of conditional expectations on a future simulation grid.
Another method to utilize the idea of  is inspired by the boosting random tree method (BRT), see, , for example. Partition the domain space 4, where is a set of disjoint sets in and consider
The choice of is important and we can use the machine learning classification techniques (or any classification rule), such as kmeans function in R programming language, in Monte Carlo simulation and related computations. Denote . It is possible to show that as long as , we only need finite number of functions, for example, , to approximate each and obtain convergence. In practice, although the domain of is , it might be centered at a small subspace , therefore facilitating the partition process. Note also that this method might require us to mollify the function , if it is not smooth. We adopt finite order Taylor expansion as the function expansion representation approach. The following theorems provide convergence analysis for this method.
Theorem 23. For an appropriate function space , we have
Theorem 24. Let be as described previously and . Then, we have
with J large enough, fixed, finite and is an approximation to , which satisfies
for any , , and is independent of k when K is sufficiently large.
Next, we propose an algorithm combining the ANN and universal approximation theorem (UAT). Suppose that is the space where we are performing the approximation. Also assume that , i.e., the information filtration is equivalently generated by X. Define an ANN with connection N by , where x is the state variables that the ANN depends on, is the vector of parameters and j is its label. We define the following nested regression approximation
where is the approximate sequence of .
In this paper, we will test and compare the performance of all of the proposed methods. A general discussion and rigorous proofs can be found in Appendix A5.
3. Applications in Derivatives Pricing
3.1. European Option Pricing
Suppose that the payoff of a European claim can be written as, similar to  and , , where is a stream of cash flows materialized at each time instance t and is a one-time terminal payoff at time T. Therefore, under no-arbitrage condition, the price of this European payoff can be written as, under risk neutral measure
where is the stochastic discount factor. If we assume a Markov structure and , then , i.e., is a function of time t and state vector . This problem is a canonical application of the evaluation of conditional expectations and we can apply the methodologies outlined in Section 2 to solve it. European claims with barrier features can be incorporated and priced in a similar way. For example, the price of a knock-in European claim can be written as
where , where . In our setting, the dynamics of X can be arbitrary, possibly stochastic differential equations with jumps, Markov chains, or even non-Markov processes. Previously, Monte Carlo based method for option pricing can be found in  and , among others.
3.2. American Option Pricing
Still use to denote the payoff structure of an American claim, whose price can be obtained via formula
Here is the space of all the stopping times in . We refer the interested readers to  and  for general derivation and explanation of Equation (24). It is also possible to derive the general BSDE that an American claim price satisfies, for example . Moreover, in  and , the authors utilize a backward induction approach to solve optimal stopping problems. The idea can be carried out using the methodologies documented in Section 2. American claims with barrier features can be incorporated and priced in a similar way. It is also known that American option prices can be related to reflected BSDEs (RBSDEs), a rigorous discussion of existence and uniqueness of such equations can be found in  and references therein.
4. Numerical Experiments
4.1. European Option Pricing
In this section, we consider a Heston model
where is a two dimensional standard Brownian motion. The parameter values are chosen as , , , , , , and . Time to maturity is set to be ,
with time discretization step and . The number of
simulation paths is . We price a plain vanilla European call option as an illustration. The QQ-plots are displayed in Figures 1-10. The first three correspond to a recursive evaluation, i.e., regressing the values at on state variables at time t. The rest of the plots correspond to direct regression, i.e., regressing the discounted payoffs at time T on state variables at time t. Figures 10-12 are for the prices of a digital call option under Black-Scholes setting and Figures 13-15 are QQ-plots for Delta values. Figure 16 and Figure 17 show the QQ-plots for method 3 under Heston model with 3 nested ANN approximations of size 4 and one ANN approximation of size 12 using R routine nnet. The absolute RMSE for the former is 0.1938% and latter 0.2581%, with the running time 10.36 seconds compared to 52.31 for ANN approximation with size 12.
Figure 1. QQ-plot for Method 1, and relative pricing error is 1.20%.
Figure 2. QQ-plot for Method 1, and relative pricing error is 1.50%.
Figure 3. QQ-plot for Method 1, and relative pricing error is 1.20%.
Figure 4. QQ-plot for Method 1, and relative pricing error is 1.66%.
Figure 5. QQ-plot for Method 1, and relative pricing error is 1.75%.
Figure 6. QQ-plot for Method 1, and relative pricing error is 3.00%.
Figure 7. QQ-plot for Method 2, and relative pricing error is 1.80%.
Figure 8. QQ-plot for Method 2, and relative pricing error is 3.50%.
Figure 9. QQ-plot for Method 2, and relative pricing error is 3.53%.
Figure 10. QQ-plot for Method 1, and relative pricing error is 0.40%.
Figure 11. QQ-plot for Method 1, and relative pricing error is 0.80%.
Figure 12. QQ-plot for Method 1, and relative pricing error is 0.60%.
Figure 13. Delta QQ-plot for Method 1, .
Figure 14. Delta QQ-plot for Method 1, .
Figure 15. Delta QQ-plot for Method 1, .
Figure 16. Price QQ-plot for Method 3, .
Figure 17. Price QQ-plot for Method 3, .
4.2. American Option Pricing
Here we refer the readers to  for the BSDE satisfied by a plain vanilla American option. For , , , , , and , the benchmark American option price at is 9.0660 and the relative difference of our Monte-Carlo price is 0.27%. The running time is less than 30 seconds.
5. Conclusion and Future Research
In this paper, we show how machine learning techniques, specifically, ANN function approximation methods, can be applied to derivatives pricing. We relate pricing problems to the evaluation of conditional expectations via BSDEJs and PIDEs. Future research topics can, potentially, be the development of reinforcement learning methodologies to solve dynamic programming problems and apply them in the context of empirical asset pricing literature. Moreover, the evaluation of energy derivatives calls for SDEJs defined in a Hilbert space. The same theoretical constructions can also be found in the evaluation of fixed income derivatives, such as the random field models proposed and studied in . One can, of course, apply Karhunen-Loéve expansion for a dimension reduction to reduce the problem to the evaluation of conditional expectations of regular SDEJs. However, the development of machine learning based methods to solve directly the conditional expectations on the stochastic processes defined in a Hilbert space is important. In addition, stochastic differential games, that arise in the context of American game options, equity swaps, and the related Mckean-Vlasov type FBSDEJs (mean-field FBSDEJ, see  ) are important topics in mathematical finance. They are also related to the theoretical analysis of high-frequency trading. Finding machine-learning based numerical methods to solve these equations is of great interest to us. Last, but not least, machine learning methods in asset pricing and portfolio optimization, which can be found in , , , ,  and , admit an elegant way to price financial derivatives under -measure. For example, we can use the method in  to calibrate the SDF process and use  to generate market scenarios. These methodologies, combined with the methods documented in this paper and , have the potential to solve for any derivative price. We leave all the development to future research.
We thank the Editor and the referee for their comments. Moreover, we are grateful to Professor Jérôme Detemple, Professor Marcel Rindisbacher and Professor Weidong Tian for their useful suggestions.
A. Convergence of the Proposed Methodologies
Proof of Theorem 10. It is known from the projection theorem of Hilbert space that and h actually exist and are unique. Moreover, as indicated by the repeated projection theorem. It is also known that . As we ask that Assumption 6 hold, we know that as .
Proof of Theorem 12. The proof follows from Assumption 6 and Theorem 8. We have
This concludes the proof.
Proof of Lemma 1. For any , we have
Therefore we have the claim announced.
Proof of Theorem 17. The proof of this theorem follows from Assumptions 1, 2, 6, 14, 15 and Theorem 10, by choosing .
Proof of Theorem 18. Essentially, Equation (7) is the result of Gauss-Markov Theorem and the consistency property of OLS estimator.
Proof of Proposition 19. This is a direct consequence of the discussion in ( , Section 3) (see Equation (5)) and Theorem 10. To elaborate, consider , , its projections h and on and defined in this proposition. Suppose that
where and is a set of orthonormal basis in . From the repeated projection theorem, we know that for any 6 and . From the property of h, we know that . Therefore, as .
Proof of Proposition 20. This is a direct consequence of the discussion in ( , Theorem 2.2), localization arguments, Theorem 10 and the proof of Proposition 19.
Proof of Theorem 23. The first, second and third equality are obvious given an appropriate choice of depending on the Markov property of X and its moment conditions in Assumption 2. Actually, because of the existence and uniqueness of such that the RHS of the first equality achieves minimum, we know that
From another perspective, we know that is a piecewise minimization. Therefore
The last equality in Equation (13) holds.
Proof of Theorem 24. The proof of this theorem is a direct consequence of Equations (13), (15) and triangle inequality.
B. Other Applications
In this section, we document other applications of our methodologies in finance.
B.1. Joint Valuation and Calibration
Suppose that there are N derivatives contracts whose prices at time can be expressed as . Their payoffs are , where X is an
r-dimensional vector of state variables. Sometimes we write to explicitly state dependence of X on its vector of parameters . Here suppose satisfies a system of stochastic differential equations with jumps
The main idea is that might contain derivatives contracts from different asset classes or hybrid ones. Therefore, we need to model X as a joint high dimensional cross-asset system. One potential problem is that is in general a high-dimensional vector, which will be hard to estimate using usual optimization routines in R or MATLAB software system. However, we can apply ADAM method, studied in  for the parameter estimation. It is based on a stochastic iteration method via the gradient of the MSE function. The key to evaluate the gradient of the MSE function is to evaluate the dynamics of . It satisfies the following system of SDEJ
The existence and uniqueness of the solution to the SDEJ system (42) can be obtained with necessary regularity conditions on the coefficients.
B.2. Option Surface Fitting
There is a strand of literature that strives to fit option panels using different dynamics for the underlying assets, for example,  on stochastic volatility models,  on local volatility models and  on local-stochastic volatility models. Models that incorporate jumps can be found in ,  and references therein.
Consider the following stochastic differential equation
Here we model by a DNN. The advantage of doing so is that it might fully capture the market volatility surface meantime ensuring a good dynamic fit, while still preserving the existence and uniqueness result for the related stochastic differential equation system (43).
B.3. Credit Risk Management: Evaluation on a Future Simulation Grid
We refer the problem definition to . It is easy to illustrate that the problem is equivalent to the evaluation of conditional expectations on a future simulation grid and our methods are suitable for this type of problems. Note that, some XVA quantities, such as KVA, require the evaluation of CVA on a future simulation grid. Our methodologies, such as the ones proposed in Sections 2 and B.7, can be applied on the evaluation of KVA, once we obtain future present values of financial claims.
B.4. Dynamic Hedging
There are references that utilize machine learning (mainly Reinforcement Learning, or RL) to solve dynamic hedging problems, e.g., ,  and . However, here in this paper we will not follow this route. Instead, we use the BSDE formulation of the problem in  and try to solve the BSDE that characterizes the hedging problem. The methodology is outlined in Appendix B.11.
B.5. Dynamic Portfolio-Consumption Choice
We use  as an example and try to solve the related coupled FBSDE with jumps. The methodology is outlined in Appendix B.11. Other examples of dynamic portfolio optimization can be found in , , , , , , , ,  and . Essentially, dynamic portfolio-consumption choice problems are stochastic programming in nature and can be related to HJB equations or BSDEs. An example of using HJB representation of the problem can be found in . The equations can be solved using the methodologies outlined in Section 0 and Appendix B.11.
B.6. Transition Density Approximation
We can generalize the theory in  and  to approximate the transition density of a multivariate time-inhomogeneous stochastic differential equation with jumps. According to  and , the transition density of a multivariate time-inhomogeneous stochastic differential equation with or without jumps can be approximated by polynomials in a weighted-Hilbert space. See ( , Equation (2.1)), for example. The key is to evaluate the coefficients , which is, again, the evaluation of conditional expectations. The resulted transition density can be used in option pricing, MLE estimation for MSDEJs and prediction, filtering and smoothing problems for hidden Markov models, see .
B.7. Evaluating Conditional Expectations via a Measure Change
Consider the following equation
where is the transition density of a stochastic differential equation with jumps, which can be simulated for arbitrary without using time discretization7 and is the transition density function of X. can be approximated by the method outlined in Appendix B.6. It is immediately obvious that we can generate random numbers from and reuse them for the evaluation of the conditional expectation on the left hand side of Equation (44) for different .
B.8. Empirical Asset Pricing with Factor Models: Evaluating Expected Returns
In this section, we propose to use machine learning, mainly, ANN techniques, to construct factor models and evaluate the conditional expected asset returns and risk-premium cross-sectionally. Related references are  and , among others.  provide a good example with basis function expansion to capture the non-linearity in asset returns. Specifically, consider the following lead-lag regression
Here and X is a set of risk factors. Then, . Linear factor models assume that . f can also be approximated by basis function expansion, using universal approximation theorem, or via ANNs. The fitted conditional expected asset returns can be fed into the mean-variance optimizer, i.e.,  and construct long-short portfolios or other trading strategies.
B.9. Recovery and Representation Theorem
In , the authors propose a model-free recovery theorem, based on a series expansion of higher order conditional moments of asset returns. Their work inspires us to exploit the ANN-factor models to represent the higher order conditional moments of the asset returns and therefore validating the recovery theorem proposed there-in. Moreover, similar to , our machine learning approximation to the conditional expectations of financial payoffs amounts to a compound option representation of arbitrary -claims in the financial economic system. Also, the second numerical method means that any financial claim, can be locally approximated by a linear combination of power derivatives, following the same idea.
B.10. Theoretical Asset Pricing via Dynamic Stochastic General Equilibrium
Note that, the equation systems proposed in ,  and  can be transformed into BSDEs and we can use time discretization and apply the techniques proposed in Section 2 and Appendix B.11 to solve them. In this paper, however, we will not test our methods on this strand of literature.
B.11. Solving High-Dimensional CFBSDEJs
A coupled forward-backward stochastic differential equation with jumps (CFBSDEJ) can be written as
where is a compensated Poisson random measure. We take the following steps to solve Equation (47) numerically.
Discretize time interval into n-equal distance sub-intervals with , and . Consider the following Euler discretized equation.
where and . Denote the solution to the time-discretized CFBSDEJ as . We need the following assumption.
Assumption 25. Under the norm introduced in , we have
Define a sequence of functions , which are bounded and have bounded derivatives of all orders and
in a point-wise sense. Also denote the solution to the CFBSDEJ with coefficients as . Then, we have the following theorem.
Theorem 26. Under Assumption 25
as for arbitrary . g is a function with at most polynomial growth in its arguments.
After the time discretization and mollification are done, we will resort to Picard fixed point iteration technique to decompose the solution to a sequence of uncoupled FBSDEJs whose solutions are denoted by , where k denotes the index of Picard iteration. For zeroth order, consider
For , define
Evaluation of Conditional Expectations
For Equation system (53), we can start from the last time interval and work backwards. The problem is transformed into the evaluation of , where u is the intermediate solution and satisfies .
B.12. Pricing Kernel Approximation
A pricing kernel is an stochastic process, adapted to the information filtration , such that
where is an payoff, and . It is obvious that , i.e., is a -martingale. Represent
where is a set of orthonormal basis in space and is the vector of coefficients of . Suppose that we have K derivative contracts, denoted by , with basis representation . Therefore
Equation (55), if truncated after J terms, formulates a linear equation system and the unknowns and can be recovered from ordinary least square optimization. After we obtain , can be recovered by , via the methodology outlined in Section 2.
Remark 27. If is not orthonormal, Equation (55) becomes nonlinear in . The evaluations remain the same, with only more complicated numerical computations. The basis can also be represented by ANNs.
Remark 28. For a specific representation via universal approximation theorem, see .
Remark 29. It is possible to allow shape constraints in the estimation (55) and formulate a constrained optimization problem, see , for example.
We can also directly utilize the method proposed in Section 2, when used with time discretization and Monte Carlo simulation. Denote M as the number of
sample paths and as M simulated final payoffs for each of the K derivatives. Define as M real numbers. Let be K derivative prices at time . Find the solution to the following optimization problem
After obtaining , we try to find function relation g such that
where is a set of simulated state variables at time T. When fitting g, we can add some shape or no-arbitrage constraints, or other regularization conditions, to the optimization problem and formulate a constrained ANN
(ACNN). We always assume that the matrix is a invertible matrix, where is the matrix transpose operator.
C. Intuition of Convergence Proof for Appendix B.11
In Appendix B.11, we propose a method to solve numerically a CFBSDEJ. As long as the time discretization step is convergent, we can argue that the methodology converges, in some sense, to the true one, as outlined above in Appendix B.11. Potentially, we need an a priori estimate formula, similar to the one in , for coupled BSDEs, to justify Picard iteration at every time discretization step.
1It is obvious that can be the basis or frame of . However, we do not assume so in this paper.
2It is the linear space spanned by the set .
3We should understand that distance can be defined in function space .
4K can be positive infinity, i.e., .
5We will only show convergence of Methods 1 and 2.
6Here we only consider the case where for any . The case with is analogous.
7For example, a Lévy process.