Modelling in Computationally Intense Field Of Finance
Transcript of Modelling in Computationally Intense Field Of Finance
Modelling in Computationally Intense Field Of Finance
Prof Dr A S Chaudhuri, Former Fellow, IIAS Shimla
Professor of Electronics and Communication Engineering
Dr B C Roy Engineering College, Durgapur West Bengal
and Dr Ranjan Chaudhuri, Associate Professor (Marketing)
National Institute of Industrial Engineering (NITIE)
Vihar Lake, PO NITIE Mumbai 400087 India
Abstract
The specification of a model almost exclusively involves purely economic considerations. The model may
be used as an aid in economic analysis, policy simulation, or policy optimization, but each case imposes
special demands on the specification. The result of such considerations generally determines the overal1
size of the model, the number of inputs and outputs, and the definition of these variables. In addition, the
outputs of the model are usually decomposed into two types: the endogenous variables which are outputs of
dynamic equations, called behavioral equations and variables which are outputs of non-stochastic
equations, called definitional equations. A choice must be made as to the use of variables in current price
(inflated) or constant price (deflated). The economic specification stage can be summarized as one in which
the following information is determined:
1) The specific purpose of the model, thereby fixing the overall size; and hence, an enumeration of all the
outputs and their type and an enumeration of all the inputs and their type.
2) The output definitions; whether it is explained by a behavioral equation together with all its explanations
(inputs to the equation), or, whether it is determined by a definitional identity.
The second stage is the most challenging of the two. This stage combines the use of a priori economic
information, hypothesis testing techniques, and cross-correlation analysis from the black box approach. In
econometric terminology, the word “structure” denotes a complete definition of the functional or stochastic
relationships between all of the endogenous and exogenous variables. The specific meaning of structure can
be ascertained by examining each equation of the structural form. Before accepting the results of any
estimation, they must be tested for their adequacy. The auto- and cross-correlation functions for the model
residuals constitute an important diagnostic check. The last diagnostic to be employed is perhaps the most
important,, namely, the model’s forecasting performance. After having successfully met the other
2
diagnostic checks a model is not accepted until it has demonstrated its ability to forecast. Forecasts are then
made with each model from the end of its sample period up to the present, using the (historical) observed
inputs over this period. Thus, forecasts are obtained outside of the sample period . Such simulations more
closely approach reality and serve as a good guide in judging the model’s adequacy in forecasting the
unknown future. This gives additional insight into the time-invariance of the model structure. The modeling
procedure described in this paper was designed to incorporate three concepts. First, employment is made of
all available a priori information provided by thus eliminating beforehand the possibility of expending
effort on fruitless searches for non-existent relationships (interconnections), Second, the basic philosophy
of the “black box” approach is then applied allowing the data to decide the exact dynamic structure. Thus,
overly complex (statistically unsubstantiated) structures are automatically eliminated. Third, diagnostics are
continually employed which are designed to both reveal inadequacies and indicate how improvements can
be made.
Introduction
Financial engineering, the most computationally intense field of finance, has only come to be recognized as
a formal profession over the last four or five years. During that time, the International Association of
Financial Engineers (IAFE) has been instrumental in helping to define the profession and in promoting the
creation of financial engineering research programs. Technological sciences recently reveals the revolution
in financial services. For more than a half-century statistics and technical analysis have been the
technologies of choice for financial analysts. However, it was not until the introduction of the Hamiltonian-
Jacobi-Bellman and Black-Scholes differential equation in the mid-70’s that more advanced forms of
mathematics were used in the field of finance. Since that time there has been a tremendous expansion in
the application of mathematics and other engineering technologies to the field of finance. The use of
mathematics has been spurred by the availability of low-cost computers and supercomputers. In addition
there has been an luminous gathering of engineers, mathematicians, and physicists into the finance and
investments industry. Many of these professionals are excited about the opportunities to apply their
quantitative and analytical skills in new ways. And finally there is the explosive growth in the financial
markets in volume, numbers and types of securities offered, and their internationalization under GATT and
3
WTO regimes which has required the intensive use of engineering technologies for development of highly
skillful computerized software.
The future of the continuing convergence of the fields of finance and engineering into a new era of
computational intelligence and its application in the financial services arena has been envisaged. There are
neural nets that emulate the Black-Scholes equation with better out-of-sample performance than the
original equation. Commercially available software using genetic algorithms provide superior results for
portfolio asset allocation. Moreover today’s computational intelligence methodologies can forecast the
inflation rate to 95 % accuracy. It is the mission of the present monograph to be a forum for new
technologies and applications in the intersection of computational intelligence and financial engineering.
The interest in application of the ingredients of computational intelligence to finance has been growing
steadily.
A few years ago, the term "financial engineering" did not exist in common usage; it is only recently that
the various models and algorithms now regarded collectively as financial engineering have coalesced into a
well-defined discipline. Three factors are largely responsible for this mighty achievement. The first is the
simple fact that the financial system is becoming more complex over time. This is an obvious consequence
of general economic growth and development in which the number of market participants, the variety of
financial transactions, and the sums involved also grow. As the financial system becomes more complex,
the benefits of more highly developed financial technology become greater and greater and, ultimately,
indispensable. The second factor is the set of professional developments in the quantitative modelling of
financial markets, e.g., financial technology, pioneered over the past three decades by the giants of
financial professionals: Black, Cox, Lintner, Markowitz, Merton, Modigliani, Miller, Ross, Samuelson,
Scholes, Sharpe, and others. Their contributions laid the remarkably durable foundations on which all of
modern quantitative financial analysis is built. The third factor is an almost parallel set of technological
breakthroughs in computer technology, including hardware, software, and data collection and organization.
Without these breakthroughs, much of the financial technology developed over the past thirty years would
be irrelevant academic observations. Precisely, Professor Merton’s work in the continuous time
mathematics of finance serves as one of the cornerstones on which the profession of financial engineering
is built. Professor Merton redefined modern financial economics definitively, and helped to launch a multi-
4
trillion-dollar industry that is still enjoying double-digit growth. While the methods and techniques range
broadly, from optimal control and filtering to neural networks to nonparametric regression, the topics are
all sharply focused on financial applications. In this way, financial engineering is following a path not
unlike those of the engineering disciplines in their formative stages: applications tend to drive the
technology, yet research and development are characterized by an intellectual entrepreneurialism that cuts
across many different methodologies. No one has illustrated this entrepreneurialism more eloquently than
Harry Markowitz, the father of modern portfolio theory and a winner of the 1990 Nobel Prize in
economics.
In the past, Wall Street and the City in London employed academics to help them develop innovative financial products
and to carry out quantitative modelling work. Besides academics, scientists and engineers were also employed by
financial institutions and trained in-house as financial engineers. In view of this evident need for engineers and
scientists to work and be trained as financial engineers, universities nowadays provide a formal educational route for
acquiring training in financial engineering. Most graduate programmes are being offered by engineering schools and
mathematics departments, though some are being offered by business schools. The understanding and mitigation of risk
are increasingly important aspects of modern finance. Running on the speed and capacity of the Internet and other
information technologies, the world’s financial markets have grown dramatically in size, interconnectedness and
complexity. With more opportunities available, investors are more willing to accept the risks associated with
entrepreneurial ventures to create new financial products and services. But the combination of technological innovation
and globalisation that created the new economy also brings new sources of risk to financial markets.
Financial engineering is defined by the International Association for Financial Engineering (IAFE), a professional body
for financial engineers based in the USA, as: ‘The development and creative application of financial technology to
solve financial problems and exploit financial opportunities. Further, financial engineering involves the design, the
development and the implementation of innovative financial instruments and processes and the formulation of creative
solutions to problems in finance.”
Financial engineering is a multidisciplinary field that combines finance, mathematics and statistics, engineering and
computer science. The related areas in these fields are financial markets, mathematical finance and econophysics and
computational finance. Quantitative and analytical principles and methods are critical in understanding financial
instruments in financial markets today, so that the new discipline of financial engineering has evolved. Financial
engineering requires a composite of skills. For example, the methodology of science and mathematics has been used in
financial engineering areas such as derivative pricing and hedging, portfolio management and risk management.
5
Stochastic calculus helps financial engineers to price exotic options and to hedge the risks. Advances in Monte Carlo
simulation have been applied to risk management. For selecting optimal portfolios, optimisation techniques have been
applied to the asset and liability management problem. The neural networks and genetic algorithms used in engineering
and physical science fields have been applied to forecasting futures prices for trading and investment. A financial
engineer typically works in areas that require a high level of quantitative skills and innovative ideas in structured
finance, derivative trading, risk management, portfolio investments, corporate financing, financial and insurance
products and financial information technology.
The domain knowledge includes financial markets, financial products and processes, price and hedge modelling,
investment technology, risk analysis, computational methods, and data support systems for trading. When institutions
create desired pay-off patterns that manage risk for their clients or use options and futures to hedge the products they
sell, they are engaging in financial engineering.
A distinct professional category called financial engineers is needed by financial institutions to develop new financial
products, to customise and trade them, to monitor risk exposure to books of complex derivatives, to devise hedging
schemes and to search for arbitrage opportunities in the markets. Therefore, there is a need for a new degree
programme for the training of financial engineers. The range of fields in which financial engineers can establish
careers includes risk management, structured financial products, quantitative trading and research.
A wide range of businesses offer career opportunities for financial engineers, including:
• commercial and investment banks
• brokerage and investment firms
• insurance companies
• accounting and consultancy firms
• treasury departments of non-financial corporations
• public institutions such as federal government agencies, state and local governments, municipalities
and international organisations
• software and technology vendors that provide products and services to the financial industry.
Financial engineering is a fairly recent field of study. It combines technical and conceptual advances from various
disciplines to create a broad interdisciplinary body of knowledge, ready to meet the challenges of a rapidly growing
market and an exciting future. John O'Brien, Executive Director of the University of California Berkeley's MFE
programme believes that 'Financial engineering is taking its place, along with traditional engineering disciplines, as a
major driving force in the global economy". A key to the growth of financial engineering has been its maintenance of a
strong interdisciplinary tradition and its appeal across all quantitative fields. The rapid assimilation of new
6
methodology into practice is both a result of and reason for this interdisciplinarity. Interest from the multidisciplinary
fields in engineering science and finance has spurred the growth of degree programmes in financial engineering.
Financial engineering is a fast-growing dynamic field driven by the needs of industry and based on the rich interplay of
and input from many disciplines, including financial economics, probability, numerical methods, operations research,
systems engineering, physics and statistics.
This monograph highlights the future of financial engineering in the making.
A model building procedure will now be outlined. The procedure follows in three stages: 1) Preliminary
Structure Determination, 2) Parameter Estimation, and 3) Diagnostics. These stages are not disjoint, for in
reality the results of any one may alter the results of another and thus require a repeated application of
some or all of the previous stages. Experience has shown that the actual construction of a model involves
repeated cycling through the stages.
The specification of a model almost exclusively involves purely economic considerations. The model may
be used as an aid in economic analysis, policy simulation, or policy optimization, but each case imposes
special demands on the specification. The result of such considerations generally determines the overal1
size of the model, the number of inputs and outputs, and the definition of these variables. In addition, the
outputs of the model are usually decomposed into two types: the endogenous variable which are outputs of
dynamic equations, called behavioral equations and variables which are outputs of non-stochastic
equations, called definitional equations. A choice must be made as to the use of variables in current price
(inflated) or constant price (deflated). The economic specification stage can be summarized as one in which
the following information is determined:
1) The specific purpose of the model, thereby fixing the overall size; and hence, an enumeration of all the
outputs and their type, and, an enumeration of all the inputs and their type.
2) The output definitions; whether it is explained by a behavioral equation together with all its explanations
(inputs to the equation), or, whether it is determined by a definitional identity.
The second stage is the most challenging of the two This stage combines the use of a priori economic
information, hypothesis testing techniques, and cross-correlation analysis from the black box approach. In
econometric terminology, the word “structure” denotes a complete definition of the functional or stochastic
relationships between all of the endogenous and exogenous variables. The specific meaning of structure can
be ascertained by examining each equation of the structural form. Before accepting the results of any
7
estimation, they must be tested for their adequacy. The auto- and cross-correlation functions for the model
residuals constitute an important diagnostic checks. The last diagnostic to be employed is perhaps the most
important,, namely, the model’s forecasting performance. After having successfully met the other
diagnostic checks a model is not accepted until it has demonstrated its ability to forecast. Forecasts are then
made with each model from the end of its sample period up to the present, using the (historical) observed
inputs over this period. Thus, forecasts are obtained outside of the sample period . Such simulations more
closely approach reality and serve as a good guide in judging the model’s adequacy in forecasting the
unknown future. This gives additional insight into the time-invariance of the model structure. The modeling
procedure described in this monograph was designed to incorporate three concepts. First, employment is
made of all available a priori information provided by thus eliminating beforehand the possibility of
expending effort on fruitless searches for non-existent. relationships (interconnections), Second, the basic
philosophy of the “black box” approach is then applied allowing the data to decide the exact dynamic
structure. Thus, overly complex (statistically unsubstantiated) structures are automatically eliminated.
Third, diagnostics are continually employed which are designed to both reveal inadequacies and indicate
how improvements can be made.
Linear Quadratic Tracking Problem to Economic Stabilization
Some of the typical policy problems of recent interest. have been if how and when taxes should be raised or
lowered, whether the money supply should grow at a constant rate or be adjusted in response to economic
conditions, when and how fast the Reserve Bank of India should change credit conditions, and whether
government expenditures should expand or contract,. The objectives of these policies have included such
successes as the minimization of unemployment and the maintenance of full-capacity output, the control of
inflation, a reasonable rate of economic growth, and perhaps more recently, the elimination of poverty
through income redistribution. The control theory has been found useful and often essential to advance the
economic theory and better regulate economies. Application of control engineering viewpoints and control
techniques to treat stabilization of economic systems using optimal control and optimal filtering theories
have attracted many economists and control engineers.
The economic stabilization has been approached as a dual tracking problem in optimal control. The
problem that is defined and solved involves tracking nominal state and nominal policy trajectories, subject
8
to a quadratic cost functional and the constraint of a linear system. This is actually quite general, and will
enable us to penalize for variations in, as well as for the levels of, the state variables and control variables.
The aim of the policy plan will be to make xi , the state vector, track as closely as possible a nominal state-
vector , but subject to ut, the control vector, tracking a nominal control vector In other words, it is desired
that variables such as GNP, investment., and unemployment to follow as closely as possible nominal or
“ideal” time paths throughout the planning period. The nominal time paths for GNP and investment for
example, would probably grow at some steady rate, while that for unemployment to follow as closely as
possible nominal or “ideal” time paths throughout the planning period. The control variables, of course, are
used to make GNP, investment, and unemployment move in the desired direction, but we are not free to
manipulate the control variables in any way whatsoever; they in turn must also stay close to a set of
nominal or ideal time paths. For example, we are not free to increase government spending or the money
supply by 100 percent in one year and decrease them by 200 percent in the next year, etc. Manipulating
policy variables has very real costs associated with it, and these costs must be embodied in the cost
functional.
The system of interest is of the form
with the initial condition
Here x i is the n-dimensional state vector at time i, ui the r-dimensional control vector at time i, and zi an s-
dimensional vector representing, at time i, s exogenous variables which are known for all i but cannot be
controlled by the policy planner. A, B, and C are time-invariant n X n, n X r, and n X s matrices.
Let xi and ui be the nominal state and control vectors that is desired to be tracked. At time i we would like
xi to be close to xi and ui to be close to ui. We assume that xi and ui have been specified for the entire
planning period, i = , 1, … , N . The cost functional is given by
where Q is an n X n positive semi-definite matrix and R is an r X r positive definite matrix. The optimal
control problem is to find a control sequence {u i , i = 0, 1, . . . , N - 1} such that
9
and the cost functional (3) is minimized. It is important to keep in mind the meaning of the cost functional (3). Both Q and R will normally be
diagonal matrices. The elements of Q give the relative costs of deviating from the nominal path of each
state variable, for example, the cost of deviating from nominal GNP relative to the cost of deviating from
nominal unemployment. Some of the elements of Q may be zero. The elements of R give the relative costs
of deviating from the nominal paths of the control variables. For example, we would expect it to be more
costly to manipulate the tax rate than to manipulate the money supply. All of the diagonal elements of R
must be nonzero. This is both meaningful in terms of the economic problem and necessary for a
mathematical solution. Finally, the comparative magnitudes of Q and R give the costs of controlling the
economy relative to the costs of having the economy deviate from its ideal path, that is, the relative costs of
means versus ends.
Solving for the optimal control
The Hamiltonian is
where p i is the vector of costates. The canonical equations for the problem then are
and these are subject to the split boundary conditions
where (10) is a result of the transversality condition. Finally, the minimization of the Hamiltonian is written as
10
yielding
The assumption that p *
i is of the form
Substituting this into ( 12 ) gives
and substituting ( 13 ) and ( 14 ) into ( 7 ) and ( 8) gives
Substituting ( 13 ) into the left hand side of ( 15b ) rearranging terms gives
It is assumed that E is nonsingular and E -1 exists. Then (16a ) can be written as
11
Equating the coefficients of x *
i
Equations ( 24 ) and ( 25 ) provide the boundary conditions for ( 21 ) and ( 22 ) respectively.
12
All of the above steps involve iterative solutions (and only N iterations) that require little more than
multiplying and adding matrices (albeit large matrices, where n might. be on the order of several hundred
for a large econometric model).
13
The economic stabilization policy is approached as a problem in optimal control. The optimal control
problem is defined as a dual discrete-time tracking problem for a linear time-invariant system with a
quadratic cost functional.
It is important to keep in mind the fact that the cost functional accumulates penalties only over a finite time
period. Optimal paths for some variables behave strangely during the last few quarters of the planning
period. Changes in the money supply, for example, affect the short-term interest rate immediately, but at
least two quarters must elapse before there is any impact on residential investment and hence on GNP.
Therefore, if the cost functiona1 does not penalize directly for interest-rate deviations from the nominal, the
optimal quarterly change in the money supply wi11 always be equal to the nominal value. The solution to
this problem is to extend the planning period beyond the time horizon of actual interest. If, for example,
one was interested in formulating an optimal stabilization policy for the next three years, he should extend
the planning period to four or five years to obtain a numerical solution.
Modelling of Stochastic Analysis and Stochastic Control
Stochastic analysis and stochastic control have entered the field of finance at a very rapid, sometimes
explosive, pace in recent years. Powerful techniques from these disciplines have been brought to bear on
almost all aspects of mathematical finance: the study of arbitrage, hedging, pricing, consumption/portfolio
optimization, incomplete and/or constrained markets, equilibrium, differential information, the term-
structure of interest rates, transaction costs and so on. At the same time, the development of sophisticated
analytical and numerical methods, based primarily on partial differential equations and on their numerical
solutions, have helped to increase the relevance of these contributions in the everyday practice of finance.
Stochastic Analysis and Stochastic Control is addressed to a wide audience of control people from
academia and industry. The mathematics of finance and financial engineering is addressed to both
engineers and applied mathematicians from the following areas: stochastic control, stochastic analysis,
applied probability, estimation and identification, systems theory, stability, adaptive control, linear and
nonlinear systems, computational methods, optimization, modelling and control applications. It is important
to bring together leading stochastic control researchers working in a new exciting area of application of
stochastic control theory. It is also important to stress that the methods used in the mathematics of finance
have been successfully used in other fields.
14
Adaptive Control in the mathematics of finance and financial engineering
Lai and Wong [1] start their paper with a brief review of recent developments in the pricing and hedging of
American options. This paper then modifies the basis function approach to adaptive control and
neurodynamic programming and applies it to develop nonparametric pricing formulas for actively traded
American options as well as simulation-based optimization strategies for complex over-the-counter options;
the corresponding optimal stopping problems are prohibitively difficult to solve numerically by standard
backward induction algorithms because of the curse of dimensionality. An important issue in this approach
is the choice of the basis functions, for which some guidelines and underlying theory are provided.
In [2], Ross writes “Despite such gaps (in the theory), when judged by its ability to explain the empirical
data, option pricing theory is the most successful theory not only in finance, but in all of economics.” A call
(put) option gives the holder the right to buy (sell) the underlying asset by a certain date (known as the
“expiration date” or maturity”) at a certain price (known as the “strike price”). European options can be
exercised only on the expiration date, whereas American options can be exercised at any time up to the
expiration date. For European options, closed-form option pricing formulas were derived in the seminal
papers [3] and [4] through a dynamic hedging argument and a no-arbitrage condition. Except for Merton’s
[4] result that American calls written on nondividend-paying stocks reduce to European calls, American
option valuation involves finite-horizon optimal stopping problems and has to be performed numerically.
One of the “gaps” in option pricing theory noted by Ross is that “surprisingly little is known about the
exercise properties” of American options. Another gap is related to “a surprisingly small empirical
literature” which “should increase; options and option pricing theory give us an opportunity to measure
directly the degree of anticipated uncertainty in the markets.” During the fifteen years that have elapsed
since Ross’ review, there have been many advances in option pricing theory and the gaps mentioned by
Ross have been narrowed considerably. However, much still remains to be done. This work modifies the
basis function approach to adaptive control and neuro-dynamic programming and apply it to theoretical and
empirical analysis of American options. For actively traded American options using market data to provide
a sufficiently large training sample for estimating the parameters of a suitably chosen learning network to
price and hedge a new option. The choice of basis functions for the learning network is based on a
decomposition formula for the value function of the optimal stopping problem that gives the price of an
15
American option. This work also considers over-the-counter (OTC) American options. Since trading is
performed “over the counter,” the options data are not available for nonparametric estimation of pricing
formulas. Therefore, the traditional parametric approach is adopted instead. However, the underlying
optimal stopping problem is often prohibitively difficult to solve numerically by backward induction
because of the “curse of dimensionality” when the payoff is associated with multidimensional state
variables. To circumvent this difficulty, the work uses basis functions to implement a simulation-based
optimization scheme. Tsitsiklis and Van Roy [5], [6] have recently introduced this “neuro-dynamic
programming” approach to pricing highly complex American options that involve multidimensional asset
prices, but have not provided a systematic way of choosing the basis functions for this approach. Lai and
Wong consider this problem, review recent developments in simulation-based optimization for American
option valuation and provide some underlying theory. The interplay of leading-edge stochastic control
theory and American option valuation is discussed.
Neural and other learning networks and simulation-based dynamic programming have a burgeoning
literature in stochastic control [7-43]. In this work Lai and Wang apply these ideas to the valuation of both
actively traded and over-the-counter American options. They point out several basic issues that have come
up in the course of applying these ideas. One is the choice of basis functions. Closely related is the balance
between statistical efficiency and computational complexity in the presence of large data sets. Another
algorithm to compute functions via regression and simulation. Although American option valuation only
involves the simplest kind of stochastic control problems, namely, optimal stopping, it already encounters
the long-standing “curse-of-dimensionality,” and how it handles this “curse” should provide useful clues
for the solution of more complicated stochastic control problems.
Simulation in Financial Engineering
Simulation algorithms in the field of financial engineering offers challenges specific to financial simulation
and approaches that researchers have developed to handle them [283-288]. Many problems in financial
engineering require numerical evaluation of an integration. Several merits make simulation popular among
professionals as a methodology for these computations.
First it is simple to apply to many problems. For most derivative securities and financial models, even those
that are complicated and high dimensional, it takes relatively easy work to create a simulation algorithm for
16
pricing the derivatives under the model. Pitfalls in numerical implementation of simulation algorithms are
relatively rare. For the most part a little knowledge and effort go a long way in financial simulations.
The second merit of simulation is its good performances on high-dimensional problems: the rate of
convergence of a Monte-Carlo estimate does not depend on the dimension of the problem. The issues of
dimension and convergence are becoming increasingly important as securities markets and financial risk
management become more sophisticated.
The third attraction of simulation is the confidence interval that it provides for the Monte-Carlo estimate.
This information makes possible of how much computational effort might be needed in order to produce an
estimate of acceptable quality.
Financial engineers frequently apply simulation techniques to derivative securities. The derivatives are
financial instruments whose payoffs derive from the values of other underlying financial variables, such as
prices and interest rates. The example is the European call option, whose payoff is max{ST – K, 0}, where
ST is the price of the stock at time T, and K is the pre-specified amount called the strike price. This option
gives its owner the right to buy the stock at time T for the strike price K: If ST > K, the owner will exercise
this right, and if not, the option expires worthless. If the future payoff of a derivate derives from the
underlying, is there a way to derive the present price of the derivative from the current value of the
underlying? Under some theoretical conditions on the payoff of the derivative, and the possibilities for
trading in the market, the answer is yes. If it is possible to replicate the derivative’s payoff by trading in a
portfolio of securities available on the open market, then the combination of executing this trading strategy
and selling the derivatives has no risk. This is known as hedging the sale of derivatives, and hedging
strategies are of great practical interest in their own right, as well as being of theoretical interest in
justifying no-arbitrage pricing. The pricing theory has this name because it postulates that there are no
arbitrages, which are opportunities to make a positive amount of money with zero risk or cost. Such
opportunities are suppose to disappear, should they exist, because unlimited demand for them would drive
their cost above zero.
The riskless combination of a derivative minus the initial portfolio of its replicating strategy must have
nonpositive cost to avoid arbitrage; assuming the same of the opposite combination, the price of the
derivative must equal the cost of its initial replicating portfolio. A basic theorem of mathematical finance
17
states that this price us the expectation of the derivative’s discounted payoff under an equivalent martingale
measure. This is a probability measure under which discounted asset prices are martingale, and it generally
does not coincide with the original probability measure which models the real world. When discounting is
done with the value of a riskless money market account, the equivalent martingale measure is known as the
risk-neutral measure, because each investor had a neutral attitude towards risk, he would demand the same
return on all risky assets as on a riskless asset.
Given this, pricing a derivative evaluating the expectation of the sum of all its discounted payoff, under a
specified measure. The discounting is crucial and allows for appropriate comparison between cash flows,
whether positive or negative, at different times. Since the probability measure of financial models typically
have density, derivative pricing is evaluating the integral of the product of payoff and probability density
over all possible paths of the underlying.
As an example, consider the European call option under the Black-Scholes model , for which the
distribution of the log stock price ln ST is normal with mean and variance T under a
probability measure P. Here S0 is the initial stock price and and are called respectively the drift and
volatility. Under the risk neutral measure Q, ln ST is normal with mean ln ST + ( /2)T and the same
variance r is the instantaneous interest rate on a riskless money market account. No-arbitrage price of the
European call option is
and are respectively the cumulative distribution and probability density function of the standard normal.
This is the famous Black-Scholes formula.
The standard Mone-Carlo approach to evaluating such expectation is to simulate under the equivalent
martingle measure a state vector which depends on the underlying variables, then evaluate the sample
average of the derivative payoff over all trials. This is an unbiased estimate of the derivative’s price, and
18
when the number of trials n is large, the central limit theorem CLT provides a confidence interval for the
estimate, based on the sample variance of the discounted payoff.
The Monte-Carlo approach is similar for other financial engineering problems, such as finding hedging
strategies and analyzing portfolio return distributions in order to assess the risk of one current portfolio or
select a portfolio with the most attractive combination of reward and risk. All of these rely on the same
basic approach of simulating many trials, each of which is a path of underlying financial variables over a
period of time, computing the values of derivative on this path, and looking at the distribution of these
values.
In some application of simulation, there is no great conceptual difficulty involved in generating simulated
path, other than that of producing pseudo-random number with a digital computer. For instance, when
estimating the steady-state mean of a random variable in a queuing system, the model specifies the
transition rate from any state, and it is not theoretically difficult to sample the next state from the correct
distribution. The situation in financial situation is not so simple. The models of mathematical finance are
usually specified by stochastic differential equation (SDE) under the equivalent martingale measure used
for pricing. Sometimes it is possible to integrate these SDEs and get a tractable expression for the state
vector, but not always.
An example that poses no difficulties is the Black-Scholes Model, which has
19
In this model, the situation is not appreciable more difficult when a path-dependent option whose payoff
depends on the value of the state vector at many times. For instance, a discretely monitored Asian call
option has the payoff max {ST -K, 0} where ST = is the average price. Now the simulation must
generate the entire path St1 , St2, . . . , Stm. Assuming tk = Tk /m kh. The way to simulate the whole path is
to generate m independent standard normal random variables Z1(i) , Z2
(i) , . . . , Zm(i) for the i-th path and set
This provide the correct multivariate distribution for (St1 , St2, . . . , Stm ) and hence the correct distribution
for ST .
Another challenge in path generation is continuous path dependent. While the payoff of the European call
option depends only on the terminal value of the state vector, and the payoff of the discretely monitored
Asian call option depends only on a finite state of observation of the state vector, some derivatives have
payoff that depend on the entire continuous-time path. An example is a down-and-out option that pays off
only if stock price stays above some barriers, or equivalently, if the minimum stock price is above the
barrier. Suppose the stock price obeys the Black-Scholes model. Because
20
Almost surely, the former is not an acceptable substitute for the latter. It is necessary to introduce a new
component Mt = minu [ 0, t ] Su into the state vector; this can be simulated since the joint distribution of
Mt and St is known.
A slightly subtler example occurs in the Hull-White model of stochastic interest rate. The SDE governing
the instantaneous interest rate rt is
Where r is the long term-interest rate, is the strength of mean reversion, and is the interest rate’s
volatility. Integration of these SDE yields the distribution of rt , which is normal. Then the simulated path
rt1 , . . . , rtm is adequate for evaluating payoff that depends only on these interest rates, but not for
evaluating the discount factor DT = ; the discrete approximation h
does not have the right distribution. Instead one must add Dt to the state vector and simulate using its joint
distribution with rt , which is easily computable.
Some financial models feature SDEs that are not easily integrable, as the Black-Scholes and Hull-White
models are. An example is the Cox-Ingersoll-Ross model, in which the SDE is
This models principal advantage over Hull-White is that the instantaneous interest rate must remain
nonnegative. However, there is no useful expression for the distribution of rt given r0 . A simulation of this
model must rely on an approximate discritisation r of the stochastic process r . Because the laws of these
processes are not the same, the Monte-Carlo estimate based on r may be biased for the true price based on
r . This bias is known as discretization error.
The standard error of a Monte-Carlo estimate decreases as 1/ C , where C is the computational budget.
This is not an impressive rate of convergence for a numerical integration method. For simulation to be
competitive for some problem, it is necessary to de3sign an estimator that has less variance than the most
obvious ones. A variance reduction technique is a strategy for producing from one Monte-Carlo estimator
another with lower variance given the same computational budget.
Portfolio Selection in the mathematics of finance and financial engineering
21
Yin and Zhou [44] focus on Markowitz’s Mean-Variance Portfolio Selection [45] in a model incorporating
“regime switching.” The objective of Markowitz’s Nobel Prize winning single-period mean-variance
portfolio selection model is to minimize the risk measured by the variance of the wealth while meeting a
target expected wealth. This model provided a foundation for modern finance theory and stimulated
numerous extensions and applications. Yin and Zhou are concerned with a discrete-time version of the
mean-variance problem under regime switching that is modeled by a Markov chain. The regime switching
can be regarded as a factor process reflecting the overall market environment (e.g., bullish or bearish) as
well as other economic factors. The incorporation of the regime switching makes the model more realistic,
but the resulting system becomes more complex. To reduce the complexity Yin and Zhou derive, using
weak convergence methods, a continuous-time limit system whose solution they have already obtained.
Based on this solution, asymptotic optimal mean-variance portfolio selection policies are obtained.
Markowitz’s Nobel-prize winning mean-variance portfolio selection model (for a single period) [56], [57]
provides a foundation of modern finance theory; it has inspired numerous extensions and applications. The
Markowitz model aims to maximize the terminal wealth, in the mean time to minimize the risk using the
variance as a criterion, which enables investors to seek highest return upon specifying their acceptable risk
level. There have been continuing efforts in extending portfolio selection from the static single period
model to dynamic multiperiod or continuous-time models. However, the research works on dynamic
portfolio selections have been dominated by those of maximizing expected utility functions of the terminal
wealth, which is in spirit different from the original Markowitz’s model. For example, the multiperiod
utility models were investigated in [52]–[55] and [58]–[60], among many others. As for the continuous-
time case, the famous Merton paper [59] studied a utility maximization problem with market factors
modeled as a diffusion process (rather than as a Markov chain). Along another line, the mean-variance
hedging problem was investigated by Duffie and Richardson [51] and Schweizer [63], where an optimal
dynamic strategy was sought to hedge contingent claims in an imperfect market. Optimal hedging policies
[51], [63] were obtained primarily based on the so-called projection theorem. Very recently, using the
stochastic linear-quadratic (LQ) theory developed in [49] and [67], Zhou and Li [70] introduced a
stochastic LQ control framework to study the continuous-time version of the Markowitz’s problem. Within
this framework, they derived closed-form efficient policies (in the Markowitz sense) along with an explicit
22
expression of the efficient frontier. In the aforementioned references, for continuous-time formulations of
mean-variance problems, stochastic differential equations and geometric Brownian motion models were
used. Although such models have been used in a wide variety of situations, they have certain limitations
since all the key parameters, including the interest rate and the stock appreciation/volatility rates, are
assumed to be insensitive to the (very likely) drastic changes in the market. Typically, the underlying
market may have many “modes” or “regimes” that switch among themselves from time to time. The market
mode could reflect the state of the underlying economy, the general mood of investors in the market, and so
on. For example, the market can be roughly divided as “bullish” and “bearish,” while the market
parameters can be quite different in the two modes. One could certainly introduce more intermediate states
between the two extremes. A system, commonly referred to as the regime switching model, can be
formulated as a stochastic differential equation whose coefficients are modulated by a continuous-time
Markov chain. Such a model has been employed in the literature to discuss options; see [47], [48], and [50].
Moreover, an investment-consumption model with regime switching was studied in [68]; an optimal stock
selling rule for a Markov-modulated Black–Scholes model was derived in [69]; a stochastic approximation
approach for the liquidation problem could be found in [65]. [71] treated the continuous-time version of
Markowitz’s mean-variance portfolio selection with regime switching and derived the efficient portfolio
and efficient frontier explicitly. Motivated by the recent developments of mean-variance portfolio selection
and Markov-modulated geometric Brownian motion formulation, a class of discrete-time mean-variance
portfolio selection models were developed and their relationship with the continuous-time counterparts was
revealed. The discrete-time case is as important as the continuous-time one. First, frequently, one needs to
deal with multiperiod, discrete-time Markowitz’s portfolio selection problems directly; see [56] for a recent
account on the topic, in which efficient strategies were derived together with the efficient frontier. In
addition, to simulate a continuous-time model, one often has to use a discretization technique leading to a
discrete-time problem formulation. Yin and Zhou [44] focus one of the main features of the problem to be
tackled is that all the market coefficients are modulated by a discrete-time Markov chain that has a finite
state–space. Owing to the presence of long-term and short-term investors, the movements of a capital
market can be divided into primary movement and secondary movement naturally leading to two-time
scales. Besides, with various economic factors such as trends of the market, interest rates, and business
23
cycles being taken into consideration, the state space of the Markov chain, representing the totality of the
possible market modes, is often large. If simply treated each possible mode as an individual one distinct
from all others, the size of the problem would be huge. A straightforward implementation of numerical
schemes may deem to be infeasible due to the curse of dimensionality. It is thus crucial to find a viable
alternative. To reduce the complexity, it has been observed that the transition rates among different states
could be quite different. In fact, there is certain hierarchy (in terms of the magnitude of the transition rates)
involved. Therefore, it is possible to lump many states at a similar hierarchical level together to form a big
“state.”
With this aggregated Markov chain, the size of the state–space is substantially reduced. Now, to highlight
the different rates of changes, a small parameter has been introduced into the transition matrix, resulting in
a singular perturbation formulation. Based on the recent progress on two-time-scale Markov chains (see
[61] and [66]), the natural connection between the discrete-time problem and its continuous-time limit has
been established. Under simple conditions, it can be shown that suitably interpolated processes converge
weakly to their limits leading to a continuous-time mean-variance portfolio selection problem with regime
switching. The limit mean-variance portfolio selection problem has an optimal solution [71] that can be
obtained in a very simple way under appropriate conditions. Using that solution, policies that are
asymptotically optimal can be designed. Current findings indicate that in lieu of examining the more
complex original problem, one could use the much simplified limit problem as a guide to obtain portfolio
selection policies that are nearly as good as the optimal one from a practical concern. The advantage is that
the complexity is much reduced. Yin and Zhou [44] remark that although the specific mean-variance
portfolio selection is treated in their work, the formulation and techniques can be generally employed as
well in the so-called hybrid control problems that are modulated by a Markov chain for many other
applications. Yin and Zhou [44] establish the natural connection between the discrete-time and continuous-
time models aiming at reducing the complexity of the underlying systems. They construct policies that are
based on optimal control of the limit problem and derives asymptotic optimal strategy via the constructed
controls. The results can be extended by allowing the Markov chain to be nonhomogeneous and/or
including transient states.
The contribution of H. M. Markowitz to economic theory
24
Prior to Markowitz’s portfolio theory [72,73] several economists had pointed to the need for diversification
in investment (“don’t put all your eggs in one basket”). This approach, however, was founded on lay
observation without due analysis and quantification. As is clear this idea without an appropriate scientific
grounding greatly irritated Markowitz and led him to his famous model on the creation of efficient
allocation and portfolio creation. His further enthusiasm in research was to a significant extent influenced
also by practice. Since 1952 he has worked in many well-known companies, which also created for him
the conditions for his further work. In the Fifties and Sixties he worked in Rand Corporation, General
Electric Corporation, Cairmand, Consolidated Analysis Centres Inc. etc. In his work he learnt in particular
optimisation techniques and methods of linear programming. He enriched logistic simulation models and
programming languages through his theoretical knowledge in the field of portfolio theory (at Rand he was a
co-creator of the programming language SIMSCRIPT, which in an updated version is still used today).
Collective investment via fund management companies, investment companies, and pension funds plays a
permanently important role in capital markets. The professional approach of these companies’ managers
does not occur without the use of portfolio creation methods. With the help of these it is possible to
determine in which securities from a selected range it is advantageous to deposit investors funds and
concurrently to also set the optimal ratios in individual assets. The work of H. M. Markowitz, published in
1952 in the article “Portfolio selection” [72] is considered the keystone of modern portfolio theory. The
work states that the price of a share represents the cash flow of future dividends discounted to the net
present value. According to this model an investor is interested in future expected dividend flows and thus
also the expected share price. Markowitz applied this conclusion also to portfolio theory when he stated
that investors are interested in the expected value of a portfolio, where this value may be quantified.
Among those values which investors are interested in are, according to him, risk and return. In creating his
selection model Markowitz works from certain abstractions and presumptions:
• investors have an aversion to risk,
• all investors invest at the same time,
• the basis of investment decisions is expected utility,
• investors make their investment decisions on the basis of expected risk and return, and
• perfect capital markets exist.
25
What does a portfolio in the financial market represent in the meaning of this theory? A portfolio is a set of
various investments, which an investor creates in order to minimize risks connected with investing and also
to find the best possible proportion between risks and returns. Since according to Markowitz’s theory the
investor is risk averse, the investor will create a portfolio with the aim of achieving the largest return for the
minimum risk. In quantifying the yield (return) of a portfolio Markowitz worked at first from determining
the expected yield of one instrument and then from the expected return of the whole portfolio.
The expected yield of the portfolio is a weighted average of the expected individual fields of individual
instruments in the portfolio, where the weightings are the shares of individual investments in the portfolio.
An investor is interested not only in the rate of return but also in the risk. In measuring risk Markowitz at
first works from the risk of one asset and then from the risk of the portfolio. The risk of a portfolio however
is not simply a weighted average of the risks of individual instruments in the portfolio. The degree of risk
of the portfolio is influenced also by other variables, in particular by the mutual relation between the yields
of individual instruments.
Markowitz stated that if an investor invests in a portfolio which perfectly positively correlated yields, then
it does not at all lower his risk, because the yields move in only one direction and the investor in such a
portfolio can suffer significant losses. The ideally compiled portfolio has negatively correlated yields, i.e.
the yields have an inverse movement. To compile such a portfolio however is in practice impossible. Assets
with non-correlated yields create a portfolio in which the yields have no relation to one another. The
benefits of diversification lie in the fact that a more efficient compensation effect of risk and return will be
achieved through an appropriate combination of assets, the correlation of which does not extend to a form
of completely positive correlation. In such cases the standard deviation of the yield of a portfolio is less
than the weighted average of the standard deviations of the assets in the portfolio. Diversification lowers
risk also in the case of a smaller number of securities – first of this risk is lowered quickly, gradually with
an increasing amount of securities, the effectiveness declines. It is thus possible to assess risk in the context
of a portfolio. We cannot judge the effective risk of any security in a way that we will examine it in
isolation. A part of the uncertainty concerning the yields of a security is “de-diversified” as soon as a
security is grouped with others in the portfolio. From Markowitz’s selection model it thus results that if an
investor wants to reduce the overall risk of the portfolio, then he must combine those assets which are not
26
perfectly positively correlated. Markowitz worked from the assumption that in the selection of a portfolio
the investor can select within the framework of the market various combinations of securities with various
yields and risks. In other words he assembled a so called feasible set of all possible combinations of
investment, which an investor is faced with in the market. The typical shape of a feasible set of portfolios
has entered financial theory under the title of an “umbrella shape”, which is depicted in the graph. From the
set of Pareto optimal combinations of expected yields and variances investors will according to Markowitz
select portfolios which: give the maximum expected rate of return at various levels of risk or offer
minimum risk in the case of various levels of expected rates of return. It was shown in a graph.In the graph
it can be seen and that given conditions fulfil the combinations S, in the case of which the investor will
achieve a maximum yield and E with the lowest risk. The set of portfolios fulfilling these two conditions is
known as the efficient set or efficient frontier. This limit depicts the points with the maximum rate of return
for a given level of risk, and which are measured by the standards deviations of the portfolio’s yields. From
the graph it can also be established that the efficient set will be located between points E and S. To this
efficient frontier he also applied indifference curves, which from the aspect of the theory of frontier utility
express the various combinations, in the case of which an investor tries to achieve the same utility. As
Markowitz states, indifference curves have a different slope in the case of a risk-averse investor and that of
a risk seeking investor. The indifference curves of an investor seeking risk have a more moderate slope and
will move closer to point S, where they will also touch the efficient frontier. In his selection model he gives
preference to the risk-averse investor.
Every portfolio manager recognises the value of the innovative approach of H. M. Markowitz in this field.
All his theoretical conclusions have become the basis and springboard for the development of other
theoretical analyses in the field of portfolio theory.
Bielecki and Pliska [74] present an application of risk sensitive, stochastic control theory to the problem of
optimally managing a portfolio of securities. The model features risky assets whose continuous-time
dynamics depend upon one or more economic factors which, in turn, are modelled as Gaussian processes.
The investor’s objective, an infinite horizon criterion, is to maximize the portfolio’s risk-adjusted growth
rate. The results in this paper are important in at least three aspects. First, by using continuous algebraic
Ricatti equations, the authors derive an explicit expression for the optimal trading strategy, thereby
27
obtaining explicit results for a new class of models. Second, by studying some particular examples, they
develop a better economic understanding about the interactions between factor processes and optimal
trading strategies. For instance, with stochastic interest rates it is rigorously shown that the strategy of
100% cash is not necessarily the least risky one. Third, this paper is one of the first to apply stochastic
control methods to fixed income portfolio management.
Beginning with the pioneering work by Merton [75], [76] and continuing through the recent books by
Karatzas and Shreve [77] and Korn [78], some very sophisticated stochastic control methods have been
applied to portfolio management. But most of these applications have been concerned with, at least
implicitly, the management of equities. In spite of an abundance of well-known mathematical interest rate
models, exemplified by the classical models of Duffie and Kan [79], Heath, Jarrow, and Morton [80], and
Asicek [81], one can find very few applications of modern control theory to fixed-income management.
Bielecki and Pliska [74] develop a stochastic control model that will provide a closed form solution to the
problem of long term optimal management of a fund comprised of equities and fixed-income instruments.
An optimization model for dynamic asset management, whether it is a discrete time or a continuous time
model, should recognize and address the following issues:
a) realistic modeling of the dynamics of asset price movements;
b) selection of planning horizon;
c) selection of optimality criterion;
d) computational feasibility.
Bielecki and Pliska [74] adopt a continuous time approach. Although in principle it is possible to obtain
existence results for optimal portfolio policies relative to selected optimality criteria under general model
dynamics such as (1), these dynamics are nevertheless too general for derivation of closed form solutions.
Some structural postulates need to be made about the parameter processes as well as about the driving
noises. One possibility is to model the parameters as Itô processes. Usually it is assumed that the price and
factor processes can be directly observed, but some models postulate only partial observations. Typically,
closed-form solutions to the corresponding optimal portfolio selection problems have only been obtained
after further specification of the functions , and Merton [76] called model an Inter-temporal Capital Asset
Pricing Model. At this point it is important to mention a key aspect of issue a), namely, model
28
completeness versus incompleteness. A non-degenerate model in which the number of sources of
randomness equals the number of underlying financial assets minus one, that is known as a complete
model; otherwise, the model is incomplete. Incomplete models are much more adequate descriptions of
financial markets. For example, if the factors are macroeconomic variables such as interest and
unemployment rates, then implicit with a complete model would be the foolish assumption that one would
know the full histories of the macroeconomic variables simply by observing the asset prices. On the other
hand, incomplete models are much more difficult to analyze, so there is an obvious tradeoff between issues
a) and d). Turning to model design issue b), the selection of the planning horizon, models in the literature
can be classified as: i) models with finite planning horizon, ii) models with random planning horizon (e.g.,
determined by the random time of bankruptcy), and iii) models with infinite time horizon. Each category
serves an appropriate purpose. In Bielecki and Pliska [74] effort is aimed at providing quantitative tools for
fund managers with long term goals of maximizing the growth rates of the portfolios they manage, keeping
the respective risks under control at the same time. In this case no specific finite horizon is appropriate, and
the model of Bielecki and Pliska [74] belongs in category iii) above. It has to be understood that, although
nobody in reality adheres to truly infinite planning horizons, a mathematical infinite horizon model
provides a convenient approximation for “long planning horizon” situations. Selection of the optimality
criterion, i.e., issue c) above, is another discriminant between competing models. The common approach,
and the one highly favored by most economists, is to choose an optimality criterion based upon a so-called
utility functional. For example, the objective might be to maximize expected utility of wealth at the (finite)
planning horizon or to maximize expected utility of consumption over the whole (possibly infinite)
planning interval. It is common practice to evaluate mutual fund by looking at criteria like average return
and volatility measured over historical periods; utility functions are never utilized. For the portfolio
problem Bielecki and Pliska [74] have decided to select the so called risk-sensitive objective criterion, a
criterion to be maximized. As a matter of fact, the model of Bielecki and Pliska [74] provides for direct
computation of optimal portfolio strategies in terms of well understood algebraic Ricatti equations. Their
results are for asset management in general, without regard to the specific application of fixed-income
management. Some recent, related papers are by Canestrelli [82], Canestrelli and Pontini [83], and Kim and
Omberg [84]; they used expected utility criteria. Detemple et al. [85] and Schroder and Skiadas [86]
29
studied somewhat more general models and also used expected utility criteria. Brennan and Schwartz [87]
and Brennan et al. [88] used numerical methods to solve the Hamilton–Jacobi–Bellman partial differential
equation for the optimal trading strategy, but they were hard pressed to derive a solution even though there
were only three factors and a similar number of assets. Explicit results on the application of the model of
Bielecki and Pliska [74] to fixed-income management were recently obtained by Bajeux-Besnainou et al.
[89], Deelstra et al. [90], Liu [91] and Sørensen [92]. Apparently working independently, they studied
some very similar, if not identical, special cases, all maximizing expected utility over a finite planning
horizon. Unfortunately, they all left unresolved some troublesome issues, such as optimal strategies calling
for unbounded positions in some assets. In particular, Deelstra et al. [90] considered a finite time terminal
utility maximization problem with the utility function. They studied a market consisting of two assets, a
stock and a discount bond maturing at a finite time. They also considered a factor process, the spot interest
rate process. They were interested in characterizing a portfolio process that maximizes the utility of the
wealth at time. Thus the planning horizon in their model coincides with the maturity of the underlying
bond, but this causes a singularity of the resulting stochastic control problem at the terminal time . They
tried to overcome this singularity problem by a judicious choice of the class of admissible controls (trading
strategies). Unfortunately, as it appears, their definition of the class of admissible controls is flawed, for it
involves the solution to the associated SDE which is supposed to be not satisfied by the wealth process,
resulting in a circular argument. Bielecki and Pliska [74] are able to incorporate in a rigorous fashion fixed-
income assets having infinite lives by utilizing the concept of rolling-horizon bonds, a concept introduced
by Rutkowski [93]. Rolling horizon bonds can be viewed, roughly, as mutual funds of zero coupon bonds,
all of which mature at about the same fixed distance in the future; these bonds are rolled over in a self
financing manner so this same fixed distance is preserved through the course of time.
Mean-variance hedging, a portfolio choice problem
Bobrovnytska and Schweizer[94] focus on mean-variance hedging, a portfolio choice problem where the
goal is to approximate (with respect to mean squared error) a given payoff by the final wealth from a self-
financing portfolio strategy. Since the wealth dynamics are linear in the chosen portfolio, it is natural to
attack the problem with linear-quadratic stochastic control methods. The novelty of this paper is that it
considers a setting with continuous semi-martingale asset prices in a general filtration which need not be
30
generated by Brownian motions. By exploiting ideas and concepts from mathematical finance, solvability
of the adjoint equations (which are generalized stochastic Riccati equations) can be studied quite precisely.
In the micro-movement of asset price, transaction data are discrete in value, irregularly-spaced in time and
extremely large in size. It has been documented in the finance literature that ignoring the discrete-type of
trading noise results in substantially inflated volatility estimates.
Bobrovnytska and Schweizer[94] show for continuous semimartingales in a general filtration how the
mean-variance hedging problem can be treated as a linear-quadratic stochastic control problem. The adjoint
equations lead to backward stochastic differential equations for the three coefficients of the quadratic value
process, and we give necessary and sufficient conditions for the solvability of these generalized stochastic
Riccati equations. Motivated from mathematical finance, this paper takes a first step toward linear-
quadratic stochastic control in more general than Brownian settings.
Stochastic control methods have a venerable history in the field of financial engineering, and a number of
Nobel Prizes bear ample witness to the fruitfulness of this interaction. One can for instance think of
Merton’s seminal contributions to portfolio optimization and option pricing, among other things. Even
earlier, Harry Markowitz was concerned with mean-variance analysis in financial markets, and this topic
has retained its popularity even after 50 years; see, for instance, the survey [95] which contains more than
200 references. However, in contrast to portfolio optimization based on utility functions, mean-variance
analysis in dynamic inter-temporal frameworks has only recently been linked to stochastic control in a
more systematic way. Bobrovnytska and Schweizer[94] explore this avenue further and to show that it
leads to results and insights in stochastic control even beyond the usual settings. In a given financial
market, the mean-variance hedging problem is to find for a given payoff a best approximation by means of
self-financing trading strategies; the optimality criterion is the expected squared error. In a series of recent
papers, this problem has been formulated and treated as a linear-quadratic (LQ) stochastic control problem
at increasing levels of generality; see for instance [96], [97], [98], [99], [100], [101], or [102] for an
overview and a historical perspective. In the general case where the market coefficients are random
processes, the adjoint equations turn out to lead to a coupled system of backward stochastic differential
equations (BSDEs) for the coefficients of the (quadratic) value functional. This has led to new interest in
and new results on general LQ stochastic control problems, and the mean-variance hedging problem has
31
been treated fairly explicitly by these methods. From the mathematical finance point of view, one drawback
of this approach is that almost all existing papers impose rather restrictive assumptions. To apply general
results from LQ stochastic control, Bobrovnytska and Schweizer[94] work with Itô processes and assume
that all their coefficients are uniformly bounded, which excludes many practically relevant models.
Moreover, the theory of BSDEs is only rarely used beyond the setting of a filtration generated by a
Brownian motion and, thus, strongly relies on a martingale representation theorem. On the other hand, the
mean-variance hedging problem has been solved in much higher generality by martingale and projection
techniques. One can allow continuous semimartingales in general filtrations and only needs an absence-of-
arbitrage condition; see [103] and [104] for recent overviews. The work of Bobrovnytska and
Schweizer[94] is a first step toward a fusion between mathematical finance and LQ stochastic control at
this more general level. For related recent results, see [105]. Bobrovnytska and Schweizer[94] presents the
basic model, explains the mean-variance hedging problem and casts it in the form of an LQ stochastic
control problem. Combining the martingale optimality principle with the natural guess that the value
process of this problem should have a quadratic structure, they then derive a system of BSDEs for the
conjectured coefficients They gives a necessary and sufficient condition for the first of these BSDEs (for
the quadratic coefficient ) to be solvable under the sole assumption that the underlying asset price process
is continuous. One can also show that is the value process of a dual control problem. Apart from continuity
of, Bobrovnytska and Schweizer[94] need that the filtration is continuous and that the variance-optimal
martingale measure satisfies the reverse Hölder inequality. They show how one can explicitly construct a
solution for the mean-variance hedging problem from the solutions of the BSDEs. This is conceptually well
known, but of course technically slightly different than in the usual case of a Brownian filtration.
Unified Bayesian estimation via filtering approach To overcome Brownian filtration., Zeng’s paper [106] presents a unified approach, via filtering, to estimate
stochastic volatility for micro-movement models. The key feature of the models is that they can be
transformed into filtering problems with counting process observations. The Markov chain approximation
method is applied to the filtering equations to construct consistent recursive algorithms, which compute the
joint posterior and the trade-by-trade Bayesian parameter estimates. To illustrate the approach, a micro-
32
movement model built on a jumping stochastic volatility geometric Brownian motion is studied in detail.
Simulation and a real data example are presented.
Stochastic volatility is well documented for asset prices in both macromovement and micromovement
[107]. Macromovement refers to daily, weekly, and monthly closing price behavior while micromovement
refers to transactional (trade-by-trade) price behavior. There is a strong connection as well as striking
distinctions between the macro and micromovements. The strong connection is observed through the
identity of the overall shapes of both, because the macromovement is an equally spaced time series drawn
from the micromovement data. Their striking distinctions are mainly due to financial noise. In
macromovement, the impact of noise is small and is usually neglected. In micromovement, however, the
impact of noise is substantial and noise must be modelled explicitly. If the noise is ignored, then the impact
of noise is transferred to volatility, and the volatility estimates are substantially inflated. This is
documented by [108], [109], and [110] for discrete noise and further in [111] and [112] for discrete plus
other types of noise. Economically, the asset price is distinguished from its intrinsic value and this
distinction is also noise. Noise, as contrasted with information, is well-documented in the market
microstructure literature. Three important types of noise have been identified and extensively studied:
discrete, clustering and nonclustering. First, intraday prices move discretely (tick by tick), resulting in
“discrete noise.” Second, because prices do not distribute evenly on all ticks, but gather more on some ticks
such as integers and halves, “price clustering” is obtained. [113] confirms that this phenomenon is
remarkably persistent through time, across assets, and across market structures. Third, the “nonclustering
noise” includes other unspecified noise, and the outliers in prices are one of the evidence for the existence
of nonclustering noise. In [112], a novel, economically well-grounded and partially observed
micromovement model for asset price is proposed to bridge the gap between the macro and micro
movements caused by noise. The most prominent feature of the proposed model is that it can be formulated
as a filtering problem with counting process observations. This connects the model to the filtering
literature, which has found great success in engineering and networking. Under this framework, the
observables are the whole sample paths of the counting processes, which contain the complete information
of price and trading time. Then, the continuous-time likelihoods and posterior, built upon the sample paths,
not only exist, but also are uniquely characterized by the unnormalized, Duncan–Mortensen–Zakai (DMZ)-
33
like filtering equation, and the normalized, Kushner–Stratonovich (KS) (or Fujisaki–Kallianpur–Kunita)-
like filtering equations respectively. Transaction (or tick) data are discrete in value, irregularly spaced in
time and extremely large in size. Despite recent advances in statistics and econometrics, obtaining
“reliable” parameter estimates for even simple, nonstochastic volatility, micromovement models are
extremely challenging. Contrasted with [114], [112] develops continuous-time Bayes estimation via
filtering with efficient algorithms for the parameter estimation of the micromovement model. That
represents a significant advance in the estimation for micromovement models, also because the continuous-
time likelihoods and posterior are utilized as the foundation for statistical inference. This foundation is
informationally better than those provided by the discrete-time likelihoods and posterior, which merely
make use of a discrete-time subset of the sample paths. In [112], however,only the parameters of a simple
model with GBM as value process is estimated. In this paper, first, a class of stochastic volatility
micromovement models is developed from the macromovement models by incorporating the three types of
noise mentioned. A new, jumping stochastic volatility (JSV) micromovement model, stemming from
geometric Brownian motion (GBM), is proposed and studied (later, it is called the JSV-GBM
micromovement model). Second, a unified approach, Bayes parameter estimation via filtering, is developed
for the mi-cromovement models, especially for estimating stochastic volatility. Stochastic volatility models
are more realistic and more interesting but more difficult to estimate than the simple model with GBM as
value process, where the parameters, the signal of interest, are fixed. In stochastic volatility model,
estimation becomes a “real” filtering problem: the stochastic volatility, the signal of interest, changes over
time and the stock prices are the observations corrupted by discrete types of noise. The JSV-GBM model is
employed to demonstrate the effectiveness of estimating stochastic volatility using Bayes estimation via
filtering. To illustrate the approach, JSV-GBMs consistent recursive algorithm, which approximates the
normalized filtering equation and calculates the joint posterior, is constructed in detail. Simulation results
show that the Bayes estimates for stochastic volatilities are close to their true volatilities, and are able to
capture the movement of volatility. Trade-by-trade volatility estimates for an actual transaction data set are
computed and they confirm that the volatility changes even more dramatically in micromovement.
A unified Bayesian estimation via filtering approach is developed for estimating stochastic volatility for a
class of micromovement models, which capture the impact of noise at the microlevel. The class of models
34
has an important feature in that it can be formulated as a filtering problem with counting process
observations. Under this formulation, the whole sample paths are observable, and the complete tick data
information is used in Bayes parameter estimation via filtering. A consistent recursive algorithm is
developed to compute the Bayes estimates for the parameters in the model, especially, the stochastic
volatility. Simulation studies show that Bayes estimates for time-invariant parameters are consistent, and
Bayes estimates for stochastic volatility are close to their true values and are able to capture the movement
of volatility quickly. The recursive algorithm is fast and feasible for large data sets and it has the recursive
feature allowing quick and easy update. The recursive algorithm is applied to Microsoft’s transaction data
and we obtain Bayes estimates and provide strong affirmative evidence that volatility changes even more
dramatically in trade-by-trade level. The model and its Bayes estimation via filtering equation can be
extended to jump-diffusion process for the value process, and other kinds of noise according to the sample
characteristics of data. The models and the Bayes estimation can be applied to other asset markets such as
exchange rates and commodity prices. It can also apply to assess the quality of security market, and to
compare information flows and noises in different periods and different markets.
Financial markets and corresponding changes in financial assets
Hanson and Westman [115] have observed that following important economic or financial announcements,
there can be large changes in the financial markets and corresponding changes in financial assets. While
these events may be scheduled or unscheduled, the amplitude of the market response may be unpredictable
or random. Often market volatility is modeled by continuous Brownian motion processes, but these are
inadequate for modelling such events and responses. Discontinuous jump processes are needed to model
these important events. Compound Poisson processes are used to model the unscheduled events and hybrid
deterministic-stochastic jump processes are used to model the time of scheduled events and the random
responses to them. The work of Hanson and Westman [115] was motivated by the “important event model”
of Rishel (116), but cast into stochastic differential equation format to facilitate generalization to
constraints, parameter values and computations. Application of the important events model is illustrated by
the optimal portfolio and consumption with the risk-adverse power utility control problem. However, the
usual separable canonical solution is not strictly valid due to added complexity of the jump events,
especially since the times of the scheduled events do not average out of the problem. Fortunately, iterations
35
about the canonical solution result in computationally feasible approximations. This illustrates that the
combined use of stochastic control theory and computations to successfully handle complex jump events.
A large number of continuous-time models of financial markets have been based upon continuous sample
path geometric Brownian motion processes, such as [117], [118], [119],and [120, Chs. 4–6]. However,
Merton [121], [119, Ch. 9], in the original jump diffusion finance model, applied discontinuous sample
path Poisson processes, along with Brownian motion processes, to the problem of pricing options when the
underlying asset returns are discontinuous. Several extensions of the classical diffusion theory of Black and
Scholes [122] were derived by minimizing portfolio variance techniques to jump diffusion models similar
to those techniques used to derive the classic Black and Scholes diffusion formulas. Earlier, Merton [119],
[120, Chs. 5 and 6] treated optimal consumption and investment portfolios with either geometric Brownian
motion or Poisson noise, and illustrated explicit solutions for constant risk-aversion in either the relative or
the absolute forms. Karatzas et al. [123] pointed out that it is necessary to enforce nonnegativity feasibility
conditions on both wealth and consumption, deriving formally explicit solutions from a consumption
investment dynamic programming model with a time-to-bankruptcy horizon, that qualitatively corrects the
Merton’s results. Sethi and Taksar [124] present corrections to certain formulas of Merton’s [119], [120,
Chs. 5 and 6] finite horizon consumption-investment model. Merton [120, Ch. 6] revisited the problem,
correcting his earlier work by adding an absorbing boundary condition at zero wealth and using other
techniques. Rishel [116] introduced a optimal portfolio model for stock prices dependent on quasi-
deterministic scheduled and stochastic unscheduled jump external events based on optimal stochastic
control theory. The jumps can affect both the stock prices directly or indirectly through parameters. The
quasi-deterministic jumps are deterministic only in the timing of the scheduled events, but the jump
responses are random in magnitude. The response to an event can be unpredictable, being based on solid
analysis, prefactored assessments, nuances or other factors external to the event. Rishel’s theoretical paper
is the motivation for this computational application paper. Much additional motivation comes from our
extensive prior research on computational stochastic control models for jump environments, such as
stochastic bioeconomic models with random disasters (see [125] and [126]) and stochastic manufacturing
systems subject to jumps from failures, repairs and other events (see [127]–[129]). Here, our model
formulation is a modification on Rishel’s [116] paper, with heavier reliance on stochastic differential
36
equations, constrained control, more general utility objectives, generalized functions, and random Poisson
measure. Many of the modifications make the model more realistic and computationally feasible. More
realism has been implemented through modifications systematically relying on linear or geometric
stochastic processes, while using control constraints on stock fractions and consumption policies. The
portfolio and consumption optimization problem is formulated and the subsequent partial differential
equation of stochastic dynamic programming is derived from a generalized Itô [130] chain rule. The
computations [122] for the portfolio optimization model have been carried out in MATLAB [131] to
demonstrate the reasonableness of the calculations.
Structure of interest rates with multiple ratings with emphasis on the arbitrage-free feature of the
model Bielecki and Rutkowski [132] provide approach to the Heath–Jarrow–Morton type modeling of defaultable
term structure of interest rates with multiple ratings. Special emphasis is put on the arbitrage-free feature of
the model, as well as on the explicit construction of the conditionally Markov process of credit migrations.
SOME of the basic elements of financial markets are discount and coupon bonds, which represent so-called
fixed income instruments. Coupon bonds can frequently be considered as portfolios of discount bonds with
various maturities. That is why, in many respects, the main object of study with regard to fixed income
instruments are discount bonds. A discount bond is an instrument that promises to pay a specified notional
amount at a specified maturity date, say . Thus, the value of a discount bond at any time is derived as a
function of the notional amount, as well as some other factors. The dependence of the price of discount
bonds on the physical time and on the maturity dates is known as the term structure of interest rates. It
needs to be said though that the term structure of interest rates can be represented in many alternative ways,
besides in terms of prices of discount bonds. If a discount bond pays the promised notional amount at the
maturity date with certainty, then such a bond is called a default free bond. A good example of such bonds
is provided by Treasury bonds. Frequently, a discount bond is not certain to pay the promised notional
amount at the maturity date. If so, such a bond is considered to be prone to default and is known as
defaultable bond. Most of the corporate bonds are defaultable bonds. The dependence of the price of
defaultable discount bonds on the physical time and on the maturity dates is known as the defaultable term
structure of interest rates.This work continues the line of research on reduced-form (or intensity-based)
modeling of defaultable term structure of interest rates originated in [133-135], [136–139], [140–142], and
37
[143]. We do not make here any attempt to classify or scrutinize various models developed in previous
works. For a detailed analysis of these approaches, the interested reader is referred to [144, Chs. 12 and 13].
From the mathematical perspective, the intensity-based modeling of random times hinges on the techniques
of modeling random times developed in the reliability theory. The key concept in this methodology is the
survival probability of a reference instrument or entity, or, more specifically, the hazard rate that represents
the intensity of default. In the most simple version of the intensity-based approach, nothing is assumed
about the factors generating this hazard rate. More sophisticated versions additionally include factor
processes that possibly impact the dynamics of the credit spreads. Important modeling aspects include: the
choice of the underlying probability measure (real-world or risk-neutral), the goal of modeling (risk
management or valuation of derivatives), and the source of intensities. In a typical reduced-form model, the
value of the firm is not included in the model. The specification of intensities is based either on the model’s
calibration to market data or on the estimation based on historical observations. Both in case of credit risk
management and in case of valuation of credit derivatives, the possibility of migrations of underlying credit
name between different rating grades is essential, as it reflects the fundamental feature of the real-life
market of credit risk sensitive instruments (corporate bonds and loans). In practice, credit ratings are the
natural attributes of credit names. Most authors (see, for instance, [140] or [142]) were approaching the
issue of modeling of the credit migrations from the Markovian perspective. In [134], we presented a
general outline of our Heath–Jarrow–Morton (HJM) type reduced-form approach to the modeling of
defaultable term structures that correspond to multiple ratings of corporate bonds. The present work
complements in many ways the previous one, mainly by providing a complete presentation of mathematical
aspects of our model. An important feature of our model is that it indeed is selfconsistent; that is, it is an
arbitrage-free model. In our previous paper, we have already indicated this arbitrage-free property of the
model. In the present work, this important property receives a full justification, since we provide a detailed
description of the enlarged probability space, which, when combined with the dynamics of the
instantaneous forward rates and the dynamics of the migration process, underly the arbitrage-free feature of
the model. It should be acknowledged that our construction can be extended in several directions, and some
of our assumptions can be weakened substantially. For instance, by combining our approach with results of
[146], Eberlein and Õzkan have shown that the model can be extended to the case of a term structure model
38
driven by a Lévy process (as opposed to the case of a standard Brownian motion examined here). The
important issue of the model’s calibration requires further studies Bielecki and Rutkowski [132]. Observe
that, in general, even if all four pieces of data—namely, the maturity date, the transition intensities, the
recovery scheme, and the initial rating—are identical for the two zero coupon bonds, the bonds themselves
may not be identical. In fact, if they are issued by two different entities, the associated migration processes
and are also distinct, in general. More specifically, if we consider the joint migration process , then the
marginal finite-dimensional distributions for and are identical, but in general . If, the credit migration
processes and may be either (conditionally) independent or dependent. In case of independent migration
processes and , no statistical dependence between credit migrations of the two bonds appears. In case of
mutually dependent migration processes, one needs to calibrate the dependence structure (or, more crudely,
the correlation structure) between and. The foregoing remarks are valid if one considers an application of
the general methodology presented in this paper to the valuation and hedging of individual defaultable
bonds—that is, corporate bonds issued by particular institutions, as well as to the valuation and hedging of
related credit derivatives. As an alternative, let us mention that the methodology presented in this paper
may be applied to a totality of alike defaultable bonds—that is, to the totality of bonds for which all four
features listed above coincide. In the latter approach, we identify all such bonds and we substitute them
with a representative bond with an associated representative migration process. This application of our
methodology aims at valuation and hedging of credit derivatives that are tied to the average market value of
corporate bonds of a given credit quality. Thus, the correlation structure between individual bonds is
deliberately disregarded. All that really matters in this interpretation are the marginal statistical properties
of individual corporate bonds, and they are identical for all bonds in a given class. Let us consider two
different defaultable bonds, and let us denote the associated migration processes as and . The respective
default events are and the respective default times are We may study two types of default correlations: the
correlation between random variables and the correlation between random variables and Various
correlation coefficients, such as Pearson’s (or linear) correlation coefficient, may be used to measure the
strength of these correlations. Likewise, we may analyze the correlations between the survival events of the
form: and . Of course, the correlation structure will typically vary depending on whether one uses the risk-
neutral probability or the real world probability .
39
Institutional money management by tracking a given stochastic benchmark
Paolo Dai Pra, Wolfgang J. Runggaldier, and Marco Tolotti [145] consider a problem from institutional
money management, where the objective of the investor/money manager is that of tracking or, better,
outperforming a given stochastic benchmark; the benchmark may be an index process such as the S&P 500
index, it may however also represent other economic quantities such as the value process of a non traded
asset like the inflation or the exchange rate. Typically, it is assumed that the investor may invest in a certain
number of risky assets in addition to a non risky one and let denote the investor’s wealth at time
corresponding to an investment strategy. The benchmark process will be denoted by and we will make the
realistic assumption that it is not perfectly correlated with the investment opportunities so that the investor
cannot completely control his risk (the market is incomplete). This portfolio problem, that is sometimes
also called active portfolio management, has been studied by various authors with particular reference to
[146]. As in [146], we will consider as relevant state variable the ratio of the investor’s wealth to the
benchmark. A natural way to proceed (see again [146]) is then to consider the process up to the exit from a
bounded domain and choose as objective the minimization of the discounted expected loss that penalizes
the deviation of from the constant 1 in the case of “benchmark tracking” and the amount by which falls
below1 in the case when the objective is that of “beating the benchmark.” While the classical criteria such
as the one described above are criteria in the mean, namely they involve expected values of costs/losses, in
this paper we aim at a stronger form of criterion, more precisely that of optimality (see, e.g., [147-150]) that
may in fact be quite appropriate for benchmark tracking/beating. The optimality criteria in use concern an
infinite horizon, which may still make sense in economic/financial applications every time an investor
makes his plans over a long horizon. To keep matters simple in this first attempt to apply a criterion to an
investment problem, we will consider here only symmetric cost functions, i.e., we will only consider the
benchmark tracking problem, thereby penalizing symmetrically both over and undershoots of with respect
to 1. An economically more meaningful asymmetric cost function that penalizes only undershoots/shortfalls
and corresponds to the typical benchmark beating/outperforming can still be dealt with in our approach, but
at the expense of less analytical tractability. More precisely, the aim is to find an investment strategy such
that for the corresponding ratio process we have . For the standard market models, namely those that that
go back to Merton [151], the price processes are geometric Brownian motions (lognormal processes). The
40
dynamics of the process are modified so that it becomes bounded (in a large domain) and still maintains the
main characteristics corresponding to lognormal models. Given the ergodic criterion, this cannot be
accomplished by simply stopping the process upon exit from a given compact set. The drift and diffusion
coefficients, which under certain assumptions can also be interpreted as a random time change and by
which the controlled process is increasingly slowed down as it gets closer and closer to a given boundary.
An interesting aspect that turns out in this context is that the optimal control and the solution of the
Hamilton–Jacobi–Bellman (HJB) equation do not depend on the particular choice of the random time
change. With the thus modified process, we will not only obtain the right ergodic behavior in order to make
the mean-optimality criterion meaningful, but it will furthermore allow us to show that the mean-optimal
control is also optimal. A final methodological aspect of [145] concerns the problem of solving the HJB
equation associated to the given (infinite-horizon) stochastic control problem with the criterion of
optimality in the mean. The traditional way is to guess a possible function which works only in specific
cases. Since a same objective may also be reached by using analytically different cost functions, we will
generalize the problem of solving the HJB by considering a class of possible cost functions and, given the
dynamics of the controlled process.
Kang Boda, Jerzy A. Filar, Yuanlie Lin, and Lieneke Spanjers [152], have studied a problem of optimal
control of ensuring that an adequate capital accumulates sufficiently quickly with sufficiently high
probability. The objective is to develop a tool that could be used to advise nonprofessional investors who
place their retirement benefits in a fund that permits only a limited number of options and offers only
limited opportunity to reallocate the money among these options; say, once a year. It is assumed that an
investor is primarily interested in maximizing the probability of being to afford the fund equal or exceed a
certain specified target amount at that terminal time. Since the mathematical framework in which we model
this problem is that of Markov decision processes (MDPs) (e.g., see [153]) and since a vast majority of
MDPs have objective criteria that depend on one of a number of “expected utility” criteria, it follows
immediately that our problem is essentially different from these classical MDP models. Instead, the
problem belongs to a class of models that are sometimes called “risk-sensitive MDPs.” The latter can,
perhaps, be traced back to [154] and constitutes an area where there has been a fair bit of research activity
in recent years (e.g., see [155], [156], [157], [158], [159], and [160]–[163]). Some of these contributions
41
tried to capture risk in terms of tradeoffs between mean and variance of suitable random variables, some
have followed [164] in considering the expected value of a suitable exponential utility criterion and some
have focused on the so-called “percentile optimality” (e.g, [155], [165], [158], and [162]). Markowitz [166]
pioneered the notion of mean-variance tradeoffs in finance literature and many more sophisticated,
dynamic and stochastic, financial models involving closely related issues have been studied in recent years
(e.g., see [167], [168], and [169]). More precisely, we consider a finite-horizon discounted MDP model in
which the decision-maker, at each stage, needs to decide what percentage of the current retirement fund to
allocate into the limited (small) number of investment options. We assume that both the initial investment
and the target retirement capital are known and that the number of stages is. Now, the first target hitting
time is a random variable whose distribution is specified by the choice of a policy. As mentioned above the
decision-maker’s goal is to find a policy which maximizes. While at first sight, this might appear to be a
very difficult problem it turns out a version of optimality principle can be shown to hold under mild
conditions when we work in an “extended” state space. However, even in the extended state space the new
process is not a Markov process under a general policy. Hence the existence and characterization of optimal
policies cannot be obtained by standard techniques. Instead, the techniques used by Kang Boda, Jerzy A.
Filar, Yuanlie Lin, and Lieneke Spanjers [152] are similar to those developed in [155] which dealt with a
related problem of minimizing the probability that the total discounted wealth is less than a specified target
level. From the preceding optimality principle, structural results about optimal policies can be easily
derived which, in turn, lead to a dynamic-programming type algorithm. It is assumed that the yield of a
given fund in every year in the future is best modeled as a random variable that takes on the past observed
yields from that fund with equal probability. This is a rather simplistic assumption that may not correspond
to reality. To try to alleviate this problem one could consider a model with rolling horizon policies. The
idea of this approach is that an optimal policy is found and the first decision rule is implemented. Then, if
new data are available, a problem with updated parameters and a new time horizon is solved. The first
decision rule from an optimal policy of the latter is then implemented and so on. Rolling horizons have
been used by many researchers (e.g., see [170]).
Appreciation rate of the stock and the volatility of the stock
42
Several authors consider a financial market where the risk free interest rate, the appreciation rate of the
stock and the volatility of the stock depend on an external finite state Markov chain [171]. The authors
investigate the problem of maximizing the expected utility from terminal wealth and solve it by stochastic
control methods for different utility functions. More precisely, they use the Hamilton–Jacobi–Bellman
equation and prove a verification theorem for the Markov-modulated portfolio optimization problem. Due
to explicit solutions it is possible to compare the value function of the problem to one where the financial
market has constant (average) market data. The case of benchmark optimization is also considered.
Stochastic control methods of utility functions for benchmark optimization
Often the financial is incomplete and consists of one bond and one risky asset. The incompleteness of the
market is due to stochastic coefficients appearing in the price process of the risky asset and the bond. More
precisely it is assumed that the interest rate of the bank account, the appreciation rate of the stock and the
volatility of the stock depend on an external continuous-time, finite state Markov chain . The state of the
Markov chain should represent the general market conditions (for a motivation; see, e.g., [173]). Models
with deterministic coefficients are only good for a relative short period of time and cannot respond to
changing conditions. In this Markov-modulated setting, it is desired to solve the classical portfolio
optimization problem where an investor wants to maximize the expected utility from terminal wealth. As
far as the information is concerned, the investor has at the time point of decision observes the stock price
and the market condition. This is due to the fact that in a diffusion price process model the quadratic
variation and thus the volatility can be approximated arbitrarily well by the price process (cf. [174]).
Therefore it is in principle sufficient to solve the optimization problem with complete observation. This is
done using stochastic control methods for a number of different utility functions, namely for logarithmic
utility, and for benchmark optimization. Motivated by [175], there is a growing literature dealing with
portfolio optimization problems under different aspects. Problems with stochastic volatility have for
example been investigated in [176], [177], [178], and [179], among others. Most of these papers assume
that the external process is a diffusion process itself, like in the established volatility model of [180]. To the
best of our knowledge, the first paper to model the volatility as a continuous-time Markov chain is [181].
As we will see this model has the advantage that many portfolio problems can be solved explicitly in
contrast to the diffusion setting (compare, for example, [176] and[179]). Moreover, a diffusion process can
43
be approximated arbitrarily closely by a continuous-time Markov chain (see [182]). Portfolio optimization
with stochastic interest rates are, e.g., treated in [183] and [184]. The authors of [184] consider the Ho-Lee
and the Vasicek model for the interest rate which are both diffusion processes. The solutions we obtain are
found with the help of stochastic control methods. More precisely, by the use of a verification theorem. For
a comprehensive presentation of this theory the reader is referred to [185] or [186], among others. In the
case of deterministic coefficients, this model has been considered in [187] and in a more general context by
[188]. In our setting, we are only partly able to solve the portfolio problem explicitly. A closed form
solution is derived when the discounted stock price process is a martingale. Portfolio optimization with
stochastic market data is more realistic than standard models with constant coefficients. The formulation of
the market condition as a continuous-time Markov chain makes the analysis simpler as in the case of a
driving diffusion. For the utility functions treated here, the maximal portfolio value can be computed as a
solution of a simple linear differential equation. More complicated is the case of benchmark optimization. It
remains open whether a closed form solution can be derived in the general Markov modulated case.
A generalized mean-variance model via optimal investment policy
For an investor to claim his wealth resulted from his multi-period portfolio policy, he has to sustain a
possibility of bankruptcy before reaching the end of an investment horizon. Risk control over bankruptcy is
thus an indispensable ingredient of optimal dynamic portfolio selection. Shu-Shang Zhu, Duan Li, and
Shou-Yang Wang [189] have proposed that a generalized mean-variance model via which an optimal
investment policy can be generated to help investors not only achieve an optimal return in the sense of a
mean-variance tradeoff, but also have a good risk control over bankruptcy. One key difficulty in solving the
proposed generalized mean-variance model is the nonseparability in the associated stochastic control
problem in the sense of dynamic programming. A solution scheme using embedding has been developed by
Shu-Shang Zhu et al [189] to overcome this difficulty and to obtain an analytical optimal portfolio policy.
Optimal dynamic portfolio selection is to redistribute successively in each time period an investor’s current
wealth among a basket of securities in an optimal way in order to maximize a measure of the investor’s
final wealth. The literature of dynamic portfolio selection has been dominated by the results of maximizing
expected utility functions of the terminal wealth [190–192], [193], [194], [195 - 199], [200]. The
Markowitz’s mean-variance model [201] has been recently extended in [202] to a multiperiod setting. The
44
analytical expression of the efficient frontier for the multiperiod portfolio selection is derived. The
continuous-time mean-variance formulation is studied in [204]. The dynamic mean-variance formulation in
[202] and [203] enables an investor to specify a risk level which he can afford when he is seeking to
maximize his expected terminal wealth or to specify an expected terminal wealth he would like to achieve
when he is seeking to minimize the corresponding risk. It is easier and more direct for investors to provide
this kind of subjective information than for them to construct a utility function in terms of the terminal
wealth. The tradeoff information between the expected return and the risk is clearly shown on the efficient
frontier, that is most useful for an investor to decide his investment decision. Performing an optimal
investment policy in accordance with a dynamic portfolio formulation does not eliminate the possibility
that an investor goes to bankruptcy in a volatile financial market before he claims his wealth at the terminal
stage. One key difficulty in solving the proposed generalized mean-variance model of Shu-Shang Zhu et al
[189] is the nonseparability in the associated stochastic control problem in the sense of dynamic
programming. A solution scheme adopting a Lagrangian dual formulation and using embedding is
developed by Shu-Shang Zhu et al [189] and a case study has been presented to gain insights about the
significance of the risk control in dynamic portfolio selection. Due to the volatility of financial markets,
bankruptcy control is an indispensable issue to be addressed in dynamic portfolio selection. In the mean-
variance formulation, reduction of the probability of bankruptcy can be achieved by increasing the
weighting coefficient for the variance. By incorporating a control of the probability of bankruptcy in the
generalized mean-variance formulation the dynamic portfolio selection problem becomes to seek a balance
among three objectives, which could lead to a more satisfactory tradeoff between the probability of
bankruptcy and the expected value of the final wealth. While the traditional stochastic control theory only
concerns a sole objective of minimizing an expected performance measure, variance control occurs
naturally in dynamic portfolio selection problems with a mean-variance formulation and many other
applications. The celebrated dynamic programming is the only universal solution scheme to achieve an
optimality for stochastic control problems. Dynamic programming, however, is only applicable to problems
that satisfy the property of separability and monotonicity. Variance control problems are not directly
solvable by dynamic programming due to its nonseparability. In this respect, variance minimization is a
notorious kind of stochastic control problems. Using an embedding scheme, a feedback optimal portfolio
45
policy can be obtained for variance control problems via parametric dynamic programming method, while
the corresponding optimal condition for the parameter can be derived by examining the relationship
between the primal and the auxiliary problems. The generalized mean-variance formulation proposed in
this note for risk control over bankruptcy in discrete-time dynamic portfolio selection can be extended to
continuous-time dynamic portfolio selection by imposing probability constraints at distinct time instants in
the continuous time horizon.
Risk-Sensitive Portfolio Optimization with Completely and Partially Observed Factors
Stettner [204] considers a market model with discrete time changes of the portfolio in which the prices of
assets depend on some economic factors. The purpose is to minimize a suitable long-horizon risk sensitive
cost functional. The factors may be completely observed, or partially observed, in particular only through
the changes in the asset prices. The form of the cost functional and the existence of both observed and
unobserved factors create a number of technical difficulties. To describe the influence of unobserved
factors a certain family of measure-valued processes is introduced. Using the so-called vanishing discount
approach the existence of the solutions to the long run risk sensitive cost Bellman equation is shown. It
turns out that the optimal portfolio strategy is a function of the observed factors and the above mentioned
measure valued processes.
Option Pricing Models in Financial Engineering
Modern option pricing techniques are often considered among the most mathematically complex of all
applied areas of finance. Financial analysts have reached the point where they are able to calculate, with
alarming accuracy, the value of a stock option. Most of the models and techniques employed by today's
analysts are rooted in a model developed by Fischer Black and Myron Scholes in 1973.
From the moment of its publication in 1973, the Black and Scholes Option Pricing Model has earned a
position among the most widely accepted of all financial models.
The idea of options is certainly not new. Ancient Romans, Grecians, and Phoenicians traded options against
outgoing cargoes from their local seaports. When used in relation to financial instruments, options are
generally defined as a "contract between two parties in which one party has the right but not the obligation
to do something, usually to buy or sell some underlying asset". Having rights without obligations has
46
financial value, so option holders must purchase these rights, making them assets. This asset derives their
value from some other asset, so they are called derivative assets. Call options are contracts giving the
option holder the right to buy something, while put options, conversely entitle the holder to sell something.
Payment for call and put options, takes the form of a flat, up-front sum called a premium. Options can also
be associated with bonds (i.e. convertible bonds and callable bonds), where payment occurs in installments
over the entire life of the bond, but this paper is only concerned with traditional put and call options.
Modern option pricing techniques, with roots in stochastic calculus, are often considered among the most
mathematically complex of all applied areas of finance. These modern techniques derive their impetus from
a formal history dating back to 1877, when Charles Castelli wrote a book entitled The Theory of Options in
Stocks and Shares. Castelli's book introduced the public to the hedging and speculation aspects of options,
but lacked any monumental theoretical base. Twenty three years later, Louis Bachelier offered the earliest
known analytical valuation for options in his mathematics dissertation "Th‚orie de la Sp‚culation" at the
Sorbonne. He was on the right track, but he used a process to generate share price that allowed both
negative security prices and option prices that exceeded the price of the underlying asset. Bachelier's work
interested a professor at MIT named Paul Samuelson, who in 1955, wrote an unpublished paper entitled
"Brownian Motion in the Stock Market". During that same year, Richard Kruizenga, one of Samuelson's
students, cited Bachelier's work in his dissertation entitled "Put and Call Options: A Theoretical and Market
Analysis". In 1962, another dissertation, this time by A. James Boness, focused on options. In his work,
entitled "A Theory and Measurement of Stock Option Value", Boness developed a pricing model that made
a significant theoretical jump from that of his predecessors. More significantly, his work served as a
precursor to that of Fischer Black and Myron Scholes, who in 1973 introduced their landmark option
pricing model.
The Black and Scholes Option Pricing Model didn't appear overnight, in fact, Fisher Black started out
working to create a valuation model for stock warrants. This work involved calculating a derivative to
measure how the discount rate of a warrant varies with time and stock price. The result of this calculation
held a striking resemblance to a well-known heat transfer equation. Soon after this discovery, Myron
Scholes joined Black and the result of their work is a startlingly accurate option pricing model. Black and
47
Scholes can't take all credit for their work, in fact their model is actually an improved version of a previous
model developed by A. James Boness in his Ph.D. dissertation at the University of Chicago. Black and
Scholes' improvements on the Boness model come in the form of a proof that the risk-free interest rate is
the correct discount factor, and with the absence of assumptions regarding investor's risk preferences.
In order to understand the model itself, we divide it into two parts. The first part, SN(d1), derives the
expected benefit from acquiring a stock outright. This is found by multiplying stock price [S] by the change
in the call premium with respect to a change in the underlying stock price [N(d1)]. The second part of the
model, Ke (-rt) N(d2), gives the present value of paying the exercise price on the expiration day. The fair
market value of the call option is then calculated by taking the difference between these two parts.
Assumptions of the Black and Scholes Model:
1) The stock pays no dividends during the option's life
Most companies pay dividends to their share holders, so this might seem a serious limitation to the model
considering the observation that higher dividend yields elicit lower call premiums. A common way of
48
adjusting the model for this situation is to subtract the discounted value of a future dividend from the stock
price.
2) European exercise terms are used
European exercise terms dictate that the option can only be exercised on the expiration date. American
exercise term allow the option to be exercised at any time during the life of the option, making american
options more valuable due to their greater flexibility. This limitation is not a major concern because very
few calls are ever exercised before the last few days of their life. This is true because when you exercise a
call early, you forfeit the remaining time value on the call and collect the intrinsic value. Towards the end
of the life of a call, the remaining time value is very small, but the intrinsic value is the same.
3) Markets are efficient
This assumption suggests that people cannot consistently predict the direction of the market or an
individual stock. The market operates continuously with share prices following a continuous Itô process.
To understand what a continuous Itô process is, you must first know that a Markov process is "one where
the observation in time period t depends only on the preceding observation." An Itô process is simply a
Markov process in continuous time. If you were to draw a continuous process you would do so without
picking the pen up from the piece of paper.
4) No commissions are charged
Usually market participants do have to pay a commission to buy or sell options. Even floor traders pay
some kind of fee, but it is usually very small. The fees that Individual investor's pay is more substantial and
can often distort the output of the model.
5) Interest rates remain constant and known
The Black and Scholes model uses the risk-free rate to represent this constant and known rate. In reality
there is no such thing as the risk-free rate, but the discount rate on U.S. Government Treasury Bills with 30
days left until maturity is usually used to represent it. During periods of rapidly changing interest rates,
these 30 day rates are often subject to change, thereby violating one of the assumptions of the model.
49
6) Returns are lognormally distributed
This assumption suggests, returns on the underlying stock are normally distributed, which is reasonable for
most assets that offer options.
After the Black and Scholes Model:
Since 1973, the original Black and Scholes Option Pricing Model has been the subject of much attention.
Many financial scholars have expanded upon the original work. In 1973, Robert Merton relaxed the
assumption of no dividends. In 1976, Jonathan Ingerson went one step further and relaxed the the
assumption of no taxes or transaction costs. In 1976, Merton responded by removing the restriction of
constant interest rates. The results of all of this attention, that originated in the autumn of 1969, are
alarmingly accurate valuation models for stock options.
Stochastic Control in Economic Theory
The control theory, in general, and stochastic control theory, in particular, are useful and sometimes even
essential to advance the state of economic theory and better regulate economies. Probabilistic and statistical
methods have been applied to economic problems such as development and planning of production and
inventory, growth models, portfolio selections, and to other aspects of theory of firms and so forth for some
time. These applications are, however, primarily to static or stationary economic situations
Application of control engineering viewpoints and control techniques to economic problems are of more
recent origin. Stochastic control theory has been applied to macroeconomic dynamic systems to answer
various policy questions in stabilization, optimal growth, planning, and others. Control of stochastic linear
dynamic macroeconomic models to minimize some quadratic social cost function is, perhaps, the best
known example of applications of stochastic control theory to macroeconomic systems The policymakers at
various government agencies and central banks are beginning to be seriously interested in modeling
national economics as large scale control systems and using econometric and control theoretic models and
techniques to evaluate alternative economic policies.
It may be emphasized that the aspect of decision making with imperfect information as the thread common
to both the stochastic control theory and economic problems. In assessing real or potential impacts of
stochastic control theory in economics, it is convenient to follow a conventional dichotomy of economics
50
into macroeconomics and microeconomics for expositional purposes. Briefly speaking, control theory has
been developed in our endeavor to guide or modify, to our advantage, time paths of objects in which we are
interested. Stochastic control theory is needed to take account of random variables or stochastic processes
which may be involved in determining, for example, the time paths of the objects of study and control. It is
not enough for policymakers to choose economic policies on a day-to-day or month-to-month basis, since
current policy decisions have impact not only in the present but also over some time in the future. This is
because the object of control, the national economy, is dynamic. The economic system must then be
modeled as a dynamic system and not just as a static system. The same can be said of basis economic
decision making units such as households and firms, which are the objects of study of microeconomic
theory, even though their decision making activities may not always be modeled as dynamic processes.
Besides the rather obvious applications of stochastic control theory to macroeconomic models such as
national economies there are applications to microeconomic situations. For example, economists
have recently constructed various models of market organization with imperfect information. Models of
search behavior, such as by the unemployed for jobs and by firms for workers, are nowadays being
modeled as ones with imperfect information. Many economic agents participate and interact directly or
indirectly with each other in markets. Their decision processes must then be modeled as dynamic decision
making processes under uncertainty or imperfect information. In models of markets or search behavior with
imperfect information, learning or formations of expectation regarding future behavior of economic
variables such as prices play important roles in explaining behavior of economic agents.
There are economic situations which illustrate the advantages of state-space representation of dynamic
models and on topics in which controllability or observability properties of dynamic systems play essential
roles. The concept of observability can be used to establish uniqueness of equilibrium solution in a quantity
adjustment (microeconomic) model.
Recursive dynamic least square instrument variable algorithm
The recursive dynamic least square instrument variable algorithm introduces the concept of the dynamics
of financial engineering based on innovative reasoning which precedes by forming an expectation and
verifying it by proper use of informational data set composed of the available knowledge and intuitive
observation. It deals with an application of the cybernetic method of recursive dynamic least square
51
instrument variable algorithm with on line parameter tracking adaptability for on line modelling of short
term national market index movement with a time slot of 1-day having interacting variables such as market
price indices of market-dominating fundamentals of industrial production. The investigation [252] unfolds
the traditions of controls and systems in estimation, identification and exploitation of structures to develop
efficient algorithms providing opportunities of significant research in financial engineering. The work
presented here formalizes a specific dynamic situation, namely the construction of a finite dimensional
process for daily movement of national market index. It has been clearly demonstrated with observed data
that the flexibility of the algorithms is remarkably broad. Indeed, it is possible to choose free variables in
such a way that the entire formal modelling process can be interpreted as a linear quadratic Gaussian
problem.
The system scientists traditionally study systems which involve continuous variables and have dynamics
which can be described by either differential or difference equations. There are many situations, in which
such models are not appropriate as the following examples will illustrate. Consider the daily movement of
all India market index. To identify the process a series of decisions tests in the form of correlation are to be
executed. The state of the system, which changes only at the discrete instant of time instead of
continuously, consists of an assortment of discrete variables. For such an important class of systems there is
a dearth of elegant and succinct identification techniques. Once the identification has been done the system
behavior can be obtained. An initial step in this direction has been exemplified in the situation one has a
discrete set of interpretation of a sequence of observed data, which form the basis for efficient and
relational assessment of system identification in order to provide the operational framework in which many
types of knowledge and information can be incorporated into the system. Development of mathematical
description of a system is often undertaken to predict performance and responses. National market is a
complex system. The market indices yield meaningful measures of systematic market development.
Returns to an index reflect averages of investor holdings and provide a performance benchmark for the
asset class included in the index. Comparison of investment returns to those of an index with the same
target weights has become the most widely accepted criterion of investment performance. The financial
news papers such as The Wall Street Journal and the London Financial Times report global, regional, local,
and industry indices on a broad range of portfolios [263],[264],[265],[266] and [267]. National market
52
indices, such as BSE Sensex, S&P 500 and London’s FTSE provide performance benchmarks for equity
investments in these national markets. Regional indices track stock returns in broad geographic regions
such as Americas, Europe, Nordic or Asia Pacific. [268].
The recursive technique is one in which an estimate is updated on receipt of fresh information. Beck [253],
[254], [255], Ljung [256], [257], [258], [259] and [260]] have given a good coverage of recursive
identification methods.
The input-output relationship of financial engineering fundamentals on national market brings in
spontaneous emergence of order of optimum complexity from the initial featureless states.
It can be clearly demonstrated that the flexibility provided by the recursive algorithms are remarkably
broad. It is evident from the mathematical description of the all India share price index that all new
financial information is quickly understood by the market state variables and the information itself becomes
immediately incorporated in the model.
Money market has been found to exhibit certain universal characteristics in which a large number of agents
interact [261], [262]. It is obvious that movements in money market are immediate and unbiased reflection
of incoming news about future earning prospects. In the present money market model there are two groups
of players: the first group ‘fundamentalists’ follows the premise of efficient market operations, in that, they
allow the price index to follow the fundamentals. The other group, which may be called noise group do not
affect the share price and are automatically filtered out in course operations of the intelligent algorithms
that have been used. A distinguishing feature of the present approach as compared to other simulation
models [269], [270] is that we have adopted a statistical formalization [271] where fundamentals react with
price index.
The fuzzy set theory in decision making in financial management
The fuzzy set theory provides a guide to and techniques for forecasting, decision making, conclusions, and
evaluations in an environment involving uncertainty, vagueness, and impression in business, finance,
management, and socio-economic sciences [273],[274], [275], [276], [277], [278]. It encompasses
applications in case studies including stock market strategy. The fuzzy membership function µ (unit share
price) for low , medium and high and the corresponding Mumbai Stock Exchange Sensitive Index (BSE
SENSEX), µ (BSE SENSEX ) for low , medium and high are described. It has been shown that the unit
53
share price in a dynamically stable market moves along with the sensitive index [272]. The application of
fuzzy control algorithms for market management may appear to be a promising domain for further
investigation.
This study [272] is devoted to the analysis of data, an endeavour that exploits the concepts, constructs, and
mechanisms of fuzzy set theory. Data analysis involves searching for stable, meaningful, easily
interpretable patterns in databases [281], [282]. Data analysis is an immensely heterogeneous research area
that embraces techniques and ideas that stem from probability and statistics, fuzzy sets. data visualization,
databases. And so forth. In spite of such a profound diversity, the focal point is constant: to reveal patterns
that are not only meaningful but also easily comprehensible. A fuzzy set can be regarded as an elastic
constraint imposed on the elements from a universe of discourse. By admitting a certain form of elasticity
when defining concepts and introducing various notions that are encountered in every day life.
Conceptually, fuzzy sets help alleviate problem with the classification of elements of boundary nature by
allowing for a notion of membership to a category. This requires a mathematical framework to handle
common usages of terms that are not Boolean in character, but rather are indistinct, vague, and fuzzy.
Rather than allow for an element to be either in or not in a particular set, each element can be assigned a
degree of membership in the set, often scaled over the range [0,1]. A membership of zero indicates that the
element is not a member of the fuzzy set. A membership of one indicates that the element definitely
belongs to the fuzzy set. Intermediate values correspond to lesser degrees of membership.
The fuzzy set theory provides a guide to and techniques for forecasting, decision making, conclusions, and
evaluations in an environment involving uncertainty, vagueness, and impression in business, finance,
management, and socio-economic sciences [279], [280]. Traditional modeling techniques do not capture
the nature of complex systems especially when humans are involved. Fuzzy logic provides effective tools
for dealing with such systems. It encompasses applications in case studies including Time Forecasting for
Project Management, New Product Pricing, Client Financial Risk Tolerance Policy, Deviation and Potential
Problem Analysis, Inventory Control Model, Stock Market Strategy.
Fuzzy systems, including fuzzy logic and fuzzy set theory, provide a rich and meaningful addition to
standard logic. The mathematics generated by these theories is consistent, and fuzzy logic may be a
generalization of classic logic. The applications which may be generated from or adapted to fuzzy logic are
54
wide-ranging, and provide the opportunity for modeling of conditions which are inherently imprecisely
defined, despite the concerns of classical logicians. Many systems may be modeled, simulated, and even
replicated with the help of fuzzy systems, not the least of which is human reasoning itself. Large data bases
on stock market indices and equity prices are available to a researcher to analyze and to apply decision
algorithms. Financial processes are becoming more complex which leads to the increased use of
computational intelligence (CI) [272] to process the mass of data produced in this field. CI-based
approaches of fuzzy algorithms are being applied to describe the linguistic interpretation of the rules in a
fuzzy system. The transparency of CI combined with good function approximation capabilities are the main
justification for applying fuzzy systems in finance and business.
Group Method of Data Handling
The group method of data handling (GMDH) is a self-contained exposition of a cybernetic approach to
develop mathematical model of an appealing environment with the behavioral equations and latent
variables as the important supporting characters. The framework presented incorporates problems of
representations, questions of parameterisation and procedure for identification algorithm for obtaining
models from observed data. The exposition deals with an application of the cybernetic method to develop
polynomials of optimum complexity with multi-layer group method of data handling algorithm of technical
cybernetics through an observation-conjecture-modelling-validation cycle. One of the goals of technical
cybernetics has been to capture the major elements of a dynamical process under the umbrella of a formal
mathematical synthesis. The work presented by Ranjan Chaudhuri [237] formalises a specific dynamic
situation, namely the construction of a finite dimensional process for daily movement of national market
index. It has been clearly demonstrated with observed data that the flexibility of the algorithms is
remarkably broad. Indeed, it is possible to choose free variables in such a way that the entire formal
modelling process can be interpreted as a loop shaping problem, where the loops are the layers in the multi-
layer selection process, and the loop-breaking takes place at the optimum layer through a certain specific
choice of variables. It has been observed that deep-lying feedback paths exist in national market operation.
To give mathematical description of daily national index movement as a function of a set of exogenous
variables interrelated with one another through deep-lying feedback paths is a complex process. Theories
based on differential or difference equations are not adequate to describe the process. In view of this
55
difficulty, the method of modelling applied here uses a technique of self-organisation. The GMDH is found
to simulate adequately the input-output relationship of the complex process of daily market index
movement as a function of industry-dependent state variables. Money market is one of the most exciting
and sobering parts of economics. It is marked by bubbles in which speculative prices are driven up far
beyond their intrinsic values. Speculative bubbles always produce crashes and often lead to economic fear.
Market is a tool where country's resources are allocated. By proper utilisation of resources a nation can
command its economy decisively. Modelling with interacting market parameters increases our appreciation
of macroeconomic analysis. This explores the exciting world where principles of cybernetics work on the
theories of economics. The Group Method of Data Handling Algorithms (GMDH) describes modeling,
forecasting, decision support and pattern recognition of complex systems [205] – [208] and [209] – [217] .
There are processes for which it is needed to know their future or to analyze inter-relations. Economy,
climate, finance, ecology, medicine, manufacturing and military systems [218] – [230] are areas where the
GMDH method has been successfully applied. This self-organizing method is based on sorting-out of
gradually complicated models and their evaluation by external criterion on data sample. It was developed
for forecasting, extrapolation of multivariate processes, knowledge discovery and data mining, decision
making by "what-if" scenario, diagnostics and pattern recognition. Linear or non-linear, probabilistic
models or clusterizations are selected by minimal value of an external criterion. Ivakhnenko’s [229], [230]
multi-layer group method of data handling is heuristic method of self-organisation of different partial
models. This method involves the generation and comparison of all possible combinations of input output
and to select the best possible ones according to the criterion of integral square error [212].
In multi-layer group method of data handling algorithms, polynomials are used as the basic means of
investigation of complex dynamical systems. The polynomials of prediction are regression equation which
connect the current values of output with the current and/or past values of input variables. Regression
analysis allows to evaluate the coefficients of the polynomial by criterion of minimum mean square error.
Then the polynomials are treated in the same manner as are seeds n the agricultural selection, an unique
mathematical concept propagated and established by Academician A.G. Ivakhnenko and his co-workers of
the Institute of Cybernetics, Kiev, Ukraine.
56
Volterra series [231] introduced to non-linear system analysis by Wiener [232] , learning filter of Gabor,
Wilby and Woodcock [233] and the perception of Rosenblatt [234] have provided the conceptual basis for
multilayer GMDH. Astrom and Eykhoff [235] pointed out that problems may arise with the use of
Volterra series or high degree polynomial to approximate non-linear functions because of the fact that there
are many coefficients to estimate, many data are needed and the computation with the resulting large
matrices may be prohibitive. Ivakhnenko’s multi-layer GMDH algorithms are free of these problems. He
models the input output relationships of complex process using multi-layer network structure of
Rosenblatt’s perception type, who designed the model of brain’s perception.
We describe GMDH-type polynomial networks. where xi is a i-th input variable, y is an output The
GMDH-type networks are the multi-layered ones. In the second and next layers r, the size Lr of the
population defined by the number F. The generation and selection of the layers are again performed.
The new layers are created while the criterion value is decreased. In Fig. we depicted an example of
the polynomial network consisting of 3 layers. The GMDH algorithm grew for m = 5 inputs and F = 4.
57
An example of polynomial network This network is described by a set of the following polynomials:
where g1, …, g6 are the transfer function of the neurons. With the help of multi-layer GMDH algorithms Ivakhnenko obtained the polynomial description of
British economy for prediction and control. The GMDH is a computer aided self-organization of
spontaneous emergence of order of optimum complexity from the initial featureless states. One of the goals
of the theory of cybernetics has been to capture major elements of a dynamical process under the umbrella
of a formal mathematical synthesis. Analysis has been done to find relationship between different
parameters of stock market indices of different stock exchanges. But the present investigation is indeed
unique in its own nature. The work presented in here formalizes a specific dynamic situation, namely the
construction of a mathematical description for a finite dimension process of the daily all India share price
index. It has been clearly demonstrated that the flexibility provided by the multi-layer group method of
data handling algorithm is remarkably broad. In fact, it is possible to choose free variables in such a way
that the entire formal modelling process can be reinterpreted not as a least square error minimization
problem but as a “loop shaping” problem where the loops are the layers in a multi-layer selection process
and the loop-breaking takes place at the optimal layer via certain specific choice of variables. Thus it may
be justifiably asserted that LQG/GMDH is a practical comparison of LQG/LTR [236] . It is evident from
the mathematical description of the all India share price index that all new financial information is quickly
58
understood by the market state variables and the information itself becomes immediately incorporated in
the model. The operating principles of an efficient market holds that market prices contain all available
information. The price movement in organized sector follows a definite pattern. Thus it can be safely
concluded that Indian market is in an efficient self-monitoring equilibrium state.
marketing applications motivate a specialized model and its adaptive control [238]. Each of r control
variables is set in each of a sequence of time periods. The process being controlled has a response (profit)
function that is the sum of a constant plus linear and quadratic forms in the control variables. The
coefficients of the quadratic form are assumed to be known constants, those of the linear form to change
with time as autoregressive process. Information about the changing coefficients is detected by performing
experiment on a sub-portion of the process being controlled. Provision is made for adding further
information from unspecified sources. Bayesian methods update the distributions of the unknown
coefficients. Dynamic programming determines the values of the control variables and experimental design
parameters to maximize the sum of discounted future profits. The probabilistic assumptions of the model
are chosen so that all distributions are normal with known variances and, for the most part, zero
covariances between variables. Partly as a result of this, optimal control turns out to involve rather simple
exponential smoothing rules.
Optimal Adaptive Control: A Multivariate Model for Marketing Applications
Marketing offers rich possibilities for adaptive control applications with its many decision variables,
reasonably clear objective functions, and dynamic environments. Measurement techniques, though
imprecise by engineering standards, offer actionable information and, by their very imprecision, challenge
the development of new theory. On the other hand, marketing is relatively lean on good descriptive models
of the processes over which control is sought and the implementation of model-based systems of substantial
complication poses formidable organizational, educational, and managerial obstacles. However, the high
importance placed on effective marketing by all companies ensures that new techniques, if truly productive,
are likely to be adopted in time. To give an example of a marketing application, consider a manufacturer
setting an advertising budget for a consumer product. He is never precisely certain how sales will respond
to advertising because response is difficult to measure and changes with time. The changes arise from
various disturbing influences; for example, competitors may introduce new products, alter prices, run
59
promotions, and generally stir the market up in ways that affect advertising response. In addition, the
advertising messages themselves may grow stale or consumer tastes may drift away from the product. Yet
the measurement of sales response to advertising is not impossible, only imprecise. Such measurements can
also form the basis of adaptive control of advertising spending. This possibility has been proposed and
analyzed by Little [239]. In at least one company the proposal has become a reality and an adaptive system
has been used to set annual advertising budgets for a product line. The model and adaptive system of [239],
however, control a single variable, whereas most marketing situations are bursting with manipulable
quantities. Therefore, a multivariate model and control system with wider applicability are sought. A
number of authors, for example, Tse and Athans [240], Deshpanda et al. [241], and Upadhyay and Lainiotis
[242] have presented approaches to optimal adaptive control of stochastic systems. Relative to their work,
the work of Little [238] presents model and control process having a rather specific structure motivated by
the class of problems addressed. As a result, a quite exact analysis is possible and, in addition, the required
calculations are simple. If the measurements yield useful data in their own right and if a meaningful
criterion function can be worked out, then an adaptive system in some more or less recognizable form has a
chance of being sold to management as a guide to setting the control variables. If then the implemented
system recommends definite changes in the control variables and these changes seem correct and not trivial
or quixotic, the system may stay in place and even grow in scope.
Macroeconomic Modeling with Modern Control Techniques
Macroeconomic Modeling [243] in the least complex structural form makes it endowable for the
application of modern control techniques. The method employed is a blend of the “black box’’ approach
where no knowledge of the inner mechanism is assumed and the classical approach of econometrics where
economic theory is used to determine structure. The technique is illustrated in detail by means of a
numerical example consisting of a four-equation sector of a larger model in which a full treatment of error
is given. This serves to reveal to engineers the difficulties of econometrics, demonstrating the many pit-
falls and acute problems that prevent a straightforward application of the methods of control theory. The
emphasis throughout has been on the development of a methodology together with the implementing of
supporting computational algorithms in the form of program packages. The objective is to make a
determined effort to establish to what extent control theory can help resolve what is acknowledged to be a
60
difficult control problem. Such an attempt is fraught with dangers; of criticism from economists on grounds
of lack of realism and from control engineers because the economy is not necessarily a causal mechanism
on which assumption all control theory is founded. The main difficulty lies in the extreme complexity of
the actual economic system. Indeed, on the contrary, a preoccupation with the “black box” approach can
lead to palpable nonsense. Some of the results given are incomplete in that they require further work from
the viewpoint of economic analysis
The modeling procedure described in [243] was designed to incorporate four concept.s. First, employment
is made of all available a priori information provided by economic theory thus limiting beforehand the
possibility of expending effort on fruitless searches for nonexistent relationships (interconnections).
Second, the basic philosophy of the “black box” approach is then applied allowing the data to decide the
exact dynamic structure. Thus, overly complex (statistically unsubstantiated) structures are automatically
eliminated. Third, diagnostics are continually employed which are designed to both reveal inadequacies and
indicate how improvements can be made. Finally, experienced judgment is always solicited whenever the
results of one stage must be used in a decision affecting the input to a succeeding stage. The overall
outcome of this modeling procedure is a model with the fewest required interconnections , and the least
possible order dynamic operators explaining the behavior of the existing interactions. Using a sector of a
small control model of the United Kingdom economy, the difficult task of structure determination has been
illustrated. In particular, it. was shown how a model structure determined from a univariate (single
equation) analysis can be extensively altered in the framework of a simultaneous model, requiring a re-
specification of the joint structure. However, repeated cycling through the estimation and diagnostic stages
allowed the determination of a reasonably well defined model with no more than 11 parameters, after an
economic re-specification to delete import prices from the endogenous variables. A final analysis of the
fitted model was required in order to determine its ability to forecast ahead of the sample period. Such
forecasting revealed a certain deficiency in the model with respect to the over-dominant role played by the
estimated means of the outputs, i.e., better, or more, explanatory variables are required. It is the considered
opinion that any attempt to investigate the applicability of control theory to the problems of
macroeconomic policy optimization needs to be an interdisciplinary effort. For the equations of the model
to have significant meaning in economic terms the importance of sympathetic collaboration with
61
professional economists cannot be over emphasised. On questions of basic relationships their assistance is
essential- a purely black-box approach cannot in this case be made to work. Most control engineers and
theorists are not sufficiently cognizant of the special economic issues involved. Conversely, most,
economist, or econometricians do not. fully understand the generality and unified approach to dynamic
systems afforded by control theory. Both areas have a considerable amount, of mutual interest and much
can be learned from the other.
The Parallel Computing Environment for Financial Engineering Applications
The parallel computing environment [ 283] is an object-oriented C++ library that uses abstractions to
simplify parallel programming for financial engineering applications. The message passing interface
ensures portability and performance over a wide range of parallel cluster and symmetric multiprocessing
machines.
Parallel computing has emerged as a cost-effective means of dealing with computationally intensive
financial and scientific problems. To effectively utilize this technology, developers need software that
reduces the complexity of the process as well as tools to support integration of parallel and desktop
machines. The parallel environment is a C++ library that facilitates development of large scale parallel
applications, particularly financial engineering applications.
Parallel computing provides domain-specific object-oriented libraries for solving partial/stochastic
differential equations using the finite-difference method and Monte Carlo simulation. These libraries factor
out the common operations required for FD and MC computations so that in most cases the user need only
provide the code required for the specific application. The architecture consists of three layers:
• The data abstraction and transportation core layer provides the transportation drivers that facilitate
communications among parallel processes, desktop machines, and databases via the message passing
interface (MPI), Extensible Markup Language (XML), base64 encoding, and open database connectivity
(ODBC). This layer simplifies the code associated with transferring complex data among different types of
processes and machines.
• The parallel application layer provides the domain-specific FD and MC libraries.
• The remote execution layer provides an interface between the parallel computation and the desktop
machine. This layer introduces remote parallel objects, which can encapsulate applications built using the
62
domain-specific libraries. Users can manipulate the handles of these parallel objects on a desktop machine
to access applications residing on the parallel platform. The core layer uses standard MPI primitives
Monte Carlo simulation is a numerical technique for solving problems that stochastic models describe by
generating numerous samples, commonly known as paths. Computation speed is a major barrier to
deploying MC simulations in many large and real-time applications. The MC library in parallel computing
environment facilitates the parallelization of MC applications, freeing users from dealing with these issues
while maintaining extensibility. The implementation works on heterogeneous clusters and uses dynamic
load balancing.
The MC library provides the Simulation abstract base class and related classes that encapsulate a parallel
simulation. To use the library, the user simply codes the computation of a single sample via an abstract
method. The MC library can then handle most other aspects associated with the parallelization and load
balancing of the MC simulation in a manner transparent to the programmer.
The MC library parallelizes the simulation by controlling each co-process to run sample generation, with
intermittent communications to schedule, coordinate, and load-balance the remaining simulation. The
Simulation class run method performs the parallel MC simulation in a series of computing and
synchronization steps.
Conclusion
Economic models are outcome of analysis and synthesis of financial engineering. Economic models are
employed for a variety of purposes; forecasting is an important application, control of the economy is
another. So in constructing an economic model it is important to know what the model is to be used for and
what the economist wants from it. The validity of the. model will require it to be readily admitted to an
economic analysis. Thus in setting up the equations it is necessary to express them in structural form with
which economists are familiar. There are some features of the economic environment that distinguishes this
particular field from others. The very imprecision of economic theory is exasperating; there is no
counterpart to the exactitude of the laws governing the behavior of electrical circuits, for example. The
nearest approach in economic theory would be a statement, such as “voltage affects or is affected by
current.” The nearest to an economic law is the statement that “demand” equals “supply;” rather as
Kirchhoff’s law tells us that, the sum of currents at a node is zero. There is even doubt about the
63
assumption of causality; not only doubt as to what is “cause” and what “affect” but also due to the fact that
action often precedes the action itself due to “smart operators”. A further outstanding feature is the extreme
brevity of the records available. Occasionally a sharp change, such as for example a devaluation, provides
the system with a jolt but owing to the complex interactions of the system its effects are soon smudged out.
The point. is quite a significant one since there are good reasons to believe that the economic system is one
whose underlying parameters are not constant, but evolving with time. Thus, in dealing with a system of
this sort it would be very important to get the maximum information from the recent past of the record.
Another distinctive feature of economic systems is the very high noise level. This arises from the way in
which highly aggregated national statistics are accumulated; a further hazard is the habit of occasionally
redefining the base of such statistics so destroying the continuity of the run and leading to a difficult
smoothing problem in handling such adulterated data. With the high noise levels disturbances do propagate
themselves through the system. The interactions between equations are very pronounced indeed and
simultaneous estimation of parameters of the complete set of equations is essential. It is in this sense that
the modelling of the economy offers new possibilities to the systems expert and represents a new challenge
to his ingenuity.
The investigation of the applicability of system theory of GMDH [205] –[210] to the problems of
macroeconomic policy is an interdisciplinary effort. For the equations of the model to have significant
meaning in economic terms the importance of sympathetic collaboration with professional economists
cannot be over emphasized. On questions of basic relationships their assistance is essential. A purely black-
box approach cannot in this case be made to work. Most system theorists are not sufficiently cognizant of
the special economic issues involved. Conversely, most, economist, or econometricians do not. fully
understand the generality and unified approach to dynamic systems afforded by the system theory of
GMDH. Both areas have a considerable amount, of mutual interest and much can be learned from the other.
We have offered some speculations about the future of interdisciplinary payoffs involving system theorists
and economists working together. Our speculations, although influenced by our own work we hope that
contrary view will help clarify the possible evolution of groups composed of system theorists and
economists working together. This work , it is hoped will have at their disposal a mathematical tool, the
group method of data handling algorithm that, when used together with econometric models, could
64
substantially advance the science of economic and financial management. The paper has also showed that
the GMDH approach has been extremely useful, easy, and computationally cheap for running tradeoff
studies that lead to alternate economic (monetary and fiscal) policies. Macroeconomics is concerned with
the cont.ro1 of economic aggregates. The system dynamics, as a whole, are described by sets of high-order
equations. The mat.hematica1 models are obtained using a combination of economic theory and regression
analysis. The developed models have tried to exploit the power of state variable representations. Once the
models have been identified by GMDH methods the time-varying parameters can be estimated using fresh
information in GMDH algorithms.
This section introduces some of the ideas and methods in the identification and estimation of simultaneous
regression equation systems in econometrics. After pointing out the special features of econometric
systems, it defines the problem of identification and presents several methods for estimating the parameters
in such systems. Hopefully this investigation will be useful to research workers in related fields of
economics and who are interested in the estimation of dynamic econometric systems.
It is important. to estimate the parameters of the structural equations. The main reason is that. economic
hypotheses such are formulated in the form of structural equations. If there is any change in economic
institutions, technological relations, or behavioral patterns of the economic agents as described by the
structural equations, one is able to assess its impact only by modifying the st.ructura1 equation affected.
Economic hypotheses arc mainly qualitative in character. They help specify the important variables which
should appear in each structural equation and frequently also the signs of their coefficients but not the
magnitudes. Limited historical data are employed for the estimation of all the unknown parameters in the
system. Before studying methods for estimating the unknown parameters it is necessary to impose
restrictions on the parameters to insure their identifiability. A set of structural parameters is said to be
identifiable if there exists no other set which will give rise to the same structural equation. If there exist two
sets of values for the structural parameters from which the same equation is deduced, the structural
parameters are unidentifiable. In this case, no consistent estimator for the set of parameters exists. The
GMDH algorithms of the methods of self organization of mathematical models based on multiple normal
regressions give most reliable estimates of the unknown parameters pertaining to the struct.ura1 equation
subject to the identifiability restrictions The nonlinear structural equations as proposed for GDP prediction
65
has suggested an application of the method of instrumental variables to certain type of nonlinear equations.
The methods just mentioned for nonlinear systems have not been tried extensively. and much more
experimentation is required to study their computational problems and their sampling properties.
Economists are probably dealing with large systems. The number of unknown parameters to be estimated is
larger, and the data on which estimates can be derived are limited. Mention should also be made of the
recent interest among econometricians in the estimation of time-varying coefficients in regression models,
a topic well-treated in GMDH algorithms. Having investigated some the methods for identification and
estimation of system theory for using in econometric it is hoped that scholars in this inter-disciplinary field
will find this work interesting and appealing.
GMDH formulates the system equations, state and space constraints, and a criterion functional for an
example for a problem in economic growth, and discusses some interpretation of the underlying economic
structure. Several examples are presented to illustrate particular features of Estimation problems in
economics and to more general work in mathematical system theory of economics involving group method
of algorithms. New developments in the theory of economic growth raise a number of issues of interest to
system theorists. This paper suggests a framework which may be helpful in studying economic growth
models and gives reference to mathematica1 discussions of the principles underlying some of the
economic problems to which system theory can usefully be applied. The description one might take of the
state of an economic system is a record, at, the specified instant, of its aggregates of variables along with a
record of flows and transaction between various agents or groups within the economy. The most recent
growth models have tended to deal with one primary factor, gross domestic product whose growth is
influenced by a number of distinct variables. The state is specified by a finite-dimensional vector whose
components represent economic aggregates. In discussions of economic growth it is usually assumed that
ultimate concern attaches to the rate of increase of GDP. The performance index for an economy can be
taken to be a function of the state so that the criterion functional really depends only on the time paths for
these variables and the initial state. It may be a more satisfactory portfolio theory into the analysis would
entail treating a growth model with important stochastic components; this topic is an open research
problem. The present investigation devises computational methods that improve the analysis of economic
phenomena for helping to improve analysis of such economic indicators of economic growth as GDP. The
66
models may become valuable tools for researchers and financial analysts. The models are directed at an
important aspect of economic movements over time, so-called non-stationary data, such as a long-lasting
effect from a temporary disturbance in economic growth patterns.
References
[1] Tze Leung Lai and Samuel Po-Shing Wong, “Valuation of American Options via Basis Functions”,
IEEE Transactions on Automatic Control, Vol. 49, No. 3, March 2004 [2] S. A. Ross, “Finance,” in The New Palgrave: A Dictionary of Economics, J. Eatwell, M. Milgate, and P. Newman, Eds. New York, NY: Stockton Press, 1987, vol. 2, pp. 322–336. [3] F. Black and M. Scholes, “The pricing of options and corporate liabilities,” J. Polit. Econ., vol. 81, pp. 637–659, 1973. [4] R. C. Merton, “Theory of rational option pricing,” Bell J. Econo.Manage. Sci., vol. 4, pp. 141–181, 1973. [5] J. N. Tsitsiklis and B. Van Roy, “Optimal stopping of Markov processes: Hilbert space theory, approximation algorithms, and an application to pricing high-dimensional financial derivatives,” IEEE
Trans. Automat.Contr., vol. 44, pp. 1840–1851, Oct. 1999. [6] , “Regression methods for pricing complex American-style options,”IEEE Trans. Neural Networks, vol. 12, pp. 694–703, July 2001. [7] F. AitSahlia and T. L. Lai, “A canonical optimal stopping problem for American options and its numerical solution,” J. Computa. Finance, vol. 3, pp. 33–52, 1999. [8] M. J. Brennan and E. S. Schwartz, “The valuation of the American put option,” J. Finance, vol. 32, pp. 449–462, 1977. [9] Apostolos-Paul N. Refenes, A. Neil Burgess, and Yves Bentz . “Neural Networks in Financial Engineering: A Study in Methodology”, IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 8, NO. 6, NOVEMBER 1997, p.1222-1268
[10] G. R. Courtadon, “A more accurate finite difference approximation for the valuation of options,” J.
Finan. Quant. Anal., vol. 17, pp. 697–703, 1982. [11] L. Wu and Y. K. Kwok, “A front-fixing finite difference method for the valuation of American options,” J. Finan. Eng., vol. 6, pp. 83–97, 1997. [12] J. Cox, S. Ross, and M. Rubinstein, “Option pricing: A simplified approach,” J. Finan. Econ., vol. 7, pp. 229–263, 1979. [13] S. D. Jacka, “Optimal stopping and the American put,” Math. Finance, vol. 1, pp. 1–14, 1991. [14] P. Carr, R. Jarrow, and R. Myneni, “Alternative characterizations of American put options,” Math.
Finance, vol. 2, pp. 87–106, 1992. [15] J. Huang, M. G. Subrahmanyam, and G. G. Yu, “Pricing and hedging American options: A recursive integration method,” Rev. Finan. Stud., vol. 9, pp. 277–300, 1996. [16] N. Ju, “Pricing an American option by approximating its early exercise boundary as a multipiece
67
exponential function,” Rev. Finan. Stud., vol. 11, pp. 627–646, 1998. [17] F. AitSahlia and T. L. Lai, “Exercise boundaries and efficient approximations to American option prices and hedge parameters,” J. Computat. Finance, vol. 4, pp. 85–103, 2001. [18] J. C. Hull, Options, Futures & Other Financial Derivatives, 4th ed. Upper Saddle River, NJ: Prentice- Hall, 2000. [19] I. Karatzas, “On the pricing of the American option,” Appl. Math.Optim., vol. 17, pp. 37–60, 1988. [20] P. Boyle and S. H. Lau, “Bumping up against the barrier with the binomial method,” J. Derivatives, vol. 2, pp. 6–14, 1994. [21] F. AitSahlia, L. Imhof, and T. L. Lai, “Fast and accurate valuation of American barrier options,” J
Computat. Finance, vol. 7, pp. 129–145, 2003. [22] T. L. Lai and T. W. Lim, “Exercise regions and efficient valuation of American lookback options,” Math. Finance, vol. 14, pp. 249–269, 2002. [23] J. M. Hutchinson, A. W. Lo, and T. Poggio, “A nonparameric approach to pricing and hedging derivative securities via learning networks,” J Finance, vol. 49, pp. 851–889, 1994. [24] M. Broadie, J. Detemple, E. Ghysels, and O. Torres, “Nonparametric estimation of American options’ exercise boundaries and call prices,” J.Econ. Dyna. Control, vol. 24, pp. 1829–1857, 2000. [25] M. Broadie and J. Detemple, “American option valuation: New bounds, approximations, and a comparison of existing methods,” Rev. Finan.Stud., vol. 9, pp. 1211–1250, 1996. [26] K. J. Hunt, D. Sbarbaro, R. Zbikowski, and P. J. Gawthrop, “Neural networks for control systems—A survey,” Automatica, vol. 28, pp. 1083–1112, 1992. [27] R. M. Sanner and J. J. Slotine, “Gaussian networks for direct adaptive control,” IEEE Trans. Neural
Networks, vol. 3, pp. 837–863, Nov. 1992. [28] T. J. Hastie and R. J. Tibshirani, Generalized Additive Models. London, U.K.: Chapman & Hall, 1990. [29] G. Wahba, Spline Models for Observational Data. Philadelphia, PA: SIAM, 1990. [30] A. R. Barron, “Universal approximation bounds for superpositions of a sigmoid function,” IEEE
Trans. Inform. Theory, vol. 39, pp. 930–945,May 1993. [31] J. H. Friedman and B.W. Silverman, “Flexible parsimonious smoothing and additive modeling,” Technometrics, vol. 31, pp. 3–21, 1989. [32] H. J. Kushner and P. G. Dupius, Numerical Methods for Stochastic Control Problems in Continuous
Time. New York: Springer-Verlag, 1992. [33] J. Barraquand and D. Martineau, “Numerical valuation of high dimensional multivariate American securities,” J. Finan. Quant. Anal., vol. 30, pp. 383–405, 1995. [34] S. B. Raymar and M. J. Zwecher, “Monte Carlo estimation of American call options on the maximum of several stocks,” J. Derivatives, vol. 5, pp. 7–23, 1997. [35] M. Broadie and P. Glasserman, “Pricing American-style securities using simulation,” J. Econ. Dyna.
Control, vol. 21, pp. 1323–1352, 1997.
68
[36] D. P. Bertsekas and J. N. Tsitsiklis, Neuro-Dynamic Programming. Belmont, MA: Athena Scientific, 1996. [37] F. A. Longstaff and E. S. Schwartz, “Valuing American options by simulation: A simple least-squares approach,” Rev. Finan. Stud., vol. 14, pp. 113–147, 2001. [38] M. Haugh and L. Kogan, “Pricing American options: A duality approach,” Oper. Res., 2004 [39] L. C. G. Rogers, “Monte Carlo valuation of American options,” Math.Finance, vol. 12, pp. 271–286, 2002. [40] M. H. A. Davis and I. Karatzas, “A deterministic approach to optimal stopping, with applications,” in Probability, Statistics and Optimization , F. P. Kelly, Ed. New York, NY: Wiley, 1994, pp. 455–466. [41] T. L. Lai and Q. Zhang, “Pricing Complex American Options by Simulation and Regression,” Dept. Statistics, Stanford University, Technical Rep., 2003. [42] P. Billingsley, Convergence of Probability Measures. New York: Wiley, 1968. [43] E. Clément, D. Lamberton, and P. Protter, “An analysis of a least squares regression method for American option pricing,” Finance Stoch., vol. 6, pp. 449–471, 2002. [44] C. J. Stone, “The use of polynomial splines and their tensor products in multivariate function estimation,” Ann. Statist., vol. 22, pp. 118–184, 1994. [45] G. Yin, Fellow and Xun Yu Zhou, “ Markowitz’s Mean-Variance Portfolio Selection With Regime Switching: From Discrete-Time Models to Their Continuous-Time Limits”, IEEE Transactions on Automatic control, vol. 49, No. 3, March 2004,p.349-360 [46] Harry M. Markowitz, “Foundations of Portfolio Theory” Nobel Lecture, December 7, 1990 [47] G. Barone-Adesi and R. Whaley, “Efficient analytic approximation of American option values,” J.
Finance, vol. 42, pp. 301–320, 1987. [48] J. Buffington and R. J. Elliott, “American options with regime switching,” Int. J. Theoret. Appl.
Finance, vol. 5, pp. 497–514, 2002. [49] S. Chen, X. Li, and X. Y. Zhou, “Stochastic linear quadratic regulators with indefinite control weight costs,” in SIAM J. Control Optim., vol. 36, 1998, pp. 1685–1702. [50] G. B. Di Masi, Y. M. Kabanov, and W. J. Runggaldier, “Mean variance hedging of options on stocks with Markov volatility,” Theory Probab.Appl., vol. 39, pp. 173–181, 1994. [51] D. Duffie and H. Richardson, “Mean-variance hedging in continuous time,” Ann. Appl. Probab., vol. 1, pp. 1–15, 1991. [52] E. J. Elton and M. J. Gruber, Finance as a Dynamic Process. Upper Saddle River, NJ: Prentice-Hall, 1975. [53] J. C. Francis, Investments: Analysis and Management. New York: Mc-Graw-Hill, 1976. [54] R. R. Grauer and N. H. Hakansson, “On the use of mean-variance and quadratic approximations in implementing dynamic investment strategies: A comparison of returns and investment policies,” Manage. Sci., vol. 39, pp. 856–871, 1993.
69
[55] N. H. Hakansson, “Multi-period mean-variance analysis: Toward a general theory of portfolio choice,” J. Finance, vol. 26, pp. 857–884, 1971. [56] D. Li and W. L. Ng, “Optimal dynamic portfolio selection: Multi-period mean-variance formulation,” Math. Finance, vol. 10, pp. 387–406, 2000. [57] H. Markowitz, “Portfolio selection,” J. Finance, vol. 7, pp. 77–91, 1952. [58] , Portfolio Selection: Efficient Diversification of Investment. New York: Wiley, 1959. [59] R. C. Merton, “An intertemporal capital asset pricing model,” Econometrica, vol. 41, pp. 867–888, 1973. [60] J. Mossin, “Optimal multiperiod portfolio policies,” J. Business, vol. 41, pp. 215–229, 1968. [61] S. R. Pliska, Introduction to Mathematical Finance: Discrete TimeModels. Oxford, U.K.: Blackwell, 1997. [62] P. A. Samuelson, “Lifetime portfolio selection by dynamic stochastic programming,” Rev. Econ.
Statist., vol. 51, pp. 239–246, 1969. [63] M. Schweizer, “Approximation pricing and the variance-optimal martingale measure,” Ann. Probab., vol. 24, pp. 206–236, 1996. [64] G. Yin and Q. Zhang, Continuous-Time Markov Chains and Applications:A Singular Perturbation
Approach. New York: Springer-Verlag, 1998. [65] G. Yin, R. H. Liu, and Q. Zhang, “Recursive algorithms for stock Liquidation: A stochastic optimization approach,” SIAM J. Optim., vol. 13, pp. 240–263, 2002. [66] G. Yin, Q. Zhang, and G. Badowski, “Asymptotic properties of a singularly perturbed Markov chain with inclusion of transient states,” Ann.Appl. Probab., vol. 10, pp. 549–572, 2000. [67] J. Yong and X. Y. Zhou, Stochastic Controls: Hamiltonian Systems andHJB Equations. New York: Springer-Verlag, 1999. [68] T. Zariphopoulou, “Investment-consumption models with transactions costs and Markov-chain parameters,” SIAM J. Control Optim., vol. 30, pp. 613–636, 1992. [69] Q. Zhang, “Stock trading: An optimal selling rule,” SIAM J. ControlOptim., vol. 40, pp. 64–87, 2001. [70] X. Y. Zhou and D. Li, “Continuous-Time mean-variance portfolio selection: A stochastic LQ framework,” Appl. Math. Optim., vol. 42, pp. 19–33, 2000. [71] X. Y. Zhou and G. Yin, “Markowitz’s mean-variance portfolio selection with regime switching: A continuous-time model,” SIAM J. ControlOptim., vol. 42, pp. 1466–1482, 2003. [72]. A.-P. N. Refenes, “Methods for optimal network design,” in A.-P. N. Refenes, Ed., Neural Networks
in the Capital Markets. Chichester, U.K.: Wiley, 1995, pp. 33–54.
[73]. H. M. Markowitz, “The utility of wealth”, JPE, 60, April 1952. [74] Tomasz R. Bielecki and Stanley R. Pliska, “Risk-Sensitive ICAPM With Application to Fixed-Income Management”, IEEE Transactions on Automatic Control, VOL. 49, NO. 3, March 2004, p. 420-432.
70
[75] R. C. Merton, “Optimum consumption and portfolio rules in a continuous time model,” J. Econ.
Theory, vol. 3, pp. 373–413, 1971. [76] , “An intertemporal capital asset pricing model,” Econometrica, vol. 41, pp. 866–887, 1973. [77] , Methods of Mathematical Finance. New York: Springer-Verlag, 1998. [78] R. Korn, Optimal Portfolios. Stochastic Models for Optimal Investment and Risk Management in
Continuous Time. Singapore: World Scientific, 1997. [79] D. Duffie and R. Kan, “A yield-factor model of interest rates,” Math.Finance, vol. 6, pp. 379–406, 1996. [80] D. Heath, R. Jarrow, and A. Morton, “Bond pricing and the term structure of interest rates: A new methodology for contingent claims valuation,” Econometrica, vol. 60, pp. 77–105, 1992. [81] O. A. Vasicek, “An equilibrium characterization of the term structure,”J. Financial Econ., vol. 5, pp. 177–188, 1977 [82] E. Canestrelli, “Applications of multidimensional stochastic processes to financial investments,” Rend.
Studi Econ. Quant., pp. 54–72, 1999. [83] E. Canestrelli and S. Pontini, “Inquiries on the applications of multidimensional stochastic processes to financial investments,” Econ. Complex., vol. 2, pp. 44–62, 1999. [84] T. S. Kim and E. Omberg, “Dynamic nonmyopic portfolio behavior,” Rev. Finan. Stud., vol. 9, pp. 141–161, 1996. [85] J. B. Detemple, R. Garcia, and M. Rindisbacher, “AMonte Carlo method for optimal portfolios,” J.
Finance, vol. 58, pp. 401–446, 2003. [86] M. Schroder and C. Skiadas, “Optimal consumption and portfolio selection with stochastic differential utility,” J. Econ. Theory, vol. 86, pp. 68–126, 2002. [87] M. J. Brennan and E. S. Schwartz, “The use of treasury bill futures in strategic asset allocation programs,” in World-Wide Asset and LiabilityModeling, J. M. Mulvey and W. T. Ziemba, Eds. Cambridge, U.K.: Cambridge Univ. Press, 1996. [88] M. J. Brennan, E. S. Schwartz, and R. Lagnado, “Strategic asset allocation,” J. Econ. Dyna. Control, vol. 21, pp. 1377–1403, 1997. [89] I. Bajeux-Besnainou, J. V. Jordan, and R. Portait, “Dynamic allocation of stocks, bonds, and cash for long horizons,” J. Business, vol. 76, pp. 263–288, 2003. [90] G. Deelstra, M. Grasselli, and P.-F. Koehl, “Optimal investment strategies in a CIR framework,” J.
Appl. Probab., vol. 37, pp. 936–946, 2000. [91] J. Liu, “Portfolio selection in stochastic environments,” Ph.D. dissertation, Stanford Univ., Stanford, CA, 1999. [92] C. Sørensen, “Dynamic asset allocation and fixed-income management,” J. Financial Quant. Anal., vol. 34, pp. 513–531, 2002. [93] M. Rutkowski, “Self-financing trading strategies for sliding, rolling-horizon, and consol bonds,” Math.
Finance, vol. 9, pp. 361–385, 1999.
71
[94] Olga Bobrovnytska and Martin Schweizer, “Mean-Variance Hedging and Stochastic Control:Beyond the Brownian Setting”, IEEE Transactions on Automatic Control, Vol. 49, NOo. 3, March 2004, p.396-408. [95] M. C. Steinbach, “Markowitz revisited: Mean-variance models in financial portfolio analysis,” SIAM
Rev., vol. 43, pp. 31–85, 2001. [96] M. Kohlmann and X. Y. Zhou, “Relationship between backward stochastic differential equations and stochastic controls: A linear-quadratic approach,” SIAM J. Control Optim., vol. 38, pp. 1392–1407, 2000. [97] A. E. B. Lim and X. Y. Zhou, “LQ control and mean-variance portfolio selections: The stochastic parameter case,” in Mathematical Finance. Workshop of the Mathematical Finance Research
Project,M. Kohlmann and S. Tang, Eds. Boston, MA: Birkhäuser, 2001, pp. 261–270. [98] M. Kohlmann and S. Tang, “Global adapted solution of one-dimensional backward stochastic Riccati equations, with application to the meanvariance hedging,” Stoch. Processes Applicat., vol. 97, pp. 255–288, 2002. [99] Y. Hu and X. Y. Zhou, “Indefinite stochastic Riccati equations,” SIAM J. Control Optim., vol. 42, pp. 123–137, 2003. [100] , “Minimization of risk and linear quadratic optimal control theory,” SIAM J. Control Optim., vol. 42, pp. 1118–1142, 2003. [101] S. Tang, “General linear quadratic optimal stochastic control problems with random coefficients: Linear stochastic Hamilton systems and backward stochastic Riccati equations,” SIAM J. Control
Optim., vol. 42, pp. 53–75, 2003. [102] , “Multidimensional backward stochastic Riccati equations and applications,” SIAM J. Control
Optim., vol. 41, pp. 1696–1721, 2003. [103] H. Pham, “On quadratic hedging in continuous time,” Math. Meth. Oper.Res., vol. 51, pp. 315–339, 2000. [104] , “A guided tour through quadratic hedging approaches,” in Option Pricing, Interest Rates and Risk
Management, E. Jouini, J. Cvitanic´, and M. Musiela, Eds. Cambridge, U.K.: Cambridge Univ. Press, 2001, pp. 538–574. [105] , “Backward stochastic PDE and imperfect hedging,” Int. J. Theore. Applied Finance, vol. 6, pp. 663–692, 2003. [106] Yong Zeng, “Estimating Stochastic Volatility via Filtering for the Micromovement of Asset Prices”, IEEE Transactions on Automatic Control, vol.. 49, no. 3, March 2004, p.338-348. [107] T. Bollerslev, R. Chou, and K. Kroner, “ARCH modeling in finance: A review of the theory and empirical evidencs,” J. Econometrics, vol. 52, pp. 5–59, 1992. [108] G. Gottlieb and A. Kalay, “Implications of the discreteness of observed stock prices,” J. Finance, vol. 40, pp. 135–153, 1985. [109] C. Ball, “Estimation bias induced by discrete security prices,” J. Finance, vol. 43, pp. 841–865, 1988. [110] D. C. Cho and E. W. Frees, “Estimating the volatility of discrete stock prices,” J. Finance, vol. 43, pp. 451–466, 1988.
72
[111] L. Harris, “Estimation of stock price variances and serial covariances from discrete observations,” J.
Finan. Quantitat. Anal., vol. 25, pp.291–306, 1990. [112] Y. Zeng, “A partially observed model for micromovement of asset prices with Bayes estimation via filtering,” Math. Finance, vol. 13, pp. 411–444, 2003. [113] L. Harris, “Stock price clustering and discreteness,” Rev. Finan. Stud.,vol. 4, pp. 389–415, 1991. [114] P. Del Moral, J. Jacod, and P. Protter, “The Monte-Carlo method for filtering with discrete time observations,” Probab. Theory Related Fields, vol. 120, pp. 346–368, 2001. [115] Floyd B. Hanson, and John J. Westman, “Optimal Portfolio and Consumption Policies Subject to Rishel’s Important Jump Events Model:Computational Methods”, ”, IEEE Transactions on Automatic Control, Vol. 49, NOo. 3, March 2004 [116] R. Rishel, “Modeling and portfolio optimization for stock prices dependent on external events,” in Proc. 38th IEEE Conf. Decision Control,1999, pp. 2788–2793 [117] F. Black and M. Scholes, “The pricing of options and corporate liabilities,” J. Polit. Econ., vol. 81, pp. 637–659, 1973. [118] H. J. Kushner, Stochastic Stability and Control. New York: Academic, 1967. [12] R. C. Merton, “Lifetime portfolio selection under uncertainty: The continuous-time case,” Rev. Econ. Statist., vol. 51, pp. 247–257, 1969. [119] Mertin , “Optimum consumption and portfolio rules in a continuous-time model,” J. Econ. Theory, vol. 3, pp. 373–413, 1971. [120] , Continuous-Time Finance. Cambridge, MA: Basil Blackwell, 1990. [121] , “Option pricing when underlying stock returns are discontinuous,”J. Finan. Econ., vol. 3, pp. 125– 144, 1976. [122] Weekly interest Rrates: Jan 1, 1994 to most recent, Federal Reserve System. (2000). [Online]. Available: http://www.frbchi.org/econinfo/finance/int-rates/rates/cw.prn [123] I. Karatzas, J. P. Lehoczky, S. P. Sethi, and S. E. Shreve, “Explicit solution of a general consumption/investment problem,” Math. Oper. Res., vol. 11, pp. 261–294, 1986. [124] S. P. Sethi and M. Taksar, “A note on Merton’s optimum consumption and portfolio rules in a continuous-time model,” J. Econ. Theory, vol. 46, pp. 395–401, 1988. 125] F. B. Hanson and D. Ryan, “Optimal harvesting with both population and price dynamics,” Math.
Biosci., vol. 148, pp. 129–146, 1998. [126] F. B. Hanson and H. C. Tuckwell, “Population growth with randomly distributed jumps,” J. Math.
Biol., vol. 36, pp. 169–187, 1997. [127] J. J. Westman and F. B. Hanson, “The LQGP problem: A manufacturing application,” in Proc. 1997
Amer. Control Conf., vol. 1, 1997, pp. 566–570. [128] , “State dependent jump models in optimal control,” in Proc. 38
th IEEE Conf. Decision Control,
1999, pp. 2378–2383.
73
[129] , “Nonlinear state dynamics: Computational methods and manufacturing example,” Int. J. Control, vol. 73, pp. 464–480, 2000. [130] K. Itô, “On stochastic differential equations,” Mem. Amer. Math. Soc., no. 4, pp. 1–51, 1972 [131] C. Moler et al., Using MATLAB, Version 5. Natick, MA: Mathworks, 1999. [132] Tomasz R. Bielecki and Marek Rutkowski, “Modeling of the Defaultable Term Structure:Conditionally Markov Approach”,
[133] A. Arvanitis, J. Gregory, and J.-P. Laurent, “Building models for credit spreads,” J. Derivatives, vol. 6, no. 3, pp. 27–43, 1999. [134] T. Bielecki and M. Rutkowski, “Multiple ratings model of defaultable term structure,” Math.
Finance, vol. 10, pp. 125–139, 2000a. [135] , “HJM with multiples,” Risk, vol. 13, no. 4, pp. 95–97, 2000b. [145] T. Bielecki and M. Rutkowski, Credit Risk: Modeling, Valuation, and Hedging. New York: Springer- Verlag, 2002. [136] D. Duffie, “Forward rate curves with default risk,” in Working Paper. Stanford, CA: Stanford Univ., 1994 . [137] , “Defaultable term structure models with fractional recovery of par,” in Working Paper. Stanford, CA: Stanford Univ., 1998. [138] D. Duffie and K. Singleton, “Ratings-based term structures of credit spreads,” Stanford Univ., Stanford, CA, 1998. [139] , “Modeling term structures of defaultable bonds,” Rev. Finan. Stud., vol. 12, pp. 687–720, 1999. [140] R. A. Jarrow, D. Lando, and S. M. Turnbull, “A Markov model for the term structure of credit risk spreads,” Rev. Finan. Stud., vol. 10, pp. 481–523, 1997. [141] R. A. Jarrow and S. M. Turnbull, “Pricing derivatives on financial securities subject to credit risk,” J.
Finance, vol. 50, pp. 53–85, 1995. [142] D. Lando, “On Cox processes and credit-risky securities,” Rev. Derivatives Res., vol. 2, pp. 99–120, 1998. [143] P. J. Schönbucher, “Term structure modeling of defaultable bonds,” Rev. Derivatives Res., vol. 2, pp. 161–192, 1998. [144] E. Eberlein and S. Raible, “Term structure models driven by general Lévy processes,” Math.
Finance, vol. 9, pp. 31–53, 1999. [145] Paolo Dai Pra, Wolfgang J. Runggaldier, and Marco Tolotti, “Pathwise Optimality for Benchmark Tracking’, IEEE Transactions on Automatic Control, Vol. 49, NOo. 3, March 2004 [146] S. Browne, “Beating a moving target: Optimal portfolio strategies for outperforming a stochastic benchmark,” Finance Stoch., vol. 3, pp. 275–294, 1999. [147] È. L. Presman, “Optimality almost surely and in probability for a stochastic linear-quadratic regulator,” Theory Probab. Applicat., vol. 42, pp. 531–535, 1997.
74
[148] È. L. Presman, V. I. Rotar, and M. Taksar, “Optimality in probability and almost surely. The general scheme and a linear quadratic regulator problem,” Stoch. Stoch. Rep., vol. 43, pp. 127–137, 1993. [149] V. I. Rotar, “Connectivity property and optimality almost surely and in probability,” in Proc. New
Trends in Probability and Statistics, Utrecht, The Netherlands, 1991, pp. 528–539. [150] P. D. Pra, G. B. Di Masi, and B. Trivellato, “Pathwise optimality in stochastic control,” SIAM J.
Control Optim., vol. 39, no. 5, pp. 1540–1557, 2000. [151] R. Merton, “Optimum consumption and portfolio rules in a continuous time model,” J. Econ. Theory, vol. 3, pp. 373–413, 1971. [152] Kang Boda, Jerzy A. Filar, Yuanlie Lin, and Lieneke Spanjers, “Stochastic Target Hitting Time and the Problem of Early Retirement”, IEEE Transactions on Automatic Control, Vol. 49, NOo. 3, March 2004. [153] M. L. Puterman, Markov Decision Processes: Discrete Stochastic Dynamic Programming. New York, NY: Wiley, 1994. [154] R. A. Howard and J. E. Matheson, “Risk-sensitive Markov decision processes,” Manage. Sci., vol. 18, pp. 356–369, 1972. [155] C. Wu and Y. Lin, “Minimizing risk models in Markov decision processes with policies depending on target values,” J. Math. Anal. Appl.vol. 231, pp. 47–67, 1999. [156] D. J. White, “Mean, variance, and probabilistic criteria in finite Markov decision processes: A review,” J. Optim. Theory Appl., vol. 56, pp. 1–29, 1988. [157] J. A. Filar, L. C. M. Kallenberg, and H.-M. Lee, “Variance-penalized Markov decision processes,” Math. Oper. Res., vol. 14, pp. 147–161, 1989. [158] J. A. Filar, D. Krass, and K. W. Ross, “Percentile performance criteria for limiting average Markov decision processes,” IEEE Trans. Automat.Contr., vol. 40, pp. 2–10, Jan. 1995. [159] M. Bouakiz andY.Kebir, “Target-level criterion in Markov decision processes,” J. Optim. Theory
Appl., vol. 86, pp. 1–15, 1995. [160] R. Cavazos-Cadena and E. Fernández-Gaucherand, “Controlled Markov chains with risk-sensitive criteria: Average cost, optimality equations, and optimal solutions,” Math. Meth. Oper. Res., vol. 49, pp. 299–324, 1999. [161] M. J. Sobel, “The variance of discounted MDPs,” J. Appl. Probab., vol.19, pp. 794–802, 1982. [162] S. X. Yu, Y. Lin, and P. Yan, “Optimization models for the first arrival target distribution function in discrete time,” J. Math. Anal. Appl., vol. 225, pp. 193–223, 1998. [163] S. C. Jaquette, “A utility criterion for Markov decision processes,” Manage. Sci., vol. 23, pp. 43–49, 1976. [164] P. Whittle, Risk-Sensitive Optimal Control. New York: Wiley, 1990. [165] , “Minimizing a threshold probability in discounted Markov decision processes,” J. Math. Anal.
Applicat., vol. 173, pp. 634–646, 1993. [166] H. Markowitz, Portfolio Selection: Efficient Diversification of Investment. New York: Wiley, 1959.
75
[167] A. D. Roy, “Safety first and the holiday of assets,” Econometrica, vol. 220, pp. 431–449, 1996. [168] R. C. Merton, Continuous-Time Finance. Oxford, U.K.: Basic Blackwell, 1990. [169] S. Roy, “Theory of dynamic portfolio for survival under uncertainty,” Math. Social Sci., vol. 30, pp. 171–194, 1995. [170] R. E. Wildeman, R. Dekker, and A. C. J. M. Smit, “A dynamic policy for grouping maintenance activities,” Eur. J. Oper. Res., vol. 99, pp. 530–551, 1997. [171] Bensoussan, “Remarks on the Pricing of Contingent Claims Under Constraints, ”, IEEE Transactions on Automatic Control, Vol. 49, NOo. 3, March 2004. [172] Nicole Bäuerle andUlrich Rieder, “Portfolio Optimization With Markov-Modulated Stock Prices and Interest Rates”, IEEE Transactions on Automatic Control, Vol. 49, NOo. 3, March 2004. [173] Q. Zhang, “Stock trading: An optimal selling rule,” SIAM J. ControlOptim., vol. 40, pp. 64–87, 2001. [174] R. Frey andW . J. Runggaldier, “A nonlinear filtering approach to volatility estimation with a view towardhigh frequency data,” Int. J. Theoret. Appl. Finance, vol. 4, pp. 199–210, 2001. [175] R. C. Merton, “Optimum consumption andportfolio rules in a continuous-time model,” J. Econ.
Theo., pp. 373–413, 1971. [176] W. H. Fleming andD. Hern´ndez-Hern´ndez, “An optimal consumption model with stochastic volatility,” Finance Stoch., vol. 7, pp. 245–262. [177] T. Zariphopoulou, “Optimal investment andconsumption models with nonlinear stock dynamics,” Math. Meth. Oper. Res., vol. 50, pp. 271–296, 1999. [178] , “A solution approach to valuation with unhedgeable risks,” FinanceStoch., vol. 5, pp. 61–82, 2001. [179] J. P. Fouque, G. Papanicolaou, andK. R. Sircar, Derivatives in FinancialMarkest with Stochastic
Volatility. Cambridge, U.K.: Cambridge Univ.Press, 2000. [180] S. Heston, “A closed-form solution for options with stochastic volatility with applications to bondand currency options,” Rev. Financial Stud., vol. 6, pp. 327–343, 1993. [181] G. B. Di Masi, Y. M. Kabanov, andW. J. Runggaldier, “Mean-variance hedging of options on stocks with Markov volatilities,” Theory Probab.Appl., vol. 39, pp. 211–222, 1994. [182] H. J. Kushner andP. G. Dupuis, Numerical Methods for Stochastic ControlProblems in Continuous
Time. New York: Springer-Verlag, 1992. [183] A. Lioui andP . Poncet, “On optimal portfolio choice under stochastic interest rates,” J. Econ. Dyna.
Control, vol. 25, pp. 1841–1865, 2001. [184] R. Korn andH. Kraft, “A stochastic control approach to portfolio problems with stochastic interest rates,” SIAM J. Control Optim., vol. 40, pp. 1250–1269, 2001. [185] W. H. Fleming andR. W. Rishel, Deterministic and Stochastic OptimalControl. New York: Springer- Verlag, 1975. [186] W. H. Fleming andH. M. Soner, Controlled Markov Processes and ViscositySolutions. New York: Springer-Verlag, 1993.
76
[187] S. Browne, “Reaching goals by deadline: Digital options and continuous-time active portfolio management,” Adv. Appl. Probab., vol. 31, pp. 551–577, 1999. [188] H. Föllmer andP . Leukert, “Quantile hedging,” Finance Stoch., vol. 3, pp. 251–273, 1999. [189] Shu-Shang Zhu, Duan Li, and Shou-Yang Wang, “Risk Control Over Bankruptcy in Dynamic Portfolio Selection:A Generalized Mean-Variance Formulation”, IEEE Transactions on Automatic Control, Vol. 49, NOo. 3, March 2004. [190] B. Dumas and E. Lucinao, “An exact solution to a dynamic portfolio choice problem under transaction costs,” J. Finance, vol. 46, pp. 577–595, 1991. [191] E. J. Elton and M. J. Gruber, “On the optimality of some multiperiod portfolio selection criteria,” J.
Business, vol. 47, pp. 231–243, 1974. [192] E. F. Fama, “Multiperiod consumption-investment decisions,” Amer.Econ. Rev., vol. 60, pp. 163– 174, 1970. [193] R. R. Grauer and N. H. Hakansson, “On the use of mean-variance and quadratic approximations in implementing dynamic investment strategies: A comparison of returns and investment policies,” Manage. Sci., vol. 39, pp. 856–871, 1993. [194] N. H. Hakansson, “On optimal myopic portfolio policies, with and without serial correlation of yields,” J. Business, vol. 44, pp. 324–334, 1971. [195] R. C. Merton, “Lifetime portfolio selection under uncertainty: The continuous-time case,” Rev.
Econ. Statist., vol. 51, pp. 247–257, 1969. [196] , Continuous-Time Finance. Cambridge, MA: Basil Blackwell, 1990. [197] J. Mossin, “Optimal multiperiod portfolio policies,” J. Business, vol. 41, pp. 215–229, 1968. [198] R. Östermark, “Vector forecasting and dynamic portfolio selection: Empirical efficiency of recursive multiperiod strategies,” Eur. J. Oper. Res., vol. 55, pp. 46–56, 1991. [199] S. R. Pliska, Introduction to Mathematical Finance. Cambridge, MA: Basil Blackwell, 1997. [200] P. A. Samuelson, “Lifetime portfolio selection by dynamic stochastic programming,” Rev. Econ.
Statist., vol. 50, pp. 239–246, 1969. [201] H. M. Markowitz, Portfolio Selection: Efficient Diversification of Investment. New York: Wiley, 1959. [202] D. Li and W. L. Ng, “Optimal dynamic portfolio selection: Multiperiod mean variance formulation,” Math. Finance, vol. 10, pp. 387–406, 2000. [203] X. Y. Zhou and D. Li, “Continuous time mean-variance portfolio selection: A stochastic LQ framework,” Appl. Math. Optim., vol. 42, pp. 19–33, 2000. [204] Lukasz Stettner, “Risk-Sensitive Portfolio Optimization With Completely and Partially Observed Factors”, IEEE Transactions on Automatic Control, Vol. 49, NOo. 3, March 2004. [205] Iakhnenko,A.G., and Osipenko,V.V. Algorithms of Transformation of Probability Characteristics into Deterministic Forecast. Sov. J. of Automation and Information Sciences, 1982, vol.15, no.2, pp.7-15.
77
[206] Farlow,S.J.,(ed.) Self-organising Methods in Modeling (Statistics: Textbooks and Monographs, vol.54), Marcel Dekker Inc., New York and Basel, 1984. [207] Spashko V.S. and Kostenko Ju.V. GMDH Algorithm for Two-Level Modeling of Multivariate Cyclic Processes, Sov. J. of Automation and Information Sciences, 1987, vol.20, No. 4. [208] Stepashko V.S. Asymptotic Properties of External Criteria for Model Selection, Sov. J. of Automation and Information Sciences, 1988, vol.21, no.6, pp.84-92.
[209] Sepashko V.S. GMDH Algorithms as Basis of Modeling Process Automation after Experimental Data, Sov. J. of Automation and Information Sciences, vol.21, no.4, 1988, pp.43- 53.
[210] Caudhuri A.S., "Observation , Conjecture and Experiment with Group Method of Data Handling Algorithms ", International Journal of System Analysis, Modeling and Simulation , vol. 6 , no. 11/12 , 1989 , p.867-906, Spinger Verlog, Berlin.
[211] J. Murata and S.Sagara, “Self-organisation of models for Power System load forecasting", International Conference on System Science, Warsaw, Poland, September 19 - 22, –1989.
[212] Belogurov,V.P. “A criterion of model suitability for forecasting quantitative processes”. Sov.
J. of Automation and Information Sciences, 1990, vol.23, no.3, p.21-25 [213] Ivakhnenko,A.G. “An Inductive Sorting Method for the Forecasting of Multidimensional Random Processes and Events with the Help of Analogues Forecast Complexing”. Pattern
Recognition and Image Analysis, 1991, vol.1, no.1, pp.99-108. [214] Zholnarsky,A.A. “Agglomerative Cluster Analysis Procedures for Multidimensional Objects: A Test for Convergence”. Pattern Recognition and Image Analysis, 1992, vol.25, no.4, pp.389-390. [215] Stepashko V.S. “ Structural Identification of Predictive Models under Conditions of a
Planned Experiment”, Sov. J. of Automation and Information Sciences, 1992, vol.25, no.1, pp.24-32.
[216] Ivakhnenko,A.G., Ivakhnenko,G.A., and Müller,J.A. “ Self-Organisation of Optimum
Physical Clustering of the Data Sample for Weakened Description and Forecasting of Fuzzy Objects”. Pattern Recognition and Image Analysis, 1993, vol.3, no.4, pp.415-421.
[217] Madala,H.R. and Ivakhnenko,A.G. “Inductive Learning Algorithms for Complex Systems Modeling”. CRC Press Inc., Boca Raton, 1994. [218] Ivakhnenko,A.G., Ivakhnenko,G.A. and Müller,J.A. “Self-Organisation of Neuronets with
Active Neurons”. Pattern Recognition and Image Analysis, 1994, vol.4, no.2, pp.177-188. [219] Ivakhnenko,G.A.“ Self-Organisation of Neuronet with Active Neurons for Effects of Nuclear
Test Explosions Forecastings”. System Analysis Modeling Simulation (SAMS), vol.20, 1995, pp.107-116.
[220] Ivakhnenko, A.G., Müller, J.-A.“ Present state and new problems of further GMDH Development”. SAMS, vol.20, no.1-2, 1995, pp.3-16. [221] Müller, J.-A., Lemke,F. “ Self-Organising modelling and decision support in economics”. Proc. of IMACS'95, Gordon and Breach Publ. 1995, pp.135-138.
[222] Lemke, F. “ Self Organize! - software tool for modelling and prediction of complex systems”.
78
SAMS, vol.20, no.1-2, 1995, pp.17-28. [223] Ivakhnenko, A.G., Müller, J.-A. “ Self-organisation of nets of active neurons”. SAMS, vol.20,
no.1-2, 1995, pp.93-106.
[224] Hild, Ch.R., Bozdogan H. “ The Use of Information-Based Model Selection Criteria in the GMDH Algorithms”. SAMS, vol.20, no.1-2, 1995, pp.29-50.
[225] Lange T. “ A New Principle of Creation of Selection Criteria for Automatic Model Selection
With Uncertainty of Data”. SAMS, vol.20, no.1-2, 1995, pp.79-92. [226] Wang Xiufeng, L.B. “ Self-Organizing Methods Modelling With "Concentrated" Variables”.
SAMS, vol.20, no.1-2, 1995, pp.51-58. [227] Müller, J.-A. “ Analysis and prediction of ecological systems”. SAMS, vol.21, 1996. [228] Ivakhnenko,A.G. and Ivakhnenko,G.A. “ Simplified Linear Programming Algorithm as
Basic Tool for Open-Loop Control”. System Analysis Modeling Simulation (SAMS), 1996, vol.22, pp.177-184.
[229] Iakhnenko A.G. “Heuristic self-organisation of problems of Engineering Cybernetics”,
Automatica, vol. 6, pp. 207-219, 1970.
[230] Ivakhnko A.G., “Polynomial theory of complex systems”, Transactions of Systems, Man and Cybernetics of Institution of Electrical and Electronic Engineering, IEEE, vol. SMC 1, NO. 4, October 1971
[231] Volterra V., “Theory of functional and integral and integro-differential equations”, London, Blackie, 1930.
[232] Wener N., “Non linear problems in random theory”, New York, Willey, 1958. [233] Gabor D., Wilby W.P.L. and WoodCock R., “A Universal Non Linear Filter Predictor And Simulator which Optimises itself by a Learning Process”, Proceedings of the Institution of Electrical Engineers, London, IEE Proc., Vol. 108, Part B, pp. 422 – 433, July 1961. [234] Rosenlblatt F., “Principles of Neurodynamics”, Spartan Books, Washington, 1962.
[235] Atrom K.J. and Eykhoff P., “System Identification , A survey”, Automatica , Vol.7, pp.123–162, March, 1971.
[236] Stein and M. Athans, “The LQG/LTR Procedure for Multivariable Feedback Control Design”,
IEEE Transactions on Automatic Control, Vol. AC –32, No. 2, February 1987, p. 105 –114 [237] R. Chaudhuri, "Identification of the Interactions of Industry-dependent State Variables on
Movement of National Market Index," Journal of the Institution of Engineers (India), Interdisciplinary Division, November 2000.
[238] John D. C. Little, “Optimal Adaptive Control: A Multivariate Model for Marketing Applications”, IEEE Transactions on Automatic Control, Vol. AC-22, No. 2, April 1977, p. 187-195. [239] J. D. C. Little, “A model of adaptive control of promotional spending,” Oper. Res., vol. 14, pp. 1075-1097, Nov. 1966. [240] E. Tse and M. Athans, “Adaptive stochastic control for a class of linear systems,” IEEE Trans.
Automat. Contr., vol. AC-17, pp. 38-52, Feb. 1972.
79
[241] J. G. Deshpande et al., “Adaptive control of linear stochastic separation theorem,” Int. J. Conrr.,
vol. 15, pp. 877-888, 1972. [242] T. N. Upadhyay and D. G. Lainiotis, “Joint adaptive plant and systems,” Automatica, vol. 9, pp. 107-1 15, Jan. 1973. [243] Wall, K.and Westcott,J. “Macroeconomic modeling for control”, Automatic Control, IEEE Transactions on , Volume: 19 Issue: 6 , Dec 1974 , Page(s): 862 -873 [244] Athans, M.; Kendrick, D., “Control theory and economics: A survey, forecast, and speculations”, Automatic Control, IEEE Transactions on , Volume: 19 Issue: 5 , Oct 1974, page(s): 518 -524 [245] Chow, G., “ Identification and estimation in econometric systems: A survey”, Automatic Control, IEEE Transactions on , Volume: 19 Issue: 6 , Dec 1974, Page(s): 855 –862. [246] H. A. Simon, “Dynamic programnling under uncertainty with a quadratic criterion function,” Ecunometriea, vol. 24, Jan. 1956. [247] H. Theil, “A note on certainty equivalence in dynamic plan- ning,” Econontetricu, vol. 25, Apr. 1957. [248] C. C. Holt, F. Modigliani, J. F. Muth, and H. A. Simon, “ Planning Production, Inventories and Work Force”, Engleaood, Prentice-Hall, 1960. [249] A. R. Dobell and Y. C. Ho, “Optimal investment policy: An example of a control problem in economic theory,’’ IEEE Trans. Automat. Contr., vol. AC-12, pp. 4-14, Feb. 1967.
[250] R. S. Pindyck, “An application of the linear quadratic tracking problem to economic stabilization policy,” IEEE Trans. Automat. Contr., vol. AC-17, pp. 287-300, June 1972.
[251] P. A. Samuelson, “Lifetime portfolio selection by dynamic stochastic programming,” Rev. Econonlics Stat., vol. LI, no. 3, 1969
[252] R. Chaudhuri (2002) , Input-Output Relationship of Industrial Production Fundamentals on National Market”, International Journal of AMSE, France, Modelling D-2002, Vol. 23 nº 1
[253] Beck M.B. (1979), “System identification, estimation and forecasting of water quality Part 1 : Theory” Working paper, WP-79-31, International Institute for Applied Systems Analysis, A-2361 Laxenberg, Austria.
[254] Beck M.B. (1977), “Real time control of water quality and quantity”, International Institute of Applied Systems Analysis, Luxembourg, Austria. [255] Beck M.B. (1977), “The identification and adaptive prediction of Urban Sewer Flows”, International Journal of Control, vol. 25, No. 3, pp. 425-440. [256] Ljung L and Soderstrom T(1983), “Theory and practice of recursive identification”, MIT Press, Boston. [257] Ljung L.(1977) , “Analysis of recursive stochastic algorithms”, IEEE transaction on Automatic Control, vol. AC-22, No. 4. [258] Ljung L.(1978),“Convergence analysis of parameter identification methods”,, IEEE Trans.,vol.AC- 23,No. 5.
80
[259] Ljung L.(1977), “A positive real transfer functions and the convergence of some recursive schemes”, IEEE transaction on Automatic Control, vol. AC-22.
[260] Ljung L. (1979), “Asymptotic Behaviour of Extended Kalman Filter as a Parameter Estimator for Linear Systems”, IEEE Trans. Vol. AC-24, No. 1, pp. 36-50. [261] Lux T. (1996) , “ Long term stochastic dependence in financial prices : evidence from the German stock market” , Appl. Econ. Lett. vol 3, p.701-706. [262] Lux T. (1997), “Time variation of second moments from a noise trader/infection model”, J. Econ. Dyn. Control , vol 22, p.1-38. [263] Lux T. (1998), “ The socio-economic dynamics of speculative market : interacting agents,chaos, and the fat tails of return distribution” , J. Econ. Behav. Organizat. , vol 33 , p. 143- 165. [264] Lux T. and Marchesi M.(1999), “ Scaling and criticality in a stochastic multi-agent model of a financial b market”, Nature, vol 397, p. 498-500.
[265] Philippe Jorion and William N. Goetzmann (!999), “Global Stock Markets in the Twentieth Century,” Journal of Finance, vol.53, June 1999. [266] Nai-Fu Chen, Richard Roll and Stephen A. Ross (1986), “Economic Forces and the Stock Market,” Journal of Business, vol 59, July 1986, pp. 383-404. [267] Eugene F. Fama and Kenneth R. French (1992), “The Cross-Section of Expected Stock Returns”, Journal of Finance, vol 47, June 1992, pp. 427-465.
[268] Narasimhan Jagadeesh and Sheridan Titman (1993), “Returns to Buying Winners and Selling Losers: Implications for Stock Market Efficiency”, Journal of Finance, vol 48, March 1993, pp. 65-91. [269] Campbell R. Harvey (1998), ”Time-varying Conditional Covariences in Test of Asset Pricing Models”, Journal of Financial Economics, vol. 24, no. 2, October 1989, p. 289- 318.
[270] Levy M., Levy H. and Solommon S.(1995); “Microscopic simulation of the stock market : the effect of microscopic diversity”, J. Phys. I (France), vol 5 , p. 1087-1107.
[271] Liu Y., Cizeau P., Meyer M., Peng C. K. and Stanley H. E. (1997), “Correlations in economic time series”, Physica A, vol 245, p. 437-440. [272]. Chaudhuri, R, “ Decision Making in Stock Market Management”, Journal of the Institution of Engineers(India), vol 85, May 2004, IE( I ) Journal-ID, pages 33-37. [273]. Zadeh, L.A,"Fuzzy sets", Information and Control, Vol. 8, 1965, pp. 338-353 [274]. Zadeh, L.A, "Fuzzy Algorithms", Information and Control, Vol. 12, 1968 ,pp. 338-353. [275]. Gougen, J.A," The logic of inexact concepts", Synthese, Vol. 19, 1969, pp 325-373. [276]. Zadeh L.A, "Outline of a New Approach to the Analysis of Complex Systems and Decision
Processes", IEEE Transaction on Systems, Man and Cybernetics, Vol. SMC-3, No.1. 1973, pp. 28-34. [277]. F. Wenstop, "Deductive verbal models of organizations," Int. Jrnl. of Man-Mach. Stud., Vol. 8, 1976, pp. 293-311. [278]. Zadeh, L.A,"PRUF - A meaning representation language for natural languages", International Journal of Man-Machine Studies, Vol. 10, 1978, pp. 395-460.
81
[279]. G.J. Klir and B. Yuan, Fuzzy Sets and Fuzzy Logic: Theory and Applications , Prentice-Hall, New Jersey, 1995
[280]. George Bojadziev (Simon Fraser University) & Maria Bojadziev, Fuzzy Logic for Business, Finance, and Management, July 1997 ,ISBN 981-02-2894-5. [281]. Kaoru Hirota and Witold Pedrycz, “ Fuzzy Computing for Data Mining”, Proceedings of the IEEE , vol. 87. No. 9, September 1999.
[282]. W. Pedrycz and F. Gomide, “ An Introduction to Fuzzy Sets : Analysis and Design”, Cambridge, MA: MIT Press. 1998. [283] Monk-Ping Leong et al. CPE: A Parallel Library for Financial Engineering Applications, IEEE Computer October 2005, vol. 38 no. 10 , pp. 70-77