Handbook of the Fundamentals of Financial Decision Making - Part 1 9789814417372

This handbook in two parts covers key topics of the theory of financial decision making. Some of the papers discuss real

390 5 36MB

English Pages 940 [916] Year 2013

Report DMCA / Copyright

DOWNLOAD PDF FILE

Recommend Papers

Handbook of the Fundamentals of Financial Decision Making - Part 1
 9789814417372

  • 0 0 0
  • Like this paper and download? You can publish your own PDF file online for free in a few minutes! Sign Up
File loading please wait...
Citation preview

HANDBOOK OF

THE FUNDAMENTALS OF

FINANCIAL

DECISION MAKING Part I

World Scientific Handbook in Financial Economic Series (ISSN: 2010-1732) Series Editor:

William T. Ziemba University of British Columbia, Canada (Emeritus); ICMA Centre, University of Reading, UK; Visiting Professor at The Korean Advanced Institute for Science and Technology and Sabanci University, Istanbul

Advisory Editors: Kenneth J. Arrow Stanford University, USA George C. Constantinides University of Chicago, USA Espen Eckbo Dartmouth College, USA

Harry M. Markowitz University of California, USA Robert C. Merton Harvard University, USA Stewart C. Myers Massachusetts Institute of Technology, USA

The Handbooks in Financial Economics (HIFE) are intended to be a definitive source for comprehensive and accessible information in the field of finance. Each individual volume in the series presents an accurate self-contained survey of a sub-field of finance, suitable for use by finance, economics and financial engineering professors and lecturers, professional researchers, investments, pension fund and insurance portfolio mangers, risk managers, graduate students and as a teaching supplement. The HIFE series will broadly cover various areas of finance in a multi-handbook series. The HIFE series has its own web page that include detailed information such as the introductory chapter to each volume, an abstract of each chapter and biographies of editors. The series will be promoted by the publisher at major academic meetings and through other sources. There will be links with research articles in major journals. The goal is to have a broad group of outstanding volumes in various areas of financial economics. The evidence is that acceptance of all the books is strengthened over time and by the presence of other strong volumes. Sales, citations, royalties and recognition tend to grow over time faster than the number of volumes published.

Published Vol. 1

Stochastic Optimization Models in Finance (2006 Edition) edited by William T. Ziemba & Raymond G. Vickson

Vol. 2

Efficiency of Racetrack Betting Markets (2008 Edition) edited by Donald B. Hausch, Victor S. Y. Lo & William T. Ziemba

Vol. 3

The Kelly Capital Growth Investment Criterion: Theory and Practice edited by Leonard C. MacLean, Edward O. Thorp & William T. Ziemba

Vol. 4

Handbook of the Fundamentals of Financial Decision Making (In 2 Parts) edited by Leonard C. MacLean & William T. Ziemba

Forthcoming The World Scientific Handbook of Insurance (To be announced) The WSPC Handbook of Futures Markets edited by A. G. Malliaris & William T. Ziemba

World Scientific Handbook in Financial Economic Series — Vol. 4

HANDBOOK OF

THE FUNDAMENTALS OF

FINANCIAL

DECISION MAKING Part I

Editors

Leonard C MacLean Dalhousie University, Canada (Emeritus)

William T Ziemba University of British Columbia, Canada (Emeritus);  ICMA Centre; University of Reading, UK; Visiting Professor at The Korean Advanced Institute for Science and Technology and Sabanci University, Istanbul

World Scientiic NEW JERSEY



LONDON



SINGAPORE



BEIJING



SHANGHAI



HONG KONG



TA I P E I



CHENNAI

Published by World Scientific Publishing Co. Pte. Ltd. 5 Toh Tuck Link, Singapore 596224 USA office: 27 Warren Street, Suite 401-402, Hackensack, NJ 07601 UK office: 57 Shelton Street, Covent Garden, London WC2H 9HE

Library of Congress Cataloging-in-Publication Data Handbook of the fundamentals of financial decision making (in 2 Parts) / edited by Leonard C. MacLean and William T. Ziemba. v. cm. -- (World scientific handbook in financial economic series, ISSN 2010-1732 ; v. 4) Includes bibliographical references and index. ISBN 978-9814417341 (Set) ISBN 978-9814417372 (Part I) ISBN 978-9814417389 (Part II) 1. Investments--Decision making. 2. Finance--Decision making. 3. Risk management. 4. Uncertainty I. MacLean, L. C. (Leonard C.) II. Ziemba, W. T. HG4515.H364 2013 332--dc23 2012037307

British Library Cataloguing-in-Publication Data A catalogue record for this book is available from the British Library.

Copyright © 2013 by World Scientific Publishing Co. Pte. Ltd. All rights reserved. This book, or parts thereof, may not be reproduced in any form or by any means, electronic or mechanical, including photocopying, recording or any information storage and retrieval system now known or to be invented, without written permission from the Publisher.

For photocopying of material in this volume, please pay a copying fee through the Copyright Clearance Center, Inc., 222 Rosewood Drive, Danvers, MA 01923, USA. In this case permission to photocopy is not required from the publisher.

In-house Editor: Alisha Nguyen

Typeset by Stallion Press Email: [email protected]

Printed in Singapore.

This part is dedicated to the memory of Professor Paul Anthony Samuelson, arguably the most influential economist of the 20th century and a major contributor to the topics in this handbook. Paul was born in Gary Indiana on May 15, 1915 and died in Belmont, Massachusetts on December 13, 2009. He was Institute Professor of Economics and Gordon Y Billard Fellow at the Massachusetts Institute of Technology.

v

This page intentionally left blank

Preface

The theory and practice of decision making in finance has a long and storied history. Many of the fundamental concepts in economics as well as the models in stochastic optimization have their origins in finance. In this handbook, a selection of the key papers in financial decision making are brought together to provide a comprehensive picture of the components and methods. The handbook is composed of two parts. Part I consists of a collection of reprints of seminal papers on the foundations of decision making in finance. Part II contains mostly original papers which explore aspects of decision models. A special emphasis is placed on models which optimize capital growth. In Section A of Part I, the concept of arbitrage is presented in a set of papers. Arbitrage is an opportunity for risk free gains, and an absence of arbitrage is a condition for a stable financial market. The seminal paper is from Ross (1976), where the Arbitrage Pricing Theory (APT) is introduced. He defines the prices on assets by a linear relation to common factors (latent random variables) each with expectation zero. After accounting for risk there is no price premium, that is, arbitrage does not exist. Another way of stating this condition is that after adjusting for risk, the price process is a martingale. The Fundamental Theorem of Asset Pricing states that “no arbitrage” is equivalent to the existence of an equivalent “martingale measure.” The papers by Schachermayer (2010a, 2010b) put the fundamental theorem in a general setting. There exists a martingale measure if and only if the price process satisfies a no free lunch with vanishing risk condition. The martingale measure is a risk-neutral measure and all assets have the same expected value (the risk-free rate) under the risk-neutral measure. Kallio and Ziemba (2007) use Tucker’s theorems of the alternative from mathematical programming to establish the existence of risk neutral probabilities for the discrete time and discrete space price process. The equivalence between no arbitrage and the existence of a martingale measure is extended to markets with various imperfections. The concept of utility as an expression of preference is developed in Section B. Fishburn’s paper (1969) gives a succinct presentation of expected utility theory. With a set of assumptions over decision maker preferences, Fishburn establishes the existence of a utility function and subjective probability distribution such that rational individuals act as though they were maximizing expected utility. Expected utility allows for the fact that many individuals are risk averse, meaning that the individual would refuse a fair gamble (a fair gamble has an expected value of zero). vii

viii

Preface

In expected utility theory, the utilities of outcomes are weighted by their probabilities. Machina (2004) in his paper describes non expected utility. It has been shown that people overweight outcomes that are considered certain relative to outcomes which are merely probable. There have been a variety of proposals for dealing with the violation of the independence axiom and the linearity in probabilities. One approach is Prospect Theory proposed by Kahneman and Tversky (1979). The value function is S-shaped, being convex for losses (x < 0) and concave for gains (x > 0). Prospect Theory has its critics. The papers by Levy and Levy (2004) cast doubt on the S-shaped value function, based on experimental results. The basis of their analysis is stochastic dominance which orders random variables. The paper by Wakker (2003) reinforces the importance of probability weighting as well as the S-shaped value in prospect theory. He re-examines the experiments, showing that the S-shape is compatible with the data when probability weighting is used correctly. Baltussen et al. (2006) provide experimental results supporting the usual risk averse (concave) function over either the S-shape or reverse S-shape. In financial decision making, the (investment, consumption) process is dynamic, with the alternatives consisting of investment and consumption decisions at points in time. Kreps and Porteus (1979) consider the inter-temporal consistency of utility for making dynamic decisions. Preferences at different times are linked by a temporal consistency axiom, so that decisions are consistent in the sense that revealed outcomes at a later time would not invalidate the optimality of earlier decisions. The paper by Epstein and Zin (1989) builds on the temporal utility form to define a recursive utility which incorporates inter-temporal substitution and risk aversion separately. The Epstein–Zin utility contains many popular utility functions such as power utility as special cases. In Section C, the preference for random variables with the order relations of stochastic dominance are considered. Hanoch and Levy (1969) show that stochastic dominance can be defined by classes of utilities as characterized by higher order derivatives of the utility, with u(k) being the k th derivative. The dominance described is for a single period, usually the final period of accumulated capital. However, the capital is accumulated from investments over time and the trajectory to final wealth could be important in the assessment of utility. Levy and Paroush (1974) extend the notions of stochastic dominance to the multi-period case. With additive utility the stochastic dominance result is extended to multiple periods. Levy and Paroush also consider the geometric process where the final return is the product of period returns and give necessary conditions for first order dominance. Efficient investment strategies, where risk is minimized for specified return, are the topic of Section D. A major advantage of Markowitz’ mean-variance analysis is the relative ease of computing optimal strategies and as such it is a practical technique. The paper by Ziemba, Parkan and Brooks-Hill (1974) considers the general risk-return investment problem for which they propose a 2-step approach: the first step is to find the proportions invested in risky assets using a fractional program; the second step is to determine the optimal ratios of risky to non-risky assets.

Preface

ix

Ziemba (1975) considers the computation of optimal portfolios when returns have a symmetric stable distribution. The normal is a stable distribution when the variance is finite, but the stable family is more general with dispersion replacing variance and has four rather than two parameters. The typical utility function in the investment models is concave to reflect risk aversion. There are aspects of risk aversion which are not captured by concavity. In the paper by Pratt (1964), an additional property of utility, decreasing absolute risk aversion, is introduced. Rubinstein (1973) develops a measure of global risk aversion in the context of a parameter-preference equilibrium relationship. Some properties of this measure in the context of risk aversion with changing initial wealth levels appear in Kallberg and Ziemba (1983). Kallberg and Ziemba establish an important property of the global risk measure: Investors with the same risk have the same optimal portfolios. Chopra and Ziemba (1993) consider the relative impact of estimation errors at various degrees of risk aversion on portfolio performance, with the estimate of mean return being most important. MacLean, Foster and Ziemba (2007) use a Bayesian framework to include the covariance in an estimate of the mean. In essence, the return on one asset provides information about the return on related assets, and the sharing of information through the covariance improves the quality of estimates. Expected utility may not be the right theory for many risk attitudes, and does not explain the modest-scale risk aversion observed in practice. Rabin and Thaler (2001) in their paper propose that the right explanation incorporates the concepts of loss aversion and mental accounting. They argue that loss aversion and the tendency to isolate each risky choice must both be components of a theory of risk attitudes.

Part II The basic components of risk are (i) the chance of a potential loss and (ii) the size of the potential loss. F¨ ollmer and Knipsel (2012) view the financial risk of X as the capital requirement ρ(X) to make the position X acceptable, with the corresponding acceptance set of positions for measure ρ. With reference to acceptance sets, Rockafeller and Ziemba (2000) have established the following result: There is a one-to-one correspondence between acceptance sets Aρ and the risk measures ρ. The concept of capital requirement to cover the losses from investment captures the financial risk idea, but the probability of loss is not taken into account. Details on measures which use the distribution are provided in the papers of Krokhmal et al. (2011) and F¨ ollmer and Knipsel (2012). Deviation measures are a generalization of variance. There is a one-to-one relationship between averse risk measures and deviation risk measures. A variation is to consider the average of the values at risk (AVaR). F¨ollmer and Knipsel show how AVaR is a building block for law-invariant risk measures. Since convexity is the desired property of a risk measure, the class of convex risk measures is considered in F¨ollmer and Knispel (2012), following Rockafeller and Ziemba (2000). F¨ ollmer and Knispel also raise the issue of model uncertainty

x

Preface

or model ambiguity. They discuss a robustification where the probability measure P is a member of a class ℘. Markowitz’s mean-variance analysis approach is generalized in Krokhmal et al. (2011). Their decision problem is a trade-off, where a weighted combination of risk and reward is optimized. The weight is viewed as a penalty on risk. An application with the use of the penalty parameter approach is provided by the financial planning model InnoALM for the Austrian pension fund of the electronics firm Siemens, developed in Geyer and Ziemba (2008). The model uses a multiperiod stochastic linear programming approach and scenarios dependent correlation matrice. There are numerous models for investment choice in a stochastic dynamic environment and some are presented in Section F. Campbell et al. (2003) propose a standard first order autoregressive model for log-returns, with the inclusion of state variables in the dynamics. The the Epstein–Zin recursive utility is used. With a log-linearization, the approximate optimal portfolio rule is the sum of two terms: a myopic component from the vector of excess returns (log optimal solution), and an inter-temporal hedging component, which accounts for the fact that asset returns are state dependent and thereby time varying. The returns from securities are related to the efficient operation of firms. The efficiency of a firms operation can be analyzed from financial ratios, and firm efficiency related to market returns. Edirisinghe et al. (2012) in their paper look at firm input dimensions and output dimensions from an efficiency perspective — maximum output for a given level of input. A Data Envelopment Analysis calculates relative efficiency scores for firms from a linear programming model, and these scores are used in selection of securities for investment In a series of papers, Browne (1999a, 1999b, 2000) considers a number of investment problems involving benchmarks. He determines a proportional strategy which maximizes the probability of exceeding a deterministic target. The strategy is equivalent to buying a European digital option with a particular strike and payoff. In the more general case of a stochastic benchmark, the optimal proportional strategy is composed of the benchmark and a hedging component. If the benchmark is the optimal growth strategy (Kelly), Browne solves the probability maximizing problem for fixed T , finding a strategy that will beat the Kelly by an arbitrary percentage. It is well known that features such as volatility of returns are time dependent. However, time periods can be segmented into regimes so that within a regime the distribution characteristics are stable. For example, the price and wealth dynamics may be defined by geometric Brownian motion within each regime, with the parameters being regime dependent. In this setup, the maximizing of the growth rate is analytic. These and other variations on the log utility model are described in MacLean and Ziemba (2012). Some of the strongest properties of the optimal growth strategy relate to its evolutionary performance in an equilibrium capital market. In a frictionless market (no transactions costs) Bashoun et al. (2012) show that the Kelly rule is globally evolutionary stable, meaning that any other essentially different strategy will become extinct with probability 1.

Preface

xi

Davis and Lleo (2012) consider a variety of alternative models for price dynamics and determine the strategy which maximizes the power utility of wealth. The first variation is an inter-temporal asset pricing model. The optimal portfolio invests in a fractional Kelly portfolio and cash. The fractional Kelly fund is a blend of funds: a time dependent Kelly portfolio and an inter-temporal hedging portfolio. Davis and Lleo also consider the pricing model where the diffusion is augmented by shocks, defined by a homogeneous Poisson process. They show that a model with shocks in the dynamics of returns and a power utility function of final wealth has an optimal portfolio strategy which has an option component. Browne shows this is also true when the objective is maximizing the probability of beating a stochastic benchmark. Thorp and Mizusawa (2012) consider the option component in two models for asset returns: log-normal, and fat tails. The strategies are a blend of stock and T-Bills versus the same blend of an option on stock and T-Bill.

This page intentionally left blank

Contents

vii

Preface About the Editors

xvii

List of Contributors

xix

Acknowledgments

xxi

Part I: Decision Making Under Uncertainty

1

Section A. Arbitrage and Asset Pricing

3

1. 2.

3. 4.

Ross, SA (1976). The arbitrage theory of capital asset pricing. Journal of Economic Theory, 13(3), 341–360.

11

Schachermayer, W (2010a). The fundamental theorem of asset pricing. In R Cont (Ed.), Encyclopedia of Quantitative Finance, 2, 792–801. New York: Wiley.

31

Schachermayer, W (2010b). Risk neutral pricing. In R Cont (Ed.), Encyclopedia of Quantitative Finance, 4, 1581–1585. New York: Wiley.

49

Kallio, M and WT Ziemba (2007). Using Tucker’s theorem of the alternative to provide a framework for proving basic arbitrage results. Journal of Banking and Finance, 31, 2281–2302.

57 79

Section B. Utility Theory 5.

6. 7. 8.

Fishburn, P (1969). A general theory of subjective probabilities and expected utilities. Annals of Mathematical Statistics, 40(4), 1419–1429.

87

Kahneman, D and A Tversky (1979). Prospect theory: An analysis of decisions under risk. Econometrica, 47(2), 263–291.

99

Levy, M and H Levy (2002). Prospect theory: Much ado about nothing? Management Science, 48(10), 1334–1349.

129

Wakker, PP (2003). The data of Levy and Levy (2002) “Prospect theory: Much ado about nothing?” Actually support prospect theory. Management Science, 49(7), 979–981.

145

xiii

xiv

9.

Contents

Levy, M and H Levy (2004). Prospect theory and mean-variance analysis. Review of Financial Studies, 17(4), 1015–1041.

149

10. Baltussen, G, T Post and PV Vliet (2006). Violations of cumulative prospect theory in mixed gambles with moderate probabilities. Management Science, 52(8), 1288–1290.

177

11. Kreps, DM and EL Porteus (1979). Temporal von Neumann–Morgenstern and induced preferences. Journal of Economic Theory, 20(1), 81–109.

181

12. Epstein, LG and SE Zin (1989). Substitution, risk aversion and the temporal behavior of consumption and asset returns: A theoretical framework. Econometrica, 57(4), 937–969.

207

13. Rabin, M (2000). Risk aversion and expected-utility theory: A calibration theorem. Econometrica, 68(5), 1281–1292.

241

14. Machina, M (2004). Non-expected utility theory. In JL Teugels and B Sundt (Eds.), Encyclopedia of Actuarial Science, 2, 1173–1179. New York: Wiley.

253

15. Tversky, A and D Kahneman (1974). Judgment under uncertainty: Heuristics and biases. Science, 185(4157), 1124–1131.

261

16. Kahneman, D and A Tversky (1984). Choices, values, and frames. American Psychologist, 39(4), 341–350.

269

Section C. Stochastic Dominance

279

17. Hanoch, G and H Levy (1969). The efficiency analysis of choices involving risk. Review of Economic Studies, 36(3), 335–346.

287

18. Levy, H (1973). Stochastic dominance, efficiency criteria, and efficient portfolios: The multi-period case. American Economic Review, 63(5), 986–994.

299

Section D. Risk Aversion and Static Portfolio Theory

309

19. Pratt, JW (1964). Risk aversion in the small and in the large. Econometrica, 32(1–2), 122–136.

317

20. Li, Y and WT Ziemba (1993). Univariate and multivariate measures of risk aversion and risk premiums. Annals of Operations Research, 45, 265–296.

333

21. Chopra, VK and WT Ziemba (1993). The effect of errors in means, variances, and co-variances on optimal portfolio choice. Journal of Portfolio Management, 19, 6–11.

365

Contents

xv

22. Ziemba, WT, C Parkan and R Brooks-Hill (1974). Calculation of investment portfolios with risk free borrowing and lending. Management Science, 21(2), 209–222.

375

23. Kallberg, JG and WT Ziemba (1983). Comparison of alternative utility functions in portfolio selection problems. Management Science, 29(11), 1257–1276.

389

24. Li, Y and WT Ziemba (1989). Characterizations of optimal portfolios by univariate and multivariate risk aversion. Management Science, 35(3), 259–269.

409

25. Ziemba, WT (1975). Choosing investment portfolios when the returns have stable distributions. In WT Ziemba and RG Vickson (Eds.), Stochastic Optimization Models in Finance, 243–266. San Diego: Academic Press.

421

26. MacLean, LC, ME Foster and WT Ziemba (2007). Covariance complexity and rates of return on assets. Journal of Banking and Finance, 31(11), 3503–3523.

445

27. Rabin, M and RH Thaler (2001). Anomalies: Risk aversion. Journal of Economic Perspectives, 15(1), 219–232.

467

Part II: From Decision Making to Measurement and Dynamic Modeling

481

Section E. Risk Measures

483

28. Geyer, A and WT Ziemba (2008). The innovest Austrian pension fund planning model InnoALM. Operations Research, 56(4), 797–810.

491

29. Rockafellar, RT and WT Ziemba (2000). Modified risk measures and acceptance sets.

505

30. F¨ ollmer, H and T Knispel (2012). Convex risk measures: Basic facts, law-invariance and beyond, asymptotics for large portfolios.

507

31. Krokhmal, P, M Zabarankin and S Uryasev (2011). Modeling and optimization of risk. Surveys in Operations Research and Management Science, 16(2), 49–66.

555

Section F. Dynamic Portfolio Theory and Asset Allocation

601

32. Edirisinghe, C, X Zhang and S-C Shyi (2012). DEA-based firm strengths and market efficiency in US and Japan.

611

xvi

Contents

33. MacLean, LC and WT Ziemba (2012). The Kelly strategy for investing: Risk and reward.

637

34. Browne, S (1999a). Reaching goals by a deadline: Digital options and continuous-time active portfolio management. Advances in Applied Probability, 31, 551–577.

683

35. Browne, S (1999b). Beating a moving target: Optimal portfolio strategies for outperforming a stochastic benchmark. Finance and Stochastics, 3, 275–294.

711

36. Browne, S (2000). Stochastic differential portfolio games. Journal of Applied Probability, 37(1), 126–147.

731

37. Davis, M, and S Lleo (2012). Fractional Kelly strategies in continuous time: Recent developments.

753

38. Bahsoun, W, IV Evstigneev and MI Taksar (2012). Growth-optimal investments and numeraire portfolios under transactions costs.

789

39. Campbell, JY, YL Chan and LM Viceira (2003). A multivariate model of strategic asset allocation. Journal of Financial Economics, 67, 41–80.

809

40. Thorp, EO, and Mizusawa, S (2012). Maximizing capital growth with black swan protection.

849

Author Index

873

Subject Index

877

About the Editors

Leonard C MacLean is Professor Emeritus in the School of Business Administration at Dalhousie University in Halifax, Canada. Dr. MacLean has held visiting appointments at Cambridge University, University of Bergamo, University of British Columbia, Simon Fraser University, Royal Roads University, University of Zimbabwe, and University of Indonesia. From 1989 to 1995, he served as Director of the School of Business Administration at Dalhousie University. His research focuses on stochastic models in finance, and models for repairable systems in aviation. This work is funded by grants from the Natural Sciences and Engineering Council of Canada and the Herbert Lamb Trust. Professor MacLean teaches in the areas of statistics and operations management. William T Ziemba is the Alumni Professor (Emeritus) of Financial Modeling and Stochastic Optimization in the Sauder School of Business, University of British Columbia where he taught from 1968–2006 and Professor at the ICMA Centre, University of Reading. His PhD is from the University of California, Berkeley. He has been a Visiting Professor at Cambridge, Oxford, London School of Economics, and Warwick in the UK, at Stanford, UCLA, Berkeley, MIT, University of Washington and Chicago in the US, Universities of Bergamo, Venice and Luiss in Italy, the Universities of Zurich, Cyprus, Tsukuba (Japan), KAIST (Korea) and the National University of Singapore. He has been a consultant to a number of leading financial institutions including the Frank Russell Company, Morgan Stanley, Buchanan Partners, RAB Hedge Funds, Gordon Capital, Matcap, Ketchum Trading and, in the gambling area, to the BC Lotto Corporation, SCA Insurance, Singapore Pools, Canadian Sports Pool, Keeneland Racetrack and some racetrack syndicates in Hong Kong and Australia. His research is in asset-liability management, portfolio theory and practice, security market imperfections, Japanese and Asian financial markets, hedge fund strategies, risk management, sports and lottery investments and applied stochastic programming. He has published widely in journals such as Operations Research, Management Science, Mathematics of Operations Research, Mathematical Programming,

xvii

xviii

About the Editors

American Economic Review, Journal of Economic Perspectives, Journal of Finance, Journal of Economic Dynamics and Control, and Journal of Financial and Quantitative Analysis and in many books and special journal issues. Recent books include Applications of Stochastic Programming with SW Wallace (SIAM-MPS, 2005), Stochastic Optimization Models in Finance, 2nd edition with RG Vickson (World Scientific, 2006) and Handbook of Asset and Liability Modeling, Volume 1: Theory and Methodology (2006) and Volume 2: Applications and Case Studies (2007) with SA Zenios (North Holland), Scenarios for Risk Management and Global Investment Strategies with Rachel Ziemba (Wiley, 2007), Handbook of Investments: Sports and Lottery Betting Markets, with Donald Hausch (North Holland, 2008), Optimizing the Aging, Retirement and Pensions Dilemma with Marida Bertocchi and Sandra Schwartz and The Kelly Capital Growth Investment Criterion, with legendary hedge fund trader Edward O Thorp and Leonard C MacLean (World Scientific, 2010). His co-written practitioner paper on the Russell–Yasuda model won second prize in the 1993 Edelman Practice of Management Science Competition. He has been a futures and equity trader and hedge fund and investment manager since 1983. He is the series editor for North Holland’s Handbooks in Finance, World Scientific Handbooks in Financial Economics and Books in Finance, and previously was the CORS editor of INFOR and the department of finance editor of Management Science, 1982–1992. He has continued his columns in Wilmott and has prepared the book Investing in the Mondern Age with Rachel Ziemba on the 2007–2012 columns plus other materials for World Scientific. Ziemba, along with Hausch, wrote the famous Beat the Racetrack book (1984) (which was revised into Dr Z’s Beat the Racetrack (1987), which presented their place and show betting system and the Efficiency of Racetrack Betting Markets (1994, 2008) — the so-called bible of racetrack syndicates. Their 1986 book Betting at the Racetrack extends this efficient/inefficient market approach to simple exotic bets. Ziemba is revising Betting at the Racetrack into Exotic Betting at the Racetrack which adds Pick3,4,5,6, etc. and updates a few things for World Scientific to be out in Spring 2013.

List of Contributors

Bahsoun, W (Loughborough University) Baltussen, G (Erasmus School of Economics and Tinbergen Institute) Brooks-Hill, FJ (University of British Columbia) Browne, S (Columbia University) Campbell, JY (Harvard University) Chan, YL (Hong Kong University of Science and Technology) Chopra, V (Frank Russell Company) Davis, M (Imperial College London) Edirisinghe, C (University of Tennessee) Epstein, LG (University of Rochester) Evstigneev, IV (University of Manchester) Fishburn, PC (Research Analysis Corporation) F¨ ollmer, H (Humboldt University) Foster, M (Dalhousie University) Geyer, A (Vienna University of Economics and Business) Hanoch, G (Hebrew University of Jerusalem) Kahneman, D (Princeton University) Kallberg, JG (New York University) Kallio, M (Helsinki School of Economics) Knispel, T (Leibniz University) Krokhmal, P (University of Iowa) Levy, H (Hebrew University of Jerusalem) Levy, M (Hebrew University of Jerusalem) Li, Y (State University of California, Fullerton) Lleo, S (University of Reims) Machina, M (University of California, San Diego) MacLean, LC (Dalhousie University) Mizusawa, S (Edward O Thorp and Associates) Parkan, C (University of Calgary) Post, T (Erasmus School of Economics) Pratt, JW (Harvard University) Rabin, M (University of California, Berkeley) Rockafellar, RT (University of Washington) Ross, SA (MIT) Schachermayer, W (University of Vienna) Taksar, MI (University of Missouri) Thorp, EO (Edward O Thorp and Associates) xix

xx

List of Contributors

Tversky, A (Stanford University) Uryasev, S (University of Florida) Viceira, L (Harvard University) Vliet, PV (Robeco) Wakker, PP (University of Amsterdam) Zabarankin, M (Stevens Institute of Technology) Zhang, X (Austin Peay State University) Ziemba, WT (University of British Columbia, ICMA Centre and University of Reading) Zin, SE (Carnegie Mellon University)

Acknowledgments

Thanks go to the following journals and publishers for permission to reproduce the papers and book chapters listed below. Academic Press Ziemba, WT (1975). Choosing investment portfolios when the returns have stable distributions. In WT Ziemba and RG Vickson (Eds.), Stochastic Optimization Models in Finance, 243–266. San Diego: Academic Press. Advances in Applied Probability Browne, S (1999a). Reaching goals by a deadline: Digital options and continuoustime active portfolio management. Advances in Applied Probability, 31, 551–577. American Economic Review Levy, H (1973). Stochastic dominance, efficiency criteria, and efficient portfolios: The multi-period case. American Economic Review, 63(5), 986–994. American Psychologist Kahneman, D and A Tversky (1984). Choices, values and frames. American Psychologist, 39(4), 341–350. Annals of Mathematical Statistics Fishburn, P (1969). A general theory of subjective probabilities and expected utilities. Annals of Mathematical Statistics, 40(4), 1419–1429. Annals of Operations Research Li, Y and WT Ziemba (1993). Univariate and multivariate measures of risk aversion and risk premiums. Annals of Operations Research, 45, 265–296. Econometrica Epstein, LG and SE Zin (1989). Substitution, risk aversion and the temporal behavior of consumption and asset returns: A theoretical framework. Econometrica, 57(4), 937–969. Pratt, JW (1964). Risk aversion in the small and in the large. Econometrica, 32(1–2), 122–136. xxi

xxii

Acknowledgments

Rabin, M (2000). Risk aversion and expected utility theory: A calibration theorem. Econometrica, 68(5), 1281–1292. Kahneman, D and A Tversky (1979). Prospect theory: An analysis of decisions under risk. Econometrica, 47(2), 263–291. Finance and Stochastics Browne, S (1999b). Beating a moving target: Optimal portfolio strategies for outperforming a stochastic benchmark. Finance and Stochastics, 3, 275–294. Journal of Applied Probability Browne, S (2000). Stochastic differential portfolio games. Journal of Applied Probability, 37(1), 126–147. Journal of Banking and Finance Kallio, M and WT Ziemba (2007). Using Tucker’s theorem of the alternative to provide a framework for proving basic arbitrage results. Journal of Banking and Finance, 31, 2281–2302. MacLean, LC, ME Foster and WT Ziemba (2007). Covariance complexity and rates of return on assets. Journal of Banking and Finance, 31(11), 3503–3523. Journal of Economic Perspectives Rabin, M and RH Thaler (2001). Anomalies: Risk aversion. Journal of Economic Perspectives, 15(1), 219–232. Journal of Economic Theory Kreps, DM and EL Porteus (1979). Temporal von Neumann–Morgenstern and induced preferences. Journal of Economic Theory, 20(1), 81–109. Ross, SA (1976). The arbitrage theory of capital asset pricing. Journal of Economic Theory, 13(3), 341–360. Journal of Financial Economics Campbell, JY, YL Chan and L Viceira (2003). A multivariate model of strategic asset allocation. Journal of Financial Economics, 67, 41–80. Journal of Portfolio Management Chopra, VK and WT Ziemba (1993). The effect of errors in mean and co-variance estimates on optimal portfolio choice. Journal of Portfolio Management, 19, 6–11. Management Science Baltussen, G, T Post and PV Vliet (2006). Violations of cumulative prospect theory in mixed gambles with moderate probabilities. Management Science, 52(8), 1288–1290.

Acknowledgments

xxiii

Kallberg, JG and WT Ziemba (1983). Comparison of alternative utility functions in portfolio selection problems. Management Science, 29(11), 1257–1276. Levy, M and H Levy (2002). Prospect theory: Much ado about nothing? Management Science, 48(10), 1334–1349. Li, Y and WT Ziemba (1989). Characterizations of optimal portfolios by univariate and multivariate risk aversion. Management Science, 35(3), 259–269. Wakker, PP (2003). The data of Levy and Levy (2002) “Prospect theory: Much ado about nothing?” actually support prospect theory. Management Science, 49(7), 979–981. Ziemba, WT, C Parkan and R Brooks-Hill (1974). Calculation of investment portfolios with risk free borrowing and lending. Management Science, 21(2), 209–222. Operations Research Geyer, A and WT Ziemba (2008). The Innovest Austrian pension fund planning model InnoALM. Operations Research, 56(4), 797–810. Review of Economic Studies Hanoch, G and H Levy (1969). The efficiency analysis of choices involving risk. Review of Economic Studies, 36(3), 335–346. Review of Financial Studies Levy, M and H Levy (2004). Prospect theory and mean-variance analysis. Review of Financial Studies, 17(4), 1015–1041. Science Tversky, A and D Kahneman (1974). Judgement under uncertainty: Heuristics and biases. Science, 185(4157), 1124–1131. Surveys in Operations Research and Management Science Krokhmal, P, M Zabarankin and S Uryasev (2011). Modeling and optimization of risk. Surveys in Operations Research and Management Science, 16(2), 49–66. Wiley Machina, M (2004). Non-expected utility theory. In JL Teugels and B Sundt (Eds.), Encyclopedia of Actuarial Science, 2, 1173–1179. New York: Wiley. Schachermayer, W (2010a). The fundamental theorem of asset pricing. In R Cont (Ed.), Encyclopedia of Quantitative Finance, 2, 792–801. New York: Wiley. Schachermayer, W (2010b). Risk neutral pricing. In R Cont (Ed.), Encyclopedia of Quantitative Finance, 4, 1581–1585. New York: Wiley.

This page intentionally left blank

Part I

Decision Making Under Uncertainty

This page intentionally left blank

Section A

Arbitrage and Asset Pricing

4

Section A. Arbitrage and Asset Pricing

The theme of this handbook is financial decision making. The decisions are the amount of investment capital to allocate to various opportunities in a financial market. The opportunity set can be very complex, with sets of equities, bonds, commodities, derivatives, futures, and currencies having trading prices changing stochastically and dynamically over time. To consider decisions in a complex market, it is necessary to impose structure. In the abstract, the assets and the participants buying and selling them are parts of a system with underlying economic states. The system’s dynamics and the factors defining the states in the system have been studied extensively in finance and economics. The dynamics of the market and the behavior of participants determine the trading prices of the various assets in the opportunity set. A simplifying assumption is that the financial market is perfectly competitive. There are conditions that must be present for a perfectly competitive market structure to exist. There must be many participants in the market, none of which is large enough to affect prices. Individuals should be able to buy and sell without restriction. All participants in the market have complete information about prices. In the competitive market, investors are price takers. These assumptions are strong, and in actual financial markets they are not exactly satisfied. However, with the assumed structure an idealized market can be characterized and that provides a standard by which existing practice can be measured. If investors are price takers, then a fundamental component of financial decision making is asset pricing. A common approach to asset pricing is to derive equilibrium prices for assets in a competitive market. This can be achieved with a model mapping the abstract states defined by a probability space into prices of assets such as equities and bonds. The Capital Asset Pricing Model (CAPM) developed independently by Sharpe (1964), Lintner (1965), Mossin (1966) and Treynor (1961, 1962) is a standard for pricing risky assets. Some clarification is provided in Fama (1968). The model proposes that the expected excess return of a risky asset over a riskless asset is proportional to the expected excess return of the market over the riskless asset. The returns on assets are assumed to be normally distributed. In this setting the financial market is in competitive equilibrium. Consistent with this structure, the optimal investment decisions are determined from the mean-variance approach developed by Markowitz (1952, 1959). The CAPM is the theoretical basis for much of the sizable index fund business. Dimension Fund Advisors alone manages $250 billion, most of which is passive. The CAPM model has a single explanatory variable, the market portfolio, in a simple linear regression. This model has been extended to include other market variables in a multivariate linear regression. For example, following Rosenberg (1974), Rosenberg, Reid and Lanstein (1985), Fama and French (1992) have added two explanatory variables: (i) small minus large capitalization; and (ii) high minus low book to market ratio. The equilibrium pricing in the CAPM type models implies that no arbitrage opportunities exist. An arbitrage is a transaction that involves no negative cash

Section A. Arbitrage and Asset Pricing

5

flow at any probabilistic or temporal state and a positive cash flow in at least one state; in simple terms, it is the possibility of a risk-free profit at zero cost. The Arbitrage Pricing Theory (APT) for asset pricing following from an assumption of no arbitrage was developed by Ross (1976). This theory defines the expected returns on assets with a linear factor model. The theory linking arbitrage to the factor model are presented in the paper “The Arbitrage Theory of Capital Asset Pricing.” The Ross argument considers a well-diversified portfolio of risky assets which uses no wealth (free lunch). The portfolio is essentially independent of noise. If the portfolio has no risk, then the random return is certain and to avoid disequilibrium the certain return must be zero. This no arbitrage condition implies that the returns on the assets are defined by a linear relation to a set of common random factors with zero expectation. This type of equilibrium arbitrage argument follows the famous Modigliani and Miller paper (1958) and is part of the reasoning in the Black–Scholes option pricing (1973) model. There are a number of differences between the CAPM and APT theories. The most significant distinction is the “factors.” In CAPM the factors/independent variables are manifest market variables (e.g., market index). With APT the factors are intrinsic (not manifest) variables, whose existence follows from diversification and no arbitrage. It is not required that the APT factors have clear definitions as entities. The APT factors are structural, without implied causation. That is, CAPM: factors → returns; APT: factors ↔ returns. So the factor model in APT is really a distributional condition on prices following from no arbitrage. The essence of arbitrage is captured in Ross’ theory. There are no assumptions in APT about the distribution of noise, whereas CAPM assumes normality. However, the use of the factor model in empirical work on pricing does use algorithms which sometime assume normality of the factors and returns. The statistical estimation would also suggest definitions/entities for the intrinsic factors, which could further link the CAPM and APT models. Factor models have been used in practice by many analysts (see Jacobs and Levy (1988) and Ziemba and Schwartz (1991), and Schwartz and Ziemba (2000)). Companies such as BARRA sell these models. The APT does not assume the existence of a competitive equilibrium. Disequilibrium can exist in the theory, but it is assumed that in aggregate the returns are uniformly bounded. The no arbitrage assumption is a natural condition to expect of a stable financial market. The existence of arbitrage free prices for assets is linked to the probability measure on which the stochastic process of prices is defined. The Fundamental Theorem of Asset Pricing states that If S = {St, t ≥ 0} are asset prices in a complete financial market, then the following statements are equivalent (i) S does not allow for arbitrage (ii) There exists a probability measure which is equivalent to the original underlying measure and the price process is a martingale under the new measure.

6

Section A. Arbitrage and Asset Pricing

A martingale is a stochastic process where the conditional expected value for the next period equals the current observed value, and does not depend on the history of the process. So a martingale is a model for a fair process and it is not surprising that the fairness of no arbitrage can be characterized by a martingale measure. Indeed the Ross (1976) argument establishes the link between arbitrage and a martingale measure using the famous Hahn–Banach theorem. The assumptions used by Ross on the underlying measure were some what limiting. In the case of an infinite probability space, the Ross result only applies to the sup norm topology. For finite dimensional space, it is not clear that the martingale measure is actually equivalent. These limitations were considered by Harrison and Kreps (1979) and Harrison and Pliska (1981). They extended the Fundamental Theorem of Asset Pricing in several ways: (i) If the price process is defined on a finite, filtered, probability space, then the market contains no arbitrage possibilities if and only if there is an equivalent martingale measure. (ii) If the price process is defined on a continuous probability space and the market admits “no free lunch,” then there exists an equivalent martingale measure. (iii) If the price process is defined on a countably generated probability space, taking values in Lp space, then the “no free lunch” condition is satisfied if and only if there is an equivalent martingale measure satisfying a q moment condition, where p1 + 1q = 1. Although the work of Kreps and colleagues made significant contributions to the theory of arbitrage pricing, there were still assumptions which limited the applicability. Ideally a more economically natural condition could replace the moment condition on the martingale measure. Delbaen and Schachermayer (2006) discuss many open questions. One particular advance links the existence of an equivalent martingale measure in processes in continuous time or infinite discrete time to a condition of “no free lunch with bounded risk.” Unfortunately, this result does not hold for price processes which are semi-martingales. Furthermore, there are strong mathematical and economic reasons to assume that the price process is a semimartingale. In that setting the no free lunch with bounded risk is replaced by a no free lunch with vanishing risk, where risk disappears in the limit. The latter is stronger than the former, but is weaker than a no arbitrage condition. So Schachermayer (2010), and Delbaen and Schachermayer (2006) have a general statement of the fundamental theorem: Assume the price process is a locally bounded real-valued semi-martingale. There is a martingale measure which is equivalent to the original measure if and only if the price process satisfies the no free lunch with vanishing risk condition. Yan (1998) brought the results even closer to the desired form. The concept of allowable trading strategies was introduced, where the trader remains liquid during

Section A. Arbitrage and Asset Pricing

7

the trading interval. The Yan formulation yields the result: Let the price process be a positive semi-martingale. There is a martingale measure which is equivalent to the original measure if and only if the price process satisfies the no free lunch with vanishing risk condition with respect to allowable trading strategies. Another term for an equivalent martingale measure is a risk-neutral measure. Prices of assets depend on their risk, with a premium required for riskier assets. The advantage of the equivalent martingale or risk-neutral measure is that risk premia are incorporated into the expectation with respect to that measure. Under the riskneutral measure all assets have the same expected value — the risk-free rate. The stock price process discounted by the risk-free rate is a martingale under the riskneutral measure. This simplification is important in the valuation of assets such as options and is a component of the famous Black–Scholes (1973) formula. Of course, the risk-neutral measure is an artificial concept, with important implications for the theory of pricing. The actual risk-neutral measure used for price adjustment must be determined from economic reasoning. The separating hyperplane arguments underlying the results linking arbitrage and no free lunch to martingale measures have an analogy in theorems of the alternative for discrete time and discrete space arbitrage pricing models. In theorems of the alternative competing systems of equlities/inequalities are posed, with only one system having a solution. A famous such theorem is due to Tucker (1954). Kallio and Ziemba (2007) used Tucker’s Theorem of the Alternative to derive known and some new arbitrage pricing results. The competing systems define arbitrage on the one hand and the existence of risk-neutral probabilities on the other hand. For a frictionless market the Fundamental Theorem of Asset pricing is established using matrix arguments for the discrete time and discrete space price process: If at each stage an asset exists with strictly positive return (there exists a trading strategy), then arbitrage does not exist if and only if there exists an equivalent martingale measure. Although the discrete time and space setting is limiting, it is used in practice as an approximation to the continuous process. Obviously there are considerable computational advantages with a discrete process, and assumptions required for its implementation are few. In the general setting the fundamental theorem posits the existence of a risk-neutral measure. Actually finding such a measure requires additional assumptions. In the discrete setting, the equations for calculating the probabilities in the measure can be solved. This is analogous to the option pricing models, where in the Black–Scholes approach strong distribution assumptions are required to get the pricing formula, but the binomial lattice approach obtains option prices with a linear programming algorithm. Even from a theory perspective, the discrete time and space extension to more complex financial markets is feasible since the mathematics is based on systems of equations. In Kallio and Ziemba (2007)

8

Section A. Arbitrage and Asset Pricing

the equivalence between no arbitrage and the existence of a martingale measure is extended to markets with various imperfections. The no arbitrage condition is fundamental to much of the theory of efficient capital markets. However, it is important to acknowledge the existence of arbitrage opportunities in actual markets. Examples are the Nikkei put warrant arbitrage discussed in Shaw, Thorp and Ziemba (1995), and the race track arbitrages discussed by Hausch and Ziemba (1990a, 1990b). Investors exhibit behavioral biases which can lead to mispricing and arbitrage. Usually over/under pricing is temporary, but correctly identifying those events and using them for financial advantage has attracted attention.

Readings Black F and M Scholes (1973). The pricing of options and corporate liabilities. Journal of Political Economy, 81(3), 637–654. Delbaen, F and W Schachermayer (2006). The Mathematics of Arbitrage. New York: Springer. Fama, EF (1968). Risk, return and equilibrium: Some clarifying comments. Journal of Finance, 23(1), 29–40. Fama, EF and F French (1992). The cross-section of expected stock returns. Journal of Finance, June, 427–466. Harrison, JM and DM Kreps (1979). Martingales and arbitrage in multiperiod securities markets. Journal of Economic Theory, 20, 381–408. Harrison, JM and SR Pliska (1981). Martingales and stochastic integrals in the theory of continuous trading. Stochastic Processes and their Applications, 11, 215–260. Hausch, DB and WT Ziemba (1990a). Arbitrage strategies for cross track betting on major horseraces. Journal of Business, 63, 61–78. Hausch, DB and WT Ziemba (1990b). Locks at the racetrack. Interfaces, 20(3), 41–48. Jacobs, BL and KN Levy (1988). Disentangling equity return regularities: New insights and investment opportunities. Financial Analysts Journal, 44, 18–43. Kallio, M and WT Ziemba (2007). Using Tucker’s Theorem of the alternative to provide a framework for proving basic arbitrage results. Journal of Banking and Finance, 31, 2281–2302. Lintner, J (1965). The valuation of risk assets and the selection of risky investments in stock portfolios and capital budgets. Review of Economics and Statistics, 47(1), 13–37. Markowitz, H (1952). Portfolio selection. Journal of Finance, 7, 77–91. Markowitz, H (1959). Portfolio Selection: Efficient Diversification of Investments. New York: Wiley. Modigliani, F and M Miller (1958). The cost of capital, corporation finance and the theory of investment. American Economic Review, 48(3), 261–297. Mossin, J (1966). Equilibrium in a capital asset market. Econometrica, 34(4), 768–783. Rosenberg, B (1974). Extra-market components of covariance in securities markets. Journal of Financial and Quantitative Analysis, 263–274. Rosenberg, B, K Reid and R Lanstein (1985). Persuasive evidence of market inefficiency. Journal of Portfolio Management, 11(3), 9–16. Ross, SA (1976). The arbitrage theory of capital asset pricing. Journal of Economic Theory, 13(3), 341–360. Schachermayer, W (2010). The fundamental theorem of asset pricing. In R Cont (Ed.), Encyclopedia of Quantitative Finance, 2, 792–801. New York: Wiley.

Section A. Arbitrage and Asset Pricing

9

Schachermayer, W (2010). Risk neutral pricing. In R Cont (Ed.), Encyclopedia of Quantitative Finance, 4, 1581–1585. New York: Wiley. Schwartz, SL and WT Ziemba (2000). Predicting returns on the Tokyo stock exchange. In DB Keim and WT Ziemba (Eds.), Security Market Imperfections in Worldwide Equity Markets, 492–511. Cambridge University Press. Sharpe, WF (1964). Capital asset prices: A theory of market equilibrium under conditions of risk. Journal of Finance, 19(3), 425–442. Shaw, J, EO Thorp and WT Ziemba (1995). Convergence to efficiency of the Nikkei put warrant market of 1989–1990. Applied Mathematical Finance, 2, 243–271. Treynor, JL (1961). Market Value, Time, and Risk. Unpublished manuscript. Treynor, JL (1962). Toward a theory of market value of risky assets. In RA Korajczyk (Ed.) Asset Pricing and Portfolio Performance: Models, Strategy and Performance Metrics. London: Risk Books. Tucker, A (1956). Dual systems of homogeneous linear relations. In H Kuhn and A Tucker (Eds.), Linear Inequalities and Related Systems, Annals of Mathematics Studies. Princeton: Princeton University Press. Yan, JA (1998). A new look at the fundamental theorem of asset pricing, 35, 659–673. Ziemba, WT and SL Schwartz (1991). Invest Japan: The Structure, Performance and Opportunities of Japan’s Stock, Bond and Fund Markets. Chicago: Probus Publishing.

This page intentionally left blank

JOURNAL OF ECONOMIC THEORY

13, 341-360 (1976)

Chapter 1 The Arbitrage Theory of Capital Asset Pricing STEPHEN

A. Ross*

Departments of Economics and Finance, University of Pennsylvania, The Wharton School, Philadelphia, Pennsylvania 19174

Received March 19, 1973: revised May 19, 1976

The purpose of this paper is to examine rigorously the arbitrage model of capital asset pricing developed in Ross [13, 14]. The arbitrage model was proposed as an alternative to the mean variance capital asset pricing model, introduced by Sharpe, Lintner, and Treynor, that has become the major analytic tool for explaining phenomena observed in capital markets for risky assets. The principal relation that emerges from the mean variance model holds that for any asset, ;, its (ex ante) expected return (l)

where p is the riskless rate of interest, .\ is the expected excess return on the market, Ern - p, and

is the beta coefficient on the market, where arn2 is the variance of the market portfolio and aim is the covariance between the returns on the ith asset and the market portfolio. (If a riskless asset does not exist, p is the zero-beta return, i.e., the return on all portfolios uncorrelated with the market portfolio.)1 The linear relation in (I) arises from the mean variance efficiency of the market portfolio, but on theoretical grounds it is difficult to justify either the assumption of normality in returns (or local normality in Wiener diffusion models) or of quadratic preferences to guarantee such efficiency, and on empirical grounds the conclusions as well as the *Professor of Economics, University of Pennsylvania. This work was supported by a grant from the Rodney L. White Center for Financial Research at the University of Pennsylvania and by National Science Foundation Grant GS-35780. 1 See Black [2] for an analysis of the mean variance model in the absence of a riskless asset.

341 Copyright Gセ@ 1976 by Academic Press, Inc. All rights of reproduction in any form reserved.

11

12

SA Ross

342

STEPHEN A. ROSS

assumptions of the theory have also come under attack. 2 The restrictiveness of the assumptions that underlie the mean variance model have, however, long been recognized, but its tractability and the evident appeal of the linear relation between return, E;, and risk, bi, embodied in (I) have ensured its popularity. An alternative theory of the pricing of risky assets that retains many of the intuitive results of the original theory was developed in Ross [13, 14]. In its barest essentials the argument presented there is as follows. Suppose that the random returns on a subset of assets can be expressed by a simple factor model X; == E;

+ (3/j + €; ,

(2)

where 8 is a mean zero common factor, and i!i is mean zero with the vector (€) sufficiently independent to permit the law of large numbers to hold. Neglecting the noise term, i!i, as discussed in Ross [14] (2) is a statement that the state space tableau of asset returns lies in a twodimensional space that can be spanned by a vector with elements 88 , (where () denotes the state of the world) and the constant vector, e セ@ (1, ... , 1). Step I. Form an arbitrage portfolio, TJ, of all the n assets, i.e., a portfolio which uses no wealth, 7Je = 0. We will also require 7J to be a well-diversified portfolio with each component, TJi , of order 1/n in (absolute) magnitude. Step 2. By the law of large numbers, for large arbitrage portfolio 7JX

+ (7]{3) 8 + YJE TJE + (7Jf3) 8.

= TJE RoO

n

the return on the

(3)

ln other words the influence on the well-diversified portfolio of the independent noise terms becomes negligible. Step 3. If we now also require that the arbitrage portfolio, 7], be chosen so as to have no systematic risk, then 7Jf3

=

0,

and from (3) 2 See Blume and Friend [3] for a recent example of some of the empirical difficulties faced by the mean variance model. For a good review of the theoretical and empirical literature on the mean variance model see Jensen [6].

Chapter 1. The Arbitrage Theory of Capital Asset Pricing CAPITAL ASSET PRICING

13

343

Step 4. Using no wealth, the random return T)X has now been engineered to be equivalent to a certain return, T)E, hence to prevent arbitrarily large disequilibrium positions we must have TJE = 0. Since this restriction must hold for all T) such that T)e ]Mセ@ TJf3 =o= 0, E is spanned by e and f3 or

E, = p

+ /o.f3,

(4)

for constants p and /o.. Clearly if there is a riskless asset, p must be its rate of return. Even if there is not such an asset, p is the rate of return on all zero-beta portfolios, ex, i.e., all portfolios with exe = 1 and exf3 = 0. If ex is a particular portfolio of interest, e.g., the market portfolio, cx 111 , with £, = ex 111 E, (4) becomes

E;

=

p

+ (E"'- p) {3;.

(5)

Condition ( 5) is the arbitrage theory equivalent of (I), and if f5 is a market factor return then /3 1 will approximate b, . The above approach, however, is substantially different from the usual mean-variance analysis and constitutes a related but quite distinct theory. For one thing, the argument suggests that (5) holds not only in equilibrium situations, but in all but the most profound sort of disequilibria. For another, the market portfolio plays no special role. There are, however, some weak points in the heuristic argument. For example, as the number of assets, n, is increased, wealth will, in general, also increase. Increasing wealth, though, may increase the risk aversion of some economic agents. The law of large numbers implies, in Step 2. that the noise term, T)E, becomes negligible for large n, but if the degree of risk aversion is increasing with n these two effects may cancel out and the presence of noise may persist as an influence on the pricing relation. In Section I we will present an example of a market where this occurs. Furthermore, even if the noise term can be eliminated, it is not at all obvious that (5) must hold, since the disequilibrium position of one agent might be offset by the disequilibrium position of another. 3 In Ross [13], however, it was shown that if (5) holds then it represents an E or quasi-equilibrium. The intent of this paper is to supply the rigorous analysis underlying the stronger stability arguments above. In Section 11 we will present some weak sufficient conditions to rule out the above exceptions (and the example of Section l) and we will prove a general version of the arbitrage result. Section II also includes a brief argument on the empirical practicality of the results. A mathematical appendix 3 Green has considered this point in a temporary equilibrium model. Essentially he argues that if subjective anticipations differ too much, then arbitrage possibilities will threaten the existence of equilibrium.

14

SA Ross

344

STEPHEN A. ROSS

contains some supportive results of a somewhat technical and tangential nature. Section l II will briefly summarize the paper and suggest further generalizations.

I. A

COUNTEREXAMPLE

In this section we will present an example of a market where the sequence of equilibrium pricing relations does not approach the one predicted by the arbitrage theory as the number of assets is increased. The counterexample is valuable because it makes clear what sort of additional assumptions must be imposed to validate the theory. Suppose that there is a riskless asset and that risky assets are independently and normally distributed as

x;

E;

=

+ i;,

(6)

where E{ii}

=

0,

and

The arbitrage argument would imply that in equilibrium all of the independent risk would disappear and, therefore, Ei ᄋセ@

(7)



Assume, however, that the market consists of a single agent with a von Neumann-Morgenstern utility function of the constant absolute risk aversion form, U(z)

=

-exp( -Az).

(8)

Letting w denote wealth with the riskless asset as the numeraire, and ex the portfolio of risky assets (i.e., ex; is the proportion of wealth placed in the ith risky asset) and taking expectations we have E{U[w(p

+ a[x- p · e])]}

=

-exp( -Awp) E{exp( -Awcx[x- p · e])}

=

-exp( -Awp){exp( -Awcx[E- p · e]

+ (a /2)(Aw)2(cx'cx))}. 2

(9)

The first-order conditions at a maximum are given by a 2 (Aw) セx Q@

= E; - p.

(10)

Chapter 1. The Arbitrage Theory of Capital Asset Pricing

15

345

CAPITAL ASSET PRICING

If the riskless asset is in unit supply the budget constraint (Walras' Law

for the market) becomes n

1\'

=

I

cx;w

+l

n

= (1/Aa 2 )

i -1

L (E; -

p)

+ l,

(11)

i=l

The interpretation of the budget constraint (11) depends on the particular market situation we are describing. Suppose, first, that we are adding assets which will pay a random total numeraire amount, ci . If Pt is the current numeraire price of the asset then

Normalizing all risky assets to be in unit supply we must have

=

Pt

CX;W,

and the budget constraint simply asserts that wealth is summed value, n

I

\\' =

Pt

+ l.

ゥセャ@

If we let

c; denote the mean of c; and c2, its variance, then (10) can be

solved for p1 as

As a consequence, the expected returns,

will be unaffected by changes in the number of assets, n, for i < n, and need bear no systematic relation to p as n increases. This is a violation of the arbitrage condition, (7). Notice, too, that as long as c1 is bounded above Ac2 , wealth and relative risk aversion, Aw, are unbounded in n. An alternative interpretation of the market situation would be that as 11 increases the number of risky investment opportunities or activities is being increased, but not the number of assets. In this case wealth, w, would simply be the number of units of the riskless asset held and would remain constant as 11 increased. The quantities cx,:W now represent the amount of the riskless holdings put into the ith investment opportunity and for the market as a whole we must have n

I

i=l

CX;

< I.

16

SA Ross

346

STEPHEN A. ROSS

Furthermore, if the random technological activities are irreversible, then each rx; ;;;: 0. From (10) it follows that E;-

p;;;: 0

and n

I

n

E, -· p =

n

I

I E;

·

pI =

a 2(Aw)

i=l

i=l

I

rx;


U(p),

but by concavity hence lim E{U[p

+ Xn]}

= U(p). Q.E.D.

A problem arises when U(-) is unbounded from below. About the weakest condition which assures convergence is uniform integrability (U.I.):

r

lim sup n ., Hlセッ@

Q.,

0 rr

I U(p

= {I U(p

+ Xn)l dr;n

= 0,

+ Xnll > c-.:},

where YJn is the distribution function of X, .



Chapter 1. The Arbitrage Theory of Capital Asset Pricing

29

359

CAPITAL ASSET PRICING

A number of familiar conditions imply U.l. If the sequence U(p + Xn) is bounded below by an integrable function the Lebesque convergence theorem can be invoked or if (38 > 0) sup £{1 U(p n

+ X nW' < 8}

oo,

then the sequence is U.I. In general, then, the convergence criterion will depend on both the utility function and the random variables. It is possible, however, to find weak sufficient conditions on the random variables alone, by taking advantage of the structure of .Yn, but the condition that X,.= (1/n) Li €;; al uniformly bounded and E; , Ej independent is not su:fficient. 9 In the text, it is assumed that all sequences satisfy the U.l. condition, and therefore (q.m.) will imply that

REFERENCES

1. P. BILLINGSLEY, "Convergence of Probability Measures," Wiley, New York, 1968. 2. F. BLACK, Capital market equilibrium with restricted borrowing, J. Business 45 (1972), 444-455. 3. M. BLUME AND I. FRIEND, A new look at the capital asset pricing model, J. Finance (March 1973), 19-33. 4. I. FRIEND, Rates of return on bonds and stocks, the market price of risk, and the cost of capital, Working Paper No. 23-73, Rodney L. White Center for Financial Research, 1973. 5. J. GREEN, Preexisting contracts and temporary general equilibrium, in "Essays on Economic Behavior under Uncertainty" (Balch, McFadden, and Wir, Eds.), North-Holland, Amsterdam, 1974. 6. M. jensセL@ Capital markets: theory and evidence, Bell. J. Econ. and Management Science 3 (1972), 357-398. 7. J. LINTNER, The valuation of risk assets and the selection of risky investments in stock portfolios and capital budgets, Rev. Econ. Statist. (February 1965), 30--55. 8. M. LOEVE, "Probability Theory," Van Nostrand, Princeton, N. J., 1963. 9. S. MYERS, A reexamination of market and industry factors in stock price behavior, J. Finance (June 1973), 695-705. 10. J, PRATT, Risk aversion in the small and in the large, Econometrica 32 (1964), 122-137. 11. H. RAIFFA, "Decision Analysis," Addison-Wesley, Reading, Mass., 1968. ' It is not difficult to construct counterexamples by having U(o) go to - oo rapidly enough as x approaches its lower bound.

30

SA Ross

360

STEPHEN A. ROSS

12. S. Ross, Comment on "Consumption and Portfolio Choices with Transaction Costs," by E. Zabel and R. Multherjee, in "Essays on Economic Behavior under Uncertainty" (Balch, McFadden, and Wir, Eds.), North-Holland, Amsterdam, 1974. 13. S. Ross, Portfolio and capital market theory with arbitrary preferences and distributions-The general validity of the mean-variance approach in large markets, Working Paper No. 12-72, Rodney L. White Center for Financial Research, 1971. 14. S. Ross, Return, risk and arbitrage, in "Risk and Return in Finance" (Friend and Bicksler, Eds.), Ballinger, Cambridge, Mass., forthcoming. 15. W. SHARPE, Capital asset prices: A theory of market equilibrium under conditions of risk, J. Finance (September 1964), 425-442. 16. .T. TREYNOR, Toward a theory of market value of risky assets, unpublished manuscript, 1961.

Chapter 2 In R Cont (Ed.), Encyclopedia of Quantitative Finance, 2, 792-801. New York: Wiley.

The Fundamental Theorem of Asset Pricing 'Walter Schachermayer Univer-sity of Vienna, Austr-ia

The subsequent theorem is one of the pillars supporting the modern theory of Mathematical Finance. Fundamental Theorem of Asset Pricing: The following two statements are essentially equivalent for a model S of a financial market: (i) S does not allow for arbitrage (NA) (ii) There exists a probability measure Q on the underlying probability space (0, :F, lP'), which is equivalent to lP' and under which the process is a martingale (EMM). \Ve have formulated this theorem in vague terms which will be made precise in the sequel: we shall formulate versions of this theorem below which use precise definitions and avoid the use of the word essentially above. In fact, the challenge is precisely to turn this vague "meta-theorem" into sharp mathematical results. The story of this theorem started - like most of modern Mathematical Finance - with the work of F. Black, M. Scholes [3] and R. :VIerton [25]. These authors consider a model S = (S1 )o 0), and such that its price 1r( m) is less than or equal to zero. The question now arises whether it is possible to extend 1r : AI ---+ JR. to a non-negative, continuous linear functional1r* : X ---+ JR.. \Vhat does this have to do with the issue of martingale measures'? This

3

34

W Schachermayer

theme was developed in detail by M. Harrison and D. Kreps [14]. Suppose that X = LP(fl, :F, IP') for some 1 セ@ p < oo, that the price process S = (St)o 0 such that 11* = lセ]ャ@ f-Ln"hn converges to a probability measure in (X, T)* = Lq(n, F, lP'), where セ@ + = 1. This yields the desired extension 11* of 11 which is strictly positive on X+ \{0}. vVe still have to specify the choice of (lvf0 , 11). The most basic choice is to take for givenS= (St)o 0. jESk

Hence, LsEs,Ys

=

1, n5 /nk

=

yjR{, and Theorem 2 implies

Theorem 3 (Risk neutral valuation). If a single period risk free asset exists at each node

k r;t T, then Arbitrage A does not exist if and only probabilities 'lk > 0, for all k > 0, such that

if there

exist conditional risk neutral

(6) For all assets i, nodes k rj T, and s E S 10 denote the total return by R; = pェセ@ from node k to s. Let Rk = (RD denote an n vector of the total returns. Then equivalently with (6), for all i,

R{ = LYsR:

'c/k r;f T.

sr;::-Sk

The risk neutral representation was first employed by Cox and and Ross (1976).

(7)

Chapter 4. Using Tucker’s Theorem of the Alternative

63

pセ@

M. Kallio, WT. Ziemba I Journal of' Banking & Finance 31 (2007) 2281-2302

2287

If for each node k >t T, there is an asset (a numeraire) with a strictly positive price > 0, for all s E Sk, then PZ > 0 in (5). Hence, we may define normalized prices Pk = pォOセ@ and P, = pLOセ@ and parameters CJ, = HョLpセIOォ@ such that (5) becomes Pk = LsESk CJsPs. As LsESk CJk = 1, and CJ,. > 0, for all s E Sk, we interpret parameters CJ, as conditional probabilities. Theorem 2 implies the following result stating that the normalized price process under no arbitrage is a martingale. Theorem 4 (Martingale valuation). If at each stage an asset exists with a strictly positive return, then Arbitrage A does not exist if and only if there exist conditional probabilities CJk > 0, for all k > 0, under which the normalized price process is a martingale; i.e.,

pk =

L

CJsPs

\:/k

>t T.

(8)

sESk

This result in a much more general framework, including vector ditlusion processes for the price vector, was developed in Harrison and Kreps (1979). 3.2. Valuation of a contingent claim

Consider a contingent claim u = (vk), where vk, fork> 0, is the cash flow provided by the claim, and for the root node, -v0 = Vis the purchase price. As an example, in case of a European put option allowing but not forcing the owner to sell a given asset at a given time and at a prespecified price, such a stochastic cash flow stream is contingent on the evolution the price of the underlying asset. For k > 0, vk is the payoff provided by the option at node k, and v0 = -Vis the price paid. A price or value of a contingent claim is admissible, if an arbitrage opportunity is not introduced while adding the claim with this price to the investment possibilities. We determine an admissible price V based on a no arbitrage assumption. Let セ@ denote the amount invested in the contingent claim. Then the net cash flow vector is c = Cx + カセ@ and an arbitrage exists, if there is x and セ@ such that c ? 0 and c i= 0. By Stiemke's theorem, there is no arbitrage if and only if the set

(9)

Il={nlnC=O, n>O, no= I}

of state price vectors n is non-empty and v0 is chosen such that nv Hence, under a no arbitrage assumption, for n E II

=

0, for some n

E II.

(10) Computations for V can be done recursively as well. For terminal nodes k E T, define vk = 0. For other nodes k, define vk = L,·ES,, (n,/nk)(v, + V,), or equivalently using risk neutral valuation (6), Vk = (1/ROL,EsJ,(v, + V,). Then the value in (10) is V= V0 . In special cases, duplicated computations can be avoided by applying the recursion in a lattice rather than in a tree; see for instance the option pricing by Cox et al. (1979). If u = Cx for some x and v0 , then v is replicated by the portfolio strategy x. Let Ck denote row vector k of C, for k = 0, 1, ... Hence C0 is the first row in C. Then v = Cx and n E II implies nv = nCx = 0, and V = -v 0 = - C0 x. Hence, if replication is possible and there is no arbitrage, then the value V = -C0 x is the initial investment needed for replication and this value is the same for all replicating portfolio strategies x.

64

M Kallio and WT Ziemba M. Kallio, W T Ziemba I Journal of Banking & Finance 31 ( 2007) 2281-2302

2288

The market is complete, if any contingent claim v can be replicated. The following observation follows from elementary linear algebra applied to nC = 0; i.e. Lk>onkCk = -no Co. Lemma 2. Assume the set II in (9) is non-empty. Then

II consists of a single state price vector n

vectors

c," .fln· k > 0, are linearly independent

the market is complete.

As a special case, if a non-singular matrix results from C when the row vector C0 of the root node is omitted, then the state price vector n E II is unique, and the market is complete. However, it is possible, that a contingent claim v can be replicated even if the market is incomplete as the following result shows. Theorem 5. If no arbitrage opportunities exist, then any admissible price V of a contingent claim u satisfies v- セ@ v セ@ v+, where v- and v+ are obtainedfrom the linear programming problems

L nkVk, v- =mil]. L nkvh v+

=

ュセ@

nEil

nEn

( 11)

k>ll

(12)

k>O

where TI is the closure of IT. If v can be replicated, then

v- = v+ = V

The replication principle for valuation was first employed by Black and Scholes ( 1973) for option pricing. By Lemma 2, if the market is complete, replication is possible and there is a single price vector in II. In case v- < v+, any price V E ( v-, V+) is admissible. However, the end points v- and v+ may or may not be admissible; see Karatzas and Kou (1996). The duals of (11) and (12) are (13) v+ =min{ -c0 lc = Cx,ck ? vk'Vk > 0}, X

v- =max{ -c0 lc =

Cx, ck セ@

vk'Vk

> 0}.

(14)

X

Hence, the bound v+ is equivalently given by the smallest initial investment -c0 of a portfolio strategy x for which the net cash flow c superreplicates the contingent claim; i.e. ck ? Vto for all k > 0. Similarly, a subreplicating portfolio strategy with ck セ@ vk defines v- in (14). Bertsimas et al. (2001) proposed an E-arbitrage approach for approximate replication, which can provide option prices free of arbitrage in an incomplete market. 3.3. General option 1:aluation

Consider an option allowing choices a in a feasible set A. If a strategy a E A is chosen for executing the option, then a contingent claim v(a) = (vk(a)) is obtained. Assume, that

Chapter 4. Using Tucker’s Theorem of the Alternative

65

M. Kallio, W T. Ziemba I Journal of Banking & Finance 31 (2007) 2281-2302

2289

is the cash flow at the set of feasible contingent claims v(a) is compact. For k > 0, カセ」H。I@ node k, and for the root node, V0 = -v0 , independent of a, is the price of the option. For equilibrium prices, besides no-arbitrage, we also assume that each buyer of the option is not willing to pay more than the smallest initial investment needed in a portfolio strategy whose cash flow c superreplicates any contingent claim v(a) (i.e., 」セ@ セ@ カセ」H。IL@ fork> 0), for a E A; A price higher than such value does not necessarily provide an arbitrage opportunity, because the seller may not know, which strategy a E A the buyer chooses. Suppose that v(a) can be replicated, for all a E A; e.g., the market is complete. Then the value of the option is uniquely determined with any n E II as

2.:= nkvk(a).

Vo = セク@

(15)

k>O

To see this, recall that Lk>onkvk(a) is the investment needed to replicate v(a), and let a* E A denote a maximizing strategy in (15). If the option price is less than V0 , then buying one option and selling short a replicating portfolio of v(a*) provides an arbitrage opportunity. On the other hand, by our assumption, no agent is willing to pay a price higher than V0 . If v( a) cannot be replicated, for all a E A, then upper and lower bounds Vri and V 0, respectively, are obtained for V0 (16)

and (17)

The arguments are as above. Let a* E A denote a maximizing strategy in (17). By ( 12) and (14), V 0 is the maximal investment subreplicating v(a*). Hence, if the option price is less than V0, then selling short such a portfolio subreplicating v(a*) and buying the option leads to an arbitrage. Furthermore, no agent is willing to pay a price higher than Vri, because any contingent claim z;(a), for a E A, can be superreplicated at a cost not exceeding Limits analogous to (16) and (17) were first proposed by Harrison and Kreps (1979). The upper limit in (16) is found by optimization, while the lower limit in (17) is obtained as a saddle value, provided one exists (Rockafellar, 1974, p. 3).

vt.

3.4. Impact ofperiodic dividends and interest payments

For each node k tf. T and for all s E sセ」L@ let D, denote a vector of dividends per share and interest payments. Then the net cash flow vector c = H」セI@ is defined by - Poxo

=

+ Dk)xk (Pk + Dk)xk (Pk

for k

c0 =

ck

=

0,

\fk E T,

- Pkxk = ck

otherwise

or by (4) with matrix C including dividends and interest payments; see Fig. 3. Theorem 2, with definition of Arbitrage A, remains valid with Eq. (5) replaced by nkPk = z=n,(D, +P,) ウesセイ@

\fk

tf. T.

(18)

66

M Kallio and WT Ziemba M. Kallio, W. T. Ziemha I Journal of Banking & Finance 31 (2007) 2281-2302

2290

-Po D, +P1

C=

-P, D3+P3

Fig. 3. The matrix C in (4) with dividends and interest payments for the scenario tree of Fig. I.

4. Imperfect markets

This section introduces various imperfections in the model. The first subsection discusses proportional transaction costs for selling and buying. Other imperfections might be studied similarly one by one. However, to shorten the discussion, we directly study the simultaneous impacts of several market frictions. Taxes are not discussed. Our simple activity analysis does not apply to account for taxes in a serious way. There are several reasons for this: rules of tax laws may involve nonlinearities, discontinuities etc.; taxation is country (region) dependent; and taxation rules change occasionally over time. 4.1. impact of transaction costs

In this subsection, we assume proportional transaction costs of buying and selling. For simplicity, transaction costs of closing terminal positions (converting terminal wealth into cash) are omitted. Such an omission may be justified by using a utility function accounting for terminal wealth. The impact of relaxing this assumptions is discussed in Section 4.2. Furthermore, transaction costs of short selling any given quantity are assumed the same as transaction costs of selling the same quantity to reduce long position. Similarly, transaction costs of reducing short positions by any given quantity are assumed the same as transaction costs of buying the same quantity to increase long positions. The vector of prices at node k is Pk. Let the vector Yt ): 0 be the quantities bought and the vector Yi: ): 0 be the quantities sold at node k, for all k if_ T. The quantity Yt can be interpreted as an increase in long position or as a reduction in short position. Similarly, the quantity Yi: can be interpreted as a reduction in long position or as an increase in short position. It will become clear shortly, that the outcome of our analysis is independent of such alternative interpretations. For all nodes k and instruments i, assume that the fractional shares c);k and A;k, with 0 :::_: 6;10 A;k < 1, are incurred in transaction costs while buying and selling, respectively. Defining diagonal scaling matrices

yields the price vectors at node k, including transaction costs, for buying and selling, respectively, as

(19) The net cash flow at node k is Ck

= -Ptyt

+ Piji:

Vk if_ T,

(20)

Chapter 4. Using Tucker’s Theorem of the Alternative M. Kallio, W. T. Ziemba I Journal of Banking & Finance 31 (2007 j 2281-2302 -

p_0

P.+ 0

-P{

p-

p3 p4

-P3 -P•

1

-

C=

p3 p4 Ps p6

-P3 -P• -Ps -P6

p,+ 2

p5 p6

67 2291

p,2

-Ps -P6

Fig. 4. The matrix C in (22) with transaction costs for the scenario tree of Fig. I.

and the wealth at a terminal node k is Ck = pk

2)Yt- y;:)

(21)

Vk E T,

kEBk

where Bk is the set of nodes preceding k E Tin the path from terminal node k to the root node 0. The net cash flows ck both in (20) and in (21) are the same irrespective of quantities y;: being interpreted as short sales or not. In matrix notation, the net cash flow vector c having one component for each node is (22)

Cy=c,

where y = (y;:, y;:) セ@ 0 is a vector with a pair y;, y;: of subvectors for each k セ@ T, and Cis a matrix defined by (20) and (21); see Fig. 4. Arbitrage B. An arbitrage opportunity exists if c セ@ 0 and c =f. 0 in (22), for some vector y セ@ 0. If Llk = Ak =I, then Arbitrage B is equivalent to Arbitrage A. To study the case with transaction costs we employ the following special case of Tucker's Theorem 1 with A 2 =I and A 3 void. Lemma 3. For each given matrix A 1, either Atz セ@

0,

Atz

ic 0

and z

セ@

0 has a solution

or nA 1

セ@

0 and n > 0 has a solution

but never both.

If an Arbitrage B does not exist, then Lemma 3 implies existence of a state price vector > 0, with n0 = 1, such that nC セ@ 0. Taking into account the structure of C in (22), yields n = (nk)

nkP;: セ@

L n,P, セ@

nkPt

Vk セ@ T,

(23)

tETk

where Tk c T is the set of terminal nodes succeeding node k. Hence, Theorem 6. Arbitrage B does not exist k, with n 0 = 1, ウ。エゥセヲケョァ@ (23).

if and only if there exists a node price ョセ」@ > 0, for

all

This result is shown by Naik (1995) assuming that a risk free asset exists and there are no transaction costs on this asset. If b;k = ).;k = 0, for all i; i.e. P;: = Pt = Pk, then (23) is equivalent with recursive application of (5). For transaction cost parameters b;10 A;k > 0, Theorem 6 involves a set of positive state price vectors n that can be used for valuation

68 2292

M Kallio and WT Ziemba ,71,[, Kallio,

W: T. Ziemha I Journal of Banking & Finance 31 (2007) 2281-2302

of contingent claims employing standard software; see e.g., Murtagh and Saunders (1978). Valuation formulas (ll) and (12) and (16) and (17) for contingent claims and general options in Section 3.2 and Section 3.3 remain valid if the set of state price vectors II in (9) is replaced by the set of vectors n satisfying (23) with n > 0, and n 0 = 1. Similarly as in (13) and (14), duality can be applied to obtain bounds v- and v+ on the price V of a contingent claim with a cash flow vector v = (vk). In this case, v+ is the smallest initial investment needed in an investment strategy defined by y > 0 satisfying (22) such that c superreplicates v; i.e. ck > v"' for all k > 0. The lower bound v- is the largest initial net cash flow c0 created by such an investment strategy y such that -c subreplicates v; i.e. -ck :( Vk, for all k > 0. For Arbitrage B, we obtain the following discrete equivalent of the result of Jouini and Kallal (1995a). While Jouini and Kallal characterize valuation of contingent claims using martingale measures, they do not explicitly consider evaluation.

?g

> O,for all k, Proposition 1. If there is an asset (a numeraire) with a strictly positive price then the absence of Arbitrage B is equivalent to the existence of an equivalent probability measure that transforms some process Pk between the bid and ask price processes PJ: and Pt of traded securities (after normalization) into a martingale. Proof. For a state price vector n > 0 with n 0 = 1 satisfying (23) define a price process Pk as follows. For each terminal node k E T, h = Pk. For all other nodes k, define Pk such that セォー@ = 'E_,.Es,n,P,. Recursively we obtain nkPk = 'E_,ET,n,P,. Hence, by (23) PJ: :( Pk :( PJ:. Repeating the arguments for Theorem 4, given there is an asset (a numeraire) with a strictly positive price pセ@ > 0, for all k, define normalized prices Pk = pォOセ@ and parameters u, = HョLpセIOォN@ Then Pk = Lscs,usPs, Lscs,uk = 1, and u,>O. Hence, the normalized price process Pk is a martingale. Consequently, if an Arbitrage B does not exist, then there is a price process F\ with PJ: :( F\ :( PJ:, such that the normalized price process Pk is a martingale. Conversely, suppose that a price process Pk and the probabilities 1'55 > 0 exist such that PJ: :( Pk :( Pt, pセ@ > 0 for the numeraire process, and the normalized price process Pk is a martingale. Then forward recursion ns = オLョォpzOセ@ for s E Sk, with n 0 = 1 yields state prices nk > 0 satisfying (23). D

4.2. Combined impact of impe1jections

Next, consider simultaneously several market imperfections: (i) transaction costs as above, (ii) an interest rate spread between borrowing and lending, (iii) charges for short positions, and (iv) restricted short selling. As in Napp (2003), additional restrictions on the net positions can be easily imposed. For example, limits may be given on the value of net positions of some class of assets relative to the value net positions in another class. Such constraints arc discussed but excluded from the model, because of their subjective nature. King (2002) studies a pricing model of imperfect markets with proportional transaction costs, restricted short selling and charges for shorting. However, periodic holding costs for short positions or benefits (interest and dividend payments) for long positions are excluded. We discuss in detail the imperfect market model on King in Appendix. Edirisinghe et a!. ( 1993) study superreplication of a contingent claim under imperfect markets in a binary tree framework involving a risky asset and a riskless bond. They propose a

Chapter 4. Using Tucker’s Theorem of the Alternative M. Kallio, W T. Ziemba I Journal of Banking & Finance 31 (2007) 2281-2302

69 2293

dynamic programming approach allowing nonlinear transaction costs, lot-size constraints, and position limits for trading. As in Section 4.1, we assume proportional transaction costs of buying and selling. Transaction costs of short selling are assumed the same as the transaction costs of reducing long positions, and transaction costs of reducing short positions are assumed the same as transaction costs of buying. However, we now apply an additional holding cost for short positions and restrict short positions relative to long positions. Unlike in Arbitrage B, for the definition of an arbitrage, all wealth at terminal nodes may be converted into cash taking into account transaction costs. We also include dividends and periodic interest payments for lending and borrowing as in Section 3.4. Independence of imperfections is not assumed. Again, for each node k, Pk is the vector of prices at node k. Price vectors Pt and F;, including transaction costs, are defined by (19) such that Pf: ? Pf:. The vector Yt ? 0 denotes the quantities bought and the vector y;: ? 0 the quantities sold at node k, for all k rf_ T. Again, the vector Yt is interpreted as an increase in long positions or as a reduction in short positions and yf: is a reduction in long positions or as an increase in short positions. It is unnecessary to keep track of these alternative interpretations. Let xk = xt - x;: denote the vector of quantities held in each instrument at node k, with xt ,x;: ? 0 referring to long and short positions, respectively. For each k rf_ T, and for each and D; as follows. For long positions x;:, ? 0 is the vector of s E S,o define vectors cash value per share provided by all instruments. This vector consists of dividends and interest, and if s is a terminal node, then also the selling price P; is included inn;. For short positions x;:, D; is composed of dividends, interest, shorting costs. For terminal nodes, the buying price is also included. The borrowing and lending components in D; and D; are interpreted as interest payments. Interest rates for short positions are assumed to be at least as large as for long positions. The shorting cost can be infinite, if the short position is prohibited. Hence, for all s =I= 0, assume that

n;

n;

P;

(24) The net cash flow at the root node is (25)

For nodes k > 0 and k rf_ T ck =

-Ptyt + Pf:yf: + Dtxt - Df:xf: .

Finally, for terminal nodes k ck =

E

(26)

T

Df:xt - Df:xf: .

(27)

Hence, the net cash flow vector c having one component for each node is

Ax+By = c,

(28)

where x = (xt, x;:) ? 0 is a vector with a pair x;:, x;: of subvectors for each k rf_ T, y = (yt,y;:) ? 0 is a vector with a pair yt,y;: of subvectors for each k rf_ T, and A and B are matrices defined by (25)-(27).

The quantities held at the root node 0 satisfy

-xt + xo +Yo -Yo

=

o,

(29)

70

M Kallio and WT Ziemba M. Kallio, W T Ziemha I .Journal of Banking & Finance 31 (2007) 2281 2302

2294

and for all other nodes k

-xt + xT;

Ft T

+ xt - xT; + Yt - Yi = 0.

(30)

ln matrix notation, Cx+Dy=O,

(31)

with C and D defined by (29) and (30). For nodes k Ft T, a margin requirement limits short positions PtxT: to a share p::;; 1 of long positions PT;xt; i.e.

>o

pPT;xt- Ptxz

'ik

et T.

(32)

For p = 1, (32) requires that the total liquidation value of all assets and liabilities is nonnegative; i.e., (32) represents a solvability condition. ln matrix notation (32) is Ex ): 0.

(33)

Obviously, redefining the matrix E in (33), we can account for additional restriction on the positions xt and xz. For example, as in Napp (2003), we may restrict positions to be nonnegative or non-positive, the value of net positions in some class of assets (such as bonds) may be restricted by a given share of total net value, the net number of some asset in short position (e.g., a European call) can be restricted by the net number of another asset in long position (e.g., the underlying stock). While the consequence of such modifications in Theorem 7 below is straightforward, it is omitted because the data involved in such constraints is largely subjective. The structure of the model of (28)-(33) is shown in Fig. 5. Arbitrage C. An arbitrage opportunity exists if c ): 0 and c =I= 0 in (28), for vectors x ): 0 andy ): 0 satisfying (33) and (31). ln a perfect market Arbitrage C is equivalent to Arbitrage A. At each node k, (31) allows long and short positions to be incremented by the same quantity, but we must ensure that both xt and X/: remain non-negative. Reasonable interpretations exist. For example, one might simultaneously reduce both long and short position without transaction costs. On the other hand, we might consider a revised model (28), (31) and (33), defining the increments in long and short positions by separate vectors. Such a revised model, which is shown in Appendix, exhibits an arbitrage if and only if the original model does.

x()

xt D+1 D+2

-I I I pPo

xi

x;-

xi

X2

Yt Yo -Po

-Dl -

-

D-2

Yi Y2

p+ p1 1 -

D+ -D:J 3 D+ -D4 4

I -I -I

Yi Y!

Po

p,2

I

-I

D+5 -D[; D+6 -D(;

I -I

p,+ 2

I

-I I

-I

I

pP2-

_p2+

-I

-Po pP,-

-P,+

Fig. 5. The structure of the model of (28), (33) and (31) for the scenario tree of Fig. I.

Chapter 4. Using Tucker’s Theorem of the Alternative M. Kallio. WT. Ziemba I Journal of Banking & Finance 31 (2007) 2281-2302

71 2295

To analyze the case with mixed imperfections, we employ another special case of Tucker's Theorem 1. Lemma 4. Let A 1 , A 2 and A 3 be given matrices, with A 1 non-vacuous. Then Atz;? 0,

Atz

# 0,

A2z;? 0,

z ;? 0,

and A3z = 0 has a solution z

or nA 1 + vA 2 + pA 3 セ@

0,

n

> 0,

v ;? 0 has a solution n, v, p

but never both.

If Arbitrage C does not exist, then Lemma 4 with A 1 = [A, B], A 2 = [E, OJ and A 3 = [C,D] with submatrices A, B, C, D and E in (28), (33) and (31), implies existence of a state price vector n = (nk) > 0, with n 0 = 1, a vector v = (vk) ;? 0, and a vector p = (!lk), with components vk and subvectors Ilk. for all k (j!. T, such that nA

+ vE + pC セ@

nB+ pD

セ@

(34)

0,

(35)

0.

Taking into account the structure of matrices A, B, C, D and E, (34) becomes (36) with Ps

=0, for s

nkYi: セiャォ@

セ@

E

T, and (35) becomes

nkPt

Vk (j!. T.

(37)

Hence, we conclude with the following new result: Theorem 7. Arbitrage C does not exist if and only if there exists a node price vector n v ;? 0 and p satisfying (36) and (37), with n0 = 1.

> 0,

We discuss valuation procedures shortly. To help understand and interpret Theorem 7, we first discuss a number of special cases, starting with the most simple ones. In these cases the margin requirement is omitted, so that vk is zero in (36). (a) The basic model (Section 3.1). We haver;= Pt = Pk. for all k, D'i: = Dt = 0, for = Pk, for k E T. The condition (37) implies flk = nkPk so (36) k (j!. T and DJ; =

n;

yields the basic state price valuation result (5). (b) The model with dividends and interest payments (Section 3.4). Again PJ; = Pt = Pk, for all k. For the cash payments, DJ; = Dt = Dk. fork (j!. T, and DJ; = Dt = Pk + Dk, fork E T. Then, condition (37) implies flk = nkh· Hence, (36) yields the result (18). (c) The model with transaction costs (Section 4.1). Let PJ; セ@ Pk セ@ Pt and DJ; = Dt = 0, fork (j!. T, and PJ; = Pt = Pk =Dr; = Dt, for terminal nodes k E T. Then, fork such that Sk c T, condition (36) implies Jlk = セウfL@ n,P,.. For all other nodes k (j!. T, (36) so that with backward recursion p, = セエetォ@ n 1P,. This with implies Ilk = セウeォャL@ (37) implies (23). Related models have been studied by Jouini and Kallal (1995a) (for more details, see Section 4.1 above), Leland (1985), Zhao and Ziemba (2007), Boyle and Vorst (1992) and Gilster and Lee (1984).

72 2296

M Kallio and WT Ziemba M. Kallio, W. T Ziemba I Journal of Banking & Finance 31 I 2007) 2281-2302

(d) Restricted short and long positions with interest rate spread. This model is a discretized version of the model of J ouini and Kallal ( 199 5b). Consider a finite set of marketed instruments i with prices pセN@ There are no transaction costs so that Pi = Y;; =Ph for all k. Instrument i = 0 refers to lending and borrowing. We set PZ = 1, for all k. Fork rf. T, the interests of lending and borrowing are paid in cash and given by DZ I = rt and = rr;' respectively. For k E T, the face value is included as well, so that dセK@ = 1 + イセ@ and = 1 + rr;. To avoid arbitrage, イセ@ :( rr;. For all other instruments i, either short positions are prohibited with dセM = oo or long positions are prohibited with dセK@ = -oo. In the former case we set dセK@ = 0, for k rf. T, and dセK@ = pセL@ for k E T, and in the latter case dセM = 0, for k rf. T, and dセM =Pi, fork E T. lfboth short and long positions are possible for some securities, such instruments are duplicated to fit in the model. lf there is no Arbitrage C, then condition (37) implies Jlk = nkPk, for all k rf. T. For i = 0, (36) yields セウesL@ n,(l + r;) :( nk :( セウe Q L@ n,(l + r;), for all k rf. T. Hence, for some interest rate process h, such that rt :( h :( rr;, we have nk = セウes」ョLHャ@ + r,), or equivalently, 1 ::::' セウeL@ O"" where O", = ( 1 + rs)n,/"!.k > 0. Define recursively a numeraire process Rk as follows. For the root node, R 0 = 1, and for other nodes, Rk = (1 + rk)Rk , where k_ denotes the predecessor node of node k. Then the nor= oo then malized price process is Pk = Pk/Rk. For instruments i =I= 0, if dセM nkPi ;? セウes」@ n,P; by (36), or equivalently, pセ@ ;? セウesL@ oBLpセ[@ i.e., pセ@ is a supermartingale. Similarly, if dセK@ = -oo, then ョォpセ@ :( セウ・」s@ ョLpセ@ and Pic :( セウ・」L@ oBLpセ@ so that P/, is a submartingale. It is straightforward to show the converse. Hence in conclusion, an Arbitrage C does not exist if and only if there is some interest rate process h, with rt :( h :( rr;, an associated numeraire process Rk. and an equivalent probability measure such that the normalized price processes pセL@ fori =I= 0, are supermartingales or submartingales. This is the main result in Jouini and Kallal (1995b). (e) The model with holding costs and benefits; no transaction costs. Assume Pi. = Pi = Pk and DJ. ;? Di, for all k. In this case, different interest rates for lending and borrowing as well as costs for short positions may be included in the model. Such a case in continuous time has been studied by Bergman (1995). Condition (37) implies Jlk = nkPk. Hence, for k such that sk c T, (36) implies セウesォョLdMG[@ :( nkPk :( セウesLョdN@ For other nodes k rf. T, (36) implies セウes」ョLHdMG[@ +P,) :( nkPk :( セウesL@ ョLHdセ@ + P,). For terminal nodes k E セ「ッエィ@ Di and DJ. ゥョセャオ、・@ the price comーッョ・セ@ Pk. Hence, there exists a process Dk such エセ。@ n,; :( Dk :( DJ., for k rf. T, D,; :( Dk + Pk :( DJ., for k E T, and nkPk = セウ・」@ n,(D, + P,), for k rf. T. The latter equation provides a frictionless condition (18) for the absence of Arbitrage C. For further discussion, see case (f) where (e) is a special case. (f) The model with transaction costs, and holding costs and benefits. Assume PI_ :( Pk :( Pi and n-;; ;? DJ., for all k. ln this case, define l\ such that nkPk = JLk, fork rf. T, and Pk = Pk. fork E T. Then condition (37) and our assumption imply Pi. :( Pk :(Pi, for all k. Hence, for k such that Sk c T, (36) implies セウeLョャN⦅AMG[@ :( nJ) :( セウes」ョd[N@ F£r other nodes k rf. T, (36) implies セウeゥィョHdMG[@ + Ps) :( nkPk :( セウesL@ ョLHdセ@ + P,). Consequently, absence of Arbitrage C implies existence of a pro-

nz-

cess jjk with Di :( jjk :( Dt;, for k

nz-

rf. T, Dt;

:( jjk

+ h :( Dt;, fork E

T, such that

Chapter 4. Using Tucker’s Theorem of the Alternative M. Kallio, W. T Ziemha I Journal

o/ Banking & Finance 31 (2007) 2281-2302

73 2297

This equation is a frictionless condition (18) for the absence of Arbitrage C One may readily check, that the converse is true with flk = nJ; b for k rf_ T, in Theorem 7, In conclusion, an Arbitrage C does not exist if and only if there is n > 0 with n 0 = 1, and a price process F\ with PJ; セ@ F\ セ@ Pt satisfying (38) for some process lh such that Dt セ@ lh セ@ DJ;, for k rf_ T, and Dt セ@ lh + i\ セ@ DJ;, for k E T Of course, interpretation of (38) in terms of risk neutral or martingale valuation is straightforward. Hence our conclusion extends earlier results. We now return to the general case of Theorem 7, Similarly as before, valuation formulas (11), (12) and (16), (17) for contingent claims and general options in Section 3.2 and Section 3,3 remain valid if the set of state price vectors II in (9) is replaced by the set of vectors n satisfying (36) and (37) with n > 0, n 0 = 1, and v ;_, 0; for an illustration, see Example 1 below. Optimization codes, such as Minos (Murtagh and Saunders, 1978), can compute bounds v- and vt on the price of a contingent claim. Because II is a convex set, every price outside the interval [ v-, 0] leads to an Arbitrage C, and there is no Arbitrage C for price levels in the interior of this interval; see also Karatzas and Kou ( 1996), As in ( 13) and (14 ), duality can be applied to obtain bounds v and vt on the admissible prices V of a contingent claim with a cash flow vector v = ( カセ」IL@ The upper limit 0 is the smallest initial investment needed in an investment strategy defined by x ;_, 0 and for all y ;_, 0 satisfying (28), (33) and (31) such that c superreplicates v; i.e. ck ;_, カセッ@ k > 0. The lower limit v- is the largest initial net cash flow c0 created by such an investment strategy x andy such that -c subreplicates v; i.e. -ck セ@ カセ」L@ for all k > 0. However, the replication result of valuation in frictionless markets is no longer valid; a result first pointed out by Bensaid et al. (1992). Example 2 demonstrates the case. We now show that a frictionless price is in the price interval obtained with frictions, Suppose an Arbitrage C does not exist in a frictionless case (b), including (a) as a special case, Consider n > 0 with n 0 = 1 and Jlk in Theorem 7 satisfying (36) and (37), Then Jlk = nkPk by (37) and nkPk = LsES, n,(D, + P,) by (36), fork rf_ T Next, consider the case (f) with frictions such that PJ; セ@ Pk セ@ Pt, for all k, Dt セ@ Dk セ@ DJ;, for k rf_ T, and Dt セ@ Dk + Pk セ@ DJ;, fork E T Then nkPJ; セ@ nkPk = flk セ@ nkPt so that (37) holds for case (f). To check that (36) holds as well, we obtain for k such that sセ」@ C T, LsESkn.,D-'; セ@ LsESk ns(D, + Ps) = nkPk = flk セ@ LsES, nsD;, and for all other k rf_ T, LsES, ( nsD; + Jl,) セ@ LsES; n,(D, + P,) = nkPk = flk セ@ LsES, (n,D; + flsl· Hence, if n and Jlk for case (b) satisfy (36) and (37) in Theorem 7, then they apply for case (f) as well. Consequently, the values (or the value, in case it is unique) obtained for a contingent claim in the frictionless case are within the bounds v and 0 obtained with transaction costs and holding costs; see Jouini and Kallal (1995a) and Karatzas and Kou ( 1996). Example 1. Consider an example of Hull (2002, pp. 395-396). A 5-month American put

option on a non-dividend paying stock, with current stock price 50 and annual volatility 40%, has an exercise price 50. The risk ti-ee annual interest rate is 10°1 v+. There is a superreplicating strategy for which the initial investment vt is less than the initial investment in any replicating strategy. Consequently, any price V based on replication results in existence of arbitrage. 5. Conclusions

Tucker's theorem of the alternative provides a simple tool to prove arbitrage results in frictionless cases and when there are transaction costs and other frictions in the finite case. The approach yields known results directly and some new results and is a useful approach for further research. Appendix. An alternative model of Section 4.2

Consider a revised model (28), (31) and (33) of Section 4.2. and define the increments in long and short positions by separate vectors. For each node k, let P; denote the price vector for shorting. As in King (2002), P; may be obtained from the selling price vector PJ: by subtracting some extra charge for shorting. Hence, we have P: ,;::; PJ: ,;::; Pt for shorting, selling and buying prices. However, if shorting charges only include holding costs of short for s E Sk), then pセ@ = PJ: as positions in each period (included in holding cost vectors dセL@ in Section 4.2. Hence, in this case shorting charges are modeled similarly as interest payments. For the revised model, the solvability conditions (33) remain unchanged, while the cash balance Eq. (28) and position dynamics Eq. (31) are reformulated as follows. For each k rt T, Yt ? 0 andy;: ? 0 denote the increase and decrease in long positions, respectively. For short positions, such increments are denoted by st ? 0 and sJ; ? 0. After such increments, ? 0 and x;: ? 0 are the long and short positions, respectively. Eq. (28) is replaced by cash balance equations, for all k, as follows

xt

(39)

xt_ x

For the root node 0, initial positions are = 0 = 0, and for terminal nodes k E T, positions are closed so that xt = x;: = 0, y;: = xt , s;: = x;:_ and Yt = st = 0. Eq. (31) is replaced by separate equations for long and short positions. For all nodes k we have

xt - xt + Yt - Yi:

=

0

(40)

Chapter 4. Using Tucker’s Theorem of the Alternative

75 2299

M. Kallio, W T. Ziemha I Journal of Banking & Finance 31 (2007) 2281-2302 x;i

X

-

xi

x;-

x;-

xi

y+

s;;-

y-

s+

-Po

-Po

Po

P.* 0

yj -

p-

_p+

p-

-PJ

3

4

p5 p-

I

p+ 1

p-

P.* 1

1

Yt

s2

-Pi -Pi

Y2

si

p-

P.* 2

2

-I I

-I I

-I I

I

-I I

-I I

-I I

-I pPo

sj

_p't

I -I

-

Y1

_p+

-I I

p+ 1

s;-

-I I

-I

-Po

pP1-

-P{ pP2-

-Pi

Fig. 6. The structure of the revised model of King (2002) for the scenario tree of Fig. I.

and

-x-;; + xf: +sf: - st For nodes k

=

0.

(41)

rf. T, the margin requirement for short positions is (42)

For this model, consider the case with P: = Pf: and periodic short pos1t1on costs included in holding cost vectors Df:. If an arbitrage opportunity exists in the revised model, then such an arbitrage strategy immediately defines an arbitrage opportunity in the original model of Section 4.2. On the other hand, suppose that an arbitrage opportunity exists in the original model. Then, because Pt ;;;, Pf:, Df: ;;;, Dt, for all k, and p セ@ I, it is straightforward to check that reducing at each node k both positions xt and x-;; by max{xt,x;;} (the maximum component-wise), results in another arbitrage strategy. In the latter strategy, no asset is held simultaneously long and short, and hence this arbitrage provides an arbitrage in the revised model. In conclusion, the revised model, exhibits an arbitrage if and only if the original model does. Finally, we consider the model of King (2002, Section 7). Assume that asset i = 0 is a numeraire (bank account) with a strictly positive price, for all k. Suppose that all prices PZ, Pf: and Pt are given in terms of the buying price of the numeraire (using this price as the monetary unit). Additionally, assume that there are no transaction costs for the numeraire, but there may be a shorting charge. Hence, the scaled prices of the numeraire satisfy pセJ@ セ@ ーセM = ーセK@ = l, for all k. All shorting charges are taken into account in the shorting prices so that P; セ@ Pf: セ@ Pt and periodic shorting costs are excluded. There are no dividend nor interest payments so that Dt = Df: = 0, for all k rf. T. For terminal nodes k E T, prices applied for closing positions define Dt = Pf: and Df: = Pt. The structure of this model is shown in Fig. 6. For this model, by Lemma 4, there is no arbitrage if and only if there exists a state prices nk > 0 for cash balance equations, for all k, with n 0 = 1, parameters vk ;;;, 0 for solvability requirements, row vectors flt for long positions, and row vectors flk' for short positions such that, for all k rf. T,

76

M Kallio and WT Ziemba M. Kallio, W T. Ziemha I Journal of Banking & Finance 31 (2007) 2281-2302

2300

(43)

fl"k セ@

v"Pt

+ RIョウヲャセ@

+ サャセIL@

(44)

sESk

with {t.;

= !(

=0, for s

E T,

and

nkP;; セ@ fli @セ nkPi, nkP; セ@ tt."k セ@ nkPi.

(45) (46)

Hence, conditions (43)-(46) are now in place of conditions (36) and (37) of Theorem 7. As in the discussion following Theorem 7, we provide the following interpretation for conditions (43)-(46). F_£r all k rf_ T, 、セヲゥョ・@ hand Pk such that flt = nkPk and fll: = nkPk. セッイ@ all k E T, define Pk = PT; and Pk = PT;. Then, with (45) and (46). we ィ。カ・イ[セ@ Pk セ@ and pセ@ セ@ Pk セpゥL@ for all k. Given that Di = D;; = 0, for all k rf_ T, and Di = PT; = Pk and DT; = Pt = 7\, fork E T, conditions (43) and (44) are restated as

?.;;

nkPk ;::, pvkP"J;

+L

nsP"

(47)

sESk

and

nkPk セ@

vkPi

+L

nsPs.

(48)

ウセsォ@

Let Q denote the (real) probability measure associated with the scenario tree, and for all k, let qk > 0 be the probability of node k. For all k, define parameters 'rk such that nk = qkYk· At time t, let y, denote the random parameter with realizations yセッ@ for all k at stage t. To interpret yセッ@ consider a bond paying 1 (the value of the numeraire) in each node at time t. Then the (scaled) value of the bond at time 0 is セォョ@ = EQ[yJ Hence, ;, may be interpreted as a stochastic discounting factor for scaled prices. Employing the definition of Ylo (47) and (48) become

qk·rkPk

> pv"P;; + L

qdi"

(49)

sESk

and

qk'fkl\ セ@

vkPt

+L

qsysPs.

(50)

sESk

In conclusion, if there is no arbitrage, then there exists price processes Pk E [P;;, Pi] and E [P;, PtJ, a stochastic discounting factory, and parameters vk ;::, 0 such that (49) and (50) hold. It is straightforward to show the converse: starting with Pk. Pk, Yk and vk, define nk = qk'Yio flt = nkPk and fll: = nkPk to satisfy conditions (43) and (46). These conclusions correspond to Theorem 5 in King (2002). To interpret this result, consider the case with solvability conditions omitted. In this case, corresponding to Theorem 4 of King (2002), the process y}, E [y,P;, y,P7J is a supermartingale and the process y,P, E [y,P;, ;,Pj] is a submartingale, where for each t, P; セ@ P, セ@ Pi and P; セ@ P, セ@ Pi denote random price vectors with realizations PJ; セ@ Pk セ@ Pt and P; セ@ Pk セ@ P;;, for k at stage t. For the numeraire, ーセM = pセ@ = ーセK@ = 1 and pセJ@ セ@ pセ@ セ@ ーセK@ = 1, for all k. Hence, (49) and (50) yield qkyk > セウesLアケ@ and qkYkPZ セ@

7\

Chapter 4. Using Tucker’s Theorem of the Alternative M. Kallio. WT. Ziemba I Journal of Banking & Finance 31 (2007) 2281-2302

アウケpセN@

77 2301

Consequently, the discounting process ·y 1 is a supermartingale, y0 = n0/q 0 = セ@ RZウesッアLケヲUセ@ セ@ lZ[ォetアケpセ@ セ@ L:;kETqkyk = EQ[; 1 ]. Hence, EQ[YtJ is noincreasing in t, and pセJ@ セ@ EQ[·;,] セ@ 1. Consider a bond paying 1$ at time t and let /1 1 denote the inverse of the $-price of the numeraire at time t with realizations f3k for nodes k at t. Then the scaled value of the bond at t is {J 1, and the value of the bond at time 0 is V, = L:;knkfJk = EQ[{J,y 1]. If /3 1 is deterministic, then l/{3 1 is the (total) risk free return from time 0 to t, and V 1 = /3 1EQ[y 1] is a risk free discounting factor, the value of the risk free bond. Due to friction costs, V1 セ@ {J 1, because EQ[y 1] セ@ 1. Following Ingersoll (2006), V1 is interpreted as a subjective risk free discounting factor. I:sES

I and

pセJ@

セ@ pセ@

References Arrow, K., 1970. Essays in the Theory of Risk Bearing. North-Holland, London. Bensaid, B., Lesne, J., Pages, H., Scheinkman, J., 1992. Derivative asset pricing with transaction costs. Mathematical Finance 2, 63-86. Bergman, Y., 1995. Option pricing with differential interest rates. Review of Financial Studies 8, 475500. Bertsimas, D., Kogan, L., Lo, A.W., 2001. Hedging derivative securities and incomplete markets: An c-arbitrage approach. Operations Research 49, 327-397. Boyle, P., Vorst, T., 1992. Option replication in discrete time with transaction costs. Journal of Finance 47,271293. Bjork, T., 1998. Arbitrage Theory in Continuous Time. Oxford University Press, Oxford. Black, F., Scholes, M., 1973. The pricing of options and corporate liabilities. Journal of Political Economy 81, 637-654. Constantinides, G., 1976. Multiperiod consumption and investment behavior with convex transaction costs. Management Science 25, 1127-1137. Cox, J., Ross, S., 1976. The valuation of options for alternative stochastic processes. Journal of Financial Economics 3, 145-166. Cox, J., Ross, S., Rubinstein, M., 1979. Option pricing: A simplified approach . .lollrnal of Financial Economics 7, 229-264. Cvitanic, J ., Karatzas, 1., 1996. Hedging and portfolio optimization under transaction costs: A martingale approach. Mathematical Finance 6. 133-165. Cvitanic, J., 2001. Theory of portfolio optimization in markets with frictions. Tn: Jouini, E., Cvitanic, J., Musiela, M. (Eds.), Option Pricing Interest Rates and Risk Management, Handbooks in Mathematical Finance. Cambridge University Press. Davis, M., Norman, A., 1990. Portfolio selection with transaction costs. Mathematics of Operations Research 15, 676-713. Debreu, G., 1959. Theory of value. Tn: Cowles Foundation Monograph, vol. 17. Yale University Press, New Haven. Dermody, J.C., Prisman, E.Z., 1993. No arbitrage and valuation in markets with realistic transactions costs. Journal of Financial and Quantitative Analysis 28, 65-80. Dybvig, P.H., Ross, S.A., 1992. Arbitrage. In: Newman, P., Milgate, M., Eatwell, J. (Eds.), The New Palgrave Dictionary of Money and Finance. Palgrave Macmillan, London. Duffie, D., 2001. Dynamic Asset Pricing Theory, third cd. Princeton University Press, Princeton. Edirisinghe, C., Naik, V., Uppal, R., 1993. Optimal replication of options with transaction costs and trading restrictions. Journal of Financial and Quantitative Analysis 28, 117-138. Garman, M.B., Ohlson, J.A., 1981. Valuation of risky assets in arbitrage free economies with transaction costs. Journal of Financial Economics 9, 271-280. Gilster, J., Lee, W., 1984. The effect of transaction costs and different borrowing and lending rates on the option pricing model. Journal of Finance 39, 1215-1221. Harrison, J.M., Kreps, D.M., 1979. Martingales and arbitrage in multi-period securities markets. Journal of Economic Theory 20, 381-408. Hull, J., 2002. Options, Futmes and Other Derivatives, fifth ed. Prentice Hall.

78 2302

M Kallio and WT Ziemba M. Kallio, W. T Ziemba I Journal of Banking & Finance 31 (2007) 2281-2302

Ingersoll, J.E., 2006. The subjective and objective evaluation of incentive stock options. Journal of Business 79, 453-487. Jouini, E., Kallal, H., 1995a. Martingales and arbitrage in securities markets with transaction costs. Journal of Economic Theory 66, 178 197. Jouini, E., Kallal, H., 1995b. Arbitrage in securities markets with short-sales constraints. Mathematical Finance 5. 197-232. Karatzas, 1., 1996. Lectures on the Mathematics of Finance. American Mathematical Society, Providence. Karatzas, 1., Kou, S.G., 1996. On the pricing of contingent claims under constraints. The Annals of Applied Probability 6, 321 369. King, A., 2002. Duality and Martingales: A stochastic programming perspective on contingent claims. Mathematical Programming, Series B 91, 543-562. King, A., Koivu, M., Pennanen, T., 2002. Calibrating Option Bounds, Working Paper, Helsinki School of Economics. Luenberger, D.G., 1998. Investment Science. Oxford University Press, New York. Leland, H., 1985. Option pricing and replication with transaction costs. Journal of Finance 40, 1283-1301. Magill, M., Constantinides, G., 1976. Portfolio selection with transaction costs. Journal of Economic Theory 13, 264-271. Mangasarian, O.L., 1969. Nonlinear Programming. McGraw-Hill, New York. Merton, R., 1973. Theory of rational option pricing. Bell Journal of Economics and Management Science 4, 141-183. Merton, R., 1992. Continuous Time Finance, seconded. Basil Blackwell, Oxford and Cambridge. Murtagh, B., Saunders. M., 1978. Large-scale linearly constrained optimization. Mathematical Programming 14. 41-72. Naik, V., 1995. Finite state securities market models and arbitrage. In: Jarrow, R.A., Maksimovic, V., Ziemba, W.T. (Eds.), Finance, Handbooks in Operations research and management Science, vol. 9. Elsevier, pp. 31-64. Napp, C., 2001. Pricing issues with investment flows, applications to market models with frictions. Journal of Mathematical Economics 35, 383-408. Napp, C., 2003. The Dalang-Morton-Willinger theorem under cone constraints. Journal of Mathematical Economics 39, 111-126. Pliska, S.R., 1997. Introduction to Mathematical Finance: Discrete Time Models. Blackwell Publishers, Oxford. Prisman, E.Z., 1986. Valuation of risky assets in arbitrage free economies with frictions. Journal of Finance 41, 545-557. Rockafellar, R.T., 1974. Conjugate Duality and Optimization. Society for Industrial and Applied Mathematics, Philadelphia. Ross, S.A., 1976. The arbitrage theory of capital asset pricing. Journal of Economic Theory 13, 341-360. Ross, S.A., 1977. Return, risk, and arbitrage. In: Friend, 1., Bicksler, J.L. (Eds.), Risk and Return in Finance. Ballinger Publishing company, Cambridge, MA. Ross, S.A., 1978. A simple approach to the valuation of risky streams. Journal of Business 51, 453-475. Ross, S.A., 1987. Arbitrage and martingales with Laxation. Journal of Political Economy 95, 371-393. Stiemke, E., 1915. Uber positive Liisungen homogener linearer Gleichungen. Mathematische Annalen 76, 340342. Tucker, A., 1956. Dual systems of homogeneous linear relations. In: Kuhn, H., Tucker, A. (Eds.), Linear Inequalities and Related Systems, Annals of Mathematics Studies, vol. 38. Princeton University Press, Princeton. Zhao, Y., Ziemba, W.T., 2003. On Leland's option pricing and hedging strategy with transaction costs. Working Paper, Nanyang Business School, Nanyang Technological University. Zhao, Y., Ziemba, W.T., 2007. On Leland's option pricing and hedging strategy with transactions costs. Finance Research Letters 4, 49-58.

Section B

Utility Theory

80

Section B. Utility Theory

Decision makers in financial markets are faced with a variety of opportunities and must decide how much capital to allocate to the various assets in the opportunity set at points in time. Investing in assets produces returns (gains/losses), resulting from the changes in trading prices. The returns on a unit of capital invested in assets are uncertain, that is, the return vector is a random variable. The basic information input to the decision on how much to invest in each asset is the distribution for the return vector. In Section A, some models for the dynamics of the stochastic return vector were considered. Assuming that the return distribution is known, the investment decisions are based on preferences for changes in wealth or accumulated wealth. To structure the decision process, a theory of preferences is required. The theory of preferences concerns the ability to represent a preference structure with a real-valued function. This has been achieved by mapping it to the mathematical index called utility. To put the preference relation for an individual into a theory of utility the following axioms were proposed by Von Neumann and Morgenstern (1944) and Savage (1954). Let S be the set (possibly infinite) of alternatives for a system each having a monetary payoff with a known probability. There are four axioms of the expected utility theory that define a rational decision maker. They are completeness, transitivity, independence and continuity. Completeness: For any two alternatives A and B in S, either A is preferred to B or B is preferred to A or there is indifference between the alternatives. Transitivity: For alternatives A, B, and C, if A is preferred to B and B is preferred to C, then A is preferred to C. Continuity: For alternatives A, B and C, if A is preferred to B is preferred to C, then there exists a probability π such that B is as good as (indifferent to) πA + (1 − π)C. Independence: For alternatives A, B and C, with A preferred to B, for α ∈ (0, 1], αA + (1 − α)C is preferred to αB + (1 − α)C. If the four preference axioms are satisfied then the preference relationship can be expressed in terms of a utility function u(X(A)), where X(A) is the random monetary payoff from an alternative A and FX(A) is the distribution for X(A). Expected Utility Theorem: For any two alternatives A and B in S, A is preferred to B if and only if Eu(X(A)) > Eu(X(B)) and there is indifference iff Eu(X(A)) = Eu(X(B)). A general proof of the expected utility theorem is provided in the article by Fishburn (1969). The fact that preferences can be defined by a utility function is very useful for the analysis of preferences and the decision problem of choosing the best alternative. In following this approach it is important to keep in mind that it is assumed that decision makers satisfy the axioms in stating their preferences. This is referred to as rational decision making (Savage, 1954).

Section B. Utility Theory

81

Expected utility theory implies that rational individuals act as though they were maximizing expected utility, and that allows for the fact that many individuals are risk averse, meaning that the individual would refuse a fair gamble (a fair gamble has an expected value of zero). If X(A) is a random outcome and X(B) is a random outcome with distribution equal to that of X(A) + ε, where ε is uncorrelated noise, then X(A) is preferred to X(B) by any risk inverse  αaverter.  −1 With the  cumulative −1 −1 −1 and FX(B) , let T (α) = 0 FX(A) (p) − FX(B) (p) dp = the area distributions FX(A) between the distributions in the α tail. Then T (α) ≥ 0, 0 ≤ α ≤ 1, T (1) = 0, which follows from the greater variability of X(B) around the same mean as X(A). So there is a class of decision makers who are averse to risk as characterized by greater uncertainty. Going back to the expected utility theorem, this class has a particular type of utility function. A utility function u for which:  α  −1  −1 FX(A) (p) − FX(B) (p) d ≥ 0, 0 ≤ α ≤ 1, implies Eu(X(A)) ≥ Eu(X(B)) 0

is a concave function (Rothschild and Stiglitz, 1970). The area/integration condition implies that the utility has decreasing first derivatives (negative second derivatives). So the risk averter has a concave utility. In fact the degree of aversion at an outcome level is defined by the size of the second derivative relative to the first derivative (Pratt, 1964; Arrow, 1965). The risk aversion implied by expected utility theory has a shortcoming in that it does not provide a realistic description of risk attitudes to modest stakes. To have realistic risk aversion for large stakes produces virtual risk neutrality for moderate ones. Rabin (2000) presents a theorem that calibrates a relationship between risk attitudes over small and large stakes. The theorem shows that, within the expected-utility model, anything but virtual risk neutrality over modest stakes implies manifestly unrealistic risk aversion overlarge stakes. For example, a person who would for any initial wealth turn down 50–50 lose $1,000/gain $1,050 bets would always turn down 50–50 bets of losing $20,000 or gaining any sum. With utility function u and initial wealth of $20,000, the first bet implies on a wager of $1,000 that 0.5u(19,000) + 0.5u(21,050) < u(20,000) and therefore on a wager of $20,000, it follows that 0.5u(0) + 0.5u(220,000) < u(20,000). In this sense, expected utility theory can be misleading when analyzing situations involving modest stakes. A decision that maximizes expected utility also maximizes the probability of the decision’s consequences being preferable to some uncertain threshold (Castagnoli and LiCalzi, 1996; Bordley and LiCalzi, 2000; Bordley and Kirkwood, 2004). In the absence of uncertainty about the threshold, expected utility maximization simplifies to maximizing the probability of achieving some fixed target. If the uncertainty is uniformly distributed, then expected utility maximization becomes expected value maximization. Intermediate cases lead to increasing risk-aversion above some fixed threshold and increasing risk-seeking below a fixed threshold. There are examples of choice problems where preferences do not satisfy the axioms. The Allias paradox (Allias, 1952), the Ellsberg paradox (Ellsberg, 1961), and the Bergen paradox (Allias and Hagen, 1979) provide well known contradictions

82

Section B. Utility Theory

to the expected utility theorem. The Allias paradox has received particular attention. From the independence axiom, for alternatives A, B and C, with A preferred to B, for α ∈ (0, 1] it follows that αA + (1 − α)C is preferred to αB + (1 − α)C. However, it has been experimentally demonstrated that the preference for αA+(1−α)C over αB + (1 − α)C can be reversed depending on the value of α. The independence axiom is the key to the linearity in probabilities of the expected utility and that property is violated in the Allias paradox. In expected utility theory, the utilities of outcomes are weighted by their probabilities. It has been shown that people overweight outcomes that are considered certain relative to outcomes which are merely probable, a phenomenon which is labelled the certainty effect. There have been a variety of proposals for dealing with the violation of the independence axiom and the linearity in probabilities. One approach is Prospect Theory (PT) proposed by Kahneman and Tversky (1979). The bilinear form of expected utility is retained, but probabilities and outcomes are transformed. The value of a prospect, denoted V , is expressed in terms of two scales, π and v. The first scale, π, associates with each probability p a subjective decision weight π(p), which reflects the impact of p on the over-all value of the prospect. The second scale, v, assigns to each outcome x a number, v(x), which reflects the subjective value of that outcome. The outcomes are defined relative to a reference point, which serves as the zero point of the value scale. Hence, v measures the value of deviations or changes from that reference point. The value function is S-shaped, being convex for losses (x < 0) and concave for gains (x > 0). This idea dates to Markowitz (1952) who commented on the Friedman–Savage (1948) utility functions leading to the S-shape. The basic equation of the theory describes the manner in which π and v are combined to determine the over-all value of regular prospects. In a simple case, if (x, p; y, q) is a prospect then the value of the prospect is V (x, p; y, q) = π(p)v(x) + π(q)v(y), where v(0) = 0, π(0) = 0, and π(1) = 1. This equation generalizes expected utility theory by relaxing the expectation principle. An axiomatic analysis of this representation is provided in Kahneman and Tversky (1979). Prospect Theory has its critics. Levy and Levy (2002, 2004) cast doubt on the S-shaped value function, based on experimental results. The class of all prospect theory value functions are S-shaped with an inflection point at x = 0. Thus, v ′ > 0, v(x)′′ > 0 for x < 0, and v(x)′′ < 0 for x > 0. This is contrasted with the class of all Markowitz (1952) value functions which are reverse S-shaped with an inflection point at x = 0. Thus, v ′ > 0, v ′′ (x) > 0 for x > 0, and v ′′ (x) < 0 for x < 0. Markowitz’s function, like the prospect theory value function, depends on change of wealth. Levy and Levy (2002) define Prospect Stochastic Dominance (PSD) and the Markowitz Stochastic Dominance (MSD). If Prospect F dominates Prospect G by PSD, then F is preferred over G by any prospect theory S-shaped value function. Markowitz stochastic dominance rule (MSD) corresponds to all reverse S-shaped value functions. PSD and MSD are opposite if the two distributions have the same

Section B. Utility Theory

Figure 1.

83

Non-concave utilities.

mean: Let F and G have the same mean. Then F dominates G by PSD if and only if G dominates F by MSD. Levy and Levy (2002) conducted a set of decision making experiments to determine if the decision behavior of subjects conforms to prospect theory. The focus of their analysis is the S-shape of the value function in prospect theory and contrasted that function with the Markowitz reverse S-shape. With the stochastic dominance approach they take the weighting function as π(p) = p so the probabilities are not transformed. In each of the Levy and Levy experiments the subjects decision behavior supported the reverse S-shape value function and they considered that as evidence against prospect theory, or at least the S-shape value function. By defining stochastic dominance (PSD, MSD) based on the value function while retaining the original probability distribution, the weighting function of prospect theory is not a factor. In Kahneman and Tversky, the bi-criteria (value, probability) are both transformed. That is, the decision maker distorts both dimensions. To illustrate this affect, Wakker (2003) analyzes the Levy and Levy experiments using decision weights to transform probabilities based on assumptions of Tversky and Kahneman. The result is that the observed decision behavior in the experiments is consistent with the S-shape of prospect theory. In the experiments conducted by Levy and Levy neither of the competing gambles were preferred by second order stochastic dominance (SSD). Baltussen et al. (2006) augmented the Levy and Levy tasks with a third gamble which by SSD is preferred to either original gamble, with all gambles having the same expectation. The empirical evidence supports SSD as opposed to either PT or MSD. The support for the S-shaped function as obtained by Kahneman and Tversky (1979) is actually due to the certainty effect. Obviously, different individuals have

84

Section B. Utility Theory

Figure 2.

Risk Matrix.

different preferences. Classes of functions: concave, S-shaped, reverse S-shaped, represent contrasting perspectives on risk. There is evidence to support different utilities for gains and losses, which are defined with respect to a reference point. When considering gains and losses separately the utility is concave for gains and convex for losses (Abdellaoui, 2000, 2007) as implied by prospect theory. However, evidence indicates individuals are less sensitive to probability differences when choosing among mixed gambles than when choosing among either gain or loss gambles (Wu and Markle, 2008). Interesting perspective on risk assessment comes from the “risk matrix” used in reliability engineering, where value and probability are on a log scale (Figure 2). The classification of risk is based on the combination of outcome and probability. There are criteria for a consistent classification defined in Cox (2008). The stochastic dominance definitions provide a way to characterize consistent preferences for risk. A standard technique for defining risk preferences is meanvariance analysis (Markowitz, 1952). Assuming a normal distribution for outcome returns, the mean-variance rule is consistent with expected utility and second order stochastic dominance (concave utility). So MV and Prospect theory are not compatible when considering the preferences between two alternatives. However, Levy and Levy (2004) establish that the efficient sets (un-dominated) from second order dominance (MV) and prospect stochastic dominance are almost identical when dealing with a mixture of sets of alternatives (diversified portfolios). The value and utility functions to this point have been atemporal, being concerned with preferences between alternatives. In financial decision making the system is dynamic and the alternatives consist of outcomes such as consumption at points in time. For example, in discrete time the outcome could be the consumption stream C = (c1 , . . . , ct , . . . , cT ). The standard approach to valuing C is to

Section B. Utility Theory

85

T have V (C) = t=1 ρt−1 u(ct ), for the discount factor ρ. With this form of intertemporal additive and homogeneous utility, the expected utility theory translates readily. However, the problems giving rise to non-expected utility apply to this intertemporal format. An additional problem is that the two distinct aspects of preference, inter-temporal substitutability and relative risk aversion, are intertwined; indeed the elasticity of substitution and the risk aversion parameter are reciprocals of one another. Epstein and Zin (1989) consider the inter-temporal utility issues. They define a general class of preferences which is sufficiently flexible to permit those two aspects of preference to be separated. The utility V is recursive, so that V satisfies the following equation: V (c1 , . . . , cT ) = W (c1 , µ(c2 , . . . , cT )), where W is an increasing aggregator function and µ is a certainty equivalent. They require that the aggregator function have a constant elasticity of substitution. The certainty equivalent can be any member of a broad class of mean value functionals. A form of continuity is required of the functional. Epstein and Zin (1989) establish the existence of the inter-temporal utility V following from the aggregator W and functional µ.

References Abdellaoui, M (2000). Parameter-free elicitation of utility and probability weighting functions. Management Science, 46(11), 1497–1512. Abdellaoui, M, H Bleichrodt and P Corina (2007). Loss aversion under prospect theory: A parameter-free measurement. Management Science, 53(10), 1659–1674. Allais, M (1953). Le comportement de l’homme rationnel devant le risque: Critique des postulatset axiomes de l’´ecole Am´ericaine. Econometrica, 21(4), 503–546. Allais, M and O Hagen (1979). Expected Utility Hypotheses and the Allias Paradox. Dordrecht, Holland: Reidel. Arrow, KJ (1965). The theory of risk aversion. In YJS Helsinki (Ed.), Aspects of the Theory of Risk Bearing. Baltussen, G, T Post and PV Vliet (2006). Violations of cumulative prospect theory in mixed gambles with moderate probabilities. Management Science, 52(8), 1288–1290. Bordley, R and C Kirkwood (2004). Preference analysis with multi-attribute performance targets. Operations Research, 52, 823–835. Bordley, RF and M LiCalzi (2000) Decision analysis using targets instead of utility functions. Decisionsin Economics and Finance, 23(1), 53–74. Castagnoli, E and M LiCalzi (1996). Expected utility without utility. Theory and Decision, 41, 281–301. Cox, LA (2008). What’s wrong with risk matrices? Risk Analysis, 28(2), 497–512. Ellsberg, D (1961). Risk, ambiguity, and the savage axioms. Quarterly Journal of Economics, 75(4), 643–669. Epstein, LG and SE Zin (1989). Substitution, risk aversion and the temporal behavior of consumption and asset returns: A theoretical framework. Econometrica, 57(4), 937–969. Fishburn, F (1969). A general theory of subjective probabilities and expected utilities, Annals of Mathematical Statistics, 40(4), 1419–1429.

86

Section B. Utility Theory

Friedman, M and LJ Savage (1948). Utility analysis of choices involving risk. Journal of Political Economy, 56(4), 279–304. Kahneman, D and A Tversky (1979). Prospect theory: An analysis of decisions under risk. Econometrica, 47(2), 263–291. Kahneman, D and A Tversky (1984). Choices, values, and frames. American Psychologist, 39(4), 314–350. Levy, M and H Levy (2002). Prospect theory: Much ado about nothing? Management Science, 48(10), 1334–1349. Levy, M and H Levy (2004). Prospect theory and mean-variance analysis. Review of Financial Studies, 17(4), 1015–1041. Pratt, JW (1964). Risk aversion in the small and in the large. Econometrica, 32, 122–136. Rabin, M (2000). Risk aversion and expected utility theory: A calibration theorem. Econometrica, 68(5), 1281–1292. Rothschild, M and J Stiglitz (1970). Increasing risk: 1. A definition. Journal of Economic Theory, 3(2), 225–243. Savage, LJ (1954). The Foundations of Statistics. New York: Wiley. Tversky, A and D Kahneman (1974). Judgment under uncertainty: Heuristics and biases. Science, 185(4157), 1124–1131. Von Neumann, J and O Morgenstern (1944). Theory of Games and Economic Behavior. Princeton, NJ: Princeton University Press. Wakker, P (2002). The Data of Levy and Levy (2002) “Prospect theory: Much ado about nothing?” Actually support prospect theory. Management Science, 49(7), 979–981. Wu, G and M Alex (2008). An empirical test of gain-loss separability in prospect theory. Management Science, 54(7), 1322–1335.

Chapter 5 T.he Annals of Mathematical Statiatics 1969, Vol. 40, No. 4, 1419-1429

A GENERAL THEORY OF SUBJECTIVE PROBABILITIES AND EXPECTED UTILITIES

BY

PETER

C.

FISHBURN

Research Analysis Corporation 1. Introduction. The purpose of this paper is to present a general theory for

the usual subjective expected utility model for decision under uncertainty. With a set S of states of the world and a set X of consequences let F be a set of functions on S to X. F is the set of acts. Under a set of axioms based on extraneous. measurement probabilities, a device that is used by Rubin [14], Chernoff [3], Luce and Rai_ffa [9, Ch. 13], Anscombe and Aumann [1], Pratt, Raiffa, and Schlaifer [11], Arrow [2], and Fishburn [5], we shall prove that there is a realvalued function u on X and a finitely-additive probability measure P* on the set of all subsets of S such that, for all j, g E F, (1)

f

セ@

g

if and only if E[u(f(s)), P*] セ@

E[u(g(s)), P*].

In (1 ), セ@ ("is not preferred to") is the decision-maker's binary preferenceindifference relation and E (y, z) is the mathematical expectation of y with respect to the probability measure z. Because we shall use extraneous measurement probabilities, (1) will be extracted from the more involved expression (2) that is presented in the next section. The axioms we shall use to derive (2) imply that P* is uniquely determined and that u is unique up to a positive linear transformation. u may or may not be bounded: however, it is bounded if there is a denumerable partition of S each element of which has positive probability under P*. Our theory places no restrictions on S and X except that they be nonempty sets with X containing at least two elements. X may or may not have a least (most) preferred consequence. In addition, no special restrictions are placed on P*. For example, if S is infinite, it may or may not be true that P* (A) = 1 for some finite subset A セ@ S, and if p* (A) < 1 for every finite A セ@ S it may or may not be true that S can be partitioned into an arbitrary finite number n of subsets such that P* = 1/n for each subset. Finally, no special properties will be implied for u apart from its uniqueness and its boundedness in the case noted above. To indicate briefly how this differs from other theories, we note first that the theories of Chernoff [3], Luce and Raiffa [9], Anscombe and Aumann [1], Pratt, Raiffa, and Schlaifer [11], and Fishburn [5] assume that S is finite. The エィ・ッセケ@ presented here is a generalization of a theory in Fishburn [5]. The theory of Davidson and Suppes [4] assumes that X is finite and implies that, if x < y and z < w and there is no consequence between x and y or between z and w then u (y) - u (x) = u (w) - u (z). The theories of Ramsey [13] and Suppes [16] Received 19 August 1968; revised 6 March 1969.

87

88

PC Fishburn 1420

PETER C. FISHBURN

place no special restrictions on S but they imply that X is infinite and that if u (x) < u(y) then there is a zc X such that u(z) = .5u (x) + .5u(y). On the other hand Savage [15] does not restrict X in any unusual way, but his theory requires S to be infinite and implies that, for any positive integer n, there is an n-part partition of S such that P* = 1/n on each part of the partition. Arrow [2) also assumes this property for p*. 2. Definitions and notation. u(w). That is, a prospect is acceptable if the utility resulting from integrating the prospect with one's assets exceeds the utility of those assets alone. Thus, the domain of the utility function is final states (which include one's asset position) rather than gains or losses. Although the domain of the utility function is not limited to any particular class of consequences, most applications of the theory have been concerned with monetary outcomes. Furthermore, most economic applications introduce the following additional assumption. (iii) Risk Aversion: u is concave (u" < 0). A person is risk averse if he prefers the certain prospect (x) to any risky prospect with expected value x. In expected utility theory, risk aversion is equivalent to the concavity of the utility function. The prevalence· of risk aversion is perhaps the best known generalization regarding risky choices. It led the early decision theorists of the eighteenth century to propose that utility is a concave function of money, and this idea has been retained in modern treatments (Pratt [33], Arrow [4]). In the following sections we demonstrate several phenomena which violate these tenets of expected utility theory. The demonstrations are based on the responses of students and university faculty to hypothetical choice problems. The respondents were presented with problems of the type illustrated below. Which of the following would you prefer? A:

50% chance to win 1,000,

B: 450 for sure.

50% chance to win nothing; The outcomes refer to Israeli currency. To appreciate the significance of the amounts involved, note that the median net monthly income for a family is about 3,000 Israeli pounds. The respondents were asked to imagine that they were actually faced with the choice described in the problem, and to indicate the decision they would have made in such a case. The responses were anonymous, and the instructions specified that there was no 'correct' answer to such problems, and that the aim of the study was to find out how people choose among risky prospects. The problems were presented in questionnaire form, with at most a dozen problems per booklet. Several forms of each questionnaire were constructed so that subjects were exposed to the problems in different orders. In addition, two versions of each problem were used in which the left-right position of the prospects was reversed. The problems described in this paper are selected illustrations of a series of effects. Every effect has been observed in several problems with different outcomes and probabilities. Some of the problems have also been presented to groups of students and faculty at the University of Stockholm and at the

Chapter 6. Prospect Theory: An Analysis of Decision Under Risk

101

265

PROSPECT THEORY

University of Michigan. The pattern of results was essentially identical to the results obtained from Israeli subjects. The reliance on hypothetical choices raises obvious questions regarding the validity of the method and the generalizability of the results. We are keenly aware of these problems. However, all other methods that have been used to test utility theory also suffer from severe drawbacks. Real choices can be investigated either in the field, by naturalistic or statistical observations of economic behavior, or in the laboratory. Field studies can only provide for rather crude tests of qualitative predictions, because probabilities and utilities cannot be adequately measured in such contexts. Laboratory experiments have been designed to obtain precise measures of utility and probability from actual choices, but these experimental studies typically involve contrived gambles for small stakes, and a large number of repetitions of very similar problems. These features of laboratory gambling complicate the interpretation of the results and restrict their generality. By default, the method of hypothetical choices emerges as the simplest procedure by which a large number of theoretical questions can be investigated. The use of the method relies on the assumption that people often know how they would behave in actual situations of choice, and on the further assumption that the subjects have no special reason to disguise their true preferences. If people are reasonably accurate in predicting their choices, the presence of common and systematic violations of expected utility theory in hypothetical problems provides presumptive evidence against that theory. Certainty, Probability, and Possibility In expected utility theory, the utilities of outcomes are weighted by their probabilities. The present section describes a series of choice problems in which people's preferences systematically violate this principle. We first show that people overweight outcomes that are considered certain, relative to outcomes which are merely probable-a phenomenon which we label the certainty effect. The best known counter-example to expected utility theory which d:ploits the certainty effect was introduced by the French economist Maurice Allais in 1953 [2]. Allais' example has been discussed from both normative and descriptive standpoints by many authors [28, 38]. The following pair of choice problems is a variation of Allais' example, which differs from the original in that it refers to moderate rather than to extremely large gains. The number of respondents who answered each problem is denoted by N, and the percentage who choose each option is given in brackets. PROBLEM

A:

1: Choose between 2,500 with probability

.33,

2,400 with probability

.66,

Owithprobability

.01;

N=72

[18]

B:

2,400 with certainty.

[82]*

102

D Kahneman and A Tversky

D. KAHNEMAN AND

266

A.

TVERSKY

PROBLEM 2: Choose between C:

2,500 with probability

.33,

0 with probability

.67;

D:

2,400 with probability

.34,

0 with probability

.66.

[83]*

N=72

[17]

The data show that 82 per cent of the subjects chose Bin Problem 1, and 83 per cent of the subjects chose C in Problem 2. Each of these preferences is significant at the .01level, as denoted by the asterisk. Moreover, the analysis of individual patterns of choice indicates that a majority of respondents (61 per cent) made the modal choice in both problems. This pattern of preferences violates expected utility theory in the manner originally described by Allais. According to that theory, with u(O) = 0, the first preference implies u(2,400) > .33u(2,500) + .66u(2,400) or .34u(2,400) > .33u(2,500) while the second preference implies the reverse inequality. Note that Problem 2 is obtained from Problem 1 by eliminating a .66 chance of winning 2400 from both prospects under consideration. Evidently, this change produces a greater reduction in desirability when it alters the character of the prospect from a sure gain to a probable one, than when both the original and the reduced prospects are uncertain. A simpler demonstration of the same phenomenon, involving only twooutcome gambles is given below. This example is also based on Allais [2]. PROBLEM 3: A:

(4,000,.80),

N=95

or

B:

(3,000). [80]*

[20]

PROBLEM 4: C: N

(4,000,.20), =

95

[65]*

or

D:

(3,000,.25). [35]

In this pair of problems as well as in all other problem-pairs in this section, over half the respondents violated expected utility theory. To show that the modal pattern of preferences in Problems 3 and 4 is not compatible with the theory, set u(O) = 0, and recall that the choice of B implies u(3,000)/u(4,000)>4/5, whereas the choice of C implies the reverse inequality. Note that the prospect C = (4,000, .20) can be expressed as (A, .25), while the prospect D = (3,000, .25) can be rewritten as (B,.25). The substitution axiom of utility theory asserts that if B is preferred to A, then any (probability) mixture (B, p) must be preferred to the mixture (A, p). Our subjects did not obey this axiom. Apparently, reducing the probability of winning from 1.0 to .25 has a greater effect than the reduction from

Chapter 6. Prospect Theory: An Analysis of Decision Under Risk PROSPECT THEORY

103 267

.8 to .2. The following pair of choice problems illustrates the certainty effect with non-monetary outcomes.

PROBLEM

5:

A:

50% chance to win a threeweek tour of England, France, and Italy;

N=72

B:

A one-week tour of England, with certainty.

[78]*

[22]

PROBLEM6:

C:

5% chance to win a threeweek tour of England, France, and Italy;

N=72

D:

[67]*

10% chance to win a oneweek tour of England.

[33]

The certainty effect is not the only type of violation of the substitution axiom. Another situation in which this axiom fails is illustrated by the following problems. PROBLEM7:

A:

(6,000, .45),

B:

N =66 [14] PROBLEM

C:

(3,000, .90).

[86]*

8: (6,000, .001),

N = 66 [73]*

D:

(3,000, .002).

[27]

Note that in Problem 7 the probabilities of winning are substantial (.90 and .45), and most people choose the prospect where winning is more probable. In Problem 8, there is a possibility of winning, although the probabilities o( winning are minuscule (.002 and .001) in both prospects. In this situation where winning is possible but not probable, most people choose the prospect that offers the larger gain. Similar results have been reported by MacCrimmon and Larsson [28]. The above problems illustrate common attitudes toward risk or chance that cannot be captured by the expected utility model. The results suggest the following empirical generalization concerning the manner in which the substitution axiom is violated. If (y, pq) is equivalent to (x, p ), then (y, pqr) is preferred to (x,pr),O (3,000, .25). [35] > (6,000, .45). [14] < (6,000, .001). [73]*

Negative prospects

Problem 3': N=95 Problem 4': N=95 Problem 7': N=66 Problem 8': N=66

(-4,000, .80) [92]* (-4,000, .20) [42] (-3,000, .90) [8] (-3,000, .002) [70]*

>

(-3,000). [8] < ( -3,000, .25). [58] < (-6,000, .45). [92]* > (-6,000, .001). [30]

In each of the four problems in Table I the preference between negative prospects is the mirror image of the preference between positive prospects. Thus, the reflection of prospects around 0 reverses the preference order. We label this pattern the reflection effect. Let us turn now to the implications of these data. First, note that the reflection effect implies that risk aversion in the positive domain is accompanied by risk seeking in the negative domain. In Problem 3', for example, the majority of subjects were willing to accept a risk of .80 to lose 4,000, in preference to a sure loss of 3,000, although the gamble has a lower expected value. The occurrence of risk seeking in choices between negative prospects was noted early by Markowitz [29]. Williams [48] reported data where a translation of outcomes produces a dramatic shift from risk aversion to risk seeking. For example, his subjects were indifferent between (100, .65; -100, .35) and (0), indicating risk aversion. They were also indifferent between (-200, .80) and (-100), indicating risk seeking. A recent review by Fishburn and Kochenberger [14] documents the prevalence of risk seeking in choices between negative prospects. Second, recall that the preferences between the positive prospects in Table I are inconsistent with expected utility theory. The preferences between the corresponding negative prospects also violate the expectation principle in the same manner. For example, Problems 3' and 4', like Problems 3 and 4, demonstrate that outcomes which are obtained with certainty are overweighted relative to uncertain outcomes. In the positive domain, the certainty effect contributes to a risk averse preference for a sure gain over a larger gain that is merely probable. In the negative domain, the same effect leads to a risk seeking preference for a loss

Chapter 6. Prospect Theory: An Analysis of Decision Under Risk PROSPECT THEORY

105

269

that is merely probable over a smaller loss that is certain. The same psychological principle-the overweighting of certainty-favors risk aversion in the domain of gains and risk seeking in the domain of losses. Third, the reflection effect eliminates aversion for uncertainty or variability as an explanation of the certainty effect. Consider, for example, the prevalent preferences for (3,000) over (4,000, .80) and for (4,000, .20) over (3,000, .25). To resolve this apparent inconsistency one could invoke the assumption that people prefer prospects that have high expected value and small variance (see, e.g., Allais [2]; Markowitz [30]; Tobin [41]). Since (3,000) has no variance while (4,000, .80) has large variance, the former prospect could be chosen despite its lower expected value. When the prospects are reduced, however, the difference in variance between (3,000, .25) and (4,000, .20) may be insufficient to overcome the difference in expected value. Because (-3,000) has both higher expected value and lower variance than (-4,000, .80), this account entails that the sure loss should be preferred, contrary to the data. Thus, our data are incompatible with the notion that certainty is generally desirable. Rather, it appears that certainty increases the aversiveness of losses as well as the desirability of gains. Probabilistic Insurance

The prevalence of the purchase of insurance against both large and small losses has been regarded by many as strong evidence for the concavity of the utility function for money. Why otherwise would people spend so much money to purchase insurance policies at a price that exceeds the expected actuarial cost? However, an examination of the relative attractiveness of various forms of insurance does not support the notion that the utility function for money is concave everywhere. For example, people often prefer insurance programs that offer limited coverage with low or zero deductible over comparable policies that offer higher maximal coverage with higher deductibles-contrary to risk aversion (see, e.g., Fuchs [16]). Another type of insurance problem in which people's responses are inconsistent with the concavity hypothesis may be called probabilistic insurance. To illustrate this concept, consider the following problem, which was presented to 95 Stanford University students. PROBLEM 9: Suppose you consider the possibility of insuring some property against damage, e.g., fire or theft. After examining the risks and the premium you find that you have no clear preference between the options of purchasing insurance or leaving the property uninsured. It is then called to your attention that the insurance company offers a new program called probabilistic insurance. In this program you pay half of the regular premium. In case of damage, there is a 50 per cent chance that you pay the other half of the premium and the insurance company covers all the losses; and there is a 50 per cent chance that you get back your insurance payment and suffer all the losses. For example, if an accident occurs on an odd day of the month, you pay the other half of the regular premium and your losses are covered; but if the accident

106

D Kahneman and A Tversky

270

D. KAHNEMAN AND A. TVERSKY

occurs on an even day of the month, your insurance payment is refunded and your losses are not covered. Recall that the premium for full coverage is such that you find this insurance barely worth its cost. Under these circumstances, would you purchase probabilistic insurance:

N=95

Yes, No. [20] [80]*

Although Problem 9 may appear contrived, it is worth noting that probabilistic insurance represents many forms of protective action where one pays a certain cost to reduce the probability of an undesirable event-without eliminating it altogether. The installation of a burglar alarm, the replacement of old tires, and the decision to stop smoking can all be viewed as probabilistic insurance. The responses to Problem 9 and to several other variants of the same question indicate that probabilistic insurance is generally unattractive. Apparently, reducing the probability of a loss from p to p/2 is less valuable than reducing the probability of that loss from p/2 to 0. In contrast to these data, expected utility theory (with a concave u) implies that probabilistic insurance is superior to regular insurance. That is, if at asset position w one is just willing to pay a premium y to insure against a probability p of losing x, then one should definitely be willing to pay a smaller premium ry to reduce the probability of losing x from p to (1- r)p, 0 < r < 1. Formally, if one is indifferent between ( w - x, p; w, 1- p) and ( w - y ), then one should prefer probabilistic insurance ( w- x, (1- r)p; w- y, rp; w - ry, 1- p) over regular insurance ( w- y ). To prove this proposition, we show that

pu(w -x)+(l-p)u(w) = u(w- y) implies (1- r)pu(w -x)+ rpu(w- y)+(l- p)u(w- ry) > u(w- y).

Without loss of generality, we can set u ( w - x) = 0 and u ( w) = 1. Hence, u ( wy) = 1- p, and we wish to show that

rp(1-p)+(l-p)u(w -ry)> 1-p

or

u(w-ry)>1-rp

which holds if and only if u is concave. This is a rather puzzling consequence of the risk aversion hypothesis of utility theory, because probabilistic insurance appears intuitively riskier than regular insurance, which entirely eliminates the element of risk. Evidently, the intuitive notion of risk is not adequately captured by the assumed concavity of the utility function for wealth. The aversion for probabilistic insurance is particularly intriguing because all insurance is, in a sense, probabilistic. The most avid buyer of insurance remains vulnerable to many financial and other risks which his policies do not cover. There appears to be a significant difference between probabilistic insurance and what may be called contingent insurance, which provides the certainty of coverage for a

Chapter 6. Prospect Theory: An Analysis of Decision Under Risk

PROSPECT THEORY

107

271

specified type of risk. Compare, for example, probabilistic insurance against all forms of loss or damage to the contents of your home and contingent insurance that eliminates all risk of loss from theft, say, but does not cover other risks, e.g., fire. We conjecture that contingent insurance will be generally more attractive than probabilistic insurance when the probabilities of unprotected loss are equated. Thus, two prospects that are equivalent in probabilities and outcomes could have different values depending on their formulation. Several demonstrations of this general phenomenon are described in the next section. The Isolation Effect In order to simplify the choice between alternatives, people often disregard components that the alternatives share, and focus on the components that distinguish them (Tversky [44]). This approach to choice problems may produce inconsistent preferences, because a pair of prospects can be decomposed into common and distinctive components in more than one way, and different decompositions sometimes lead to different preferences. We refer to this phenomenon as the isolation effect. PROBLEM 10: Consider the following two-stage game. In the first stage, there is a probability of .75 to end the game without winning anything, and a probability of .25 to move into the second stage. If you reach the second stage you have a choice between (4,000, .80)

and

(3,000).

Your choice must be made before the game starts, i.e., before the outcome of the first stage is known. Note that in this game, one has a choice between .25 x .80 = .20 chance to win 4,000, and a .25 x 1.0 = .25 chance to win 3,000. Thus, in terms of final outcomes and probabilities one faces a choice between (4,000, .20) and (3,000, .25), as in Problem 4 above. However, the dominant preferences are different in the two problems. Of 141 subjects who answered Problem 10,78 per cent chose the latter prospect, contrary to the modal preference in Problem 4. Evidently, people ignored the first stage of the game, whose outcomes are shared by both prospects, and considered Problem 10 as a choice between (3,000) and (4,000, .80), as in Problem 3 above. The standard and the sequential formulations of Problem 4 are represented as decision trees in Figures 1 and 2, respectively. Following the usual convention, squares denote decision nodes and circles denote chance nodes. The essential difference between the two representations is in the location of the decision node. In the standard form (Figure 1), the decision maker faces a choice between two risky prospects, whereas in the sequential form (Figure 2) he faces a choice between a risky and a riskless prospect. This is accomplished by introducing a dependency between the prospects without changing either probabilities or

108

D Kahneman and A Tversky

272

D. KAHNEMAN AND A. TVERSKY

3000

0 4000

FIGURE

FIGURE

1.-The representation of Problem 4 as a decision tree (standard formulation).

2.-The representation of Problem 10 as a decision tree (sequential formulation).

outcomes. Specifically, the event 'not winning 3,000' is included in the event 'not winning 4,000' in the sequential formulation, while the two events are independent in the standard formulation. Thus, the outcome of winning 3,000 has a certainty advantage in the sequential formulation, which it does not have in the standard formulation. The reversal of preferences due to the dependency among events is particularly significant because it violates the basic supposition of a decision-theoretical analysis, that choices between prospects are determined solely by the probabilities of final states. It is easy to think of decision problems that are most naturally represented in one of the forms above rather than in the other. For example, the choice between two different risky ventures is likely to be viewed in the standard form. On the other hand, the following problem is most likely to be represented in the sequential form. One may invest money in a venture with some probability of losing one's capital if the venture fails, and with a choice between a fixed agreed return and a percentage of earnings if it succeeds. The isolation effect implies that the contingent certainty of the fixed return enhances the attractiveness of this option, relative to a risky venture with the same probabilities and outcomes.

Chapter 6. Prospect Theory: An Analysis of Decision Under Risk

PROSPECT THEORY

109

273

The preceding problem illustrated how preferences may be altered セケ@ different representations of probabilities. We now show how choices may be altered by varying the representation of outcomes. Consider the following problems, which were presented to two different groups of subjects. PROBLEM 11: In addition to whatever you own, you have been given 1 ,000. You are now asked to choose between A:

(1,000, .50),

N=70

and

B:

(500).

[84]*

[16]

PROBLEM 12: In addition to whatever you own, you have been given 2,000. You are now asked to choose between C:

(-1,000, .50),

N = 68

and

D:

[69*]

(-500). [31]

The majority of subjects chose B in the first problem and C in the second. These preferences conform to the reflection effect observed in Table I, which exhibits risk aversion for positive prospects and risk seeking for negative ones. Note, however, that when viewed in terms of final states, the two choice problems are identical. Specifically, A= (2,000, .50; 1,000, .50)= C,

and

B = (1,500) =D.

In fact, Problem 12 is obtained from Problem 11 by adding 1,000 to the initial bonus, and subtracting 1,000 from all outcomes. Evidently, the subjects did not integrate the bonus with the prospects. The bonus did not enter into the comparison of prospects because it was common to both options in each problem. The pattern of results observed in Problems 11 and 12 is clearly inconsistent with utility theory. In that theory, for example, the same utility is assigned to a wealth of $100, 000, regardless of whether it was reached from a prior wealth of $95,000 or $105,000. Consequently, the choice between a total wealth of $100,000 and even chances to own $95,000 or $105,000 should be independent of whether one currently owns the smaller or the larger of these two amounts. With the added assumption of risk aversion, the theory entails that the certainty of owning $100,000 should always be preferred to the gamble. However, the responses to Problem 12 and to several of the previous questions suggest that this pattern will be obtained if the individual owns the smaller amount, but not if he owns the larger amount. The apparent neglect of a bonus that was common to both options in Problems 11 and 12 implies that the carriers of value or utility are changes of wealth, rather than final asset -positions that include current wealth. This conclusion is the cornerstone of an alternative theory of risky choice, which is described in the following sections.

110

D Kahneman and A Tversky

274

D. KAHNEMAN AND A. TVERSKY

3.

THEORY

The preceding discussion reviewed several empirical effects which appear to invalidate expected utility theory as a descriptive model. The remainder of the paper presents an alternative account of individual decision making under risk, called prospect theory. The theory is developed for simple prospects with monetary outcomes and stated probabilities, but it can be extended to more involved choices. Prospect theory distinguishes two phases in the choice process: an early phase of editing and a subsequent phase of evaluation. The editing phase consists of a preliminary analysis of the offered prospects, which often yields a simpler representation of these prospects. In the second phase, the edited prospects are evaluated and the prospect of highest value is chosen. We next outline the editing phase, and develop a formal model of the evaluation phase. The function of the editing phase is to organize and reformulate the options so as to simplify subsequent evaluation and choice. Editing consists of the application of several operations that transform the outcomes and probabilities associated with the offered prospects. The major operations of the editing phase are described below. Coding. The evidence discussed in the previous section shows that people normally perceive outcomes as gains and losses, rather than as final states of wealth or welfare. Gains and losses, of course, are defined relative to some neutral reference point. The reference point usually corresponds to the current asset position, in which case gains and losses coincide with the actual amounts that are received or paid. However, the location of the reference point, and the consequent coding of outcomes as gains or losses, can be affected by the formulation of the offered prospects, and by the expectations of the decision maker. Combination. Prospects can sometimes be simplified by combining the probabilities associated with identical outcomes. For example, the prospect (200, .25; 200, .25) will be reduced to (200, .50). and evaluated in this form. Segregation. Some prospects contain a riskless component that is segregated from the risky component in the editing phase. For example, the prospect (300, .80; 200, .20) is naturally decomposed into a sure gain of 200 and the risky prospect (100, .80). Similarly, the prospect (-400, .40; -100, .60) is readily seen to consist of a sure loss of 100 and of the prospect (-300, .40). The preceding operations are applied to each prospect separately. The following operation is applied to a set of two or more prospects. Cancellation. The essence of the isolation effects described earlier is the discarding of components that are shared by the offered prospects. Thus, our respondents apparently ignored the first stage of the sequential game presented in Problem 10, because this stage was common to both options, and they evaluated the prospects with respect to the results of the second stage (see Figure 2). Similarly, they neglected the common bonus that was added to the prospects in Problems 11 and 12. Another type of cancellation involves the discarding of common constituents, i.e., outcome-probability pairs. For example, the choice

Chapter 6. Prospect Theory: An Analysis of Decision Under Risk PROSPECT THEORY

111

275

between (200, .20; 100, .50; -50, .30) and (200, .20; 150, .50; -100, .30) can be reduced by cancellation to a choice between (100, .50; -50, .30) and (150, .50; -100, .30). Two additional operations that should be mentioned are simplification and the detection of dominance. The first refers to the simplification of prospects by rounding probabilities or outcomes. For example, the prospect (101, .49) is likely to be recoded as an even chance to win 100. A particularly important form of simplification involves the discarding of extremely unlikely outcomes. The second operation involves the scanning of offered prospects to detect dominated alternatives, which are rejected without further evaluation. Because the editing operations facilitate the task of decision, it is assumed that they are performed whenever possible. However, some editing operations either permit or prevent the application of others. For example, (500, .20; 101, .49) will appear to dominate (500, .15; 99, .51) if the second constituents of both prospects are simplified to (1 00, .50). The final edited prospects could, therefore, depend on the sequence of editing operations, which is likely to vary with the structure of the offered set and with the format of the display. A detailed study of this problem is beyond the scope of the present treatment. In this paper we discuss choice problems where it is reasonable to assume either that the original formulation of the prospects leaves no room for further editing, or that the edited prospects can be specified without ambiguity. Many anomalies of preference result from the editing of prospects. For example, the inconsistencies associated with the isolation effect result from the cancellation of common components. Some intransitivities of choice are explained by a simplification that eliminates small differences between prospects (see Tversky [43]). More generally, the preference order between prospects need not be invariant across contexts, because the same offered prospect could be edited in different ways depending on the context in which it appears. Following the editing phase, the decision maker is assumed to evaluate each of the edited prospects, and to choose the prospect of highest value. The overall value of an edited prospect, denoted V, is expressed in terms of two scales, 1r and v. The first scale, 1r, associates with each probability p a decision weight 1r(p ), which reflects the impact of p on the over-all value of the prospect. However, 1r is not a probability measure, and it will be shown later that 1r(p) + 1r(l- p) is typically less than unity. The second scale, v, assigns to each outcome x a number v (x ), which reflects the subjective value of that outcome. Recall that outcomes are defined relative to a reference point, which serves as the zero point of the value scale. Hence, v measures the value of deviations from that reference point, i.e., gains and losses. The present formulation is concerned with simple prospects of the form (x, p; y, q), which have at most two non-zero outcomes. In such a prospect, one receives x with probability p, y with probability q, and nothing with probability 1- p- q, where p + q セ@ 1. An offered prospect is strictly positive if its outcomes are all positive, i.e., if x, y > 0 and p + q = 1; it is strictly negative if its outcomes

112

D Kahneman and A Tversky

276

D. KAHNEMAN AND A. TVERSKY

are all negative. A prospect is regular if it is neither strictly positive nor strictly negative. The basic equation of the theory describes the manner in which 71' and v are combined to determine the over-all value of regular prospects. If (x, p; y, q) is a regular prospect (i.e., either p + q < 1, or x セ@ 0 セ@ y, or x セ@ 0 セ@ y), then (1)

V(x, p; y, q)

= 7r(p)v(x)+7r(q)v(y)

where v(O) = (}, 7r(O) = 0, and 71'(1) = 1. As in utility theory, V is defined on prospects, while v is defined on outcomes. The two scales coincide for sure prospects, where V(x, 1.0) = V(x) = v(x). Equation (1) generalizes expected utility theory by relaxing the expectation principle. An axiomatic analysis of this representation is sketched in the Appendix, which describes conditions that ensure the existence of a unique 71' and a ratio-scale v satisfying equation (1). The evaluation of strictly positive and strictly negative prospects follows a different rule. In the editing phase such prospects are segregated into two components: (i) the riskless component, i.e., the minimum gain or loss which is certain to be obtained or paid; (ii) the risky component, i.e., the additional gain or loss which is actually at stake. The evaluation of such prospects is described in the next equation. If p + q = 1 and either x > y > 0 or x < y < 0, then (2)

V(x, p; y, q)

= v(y)+7r(p)[v(x)-v(y)].

That is, the value of a strictly positive or strictly negative prospect equals the value of the riskless component plus the value-difference between the outcomes, multiplied by the weight associated with the more extreme outcome. For example, V(400, .25; 100, .75)= v(100)+7r(.25)[v(400)-v(100)]. The essential feature of equation (2) is that a decision weight is applied to the value-difference v (x)- v (y ), which represents the risky component of the prospect, but not to v (y ), which represents the riskless component. Note that the right-hand side of equation (2) equals 7r(p)v(x)+[1-7r(p)]v(y). Hence, equation (2) reduces to equation (1) if 7r(p) + 71'(1- p) = 1. As will be shown later, this condition is not generally satisfied. Many elements of the evaluation model have appeared in previous attempts to modify expected utility theory. Markowitz [29] was the first to propose that utility be defined on gains and losses rather than on final asset positions, an assumption which has been implicitly accepted in most experimental measurements of utility (see, e.g., [7, 32]). Markowitz also noted the presence of risk seeking in preferences among positive as well as among negative prospects, and he proposed a utility function which has convex and concave regions in both the positive and the negative domains. His treatment, however, retains the expectation principle; hence it cannot account for the many violations of this principle; see, e.g., Table I. The replacement of probabilities by more general weights was proposed by Edwards [9], and this model was investigated in several empirical studies (e.g.,

Chapter 6. Prospect Theory: An Analysis of Decision Under Risk PROSPECT THEORY

113

277

[3, 42]). Similar models were developed by Fellner [12], who introduced the concept of decision weight to explain aversion for ambiguity, and by van Dam [46] who attempted to scale decision weights. For other critical analyses of expected utility theory and alternative choice models, see Allais [2], Coombs [6], Fishburn [13], and Hansson [22]. The equations of prospect theory retain the general bilinear form that underlies expected utility theory. However, in order to accomodate the effects described in the first part of the paper, we are compelled to assume that values are attached to changes rather than to final states, and that decision weights do not coincide with stated probabilities. These departures from expected utility theory must lead to normatively unacceptable consequences, such as inconsistencies, intransitivities, and violations of dominance. Such anomalies of preference are normally corrected by the decision maker when he realizes that his preferences are inconsistent, intransitive, or inadmissible. In many situations, however, the decision maker does not have the opportunity to discover that his preferences could violate decision rules that he wishes to obey. In these circumstances the anomalies implied by prospect theory are expected to occur.

The Value Function An essential feature of the present theory is that the carriers of value are changes in wealth or welfare, rather than final states. This assumption is compatible with basic principles of perception and judgment. Our perceptual apparatus is attuned to the evaluation of changes or differences rather than to the evaluation of absolute magnitudes. When we respond to attributes such as brightness, loudness, or temperature, the past and present context of experience defines an adaptation level, or reference point, and stimuli are perceived in relation to this reference point [23]. Thus, an object at a given temperature may be experienced as hot or cold to the touch depending on the temperature to which one has adapted. The same principle applies to non-sensory attributes such as health, prestige, and wealth. The same level of wealth, for example, may imply abject poverty for one person and great riches for another-depending on their current assets. The emphasis on changes as the carriers of value should not be taken to imply that the value of a particular change is independent of initial position. Strictly speaking, value should be treated as a function in two arguments: the asset position that serves as reference point, and the magnitude of the change (positive or negative) from that reference point. An individual's attitude to money, say, could be described by a book, where each page presents the value function for changes at a particular asset position. Clearly, the value functions described on different pages are not identical: they are likely to become more linear with increases in assets. However, the preference order of prospects is not greatly altered by small or even moderate variations in asset position. The certainty equivalent of the prospect (1,000, .50), for example, lies between 300 and 400 for most people, in a wide range of asset positions. Consequently, the representation

114

D Kahneman and A Tversky

278

D. KAHNEMAN AND A. TVERSKY

of value as a function in one argument generally provides a satisfactory approximation. Many sensory and perceptual dimensions share the property that the psychological response is a concave function of the magnitude of physical change. For example, it is easier to discriminate between a change of 3° and a change of 6° in room temperature, than it is to discriminate between a change of 13° and a change of 16°. We propose that this principle applies in particular to the evaluation of monetary changes. Thus, the difference in value between a gain of 100 and a gain of 200 appears to be greater than the difference between a gain of 1,100 and a gain of 1,200. Similarly, the difference between a loss of 100 and a loss of 200 appears greater than the difference between a loss of 1,100 and a loss of 1,200, unless the larger loss is intolerable. Thus, we hypothesize that the value function for changes of wealth is normally concave above the reference point (v"(x) < 0, for x > O) and often convex below it (v"(x) > 0, for x < 0). That is, the marginal value of both gains and losses generally decreases with their magnitude. Some support for this hypothesis has been reported by Galanter and Pliner [17], who scaled the perceived magnitude of monetary and non-monetary gains and losses. The above hypothesis regarding the shape of the value function was based on response& to gains and losses in a riskless context. We propose that the value function which is derived from risky choices shares the same characteristics, as illustrated in the following problems. PROBLEM

13: (6,000, .25),

N=68 PROBLEM

or

(4,000, .25; 2,000, .25). [82]*

[18]

13': (-6,000, .25),

N =64

[70]*

or

(-4,000, .25; -2,000, .25). [30]

Applying equation 1 to the modal preference in these problems yields 1T(.25)v(6,000) < 1T(.25)[v(4,000) + v(2,000)]

and

1T(.25)v( -6,000) > 1T(.25)[v (-4,000) + v( -2,000)].

Hence, v(6,000) < v(4,000) + v(2,000) and v(-6,000) > v(-4,000) + v(-2,000). These preferences are in accord with the hypothesis that the value function is concave for gains and convex for losses. Any discussion of the utility function for money must leave room for the effect of special circumstances on preferences. For example, the utility function of an individual who needs $60,000 to purchase a house may reveal an exceptionally steep rise near the critical value. Similarly, an individual's aversion to losses may increase sharply near the loss that would compel him to sell his house and move to

Chapter 6. Prospect Theory: An Analysis of Decision Under Risk PROSPECT THEORY

115

279

a less desirable neighborhood. Hence, the derived value (utility) function of an individual does not always reflect "pure" attitudes to money, since it could be affected by additional consequences associated with specific amounts. Such perturbations can readily produce convex regions in the value function for gains and concave regions in the value function for losses. The latter case may be more common since large losses often necessitate changes in life style. A salient characteristic of attitudes to changes in welfare is that losses loom larger than gains. The aggravation that one experiences in losing a sum of money appears to be greater than the pleasure associated with gaining the same amount [17]. Indeed, most people find symmetric bets of the form (x, .50; -x, .50) distinctly unattractive. Moreover, the aversiveness of symmetric fair bets generally increases with the ·size of the stake. That is, if x > y ;30, then (y, .50; -y, .50) is preferred to (x, .50; -x, .50). According to equation (1), therefore, v(y) + v(-y) > v(x) + v( -x)

and

v(-y)- v( -x) > v(x)- v(y ).

Setting y=O yields v(x) p for low probabilities, there is evidence to suggest that, for all O 1T(p 1 )v(x) + 1T(q 1 )v(y ),

or 7T(p)-7r(pr)

v(y)

--''--'-----'=--'-> - - .

1T(q )-1T(q) 1

v(x)

Hence, as y approaches x, 1r(p) -7T(p 1 ) approaches 1r(q 1 ) -1r(q). Since p- p' = qr- q, 1r must be essentially linear, or else dominance must be violated. Direct violations of dominance are prevented, in the present theory, by the assumption that dominated alternatives are detected and eliminated prior to the evaluation of prospects. However, the theory permits indirect violations of dominance, e.g., triples of prospects so that A is preferred to B, B is preferred to C, and C dominates A. For an example, see Raiffa [34, p. 75]. Finally, it should be noted that the present treatment concerns the simplest decision task in which a person chooses between two available prospects. We have not treated in detail the more complicated production task (e.g., bidding) where the decision maker generates an alternative that is equal in value to a given prospect. The asymmetry between the two options in this situation could introduce systematic biases. Indeed, Lichtenstein and Slavic [27] have constructed pairs of prospects A and B, such that people generally prefer A over B, but bid more for B than for A. This phenomenon has been confirmed in several studies, with both hypothetical and real gambles, e.g., Grether and Plott [20]. Thus, it cannot be generally assumed that the preference order of prospects can be recovered by a bidding procedure. Because prospect theory has been proposed as a model of choice, the inconsistency of bids and choices implies that the measurement of values and decision weights should be based on choices between specified prospects rather than on bids or other production tasks. This restriction makes the assessment of v and 1r more difficult because production tasks are more convenient for scaling than pair comparisons.

4.

DISCUSSION

In the final section we show how prospect theory accounts for observed attitudes toward risk, discuss alternative representations of choice problems induced by shifts of reference point, and sketch several extensions of the present treatment.

Risk Attitudes The dominant pattern of preferences observed in Allais' example (Problems 1 and 2) follows from the present theory iff 7T(.33) 7T(.34)

--->

v(2,400) 7r(.33) > --'---'-v (2,500) l-7T(.66)'

Chapter 6. Prospect Theory: An Analysis of Decision Under Risk PROSPECT THEORY

121

285

Hence, the violation of the independence axiom is attributed in this case to subcertainty, and more specifically to the inequality 7T(.34) < 1- 1r(.66). This analysis shows that an Allais-type violation will occur whenever the v-ratio of the two non-zero outcomes is bounded by the corresponding 1r-ratios. Problems 3 through 8 share the same structure, hence it suffices to consider one pair, say Problems 7 and 8. The observed choices in these problems are implied by the theory iff 7T(.001) 7T(.002)

--->

v (3,000) 7T(.45) >---. v (6,000) 7T(.90)

The violation of the substitution axiom is attributed in this case to the subproportionality of 7T. Expected utility theory is violated in the above manner, therefore, whenever the v- ratio of the two outcomes is bounded by the respective 7T-ratios. The same analysis applies to other violations of the substitution axiom, both in the positive and in the negative domain. We next prove that the preference for regular insurance over probabilistic insurance, observed in Problem 9, follows from prospect theory-provided the probability of loss is overweighted. That is, if (- x, p) is indifferent to (-y ), then (-y) is preferred to (-x, p/2; -y, p/2; -y/2, 1-p). For simplicity, we define for f(x) = - v( -x). Since the value function for losses is convex, f is a concave x セPL@ function of x. Applying prospect theory, with the natural extension of equation 2, we wish to show that 1r(p )f(x) = f(y)

implies

f(y),;;, f(y/2) + 7T(p/2)[f(y)-f(y/2)] + 7T(p/2)[f(x)- f(y/2)]

= 7T(p/2)f(x) + 7T(p/2)f(y) + [1- 27T(p/2)Jf(y/2). Substituting for f(x) and using the concavity off, it suffices to show that 7T(p/2) f(y),;;, 7T(p) f(y) + 7T(p/2)f(y) + f(y )/2- 7T(p/2)f(y)

or 7T(p )/2,;;, 7T(p/2),

which follows from the subadditivity of 7T.

According to the present theory, attitudes toward risk are determined jointly by

v and 7T, and not solely by the utility function. It is therefore instructive to examine the conditions under which risk aversion or risk seeking are expected to occur. Consider the choice between the gamble (x, p) and its expected value (px ). If x > 0, risk seeking is implied whenever 1r(p) > v (px )/ v (x ), which is greater than p if the value function for gains is concave. Hence, overweighting (1r(p) > p) is necessary but not sufficient for risk seeking in the domain of gains. Precisely the same condition is necessary but not sufficient for risk aversion when x < 0. This analysis restricts risk seeking in the domain of gains and risk aversion in the domain of losses to small probabilities, where overweighting is expected to hold.

122

D Kahneman and A Tversky

286

D. KAHNEMAN AND A. TVERSKY

Indeed these are the typical conditions under which lottery tickets and insurance policies are sold. In prospect theory, the overweighting of small probabilities favors both gambling and insurance, while the S-shaped value function tends to inhibit both behaviors. Although prospect theory predicts both insurance and gambling for small probabilities, we feel that the present analysis falls far short of a fully adequate account of these complex phenomena. Indeed, there is evidence from both experimental studies [37], survey research [26], and observations of economic behavior, e.g., service and medical insurance, that the purchase of insurance often extends to the medium range of probabilities, and that small probabilities of disaster are sometimes entirely ignored. Furthermore, the evidence suggests that minor changes in the formulation of the decision problem can have marked effects on the attractiveness of insurance [37]. A comprehensive theory of insurance behavior should consider, in addition to pure attitudes toward uncertainty and money, such factors as the value of security, social norms of prudence, the aversiveness of a large number of small payments spread over time, information and misinformation regarding probabilities and outcomes, and many others. Some effects of these variables could be described within the present framework, e.g., as changes of reference point, transformations of the value function, or manipulations of probabilities or decision weights. Other effects may require the introduction of variables or concepts which have not been considered in this treatment. Shifts of Reference

So far in this paper, gains and losses were defined by the amounts of money that are obtained or paid キィ・セ@ a prospect is played, and the reference point was taken to be the status quo, or q_ne's current assets. Although this is probably true for most choice problems, there are situations in which gains and losses are coded relative to an expectation or aspiration level that differs from the status quo. For example, an unexpected tax withdrawal from a monthly pay check is experienced as a loss, not as a reduced gain. Similarly, an entrepreneur who is weathering a slump with greater success than his competitors may interpret a small loss as a gain, relative to the larger loss he had reason to expect. The reference point in the preceding examples corresponded to an asset position that one had expected to attain. A discrepancy between the reference point and the current asset position may also arise because of recent changes in wealth to which one has not yet adapted [29]. Imagine a person who is involved in a business venture, has already lost 2,000 and is now facing a choice between a sure gain of 1,000 and an even chance to win 2,000 or nothing. If he has not yet adapted to his losses, he is likely to code the problem as a choice between (-2,000, .50) and (-1,000) rather than as a choice between (2,000, .50) and (1,000). As we have seen, the former representation induces more adventurous choices than the latter. A change of reference point alters the preference order for prospects. In particular, the present theory implies that a negative translation of a choice

Chapter 6. Prospect Theory: An Analysis of Decision Under Risk PROSPECT THEORY

123

287

problem, such as arises from incomplete adaptation to recent losses, increases risk seeking in some situations. Specifically, if a risky prospect (x, p; - y, 1-p) is just acceptable, then (x- z, p; - y- z, 1- p) is preferred over (- z) for x, y, z > 0, with x>z. To prove this proposition, note that V(x,p; y, 1-p)=O

iff

1r(p)v(x) = -7T(1-p)v(-y).

Furthermore, V(x-z,p;-y-z, 1-p)

= 1r(p )v(x- z)+ 7T(1- p)v(- y- z) > 1r(p )v(x) -7r(p)v(z) + 7T(1- p )v(- y) +7T(1-p)v(- z)

by the properties of v,

= -7T(1- p )v(- y) -1r(p )v(z) + 7T(1- p)v( -y)

+7T(1- p)v(-z)

by substitution,

= -1r(p )v(z) + 7T(l- p )v( -z)

> v(- z )[1r(p) + 7T(1- p )] >v( -z)

since v(- z) < -v(z ),

by subcertainty.

This analysis suggests that a person who has not made peace with his losses is likely to accept gambles that would be unacceptable to him otherwise. The well known observation [31] that the tendency to bet on long shots increases in the course of the betting day provides some support for the hypothesis that a failure to adapt to losses or to attain an expected gain induces risk seeking. For another example, consider an individual who expects to purchase insurance, perhaps because he has owned it in the past or because his friends do. This individual may code the decision to pay a premium y to protect against a loss x as a choice between (-x + y, p; y, 1- p) and (0) rather than as a choice between (-x, p) and (-y ). The preceding argument entails that insurance is likely to be more attractive in the former representation than in the latter. Another important case of a shift of reference point arises when a person formulates his decision problem in terms of final assets, as advocated in decision analysis, rather than in terms of gains and losses, as people usually do. In this case, the reference point is set to zero on the scale of wealth and the value function is likely to be concave everywhere [39]. According to the present analysis, this formulation essentially eliminates risk seeking, except for gambling with low probabilities. The explicit formulation of decision problems in terms of final assets is perhaps the most effective procedure for eliminating risk seeking in the domain of losses.

124

D Kahneman and A Tversky

288

D. KAHNEMAN AND A. TVERSKY

Many economic decisions involve transactions in which one pays money in exchange for a desirable prospect. Current decision theories analyze such problems as comparisons between the status quo and an alternative state which includes the acquired prospect minus its cost. For example, the decision whether to pay 10 for the gamble (1,000, .01) is treated as a choice between (990, .01; -10, .99) and (0). In this analysis, readiness to purchase the positive prospect is equated to willingness to accept the corresponding mixed prospect. The prevalent failure to integrate riskless and risky prospects, dramatized in the isolation effect, suggests that people are unlikely to perform the operation of subtracting the cost from the outcomes in deciding whether to buy a gamble. Instead, we suggest that people usually evaluate the gamble and its cost separately, and decide to purchase the gamble if the combined value is positive. Thus, the gamble (1,000, .01) will be purchased for a price of 10 if 7r (.01)v(l,OOO) + v( -10) > 0. If this hypothesis is correct, the decision to pay 10 for (1,000, .01), for example, is no longer equivalent to the decision to accept the gamble (990, .01; -10, .99). Furthermore, prospect theory implies that if one is indifferent between (x(1p ), p; -px, 1- p) and (0) then one will not pay px to purchase the prospect (x, p ). Thus, people are expected to exhibit more risk seeking in deciding whether to accept a fair gamble than in deciding whether to purchase a gamble for a fair price. The location of the reference point, and the manner in which 'choice problems are coded and edited emerge as critical factors in the analysis of decisions. Extensions In order to encompass a wider range of decision problems, prospect theory should be extended in several directions. Some generalizations are immediate; others require further development. The extension of equations (1) and (2) to prospects with any number of outcomes is straightforward. When the number of outcomes is large, however, additional editing operations may be invoked to compound simplify evaluation. The manner in which complex options, ・Nァセ@ prospects, are reduced to simpler ones is yet to be investigated. Although the present paper has been concerned mainly with monetary outcomes, the theory is readily applicable to choices involving other attributes, e.g., quality of life or the number of lives that could be lost or saved as a consequence of a policy decision. The main properties of the proposed value function for money should apply to other attributes as well. In particular, we expect outcomes to be coded as gains or losses relative to a neutral reference point, and losses to loom larger than gains. The theory can also be extended to the typical situation of choice, where the probabilities of outcomes are not explicitly given. In such situations, decision weights must be attached to particular events rather than to stated probabilities, but they are expected to exhibit the essential properties that were ascribed to the weighting function. For example, if A and B are complementary events and neither is certain, 7r(A) + 7r(B) should be less than unity-a natural analogue to subcertainty.

Chapter 6. Prospect Theory: An Analysis of Decision Under Risk PROSPECT THEORY

125

289

The decision weight associated with an event will depend primarily on the perceived likelihood of that event, which could be subject to major biases [45]. In addition, decision weights may be affected by other considerations, such as ambiguity or vagueness. Indeed, the work of Ells berg [10] and Fellner [12] implies that vagueness reduces decision weights. Consequently, subcertainty should be more pronounced for vague than for clear probabilities. The present analysis of preference between risky options has developed two themes. The first theme concerns editing operations that determine how prospects are perceived. The second theme involves the judgmental principles that govern the evaluation of gains and losses and the weighting of uncertain outcomes. Although both themes should be developed further, they appear to provide a useful framework for the descriptive analysis of choice under risk.

The University of British Columbia and Stanford University Manuscript received November, 1977; final revision received March, 1978.

APPENDIX2 In this appendix we sketch an axiomatic analysis of prospect theory. Since a complete self-contained treatment is long and tedious, we merely outline the essential steps and exhibit the key ordinal properties needed to establish the bilinear representation of equation (1), Similar methods could be extended to axiomatize equation (2). Consider the set of all regular prospects of the form (x, p; y, q) with p + q < 1. The extension to regular prospects with p + q = 1 is straightforward. Let 2: denote the relation of preference between prospects that is assumed to be connected, symmetric and transitive, and let = denote the associated relation of indifference. Naturally, (x, p; y, q) = (y, q; x, p). We also assume, as is implicit in our notation, that (x, p; 0, q) = (x, p; 0, r), and (x, p; y, 0) = (x, p; z, 0). That is, the null outcome and the impossible event have the property of a multiplicative zero. Note that the desired representation (equation (1)) is additive in the probability-outcome pairs. Hence, the theory of additive conjoint measurement can be applied to obtain a scale V which preserves the preference order, and interval scales f and g in two arguments such that V(x, p; y, q) = f(x, p)+ g(y, q).

The key axioms used to derive this representation are: Independence: (x, p; y, q) 2: (x, p; y'q') iff (x', p'; y, q) 2: (x', p'; y', q'). Cancellation: If (x, p; y' q') 2: (x', p'; y, q) and (x', p'; y", q") 2: (x", p"; y', q'), then (x, p; y", q") 2: (x", p"; y, q). Solvability: If (x, p; y, q) 2: (z, r)2: (x, p; y' q')for some outcome z and probability r, then there exist y", q" such that (x,p; y"q")=(z, r).

It has been shown that these conditions are sufficient to construct the desired additive representation, provided the preference order is Archimedean [8, 25]. Furthermore, since (x, p; y, q) = (y, q; x, p ), f(x, p) + g(y, q) = f(y, q) + g(x, p ), and letting q = 0 yields f = g. Next, consider the set of all prospects of the form (x, p) with a single non-zero outcome. In this case, the bilinear model reduces to V(x, p) = 1r(p)v(x). This is the multiplicative model, investigated in [35]

and [25]. To construct the multiplicative representation we assume that the ordering of the probability-outcome pairs satisfies independence, cancellation, solvability, and the Archimedean axiom. In addition, we assume sign dependence [25] to ensure the proper multiplication of signs. It should be noted that the solvability axiom used in [35] and [25] must be weakened because the probability factor permits only bounded solvability. 2 We are indebted to David H. Krantz for his help in the formulation of this section.

126

D Kahneman and A Tversky

290

D. KAHNEMAN AND A. TVERSKY

Combining the additive and the multiplicative representations yields V(x, p; y, q) = f[ 1T(p)v(x)] + f[ 1T(q)v(y)]. Finally, we impose a new distributivity axiom: (x, p; y, p) = (z, p)

iff

(x, q; y, q) = (z, q).

Applying this axiom to the above representation, we obtain

f[ 1T(p) v (x)] +f[ 1T(p )v (y )] = f[ 1T(p )v(z)] implies f[ 1T(q)v(x )] + f[ 7T(q)v(y )] = f[ 1T(q )v (z)]. Assuming, with no loss of generality, that 7T(q) < 1T(p), and letting a= 7T(p)v(x), (3 = 1T(p)v(y), y= 17 (p)v(z), and 0=7T(q)/7T(p), yields f(a)+f((3)=f(y) implies f(Oa)+f(0(3)=f(Oy) for all 0:: 0 for x > 0, and V" : >: 0 for x < 0. Notice that since the value function is nondecreasing (V' : >: 0), V KT is a subset of U 1 . We call V a value function, rather than a utility function, to be consistent with the terminology of Kahneman and Tversky. In the discussion of Markowitz utility functions we confine ourselves to the range between the two extreme inflection points of the Markowitz utility function (Points A and B in Figure 1c), which are expected to be at extreme wealth levels (see Markowitz 1952b). Thus we define: VM-the class of all Markowitz utility functions (where the subscript M denotes Markowitz): functions which are reverse S-shaped with an inflection point at x=O. Thus, VEVM if V' ::>::Oforallx,PO/ V" ::>::0 for x > 0, and V" ::0: 0 for x < 0. Notice that since the value function is nondecreasing (V' ::>:: 0), VM is also a subset of U 1• As Markowitz's function, like the prospect theory value function, depends on change of wealth, we denote it by VM, rather than UM. We turn now to a few decision rules which are known as stochastic dominance rules. As explained above, because all of these rules (not only FSD as in the example above) are invariant to the initial wealth, they can be written in terms of the change of wealth x. We briefly describe the well-known FSD and SSD rules. We then present the prospect stochastic dominance rule (PSD), which corresponds to all S-shaped value functions, and the Markowitz stochastic dominance rule (MSD), which corresponds to all reverse S-shaped value functions. All these rules will be examined in the experiment. We focus on PSD and MSD, which may indicate whether the prospect theory value function is valid in the general case, where the subjects have to make choices between mixed prospects with negative as well as positive outcomes, and in which there is no certainty effect. FSD (First-Degree Stochastic Dominance). Let F and G be two distinct prospects with cumulative distributions F and G, respectively. Then, F(x) ::0: G(x) for all x {}

EFU(x) "':: EGU(x) for all

u E ul,

(1)

As in the case of 5-shaped functions, in the general case we allow for V' (O) to be nonexistent.

7

MANAGEMENT SciENCE/Val. 48, No. 10,

October 2002

where there is a strict inequality for some x = x0 , and a strict inequality for some UQ E U 1 . SSD (Second-Degree Stochastic Dominance). Define F and G as above. Then, {)c(t)-F(t)]dt::>::O forallx {}

EFU(x) : >: EGU(x) for all U E U2 ,

(2)

where there is a strict inequality for some x = x0 , and a strict inequality for some UQ E U2 • For proof of the FSD and SSD rules, see Hadar and Russell (1969), Hanoch and Levy (1969), and Rothschild and Stiglitz (1970). For a survey on stochastic dominance rules, see Levy (1992). PSD (Prospect Stochastic Dominance). Define F and G as above. Then F dominates G for all S-shape utility /value functions, V E V KT• if and only if

jy

0

[G(t)- F(t)] dt : >: 0 for ally ::0:0 and

f

[G(t)- F(t)] dt ::>:: 0 for all x : >: 0.

(3)

(Once again, we require a strict inequality for some pair (y0 , x0 ) and for some V0 E VKr·) A proof of PSD and more detail can be found in Levy (1998), and Levy and Wiener (1998). MSD (Markowitz Stochastic Dominance). Define F and G as above. Then F dominates G for all reverse S-shaped value functions, V E VM, if and only if

L:

[G(t)- F(t)] dt : >: 0 for ally:::: 0 and £oo!G(t)-F(t)]dt::>::O forallx::>::O,

(4)

(with at least one strict inequality). We call this dominance relation MSD-Markowitz Stochastic Dominance. PROOF. See Appendix A. Although Conditions (3) and (4) look similar, we will show below that it is possible that F dominates G by PSD, but G dominates F by MSD. While the intuitions of FSD and SSD are relatively straightforward, the intuition for PSD and MSD is less transparent. The intuition for SSD and the riskseeking dominance (RSD) rule (presented below) are important in particular because both PSD and MSD

1339

Chapter 7. Prospect Theory: Much Ado About Nothing?

135

LEVY AND LEVY

Prospect Theory Revisited

preferences contain both risk-averse and risk-seeking segments. Therefore, once we understand the intuition for SSD and RSD, the intuition for PSD and MSD becomes straightforward. FSD. F(x) :S G(x) for all x implies that 1- F(x) ::: 1- G(x) for all x, or PrF(X::: x)::: Pre( X::: x) for all x; Thus, the probability of getting x or more is higher under F than under G for all x; consequently; any individual who prefers more money over less money will prefer F. SSD. For any utility function the difference in expected utility between prospects F and G is given by: !J.=EFU(x)-EGU(x)= /_:[G(t)-F(t)]U'(t)dt. (5)

Equation (5) is obtained by integration by parts of the difference in expected utility of the two prospects, see Hadar and Russell (1969), Hanoch and Levy (1969), Rothschild and Stiglitz (1970), and Levy (1998). The condition for SSD dominance of F over G is that foo[G(t) -F(t)] dt::: 0 for all x. Graphically, this means that the area enclosed between G and F from -oo to any value x is positive. Thus, the SSD condition states that for any negative area (F > G) there is a larger preceding positive area (G >F). In calculating !J. these areas are multiplied by U' (see Equation (5)), and as U' is a declining function for risk averters (U" :S 0) the SSD condition ensures that the positive contribution to !J. of the positive areas is greater than the negative contribution to !J. of the negative areas, and therefore !J. :;:0. RSD. The dominance condition ofF over G for riskseeking investors (U'::: 0, U"::: 0 in the whole domain of x) is that fxoo[G(t)- F(t)] dt:;: 0 for all x (see Levy 1998, pp. 33Q-332). Graphically this means that the last area (largest x) enclosed between G and F is positive (which is not required by SSD). There may be a negative area before the last positive area, but it must be smaller than the last positive area, and so forth-for each negative area there is a larger positive area which follows (at larger x). The explanation for the RSD criterion is that we have, as before, !J. = EFU(x)- EGU(x) = ヲセッ{gHエI@ -F(t)]U'(t) dt (see Equation (5)), but this time U' is an increasing function (U"::: 0). If the RSD criterion holds, the contribution of the positive areas to !J. is larger than the negative contribution of the negative areas, and we have !J. ::: 0 or

1340

EFU(x) ::: EGU(x) for all risk-seeking investors. Having the intuitive explanation for risk-averse and riskseeking dominance, the intuition of PSD and MSD becomes straightforward. PSD. In this case we have U" ::: 0 for x < 0, and U" :S 0 for x > 0, and the condition for the dominance ofF over G by PSD is: fox[G(t)- F(t)]dt::: 0 for all x > 0 and f:[G(t) -F(t)]dt::: 0 for ally< 0 (see Equation (3)). The intuition for this condition relates directly to the preceding explanations about SSD and RSD. We want F to be preferred over G by any S-shaped utility function. Consider an 5-shaped utility function which is almost linear and with a very small slope for y < 0 (i.e., U' "" 0 for y < 0). For such a function the contribution of the negative domain to !J. of Equation (5) is negligible, and we should only consider the risk-averse condition for the positive range, which is f 0x[G(t) -F(t)]dt:;: 0 for all x > 0. Similarly, considering an S-shaped function which is almost linear and flat in the positive range yields the risk-seeking condition for the negative range, which is f:[ G(t)- F(t)] dt::: 0 for ally < 0. Therefore, for F to dominate G by PSD for all 5-shaped functions, both of these conditions must hold. MSD. Similar to the logic of the PSD intuition, consider a reverse 5-shaped function which is almost linear and flat in the negative domain. For such a function the contribution of the negative domain to !J. is negligible, and we should only consider the riskseeking condition for the positive range, which is fxoo[G(t)- F(t)] dt::: 0 for all x > 0. A reverseS-shaped function which is almost linear and flat in the positive domain dictates the risk-averse condition for the negative range, which is f!oo[G(t) -F(t)]dt::: 0 for all y :S 0. As we encompass all reverse 5-shaped functions by MSD, the MSD rule is a combination of these two conditions (for formal proof of the MSD rule, see Appendix A). MSD is generally not "the opposite" of PSD. In other words, if F dominates G by PSD, this does not necessarily mean that G dominates F by MSD. This is easy to see, because having a higher mean is a necessary condition for dominance by both rules. 8 Therefore, if F dominates G by PSD, and F has a higher To see this, choose y = -oo and x = +oo in Equation {3), and y = x = 0 in Equation (4), and employ Equation (5) with U(x) = x.

8

MANAGEMENT SciENCE/Vol. 48, No. 10, October 2002

136

M Levy and H Levy LEVY AND LEVY

Prospect Theory Revisited

mean than G, G cannot possibly dominate F by MSD. However, as Corollary 1 below states, PSD and MSD are opposite if the two distributions have the same mean. CoROLLARY 1. Let F and G have the same mean. Then F dominates G by PSD if and only if G dominates F by MSD.

PROOF. From the relation EF(x) -Ec(x) = f"oo[G(t)F(t)] dt (Equation (5) with U(x) = x) we know that if F and G have the same mean, then f_:"oo[G(t)F(t)] dt = 0, or ヲセッ{gHエIM F(t)] dt + f 000 [G(t) -F(t)] dt = 0. If F dominates G by PSD, then each of these two terms must equal zero. To see this claim, recall that by the equal-mean condition, both terms cannot be negative. If one is positive, by the equal mean condition the other must be negative, and Equation (3) does not hold, so F cannot dominate G by PSD. By the same argument, if G dominates F by MSD we must also have too[G(t)- F(t)] dt = fooo[G(t)- F(t)] dt = 0. Thus, if one distribution dominates the other by either PSD or MSD, this implies too[G(t)- F(t)] dt = 0 and f 000 [G(t)- F(t)] dt = 0. Now, let us prove the claim of Corollar6. 1. F dominates by PSD only if for every y < O,JY[G(t) -F(t)]dt ::0:0. But because in the case of equal means we have too[G(t)- F(t)] dt = 0, this implies that f!oo[G(t) -F(t)]dt :0:0 for every y < 0, or J!oo[F(t)- G(t)] dt ::0: 0 for every y < 0, which is the condition for the MSD dominance of G over F in this range (see Equation (4)). Similarly, f 0x[G(t)- F(t)] dt ::0: 0 for all x > 0 implies fxoo[G(t) -F(t)] dt :0:0 for all x > 0, which in turn leads to froo[F(t)- G(t)] dt ::0:0 for all x > 0. Thus, for F and G with equal mean, F dominates G by PSD if and only if G dominates F by MSD. D

3. The Experiments and the Results The subjects in our experiments were students, university professors, and practitioners. The students are graduate and undergraduate students in the business schools at UCLA, the University of Washington, and the Hebrew University. The university professors are faculty in the business and economics departments at Baruch College, UCLA, and the Hebrew University. The practitioners are financial analysts, mutual funds managers, and portfolio managers.

MANAGEMENT SCIENCE/Yo!. 48, No. 10, October 2002

The subjects were asked to answer a questionnaire with one or more tasks. In each task the subjects had to choose between two investments denoted by F and G. We have conducted three experiments, with questionnaires as given in Tables 1, 2, and 3. All probabilities given in the experiments are relatively large (p ::0: 0.25),9 hence it is unlikely that subjective probability distortion plays an important role in the decision-making process. 10 Since we have no certain outcome, the certainty effect is neutralized. Experiment 1: Design Experiment 1 is designed to investigate which type of preference class (risk aversion, 5-shaped functions, or reverse S-shaped functions) best describes individual choice under risk. Subjects were presented with four tasks, as given in Table 1. In Task I, G dominates F by PSD, but F dominates G by MSD, as shown by Figure 3. Thus, G dominates F for all value functions V E VKr because Condition (3) holds with F and Gin switched roles. At the same time, Condition (4) holds, and hence F dominates G for all reverse S-shaped functions V E VM. Thus, Task I is a head-to-head competition between these two alternative theories. Note that neither prospect dominates the other by SSD. Task IV is designed to test whether risk aversion provides a good description of behavior. In this task, G dominates F by SSD. In the remaining two tasks, Tasks II and III, one prospect dominates the other by FSD. These tasks are "controls" designed to test whether the subjects comprehend the experimental setup, and whether they are rational in the sense of preferring more money to less money. The subjects participating in Experiment 1 were 132 students, 66 professors, and 62 practitioners, for a total of 260 subjects. Experiment 1: Results The results of the experiment are given in Table lb. As there are no significant differences between the results Except in Task 3 of Experiment 1, which tests for FSD, and is a "control" task in our experiment.

9

It is found experimentally that small probabilities, e.g., those related to the chance of winning a lottery, are subjectively distorted by subjects. However, this distortion typically takes place for extreme probabilities, and not for the moderate probabilities employed here. 10

1341

Chapter 7. Prospect Theory: Much Ado About Nothing?

137

LEVY AND LEVY

Prospect Theory Revisited

Table 1a

Experiment 1: The Choices Presented to the Subjects

Table 1b

Suppose that you decided to invest $10,000 enher in stock For in stock G. Which stock would you choose, F, or G, when it is given that the dollar gain or loss one month from now will be as follows. TASK 1:

G Gain or loss

Probability

-3.000 4,500

112

Please write F for G:

1/2

Gain or loss

Probability

-6,000 3,000

114 3/4

G Gain or loss

-500 +2,500

Probability

Gain or loss

Probability

1/3 213

-500 2,500

112 112

o

TASK Ill: Which would you prefer, F or G, if the dollar gain or loss one month from now will be as follows:

G Gain or loss

Probability

Gain or loss

Probability

+500 +2.000 +5,000

3/10 3/10 4/10

-500 0 +500 +1.000 +2.000 +5.000

1/10 1/10 1/10 2110 1110 4/10

Please write For G:

o

TASK IV: Which would you prefer, F or G, if the dollar gain or loss one month from now will be as follows:

G Gain or loss

Probability

-500 +500 +1,000 +2.000

1/4 1/4 1/4 114

Please write F or G: 0

1342

I (G >-pso F, F '""""G) II (F >-FSo G) ill (F >-FSo G) IV (G >-sso F)

G 71

27

96 82 47

18 51

Indifferent

Total

100 100 100 100

Number of subjects: 260 •Numbers in the tables are in percent. rounded to the nearest integer. The notations,...... >-sso· '""""'and '""•so indicate dominance by FSD, SSD, PSD, and MSD, respectively.

o

TASK II: Which would you prefer, F or G. if the dollar gain or loss one month from now will be as follows:

Please write F for G:

The Results of Experiment 1•

Task No.

Gain or loss

+1.500

obtained in different institutions, or across the subject categories, we report only the aggregate results. In Tasks II and III F dominates G by FSD because F(x) ::: G(x) for all values x. Indeed, in all groups of subjects we find a strong preference for F, the FSD dominating prospect (see Task II and Task III). In Task II there is an obvious FSD superiority of F over G, and therefore we are not surprised with the experimental findings. In Task III the FSD preference of F over G is a little less transparent in comparison to Task II; still, almost all subjects identified the superior FSD investment in this task. Given that investors are rational, and always prefer more money to less money, the findings in Tasks II and III are expected regardless of the investors' specific preferences (at least with our sophisticated group of subjects, where we do not expect to find many errors in the decision making). This is because FSD is a decision rule which corresponds to all types of nondecreasing utility or value functionsY The findings of Tasks II and III neither confirm nor reject prospect theory, since there is FSD dominance In their 1979 paper, Kahneman and Tversky propose possible distortion in probabilities by the decision makers. Such a distortion

11

Probability

may imply that an option which is inferior to the other by FSD (with the objective probabilities) may be selected and the supe-

1/2

rior FSD option may be rejected. Kahneman and Tversky realize that FSD dominance is a cornerstone for the decision-making process, hence suggested in their 1992 paper the cumulative version of prospect theory such that if F dominates G with the objective probabilities, the dominance will not be violated even after the probabil-

112

ity distortion (see also Quiggin 1982 and Yaari 1987). Indeed, FSD is sometimes used as one of the axioms of expected utility theory (see Fishburn 1982 and Levy 1998).

MANAGEMENT ScJENCE/Vol. 48, No. 10, October 2002

138

M Levy and H Levy LEVY AND LEVY

Prospect Theory Revisited

Figure 3

Task I of Experiment 1- G Dominates F by PSD, F Dominates G by MSD

1.00

セ@

-

0.75

:0

as .c a.

1500x0.5

e CD

.セ@

.....oF

0.50

ro :;

+

+

3000x0.25

3000x0.25

E :::J

u

G

0.25

-

3000x0.25

0.00 -6000



!

4000 -3000 -2000

0

2000 3000 40004500

6000

payoff($}

of F over G, and therefore a subject with any nondecreasing utility or value function should prefer F to G, whether he has a "standard" concave utility function, an 5-shaped value function, or a reverse 5-shaped function. Thus, any choice of G over F is due to irrationality or due to a possible human error. Fortunately, we did not find many such errors, indicating that the subjects were not answering the questionnaire irrationally or arbitrarily. This point is important for the interpretation of the results of Tasks I and IY, which are the focus of this experiment. In Task I neither F nor G dominates the other by F5D or by 550 (see Figure 3). However, G dominates F by PSD, thus we would expect any subject with 5-shaped preferences to select Prospect G in Task I. At the same time, F dominates G by M5D. Thus Task I constitutes a head-to-head race between two competing theories about preferences: the prospect theory 5-shaped function and the Markowitz reverse 5-shaped function. The striking result is that the majority of subjects (71%) preferred prospect F, strongly contradicting the 5-shaped function suggested by prospect theory, and supporting the reverse 5-shaped function hypothesis. It is interesting to note that the preference of F over G in Task I is even stronger among the more sophisti-

MANAGEMENT Sc!ENCE/Vol. 48, No. 10, October 2002

cated investors-practitioners (79%) and faculty (86%) (for brevity's sake not reported in the table). The highest proportion rejecting at least one segment of the 5-shaped value function in the Kahneman and Tversky experiment is 20% (see Kahneman and Tversky 1979, p. 268). We have on the aggregate 71% of the choices rejecting the 5-shaped function. A statistical test reveals that the difference in the proportion of subjects rejecting the 5-shaped function in these two studies is highly significant with a z-statistic of 8.59. 12 12

The sample statistic is:

z=

PEI- fiKr p•q•

+ p•q•

NEt

NKT

where PEt and fin are the sample mean proportions rejecting the Sshaped function in Experiment 1 (Task I} and the Kahneman and Tversky experiment, respectively, NEt and NKT are the number of observations in these two experiments, p* is the overall sample proportion: * PEl NEt +PKTNKT p = NEl+NKT 1 and q* = 1 - p". For Task I of Experiment 1 we obtain: Z=

0.71-0.20

=8.59. J0.57 ·0.43 0.57. 0.43 ----uo+ _9_5_

1343

Chapter 7. Prospect Theory: Much Ado About Nothing?

139

LEVY AND LEVY

Prospect Theory Revisited

The results in Task I constitute a rejection of the S-shaped function because they imply that at least 71% of the subjects do not have this type of preference.13 This result lends support to the Markowitz reverse S-shaped function hypothesis, yet proving this hypothesis is much more difficult than rejecting the S-shape hypothesis. This is because one could claim that individuals are driven to choose F by preferences other than a reverse S-shaped function. For example, it is possible that subjects are risk averse, and while F does not dominate G by SSD, it could be that most subjects have higher expected utility under F. One could also suspect that the subjects (explicitly or implicitly) employ the Markowitz (1952a, 1959, 1987) well-known mean-variance rule, because,

hence, F is slightly preferable to G by the meanvariance rule. Thus, the subjects may employ the mean-variance rule even if it is not justified in this specific case because the distributions are not normal; see Tobin (1958) and Hanoch and Levy (1969). Task IV is designed to address these issues. In Task IV, there is no FSD and no PSD or MSD dominance, but G dominates F by SSD. Thus, if subjects are risk-averse we would expect strong support for G. We find that only about half of the subjects selected Prospect G. 14 Thus, almost half of the subjects made choices contradicting risk aversion, casting grave doubt on the ability of preferences which are concave throughout (as in Figure 1a) to describe the subjects' behavior. This contradiction of risk aversion is consistent with the findings of Levy and Levy (2001). The results of Task IV also casts doubt on the hypothesis that the subjects employed the meanvariance rule in their decision making. We tend to reject this hypothesis for the following reason: In 13 We stress "at least" because any individual with S-shaped preferences should prefer G, but choosing G does not imply that preferences are necessarily 5-shaped.

71% of the faculty preferred G in Task IV. It is interesting that while the faculty (most of whom are conscious of the normative theory) chose G, most students (59%) who decide based on "gut feeling" prefer F. 14

1344

Task IV, Prospect G dominates prospect F not only by SSD but also by the mean-variance rule: EF(x)

= Ec(x) = 750

O'F(x)

= 901 >

O'c(x)

= 750.

Moreover, the mean-variance dominance of G over F is much more pronounced in Task IV than the mean-variance dominance of F over G in Task I. If the subjects employ the mean-variance rule (even if there is no theoretical justification to employ this rule here, as suggested above), we would expect to find a strong preference for Prospect G in Task IV. This did not occur, and approximately half of the subjects preferred Prospect F in Task IV. Thus, the decision making of the subjects is not driven by the mean-variance rule. Experiment 2: Design Experiment 2 is designed to create another headto-head competition between the prospect theory S-shaped function and the Markowitz reverse S-shaped function. The subjects in this experiment were 84 business school students. Table 2a provides the task in this experiment. It is straightforward to verify that F dominates G by PSD, but G dominates F by MSD. The task in this experiment is different than Task I of Experiment 1 in two respects: First, it is more complex. In this experiment there are four possible outcomes for each prospect, rather than the two outcomes of Task I in Experiment 1. Second, in Experiment 2 all the outcomes are equally likely. This is an attractive feature, because it makes any subjective probability distortion very unlikely. 15 Experiment 2: Results The results of Experiment 2 are given in Table 2b. The table reveals that 62% of the subjects chose Prospect G, which is inferior by PSD but dominant by MSD. Thus, we can state that at least 62% of the choices are inconsistent with prospect theory. While this result again rejects the S-shaped function and This point is made by Quiggin (1982). In addition, any subjective transformation performed directly on the probabilities (as in prospect theory) will still attach an equal probability weight to each outcome. This is also true of Viscusi's (1989) prospective reference theory with a symmetric reference point. 15

MANAGEMENT SCJENCE/Vol. 48, No. 10, October 2002

140

M Levy and H Levy LEVY AND LEVY

Prospect Theory Revisited

Tabla 2a

The Choices Presented to the Subjects

Suppose that you decided to invest $10,000 etther in stock For in stock G. Which stock would you choose, F, or G, when it is given that the dollar gain or Joss one month from now will be as follows:

G Gain or loss

Probability

Gain or loss

Probability

-1,600 -200 1,200 1,600

1/4 1/4 1/4 1/4

-1,000 -800 800 2,000

1/4 1/4 1/4 1/4

Please write For G: 0 Tabla 2b

The Results of Experiment 2•

(F >-pso G, G >-MSo F)

38%

G

Indifferent

Total

62%

0%

100%

Number of subjects: 84. •Numbers in the tables are in percent, rounded to the nearest integer. The notations >-pso and >-MSo indicate dominance by PSD and MSD, respectively.

prospects. By this procedure we can pinpoint whether having prospects with mixed outcomes is of crucial importance. Experiment 3: Design The tasks in Experiment 3 are given in Table 3a. Tasks I and II are precise replications of the tasks in the Kahneman and Tversky (1979) experiment (see p. 268). Task I compares a positive prospect with a certain positive outcome, and Task II compares a negative prospect with a certain negative outcome. Task III involves two uncertain prospects with mixed outcomes. In this third task, Prospect G dominates by P5D, but Prospect F dominates by M5D. The subjects participating in this experiment were 51 practitioners and 129 business school students.

Tabla 3a

lends support to the reverse 5-shaped function, it is somewhat less dramatic than the 71% who chose the PSD-inferior (and M5D-dominant) prospect in Task I of Experiment 1. This difference may be due to the increased complexity of the task in Experiment 2 relative to Task I of Experiment 1. Indeed, a similar drop (from 96% to 82%) is observed in the choices conforming with F5D as the complexity of the tasks was increased (see Tasks II and III in Experiment 1). Yet the proportion rejecting the 5-shaped function here is significantly larger than in the Kahneman and Tversky (1979) experiment, with a z-value of 5.72 (see Footnote 12). The strong results rejecting the prospect theory 5-shaped function that was revealed in Experiments 1 and 2 raise the following questions: How do these results fit with the wide support documented for prospect theory? Is there perhaps some special attribute to the subject population in our experiments which systematically biases our results? To answer these questions, we have conducted a third experiment, in which the same group of subjects is presented with tasks identical to the original Kahneman and Tversky certainty equivalent tasks, and also with a task involving mixed and uncertain

MANAGEMENT SciENCE/Vol. 48, No. 10, October 2002

The Choices Presented to the Subjects

Task 1: Imagine that you face the following two alternatives and you must choose one of them. Which one would you select?

G Gain or loss

Probability

Gain or loss

4,000

0.80 0.20

3,000

0

Probability

Please write F for G: 0 TASK II: Imagine that you face the following two alternatives and you must choose one of them. Which would you select?

G Gain or loss

Probability

Gain or loss

-4,000 0

0.80 0.20

-3,000

Probability

Please write F for G: 0 TASK Ill: Suppose that you decided to invest $10,000 either in stock F or in Stock G. Which stock would you choose, F, or G, when it is given that the dollar gain or foss one month from now will be as follows:

G Gain or loss

Probability

Gain or loss

Probability

-1,500

1/2 1/2

-3,000 3,000

3/4

4,500

1/4

1345

Chapter 7. Prospect Theory: Much Ado About Nothing?

141

LEVY AND LEVY Prospect Theory Revisited

Table 3b

The rasuiiS ol Experiment3

Task

Ill (G >-pso F, F >-Mso G)

19% 69% 76%

G

Indifferent

Total

81% 30% 23%

0%

100% 100% 100%

1% 1%

Number of subjects: 180. Numbers in the tables are in percent, rounded to the nearest integer. The notations >-pgo and >-Mso indicate dominance by PSD and MSD, respectively.

Experiment 3: Results Table 3b reports the results of Experiment 3. Because there are no significant differences between the subject populations, we report only the aggregate results. In Task I we find that 81% of the choices are consistent with risk aversion for gains. In Task II, 69% of the choices are consistent with risk seeking for losses. Thus, the results in these two tasks are very similar to Kahneman and Tversky' s results, albeit our results in the negative domain are somewhat weaker (Kahneman and Tversky obtain 80% and 92%, respectively, see Kahneman and Tversky 1979, Table 1). The striking result is related to Task ill. Here, 76% of the subjects chose F, the prospect which is inferior by PSD! 16 The same subjects which seem to support risk aversion for gains and risk seeking for losses when they are confronted with nonmixed gambles in the certainty equivalent framework, reject the S-shaped function when mixed and uncertain prospects are compared. Only 17% of the subjects made choices consistent with the S-shaped function throughout: a choice of G in Task I, F in Task II, and G in Task III. This leads us to believe that the support found in previous studies for the S-shaped function may be due to the use of nonmixed gambles and the certainty equivalent framework, with one certain outcome. As these constitute very special and atypical situations, we believe that the results in these tasks may be biased due to their special setup or to the well-known "certainty effect". Note that in Task ill, F dominates G by MSD. Thus, when the two competing theories of the S-shaped function and the reverse S-shaped function race head This is significantly larger than the proportion of subjects rejecting the S-shaped value function in the Kahneman and Tversky experiment, with a z-value of 8.92

16

1346

to head, 76% of the choices conform with MSD and contradict PSD.

4.

Concluding Remarks

Prospect theory is a paradigm challenging expected utility theory. The four main components of prospect theory are: Individuals make decisions based on change of wealth rather than total wealth; preferences are described by an S-shaped value/utility function, V(x), with an inflection point at x = 0; individuals distort small probabilities; and the "framing" of alternatives affects individuals' choices. The main justification for prospect theory is the results of experimental studies in which the subjects had to declare the certainty equivalent of either negative or positive bets, bets which virtually do not exist in practice, and certainly are not common in the financial markets. This unrealistic "framing" of the bets, and possible biases induced by the "certainty effect" could account for the support for the S-shaped value function which is found in previous experiments, and which is rejected in our experiment. In our experiment, we employ stochastic dominance investment criteria. Because the stochastic dominance criteria are invariant to the initial wealth, we can focus on change of wealth rather than total wealth. We deal with relatively large probabilities (p 2:: 0.25), hence it is very unlikely that probability distortion plays an important role in our experiment. Furthermore, in our study we have prospects with no certain outcome, hence the certainty effect is irrelevant here. Neutralizing these three factors allows us to focus on the preference and to examine whether it is indeed 5-shaped, as claimed by prospect theory. In the previous experiments the certainty equivalent approach was employed, hence, in order to characterize the properties of the value function there was no practical choice but to use nonmixed bets (either negative or positive). The stochastic dominance approach allows us to employ mixed bets, which are much more typical of real investment situations. For the first time, we use a recently developed investment criterion called Prospect Stochastic Dominance (PSD), and a criterion which is developed here, called Markowitz Stochastic Dominance (MSD).

MANAGEMENT SCIENCE/Yo!. 48, No. 10, October 2002

142

M Levy and H Levy LEVY AND LEVY

Prospect Theory Revisited

Using the PSD criterion and mixed bets, based on the experimental results of three distinct groups of subjects (students, university professors, and practitioners), we conclude that the S-shaped preference is rejected. Moreover, employing the MSD criterion reveals support for the reverse S-shaped value function, which was suggested by Markowitz (1952b). To be more specific, in tasks where one prospect dominates the other by PSD but an opposite dominance holds for MSD, three experiments revealed that 71%, 62%, and 76% of the subjects preferred the MSD dominating prospect. Moreover, this preference is even stronger with the more sophisticated investors: 79% of the practitioners and 86% of the professors preferred F over G, where G is the PSD dominating option! Thus, we reject the prospect theory S-shaped function. It is interesting to note that when faced with outcomes restricted either to the positive domain or to the negative domain, as in the certainty equivalent approach, most subjects make choices according to prospect theory. However, the same subjects reject the PSD superior option with mixed bets. This leads us to believe that the support for the S-shaped function as obtained by Kahneman and Tversky is actually due to the well-known certainty effect and does not represent investors' preferences in a realistic setting of mixed bets. Obviously, different individuals may have preferences of different types, i.e., some may be characterized by risk aversion, others by an S-shaped value function, etc. Some individuals may not even be consistent in their decision making. Thus, one cannot hope to find a preference class which perfectly characterizes all individuals. However, our experimental evidence indicates that the best characterization of subjects' behavior is given by the class of reverse Sshaped functions, as suggested by Markowitz, which is just the opposite of the 5-shaped function advocated by prospect theory. Acknowledgments The authors acknowledge the helpful comments of the associate editor and the two anonymous referees. This study was financially supported by the Krueger Center of Finance and by the Zagagi Fund.

MANAGEMENT ScrENCE/Vol. 48, No. 10, Cktober 2002

Appendix A. A Proof of the Markowitz Stochastic Dominance (MSD) Criterion THEOREM 1. Let V(x) E VM where VM is the class of all continuously and twice differentiable Milrkowitz utility functions such that V' 0> 0 for all x, with V" :50 for x < 0 and V" 2: 0 for x > 0. Then F dominates G for all V E VM if and only if:

/_: [G(t)- F(t)] dt 0> 0 for ally< 0 and ioo[G(t)-F(t)]dto-0 forallx>O. PROOF. Let us first formulate our proof in terms of change of wealth, x, rather than total wealth, w + x, and then show that the dominance is invariant to the value w. For convenience, assume that the outcomes of Prospects F and G have lower and upper bounds a and b. One can later generalize the result to the case of unbounded prospects by taking the limits 。セ@ -oo b セ@ oo. Define:

b.= E, V(x)- Ec V(x) =f.

b

V(x) dF(x)-

f.

b

V(x) dG(x).

Integrating by parts, we have

b.= [F(x)- G(x)]V(x{- J.' (F(x)- G(x))V'(x) dx. As F(b) = G(b) =I and F(a) = G(a) = 0, we have:

b.= J.'[G(x)-F(x)]V'(x)dx = f[G(y)-F(y)]V'(y)dy+ {[G(x)-F(x)]V'(x)dx,

where we use the notation y for variable values in the negative domain. Integrating once again by parts, the tvvo terms on the righthand side yield:

b.= V'(y) f.

y

10

(G(t)-F(t))dt,-

f. 0V"(y) f.

y

(G(t)-F(t))dtdy

+V'(x) {(G(t)-F(t))dtl'- { V"(x) {(G(t)-F(t))dtdx. 0

0

0

0

As some of the terms (i.e., the cases y =a, and x = 0) are equal to zero, A can be rewritten as:

b.= V'(O) f[G(t)-F(t)]dt-

t

V"(y) t(G(t)-F(t))dtdy

+ V'(b) {[G(I)-F(t)]dt- { V"(x) [(G(t)-F(t))dtdx. Because V' 0> 0 and V" :0 0 for y < 0, the condition J;'[G(t)F(t)]dt 0> 0 ensures that the first two terms on the right-hand side of A are nonnegative. (Note that we assume that the utility function is twice differentiable, and that V' 2: 0 for all x. If the utility function is not differentiable at a given point x0 , approximations can be used without altering the results.) One is tempted to believe that for

1347

Chapter 7. Prospect Theory: Much Ado About Nothing?

143

LEVY AND LEVY Prospect Theory Reuisited

x > 0, the condition for dominance should be fo'[G(t)-F(t)]dt : 0, the conditions of the theorem guarantee that セBPN@

Finally, note that if the utility function is V ( w + x) and the inflection point is at x = 0, the proof is kept unchanged because F(w+x) and G(w+x) are simply shifted to the right by w with no change in the area enclosed between F and G. Necessity. It can be easily shown that if 0 [G(t)- F(t)]dt < 0 for some x 0 < 0, then there is some V E V M for which .6. 0 < 0. To show this, employ the same necessity proof of Hanoch and Levy (1969) for second-degree stochastic dominance. By a similar argument, one can show that J:[G(t)- F(t)]dt > 0 for x > 0 is also a necessary condition for MSD dominance.

J..'

References Allais, M. 1953. Le Comportement de l'homme rationnel devant le risque: Critique des postulates et axioms de I'ecole Americaine. Econometrica 21 503-546.

1348

b。イ「・ゥウセ@

N., M. Huang, T. Santos. 2001. Prospect theory and asset prices. Quart. J. Econom. 116 1-53. Benartzi. S., R. Thaler. 1995. Myopic loss aversion and the equity premium puzzle. Quart. J. Econom. 110(1) 73-92. - , - - 1999. Risk aversion or myopia? Choices in repeated gambles and retirement investments. Management Sci. 45(3) 364-381. Edwards, K. D. 1996. Prospect theory: A literature review. Internat. Rev. Financial Anal. 5(1), 13-38. Fishburn, P. C. 1982. Nontransitive measurable utility. J. Math. Psych. 26 31-67. Friedman, M., L. J. Savage. 1948. The utility analysis of choices involving risk. J. Political Econom. 56 279-304. Hadar, J., W. Russell. 1969. Rules for ordering uncertain prospects. Amer. Econom. Rev. 59 25-34. Hanoch, G., H. Levy. 1969. The efficiency analysis of choices involving risk. Rev. Econom. Stud. 36 335-346. Kahneman, D., A. Tversky. 1979. Prospect theory of decisions under risk. Econometrica 47(2) 263-291. Levy, H. 1992. Stochastic dominance and expected utility: Survey and analysis. Mimagement Sci. 38(4) 555-593. - - . 1998. Stochastic Dominance: Investment Decision Making Under Uncertainty. Kluwer Academic Publishers, - - . 2000. Cumulative prospect theory and the CAFM. Working paper, Hebrew uョゥカ・イウエケセ@ Jerusalem, Israel. - , Z. Wiener. 1998. Stochastic dominance and prospect dominance with subjective weighting functions. ]. Risk Uncertainty 16 147-163. Levy, M., H. Levy. 2001. Testing for risk-aversion: A stochastic dominance approach. Econom. Lett. 71 233-240. - - , - - , S. Solomon. 2000. Microscopic Simulation of Financial Markets: From Investor Behavior to Market Phenomena. Academic Press, San Diego, CA. Markowitz, H. M. 1952a. Portfolio selection. f. Finance 7 77-91. - - . !952b. The utility of wealth. J. Political Econom. 60 151-156. - . 1959. Portfolio Selection. John Wiley and Sons, New York. - - . 1987. Mean Variance Analysis, Portfolio CJwice and Capital Markets. Basil Blackwell, New York. Quiggin, J. 1982. A theory of anticipated utility, J. Econom. Behavior Organ. 3 323-343. - - . 1993. Generalized Expected Utility Theory: The Rilnk Dependent Model. Kluwer Academic Publishers, Boston, MA. Rothschild, M., J. Stiglitz. 1970. Increasing risk I. A definition. J. Econom. Theory 2 225-243. Shefrin, H., M. Statman 1993. Behavioral aspect of the design and marketing of financial products. Financial Management. 22(2) 123-134. Swaim, R. 0. 1966. Utility theory-Insights into risk taking. Harvard Bus. Rev. 44 123-136. Thaler, R. 1985. Mental accounting and consumer choice. Marketing Sci. 4(3) 199-214. Tobin, J. 1958. Liquidity preferences as behavior toward risk. Rev. Econom. Stud. 25 65-86.

MANAGEMENT SciENCE/Val.

48, No. 10, October 2002

144

M Levy and H Levy LEVY AND LEVY

Prospect Theory Revisited

Tversky, A., D. Kahneman. 1981. The framing of decisions and the psychology of choice. Science 211 453-480. - - , - - . 1986. Rational choice and the framing of decision. J. Bus. 59(4) 251-278. - , - - . 1992. Advances in prospect theory: Cumulative representation of uncertainty. ]. Risk Uncertainty 5 297-323.

Viscusi, W. K. 1989. Prospective reference theory: Toward an Explanation of the paradoxes.]. Risk Uncertainty. 2 235-264.

von Neuman, J., 0. Morgenstern. 1944. Theory of Games and Economic Behavior. Princeton University Press, Princeton, NJ. Yaari, M. 1987. The dual theory of choice under risk. Econometrica 55(1) 95-115.

Accepted by Phelim. P. Boyle; received February 2001. This paper was with the authors 2 months for 2 revisions.

MANAGEMENT SciENCE/Vol. 48, No. 10, October 2002

1349

Chapter 8

The Data of Levy and Levy (2002) "Prospect Theory: Much Ado About Nothing?" Actually Support Prospect Theory Peter P. Wakker Department of Economics, University of Amsterdam, Roetersstraat 11, Amsterdam 1018 WB, The Netherlands [email protected]

L

evy and Levy (Management Science 2002) present data that, according to their claims, violate prospect theory. They suggest that prospect theory's hypothesis of an S-shaped value function, concave for gains and convex for losses, is incorrect. However, all the data of Levy and Levy are perfectly consistent with the predictions of prospect theory, as can be verified by simply applying prospect theory formulas. The mistake of Levy and Levy is that they, incorrectly, thought that probability weighting could be ignored.

(Prospect Theory; Stochastic Dominance; Utility; Probability Weighting; Inverses)

Levy and Levy (Management Science 2002, henceforth LL) present data that, according to their claims, violate the S-shaped value function posited by prospect theory. This comment will show, however, that LL's data are in perfect agreement with prospect theory. Following LL, we will throughout restrict attention to outcomes that are not very extreme, say between $6,000 and -$6,000. The classical views on risk attitudes assumed universal risk aversion. Empirical studies have revealed a more complex, fourfold pattern in behavior. For gains, people are mostly risk averse; but for specific prospects, yielding a best outcome with a low probability (below 1/3), we often find risk-seeking behavior, as observed in gambling for instance. The pattern for losses is less clear but seems to be reversed. People are mostly risk seeking, but for prospects yielding a worst outcome with a low probability (below 1 /3), risk aversion can occur, as observed for instance in insurance. This fourfold pattern is based on extensive empirical evidence (reviewed in Starmer 2000 and Luce 2000) and entails extremity-orientedness whereby the best and worst

outcomes of a prospect are overweighted and middle outcomes are underweighted. Prospect theory models the fourfold pattern. We focus on the most recent, cumulative version of prospect theory (Tversky and Kahneman 1992). It assumes a utility or value Junction v(x) that isS-shaped: increasing for all amounts, concave for gains, and convex for losses. This assumed shape reflects the psychological phenomenon of diminishing sensitivity as one moves away from the "reference point." The reference point divides gains from losses, and is taken to be zero in this comment. Consider a prospect (or gamble) with outcomes x, :S · · · :S x1 :S 0 :S xk+l :S · · · :S x, having probabilities p 1 , . . . , p,. Prospect theory predicts that people will choose prospects according to the value given by k

I: 7T;Av(x;) + I:

v(x;),

7T1

(1)

j=k+l

where A > 0 is a loss-aversion parameter, and the 7TS are decision weights that are calculated based on the "cumulative" probabilities associated with the

0025-1909/03/4907 I 0979$05.00

MANAGEMENT SCIENCE © 2003 INFORMS Vol. 49, No. 7, July 2003, pp. 979-981

1526-5501 electronic TSSN

145

146

PP Wakker WAKKER

Data of Levy and Levy (2002) Jlctually Support Prospect Theory

outcomes. The details of this model are provided in the appendix. Following Markowitz (1952), LL posit that the value function is convex for gains and concave for losses, implying a reverse 5-shape-the opposite of that assumed by prospect theory. In their theoretical analysis, LL assume expected utility theory and, thus, assume that the decision weights in Equation (1) are simply equal to the probabilities associated with the outcomes. This coupled with their assumption about the form of the value function implies risk aversion (and, more strongly, secondorder stochastic dominance) for losses and risk seeking (and reversed second-order stochastic dominance) for gains, in contrast to the more complex fourfold pattern of observed behavior. LL develop a stochastic dominance rule that they call Markowitz stochastic dominance. Given expected utility theory, this rule allows them to show that some gambles are preferred to others for all value functions having the reverse 5-shape. LL test their hypothesis about the shape of the value function through three choice exercises that they refer to as head-to-head competitions, and in which they claim that an S-shaped value function would lead to one choice and a reverse S-shaped value function to another. In all three experiments, the majority choice corresponds to the choice that supposedly supports the reverse 5-shape, and LL interpret this as evidence contradicting prospect theory. Simple calculations show, however, that prospect theory with the functional forms and parameter estimates of Tversky and Kahneman (1992) correctly predicts the majority choice in all head-to-head competitions; these calculations are displayed in Table 1, with explanations given in the appendix. This finding is contrary toLL's claims (p. 1344, "Thus, we can state that at least 62% of the choices are inconsistent with prospect theory"). We conclude that LL's data actually support prospect theory. The error in LL's analysis is that they neglect the probability weighting function of prospect theory. They argue, "All probabilities given in the experiments are relatively large (p ::> 0.25), hence it is unlikely that subjective probability distortion plays an important role in the decision-making

980

Table 1

Prospects Yield Outcome

x with Probability p.

(a) LL's Experiment 1, Task 1 Prospect

p

0.50 0.50 0.25 0.75

X

-3,000 4,500 -6,000 3,000

(b) 0.25 0.25 0.25 0.25 0.25 0.25 0.25 0.25

-1,600 -200 1,200 1,600 -1,000 -800 800 2,000

0.50 0.50 0.25 0.75

-1,500 4,500 -3,000 3,000

lセウ@

0.45 0.42 0.29 0.57

-1,148 1,640 -2,112 1,148

PT

Choice

-483

71%

-743

27%

-216

38%

-138

62%

53

76%

-106

23%

Experiment 2 0.29 0.16 0.13 0.29 0.29 0.16 0.13 0.29

-660 -106 512 660 -437 -359 359 803

(c) LL's Experiment 3, Task 3 0.45 0.42 0.29 0.57

-624 1,640 -1,148 1,148

Note: Participants in Levy and Levy (2002) chose between the head-to-head prospect pairs F and G. The PT column gives the values of the prospects according to prospect theory, using the parameters estimated by Tversky and Kahneman (1992). Each " is the decision weight of outcome x, and v its utility/value. Bold printing indicates the majority choice, which is always the option preferred according to prospect theory because it has the higher value under that theory.

process" (p. 1341; a similar statement appears on p. 1346) 1 As Table 1b shows, the extreme outcomes of F and G in Experiment 2 have decision weights about twice as much as the intermediate outcomes 1 ln addition to this argument, LL suggest (p. 1344) that taking the outcomes to be equally likely in their experiment 2 "makes

any subjective probability distortion very unlikely." In their Footnote 15, they expand on this and say "This point was made by Quiggin (1982). Tn addition, any subjective transformation performed directly on the probabilities (as in prospl'ct thl'ory) will still attach an equal probability weight to each outcome." While in the original version of prospect theory (Kahneman and Tvcrsky 1979) the transformation was performed directly on the probabilities alone, this is not so in the current version and equally likely outcomes can be weighted differently. Contrary to "'What LL claim, Quiggin (1982) also argued that equally likely outcomes may be weighed differently (end of §1 ).

MANAGEMENT SciENCE/ Vol.

49, No.

7,

july 2003

Chapter 8. The Data of Levy and Levy (2002) Actually Support Prospect Theory

147

WAKKER

Data of Levy and Levy (2002) Actually Support rrospect Theory

under prospect theory (weights 0.29 versus 0.13 or 0.16), which deviates considerably from the equal weighting assumed by LL. While these calculations are based on the specific parameter assumptions suggested by Tversky and Kahneman (1992), LL's experimental results are also qualitatively consistent with the extremity-orientedness predicted by prospect theory. In each of LL's three head-to-head competitions, the majority chose the gamble that had both the best maximal outcome and the best minimal outcome, as would be done if only the extreme outcomes mattered. In conclusion, the data of LL support the predictions of Tversky and Kahneman's (1992) prospect theory. The incorrect claims of LL are mostly due to their overlooking the crucial role of probability weighting in prospect theory. While the data could also be consistent with other theories, it is extremely misleading to interpret this as evidence against prospect theory or to suggest that prospect theory is "much ado about nothing." In particular, the results of LL do not provide new insights into the shape of the value/utility function. Their hypothesis of convex utility for gains is contrary to the diminishing marginal utility assumed in classical analyses, the diminishing sensitivity assumed in prospect theory, and virtually all empirical findings of the vast literature on this topic.

Acknowledgments The author thanks John Payne, Thierry Post, Pim van Vliet, George Wu, and three anonymous referees for helpful comments, and the department editor for numerous thoughtful and constructive suggestions.

Appendix:

Prospect Theory

Prospect theory assumes, besides the utility or value function and the loss-aversion parameter, a probability weighting function w+: [0, 1J ---+ [0, 1] for gains, and a probability weighting jimction w [0, 1] -+ [0, 1] for losses. The

decision ·weights 7T in Equation (1) are defined as follows. If k ::: 1 then 7T1 = ur(p 1), and 7T; = w-(p 1 + · · · + p,) -w-(p1 +···+PH) for 2 :S i :S k. If k < n then 7T11 = w-(Pn) and 7Ti = w+(pll +···+pi)- zu+(p 11 +· · · +pw for n-1 セ@ j > k. Tversky and Kahneman (1992) estimated the following parametric form: v(x) = for x セ@ 0, v(x) = -(-x) 0 -Rs for x :S 0, 1)

セ@

po'' /(po'' + (1 - p)"')''06'), w-(p) セ@

pooo /(po"+

For prospect F in Table 1 b, the decision ,.veights are 7T1 = w-(0.25) = 0.29 for outcome x1 = -1,600, 7T2 = w-(0.50)w-(0.25) セ@ 0.16 for outcome セ@ -200, 1r, セ@ w-(0.50)- w+(0.25) セ@ 0.13 for outcome x3 = 1,200, and 7T-t = w+(0.25) = 0.29 for outcome x4 =1,600. The value ofF is 7T 1Av(-1,600)+7T 2 Av(-200)+ 7T3 v(1,200) + 7T4 v(1,600) = -215.70. The other prospects are evaluated similarly. A program to calculate prospect-theory values, -.;;vritten by Veronika KObberling, is available at http://wwwl. fee.uva.nl/ creed/-o;;'\rakker / miscella/ calculate.cpt.kobb / index.htm.

References Kahneman, Daniel, Amos Tversky. 1979. Prospect theory: An analysis of decision under risk. Econometrica 47 263-291. Levy, Moshe, Haim levy. 2002. Prospect theory: Much ado about nothing. Management Sci. 48 1334-1349. Luce, R. Duncan. 2000. Utility of Cains and Losses: MeasurementTheoretical and Experimental Approaches. Lawrence Erlbaum Publishers, London, UK. Markowitz, Harry M. 1952. The utility of wealth. [. Political Econom. 60 151-158. Quiggin, John. 1982. A theory of anticipated utility. J. Econom. Behaviour Organ. 3 323-343. Starmer, Chris. 2000. Developments in non-expected utility theory: The hunt for a descriptive theory of choice under risk. }. Econom. Literature 38 332-382. Tversky, Amos, Daniel Kahneman. 1992. Advances in prospect theory: Cumulative representation of uncertainty. J. Ri:;k Uncertainty 5 297-323.

Note from the Editor-in-Chief. Levy and Levy, the authors of the paper "Prospect Theory: Much Ado About Nothing," also wrote the following paper: "Experimental Test of the Prospect Theory Value Function: A Stochastic Dominance Approach" (Orgnnizatio11 Behavior and Huma11 Decision Processes 89, 2002, pp. 1058-1081). The two papers present very similar experiments and results. The failure of Levy and Levy to cross-cite these papers is a violation of proper scholarly practice and may have contributed to the controversy surrounding their work.

Accepted by James E. Smith; received Nmxmbcr 27, 2002. This paper ·was 'With the author 1 month for 2 revisions.

MANAGEMENT SCTENCE/Vol. 49, No. 7, july 2003

981

This page intentionally left blank

Chapter 9

Prospect Theory and Mean-Variance Analysis Haim Levy Hebrew University

Moshe Levy Hebrew University The experimental results of prospect theory (PT) reveal suggest that investors make decisions based on change of wealth rather than total wealth, that preferences are S-shaped with a risk-seeking segment, and that probabilities are subjectively distorted. This article shows that while PT's findings are in sharp contradiction to the foundations of mean-variance (MV) analysis, counterintuitively, when diversification between assets is allowed, the MV and PT-efficient sets almost coincide. Thus one can employ the MV optimization algorithm to construct PTefficient portfolios.

The Markowitz (1952a)-Tobin (1958) mean-variance (MV) rule is probably the most popular investment decision rule under uncertainty in economics and in finance, and it is widely employed by both academics and practitioners. The strength of the MV analysis is that in the case of normal return distributions the choice of any expected utility maximizing risk-averse individual will be according to the MV rule. 1 The MV framework is the foundation of the Sharpe (1964)-Lintner (1965) capital asset pricing model (CAPM), which is a cornerstone of modern finance. Moreover, the MV framework provides a very important practical procedure for the construction of efficient portfolios. While standard economic theory and, in particular, MV analysis assume expected utility maximization and risk aversion, in a breakthrough article Kahneman and Tversky (1979) show that the actual behavior of individuals systematically and consistently violates these We are grateful to Campbell Harvey aud to the auouymous referee of this jourual for their mauy helpful commeuts aud suggestions. Fiuaucial support for this study was provided by the Krueger Fuud aud by the Zagagi Fund. Address correspondence to Haim Levy, Jerusalem School of Business Administration, The Hebrew University, Jerusalem 91905, Israel, or e-mail: [email protected]. 1

Formally, when the returu distributions are normal, the MV rule coincides with second-degree stochastic dominance (SSD) [see Hanoch and Levy (1969) and Levy (1998)]. The MV rule is also valid if one replaces the assumption of normality with the assumption of quadratic utility [Tobin (1958)]. However, this latter assumption is problematic as it implies increasing absolute risk aversion and utility which is decreasing beyond a certain wealth level [see Arrow (1971)]. In contrast, the assumption of normal returu distributions is approximately valid for investment horizons longer than one month but shorter than several years. [For very short horizons, the returu distribution is "fat tailed," as shown by Fama (1963), Mandelbrot (1963), Mantegna and Stanley (1995), and others]. The Review of Financial Studies Vol. 17, No. 4 © 2004 The Society for Financial Studies; all rights reserved. doi:l0.1093/rfs/hhg062 Advance Access publication October 15, 2003

149

150

H Levy and M Levy The Review of Financial Studies I v 17 n 4 2004

assumptions. The alternative framework they present is well known as prospect theory (PT). The three main findings of PT are Individuals make decisions based on change of wealth rather than the total wealth, which is in direct contradiction to expected utility theory. (b) Risk aversion does not globally prevail-individuals are riskseeking regarding losses. (c) Individuals distort objective probabilities and subjectively transform them in a systematic way. Tversky and Kahneman (1992) extend PT and suggest cumulative prospect theory, in which the transformation is of the cumulative probability rather than the probability itself. The transformation of cumulative probability allows cumulative prospect theory to avoid violations of firstdegree stochastic dominance (FSD). In this study we refer to the cumulative version of PT. (a)

Many studies support the above findings of PT. 2 Only fairly recently, however, researchers have begun to explore the implications of PT for economics and finance. 3 This article investigates the implications of PT to Markowitz's portfolio theory. It is clear that each ofPT's main elements is in sharp contradiction to the assumptions of MV analysis: the MV framework assumes expected utility maximization and risk aversion while the above-mentioned elements (a) and (b) are at odds with both of these assumptions. Furthermore, if individuals subjectively distort probabilities, the MV analysis seems inappropriate because, even if the objective probability distributions are normal, the subjective distributions are not. Thus it would seem that the MV framework is completely incompatible with PT. Does PT imply that the MV framework, which is so central in finance, should be abandoned? Can any of the effective MV machinery (e.g., the efficient diversification algorithm it provides) be employed to benefit investors with PT preferences? This article addresses these issues. We employ stochastic dominance rules to prove our main claim, which is that the PT and MV efficient sets almost coincide. To be more specific, assume e 1 and e 2 are two random variables. Then the theory of stochastic 2

For an extensive review of studies supporting PT, see Edwards (1996). A number of studies reject PT or some of its elements; see, for example, Battalio, Kagel, and Jiranyakul (1990), Casey (1994), Harless and Camerer (1994), and Luce (2000). In two recent experimental studies, Levy and Levy (2002a,b) reject the PT S-shape value function and find support for the Markowitz (1952b) value function, which has a reverse S-shape. It is interesting to note the main results presented here hold also for the set of reverse S-shape value functions. The proof of this statement is available from the authors upon request.

3

For example, Shefrin and Statman (1985) and Ferris, Haugen, and Makhija (1988) show that PT can explain the empirically observed "disposition effect"- the disposition to sell winning stocks too early and to ride losing stocks too long. Thaler (1985) investigates the implications of PT to marketing. Benartzi and Thaler (1995) show that PT may explain the equity premium puzzle. Gomes (2000) investigates the implication ofPT to trading volume, and Barberis, Huang, and Santos (2001) and Levy, DeGiorgi, and Hens (2003) investigate the consequences of PT for asset pricing.

1016

Chapter 9. Prospect Theory and Mean-Variance Analysis

151

Prospect Theory and Mean- Variance Analysis

dominance is concerned with the conditions on the CDFs of e 1 and e 2 under which E[u(e 1 )] < E[u(e 2 )]

for some set u E U. The standard concepts of stochastic dominance refer to three situations: (a) U is the set of all increasing functions (firstdegree stochastic dominance, FSD); (b) U is the set of increasing and concave functions (second-degree stochastic dominance, SSD); (c) U is the set of increasing and concave functions with a positive third derivative, u"' > 0, which reflects a preference for positive skewness (third-degree stochastic dominance, TSD). Recently the theory of stochastic dominance has been extended to the set of functions that are S-shaped, and that include the family of prospect theory value functions (prospect stochastic dominance, PSD), which have been suggested by Kahneman and Tversky (1979) as a better description of individuals' decisions under uncertainty. In this article we link the standard and familiar concepts of stochastic dominance to the new concept of prospect stochastic dominance. In particular, what are the implications of PSD for the traditional meanvariance analysis of Markowitz? It is well known that SSD is equivalent to mean-variance efficiency when the CDFs are normal. We show that in this case PT efficiency is "almost" identical to mean-variance efficiency. We focus on the canonical case of normal distributions, and offer two results: First, a characterization of PSD efficiency versus MV efficiency when the initial CDFs are not subjectively transformed and the focus on the first two moments is warranted. Second, we characterize PSD efficiency versus MV efficiency when we allow for subjective transformations of the CDFs that preserve FSD. Thus, while in most empirical studies in finance and economics probability distortion is ignored, in this article we analyze MV and PSD efficiency once with no probability distortion and once with probability distortion exactly as suggested by CPT. Note that in this second case the transformed CDFs are not necessarily normal, and may have nonzero skewness and higher odd moments. We show that in both cases the PSD-efficient set is a subset of the (objective) MV -efficient frontier, and it typically almost completely coincides with the MV -efficient set. The implication of this result is that one can employ the well-known MV optimization techniques to derive the efficient set for PT investors. Both frameworks, MV analysis and PT, are strengthened by this result: the MV framework is shown to be valid for a broader class of preferences, and PT is provided with an algorithm for the construction of PT -efficient portfolios. Thus, despite the sharp contradiction between the assumptions in the PT and MV frameworks, MV analysis turns out to be very central in portfolio optimization even under PT preferences.

1017

152

H Levy and M Levy The Review of Financial Studies I v 17 n 4 2004

The focus of this article is on the canonical case of normal return distributions. However, as long-horizon empirical return distributions tend to be positively skewed, we also analyze the case of lognormal return distributions. The results for the lognormal case, which are derived in Appendix B, are very similar to those of the normal case. Thus it seems that the results of this article are quite robust and do not hinge on specific assumptions regarding the functional form of the return distribution. The structure of this article is as follows: In Section 1 we provide a review of the various decision rules employed in the article, and in particular the prospect stochastic dominance rule. Section 2 contains the theoretical results. Section 3 provides an empirical analysis. Section 4 concludes. 1. Investment Decision Rules

This section briefly reviews the main investment decision rules discussed in the article, with a special emphasis on the recently developed and less familiar PSD rule, which is a criterion for dominance by all S-shaped value functions. 1.1 The MV rule

Let F and G be two investments with stochastic outcomes. Investment F dominates G by MV iff ILF ;:::: !La

and

£Tp

:s; £Ta,

(1)

with at least one strict inequality, where /LF and /LG denote the expected values of investments F and G, and £Tp and £Ta denote the respective standard deviations. Under normal distributions and risk aversion the MV rule coincides with EU maximization [see Markowitz (1952a) and Tobin (1958)]. 1.2 First-degree stochastic dominance

Let F and G be two stochastic investments with cumulative distributions F(x) and G(x), and let U 1 denote the set of all nondecreasing utility functions. Then, F dominates G for all utility functions U E U 1 (i.e., EUF;:::: EUa) iff F(x) :s; G(x)

for all x,

(2)

with at least one strict inequality. When the two distributions are normal and £T 1 = £T2 , a dominance by the MV rule coincides with FSD dominance, namely MV {c} FSD. Note that expected utility can be expanded to

1018

Chapter 9. Prospect Theory and Mean-Variance Analysis

153

Prospect Theory and Mean- Variance Analysis

a Taylor series such that choices depend on all of the distribution moments, for example, mean, variance, skewness, etc. However, if F(x) :s; G(x) for all x, then F is preferred over G by all nondecreasing utility functions, regardless of preference for variance, skewness, and higher moments. 1.3 Second-degree stochastic dominance Let F and G be two stochastic investments with cumulative distributions F(x) and G(x), and let U 2 denote the set of all nondecreasing riskaverse utility functions (U';:::: 0, U" < 0). Then, F dominates G for all U EU 2 , iff

lssn(x) =

1:

[G(z)- F(z)]dz;:::: 0 for all values x,

(3)

with at least one strict inequality. When the distributions are normal, Markowitz's MV rule coincides with SSD, namely SSD {c} MV. In other words, in the EU framework with normal distributions and risk aversion, MV is an optimal rule. For proofs and discussion ofFSD and SSD, see Fishburn (1964), Hadar and Russell (1969), Hanoch and Levy (1969), and Rothschild and Stiglitz (1970). For a survey of SD rules, see Levy (1998). While the above rules are widely used, the recently developed PSD rule [Levy and Wiener (1998)] is still well known. We therefore discuss this rule in some detail below. 1.4 Prospect stochastic dominance Let F and G be two stochastic investments with cumulative distributions F(x) and G(x), and let Us denote the set of all S-shaped utility (or value) functions (U';:::: 0 for all x # 0, U" > 0 for x < 0, and U" < 0 for x > 0). F dominates G for all U E Us iff

lpsn= 1x[G(z)-F(z)]dz;:::O

ヲッイ。ャセZウ[NクL@

(4)

with at least one strict inequality. Graphically F dominates G by PSD if and only if the area enclosed between the two cumulative distributions G and F is positive for any range [x, x] with x :s; 0 and x;:::: 0. For a discussion and proof of the PSD rule, see Levy (1998) and Levy and Wiener (1998). To illustrate the PSD rule, consider the two prospects given in Table 1. Which of these two prospects will various individuals with S-shaped value functions prefer? Tversky and Kahneman (1992) suggest the following

1019

154

H Levy and M Levy The Review of Financial Studies I v 17 n 4 2004

Table 1 Two alternative investments F

Probability

G

Outcome

Probability

Outcome

-12 3.5 10 14

1/3 1/3 1/3

-10 5 15

1/4 1/4 1/4 1/4

specific form for the value function: V(x)

x" { -A(

=

-xl

if X;:::: 0 if X< 0'

(5)

and they experimentally estimate the parameters of the "typical" individual as a= 0.88, f3 = 0.88, and A= 2.25. It is straightforward to calculate the expected value of the two prospects in Table 1 for such an individual: 4 =

セHMRNUIQ

PX@

X@ X@ X@ + セSNUᄚ + セQPᄚ + セQTᄚ EVa= t(-2.25)(-(-10))088 + t5o88 + t15o88 = -0.703.

EVF

=

0.190

Thus, as EVF >EVa, this implies that an individual with the S-shaped preferences of Equation (5) with the above parameters would prefer prospect F over G (see footnote 4). What about individuals with different parameters, or with an altogether different functional form of the Sshaped value function? Would they all prefer F, or would some individuals with S-shaped value functions prefer prospect G? The PSD criterion provides an answer. Figure 1 depicts the cumulative distributions of the two prospects in Table 1. The numbers in the figure denote the areas enclosed between the two cumulative distributions. It is easy to verify from this figure that Equation (4) holds, and therefore F dominates G by PSD. This implies that for these two specific prospects, individuals with any S-shaped function, and not only the specific function given by Equation (5), would prefer F over G. As in the other stochastic dominance rules, the strength of the PSD rule is that it provides a criterion for the preference of all individuals with S-shaped preferences, with no need to know the specific parameters or the exact functional form of the value function. This 4

For simplicity, in this example we employ the objective probabilities. The possible distortion of probabilities is discussed in the next section, where our general results are derived.

1020

Chapter 9. Prospect Theory and Mean-Variance Analysis

155

Prospect Theory and Mean- Variance Analysis

1. 0

セ@

0.8

Fr - - - -

:0 Ill .0 0

c.

-0.666

G 0.6

... o.a33

Ql

セ@

>

:::J

E

f F, and the"-" areas are those where the opposite holds. The PSD condition [Equation (4)] states that F dominates G by PSD iff the area enclosed between G and F ヲイッュセ@ to x is positive for any セM\Z@ 0- £T0 ). 6 In a portfolio context, however, when one can diversify across assets, surprisingly, it turns out that there is a very close relationship between the MYand PSD-efficient sets. 7 The two theorems below state this relationship. Both theorems are based on the following three standard assumptions: Assumption 1. Returns are normally distributed. Assumption 2. Portfolios can be formed without restrictions. Assumption 3. No two assets are perfectly correlated, IRiil < 1 for all i,j. 6

Conflicting dominance, that is, F dominates G by MV and G dominates F by PSD, can only be obtained in the special case of /.LF = /.LG· This is because /.LF > /.LG is a necessary condition for the dominance ofF over G by both rules. /.LF :> /.LG is obviously a necessary condition for MV dominance. It is also necessary for PSD dominance, because /.Lp- !La= J0000 [G(z)- F(z)[dz [see Hanoch and Levy (1969)]. Taking -oo and x oo in Equation (4), we see that /.LF :> /.LG is a necessary condition for the PSD dominance ofF over G.

"'= 7

=

Strictly speaking, PT as postulated by Kahneman and Tversky refers only to "one-shot" lotteries and is silent about individuals' behavior when confronted with a simultaneous series of lotteries, as in the portfolio context. Here we assume that the investor is faced with a universe of portfolios (e.g., those offered by a multitude of mutual funds) and individuals assets alike, and treats each of these as a lottery according to PT.

1022

Chapter 9. Prospect Theory and Mean-Variance Analysis

157

Prospect Theory and Mean- Variance Analysis

(a) : F dominates G by MV , but not by PSD 1.0

cr,=0.4

f1, =0. 1 flG=0.05

g

crG=0 .8 - area

0.8

:0

セ@ FsD G =? T(F) >FsD T(G); see Quiggin (1982) and Levy and Wiener (1998)]. Indeed, the extension of PT called cumulative prospect theory (CPT) employs a transformation of the cumulative distributions which maintains FSD [see Tversky and Kahneman (1992)]. Theorem 2 states the relationship between the MV- and PSD-efficient sets when probability distortion takes place.

1025

160

H Levy and M Levy The Review of Financial Studies I v 17 n 4 2004

(a) F' dominates F by PSD (and by FSD) 10

q .0

08

cu

.0 0

ct

Q)

06

>

'@ (ij

E ::0

04

0

Outcome

(b) 0' dominates 0 by PSD (but not by SSD) 1.0

0

q

0.8

:0

cu

.0 0

ct

0.6

Q)

セ@

>

(ij

E ::0

0.4

0

0.2

Outcome Figure 4 Graphical explanation of Theorem 1 (a) Depicts the CDFs ofF' and F, where F has the same standard deviation as F', but a higher mean. In this case F dominates F' by PSD (and actually also by FSD). (b) Shows the CDFs of portfolios 0 and 0', where 0 is the minimum variance portfolio and O' is a portfolio slightly above it on the MY-efficient frontier. O' has a higher mean, but only a slightly higher standard deviation compared with 0, and therefore O' dominates 0 by PSD (but not by SSD).

Theorem 2. Suppose that the objective probabilities are subjectively distorted by any transformation that does not violate FSD, for example, the cumulative prospect theory transformation. Then, the PSD-efficient set is a subset of the MV-efficient set.

1026

Chapter 9. Prospect Theory and Mean-Variance Analysis

161

Prospect Theory and Mean- Variance Analysis

Proof The proof that the PSD-efficient set is a subset of the MY-efficient frontier is as in Theorem 1. No portfolio interior to the MY-efficient frontier, such as portfolio F' in Figure 3, can be PSD efficient, because it is FSD dominated by portfolio F on the frontier (see the proof of Theorem 1 in the appendix). As we are considering FSD-maintaining probability transformations, the FSD dominance of F over F' with the objective probabilities implies that F dominates F' for every individual with an increasing utility/value function, even if he subjectively distorts the probability distributions. Namely, if F dominates F' by FSD, then also T(F) dominates T(F') by FSD. As FSD =? PSD, portfolio F' is dominated by portfolio F, also by PSD. Thus the PSD-efficient set is a • subset of the MY -efficient set.

Theorem 2 proves that the PSD-efficient set cannot be larger than the MY-efficient set, but like in Theorem 1, it could be strictly smaller. However, unlike the case of Theorem 1, where the objective probabilities are employed, when probabilities are subjectively distorted we cannot generally restrict the location of the PSD-inefficient portfolios to a specific segment of the MY-efficient frontier, that is, the set of MY-efficient portfolios that are not PSD efficient cannot be characterized neatly as a segment between the minimum variance portfolio and the tangency point from the origin. The reason for this is that the probability transformation may change the perceived portfolio mean and standard deviation. Therefore the intuition behind result (ii) of Theorem 1 does not hold, because the perceived efficient frontier is not necessarily smooth and with a diminishing slope. As a result, some portfolios that are located on the MY-efficient frontier and which are PSD efficient with the objective probabilities may become PSD inefficient when the probabilities are transformed. Similarly some portfolios that are located on the MYefficient frontier and are PSD inefficient with the objective probabilities may become PSD efficient with the transformed probabilities. As this claim is not obvious, let us demonstrate it with an example. Consider two portfolios, F and G, located on the MY-efficient frontier, with normal CDFs and the following parameters: /LF = 0.25, O"p = 0.50; /La= 0.23, O"a = 0.05. First, note that when the objective probabilities are employed, F is PSD efficient. This follows from the fact that F is necessarily on the segment of the MY-efficient frontier, which is to the right of the point of tangency from the origin to the frontier (point a in Figure 3), because F is on the segment of the frontier キィ・イセ@ is diminish< セᄋ@ (IfF and G were located below point a ing, that is, /LF >/La 。ョ、セZ@ than we would have ィ。、セZ@ > セLキィゥ」@ does not hold for the parameters in this example.) Thus F is PSD efficient with the objective probabilities. (G may be PSD efficient or inefficient, depending on its location on the frontier.) However, when a subjective probability transformation is

1027

162

H Levy and M Levy The Review of Financial Studies I v 17 n 4 2004

employed, F may become PSD inefficient. For example, consider the two CDFs, F and G, with objective probabilities as given in Figure Sa, and the monotonic (and FSD maintaining) probability transformation T(F) = F 0 · 2 . With this transformation F becomes PSD inefficient-G dominates it. This dominance is revealed in Figure 5b, which depicts the transformed CDFs. The figure shows that with the transformed cumulative distributions F is PSD dominated by G: T(F) is above T(G) for x < 0, and for x > 0 the"+" area is larger than the"-" area [see Figure 5b and Equation (4) with T(F) and T(G) replacing F and G]. Note that as we have a PSD dominance of Gover F with the transformed probabilities, it must be that G's perceived mean is larger than the perceived mean of F, (/LT(F) < fLT(G)) (see footnote 6). Thus when subjective probability transformation is employed, portfolios that are PSD efficient with the objective probabilities may become inefficient, and the segment of the MY-efficient frontier that is PSD inefficient is generally no longer necessarily restricted to be between the minimum variance portfolio and the point of tangency from the origin. 2.1 The role of skewness We focus in this article on the MY frontier. Some researchers emphasize the importance of skewness and advocate the mean-variance skewness (MYS) criterion [see, e.g., Arditti (1967), Rubinstein (1973), Levy and Sarnat (1984), Harvey and Siddique (2000), and Harvey et al. (2002)]. This section discusses the role of skewness in our analysis. The article discusses three cases: normal distributions (Theorem 1), transformed normal distributions (Theorem 2), and lognormal distributions (both transformed or not; Theorem 3 in Appendix B). Let us deal with each case separately. 2.1.1 Normal distributions. Here there is no skewness. Theorem 1 proves that all PT investors will choose a portfolio from the MY-efficient frontier. Moreover, in the case of normal distributions, the FSD-, SSD-, and TSD-efficient sets exactly coincide with the MY -efficient frontier: Theorem 1 first shows that the FSD-efficient set is a subset of the MY frontier, and as FSD =? SSD =? TSD, the SSD- and TSD-efficient sets are also subsets of the MY frontier. Second, the fact that, for any two normal distributions with fLl > fL 2 and £T 1 > £T 2 , there is no FSD, SSD, or TSD dominance implies that all portfolios on the MY efficient frontier are FSD, SSD, and TSD efficient. Thus the FSD-, SSD-, TSD-, and MY -efficient sets are exactly identical. 2.1.2 Transformed normal distributions. This case is dealt with in Theorem 2. The theorem proves that all portfolios located below the MY -efficient frontier are inefficient by FSD even if the distributions are

1028

Chapter 9. Prospect Theory and Mean-Variance Analysis

163

Prospect Theory and Mean- Variance Analysis

(a) : Objective Probabilities - No PSD

セ@

:.0

0.8

rn

.0

e0..

0.6

"S

E

:l

u

0.4

0

-1

outcome (b): Transformed Probabilities - G dominates F by PSD

1.0

0.8

セ@

:.0

T(F)

rn

.0

e0..

0.6

@セ "S E

0.4

::J

u

0. 2

0.0 -1

0

outcome Figure 5

Cumulative probability transformation and PSD A portfolio that is PSD efficient when the objective probabilities are employed may become inefficient when the transformed probabilities are employed. The figure depicts the CDFs of two hypothetical portfolios on the MY-efficient frontier, F and G, where /LF = 0.25, ap = 0.50; /LG = 0.23, a a = 0.05. (a) Depicts the objective probabilities, where no PSD exists. (b) Shows the transformed distributions with the monotonic transformation T(F) = F 0 ·2 , T(G) = G 0 ·2 While F is PSD efficient with the objective probabilities (as shown in the text), when the transformed probabilities are employed, G dominates F by PSD: Equation (4) holds with the reversed roles ofF and G (i.e., the"+" area in the figure is larger than the "-" area). Thus, with the transformed probabilities, F is PSD inefficient.

1029

164

H Levy and M Levy The Review of Financial Studies I v 17 n 4 2004

transformed. As FSD =? SSD =? TSD, TSD investors will also not choose a portfolio that is not on the MY-efficient frontier. However, in this case the transformed distributions may be skewed and MYS and MY investors may choose different portfolios on the frontier. We cannot tell which portfolios will be chosen unless we know the exact preferences as well as the precise parameters of the transformed distributions (i.e., skewness, which is determined by the specific transformation employed). Moreover, suppose that all investors "switch" from quadratic preferences to cubic preferences. This will effect portfolio demands and the equilibrium prices, and thus the shape of the frontier, but still all investors will choose their portfolios from the MY frontier. In this case it is possible that the SSD-, and in particular the TSD-efficient sets, will be different from the MYefficient frontier, but Theorem 2 shows that these efficient sets are always subsets of the objective MY-efficient frontier. 2.1.3 Lognormal distributions. Here skewness prevails with the objective distributions, and of course may also prevail with the transformed distributions. If investors have a preference for skewness, skewness may be very important and will be priced. However, Theorem 3 shows that when distributions are lognormal the FSD-efficient set is a subset of the MYefficient frontier (this result holds both with and without probability transformation; see Appendix B). Hence no investor with increasing utility will choose an MY-inefficient portfolio. In particular, all MYS investors will choose their portfolios from the MY-efficient frontier. To sum up, in all of the above three cases, MYS and PT investors will generally choose different portfolios, but all of them will choose their portfolios from the objective MY-efficient frontier. Theorem 1 states that at most the lower part of the MY -efficient frontier (segment Oa in Figure 3) is PSD inefficient. Theorem 2 states that the PSD-efficient set is a subset of the MY-efficient frontier, but it does not state which parts of the frontier are PSD inefficient. The important result of these theorems is that in the PT framework, with or without probability transformation, one will always choose his optimal portfolio from Markowitz's MY-efficient set. In the next section we numerically investigate which part of the MY -efficient set is PT efficient by employing the empirical MY frontier and the specific probability transformation suggested by Tversky and Kahneman (1992).

3. The Empirical PT-Efficient Set Theorems 1 and 2 state that the PT -efficient set is a subset of the MYefficient frontier, but they provide only partial information (in the case that the objective probabilities are employed) or no information (in the case of probability distortion) regarding the exact part(s) of the MY-efficient

1030

Chapter 9. Prospect Theory and Mean-Variance Analysis

165

Prospect Theory and Mean- Variance Analysis

frontier that is PSD inefficient. In this section we numerically find the empirical PSD-efficient set with and without probability distortion. We select a random sample of 50 stocks from the Center for Research in Security Prices (CRSP) data set, and employ their monthly rates of return over the period January 1980 to January 2000 to estimate the means and the variance-covariance matrix. These parameters, in turn, are employed to derive the MY-efficient set, as shown in Figure 6. To find out which portfolios on the frontier are PSD inefficient, we go over each portfolio (with discrete increments of 0.0001 in f.L) and check its efficiency by comparing it with all portfolios located on the efficient frontier, and numerically testing whether it is PSD dominated. 9 We conduct this analysis twice: once assuming that the return distributions are normal, and once by transforming the normal return distributions with the subjective probability transformation suggested by Tversky and Kahneman (1992): T(F)F" - (F" + (1- F)") 1/r

(6)

with y experimentally estimated as 0.6. 10 The results are shown in Figure 6. In the case where the objective probabilities are employed, the PSD-efficient set is the MY-efficient frontier except segment Oc. In the case where the PT probability transformation of Equation (6) is employed, the PSD-efficient set is the MY-efficient frontier except segment Ob. In both cases, only a small lower part of the MY -efficient frontier is relegated to the PSD-inefficient set. It is perhaps surprising that when the probabilities are distorted, the part of the MY frontier that is PSD inefficient is even smaller than when the objective probabilities are employed, that is, empirically the probability transformation actually makes the MY- and PSD-efficient sets even more similar. 11 It is important to point out that while the main focus of this article is on the standard case of normal return distributions, our theoretical and empirical results can be extended to the case of lognormal distributions

9

The numerical testing of PSD is greatly simplified in the case of normal CDFs (see the lemma in the Appendix). As two normal CDFs with different standard deviations cross exactly once, and as Equation (6) is monotonic, the subjectively transformed cumulative probability distributions also cross exactly once (and at the same x 0). Thus we can employ Equations (A. I) and (A.2) of Appendix A to check for PSD between the transfonned distributions without worrying about 1nultiple crossings of the distributions.

10

Tversky and Kahneman estimate y separately for gains and losses. For gains they estimate y = 0.61, while for losses they estimate y = 0.69. Camerer and Ho (1994) estimate the parameter y as about 0.6.

11

A possible explanation for this may be as follows: PSD relegates the low-mean, low-variance part of the MY-efficient frontier to the inefficient set. The PT probability transformation of Equation (6) has the effect of increasing the perceived likelihood of low-probability extreme events [see Tversky and Kahneman (1992)]. Thus this transformation shifts probability weight to the extremes and therefore tends to increase the perceived variance. As a result, the range of perceived low-mean, low-variance portfolios becomes smaller under this probability transformation.

1031

166

H Levy and M Levy The Review of Financial Studies I v 17 n 4 2004

0.030

0.025

0.020 -

Inefficient with objective probabilities

Inefficient with transformed probabilities

0.015 +------.-------,---------r------.-0.030 0.035 0.040

a Figure 6 The empirical MV- and PSD-efficient sets The MY-efficient set is derived for a random sample of 50 stocks from the CRSP monthly return dataset. The PSD-efficient set is found by employing Equations (A.!) and (A.2). 0 is the minimum variance portfolio. When the objective probabilities are employed, the PSD-efficient set is the MY-efficient frontier excluding segment Oc. This is the same as Baumel's efficient set with k = 1.5. When the transformation of cumulative probability, as suggested by cumulative PT, is employed [Equation (6)], the PSD-efficient set is the MY-efficient frontier excluding segment Ob.

as well (see Appendix B). The results for the lognormal case are very similar to those obtained in the normal case, which suggests that the results are robust to the exact mathematical form of the return distributions. Finally, it is interesting to note that as early as 1963, Baumol suggested an intuitive investment criterion that does not rely on expected utility maximization. Baumol's criterion asserts that F dominates G iff

f.Lp- kcrp ;:::: f.La- kcra,

where k is a subjective positive value related to Chebechev's inequality. It turns out that Baumol's efficient set is a subset of the MY-efficient set, relegating to the inefficient set the lower part of the MV -efficient set (like segments Oc or Ob in Figure 6). The size of the efficient set relegated to the inefficient set by Baumol's criterion depends on the value k one selects: the larger k, the smaller Baumol's inefficient set; fork---+ oo, the Baumol and the MV -efficient sets coincide. With the actual empirical data corresponding to Figure 6, we investigate what value k is needed in Baumol's criterion

1032

Chapter 9. Prospect Theory and Mean-Variance Analysis

167

Prospect Theory and Mean- Variance Analysis

such that Baumol's efficient set and the PSD-efficient set with no probability distortion coincide. As we can see from Figure 6, the two efficient sets coincide fork セ@ 1.5, namely Baumol's lower bound is 1.5 standard deviations from the mean. 12 Of course, this particular value of k is for illustration purposes only, as k is in general a function of the particular dataset under consideration. Thus we see some similarity between Baumol's and the PSD-efficient sets, where both relegate to the inefficient set portfolios with low mean and low variance. However, recall that there is one important difference; while Baumol's rule has no theoretical justification, the PSD-efficient set is based on the foundations of PT. 4. Conclusion

The MV rule is probably the most common investment rule under uncertainty among both academics and practitioners. Moreover, the MV rule is very important as it is the foundation of the cornerstone Sharpe-Lintner capital asset pricing model (CAPM) and other theoretical models. Kahneman and Tversky (1979), and Tversky and Kahneman (1992), based on experimental results, suggest PT as an alternative to the EU paradigm. According to PT, the value function is S-shaped with a riskseeking segment, it is a function of change of wealth rather than total wealth, and in addition, individuals subjectively distort probabilities by a transformation of the cumulative probability. These experimental results seem to be fatal to EU in general and to the MV framework in particular. First, employing change of wealth rather than total wealth contradicts the EU paradigm, and thus casts doubt on the validity of the MV analysis, which is derived in the EU framework. Second, the risk-seeking segment of the PT value function contradicts the risk-aversion assumption, which is the foundation of the MV rule. Finally, when individuals distort probabilities, the perceived distributions are generally not normal, even if the objective distributions are, which again seems to be fatal to the MV analysis. We show in this study that, in general, when considering two alternative investments F and G, there is no relationship between the PT- and MYefficient sets: F may dominate G by PSD (which corresponds to PT) and not by MV, and vice versa. This is true even when normality is assumed, let alone in the general case when no restrictions are imposed on the distributions. However, in a portfolio context when diversification between assets is allowed, as is common in the security market, and as suggested by Markowitz (1952a), and when individuals face portfolios 12

For any given value of k, Baumel's efficient set is the segment of the MY-efficient frontier to the right of the point where a straight line of slope k is tangent to the hyperbola. For the PSD-efficient set obtained with the transformed probability distributions we obtain a slightly higher value of k (k = 1.7).

1033

168

H Levy and M Levy The Review of Financial Studies I v 17 n 4 2004

(mutual funds) as well as individual assets, we find that the PT-efficient set is a subset of the MY -efficient frontier. This result holds both for the case where objective probabilities are employed, as well as for the case where probabilities are subjectively transformed, as suggested by cumulative PT. When the objective probabilities are employed, we theoretically show that at most a small segment of low-mean low-variance portfolios is relegated to the PSD-inefficient set. With probability transformation the segment of the MY-efficient set relegated to the PT-inefficient set cannot be determined in general. However, empirical analysis reveals that the PSD rule (with and without probability distortion) relegates to the inefficient set a very small segment of the MY-efficient set corresponding to very riskaverse investors. Thus, surprisingly, even with the S-shaped value function, which contains a risk-seeking segment and which is a function of change of wealth rather than total wealth, and even with probability distortion, the PT- and MY-efficient sets almost coincide. It is interesting to note that PT relegates to the inefficient set some segment of the MY-efficient set in a similar way to the relegation of portfolios to the inefficient set by Baumol's (1963) intuitive criterion. In Baumol's criterion risk is measured by E-klr, where k is a positive constant reflecting the safety requirement of the individual investor. We find in the empirical analysis that the PSD rule relegates to the efficient set the same segment relegated by Baumol's criterion corresponding to the value of approximately k= 1.5. Note that the similarity of the MY- and PT-efficient sets does not necessarily imply that the CAPM equilibrium holds when investors have PT preferences. In particular, for some PT preferences the risk-seeking segment of the utility function may imply infinite borrowing, which contradicts the notion of market equilibrium. The main results of this article are derived for the canonical case of normal return distributions. However, it is well known that for long investment horizons the return distributions are typically positively skewed. To what extent do the results of this article hold when the return distributions are positively skewed? In order to answer this question we analyze the PT-efficient set in the case of lognormal return distributions (see Appendix B). We find that even in this case the PT- and MY-efficient sets almost coincide. Thus while skewness may be priced (and may perhaps be even more important in the case of PT investors with high sensitivity to losses), the set of PSD-cfficicnt portfolios is very similar to the MY -efficient frontier. In conclusion, the cumulative PT- and the well-known MY-efficient sets almost coincide. What seems a priori to be a severe contradiction between these two cornerstone paradigms, turns out to be a minimal one. Thus one can employ the existing MY diversification algorithm to construct the PTefficient set. This is an important step toward bridging between subjects'

1034

Chapter 9. Prospect Theory and Mean-Variance Analysis

169

Prospect Theory and Mean- Variance Analysis

observed behavior, as manifested by PT, and the standard MV portfolio theory. Appendix A: Proof of Theorem 1 Before we turn to the proof of Theorem 1, consider the following lemma, which is employed in the proof. Lemma. Assume that both F and G are normal CDFs, such that dominates G by PSD iff the following two conditions hold:

f.LF

xo

-oo

[F(z)- G(z)]dz::.;

f.LG,

o-p >

o-G

Then F

(A.1)

xo

10 xo

[G(z)- F(z)]dz,

(A.2)

where x 0 is the crossing point ofF and G (i.e., F(x 0) = G(x0), recall that two normal CDFs with different variances cross exactly once). In other words, for normal distributions, Equation (4) is equivalent to Equations (A.1) and (A 2). Proof Necessity Equation (A.1) is required for the PSD dominance ofF over G. To see this recall that for normal distributions with o-p > o-G, F crosses G from above (as in Figure 2b). If x 0 2:0 (unlike what is shown in Figure 2b), then lpsn jセ P {gHコIM F(z)]dz < 0, in violation of the PSD criterion of Equation ( 4). Equation (A.2) also follows directly from Equation ( 4) by taking .r=-oo and x = 0, because when x 0 a-G). x 0 < 0 ensures that (i) jセ{gHコIM F(z)]dz > 0 for all x > 0 (because G >Fin the positive range; see 0 Figure 2b). Equation (A.2) ensures that (ii) .fx [G(z)- F(z)]dz > 0 for all.r < 0, because this expression is minimal for .r = -oo, and Equation (A.2) states that it is positive even in this case. Taken together, (i) and (ii) imply Equation (4). •

Note that even though the PSD condition of Equation (4) requires verification for all 0::.; x, in the case of normal CDFs with f.LF > f.LG and o-p > o-G, this boils down to the verification of only the two conditions [Equations (A.1) and (A.2)]. This result can be extended to the case of any two CDFs that cross only once. Proof of Theorem 1. Let us first prove that the PSD-efficient set is a subset of the MY-efficient set [claim (a)], and then go on to derive the segment of the MY-efficient frontier which may be PSD inefficient [claim (b)].

MY-efficient set (a) PSD-efficient ウ・エセ@ The MY-efficient set is well known to have the shape of a hyperbola, which is called the efficient frontier [see Roy (1952), Merton (1972), and Roll (1977)]. We claim that any portfolio that is interior to the MY-efficient frontier (i.e., which is MY inefficient) is also PSD inefficient. To see this, consider any portfolio F' that is MY inefficient (see Figure 3). This portfolio is FSD dominated by portfolio F, which has the same standard deviation, but higher mean (see the FSD condition, [Equation (2)] for the case of normal distributions). As FSD implies dominance by all individuals with increasing utility/value functions, it also implies preference by all PT investors (FSD PSD). Hence all PT investors prefer F over

'*

1035

170

H Levy and M Levy The Review of Financial Studies I v 17 n 4 2004

F', and therefore F' is PSD inefficient. As the same argument holds for all portfolios located below the MY-efficient frontier, we conclude that all the portfolios below the MY-efficient frontier are PSD inefficient. (b) The segment of the MY frontier that may be PSD inefficient While the MY-efficient frontier coincides with the FSD-efficient set for normal distributions, some portfolios located on the M V-etlicient frontier may be relegated to the PSDinefficient set, because the PSD-efficient set is by definition a subset of the FSD-efficient set. We show below that the PSD-efficient set is the entire MY-efficient frontier excluding at most the segment between the minimum variance portfolio and the point of tangency from the origin to the frontier (segment Oa in Figure 3). To see this, consider any two portfolios F and G on the efficient frontier, and assume without loss of generality that f.LF > f.LG, O"p > O"Q. Figure A.l depicts such a situation. As mentioned above, any two cumulative normal distributions with O"p # D"G cross exactly once. Denote the crossing point ofF and G by x 0 (see Figure 2). G cannot dominate F by PSD because f.LF > f.LG (see footnote 6). By the above lemma, F dominates G by PSD iffEquations (A.l) and (A.2) hold. The condition x 0 < 0 has a direct geometric interpretation in the MY plane. For two normal distributions the intersection point x 0 of the cumulative distributions is given by the solution of the following equation: Xo- f.Lp

Xo- f.LG

O"p

O"G

(A.3)

:::t

0 Figure A.l PSD inefficiency and the MV frontier Necessary conditions for the PSD dominance ofF over G are that F has a higher mean than G, and that the slope from the origin (0,0) to F is higher than the slope to G. Thus all portfolios on the MV frontier to the right of point a are PSD efficient: for any such portfolio there is no other portfolio with both higher mean and higher slope. For example, in the figure neither F nor G dominate the other by PSD.

1036

Chapter 9. Prospect Theory and Mean-Variance Analysis

171

Prospect Theory and Mean- Variance Analysis

[see, e.g., Hanoch and Levy (1969)]. Thus x 0 is given by xo

The condition tion), or

x 0

=

< 0 therefore implies

f.LG(J"F- f.LF(J"G

f.LG(J"F- f.LF(J"G

< 0 (recall that

(A.4) (J"F

> (J"G by construc(A.5)

Thus, Equation (A.5) can replace Equation (A.l) as a condition for the PSD dominance ofF over G. g・ッュエイゥ」。ャケセ@ is the slope from the origin to portfolio F in the MV plane. Thus a necessary condition for the PSD dominance ofF over G is that the slope from the origin to F is higher than the slope toG (see FigureA.l). For the specific F and G depicted in Figure A.l there is no PSD dominance, because the slope ofF is smaller than the slope ofG (and G also cannot dominate because f.LF > f.La). In contrast, F may dominate portfolio G' (see Figure A.l), because F has both higher mean and higher slope [however, this dominance is not guaranteed because one still has to check whether Equation (A.2) holds]. In general, all the portfolios on the Markowitz efficient frontier to the right of the tangency point from the origin (point a in Figure A.l) are PSD efficient. Consider, for example, portfolio F: this portfolio is not dominated by any portfolio to its left, because F has a higher mean. It is also not dominated by any portfolio to its right, because F has a higher slope. Thus portfolio F, and all other portfolios to the right of point a, are PSD efficient. PSD relegates to the PSD • inefficient set at most segment Oa of the efficient frontier.

Appendix B: The PSD- and MY-Efficient Sets in the Case of Lognormal Distributions We show in this appendix that when the distributions are lognormal, the FSD- (and hence the PSD-) efficient set is a subset of the MY-efficient set, despite of the fact that the lognormal distributions are skewed. Let us elaborate. In the text we analyze the relationship between MV analysis and PT for the case of normal distributions. It is well known, however, that for long investment horizons the return distributions are positively skewed, and the lognormal distribution is a better approximation for the empirical return distribution. A drawback of the lognormal distribution framework is that the sum of two random variables, which are lognormally distributed, is not precisely lognormal. However, Lintner (1972) shows that the sum of two such lognormal random variables is almost lognormal. He concludes, "the approximation to lognormally distributed portfolio outcomes of lognormally distributed stocks is sufficiently good that theoretical models based on these twin premises should be useful in a wide range of applications and empirical investigations." [See also Ohlson and Ziemba (1976), and Dexter, Yu, and Ziemba (1980)]. In Theorem 3 we extend our main theoretical results given in the text to the case of lognormal distributions. Theorem 3. Consider n assets with lognormal return distributions and correlations IRyl < 1 for all i, j. Portfolios can be formed without restrictions, and all portfolio returns are also lognormally distributed. Objective probabilities may be subjectively distorted by any transformation which does not violate FSD. Then, the PSD-efficient set is a subset of the MVefficient frontier, excluding at least the segment between the minimum variance portfolio and the point of tangency from coordinate (0, -1) in the MV plane to the frontier (segment Oa' in Figure A.2).

1037

172

H Levy and M Levy The Review of Financial Studies I v 17 n 4 2004

0.0 /

/

-1.00

Moreover,

UQ"D(b) is dense in D(b).

The interpretation of the Theorem is evident in light of the preceding discussion. (The difference between M(D(b)) and M(D(b)) is clarified in Appendix 1 following the proof of the theorem.) 3. RECURSIVE UTILITY

All utility functions are defined on D( b) and are recursive there. It may be helpful, therefore, to begin by considering briefly the structure of recursive utility over deterministic consumption streams. If V is such a utility function, then (Koopmans (1960)) in the obvious notation (3.1)

V(c 0 ,c1 , ••• ) = W(c 0 , V(c1 ,c 2 , ••. )),

for some function W. This structure has been explored also by Lucas and Stokey (1984) and Boyd (1987) where W is termed an aggregator, as it combines current consumption and future utility to determine current utility. In contemplating an extension of (3.1) to the stochastic case we note that future utility is random. It seems natural, in that case, to compute a certainty equivalent for random future utility and then to combine the certainty equivalent utility level with c0 via an aggregator. Thus we are led to consider certainty equivalent (or generalized mean value) functionals p,. Each such mean value is a map, p,: dom p, c.;;;, M(R+)-+ R+,

which is consistent with first and second degree stochastic dominance and satisfies

(3.2)

p,(8x)=x 'VX ER+,

i.e., if a gamble yields the outcome x with certainty, then x is the certainty equivalent of the gamble. Given a utility function V: D(b)-+ R+ and (c 0 , m) E D(b), denote by V[m] the probability measure for future utility implied by V and m E M( D( b)) c

Chapter 12. Substitution, Risk Aversion and the Temporal Behavior of Consumption TEMPORAL BEHAVIOR OF CONSUMPTION

215

945

temporal lottery d, rna and mb eM(D)

temporal lottery d,

ma ana mbe M(D)

Consistency:

i=a,b

=> d>; d FIGURE

2

M(D(b)), i.e., (3.3)

V[m](Q)=m{dED: V(d)EQ},

QEB(R+).

The utility function V is called recursive if it satisfies the following equation on its domain:

for some increasing aggregator function W: Ri-+ R+ and some certainty equivalent IL· This relation is the cornerstone of our analysis. Of course, it generalizes the more familiar structure (3.1). Note also that the recursive structure immediately implies the intertemporal consistency of preference (in the sense of Johnsen and Donaldson (1985) or Figure 2) and the stationarity of preference (in the sense of Koopmans (1960), for example).

216

LG Epstein and SE Zin

946

LARRY G. EPSTEIN AND STANLEY E. ZIN

The question which immediately arises is whether and under what circumstances there exist utility functions V satisfying (3.4). 3 To answer this question we restrict the admissible aggregators and certainty equivalents. First, we require that W have the CES form (3.5)

W(c,z)=[cP+,BzP] 11 P,

O*p is continuous, increasing, concave, and f/>(1) = 0. Note that if (3.12)

4J(x)=(xa-1)ja,

O*a are possible. For example, one that is investigated empirically in Epstein and Zin (1989) is (3.13)

f/>(x)=(xa-1)/a+a(x-1),

O*a that -xf/>"(x)/f/>'(x) is (strictly) decreasing. This restriction can be readily incorporated into specifications for 4>, e.g., it is satisfied by (3.13) if a> 0. The above examples do not exhaust the class of recursive intertemporal utility functions covered by Theorem 3.1. Other specifications for J.L, taken from the atemporal non-expected utility literature for example, could be adopted if the seemingly mild continuity condition MV.1 is satisfied. Thus Theorem 3.1 should permit the integration into a temporal setting of a substantial portion of the non-expected utility literature. 4. SUBSTITUTION, RISK AVERSION, AND TIMING

The key properties of recursive utility functionals will be discussed here. It has already been noted that the specification (3.5) for the aggregator implies that deterministic consumption sequences are ranked by an intertemporal CES utility function with elasticity of substitution o = (1 - p) -l. Thus we interpret p as a parameter reflecting substitutability. Next tum to risk aversion and in particular to comparative risk aversion. Let V and V* be two recursive utility functions with possibly distinct aggregators W and W* conforming to (3.5). 7 We wish to define what it means for V* to be more risk averse than V. To do that, define c = 'A(m) for any (c 0 , m) by V(c 0 , m) = V(c 0 , c, c, ... ). 7 The ensuing discussion could be carried out without restricting aggregators to conform with (3.5). But such generality would complicate the exposition somewhat for the following reason: If V is functional p., then any monotouic transform of recursive with aggregator W and certainty ・アオゥセ⦅A@ V, say h(V), is also recursive with aggregator W, W(c, z) = hW(c, h- 1(z)) and certainty equivalent functional Ji., Ji.(Px) = h( u(Ph-'(x))}. Thus a given preference ordering of temporal lotteries can be represented by many ( W, p.) pairs. By fixing the representation (3.5) for the aggregator, we avoid the need to refer to the entire class of "equivalent" ( W, p.) pairs.

220

LG Epstein and SE Zin

950

LARRY G. EPSTEIN AND STANLEY E. ZIN

Interpret the "nearly" constant and deterministic path ( c0 , c, c, ... ) which is indifferent to ( c0 , m) as the certainty equivalent for the latter. (Note that c depends only on m and not on c0 .) The path ( c0 , A*( m ), A*( m ), ... ) is defined analogously given V*. It is natural to say that V* is more risk averse than V if and only if A*(m) セ@ A(m) for all (c0 , m) in some common domain for V and V*. Evidently, if V and V* are comparable in the above sense, then they must rank nonstochastic consumption programs identically, i.e., W = W* or equivalently, p = p* and /3 = /3*. Moreover, V* is more risk averse than V if and only if W= W* and (4.1)

ーNJHᄋIセ@

on the appropriate domain. (Necessity of these conditions is obvious. For sufficiency, note that they imply, given the construction in the proof of Theorem 3.1, that V*( ·) セ@ V( · ).) Thus the certainty equivalent functional determines the degree of risk aversion of the corresponding intertemporal utility function, at least for comparative purposes. Further support for this interpretation of p. is provided at the end of Section 5. Since, by assumption, mean value functionals p. * exhibit risk aversion in the sense of second degree stochastic dominance, it follows that ーNJHᄋIセeL@

where E( ·) denotes the expected value operator. Thus the least risk averse intertemporal utility function is the one for which p.( ·) = E( ·).Moreover, there is a sense in which the latter specification implies risk neutrality, e.g., in the context of timeless wealth gambles or in the portfolio choice context described at the end of Section 5. It is apparent, therefore, that "low" or "moderate" risk aversion can coexist with a small elasticity of substitution, which is impossible in the expected utility specification. a. In the case of KP functionals (3.8), the condition (4.1) is equivalent to 。Jセ@ Thus we interpret a as a measure of risk aversion for comparative purposes with smaller a's indicating greater risk aversion. A separation between the risk aversion parameter (a) and the substitution parameter ( p) is achieved. For the Chew-Dekel class based on (3.11), (4.1) is equivalent to

4>*"(. )/4>*'(-)

セ@

4>"(. )/4>'(.).

a and 。Jセ@ a. This is satisfied in the parametric class defined by (3.13) if 。Jセ@ A comparable separation between risk aversion and substitution does not appear possible within the expected utility model. To see this, consider consumption programs (c 0, m), with mE M(Y(b; /)) c M(R"':.) and let

V( c0 , m) =Em{ cgjp +

セOS Q 」ヲ@

/p},

which is ordinally equivalent to (3.6). The general multicommodity analysis of Kihlstrom and Mirman (1974) suggests that we take a monotonic transform h of

Chapter 12. Substitution, Risk Aversion and the Temporal Behavior of Consumption TEMPORAL BEHAVIOR OF CONSUMPTION

221

951

the von Neumann-Morgenstem utility index and define

Then V* is more risk averse than V in the sense defined above, or equivalently in the sense of Kihlstrom and Mirman (1974), if and only if h is concave. At first glance, therefore, it would seem that comparative risk aversion analysis is feasible within an expected utility framework. But in a temporal setting, this familiar approach encounters serious difficulties. To see this, consider an individual with the utility function (4.2) who arrives at period T and contemplates the remaining future. If past consumption levels were c0 , ... , Cr_ 1, then the utility function for the remaining future is presumably 8

Thus the preference ordering at T depends upon past consumption values unless h has constant absolute risk aversion. Dependence upon the past is in principle sensible but the form which this dependence takes above is implausible, since (4.3) and 0 < {3 < 1 imply that the dependence on past consumption is greater as the past becomes more distant. (For example, denote by H( c0 , ... , Cr_ 1 ; cr, cr+ 1, ... ) the von Neumann-Morgenstern utility index on the right side of (4.3). If derivatives are evaluated at a point where c0 = · · · = cr_ 1, then

Thus the risk premium for a small gamble in period (T + 1) consumption is affected more by a small change in c0 than by a small change in Cr_ 1.) On the other hand, if h(z) = -exp( -Az), A> 0, then the nonstationarity of preferences is implied as period T preferences are represented by vr( cr, m) = Em exp(- Afjru( Cr, Crn ... )), where u is the additive functional f. 0{3 1cfl p. Declining risk aversion with T is imposed a priori. Though such a "changing tastes" specification may be appropriate in some modelling exercises, it would appear to be a hindrance rather than a help to exploring the questions outlined in the introduction and in Section 7. More importantly, if attitudes towards future gambles are changing with the passage of time as above, then plans will generally 8 Otherwise tastes are changing through time. This is unappealing as an a priori specification and, moreover, implies that preferences are intertemporally inconsistent. Hall's (1981) specification suffers from these problems.

222

LG Epstein and SE Zin

952

LARRY G. EPSTEIN AND STANLEY E. ZIN d

d, eED

early resolution

d

(co, '\c 1,m))

m =nod+ (1-n)oe e M(D)

late resolution FIGURE

3

not be intertemporally consistent. The above unappealing features of the familiar Kihlstrom and Mirman approach to comparative risk aversion are not restricted to the case where the von Neumann-Morgenstem utility index is additively separable. For example, they may be confirmed also for the nonadditive indices axiomatized in Epstein (1983) which feature variable discount rates. Finally, with regard to attitudes towards the timing of the resolution of uncertainty, see Figure 3. The two temporal lotteries portrayed there differ precisely in the timing of the resolution of uncertainty as defined by KP (1978). A recursive utility function is indifferent to the timing of resolution (in all such pairs of lotteries) if and only if it is an expected utility functional such as (3.6). (This follows by a straightforward extension of the finite horizon arguments in Chew and Epstein (1989).) In particular, for the KP class, the curvature of H( c0 , ·) defined in (3.9) is the determinant of attitudes towards timing with indifference towards timing prevailing only if H(c 0 , ·)is linear (KP (1978)). We can conclude, in fact, that given (3.8) early (late) resolution is preferred if a )p. For more general recursive utility functions, we have not found a characterization in terms of W and p. of the conditions under which early or late resolution is preferred. But the characterization for the KP class raises an issue which we suspect is relevant more generally and which calls for some attention. We have interpreted a as a risk aversion parameter. But with p fixed, a reduction in a not only increases risk aversion but also may transform a preference for late resolution into a preference for early resolution. One is left wondering how to interpret the comparative statics effects of a change in a. Similarly, a change in p for given a affects both substitutability and attitudes towards timing. Thus the latter aspect of preference seems intertwined with both substitutability and risk aversion.

Chapter 12. Substitution, Risk Aversion and the Temporal Behavior of Consumption TEMPORAL BEHAVIOR OF CONSUMPTION

223

953

We offer three comments in response. First, from the perspective of potential empirical applications, the specifications (3.8) and a fortiori (3.4) are still more flexible than the common expected utility functional form. Second, the behavioral analysis in the next section will provide further support for our interpretation of a, or more generally p., as a risk aversion parameter since a or p. will determine the degree of risk taking in certain portfolio choice problems. Finally, we suspect that the lack of separation noted above reflects the inherent inseparability of these three aspects of preference rather than a deficiency of our theoretical framework. Further study of this issue is required. To conclude this section, we observe that attitudes towards timing can be used to distinguish, within the family of recursive utility functions, each of the three subclasses defined in Section 3. It has already been pointed out that timing indifference implies an expected utility ordering. Next consider the Chew-Dekel subclass. Suppose that V is such that the lotteries in Figure 3 are indifferent to one another whenever V( d) = V( e); that is, the timing of resolution is a matter of indifference if the two future prospects regarding which information is being provided, are themselves indifferent. Refer to this property as quasi-timing indifference (QTI). A straightforward extension of the finite horizon arguments in Chew and Epstein (1989) shows that the only recursive utility functions satisfying QTI are those based on (3.10). If an appropriate homotheticity assumption is imposed on V, then p. must satisfy MV.2 and (3.11) is obtained. (See Chew and Epstein (1989) for the basis for a comparable argument for the KP class (3.8) and for discussion of QTI. An alternative basis for an axiomatization of KP preferences may be found in KP (1978).) Thus a theoretical case can be made for interest in the KP and Chew-Dekel subclasses of recursive utility functions. Accordingly, we do not apologize for the fact that some of the discussion of the asset pricing implications of our framework is limited to these subclasses. 5. THE REPRESENTATIVE AGENT

The remainder of this paper derives relations between aggregate consumption and real rates of return which must hold in a competitive equilibrium. The procedure adopted is that of the rational expectations literature on aggregate consumption (Hall (1978)). In this section we determine the optimal consumption and portfolio behavior of an individual who faces exogenous rates of return to saving. Then, in the next section, we take the individual to be a representative agent in the economy so that the Euler equations corresponding to his intertemporal plan define relations between aggregate consumption and rates of return that must hold in equilibrium. We deviate from earlier literature in the specification of a recursive (but not necessarily expected utility) specification for preferences. Consequently, the derivation of the Euler equations is nonstandard. Our representative agent operates in a standard environment. There are K assets. The gross return to holding the kth asset between t and (t + 1) is described by the random variable rkl' - 00 < t < 00, where each rkt has support

224

LG Epstein and SE Zin

954

LARRY G. EPSTEIN AND STANLEY E. ZIN

"=

in [r, r], r > 0. Let (111 , ••• ,fK1 ). We assume that (", コ Q IGセ@ oo is a stationary stochastic process. The role of z 1 E R 2 is to provide information regarding the future. It is assumed that ( "' 1 ) is observed at the start of period ( t + 1), just prior to the time at which period (t + 1) consumption and portfolio decisions are made. Without loss of generality we may take the underlying probability space to be (fl, B(fl), >¥) where fl is the set of doubly infinite sequences with tth component (rP z 1 ) and B(fl) is the (product) Borel a-algebra. The state of the world at time t is defined by the existing wealth level x 1 and by the history of realized past values of ( r;, z;)' s. Thus let

z

t-l

fl 1 =[0,oo)x

X ([r,r]KxR 2 )=[0,oo)xii' i= -oo

where the (product) Borel a-algebra is adopted for I 1 and fl 1 . A consumptionportfolio plan (from t = 0 onwards) is a sequence (h 0 , ••• , hn ... ) of measurable functions ht: [lt セ@ [0, oo) X sK, where sK is the unit simplex in RK. The interpretation of ht(xl' It)= (ct, w1 ) is that given period t wealth xt and history It E It, the agent consumes ct and invests the proportion wkt in the k th asset, wt = ( wln ... ' wKt). A plan is homogeneous if "iit>O and "ii(x,It)Eflt,hl1,It)=(cl'wJ=> h l x, It) = ( c1 x, wt ). Because of the homotheticity of preferences and the linearity of "technology," it is natural to restrict oneself to homogeneous plans which are henceforth simply plans. A plan is stationary if 3h such that ht = h "iit. Finally, a plan is feasible if "iit セ@ 0 and "ii(xf' I 1 ) E fln c1 セ@ xt where ct is the first component of ht(xn IJ and where wealth evolves according to (5.1)

X1

= (x 1 _ 1 -

C1 _ 1

)w/_ 1"_ 1 ,

t セ@

1,

x 0 > 0 given. 9 Each plan implies an infinite probability tree in consumption levels. Moreover, if the plan is feasible, then the corresponding probability tree itself corresponds to a temporal lottery in D(r; x 0 ) c D(r). Formally, denote by FP the set of feasible plans for a given (x 0 , I 0 ). Appendix 4 shows that FP can be embedded in a "natural fashion" in D(r) by the map e. Moreover, D(r) is a subspace of D(b) for any b > r. Thus, if Vis defined on D(b), the problem (5.2)

J(I0 , x 0 )

=sup{V(d): dE e(FP)}

is well-defined if r < b. The conditions under which the supremum is attained are specified in the next theorem. THEOREM 5.1: Let V be the recursive utility function constructed in Theorem 3.1, defined on D( b) and having aggregator (3.5) and a mean value functional satisfying 9 This budget constraint excludes exogenous sources of income and labor income. This exclusion is important for the homogeneity property (5.4) below and subsequently for our derivation of Euler equations. Some discussion of the modifications necessary to accommodate these sources of income may be found in Epstein and Zin (1989).

Chapter 12. Substitution, Risk Aversion and the Temporal Behavior of Consumption TEMPORAL BEHAVIOR OF CONSUMPTION

225

955

MV.1 and MV.2. Then (5.2) possesses a maximum, achieved by a stationary and homogeneous plan if (a) p > 0, and r < b, or (b) p < 0 and f3rP < 1. In either case, J>O.

Note that the homogeneity of JL (MV.2) has been added as an assumption. Also note that the strict positivity of the value function of J is used below. We now turn to the implications of optimality and more particularly to the appropriate set of Euler equations for interior optima. The recursive structure of utility functions immediately implies the "Bellman equation" (5.3)

J(Io,Xo)=

max c0 ;;.0, "'JES

Jct;+f3JLP[PJ(i,,(x 0 -c0 )w0r0 )/lolr/p'

where the argument of JL is the probability measure for jHセLク P M c0 )w0r0 ) conditional on 10 . Moreover, the maximizing values of c0 and w0 correspond to the utility maximizing plan in the customary fashion. It is evident from the homotheticity of utility, that J can be expressed in the form

(5.4)

J(I,x)=A{I)x.

Thus (5.3) can be written (5.5)

A{I0 )x 0 =

max 」 P セP@

Jct;+f3(x 0 -c0 )PJLp[PA(i,)woro/Io]r 1P.

w0 ES

An immediate implication is the portfolio separation property and more particularly that the portfolio decision is determined by the solution to

(5.6)

maxKJL [PA(i,)w6ro/Io] · w0 ES

Write c6 = a 0 x 0 , where an asterisk denotes the maximizing value. Substitution into (5.5) yields

AP{I0 ) =at;+ /3(1- a 0 )PJL *P, and the first order condition for consumption in (5.5) yields

ag- 1 =

(5.7)

(1- a 0 )p- 1f3JL*P.

These last two equations can be combined to yield A(I0 ) = a&p- 1)/p = (c6 jx 0)-'l/> ·(F,- ;,)jfolセ@

0,

k=2, ... ,K.

=

10 LetS be a set of probability measures and J.L* max{J.L(p): pES} where J.L is given by (3.10). Then J.L* =J.L(p*) =>p* solves max(EpF(·,J.L*): pES}. See Epstein (1986, Section 3).

228

LG Epstein and SE Zin

958

LARRY G. EPSTEIN AND STANLEY E. ZIN

Also, substitute (5.9) into (3.11) to deduce

Equations (6.2) and (6.3) constitute the Euler equations for the model based on Chew-Dekel preferences. The former reveal more clearly than (5.10) the joint role played by consumption and the market return in determining systematic risk. A further specialization to KP preferences is obtained by imposing (3.12). The resulting Euler equations take the form (k = 2, ... , K),

( 6.4)

Alternatively, these equations are equivalent to (6.6)

- ]a(p-1)/p /] f3EP!a [[ :: MJ/'-p)/Pfko Io = 1

(k=1, ... ,K).

If the further specialization a = p is adopted, then we obtain

(6.7)

(k=1, ... ,K),

which are the familiar Euler equations of the expected utility model (see Hansen and Singleton (1983), for example). Earlier we pointed out that one consequence of the generalization from expected utility to recursive utility is the emergence of the market return as a factor in explaining excess mean returns. The significance of the market return is apparent from (6.4), which can be rewritten

Thus both consumption and the market return enter into the covariance that defines systematic risk. The consumption-CAPM is obtained if o: = p. But if the substitution o: = 0 is adopted instead, then covariance with the market return

Chapter 12. Substitution, Risk Aversion and the Temporal Behavior of Consumption TEMPORAL BEHAVIOR OF CONSUMPTION

229

959

alone determines systematic risk which is the prediction of the static CAPM. 11 (Though a = 0 lies outside the scope of our formal analysis, one can show that the a = 0 version of (6.4) is the model of mean excess returns that corresponds to the a= 0 version of KP preferences.) Another specialization of (6.4) which yields the market based CAPM model is p = 1, or infinite elasticity of substitution. This case was excluded from the preceding analysis because it would generally rule out interior optima for consumption (see (5.5)). But the appropriate form of (6.4) is still valid. Intuitively, the emergence of the static CAPM here is presumably due to the perfect substitutability of consumption across time. We have described some relations between aggregate consumption and asset returns which must hold in a competitive equilibrium, but we have not demonstrated the consistency of our analysis with a general equilibrium framework such as Lucas' (1978) stochastic pure endowment economy. Such an extension would need to confront the questions of existence and uniqueness of equilibrium asset prices. Moreover, Lucas' contraction mapping techniques would not suffice for the same reasons that those techniques were inadequate in establishing Theorem 3.1. Thus we leave such an extension to a separate paper. However, we conclude this section by describing some asset pricing implications of our analysis which are valid for any general equilibrium extension. For simplicity, consider KP preferences, though comparable formulae may be derived for Chew-Dekel preferences. Consider an asset which pays the dividend ij1 in period t. The real gross return to holding the asset during period 0 is + ij1 ) /P0 , where P0 and i\ denote the current and random period 1 prices respectively. Then substitution into (6.6) implies that the asset price satisfies the recursive relation

0\

which has a solution

if the right-hand side is finite. Price equals the discounted expected value of 11 The p = 0 version of the expected utility model, i.e., logarithmic within period utility function, leads to E[M0 1(i'ko- i'10 )jl0 ] = 0 Vk, a CAPM model. But then it is also true that

so that the consumption-CAPM also applies. In contrast, our specialization a = 0 implies that only the market return determines systematic risk. A related observation is that we generate a CAPM model without restricting the intertemporal elasticity of substitution in consumption. See Grossman and Laroque (1987) for an alternative intertemporal model which leads to the static CAPM.

230

LG Epstein and SE Zin

960

LARRY G. EPSTEIN AND STANLEY E. ZIN

future dividends where the discount factors involve both consumption and market returns. In the case of a one-period pure discount bond, ij1 = 1 and ij1 = 0 for t > 1,

Given fixed marginal distributions for i\ and M0 , the bond price increases with the covariance between fi(p-l)/p and M/)"--P)IP. The two exponents have the same signs if and only if 0 < a < p or p < a < 0. In those cases, the bond price increases as consumption c1 and the market return M0 become more correlated. 12 Otherwise, the bond price falls in response to an increase in correlation. 7. CONCLUDING REMARKS

The intertemporal utility functions we have formulated have three very appealing features: (1) intertemporal substitution and risk aversion are disentangled; (2) they integrate atemporal non-expected utility theories into a temporal framework; and (3) they generate implications for the temporal behavior of consumption and asset returns. Moreover, these implications may be investigated empirically by existing econometric techniques as demonstrated in Epstein and Zin (1989). Some empirical work is done in the latter paper but further empirical investigation, exploring alternative data sets and functional forms, would be worthwhile. A promising application on the theoretical front is to recursive dynamic GE modelling (Sargent (1987)). For example, we have already mentioned the need to integrate our model into a general equilibrium framework such as Lucas' (1978). Because of the inseparability of substitution and risk aversion in his expected utility model, Lucas is unable to provide a clear interpretation for some of his comparative statics results. Our utility functions should clarify those results and thus provide a clearer understanding of the determinants of asset prices. (See Epstein (1988) for such a comparative statics analysis in a stochastic pure endowment economy where endowments are i.i.d.). In addition, the separation which they provide should make them useful in exploring the role played by differences in risk aversion in influencing the distribution of wealth across agents. Such an investigation would complement existing theories of distribution that are based on differences in time preference (Epstein (1987)). In these and other theoretical developments, the specific functional form (3.5) for the aggregator could be useful in providing some initial insights. Indeed many of the early multiperiod expected utility models of consumption/portfolio behavior are based on the homogeneous parameterization (3.6). But it would clearly be desirable to generalize to a larger class of recursive utility functions. It is hoped that several elements of this paper, such as the proofs of Theorems 3.1 and 5.1 12 A

notion of "greater correlation" that-suffices here is described in Epstein and Tanny (1980)_

Chapter 12. Substitution, Risk Aversion and the Temporal Behavior of Consumption TEMPORAL BEHAVIOR OF CONSUMPTION

231

961

and especially our formalization of the space of infinite horizon temporal lotteries, will be useful in developing such generalizations. Department of Economics, University of Toronto, Toronto, Ontario, Canada M5S JAJ and Graduate School of Industrial Administration, Carnegie-Mellon University, Pittsburgh, PA 15213, U.S.A. Manuscript received July, 1987; final revision received September, 1988.

APPENDIX I In this appendix we prove Theorems 2.1 and 2.2. The notation of Section 2 is adopted. LEMMA ALl: The function f defined in (2. 7) is continuous.

PROOF: By Parthasarathy (1967, Theorem 6.1, p. 40) it suffices to show that m.--+ m

=

f(m.)(B)-+ f(m)(B)

'T/ BE 8( R";) such that j(m)( a B)= 0. But the latter condition implies v { y E R";: ( c0 , y) E aB}

= 0 a.e. [m] and hence (Billingsley (1968, Theorem 5.5)) that T0 is continuous a.e. [m]. By Billingsley (1968, Theorem 5.1), Em,,Tn-> EmTB as desired. Q.E.D.

LEMMA AL2: For all t;;. 1 the functions ft and g, defined in (2.8) are continuous and satisfy (2.9).

PROOF: Continuity follows from Lemma Au and Billingsley (1968, p. 29). Condition (2.9) is readily verified. Q.E.D. Recall the projection maps .,, defined in Section 2. LEMMA A1.3: (a) For each t ;;.1 and B, E 8" .,,- 1(B,) t;;. L {c) Uj"'.,,- 1(81 ) is an algebra. (d) 8 = HuヲGNLMッ ur.,,-1 d, then m;-> m, 'T/t;;. 1 and so by (ALl) m"(A)-> m(A) for any A EUj"'.,,- 1(8,) such that m(aA)=O.

232

LG Epstein and SE Zin

962

LARRY G. EPSTEIN AND STANLEY E. ZIN

But the algebraUi.,.- 1(B,) is convergence determining in the sense of Billingsley (1968, exercise 7, p. 22) and thus m"-> min M(D). This proves that e is a homeomorphism. D is a metric space by construction. It is also a closed subspace of the separable space Xj'D, by the continuity of the g,'s. Thus D is separable. Finally, the desired denseness is immediate. Q.E.D. PROOF OF THEOREM 2.2: One can show using Billingsley (1968, Theorems 6.2 and 6.4) that D(b; l) from (2.5) is a compact metric space and hence separable. Moreover, D(b) =U 1D(b; l) and so D(b) is separable. It remains to prove the asserted homeomorphism. Let be the map constructed in the proof of Theorem 2.1. We need to show that

e

B(D(b))

セ@

R+X M(D(b)),

or equivalently, that mE M(D) defined in (ALl) satisfies m(D(b)) = 1. Let d = (d1 , ... , d,, . .. ) E D(b), d, = (c0 , m 1 ) 'Vt. It suffices to show that

(A1.2) Since dE D( b), we know that

(A1.3) Prove the t

=

1 version of (A1.2), i.e.,

(A1.4) Of course, m 2 such that

E

M(R+x M(R";)) is given. By (A1.3) and the definition (2.7) of j, there exists ャセ@

f(m 2 )(Y(b;l))=1

0

=>

v{yER";:(c,y)EY(b;/)}=1 a.e.[m 2 ]

=>

m2 {(c,v)ER+XM(R";): v{yER";: (c,y)E Y(b;l)} =1} =1

=>

m2 {(c,v)ER+XM(R";):v(Y(b;bl))=1}=1·= m 2 (R+xM(Y(b;bl)))=1

=>

(A.1.4).

In a similar fashion (A1.2) can be proven by induction. Finally, note that the subspace M( D( b)) in the statement of the theorem satisfies

B(D(b)) = R+x M(D(b)).

Q.E.D.

To clarify the difference between M(D(b)) and M(D(b)), consider the following: Let D0 (b) = {(co, y): Co> 0, y E u/>0 Y(b; l)}, which qm be identified with a セオ「ウー。」・@ of D(b). Then a given mE M(D0 (b)) c M(D(b)) will 。ャセッ@ lie in M(D(b)) if and only if 3[> 0 for which the support of m lies in Y(b; l), i.e., mE M(Y(b; l)). That will not generally be the case for the same reason that U 1 > 0 M(Y(b; l)) is a proper subset of M(U 1 > 0 Y(b; l)).

APPENDIX2

Two lemmas are provided here. The first is used in the proof of Theorem 3.1 and the second establishes that the Chew-Dekel mean value (3.10) satisfies MV.l. LEMMA A2.1: Let X be a metric space and m n X-> R u.s.c. and bounded above.

->

m in M( X). Then lim sup ff dm n

.;;

ff dm for all f:

Chapter 12. Substitution, Risk Aversion and the Temporal Behavior of Consumption

233

963

TEMPORAL BEHAVIOR OF CONSUMPTION

PROOF: When f is bounded below, see Billingsley (1968, exercise 7, p. 17). In general, for each integer K < 0, define fK(x) =max{f(x),

K}.

Then fK is bounded and u.s.c. Thus (by above) q,K; M(X)-+ R is u.s.c. where q,K(p) fK(x) hf(x), (p) = ffdp = inf q,K(p) and so is u.s.c.

= ffK dp. Since Q.E.D.

LEMMA A2.2: The implicit weighted functional p. in (3.10) satisfies MV.l. PROOF: (a) Let P.n =p.(p"), p.=p.(p). The p./s and p. all lie in [0, a]. Suppose (for some subsequence) P.n-+ 8. Then since F is uniformly continuous on [0, 1] 2 ,

Since F( x, z) is increasing in x,

It follows that 0=/F(·,p.")dpn-+fF(·,o)dp, which implies 0=p.n-+fF(·,8)dp. But then 8=p. since F(x, z) is decreasing in z. Similarly for (b). Q.E.D.

APPENDIX3 We prove Theorem 3.1 regarding the existence of recursive utility functions. Denote by s+ ( D( b)) the set of functions from D( b) into R +. Let h E s+ ( D( b)) be strictly positive and define sh+(D(b)) = { vE s+(D(b)): llvllh

=supv(d)/h(d) < oo }.

s,; (

s,; (

With the norm ll·llh, D( b)) is a complete metric space. A transformation T: D( b)) -+ sh+(D(b)) is a strict contraction if IITv- Tullh セ@ euv- ullh withe< 1. Every strict contraction on a complete metric space has a unique fixed point. The following is an immediate corollary which is adapted from Boyd. WEIGHTED CONTRACTION MAPPING THEOREM (WCMl): LetT: Sh+(D(b))-+ s+(D(b)) be such T(u) + E>Ah that (WCM.1) u セ@ v => T(u) セ@ T(v); (WCM.2) T(O) E sh+ (D); (WCM.3) T(u +A ィIセ@ for some constant e < 1 and 'ItA > 0. Then T has a unique fixed point v*. Moreover, TNO-+ v* in sh+ (D). We are able to base the following proof on WCMT in the KP case when a and p have identical signs. An advantage of such a proof is that it leads to the uniqueness of the solution to (3.4) and also to stronger continuity properties for V than described in the theorem. Moreover, it facilitates the proof of existence of optimal plans. But we could not find a way to apply WCMT to the remaining cases for the KP functional or to more general specifications for p.. Thus we use it below only where absolutely necessary, namely in Case 1, and otherwise we present a shorter, simpler argument. PROOF OF THEOREM 3.1: CASE 1: KP functional, 0 < p セ。]@ by h(d)

= [ 1 + c0 +

eュ Q セxGH@

c,/b')

r

1, {fbP < 1. Define h: D(b)-+ R++

where X E (0,1) is to be determined below and d= (c0 , m), m 1 = f(P2 m). (See (2.3) or (2.7) and (2.11); m 1 is the "atemporal" probability measure on future consumption induced by m.) The above expected value exists since 'i,'f)N(c,/b') is bounded on each Y(b; l) and since m1 EU1 > 0 M(Y(b; !)).

234

LG Epstein and SE Zin

964

LARRY G. EPSTEIN AND STANLEY E. ZIN

Rewrite (3.8) in terms of U = VP j

(A3.1)

cg

H(c 0 ,z)=-+{3z, p

p

as follows:

y=ljp;;.l.

Use (A3.1) to define the transformation T: sh+(D(b))-+ s+(D(b)), where

(Note that [ EmuY( · )] 1/y.;;; llullh · [ EmhY( · )] 1/y

if u E sh+ ( D( b)). Moreover, 00

EmhY( ·)

=

1 + c0

+ Em1 LXc,jb' 1

which is finite for the same reasons provided above for the finiteness of h(d).) We wish to prove that T has a fixed point and so we verify the conditions of WCMT. The first two conditions are immediate. For the third, note that T(u +Ah)(c0 , m)

=

H( c0 , [ Em(u( ·) +Ah(·))Y] 1

.;;; H( co, ( EmuY( · )j 11Y)

h)

+ f3A(EmhY( ·)flY (by Minkowski's inequality and y;;, 1)

=Tu(c0 ,m)+f3A[Emg(-)fh

(g=hY)

bP

=

Tu( c0 , m)

+ f3 ;\_P A(g( c0 , m)] 1h

( g ( c, ·) satisfies the independence axiom on M ( D (b))) =

Tu( c, m)

+ f3bPAh( c0 , m)jV.

Thus WCM.3 is satisfied with e = f3bP jV if ;\. is any number such that {3bP < V. By WCMT, rNo-+ U= VP jp in the ll·llh topology on st(D(b)). Moreover, his bounded on any D(b; /). Thus sup { iTNO(d)- U(d)l: dE D(b; /)}-+ 0 as N-+ oo. Each TNO( ·) is continuous on D( b; /). Thus U and V are continuous there. It can be shown that D(b; /)is compact. Thus (A3.2)

max{V(d): dED(b;l)} < oo,

a fact which is used below. CASE 2: General p., p > 0, {3bP < 1. Refer to the previous case and let v• be the corresponding utility functional which satisfies (A3.3) Define T: sh+ ( D( b)) -+ Tv (Co ' m)

s+ ( D( b)) by =

w( Co' IL ( v [ m])) .

(See the notation introduced in (3.3).) Then TNv•;;, 0 so the sequence {TNv*(c0 , m)} is bounded

Chapter 12. Substitution, Risk Aversion and the Temporal Behavior of Consumption

235

965

TEMPORAL BEHAVIOR OF CONSUMPTION

below. It is also declining in N:

W( c0 , Emv* ( ·)) (since p. is risk averse)

Tv* ( c0 , m) .,;

=v*(c0 ,m)

(by(A3.3)).

By induction, TNv*(c 0 , m).,; TN- 1v*(c0 , m) Gvnセ@ It remains to show that V solves (3.4 ). Since TNv•( Co, m)

=

1. Define V(c0 , m) =lim TNv*(c0 , m).

w( Co, JL( TN-lv•[m l))

and the left side converges to V(c0 , m), it suffices to show that JL(TN- 1v*[ml) ...... JL(V[m]).

But this is true by (A2.2), MV.1, and the monotone convergence theorem (Billingsley (1979, p. 179)), the latter of which implies

JI ( T"- v* (-)) dm (') ...... JI ( V (')) dm (') 1

for all increasing

f.

Since u* was shown to be continuous, on D( b; I) each T"v* is continuous there by induction. Thus, as the infimum of continuous functions, V is u.s.c. CASE 3: General JL, p < 0. Define the sequence { u"} of nonpositive extended real valued functions on D( oo) inductively: u1 ( c0 , m) = c81P.

u"(c0 ,m)=c81p+

HiセpIjlーオBG



IQOp{ュjL@

n;;.2.

(Recall that (pu" -I ) 11P[m] denotes the distribution of ( pu"- 1 ( · )) 1/P induced by m. Define JL = oo if there is positive probability that (pu"- 1 ) 1/P = oo .) Then u".,; u"- I So the extended real valued function U, U(d) =lim u"(d), is well-defined. Let V solve VP = pU. That V solves (3.4) is demonstrated as in Case 2. Each u" is u.s.c. on D(b; /)by recursive applications of Lemma A2.1 and MV.1(b). Thus the same is true of the infimum. Q.E.D.

APPENDIX 4

This appendix defines the map e employed in (5.2) and then proves Theorem 5.1. Denote by F;+ the 0'-algebra generated by (1,. i, )'{'.We assume that associated with the stochastic process ( r,, i, )o 0, B E F; + and I, E / 1 , (A4.1)

.Y (·I I,) is a probability measure on (fl. F; +) and

(A4.2)

.Y( Bl ·)is f;_-measurable.

The fact that .Y does not vary with t reflects the assumed stationarity of the process. Each plan, in conjunction with the wealth accumulation equation, defines a random variable (r.v.) _)i: !? ...... Y( b)= U, > 0 Y( b; I) such that .Y = ( E0 , ... , E,, ... ) and E, is measurable with respect to f;_, the For given I0 and x 0 , E0 is nonstochastic and is written simply c0 . 0'-algebra generated by (1,. ゥ[IG⦅MセN@ Feasibility implies that oc

(A4.3)

LE,/r'.,; x 0 a.e. ['1']. 0

For given I 0 and x 0 , we associate with each feasible r.v. consumption program y a temporal lottery d = (d1 , ... , d,, ... ), d, = (c0 , m,), m, E M(D,_ d for t > 2 and m 1 E M(Y(i'; x0 )). Each such dE D(i'; x 0 ) c D(b; x 0 ) c D(b) 'Vb > r. Roughly speaking, the association from .Y to dis as follows. The measure m 1 is that implied by the function .Y and the probability distribution for ( i, )g' by treating all of the latter as though they were realized at t = 1. In this way one obtains a consumption Similarly, if we lottery in which all uncertainty is resolved by t = 1, i.e., m 1 E M(Y(r; x0 )) c mHrセIN@

r,,

236

LG Epstein and SE Zin

966

LARRY G. EPSTEIN AND STANLEY E. ZIN

recognize that (r0 , z0 ) is realized at r = 1 and regard (T,, z,), t ;;.1, as being realized at t- 2, we obtain m 2 e M(R+x M(Ro;')) = M(D1 ); and so on for the other m,'s. These measures differ only in the way the (common) uncertainty regarding future consumption is resolved through time. Thus an element de D is obtained. Formally, define m1 by

m1 (Q/I0 )='1i{{we0: Y(w)eQ}/I0 ), QeB(Ro;'). By (A4.3), m 1 e M(Y(r; x 0 )). Suppose that we have constructed m 1 (·/l1 )EM(D,_I) for some r;;.l. (Actually, for t=1, m1 (·/l0 ) e M(Ro;').) Then m1+ 1(·/l) is defined by m,+l(Q/l) =

>Y( {we 0: {c1 (w ), m,( ·j(I, "'o)}) e

Q} ji).

Here, "'o = (r0 , z0 ) and (I, "'o) represents realized history at t = 1. In this way d = (( c0 , mi), ... , ( c0 , m, ), ... } e D{r; x 0 )

is constructed. The desired map e: FP .... D(r; x 0 ) is the composition: plan-+ r.v. consumption prograni-+ temporal lottery. PROOF OF THEOREM 5.1: Define Cases 1-3 as in the proof of Theorem 3.1. In those cases where a contraction mapping is applicable, existence follows in the customary fashion. For the remaining cases we employ parallel arguments to those in the proof of Theorem 3.1. An alternative route would be to prove the compactness of e(FP) and apply the u.s.c. of utility. But we were able to show only that the closure of e(FP) is compact and not that e(FP) is closed. Thus only the finiteness of the supremum in (5.2) could be inferred in this way. CASE 1: Since WCMT applies, the existence of a maximum follows from Denardo (1967). That it can be achieved by a stationary plan follows from Sobel (1975). Oearly, J > 0 since c0 = x 0 is feasible and yields utility= x0 . CASE 2: Let v* and J* be the utility and value functions respectively for Case 1. Because of the homotheticity of preferences we may write J*(l, x) = A*(l)x. A*(·) is bounded above since v* is bounded on D(r; x 0 ) (see (A3.2)). Moreover, we have just seen that J*;;. x 0 . We conclude that (A4.4)

1.;>A*(·).;>a

max

w( a,(1-a)E[A•(i;,)w'r0 ji]),

aE[O,l], wESK

which in tum equals A*(l0 ) by the definition of A*. Thus (FA*)(l) .;>A*(l).

By induction, prove that FNA*(l)J, in N. Since the sequence is bounded below by 0, A{l)= lim FNA*(l) N-oo

is well-defined. We now show that A(l0 )x0 is at least as ャ。イセ・@ as the supremum in (5.2): For any feasible lottery (c0 , m), v*(c0 , m) .;> A*(l0 )x0 • By induction, T v*(c0 , m) .;> FNA*(l0 )x0 , where Tis defined following (A3.3). Thus taking limits yields V(c0 , m) .;> A(l0 )x0 • We need to show that the supremum is attained. For any plan h=(h 0 , ... ,h1, ... ) denote by e(h; [0 , x 0 ) the temporal lottery in D(b), where e is defined above and its dependence on I0 and x 0 is made explicit. Then V(e(h; l 0 , x0 )) is the utility of the plan h given initial conditions (l0 , x 0 ). By

Chapter 12. Substitution, Risk Aversion and the Temporal Behavior of Consumption TEMPORAL BEHAVIOR OF CONSUMPTION

237

967

homogeneity, (A4.6) Let h** be any (not necessarily feasible) plan such that e(h**; 10 , x 0 )

E

D(b; I) for some I> 0 and

(Note that h** is not the optimal plan of Case 1 since the utility function Vis not the utility function of Case 1.) For example, h** could be the plan in which c0 = A*(/0 )x 0 and consumption is set equal to zero in all subsequent periods. Denote by ( a0 (!; H), w0 (!; H)) the solution to (A4.5) when A*(·) in the maximand is replaced by H( · ). Define the transformation of plans G by G(h 0 , ... , h,, ... ) = ( h 0 , h 0 , .. . , h,, ... ),

where h0 (l, x) = (xa 0 (!; H), w0 (!; H)) and His defined in (A4.6). Consider the sequence of plans { GN(h**); N セ@ 1 }. Following is a list of facts regarding this sequence. Write GN(h**) = HィセL@ ... , h{', ... ) and h{' =(he{', hp,N). FACT 1: V(e(GN(h**): 10 , x 0 )) = FNA*(/0 ) · x 0 セ@ A(l0 )xo. FACT 2: For each t and N セ@ t + 1, he{'(·) is increasing in N if p > 0 and decreasing if p < 0. In either case, he, ( ·) = lim he{' ( ·) exists. FACT 3: For each I, [and N セiK@ l,

FACT 4: The plan h = (( hc0 , hp0 ), ... , (he" hp1), ... ) is feasible and stationary. FACT 5: e(GN(h**); 10 , x0 )--> e(h; 10 , x0 ) in D(b; I). (To prove the latter use the fact that a.e. pointwise convergence of a sequence of random variables implies the weak convergence of the corresponding sequence of probability measures.) Combine Facts 1, 5, and the u.s.c. of utility to cpnclude that

But A (!0 )x0 is no smaller than the utility supremum over feasible paths. Thus by Fact 4, equality must prevail and the supremum is achieved. The positivity of J is clear as in Case l. The argument for Case 3 is sintilar to that for Case 2. To prove positivity when p < 0, note that there exists i\ セ@ 1 and a sufficiently small e > 0 such that the plan of consunting e · (r/i\) 1 in each t (and making any portfolio decisions whatsoever) is feasible. This plan yields utility

Q.E.D.

REFERENCES BILLINGSLEY, P. (1979): Probability and Measure. New York: John Wiley. - - (1968): Convergence of Probability Measures. New York: John Wiley. BoYD, J. H. (1987): "Recursive Utility and the Ramsey Problem," University of Rochester, ntimeo. CHEW, S. H. (1983): "A Generalization of the Quasilinear Mean with Applications to the Measurement of Income Inequality and Decision Theory Resolving the Allais Paradox," Econometrica, 51, 1065-1092. - - (1989): "Axiomatic Utility Theories with the Betweenness Property," Annals of Operations Research: Choice Under Uncertainty, forthcoming.

238

968

LG Epstein and SE Zin LARRY G. EPSTEIN AND STANLEY E. ZIN

CHEW, S. H., AND L. G. EPSTEIN (1987): "Non-Expected Utility Preferences in a Temporal Framework with an Application to Consumption-Savings Behavior," University of Toronto, Working Paper 8701. To appear in Journal of Economic Theory. - - (1989): "The Structure of Preferences and Attitudes Towards the Timing of the Resolution of Uncertainty," University of Toronto, Working Paper 8706, forthcoming in the International Economic Review, 30, 103-117. CHEW, S. H., L. G. EPSTEIN, AND I. ZILCHA (1988): "A Correspondence Theorem Between Smooth Utility and Expected Utility," Journal of Economic Theory, 46, 186-193. DEKEL, E. (1986): "An Axiomatic Characterization of Preferences under Uncertainty," Journal of Economic Theory, 40, 304-318. DENARDO, E. V. (1967): "Contraction Mappings in the Theory Underlying Dynamic Programming," SIAM Review, 9, 165-177. EPSTEIN, L. G. (1988): "Risk Aversion and Asset Prices," Journal of Monetary Economics, 22, 179-192. - - (1987): "Global Stability and Efficient Allocations," Econometrica, 55, 329-355. - - (1986): "Implicitly Additive Utility and the Nature of Optimal Economic Growth," Journal of Mathematical Economics, 15, 111-128. - - (1983): "Stationary Cardinal Utility and Optimal Growth Under Uncertainty," Journal of Economic Theory, 31, 133-152. EPSTEIN, L. G., AND S. M. TANNY (1980): "Generalized Increasing Correlation: A Definition and Some Economic Consequences," Canadian Journal of Economics, 13, 16-34. EPSTEIN, L. G., AND S. E. ZIN (1989): "Substitution, Risk Aversion and the Temporal Behavior of Consumption and Asset Returns II: An Empirical Analysis," Working Paper 8718, University of Toronto, revised. FARMER, R. E. A. (1987): "Closed-Form Solutions to Dynamic Stochastic Choice Problems," University of Pennsylvania. GROSSMAN, S., AND G. LAROQUE (1987): "Asset Pricing and Optimal Portfolio Choice in the Presence of Illiquid Durable Consumption Goods," mimeo. GROSSMAN, S., AND R. SHILLER (1981): "The Determinants of the Variability of Stock Market Prices," American Economic Review, 71, 222-227. HALL, ROBERT E. (1978): "Stochastic Implications of the Life Cycle-Permanent Income Hypothesis: Theory and Evidence," Journal of Political Economy, 86, 971-987. - - (1981): "Intertemporal Substitution in Consumption," NBER Working Paper 720. - - (1985): "Real Interest and Consumption," NBER Working Paper 1694. HANSEN, L., AND K. SINGLETON (1983): "Stochastic Consumption, Risk Aversion, and the Temporal Behavior of Asset Returns," Journal of Political Economy, 91, 249-265. - - (1982): "Generalized Instrumental Variables Estimation of Nonlinear Rational Expectations Models," Econometrica, 50, 1269-1286. JOHNSEN, T. H., AND J. B. DONAWSON (1985): "The Structure of Intertemporal Preferences under Uncertainty and Time Consistent Plans," Econometrica, 53, 1451-1458. KIHLSTROM, R. E., AND L. J. MIRMAN (1974): "Risk Aversion with Many Commodities," Journal of Economic Theory, 8, 361-88. KOOPMANS, T. C. (1960): "Stationary Ordinal Utility and Impatience," Econometrica, 28, 287-309. KREPS, D. M., AND E. L. PORTEUS (1978): "Temporal Resolution of Uncertainty and Dynamic Choice Theory," Econometrica, 46, 185-200. - - (1979a): "Temporal von Neumann-Morgenstem and Induced Preferences," Journal of Economic Theory, 20, 81-109. - - (1979b): "Dynamic Choice Theory and Dynamic Programming," Econometrica, 47, 91-100. LUCAS, R. E. (1978): "Asset Prices in an Exchange Economy," Econometrica, 46, 1426-1445. LUCAS, R. E., AND N. STOKEY (1984): "Optimal Growth with Many Consumers," Journal of Economic Theory, 32, 139-171. MACHINA, M. J. (1982): "'Expected Utility' Analysis Without the Independence Axiom," Econometrica, 50, 277-323. - - (1984): "Temporal Risk and the Nature of Induced Preference," Journal of Economic Theory, 33, 199-231. MEHRA, R., AND E. PRESCOTT (1985): "The Equity Premium: A Puzzle," Journal of Monetary Economics, 15, 145-161. MERTENS, J. F., AND S. ZAMIR (1985): "Formalization of Bayesian Analysis for Games with Incomplete Information," International Journal of Game Theory, 14, 1-29.

Chapter 12. Substitution, Risk Aversion and the Temporal Behavior of Consumption TEMPORAL BEHAVIOR OF CONSUMPTION

239

969

MYERSON, R. (1985): "Bayesian Equilibrium and Incentive Compatability: An Introduction" in Social Goals and Social Organizations, Essays in Memory of Elisha Pazner, ed. by L. Hurwicz, D. Schmeidler, and H. Sonnenschein. Cambridge: Cambridge University Press. PARTHASARATHY, K. R. (1967): Probability Measures on Metric Spaces. New York: Academic Press. SARGENT, T. (1987): Dynamic Macroeconomic Theory. Cambridge, Mass.: Harvard University Press. SELDEN, L. (1978): "A New Representation of Preference over 'Certain X Uncertain' Consumption Pairs: The 'Ordinal Certainty Equivalent' Hypothesis," Econometrica, 46, 1045-1060. SOBEL, M. J. (1975): "Ordinal Dynamic Programming," Management Science, 21, 967-975. WEIL, P. (1987a): "Asset Prices with Non-Expected Utility Preferences," Harvard University, mimeo. - - (1987b): "Non-Expected Utility in Macroeconomics," Harvard University, mimeo. YAARI, M. E. (1987): "The Dual Theory of Choice Under Risk," Econometrica, 55, 95-115. ZIN, S. (1987): "Intertemporal Substitution, Risk and the Time Series Behavior of Consumption and Asset Returns," Ph.D. Thesis, University of Toronto.

This page intentionally left blank

Econometrica, Vol. 68, No. 5 (September, 2000), 1281-1292

NOTES AND COMMENTS

Chapter 13 RISK AVERSION AND EXPECTED-UTILITY THEORY: A CALIBRATION THEOREM

BY

MATIHEW RABIN 1

1. INTRODUCTION USING EXPECTED-UTILITY THEORY, economists model risk aversion as ansmg solely because the utility function over wealth is concave. This diminishing-marginal-utility-ofwealth theory of risk aversion is psychologically intuitive, and surely helps explain some of our aversion to large-scale risk: We dislike vast uncertainty in lifetime wealth because a dollar that helps us avoid poverty is more valuable than a dollar that helps us become very rich. Yet this theory also implies that people are approximately risk neutral when stakes are small. Arrow (1971, p. 100) shows that an expected-utility maximizer with a differentiable utility function will always want to take a sufficiently small stake in any positive・クー」エ、Mカセャオ@ bet. That is, expected-utility maximizers are (almost everywhere) arbitrarily close to risk neutral when stakes are arbitrarily small. While most economists understand this formal limit result, fewer appreciate that the approximate risk-neutrality prediction holds not just for negligible stakes, but for quite sizable and economically important stakes. Economists often invoke expected-utility theory to explain substantial (observed or posited) risk aversion over stakes where the theory actually predicts virtual risk neutrality. While not broadly appreciated, the inability of expected-utility theory to provide a plausible account of risk aversion over modest stakes has become oral tradition among some subsets of researchers, and has been illustrated in writing in a variety of different contexts using standard utility functions. 2 In this paper, I reinforce this previous research by presenting a theorem that calibrates a relationship between risk attitudes over small and large stakes. The theorem shows that, within the expected-utility model, anything but virtual risk neutrality over modest stakes implies manifestly unrealistic risk aversion over 1 Many people, including David Bowman, Colin Camerer, Eddie Dekel, Larry Epstein, Erik Eyster, Mitch Polinsky, Drazen Prelec, Richard Thaler, and Roberto Weber, as well as a co-editor and two anonymous referees, have provided useful feedback on this paper. I thank Jimmy Chan, Erik Eyster, Roberto Weber, and especially Steven Blatt for research assistance, and the Russell Sage, MacArthur, National Science (Award 9709485), and Sloan Foundations for financial support. I also thank the Center for Advanced Studies in Behavioral Sciences, supported by NSF Grant SBR-960123, where an earlier draft of the paper was written. 2 See Epstein (1992), Epstein and Zin (1990), Hansson (1988), Kandel and Stambaugh (1991), Loomes and Segal (1994), and Segal and Spivak (1990). Hansson's (1988) discussion is most similar to the themes raised in this paper. He illustrates how a person who for all initial wealth levels is exactly indifferent between gaining $7 for sure and a 50-50 gamble of gaining either $0 or $21 prefers a sure gain of $7 to any lottery where the chance of gaining positive amounts of money is less than 40%-no matter how large the potential gain is.

1281

241

242

M Rabin

1282

MATIHEW RABIN

large stakes. The theorem is entirely "nonparametric," assuming nothing about the utility function except concavity. In the next section I illustrate implications of the theorem with examples of the form "If an expected-utility maximizer always turns down modest-stakes gamble X, she will always turn down large-stakes gamble Y." Suppose that, from any initial wealth level, a person turns down gambles where she loses $100 or gains $110, each with 50% probability. Then she will turn down 50-50 bets of losing $1,000 or gaining any sum of money. A person who would always turn down 50-50 lose $1,000/gain $1,050 bets would always turn down 50-50 bets of losing $20,000 or gaining any sum. These are implausible degrees of risk aversion. The theorem not only yields implications if we know somebody will turn down a bet for all initial wealth levels. Suppose we knew a risk-averse person turns down 50-50 lose $100/gain $105 bets for any lifetime wealth level less than $350,000, but knew nothing about the degree of her risk aversion for wealth levels above $350,000. Then we know that from an initial wealth level of $340,000 the person will turn down a 50-50 bet of losing $4,000 and gaining $635,670. The intuition for such examples, and for the theorem itself, is that within the expected-utility framework, turning down a modest-stakes gamble means that the marginal utility of money must diminish very quickly for small changes in wealth. For instance, if you reject a 50-50 lose $10/gain $11 gamble because of diminishing marginal utility, it must be that you value the eleventh dollar above your current wealth by at most (10/11) as much as you valued the tenth-to-last-dollar of your current wealth. 3 Iterating this observation, if you have the same aversion to the lose $10/gain $11 bet if you were $21 wealthier, you value the thirty-second dollar above your current wealth by at most (10/11)X(10j11)"'(5/6) as much as your tenth-to-last dollar. You will value your two-hundred-twentieth dollar by at most (3 /20) as much as your last dollar, and your eight-hundred-eightieth dollar by at most (1/2,000) of your last dollar. This is an absurd rate for the value of money to deteriorate-and the theorem shows tAe rate of deterioration implied by expected-utility theory is actually quicker than this. Indeed, the theorem is really just an algebraic articulation of how implausible it is that the consumption value of a dollar changes significantly as a function of whether your lifetime wealth is $10, $100, or even $1,000 higher or lower. From such observations we should conclude that aversion to modest-stakes risk has nothing to do with the diminishing marginal utility of wealth. Expected-utility theory seems to be a useful and adequate model of risk aversion for many purposes, and it is especially attractive in lieu of an equally tractable alternative model. "Extremely-concave expected utility" may even be useful as a parsimonious tool for modeling aversion to modest-scale risk. But this and previous papers make clear that expected-utility theory is manifestly not close to the right explanation of risk attitudes over modest stakes. Moreover, when the specific structure of expected-utility theory is used to analyze situations involving modest stakes-such as in research that assumes that large-stake and modest-stake risk attitudes derive from the same utility-for-wealth 3 My wording here, as in the opening paragraph and elsewhere, gives a psychological interpretation to the concavity of the utility function. Yet a referee has reminded me that a common perspective among economists studying choice under uncertainty has been that the concavity of the utility function need be given no psychological interpretation. I add such a psychological interpretation throughout the paper as an aid to those readers who, like me, find this approach to be the natural way to think about utility theory, but of course the mathematical results and behavioral analysis in this paper hold without such interpretations.

Chapter 13. Risk Aversion and Expected-Utility Theory RISK AVERSION

243

1283

function-it can be very misleading. In the concluding section, I discuss a few examples of such research where the expected-utility hypothesis is detrimentally maintained, and speculate very briefly on what set of ingredients may be needed to provide a better account of risk attitudes. In the next section, I discuss the theorem and illustrate its implications.

2.

SOME CALIBRATIONS BASED ON A THEOREM

Consider an expected-utility maximizer over wealth, w, with Von Neumann-Morgenstem preferences U( w ). Assume that the person likes money and is risk-averse: For all w,U(w) is (strictly) increasing and (weakly) concave. Suppose further that, for some range of initial wealth levels and for some g > l > 0, she will reject bets losing $1 or gaining $g, each with 50% chance. 4 From the assumption that these bets will be rejected, the theorem presented in this paper places an upper bound on the rate at which utility increases above a given wealth level, and a lower bound on the rate at which utility decreases below that wealth level. Its proof is a short series of algebraic manipulations; both the theorem and proof are in the Appendix. Its basic intuition is straightforward, as described briefly in the introduction. The theorem handles cases where we know a person to be averse to a gamble only for some ranges of initial wealth. A simpler corollary, also in the Appendix, holds when we know a lower bound on risk aversion for all wealth levels. Table I illustrates some of the corollary's \implications: Consider an individual who is known to reject, for all initial wealth levels, 50-50, lose $100/gain g bets, for g = $101, $105, $110, and $125. The table presents implications of the form "the person will turn down 50-50 gambles of losing L and gaining G," where each L is a row in the table and the highest G (using the bounds established by the corollary) making the statement true is the entry in the table. 5 All entries are rounded down to an even dollar amount. So, for instance, if a person always turns down a 50-50 lose $100jgain $110 gamble, she will always turn down a 50-50 lose $800jgain $2,090 gamble. Entries of oo are literal: Somebody who always turns down 50-50 lose $100/gain $125 gambles will turn down any gamble with a 50% chance of losing $600. This is because the fact that the bound on risk aversion holds everywhere implies that U( w) is bounded above. 4 The assumption that U is concave is not implied by the fact that an agent always turns down a given better-than-fair bet; if you know that a person always turns down 50-50 lose $10jgain $11 bets, you don't know that her utility function is concave everywhere-it could be convex over small ranges. (For instance, let U(w) = 1- C±)w for w $ (19,20), but

1 ) 19 + U(w)=1- ( 2

[( 1 ) 19 - ( 1 ) 20] (w-19) 2 2

2

for wE (19, 20).) Concavity is an additional assumption, but I am confident that results hold approximately if we allow such small and silly nonconvexities. 5 The theorem provides a lower bound on the concavity of the utility function, and its proof indicates an obvious way to obtain a stronger (but uglier) result. Also, while the theorem and applications focus on "S0-50 bets," the point is applicable to more general bets. For instance, if an expected-utility maximizer dislikes a bet with a 25% chance of losing $100 and a 75% chance of winning $100, then she would turn down 50-50 lose $100jgain $300 bets, and we could apply the theorem from there.

244

M Rabin

1284

MATTHEW RABIN TABLE I IF AVERSE TO 50-50 LOSE $100 I GAIN g BETS FOR ALL WEALTII LEVELS, WILL TuRN DowN 50-50 LosE L 1 GAIN G BETS; G's ENTERED IN TABLE. g

L

$400 $600 $800 $1,000 $2,000 $4,000 $6,000 $8,000 $10,000 $20,000

$101

$105

$110

$125

400 600 800 1,010 2,320 5,750 11,810 34,940

420 730 1,050 1,570

550 990 2,090

1,250 00

00

00

00

00

00

00

00

00

00

00

00

00

00

00

00

00

00

00

00

00

00

The theorem and corollary are homogeneous of degree 1: If we know that turning down 50-50 lose ljgain g gambles implies you will turn down 50-50 lose Ljgain G, then for all x > 0, turning down 50-50 lose xl I gain xg gambles implies you will turn down 50-50 lose xLjgain xG. Hence the L = $10,000, g = $101 entry in Table I tells us that turning down 50-50 lose $10jgain $10.10 gambles implies you will turn down 50-50 lose $1,000/gain co gambles. The reader may worry that the extreme risk aversion shown in Table I relies heavily on the assumption that the person will turn down the given gamble for all initial wealth levels. It doesn't. While without knowing a global lower bound on a person's modest-stakes risk aversion we cannot assert that she'll turn down gambles with infinite expected return, Table II indicates that the lack of a lower bound does not salvage the plausibility of expected-utility theory. Table II shows calibrations if we know the person will turn down 50-50 lose $100jgain g gambles for initial wealth levels less than $300,000, indicating which gambles she'll turn down starting from initial wealth level of $290,000. Large entries are approximate. TABLE II TABLE I REPLICATED, FOR INITIAL WEALTH LEVEL $290,000, WHEN ilg BEHAVIOR IS ONLY KNOWN TO HoLD FOR W :5:$300,000. g L

$400 $600 $800 $1,000 $2,000 $4,000 $6,000 $8,000 $10,000 $20,000

$101

$105

$110

$125

400 600 800 1,010 2,320 5,750 11,510 19,290 27,780 85,750

420 730 1,050 1,570 69,930 635,670 1,557,360 3,058,540 5,503,790 71,799,110

550 990 2,090 718,190 12,210,880 60,528,930 180,000,000 510,000,000 1,300,000,000 160,000,000,000

1,250 36,000,000,000 90,000,000,000 160,000,000,000 850,000,000,000 9,400,000,000,000 89,000,000,000,000 830,000,000,000,000 7, 700,000,000,000,000 540,000,000,000,000,000,000

Chapter 13. Risk Aversion and Expected-Utility Theory

245

1285

RISK AVERSION TABLE III

IF A PERSON HAs CARA UTILITY FuNCTION AND Is AvERSE TO 50150 LosE$/ 1 GAIN $g BETS FOR ALL WEALTH LEVELS, THEN (i) SHE HAs COEFFICIENT OF ABSOLUTE RISK AVERSION No SMALLER THAN p AND (ii) INVESTS $X IN THE STOCK MARKET WHEN STOCK YIELDS ARE NORMALLY DISTRIBUTED WITH MEAN RE/\L RETURN 6.4% i\ND STANDi\RD DEVIATION 20%, AND BONDS YIELD A RISKLESS RETURN OF 0.5%. ljg

$1001$101 $1001$105 $1001$110 $1001$125 $100j$150 $1,0001$1,050 $1,0001$1,100 $1,0001$1,200 $1,0001$1,500 $1,0001$2,000 $10,0001$11,000 $10,0001$12,000 $10,0001$15,000 $10,0001$20,000

f>

X

.0000990 .0004760 .0009084 .0019917 .0032886 .0000476 .0000908 .0001662 .0003288 .0004812 .0000090 .0000166 .0000328 .0000481

$14,899 $3,099 $1,639 $741 $449 $30,987 $16,389 $8,886 $4,497 $3,067 $163,889 $88,855 $44,970 $30,665

If we only know that a person turns down 50-50 lose $100jgain $125 bets when her lifetime wealth is below $300,000, we also know she will turn down a 50-50 lose $600/gain $36 billion bet beginning from lifetime $290,000. 6 The intuition is that the extreme concavity of the utility function between $290,000 and $300,000 assures that the marginal utility at $300,000 is tiny compared to the marginal utility at wealth levels below $290,000; hence, even if the marginal utility does not diminish at all above $300,000, a person will not care nearly as much about money above $300,000 as she does about amounts below $290,000. The choice of $290,000 and $300,000 as the two focal wealth levels is arbitrary; all that matters is that they are $10,000 apart. As with Table I, Table II is homogeneous of degree 1, where the wedge between the two wealth levels must be multiplied by the same factor as the other entries. Hence-multiplying Table II by 10-if an expected-utility maximizer would turn down a 50-50 lose $1,000/gain $1,050 gamble whenever her lifetime wealth is below $300,000, then from an initial wealth level of $200,000 she will turn down a 50-50 lose $40,000/gain $6,356,700 gamble. While these "nonparametric" calibrations are less conducive to analyzing more complex questions, Table III provides similar calibrations for decisions that resemble realworld investment choices by assuming conventional functional forms of utility functions. It shows what aversion to various gambles implies for the maximum amount of money a person with a constant-absolute-risk-aversion (CARA) utility function would be willing to

6 Careful examination of Tables I and II shows that most entries that are not oo in Table I show up exactly the same in Table II. The two exceptions are those entries that are above $10,000-since Table II implicitly makes no assumptions about concavity for gains of more than $10,000, it yields lower numbers.

246

M Rabin

1286

MATTHEW RABIN

keep invested in the stock market, for reasonable assumptions about the distribution of returns for stocks and bonds. Hence, an expected-utility maximizer with CARA preferences who turns down 50j50 lose $1,000/gain $1,200 gambles will only be willing to keep $8,875 of her portfolio in the stock market, no matter how large her total investments in stocks and bonds. If she turns down a 50j50 lose $100/gain $110 bet, she will be willing to keep only $1,600 of her portfolio in the stock market-keeping the rest in bonds (which average 6% lower annual return). While it is widely believed that investors are too cautious in their investment behavior, no one believes they are this risk averse.

3. DISCUSSION AND CONCLUSION

Expected-utility theory may well be a useful model of the taste for very-large-scale insurance. 7 Despite its usefulness, however, there are reasons why it is important for economists to recognize how miscalibrated expected-utility theory is as an explanation of modest-scale risk aversion. For instance, some research methods economists currently employ should be abandoned because they rely crucially on the expected-utility interpretation of modest-scale risk aversion. One example arises in experimental economics. In recent years, there has been extensive laboratory research in economics in which subjects interact to generate outcomes with payoffs on the order of $10 to $20. Researchers are often interested in inferring subjects' beliefs from their behavior. Doing so often requires knowing the relative value subjects hold for different money prizes; if a person chooses $5 in event A over $10 in event B, we know that she believes A is at least twice as likely as B only if we can assume the person likes $10 at least twice as much as $5. Yet economic theory tells us that, because of diminishing marginal utility of wealth, we should not assume people like $10 exactly twice as much as $5. Experimentalists (e.g., Davis and Holt (1993, pp. 472-476)) have developed a clever scheme to avoid this problem: Instead of prizes of $10 and $5, subjects are given prizes such as 10% chance of winning $100 vs. 5% chance of winning $100. Expected-utility theory tells us that, irrespective of the utility function, a subject values the 10% chance of a prize exactly twice as much as the 5% chance of winning the same prize. The problem with this lottery procedure is that it is known to be sufficient only when we maintain the expected-utility hypothesis. But then it is not necessary-since expected-utility theory tells us that people will be virtually risk neutral in decisions on the scale of laboratory stakes. If expected-utility theory is right, these procedures are at best redundant, and are probably harmful. 8 On the other hand, if we think that subjects in experiments are risk averse, then we know they are not expected-utility maximizers. Hence the lottery procedure, which is motivated solely by expected-utility theory's assumptions that preferences are linear in probabilities and that risk attitudes come only 7 While there is also much evidence for some limits of its applicability for large-scale risks, and the results of this paper suggest an important flaw with the expected-utility model, the specific results do not of course demonstrate that the model is unuseful in all domains. 8 If expected-utility theory explained behavior, these procedures would surely not be worth the extra expense, nor the loss in reliability of the data from making experiments more complicated. Nor should experimentalists who believe in expected-utility theory ever be cautious about inferences made from existing experiments that don't use the lottery methods out of fear that the results are confounded by the subjects' risk attitudes.

Chapter 13. Risk Aversion and Expected-Utility Theory RISK A VERSION

247

1287

from the curvature of the utility-of-wealth function, has little presumptive value in "neutralizing" risk aversion. Perhaps there are theories of risk attitudes such that the lottery procedure is useful for neutralizing risk aversion-but expected-utility theory isn't one of them. 9 A second example of problematic research methods relates to the logic underlying the theorem: Expected-utility theory makes wrong predictions about the relationship between risk aversion over modest stakes and risk aversion over large stakes. Hence, when measuring risk attitudes maintaining the expected-utility hypothesis, differences in estimates of risk attitudes may come from differences in the scale of risk comprising data sets, rather than from differences in risk attitudes of the people being studied. 10 Data sets dominated by modest-risk investment opportunities are likely to yield much higher estimates of risk aversion than data sets dominated by larger-scale investment opportunities. So not only are standard measures of risk aversion somewhat hard to interpret given that people are not expected-utility maximizers, but even attempts to compare risk attitudes so as to compare across groups will be misleading unless economists pay due attention to the theory's calibrational problems. The problems with assuming that risk attitudes over modest and large stakes derive from the same utility-of-wealth function relate to a long-standing debate in economics. Expected-utility· theory makes a powerful prediction that economic actors don't see an amalgamation of independent gambles as significant insurance against the risk of those gambles; they are either barely less willing or barely more willing to accept risks when clumped together than when apart. This observation was introduced in a famous article by SamueJson (1963), who showed that expected-utility theory implies that if (for some sufficiently wide range of initial wealth levels) a person turns down a particular gamble, then she should also tum down an offer to play n > 1 of those gambles. Hence, in his example, if Samuelson's colleague is unwilling to accept a 50-50 lose $100/gain $200 gamble, then he should be unwilling to accept 100 of those gambles taken together. Though Samuelson's theorem is "weaker" than the one in this paper, it makes manifest the fact that expected-utility theory predicts that adding together a lot of independent risks should not appreciably alter attitudes towards those risks.

9 Indeed, the observation that diminishing marginal utility of wealth is irrelevant in laboratory experiments raises questions about interpreting experimental tests of the adequacy of expectedutility theory. For instance, while showing that existing alternative models better fit experimental data than does expected-utility theory, Harless and Camerer (1994) show that expected-utility theory better fits experimental data than does "expected-value theory"-risk-neutral expected-utility theory. But because expected-utility theory implies that laboratory subjects should be risk neutral, such evidence that expected-utility theory explains behavior better than expected-value theory is evidence against expected-utility theory. 10 Indeed, Kandel and Stambaugh (1991, pp. 68-69) discuss how the plausibility of estimates for the coefficient of relative risk aversion may be very sensitive to the scale of risk being examined. Assuming constant risk aversion, they illustrate how a coefficient of relative risk aversion needed to avoid predicting absurdly large aversion to a 50j50 lose $25,000jgain $25,000 gamble generates absurdly little aversion to a 50/50 lose $400/gain $400 gamble. They summarize such examples as saying that "Inferences about [the coefficient of relative risk aversion] are perhaps most elusive when pursued in the introspective context of thought experiments." But precisely the same problem makes inferences from real data elusive.

248

1288

M Rabin MAITHEW RABIN

Yet virtually everybody would find the aggregated gamble of 100 50-50 lose $100/gain $200 bets attractive. It has an expected yield of $5,000, with negligible risk: There is only a (1/700) chance of losing any money and merely a Cl/25,000) chance of losing more than $l,OOO.U While nobody would turn down this gamble, many people, such as Samuelson's colleague, might reject the single 50-50 lose $100jgain $200 bet. 12 Hence, using expected-utility theory to make inferences about the risk attitudes towards the amalgamated bet from the reaction to the one bet-or vice versa-would be misleading. What does explain risk aversion over modest stakes? While this paper provides a "proof by calibration" that expected-utility theory does not help explain some risk attitudes, there are of course more direct tests showing that alternative models better capture risk attitudes. There is a large literature (see Machina (1987) and Camerer (1992) for reviews) of formal models of such alternatives. Many of these models seem to provide a more plausible account of modest-scale risk attitudes, allowing both substantial risk aversion over modest stakes and nonridiculous risk aversion over large stakes, and researchers (e.g., Segal and Spivak (1990), Loomes and Segal (1994), Epstein and Zin (1990)) have explicitly addressed how non-expected-utility theory can help account for small-stake risk aversion. Indeed, what is empirically the most firmly established feature of risk preferences, loss aversion, is a departure from expected-utility theory that provides a direct explanation for modest-scale risk aversion. Loss aversion says that people are significantly more averse to losses relative to the status quo than they are attracted by gains, and more generally that people's utilities are determined by changes in wealth rather than absolute levels.U Preferences incorporating loss aversion can reconcile significant small-scale risk aversion with reasonable degrees of large-scale risk aversion. 14 A loss-averse person will, for instance, be likely to turn down the one 50/50 lose $100/gain $200 gamble Samuelson's colleague turned down, but will surely accept one hundred such gambles pooled' together. Variants of this or other models of risk attitudes can provide useful alternatives to

II The theorem in this paper predicts that, under exactly the same assumptions invoked by Samuelson, turning down a 50-50 lose $100jgain $200 gamble implies the person turns down a 50-50 lose $200jgain $20,000 gamble. This has an expected return of $9,900-with zero chance of losing more than $200. IZ As Samuelson noted, the strong statement that somebody should turn down the many bets if and only if she turns down the one is not strictly true if a person's risk attitudes change at different wealth levels. Indeed, many researchers (e.g., Hellwig (1995) and Pratt and Zeckhauser (1987)) have explored features of the utility function such that an expected-utility maximizer might take a multiple of a favorable bet that they would turn down in isolation. But characterizing such instances isn't relevant to examples of the sort discussed by Samuelson. We know from the unwillingness to accept a 50j50 lose $100jgain $200 gamble that Samuelson's colleague was not an expected-utility maximizer. 13 Loss aversion was introduced by Kahneman and Tversky (1979) as part of the more general "prospect theory," and is reviewed in Kahneman, Knetch, and Thaler (1991). Tversky and Kahneman (1991) and others have estimated the loss-aversion-to-gain-attraction ratio to be about 2:1. I 4 While most formal definitions of loss aversion have not made explicit the assumption that people are substantially risk averse even for very small risks (hut see Bowman, Minehart, and Rabin (1999) for an explicit treatment of this issue), most examples and calibrations of loss aversion imply such small-scale risk aversion.

Chapter 13. Risk Aversion and Expected-Utility Theory

249

1289

RISK AVERSION

expected-utility theory that can reconcile plausible risk attitudes over large stakes with nontrivial risk aversion over modest stakes. 15 Dept. of Economics, University of California-Berkeley, Berkeley, CA 94720-3880, U.S.A.; rabin@econ. berkeley. edu; http: I I emlab. berkeley. edul usersI rabin1index.html Manuscript received October, 1997; final revision received May, 1999.

APPENDIX:

THE THEOREM AND A CoROLLARY

THEOREM: Suppose that for all w, U(w) is strictly increasing and weakly concave. Suppose that there exists w > }:!:', g >I> 0, such that for all wE[):!:', w], .5U(w -I)+ .5U(w +g)< U(w ). Then for all WE[!:!:', w], for all X> 0, (i) if g :s; 21, then

(ii)

U(w +x)- U(w) :s;



ォイ⦅クIHセゥキ@

if x:s;w-w, ゥセo@

.

g

.

k"'(W) ( [ )'

ゥセo@

L

-

g

( [

r(w)+[x-w]-

)k'"(W)

r(w)

g

where, letting int(y) denote the smallest integer less than or equal toy, k*(x) int((xjg) + 1), and r(w) U(w)- U(w -I).

=

if

X

2: W-

W,

=int(xj2/), k**(x) =

15 But Kahneman and Lovallo (1993), Benartzi and Thaler (1995), and Read, Loewenstein, and Rabin (1999) argue that an additional departure from standard assumptions implicated in many risk attitudes is that people tend to isolate different risky choices from each other in ways that lead to different behavior than would ensue if these risks were considered jointly. Samuelson's colleague, for instance, might reject each 50/50 lose $100/gain $200 gamble if on each of 100 days he were offered one such gamble, whereas he might accept all of these gambles if they were offered to him at once. Benartzi and Thaler (1995) argue that a related type of myopia is an explanation for the "equity premium puzzle"-the mystery about the curiously large premium on returns that investors seem to demand to compensate for riskiness associated with investing in stocks. Such risk aversion can be explained with plausible (loss-averse) preferences-if investors are assumed to assess gains and losses over a short-run (yearly) horizon rather than the longer-term horizon for which they are actually investing.

250

M Rabin

1290

MATIHEW RABIN

PROOF OF PART (i) OF THEOREM: For notational ease and without loss of generality, let r( w) = U( w) - U( w -f) = 1. Then clearly U(w -f) - U(w - 2/) ?= 1, by the concavity of UO. Also, since 2/ > g > l, we know that w- 21 + g E ( w -I, w ), and by the concavity of UO we know that

U( w - 21 +g) - U( w -I) ?= Mセ

g-l

g

=

I -

1.

Hence, g g U(w- 21 +g)- U(w- 2/) >- -1 + 1 = - .

- I

l

Hence, if w- 21 _]セNキ・@ know that U(w- 2/)- U(w- 3/) ?= (gjl) since by assumption, U(w2l-l)+U(w-2l+g),s,2u(w-2l). By concavity, we also know that U(w-31)-U(w-4/)2 (gjl). More generally, if w - 2kl 2 セN@ then U(w- (2k -l)l)- U(w- 2kl) 2 U(w- 2(k -1)1)- U(w- (2k- l)l)

=

g

U(w- 2kl +g)- U(w- 2kl) ?= l[U(w- 2(k -1)1)- U(w- (2k -l)l)] g

= U(w- 2kl)- U(w- (2k + l)l) 2/[U(w- 2(k -l)l)- U(w- (2k -1)1)]. These lower bounds on marginal utilities yield the lower bound on total utilities U(w)- U(w- x) in part (i) of the Theorem. 16

PROOF OF PART (ii) OF THEOREM: Again let r( w) = U( w) - U( w -I) = 1. Then U( w +g) - U( w) ,s,l. By the concavity of U, U(w+g)-U(w+g-l),s,(ljg). But if w+g,s,w, this implies by assumption that U(w + 2g)- U(w +g) ,s, (/jg) (since U(w + g -I)+ U(w + 2g) ,s, 2U(w +g)). More generally, we know that if w + mg ,s, w, then U(w +mg +g)- U(w +mg) 2 (/jg)[U(w + mg)- U(w +mg- g)].

These upper bounds on marginal utilities yield the upper bound on utilities U( w + x) - U( w) in part (ii) of the Theorem.

16

The theorem is weaker than it could be. If we observe, for all m such that 2 ,s, m ,s, w -

w+ 1,

that U( w - m) - U( w - m - 1) 2 U( w - m

+(

+ 1) -

T- 1)

U( w - m)

[U(w- m + 2)- U( w- m + 1)],

we can prove a stronger (but far messier) theorem. (The current theorem merely invokes U(w- m) - U(w- m -l) 2 U(w- m + 1)- U(w- m) for even m's.)

Chapter 13. Risk Aversion and Expected-Utility Theory

251

1291

RISK AVERSION

COROLLARY: Suppose that for all w, U(w) is strictly increasing and weakly concave. Suppose there exists g > 1 > 0 such that, for all w, .5U( w -I) + .5U( w +g) < U( w ). Then for all positive integers k, for all m < m(k), .S(w- 2kl) + .5U(w + mg) < U(w ), where

(fr]

ln[1- (1-IRLセ@ セ@ m(/c)

l In-

=

-1

(g)' I > o,

if 1 -

(1 - :gl) 2 セ@

if 1 -

(1 - :g1) 2 セ@ (g)' I

k

g

k

: 0.10U($5M) + 0.89U($1M) + 0.01 U ($0), whereas the second ranking implies the inconsistent inequality O.IOU($5M) + 0.90U($0) > 0.11 U($1M) + 0.89U($0). By setting x1 = $0, x2 = $1M, and x3 = $5M, the four prospects can be plotted in the probability triangle as in Figure 2, where they are seen to form a parallelogram. The typical preference for a 1 over a 2 and a 3 over a 4 suggests that indifference curves depart from the expected utility property of parallel straight lines in the direction of 'fanning out', as in the figure. The Allais Paradox is merely one example of a widely observed phenomenon termed the Common Consequence Effect,

Increasing preference

0

Figure 2 Allais Paradox choices and indiflerence curves that 'fan out'

3

which states that when P* is riskier than P, individuals tend to depart from the Independence Axiom by preferring aP* + (1 - a)P** to aP + (1- a)P** when P** involves low payoffs, hut reverse this preference when P** instead involves high payoffs. (In the Allais example, P =($1M,l), P* =($5M, 10/11;$0,1/11), a= 0.11, and P** is ($1M,l) in the first pair and ($0,1) in the second pair.) The second broad class of systematic violations, also originally noted by Allais, can he illustrated by the following example from Kahneman and Tversky [18]: b . { 0.45 chance of 1' 0.55 chance of versus b . { 0.90 chance of 2· 0.10 chance of

$6000 $0 $3000 $0

b . { 0.00 I chance of $6000 3·

0.999 chance of $0 versus b . { 0.002 chance of $3000 4' 0.998 chance of $0 Most individuals express a preference for b 2 over b 1 and for b3 over b4 , which again violates expected utility theory since they imply the inconsistent inequalities 0.45U($6000) + 0.55U($0) < 0.90 U($3000) + O.IOU($0) and O.OOIU($6000) + 0.999U($0) > 0.002U($3000) + 0.998U($0). Setting x1 = $0, x2 = $3000 and x3 = $6000, the prospects can be plotted as in Figure 3, where the above preference rankings again suggest that indifference curves depart from expected utility by fanning out. This is one example of a widely observed phenomenon termed the Common Ralio Effect, which states that for positive-outcome prospects P* which is risky and P which is certain, with P** being a sure chance of $0, individuals tend to depart from the Independence Axiom by preferring aP* + (1 - a )P** to aP + (1 - a)P'* for low values of a, but reverse this preference for high values of a. Kahneman and Tversky [18] observed that when the positive outcomes $3000 and $6000 are replaced by losses of $3000 and then $6000 to create the prospects b'1, bS, b; and 「セL@ preferences typically 'reflect', to prefer b; over bS and 「セ@ over b;. Setting x1 = -$6000, x2 = -$3000 and x3 = $0 (to preserve the ordering x 1 < x2 < x3 ) and plotting as in Figure 4, such preferences again

256

MJ Machina

4

Nonexpected Utility Theory Most of these forms have been axiomatized, and given proper curvature assumptions on their component functions v(-), r(·), G(-), K(·, ·), ... , most are capable of exhibiting first-order stochastic dominance preference, risk aversion, and the common consequence and common ratio effects. The rankdependent expected utility form, in particular, has been widely applied to the analysis of many standard questions in economic choice under uncertainty.

Increasing preference

P3

Generalized Expected Utility Analysis An alternative approach to nonexpected utility, which yields a direct extension of most of the basic results

0

Figure 3 Common ratio effect and indifference curves that fan out

7 suggest that indifference curves in the probability triangle fan out. Further descriptions of these and other violations of the expected utility hypothesis can be found in [6, 18, 21, 23, 24, 34, 35, 38],

Increasing preference

b'1

Nonexpected Utility Functional Forms Researchers have responded to the above types of departures from linearity in the probabilities in two manners. The first consists of replacing the expected utility preference function VEU (-) by some more general functional form for V (P) = V (xt, Pt; ... ; X 11 , p 11 ), such as the forms in the Table below.

0

Figure 4 Couuuon ratio etlect for negative payotls and inditlerence cnrves that fan out

iZ[Gセ

'Prospect theory' HlZ[Gセ

f

Moments of utility

Q@ v(xi)p;, iZ[Gセ iZ[Gセ

Weighted utility

iZ[Gセ

iZ[Gセャ@

Dual expected utility Rank-dependent expected utility Quadratic in probabilities Ordinal independence

Q@ v(x;):n:(p;)

iZ[Gセ

Q@

X; [

G

HiZセ@

v(x;) [ G ( iZGゥセャ@

[11, 12, 18]

Q@ v(x;) 2p;, .. .)

Q@ v(x;)p; Q@ T(x;)p;

I PJ)- G HRZセ

Pi) - G ( iZGゥセ|@

[17]



[7]

Q@

PJ)]

Pi)]

[39]

[28]

[8]

[33]

Chapter 14. Nonexpected Utility Theory Nonexpected Utility Theory of expected utility analysis, applies calculus to a general 'smooth' nonexpected utility preference function V(P) = V(x1, PI; ... ;xn,JJn), concentrating in particular on its probability derivatives av (PJ I a prob (x ). This approach is based on the observation that for the expected utility function VEu (x 1, PI; ... ; x,, Pn) = U(xt)p 1 + · · · + U(x,)pn, the value U(x) can be interpreted as the coefficient of prob(x), and that many theorems relating to a linear function's 」ッセヲᆳ ficients continue to hold when generalized to a nonlinear function's derivatives. By adopting the notation U(x;P) = iJV(P)/iJ prob(x) for the probability derivative, the above expected utility characterizations of first -order stochastic dominance preference, risk aversion, and comparative risk aversion generalize to any smooth preference function V (P) in the following manner: V(·) will exhibit first-order stochastic dominance preference if and only if at each lottery P, U (x ;P) is an increasing function of x

V(·) will exhibit risk aversion (an aversion to all meanpreserving increases in risk) if and only if at each lottery P, U (x ;P) is a concave function of x V*(-) will be at least as risk averse as V(·) if and only if at each lottery P, U*(x;P) is a concave transformation of U (x ;P) It can be shown that the indifference curves of a smooth preference function V (·) will fan out in the probability triangle and its multiple-outcome generalizations if and only if U (x;P') is a concave transformation of U (x;P) whenever P* first order stochastically dominates P; see [3, 9, 19, 22, 23, 37] for the development and further applications of generalized expected utility analysis.

Applications to Insurance The above analysis also permits the extension of many expected utility-based results in the theory of insurance to general smooth nonexpected utility preferences. The key step in this extension consists of observing that the formula for the expected utility outcome derivative -upon which most insurance results are based - is linked to its probability coefficients via the relationship iJVEu(P)jiJx = prob(x)·U'(x), and that this relationship generalizes to any smooth nonexpected utility preference function V (·),where it takes the form iJV(P)jiJx = prob(x)·U'(x;P) (where U'(x)

257 5

and U'(x;P) denote derivatives with respect to the variable x ). One important expected utility result that does not directly extend is the equivalence of risk aversion to the property of outcome-convexity, which states that for any set of probabilities (PI, ... , p, ), if the individual is indifferent between the lotteries (x1, P1: .. . ;x,, p,) and , p1; ... ;x,;, p,) then they will strictly prefer any outcome mixture of the form (f3xt +(I- f3)x1, p1; ... [ヲSクLセ@ +(I- f3)x1, Pnl (for 0 < f3 < I). Under nonexpected utility, outcomeconvexity still implies risk aversion but is no longer implied by it, so for some nonexpected utility insurance results, both risk aversion and outcome-convexity must be explicitly specified. (Nevertheless, the hypothesis that the individual is risk averse and outcomeconvex is still wealcer than the hypothesis that they are a risk averse expected utility maximizer.) One standard insurance problem - termed the demand for coinsurance - involves an individual with initial wealth w who faces a random loss i. with probability distribution (1 1. p 1; ... ; 1,, p,) (/; :;. 0), who can insure fraction y of this loss by paying a premium of .Ay E[i.], where E[i.] is the expected loss and the load factor A equals or exceeds unity. The individual thus selects the most preferred option from the family of random variables {w - ( 1 y)i.- AyEliJIO :5: y :5: 1]. Another standard problem- termed the demand for deductible insurance involves fully covering any loss over some amount d, so the individual receives a payment of max{£d, 0), and paying a premium of AElmax{i.- d, O]j. The individual thus selects the most preferred option from the family of random variables {w- min{£, d)- AEfmax{i.- d, OJ lid:;. 0]. In each case, insurance is said to be actuarially }i1ir if the load factor A equals unity, and actuarially unj(zir if A exceeds unity. Even without the assumption of outcomeconvexity, the following results from the classic expected utility-based theory of insurance can be shown to extend to any risk averse smooth nonexpected utility preference function V (-): V ( ·) will purchase complete coinsurance (will choose to set the fraction y equal to one) if and only if such insurance is actuarially fair V 0 will purchase complete deductible insurance (will choose to set the deductible d equal to zero) if and only such insurance is acluarially fair If V*(-) is at least as risk averse as V (·),then whenever they each have the same initial wealth and

258

6

MJ Machina Nonexpected Utility Theory

face the same random loss, V '(-) will purchase at least as much coinsurance as V ( ·) If V*(-) is at least as risk averse as V(·), then whenever they each have the same initial wealth and face the same random loss, V * ( ·) choose at least as low a deductible level as V(·) If we do assume outcome-convexity, many additional results in the theory of insurance and optimal risk-sharing can be similarly extended from expected utility to smooth nonexpected utility preferences; see [25] for the above and additional extensions, and [10, 14, 16, 20, 32, 40] as well as the papers in [15], for additional results in the theory of insurance under nonexpected utility risk preferences.

References [I]

[2]

[3]

[4] [5]

[6]

[7]

[8] [9]

[10]

[II] [ 12]

Allais, M. (1953). Le Comportement de !'Homme Rationnel devant le Risque, Critique des Postulats et Axiomes de I'Rcole Am6ricaine, Rconometrica 21, 503-546. Allais, M. & Hagen, 0., eds (1979). Expected Utility Hypotheses and the Allais Paradox. D. Reidel Publishing, Dordrecht. Allen, B. (1987). Smooth preferences and the local expected utility hypothesis, Journal of Economic Themy 41. 340-355. Arrow, K. (1963). Comment, Review of' Economics and Statistics 45, (Supplement), 24-27. Bernoulli, D. (1738). Specimen Theoriae Novae de Mensura Sortis, Commentarii Academiae Scientiarum lmperialis Petropolitanae V, 175-192; English translation (1954). Exposition of a New theory on the measurement of risk, liconometrica 22, 23-36. Camerer, C. (1989). An experimental test of several generalized utility theories, Journal r4Risk and Uncertainty 2, 61-104. Chew, S. (1983). A generalization of the quasilincar mean with applications to the measurement of income inequality and decision theory resolving the allais paradox, Econometrica 51, 1065-1092. Chew, S., Epstein, L. & Segal, U. (1991). Mixture symmetry and quadratic utility, Econometrica 59, 139-163. Chew, S., Epstein, L. & Zilcha, I. ( 1988). A correspondence theorem between expected utility and smooth 'theory 46, 186-193. utility, Journal ッャセG」ョュゥ@ Doherty, N. & Eeckhoudt, L. (1995). Optimal insurance without expected utility: the dual theory and the linearity of insurance contracts, Journal イセヲB@ Ri.•Jk and Uncertainty 10, 157-179. Edwards, W. (1955). The prediction of decisions among bets, Journal ofExperimental Psychology 50, 201-214. Edwards, W. ( 1962). Subjective probabilities inferred from decisions, Psychological Revinv 69, I 09-135.

[13]

[14]

[15]

[16]

[17J [18]

[19]

[20]

[21] [221 [23]

[24]

[25j

[26] [27] [28j

[29] [30]

[311 [32j

Friedman, M. & Savage. L. (1948). The utility analysis of choices involving risk, Journal ofPoUt;caf Economy 56, 279-304. Gollier, C. (2000). Optimal insurance design: What can we do without expected utility, in Handbook £?!" Insurance, G. Dionne, ed., Kluwer Academic Publishers, Boston. Gallier, C. & Machina, M., eds (1995). Non-Expected UtWty and R;sk Management, Kluwer Academic Publishers, Boston. Gallier, C. & Schlesinger, H. (1996). Arrow's theorem on the optimality of deductibles: a stochastic dominance approach, Economic Theory 7, 359-363. Hagen, 0. (1979). Towards a Positive Theory of Preferences Under Risk, in [2]. Kahneman, D. & Tversky, A. (1979). Prospect theory: an analysis of decision under risk, Econometrica 47, 263-291. Karni, E. (1989). Generalized expected utility analysis of multivariate risk aversion, International Economic Review 30, 297-305. Konrad, K. & Skaperdas, S. (1993). Self-insurance and self-protection: a nonexpected utility analysis, Geneva Papers on Risk and Insurance Theor.v 18, 131-146. MacCrimmon, K. & Larsson, S. (1979). Utility Theory: Axioms Versus "Paradoxes", in [2]. Machina, M. (1982). "Expected Utility" analysis without the independence axiom, Econometrica 50, 277-323. Machina, M. (1983). Generalized expected utility analysis and the nature of observed violations of the independence axiom, in Foundations of Utility and Ri.•Jk Theor.v with Applications, B. Stigum & F. Wenstop, eds, D. Reidel Publishing, Dordreeht. Machina, M. (1987). Choice under uncertainty: problems solved and unsolved, Journal of Economic Perspectives 1, 121-154. Machina, M. (1995). Non-expected utility and the robustness of the classical insurance paradigm, Geneva Papers on Risk and Insurance Theory 20, 9-50. Markowitz, H. (1952). The utility of wealth, Journal of Political Economy 60, 151-158. Pratt, J. (1964). Risk aversion in the small and in the large, Econometrica 32, 122-136. Quiggin, .1. (1982). A theory of anticipated utility. Journal ql Economic Behavior and Organization 3, 323-343. Rothschild, M. & Stiglitz, J. ( 1970). Increasing risk: I. A definition, Journal of Economic Theory 2, 225-243. Rothschild, M. & Stiglitz, J. (1971 ). Increasing risk: II. Its economic consequences, Journal of Economic Theory 3, 66-84. Savage, L. (1954). The Foundations of Statistics, John Wiley & Sons, New York. Schlesinger, H. (1997). Insurance demand without the expected utility paradigm, Journal of Risk and Insurance 64, 19-39.

Chapter 14. Nonexpected Utility Theory

259

Nonexpected Utility Theory [33]

[34]

[351

[36]

[37]

Segal, U. (1984). Nonlinear Decision Weights with the Independence Axiom, Manuscript, University of California, Los Angeles. Starmer, C. (2000). Developments in non-expected utility theory: the hunt for a descriptive theory of choice under risk, Journal o.f EconomJc Literature 38, 332-382. Sugden, R. (1986). New developments in the theory of choice under uncertainty, Bulletin of Economic Research

38. 1-24. von Neumann, J. & Morgenstern, 0. (1947). Theory of' Games and Economic Behavior, 2nd Edition, Princeton University Press, Princeton. Wang, T. (1993). Lp-Frechet differentiable preference and "local utility" analysis. Journal of Economic Theon-' 61, 139-159.

[38]

[39] [40]

7

Weber, M. & Camerer, C. (1987). Recent developments in modeling preferences under risk, OR Spektrum 9, 129-151. Yaari, M. (1987). The dual theory of choice under risk, Hconometrica 55, 95- 115. Young, V. & Browne, M. (2000). Equilibrium in competitive insurance markets under adverse selection and Yaari's dual theory of risk, Geneva Papers on Risk and Insurance Theory 25, 141-157.

(See also Risk Measures) MARK

J.

MACHINA

This page intentionally left blank

Chapter 15 Science,Vol.I85,No.4157, 1974,1124-1131

Judgment under Uncertainty: Heuristics and Biases Biases in judgments reveal some heuristics of thinking under uncertainty. Amos Tversky and Daniel Kahneman

Many decisions are based on beliefs concerning the likelihood of uncertain events such as the outcome of an election, the guilt of a defendant, or the future value of the dollar. These beliefs are usually expressed in statements such as "I think that . . . ," "chances are . . . , " "it is unlikely that . . . ," and so forth. Occasionally, beliefs concerning uncertain events are expressed in numerical form as odds or subjective probabilities. What determines such beliefs? How do people assess the probability of an uncertain event or the value of an uncertain quantity? This article shows that people rely on a limited number of heuristic principles which reduce the complex tasks of assessing probabilities and predicting values to simpler judgmental operations. In general, these heuristics are quite useful, but sometimes they lead to severe and セケウエ・ュ。ゥ」@ errors. The subjective assessment of probability resembles the subjective assessment of physical quantities such as distance or size. These judgments are all based on data of limited validity, which are processed according to heuristic rules. For example, the apparent distance of an object is determined !n part by its clarity. The more sharply the object is seen, the closer it appears to be. This rule has some validity, M「・」。オウセ@ in any given scene the more distant objects are seen less sharply than nearer objects. However, the reliance on this rule leads to systematic errors in the estimation of distance. Specifically, distances are often overestimated when visibility is poor because the contours of objects are blurred. On the other hand, distances are often underestiThe authors are members of the department of psychology at the Hebrew University, Jerusalem, Tsrael.

mated when visibility is good because the objects are seen sharply. Thus, the reliance on clarity as an indication of distance leads to common biases. Such biases are also found in the intuitive judgment of probability. This article describes three heuristics that are employed to assess probabilities and to predict values. Biases to which these heuristics lead are enumerated, and the applied and theoretical implications of these observations are discussed.

Representativeness Many of the probabilistic questions with which people are concerned belong to one of the following types: What is the probability that object A belongs to class B? What is the probability that event A originates from process B? What is the probability that process B will generate event A? In answering such questions, people typically rely on the representativeness heuristic, in which probabilities are evaluated by the degree to which A is representative of B, that is, by the degree to which A resembles B. For example, when A is highly representative of B, the probability that A originates from B is judged to be high. On the other hand, if A is not similar to B, the probability that A originates from B is judged to be low. For an illustration of judgment by representativeness, consider an individual who has been described by a former neighbor as follows: "Steve is very shy and withdrawn, invariably helpful, but with little interest in people, or in the world of reality. A meek and tidy soul, he has a need for order and structure, and a passion for detail." How do people assess the probability that Steve is engaged in a particular

occupation from a list of possibilities (for example, farmer, salesman, airline pilot, librarian, or physician)? How do people order these occupations from most to least likely? In the representativeness heuristic, the probability that Steve is a librarian, for example, is assessed by the degree to which he is representative of, or similar to, the stereotype of a librarian. Indeed, research with problems of this type has shown that people order the occupations by probability and by similarity in exactly the same way (J). This approach to the judgment of probability leads to serious errors, because similarity, or representativeness, is not mftuenced by several factors that should affect judgments of probability. Insensitivity to prior probability of outcomes. One of the factors that have no effect on representativeness but should have a major effect on probability is the prior probability, or base-rate frequency, of the outcomes. In the case of Steve, for example, the fact that there are many more farmers than librarians in the population should enter into any reasonable estimate of the probability that Steve is a librarian rather than a farmer. Considerations of base-rate frequency, however, do not affect the similarity of Steve to the stereotypes of librarians and farmers. If people evaluate probability by representativeness, therefore, prior probabilities will be neglected. This hypothesis was tested in an experiment where prior probabilities were manipulated (1). Subjects were shown brief personality descriptions of several individuals, allegedly sampled at random from a group of 100 professionals-engineers and lawyers. The subjects were asked to assess, for each description, the probability that it belonged to an engineer rather than to a lawyer. In one experimental condition, subjects were told that the group from which the descriptions had been drawn consisted of 70 engineers and 30 lawyers. In another condition, subjects were told that the group consisted of 30 engineers and 70 lawyers. The odds that any particular description belongs to an engineer rather than to a lawyer should be higher in the first condition, where there is a majority of engineers, than in the second condition, where there is a majority of lawyers. Specifically, it can be shown by applying Bayes' rule that the ratio of these odds should be (.7/.3)2, or 5.44, for each description. In a sharp violation of Bayes' rule, the subjects in the two conditions produced essenSCIENCE, VOL, 185

1124

261

262

A Tversky and D Kahneman

tially the same probability judgments. Apparently, subjects evaluated the likelihood that a particular description be-

sample size. Indeed, when subjects assessed the distributions of average height for samples of various sizes,

longed to an engineer rather than to a

they produced identical distributions. For example, the probability of obtaining an average height greater than 6 feet was assigned the same value for samples of 1000, 100, and 10 men (2).

lawyer by the degree to which this description was representative of the two stereotypes, with little or no regard

for the prior probabilities of the categories.

The subjects used prior probabilities correctly when they had no other information. In the absence of a personality sketch, they judged the probability that an unknown individual is an engineer

to be .7 and .3, respectively, in the two base-rate conditions. However,

prior

probabilities were effectively ignored when a description was introduced, even when thjs description was totally uninformative. The responses to the

following description illustrate this phenomenon: Dick is a 30 year old man. He is married with no children. A man of high ability and high motivation, he promises

to be quite successful in his field. He is well liked by his colleagues. This description was intended to convey no information relevant to the question of whether Dick is an engineer or a

lawyer. Consequently, the probability that Dick is an engineer should equal the proportion of engineers in the group, as if no description had been given. The subjects, however, judged the probability of Dick being an engineer to be .5 regardless of whether the stated proportion of engineers in the

group was .7 or .3. Evidently, people respond differently when given no evidence and when given worthless evidence. When no specific evidence is given, prior probabilities are properly utilized; when worthless evidence is given, prior probabilities are ignored (I). Insensitivity to sample size. To eval-

uate the probability of obtaining a particular result in a sample drawn from

a specified population, people typically apply the representativeness heuristic.

That is, they assess the likelihood of a sample result, for example, that the average height in a random sample of ten men will be 6 feet (180 centimeters), by the similarity of this result to the corresponding parameter (that is, to the average height in the population of men). The similarity of a sample statistic to a population parameter does not depend on the size of the sample. Consequently, if probabilities are assessed by representativeness, then the judged probability of a sample statistic will be essentially independent of 21 SEPTEMBER 1974

In this problem, the correct posterior odds are 8 to 1 for the 4 : 1 sample and 16 to 1 for the 12 : 8 sample, assuming equal prior probabilities. However, most people feel that the first sample provides much stronger evidence

for the hypothesis that the urn is predominantly red, because the proportion

of red balls is larger in the first than in

Moreover, subjects failed -to appreciate the role of sample size even when it was emphasized in the formulation of

judgments are dominated by the sample

the problem. Consider the following

proportion and are essentially unaffected

question: A certain town is served by two hospitals. In the larger hospital about 45 babies are born each day, and in the smaller hospital about 15 babies are born each day. As you know, about 50 percent of all babies are boys. However, the exact percentage varies from day to day. Some· times :i.t may be higher than 50 percent, sometimes lower. For a period of 1 year, each hospital recorded the days on which more than 60 percent of the babies born were boys.

Which hospital do you think recorded more such days? セ@ セ@

The larger hospital (21) 'Jlhe smaller hospital (21)

,.. About the same (that is, within 5

percent of each other) (53) The values in parentheses are the ョオュセ@ ber of undergraduate students who chose each answer.

Most subjects judged the probability of obtaining more than 60 percent boys to be the same in the small and in the large hospital, presumably because these events are described by the same statistic and are therefore equally representative of the general population. In contrast, sampling theory entails that the expected number of days on which more than 60 percent of the babies are boys is much greater in the small hospital than in the large one, because a large sample is less likely to stray from 50 percent. This fundamental notion of statistics is evidently not part of people's repertoire of intuitions. A similar insensitivity to sample size

has been reported in judgments of posterior probability, that is, of the probability that a sample has been drawn from one population rather than from another. Consider the following example: Imagine an urn fiiled with 「。ャウセ@ of which % are of one color and lh of another. One individual has drawn 5 balls .from the urn, and found that 4 were red and 1 was white Another individual has drawn 20 balls and found that 12 were red and 8 were white. Which of the two individuals should feel more confident that the urn contains 2h red balls and V3 white balls, rather than the opposite? What odds should each individual give?

the second sample. Here again, intuitive

by the size of the sample, which plays a crucial role in the determination of

the actual posterior odds (2). In addition, intuitive estimates of posterior odds are far less extreme than the cor· rect values. The underestimation of the impact of evidence has been observed repeatedly in problems of this type (3, 4). It has been labeled "conservatism." Misconceptions of chance. People exM pect that a sequence of events generated

by a random process will represent the essential characteristics of that process even when the sequence is short. In considering tosses of a coin for heads

or tails, for example, people regard the sequence H-T·H-T-T-H to be more likely than the sequence H-H-H-T-T-T, which does not appear random, and

also more likely than the sequence H-HH-H-T-H, which does not represent the fairness of the coin (2). Thus, people expect that the essential characteristics

of the process will be represented, not only globally in the entire sequence, but also locally in each of its parts. A locally representative sequence, how· ever, deviates systematically from chance expectation: it contains too many al· ternations and too few runs. Another

consequence of the belief in local representativeness is the well-known gam·

bier's fallacy. After observing a long run of red on the roulette wheel. for example, most people erroneously be-

lieve that black is now due, presumably because the occurrence of black will result in a more representative sequence

than the occurrence of an additional red. Chance is commonly viewed as a self-correcting -process in which a devi· ation in one direction induces a deviation in the opposite direction to restore the equilibrium. In fact. deviations are not "corrected" as a chance process

unfolds, they are merely diluted. Misconceptions of chance are not

limited to naive subjects. A study of the statistical intuitions of experienced

research psychologists (5) revealed a lingering belief in what may be called the "law of small numbers," according

to which even small samples are highly 1125

Chapter 15. Judgment Under Uncertainty: Heuristics and Biases representative of the populations from which they are drawn. The responses of these investigators reflected the expectation that a valid hypothesis about a population will be represented by a statistically significant result in a sample-with little regard for its size. As a consequence, the researchers put too much faith in the results of small samples and grossly overestimated the replicability of such results. In the actual conduct of research, this bias leads to the selection of samples of inadequate size and to overinterpretation of findings. Insensitivity to predictability. People are sometimes called upon to make such numerical predictions as the future value of a stock, the demand for a commodity, or the outcome of a football game. Such predictions are often made by representativeness. For example, suppose one is given a description of a company and is asked to predict its future profit. If the description of the company is very favorable, a very high profit will appear most representative of that description; if the description is mediocre, a mediocre performance will appear most representative. The degree to which the description is favorable is unaffected by the reliability of that description or by the degree to which it permits accurate prediction. Hence, if people predict solely in terms of the favorableness of the description, their predictions will be insensitive to the reliability of the evidence and to the expected accuracy of the prediction. This mode of judgment violates the normative statistical theory in which the extremeness and the range of predictions are controlled by considerations of predictability. When predictability is nil, the same prediction should be made in all cases. For example, if the descriptions of companies provide no information relevant to profit, then the same value (such as average profit) should be predicted for all companies. If predictability is perfect, of course, the values predicted will match the actual values and the range of predictions will equal the range of outcomes. In general, the higher the predictability, the wider the range of predicted values. Several studies of numerical prediction have demonstrated that intuitive predictions violate this rule, and that subjects show little or no regard for considerations of predictability (I). In one of these studies, subjeczts were presented with several paragraphs, each describing the performance of a stu1126

dent teacher during a particular practice lesson. Some subjects were asked to evaluate the quality of the lesson described in the paragraph in percentile scores, relative to a specified population. Other subjects were asked to predict, also in percentile scores, the standing of each student teacher 5 years after the practice lesson. The judgments made under the two conditions were identical. That is, the prediction of a remote criterion (success of a teacher after 5 years) was identical to the evaluation of the information on which the prediction was based (the quality of the practice lesson). The students who made these predictions were undoubtedly aware of the limited predictability of teaching competence on the basis of a single trial lesson 5 years earlier; nevertheless, their predictions were as extreme as their evaluations. The illusion of validity. As we have seen, people often predict by selecting the outcome (for example, an occupation) that is most representative of the input (for example, the description of a person). The confidence they have in their prediction depends primarily on the degree of representativeness (that is, on the quality of the match between the selected outcome and the input) .with little or no regard for the factors that limit predictive accuracy. Thus, people express great confidence in ·the prediction that a person is a librarian when given a description of his personality which matches the stereotype of librarians, even if the description is scanty, unreliable, or outdated. The unwarranted confidence which is produced by a good fit between the predicted outcome and the input information may be called the illusion of validity. This illusion persists even when the judge is aware of the factors that limit the accuracy of his predictions. It is a common observation that psychologists who conduct selection interviews often experience considerable confidence in their predictions, even when they know of the vast literature that shows selection interviews to be highly fallible. The continued reliance on the clinical interview for selection, despite repeated demonstrations of its inadequacy, amply attests to the strength of this effect. The internal consistency of a pattern of inputs is a major determinant of one's confidence in predictions based on these inputs. For example, people express more confidence in predicting the final grade-point average of a student

263

whose first-year record consists entirely of B's than in predicting the gradepoint average of a student whose firstyear record includes many A's and C's. Highly consistent patterns are most often observed when the input variables are highly redundant or correlated. Hence, people tend to have great confidence in prediciions based on redundant input variables. However, an elementary result in the statistics of correlation asserts that, given input variables of stated validity, a prediction based on several such inputs can achieve higher accuracy when they are independent of each other than when they are redundant or correlated. Thus, redundancy among inputs decreases accuracy even as it increases confidence, and people are often confident in predictions that are quite likely to be off the mark (I) . Misconceptions of regression. Suppose a large group of children has been examined on two equivalent versions of an aptitude test. If one selects ten children from among those who did best on one of the two versions, he will usually find their performance on the second version to be somewhat disappointing. Conversely, if one selects ten children from among those who did worst on one version, they will be found, on the average, to do somewhat better on the other version. More generally, consider two variables X and Y which have the same distribution. If one selects individuals whose average X score deviates from the mean of X by k units, then the average of their Y scores will usually deviate from the mean of Y by less than k units. These observations illustrate a general phenomenon known as regression toward the mean, which was first documented by Galton more than 100 years ago. In the normal course of life, one encounters many instances of regression toward the mean, in the comparison of the height of fathers and sons, of the intelligence of husbands and wives. or of the performance of individuals on consecutive examinations. Nevertheless, people do not develop correct intuitions about this phenomenon. First, they do not expect regression in many contexts where it is bound to occur. Second, when they recognize the occurrence of regression, they often invent spurious causal explanations for it (I). We suggest that the phenomenon of regression remains elusive because it is incompatible with the belief that the predicted outcome should be maximally SCIENCE, VOL. 185

264 representative of the input, and, hence, that the value of the outcome variable should be as extreme as the value of the input variable. The failure to recognize the import of regression can have pernicious consequences, as illustrated by the following observation ( 1). In a discussion of flight training, experienced instructors noted that praise for an exceptionally smooth landing is typically followed by a poorer landing on the next try, while harsh criticism after a rough landing is usually followed by an improvement on the next try. The instructors concluded that verbal rewards are detrimental to learning, while verbal punishments are beneficial, contrary to accepted psychological doctrine. This conclusion is unwarranted because of the presence of regression toward ,the mean. As in other cases of repeated examination, an improvement will オウセ@ ally follow a poor performance and a deterioration will usually follow an outstanding performance, even if the instructor does not respond to the trainee's achievement on the first attempt. Because the instructors had praised their trainees after good landings and admonished them after poor ones, they reached the erroneous and potentially harmful conclusion that punishment is more effective than reward. Thus, the failure to understand the effect of regression leads one to overestimate the effectiveness of punishment and to underestimate the effectiveness of reward. In social interaction, as well as in training, rewards are typically administered when performance is good, and punishments are typically administered when performance is poor. .By regression alone, therefore, behavior is most likely to improve after punishment and most likely to deteriorate after reward. Consequently, the human condition is such that, by chance alone, one is most often rewarded for punishing others and most often punished for rewarding them. People are generally not aware of this contingency. In fact, the elusive role of regression in determining the apparent consequences of reward and punishment seems to have escaped the notice of students of this area.

Availability

There are situations in which people assess the frequency of a ciass or the probability of an event by the ease with 27 SEPTEMBER 1974

A Tversky and D Kahneman which instances or occurrences can be brought to mind. For example, one may assess the risk of heart attack among middle-aged people by recalling such occurrences among one's acquaintances. Similarly, one may evaluate the probability that a given business venture will fail by imagining various difficulties it could encounter. This judgmental heuristic is called availability. Availability is a useful clue for assessing frequency or probability, because instances of large classes are usually recalled better and faster than instances of less frequent classes. However, availability is affected by factors other than frequency and probability. Consequently, the reliance on availability leads to predictable biases, some of which are illustrated below. Biases due to the retrievability of ゥョセ@ stances. When the size of a class is judged by the availability of its instances, a class whose instances are easily retrieved will appear more numerous than a class of equal frequency whose instances are less retrievable. In an elementary demonstration of this ・ヲセ@ feet, subjects heard a list of well-known personalities of both sexes and were subsequently asked to judge whether the list contained more names of men than of women. Different lists were presented to different groups of subjects. In some of the lists the men were relatively more famous than the women, and in others the women were relatively more famous than the men. In each of the lists, the subjects erroneously judged that the class (sex) that had the more famous personalities was the more numerous (6).

In addition to familiarity, there are other factors, such as salience, which affect the retrievability of instances. For example, the impact of seeing a house burning on the subjective probability of such accidents is probably greater than the impact of reading about a fire in the local paper. Furthermore, recent occurrences are likely to be relatively more available than earlier occurrences. It is a common experience that the subjective probability of traffic accidents rises temporarily when one sees a car overturned by the side of the road. Biases due to the effectiveness of a search set. Suppose one samples a word (of three letters or more) at random from an English text. Is it more likely that the word starts with r or that r is the third letter? People approach this problem by recalling words that

begin with r (road) and words that have r in the third position (car) and assess the relative frequency by the ease with which words of the two types come to mind. Because it is much easier to search for words by their first letter than by their third letter, most people judge words that begin with a given consonant to be more numerous than words in which the- same consonant apM pears in the third position. They do so even for consonants, such as r or k, that are more frequent in the third position than in the first (6). Different tasks elicit different search sets. For example, suppose you are asked to rate the frequency with which abstract words (thought, love) and concrete words (door, water) appear in written English. A natural way to answer this question is to search for contexts in which the word could appear. It seems easier to think of contexts in which an abstract concept is mentioned (love in love stories) than to think of contexts in which a concrete word (such as door) is mentioned. If the frequency of words is judged by the availability of the contexts in which they appear, abstract words will be judged as relatively more numerous than concrete words. This bias has been observed in a recent study (7) which showed that the judged frequency of occurrence of abstract words was much higher than that of concrete words, equated in objective frequency. Abstract words were also judged to appear in a much greater variety of contexts than concrete words. Biases of imaginability. Sometimes one has to assess the frequency of a class whose instances are not stored in memory but can be generated according to a given rule. In such situations, one typically generates several instances and evaluates frequency or probability by the ease with which {he relevant instances can be constructed. However, the ease of constructing instances does not always reflect their actual frequency, and this mode of evaluation is prone to biases. To illustrate, consider a group of I 0 people who form committees of k members, 2 :=:;; k ::;; 8. How many different committees of k members can be formed? The correct answer to this problem is given by the binomial coefficient (lg) which reaches a maximum of 252 for k = 5. Clearly, the number of committees of k members equals the number of committees of (10- k) members, because any committee of k 1127

Chapter 15. Judgment Under Uncertainty: Heuristics and Biases members defines a unique group of (10- k) nonmembers. One way to answer this question without computation is to mentally construct committees of k members and to evaluate ,their number by the ease with which they come to mind. Committees of few members, say 2, are more available than committees of many members, say 8. The simplest scheme for the construction of committees is a partition of the group into disjoint sets. One readily sees that it is easy to construct five disjoint committees of 2 members, while it is impossible to generate even two disjoint committees of 8 members. Consequently, if frequency is assessed by imaginability, or by availability for construction, the small commiltces will appear more numerous than larger committees, in contrast to the correct bell-shaped function. Indeed, when naive subjects were asked to estimate the number of distinct committees -of various sizes, their estimates were a decreasing monotonic function of committee size (6). For example, the median estimate of the number of committees of 2 members was 70, while the estimate for committees of 8 members was 20 (the correct answer is 45 in both cases). Tmaginability plays an important role in the evaluation of probabilities in reallife situations. The risk involved in an adventurous expedilion, for example, is evaluated by imagining contingencies with which the expedition is not equipped to cope. If many such difficulties are vividly portrayed, the expedition can be made to appear exceedingly dangerous, although the case with which disasters are imagined need not reflect their actual likelihood. Conversely, the risk involved in an undertaking may be grossly underestimated if some possible dangers are either difficult to conceive of, or simply do not come to mind. Illusory correlation. Chapman and Chapman (8) have described an interesting bias in the judgment of the frequency with which two events co-occur. They presented naive judges with information concerning several hypothetical mental patients. The data for each patient consisted of a clinical diagnosis and a drawing of a person made by the patient. Later the judges estimated the frequency with which each diagnosis (such as paranoia or suspiciousness) had been accompanied by various features of the drawing (such as peculiar eyes). The subjects markedly overestimated the frequency of co-occurrence of H28

natural associates, such as suspiciOusness and peculiar eyes. This effect was labeled illusory correlation. In their erroneous judgments of the data to which they had been exposed, naive subjects "rediscovered" much of the common, but unfounded, clinical lore concerning the interpretation of the draw-aperson test. The illusory correlation effect was extremely resistant to contradictory data. It persisted even when the correlation between symptom and diagnosis was actually negative, and it prevented the judges from detecting relationships that were in fact present. Availability provides a natural account for the illusory-correlation effect. The judgment of how frequently two events co-occur could be based on the strength of the associative bond between them. When the association is strong, one is likely to conclude that the events have been frequently paired. Consequently, strong associates will be judged to have occurred together frequently. According to this view, the illusory correlation between suspiciousness and peculiar drawing of the eyes, for example, is due to the fact that suspiciousness is more readily associated with the eyes than with any other part of the body. Lifelong experience has taught us that, in general, instances of large classes are recalled better and faster than instances of less frequent classes; that likely occurrences are easier to imagine than unlikely ones; and that the associative connections between events are strengthened when the events frequently co-occur. As a result, man has at his disposal a procedure (the availability heuristic) for estimating the numerosity of a class, the likelihood of an event, or the frequency of co-occurrences, by the ease with which the relevant mental operations of retrieval, construction, or association can be performed. However, as the preceding examples have demonstrated, this valuable estimation procedure results in systematic errors.

Adjustment and Anchoring

In many si-tuations, people make estimates by starting from an initial value that is adjusted to yield the final answer. The initial value, or starting point, may be suggested by the formulation of the problem, or it may be the result of a partial computation. In either case, adjustments arc typically insufficient (4).

265

That is, different starting points yield different estimates, which are biased toward the initial values. We call this phenomenon anchoring. Insufficient adjustment. In a demonstration of the anchoring effect, subjects were asked to estimate various quantities, stated in percentages (for example, the percentage of African countries in the United Nations). For each quantity, a number between 0 and 100 was determined by spinning a wheel of fortune in the subjects' presence. The subjects were instructed to indicate first whether that number was higher or lower than the value of the quantity, and then to estimate the value of the quantity by moving upward or downward from the given number. Different groups were given different numbers for each quantity, and these arbitrary numbers had a marked effect on estimates. For example, the median estimates of the percentage of African countries in the United Nations were 25 and 45 for groups that received 10 and 65, respectively, as starting points. Payoffs for accuracy did not reduce the anchoring effect. Anchoring occurs not only when the starting point is given to the subject, but also when the subject bases his estimate Dn the result of some incomplete computation. A study of intuitive numerical estimation illustrates this effect. Two groups of high school students estimated, within 5 seconds, a numerical expression that was written on the blackboard. One group estimated the product 8X7X6X5X4X3X2Xl

while another product

group

estimated the

IX2>O for some u in U. In this section, U is the class of all non-decreasing functions (assumed to have finite values for any finite value of x). 1 The variables X and Y 2 are defined here as the money payoffs of a given venture, which are additions (or reductions) to the individual's (constant) wealth, so that no restriction to positive values of X or Y is necessary. (The utility function u depends, of course, on the individual's initial wealth position). A sufficient condition for FDG (or X D Y) is an efficiency criterion. 3 The set of all risks which are not dominated by another risk according to the given criterion, is an efficient set. The weaker the sufficient condition, the smaller the efficient set. The minimal efficient set E is generated when the criterion is optimal, that is, when the condition for FDG is both sufficient and necessary. If the class of admissible utility functions is restricted to a proper subset of U, the minimal efficient set may be reduced to a subset of E, and the corresponding optimal efficiency criterion may be weakened. Before proceeding to state and prove the general criterion, we need the following: Lemma 1. Let G, F, be two (cumulative) distributions, and u(x) a non-decreasing function, with finite values for any finite x; then

t1Eu = EFu(x)-Eau(x) = Proof By definition, 5

t1Eu =

=

[f

f

d(u.F)-

[G(x)-F(x)]du(x). 4

f f f [J f

t1Eu =

Integrating (1) by parts gives:

f

udF-

Fdu]-

d[u .(F-G)]+

...(1)

udG.

d(u.G)-

f

Gdu]

(G-F)du.

In order to show that the first term on the right vanishes, we define a sequence of functions u.(x), converging to u(x): u( -n) for x< -n { u.(x) = u(x) -n セ@ x セ@ n u(n)

x>n.

I Quirk and Saposnik [18] proved a similar criterion for the discrete, finite case, and assumed a strictly monotone utility. They also sketched a proof for the (pure) continuous case, requiring that u(x) be bounded and piecewise differentiable (p. 144). All these requirements are not necessary. Similar comments apply to Hadar and Russel [7), who assumed u(x) to be twice continuously differentiable, and restricted to a finite range. Cf. also Hammond [8]. 2 In much of the following discussion, we do not distinguish in notation between the variables X and Y, restricting our attention to the two distributions, F(x) and G(x). 3 An additional requirement is the transitivity of the sufficient condition; but this is obviously assured by the given definition of dominance, since EFu-EGu セ@ 0 and EGu-EHu セ@ 0, imply EFu-EHu セ@ 0, andFDH. 4 The integrals throughout are Stieltjes-Lebesgues integrals, ranging on all real values of x, unless specified otherwise; cf. Cramer [4), p. 62. 5 The arguments in the functions appearing in integrals are omitted in cases where no misunderstanding should arise.

Chapter 17. The Efficiency Analysis of Choices Involving Risk

EFFICIENCY ANALYSIS OF CHOICES INVOLVING RISK

289

337

But now,

f

d[u(F-G)] =

Nセ@

f d[u.(F-G)]

=lim {u(n)[F(oo)-G(oo)]-u(-n)(F(-oo)-G(-oo)]}

= Hence t.Eu

lim {u(n).O-u( -n).O}

= 0.

= f (G- F)du, if the integral exists. 1 Q.E.D.

The optimal criterion for FDG is given in Theorem 1.

Theorem 1. Let F, G and u be as in Lemma 1. A necessary and sufficient condition for FDG is: F(x)

セ@

G(x)for every x, and F(x 0 )0, due to the right-continuity ofF and G, there is an interval x 0 セク\@ x 0 +[3 where G(x)-F(x)>O. To show that there exist some u 0 for which t.Eu 0 >0, choose u0 (x) as follows: U0 (x) =

{

Xo

xセ@

X

X 0 セx@

Xo X 0 +f3

x 0 +f3 X'?;._ x 0 +f3.

Then

Uo E

u, and f

(G-F)du =

r:+/l

(G-F)dx>O.

(b) The necessity is proved similarly. If, for some xi, G(xi)-F(x 1) xi +e), where G(x)-F(x)O, and another function v where ilEv Ea Y, i.e., a larger mean value of X, is a necessary condition for dominance of X. The variance, however, plays no direct role in the efficiency criterion. For example, consider two random variables with rectangular distributions, i.e. X has a constant probability density in the range x セ@ x 2 , and Y a constant density in the range y 1 セ@ y セ@ Yz. 1 If x 1 > y 1 and x 2 > y 2 , X 1 セ@ X is preferred to Y with every possible utility function. The variance of X (which equals / 2 (x 2 - x 1) 2 ) may be much larger than var Y, and the degree of risk aversion implied by u(x) may be as high as one wants to assume, still X is better than Y, if individuals do not prefer less money to more money (i.e., if u(x) is non-decreasing). If one wishes to consider utility functions which vary only in a bounded range, 2 it can easily be verified that the optimal criterion for dominance of F over G, is that F(x) セ@ G(x) for all x and F(x 0 )E6 (Y)

=

s:

(G-F)dt

0 0·21 0·20

0·25,

varp (X)= 0·4475>var6 (Y) = 0·4075, but X is preferred to Y for any concave utility function. Since F(l)>G(1), the general criterion of theorem 1 is not satisfied. Indeed, choosing a convex u(x) = l0X 2 , we get Y preferred to X: !J.Eu = Epu(x)-E0 u(y) セ@ 1003-1101O], exceeds the area to the right, where (G-F)p,2 ; For all x, x-p, 1 < x-p, 2 ; thus F(x) セ@ G(x) 0' 0' for all x, and there is no intersection point. F dominates G by Theorem 1. If p, 1 = p, 2 , F and G are identical. ·

If u 1 #= u2 , we have an intersection point at x 0 , where

(a) If u1 >u 2 , then for xx0 ,

x-p,

1 -0'1

セ@

x-p,2

--,

x-p,

1 -0'1

and F(x)

0'2

セ@

x-p,

> - -2 , and fHクIセg@ 0'2

(F>G for x1); and

G(x). Thus, the condition of Theorem 3 is not

satisfied, and F cannot dominate G. (b) If u 1 x-p,2, and F セ@ G. Since FO, cf>"(x)O, cf>"(x) 0.

A. General Efficiency Criterion

In the most general case, where investors are assumed to have no systematic preferences with respect to risk (i.e., no restrictions are placed on investors' utility function beyond the assumption that their first derivatives be nonnegative), it has been proved (see articles by James Quirk and Rubin Saposnik, Joseph Hadar and William Russell, and Giora Hanoch and Levy) that option F eliminates option G from the efficient set (i.e., F dominates G, or F DG) for all (utility) functions cf>(x) (where

f_:

[G(I) -

F(t)]q,'(t)dt?:: 0

functions, is: 8 F;DGi (i= I, 2), where F, and G; denote the single period distributions. PROOF: Let x 1 and x2 be two continuous random variables for periods 1 and 2, respectively, and x = x 1x 2 is the random variable (return) for the two periods. Since F,DG; (i= 1, 2), it is given that: for every x 1

and F 1 (x 10 )

for every nondecreasing concave ¢(x).

II. Multi-Period Dominance in the Case of an Unrestricted Utility Function

In the rest of the paper we shall prove several theorems, first for the two-period horizon, and then the results will be generalized for an n-period horizon. Like Tobin, we assume independence over time. More precisely, we accept the Random Walk Theory (see, for example, Paul Cootner). Let R; denote the return on one dollar invested in period i. The final wealth for an n-period horizon is therefore given by 1 R ]itセ@ (l+R;). Denoting 7 1+R;=x;, the joint density function ッヲャセ@ X; isf(x,, ... , Xn)- Accepting the Random Walk hypothesis, the joint density function can be written as f(x,, ... , Xn-t)f(xn), since f(xn) is independent of f(x 1, . . . , x,_1). For the twoperiod case, independence implies f(x 1 , x2 ) = f(x,)f(x,).

+

THEOREM 1: A sufficie,nt condition for a two-period risk F(x) to dominate another twoperiod risk G(x), for all nondecreasing utility

( 4)

F 2 (x 2)

< G1 (x10 )

for some x 10

G,(x,)

for every x 2

:::;

and F 2 (x2 ,) < G2 (x2 ,)

for some value x20

Assuming independence over time, the distribution function of the two-period return is given by: (5)

F(x)

P,(x :::; x)

=

i"' fo

=

x/xr[t(t,)j,(t,)dt,dt,

Jo"' F 2 (xjt 1 )f1 (t 1 )dt 1

=

Similarly, the distribution function of option G is given by:

(6)

G(x)

=

i"'

G 2 (x/t 1 )g 1 (t 1 )dtt,

where f;, g; stand for the density functions of the two risks in period i. We must prove that the two-period distribution F, dominates the two-period distribution G, i.e., F(x) :s;G(x) for every x, and F(xo) 0 for some value (x 0/t 10), and hence there is a strict inequality in (7) at least for one value. Thus, F(x) S G(x) and F(x 0 ) < G(x 0) for some value x 0 which implies that the two-period distribution F dominates the two-period distribution G.

COROLLARY 1: If, in addition to the assumption of independence, we also assume stationarity over time of the distributions under consideration, one may conclude that for any two horizons, n 1 and n 2 (n, > n 1), the number of options in the e_flicient set which is appropriate to horizon n 2 is not larger than the number of options in the efficient set which is appropriate to horizon n 1, where the e_flicient sets are defined for all nondecreasing utility 0 Note that in Theorem 1 we did not assume stationarity over time of the distributions. The 」ッュー。イゥセョ@ of the pair (Fn-I, Fn) with the pair (Gn-l, Gn) does not violate the conditions under which Theorem 1 holds.

Chapter 18. Stochastic Dominance, Efficiency Criteria, and Efficient Portfolios 990

THE AMERICAN ECONOMIC REVIEW

junctions. T/u:s conclusion stems directly from Theorem 1. Suppose we have two options, F and G.

The stationarity assumption implies that, F1=F2 ... =Fn and G1=G2= ... Gn. For the sake of simplicity, suppose that n 1 = 1 and n 2 =n. If F 1DG1, then, by the stationarity assumption, Fi also dominates G, for all other periods i = 2, ... , n. 10 Hence, by Theorem 1, F" dominates Gn. To sum up, an option which is eliminated from the one-period efficient set, is also eliminated from the n-period efficient set. Hence, the number of options in the efficient set of the long horizon cannot exceed the number of options in the efftcient set of the short horizon. It is interesting to note that one can find numerical examples for which the number of elements in the efficient set is striclly decreasing when the horizon increases. Consider the following two-period example, where F1 = F2 and G1 = G2. Example 1 G1 (or G2)

F, (or F,)

Outcome

Probability

1 4

1/4

2

3/4

10

Outcome

Probability

1/2 1/2

G'

II. Multi-period Dominance in the Case of Risk Aversion

It has been shown that for the one-period investment case, a necessary and sufficient condition for a risk with cumulative probability distribution function F 1 to dominate one with Gt(F1DG1) as its cumulative distribution function, for all nondecreasing concave utility functions, is:

for every x, and GJ7"'Ft for some x 0 , where the subscript 1 indicates the distribution for period one. (See articles by Hadar and Russell, Hanoch and Levy, and Rothschild and Stiglitz.) We shall now prove that if F1DGt and F,DG2, then, F dominates G, where F and G are the two-period distributions. Before proceeding to state and prove a theorem for the multiperiod risk averswn case we need the following: LEMMA: Let T(x, It)= fJF(t/lt)dt, where F is some ( cumulatit•e) distribution function. Then, T' (x, It) :::; 0

at!

a T(x, 1 2

F'

Probability

Outcome

Probability

1 4 16

1/16 6/16 9/16

4 20 100

1/4 1/2 1/4

Since F 1 and G1 (or F 2 and G2) intersect, both of them are included in the one-period efficient set. But F 2 (x) :s;G 2 (x) for every x (with a strict inequality for some values) and, hence, F 2 DG 2 • This numerical example shows that portfolios which are not eliminated from the one-period efficient set might

10 The stationarity assumption implies that the number of elements in the efficient sets constructed for each single period (e.g., year) is the same.

DECEMBER 197 J

be eliminated from the multi-period efficient set. 11

Assuming independence over time, the twoperiod distributions F 2 and G 2 are:

Outcome

303

1)

---(att) 2

T"(x, It)

PROOF:

a

Jo

x

(12) =

J

x (

o

and

2: 0

F(l/tl)dl

att aF(I/

11l) dt

att

Let j(l/11) be the density function of F(l/11 ). Then, 11 The numerical example confirms that the conditions for dominance given in Theorem 1 are sufficient, but not necessary.

304

H Levy 991

LEVY: EFFICIENT PORTFOLIOS

VOL. 63 NO.5

F DC, for every nondecreasing concare utility function (cj/(x) 2_0, q:,"(x):::; 0) is:

since the integrand is always nonnegative. The proof that T"(x, t1 )2_0 is somewhat more involved:

(16)

fox; [Gi(ti)- Fi(ti)]dti

2. 0 (i

=

1, 2)

for every value x;, with a strict inequality for at least one mlue.

PROOF: We have to prove that (17) holds: (17)

(with a strict inequality for some value x 0). Using (5) and (6), equation (17) can be rewritten:

Hence, (13)

T"(x,

t

1) = -

[f)'(t/t

1)(

MエOセIH、@

(18)

+

2. fox fo"' F ,(t/tt)ft(tt)dt,dt By condition (16), for the second-period distribution we have:

c/ OエセI、@ =fox

(19)

{ェGHエLIQNOセ}、@

Integrating by parts (with respect to x), we get:

(14)

=

サ{ヲHエェLIOZjャセM f(x/t 1 )x

f 2

/1 13

-

f

X

0

0

fox G,(t/tt)dl

=fox F 2 (t/t 1 )dt

+ o(x/t 1)

where ii(x/lt)::::: 0 and o(xo/ltn) > 0 for at least one value (x 0/tt 0). Substituting (19) in (18), we obtain:

J:x [J'(t/t,)(1/l,) ·t 2 /t:]dl

=

fox fo"' G2(tjt,)g 1(t 1)dt 1dt

f)wt,)C -2t/t:)dt]

Let us expand the fust term on the right side of (13): fo xj'(tjt,).

J:x [G(t) - F(t)]dt :;::,_ 0

xf(tjt,)21/t:dt 3

f(t/t 1 )2t/t 1 dt

Substituting these results in (13), we get:

We can now prove the following theorem:

THEROEM 3: Let F and G be the cumulatit'e distributions for the two- period horizon of two alternative risks. A sufficient condition for

Since the second term is nonnegative, it will suffice to prove that the f1rst term in (20) is nonnegative. Integrating the first term in (20) by parts, with respect tot, we obtain:

305

Chapter 18. Stochastic Dominance, Efficiency Criteria, and Efficient Portfolios THE AMERICAN ECONOMIC REVIEW

992

-f

J

a

00

[(Gl(ti)-Fl(t,)]

Since

the

forward; since pn-l dominates cn-l, one can conclude that

"F,(t/tl)dt

、エQセP@

0

。エセ@

o

DECEMBER 1973

first

term equals zero, and t,) (by definition), it remains to prove that the following term is nonnegative.

f 0'F,(t/t2)dt== T(x,

(22)

fo

oo

[G1(t1)- F 1 (t 1 )]dt 1

X;

i=l

0

Xote that OセG{gエHャIMf for Q I}、ゥエセo@ every x 1 (by (16)), and -T(x, 11) is nondecreasing concave (utility) function in 11 (bY the Lemma). Applying (2') to this case we get,

Jon

II

Since by assumption of the theorem,

j[G,(t,) - FtCtt)]}

( - aT(x, It)\ dt 1 セ@ at1 )

(23)

7t-l

for every

セ@

for every x,, we may again employ Theorem 3 1 ' and prove that n

f ,_

II

_,

Xi

1

[G"(t) - F"(t)]dt

セ@

0

Il?-1

0

Hence inequality (23) holds. We also have in (20) at least one strict inequality, since for some value x 0 jl1" IJ(xo/t,.,) > 0. Hence, F DC.

THEOREM 4: Let Fn and G" be the cumulath•e distributions of two n-period risks. A sufficient condition for F" DC" for all nondecreasing conca1•e utility functions ゥウOセG@ [G J!J MfLHエャI}、[セo@ for every x, (with a strict inequality for some X; 0 ) , where i=l, 2, ... , n. PROOF: The proof is by induction and similar to that given in Theorem 2. For n= 2, it has been proved in Theorem 3 that }セG@ [G;(I;) - F.,(t;) }、エ[セ@ 0 (for every X; where i = 1, 2) implies that F 2 dominates G2, for all risk averters. Assume, that the statement of the theorem holds for n -1 period, i.e., fo' [G;(l,) Mf[HエI}、Lセo@ (for every X; where i=1, 2, ... , n- 1) implies that Fn-l dominates G"- 1 • vVe have to prove that the theorem also holds for n-periods. This is straight-

for all values X;, i.e., pn dominates Gn for all risk averters. COROLLARY 2 If we are prepared to assume stationarity mer time, in addition to independence of the returns, i.e., F 1 =F.= ... =Fn and G 1 =G 2 = ... =G,, we may conclude that t!ze number of elements in the efficient set which is constructed for all risk a1•erters does not increase when the im·estment horizon increases. This conclusion follows from Theorem 3. Suppose that in period 1, F 1 dominates G1 , i.e., Oセ Q {g Q HQIMf for all values Q HQI@ }、エセo@ x 1 , then, by the stationarity assumption, F; dominates G; for all other periods i (i = 1, 2, ... , n), and by Theorem 3, F" dominates G". In other words, any option which is eliminated from the risk averters' oneperiod efficient set is also eliminated from the n-period efficient set. The efficient set in each single period is identical, and the nperiod efficient set is no greater than the single period efficient set. Again, as in the previous case, we can find examples where the efficient set is strictly decreasing when 12 Note again that, by assumption, Xn and IJt.:i Xi are independent, i.e., by looking at the results of the first n-1 periods one cannot improve prediction with respect to the nth period.

306

H Levy

VOL. 63 NO.5

993

LEVY: EFFICIENT PORTFOLIOS

the investment horizon increases. 13 \Ve shall again use Example 1 above to illustrate this argument: F 1 does not dominate Gt (i.e., F 1 DG1) for all risk averters since,

But, also, G1 docs not dominate F 1

Hence, F 1 and G1 are included in the ッョ・セ@ period efficient set. Assuming stationarity over time F 1 = F 2 , G1 = G2, and hence also F 2 DG2 , G2 DFz, which means that Fz and Gz are included in the second period efficient set. Looking back at Example 1, we can see that, in the エキッセー・イゥ、@ horizon, F 2 (x) ::; G2 (x), for every x and for some value x a strict inequality holds. Consequently ft(G 2 (t)-F 2 (t)dt]2:0 (with a strict ゥョセ@ equality for some value x 0 ) and, hence, F 2 dominates G2 . Therefore, in this example the risk averters' efficient set decreases when the investment horizon increases. IV. Concluding Remarks

This study investigated the role of the investment horizon in determining the ・ヲゥセ@ cient set of portfolios. We found that, unlike the ッョ・セーイゥ、@ case, the efficiency criteria for the ュオャエゥセー・イッ、@ case are sufficient but not necessary for dominance. If a portfolio is eliminated from each single period efficient set it will also be eliminated from the ュオャエゥセー・イッ、@ efficient set, but the ゥョ」ャオセ@ 13 Thi;; means that the longer the horizon the smaller the efficient set. Similar results have been obtained by Nils Hakansson, who assumes that the objective ヲオョ」セ@ tion of the investors is to maximize the geometric average return over time. In Hakansson's analysis, when n-> oo, the efficient set includes only one portfolio, which is also the optimal one. It is of interest to note that Menachem Yaari, who analyzes the individual's multi-period investment-consumption decision, obtains similar results: the longer the horizon, the more certain the optimal consumption and when n-+"", the ゥョ、カセ@ ual ゥョカ・ウエュセ」ッオー@ decisions are delern1ined as with certainty.

sian of a portfolio in each single period efficient set does not guarantee that this portfolio will be included in the multi-period efficient set. Thus, an investment consultant who screens all the available portfolios should construct the efficient set for each group of investors according to their investment horizon. However, the consultant may construct the efficient set for the shortest relevant horizon, knowing that the efficient set which is appropriate for longer horizons must be a subset of (or identical to) the efficient set which has been obtained for the short horizon. N everthelcss, one should keep in mind that the suggested procedure is valid only under the case in which probability functions are stationary over time. Otherwise, the consultant can not limit his attention only to the first period distributions. The magnitude of the decrease in the efficient set resulting from an increase in the investment horizon is left to be investigated either empirically or by simulation. REFERENCES K. Arrow, Aspects of the Theory of Risk Bearing, Helsinki 1965. P. H. Cootner, The Random Character of Stock Market Prices, rev. ed., Cambridge, Mass. 1967. J. Hadar and W. R. Russell, "Rules for Ordering Uncertain Prospects," Amer. Econ. Rev., Mar. 1969, 59, 25-34. N. H. Hakansson, "Multi-Period :\1ean-Variance Analysis: Toward a General Theory of Portfolio Choice," J. Finance, Sept. 1971, 26, 85 7-84. G. Hanoch and H. Levy, "The Efficiency Analysis of Choices Involving Risk." Rev. Econ. Stud., July 1969, 36, 335-46. H. Levy and M. Sarnat, hn,estment and Portfolio Analysis, New York 1972. ---and-·--, "International Diversification of Investment Portfolios," Amer. Econ. Rev., Sept. 1970, 60, 668-7 5. J. Lintner, "Security Price, Risk, and Maximal Gains from Diversification," J. Finance, Dec. 1965, 20, 587-615. H. M. Markowitz, Portfolio Selection, New York 1959.

Chapter 18. Stochastic Dominance, Efficiency Criteria, and Efficient Portfolios 994

J.

THE AMERICAN ECONOMIC REVIEW

Mossin, "Optimal iVIulti-Period Portfolio Policies," f. Bus., Univ. Chicago, Apr. 1968, 41, 215-29. J. Pratt, "Risk Aversion in the Small and in the Large," Econometrica, Jan. 1964, 32, 122-36. J. P. Quirk and R. Saposnik, "Admissibility and Measurable Utility Functions," Rev. Econ. Stud., Feb. 1962, 29, 140-46. M. Rothschild and J. E. Stiglitz, "Increasing Risk: I. A Definition," f. Econ. Tlzeor., Sept. 1970, 2, 225-43.

307

DECEMBER 1973

W. F. Sharpe, "Capital Assets Prices: A Theory of Market Equilibrium Under Conditions of Risk," f. Finance, Sept. 1964,19, 42 5-42. J. Tobin, "The Theory of Portfolio Selection," in F. H. Hahn and F. P.R. Brechlings, eds., The Theory of Interest Rates, London 1965. M. E. Yaari, "A Law for Large Numbers in the Theory of Consumer's Choice Under Uncertainty," working pap. no. CP-330, Berkeley Center Res. Manage. Sci., Univ. California, Mar. 1971.

This page intentionally left blank

Section D

Risk Aversion and Static Portfolio Theory

310

Section D. Risk Aversion and Static Portfolio Theory

The components of the financial decision problem have been presented: assets, prices, utilities, and efficiency. However, there are important considerations in the decision models and the data inputs which affect the quality of investment decisions. One significant issue is the computability of an efficient investment strategy. A major advantage of Markowitz’ mean-variance analysis is the relative ease of computing optimal strategies and as such it is a practical technique. Ziemba, Parkan and Brooks-Hill (1974) consider the general problem max Eu(R′ x),

st . e′ x = 1, x ∈ K,

where E represents expectation with respect to the uncertainty in R′ = (R1 , . . . , Rn ), the vector of risky asset returns. K is a convex set that represents additional constraints on the choice of x. It is assumed that R is joint normal and that Tobin’s riskless asset exists, so that borrowing and lending can occur at the risk free rate r. Ziemba et al. propose a 2-step approach based on Tobin’s (1958) result that the distribution of proportions x∗ in the risky assets is independent of the utility function assuming it is concave. So the first step is to find the proportions invested in risky assets using a fractional program:   (R − re)′ x − r ′ max |e x = 1, x E K . 1 x (x′ Σx) 2 The fractional program always has a unique solution and it may be solved via Lemke’s algorithm for the linear complementarity problem or as a quadratic program. The second step is to determine the optimal ratios of risky to non-risky assets. ′ The return on the risky portfolio is R∗ ∼ N (µ′ x∗ , x∗ Σx∗ ), where (µ, Σ) are the (mean, variance) of R. The optimal mix of R∗ and the risk free asset may be found by solving max

−∞ 0. If UD is the class of non-decreasing, concave, decreasing absolute risk version utilities, then UD ⊂ U3 . Vickson (1975) considered stochastic dominance tests for decreasing absolute risk aversion. The class UD is difficult to characterize. Pratt defines a measure of relative risk xu(2) (x) , and identifies the class of constant relative risk aversion ρ∗ (x) = xρ(x) = − (1) u (x) aversion utilities (CRRA). If UC denotes the CRRA family of utilities, then UC ⊂ xu(2) (x) = α can be solved for the function u, UD . Furthermore, the equation − (1) u (x) 1 with the solution: u(x) = 1−α x1−α , α < 1. When α = 1 the utility is u(c) = log(c). The CRRA class of utilities is often used in the analysis of investment decisions. The parameter α captures the aversion to risk, and the impact of risk aversion on investment decisions can be analyzed analytically. The Arrow–Pratt risk aversion measure is local, that is, it depends on the level of wealth. Rubinstein (1973) developed a measure of global risk aversion in the context of a parameter-preference equilibrium relationship: R(x0 ) =

−x0 E{u(2) (x)} E{u(1) (x)}

where x0 is the initial wealth level and expectation is with respect to the distribution of wealth. Some properties of this measure in the context of risk aversion with changing initial wealth levels appear in Kallberg and Ziemba (1983). This measure is less tractable and less familiar than the Arrow–Pratt measure. In contrast to the Arrow–Pratt measure which is a function of wealth x, R is a constant. Except for a few special cases, R does not have a simple form. For quadratic utility, R(1) = ρ(¯ x); thus the Rubinstein measure is the Arrow–Pratt measure evaluated at the expected final wealth level. For the optimal investment problem, Kallberg and Ziemba establish an important property of the global risk measure. The investor’s problem is max Eu(R′ x),

s.t . e′ x = 1, x ∈ K,

where E represents expectation with respect to the randomness in R′ = (R1 , . . . , Rn ), the vector of risky asset returns. K is a convex set that represents additional constraints on the choice of x. Optimality of Rubinstein’s Risk Aversion Measure. Investors with the same R have the same optimal portfolios.

Section D. Risk Aversion and Static Portfolio Theory

313

They also show with empirical studies that investors with similar average ArrowPratt risk aversion have similar portfolios, regardless of their utility function. The results also yield the following conclusions: • The special exponential and negative power utility functions yield very risk averse portfolios, while the positive power (CRRA) utility function which yields highly risky portfolios and moderately risk averse portfolios for different values of α. • The arctangent utility function usually yields highly risky portfolios. • The quadratic, exponential and logarithmic utility functions yield the largest range in variation of ρ and yield the safest and riskiest portfolios. • The quadratic utility function may well play a useful role as a computational surrogate for more plausible utility functions when the number of possible investment securities is large, say n > 50. • With horizons of a year or less one can substitute easily derived surrogate utility functions that are mathematically convenient for more plausible but mathematically more complicated utility functions and feel confident that the errors produced in the calculation of the optimal portfolios are at most of the order of magnitude of the errors in the data. • Calculations indicate that the maximum expected utility and optimal portfolio composition are relatively insensitive to errors in estimation of the variancecovariance matrix. However, errors in estimating the mean return vector do significantly change these quantities. Chopra and Ziemba (1993), following the earlier papers of Kallberg and Ziemba (1981, 1984), consider the relative impact of estimation errors and the impact of risk aversion on portfolio performance. Mean percentage cash equivalent loss due to errors in means, variances and covariances in a mean-variance model are found to be in relative terms roughly 20:2:1 times as important, respectively. The error depends on the risk tolerance, the reciprocal of the Arrow–Pratt risk aversion ρ(x). With low risk aversion, like log, the ratios can be 100:2:1. So good estimates of asset return distribution moments are the most crucial aspect for successful application of a mean-variance analysis, and in all other stochastic modeling approaches. The sensitivity of the mean carries into multiperiod models. There the effect is strongest in period 1 then less and less in future periods, see Geyer and Ziemba (2008). The impact of modeling and estimation errors on forecasts for securities prices and the resulting effect on portfolio decisions and capital accumulation have been considered in other studies. Pastor and Stambaugh (1999) conclude that model error is less important than estimation error; see also Kallberg and Ziemba (1981) who conclude the same. With regard to estimation error, alternative estimates for the mean return have been considered in a long series of asset prices (Grauer and Hakansson, 1995), with improved results from shrinkage (Stein) estimators. MacKinlay and Pastor (2000) use a restriction, which incorporates the covariance of returns to calculate an estimate of expected returns which is superior to the shrinkage estimator.

314

Section D. Risk Aversion and Static Portfolio Theory

MacLean, Foster and Ziemba use a Bayesian framework to include the covariance in an estimate of the mean. In essence, the return on one asset provides information about the return on related assets, and the sharing of information through the covariance improves the quality of estimates (Jones and Shanken, 2003). The significance of quality estimates for model parameters is highlighted when those values become inputs to portfolio decisions and the accumulation of wealth over time. This process is illustrated for the portfolio selection problem with constant relative risk aversion utility. That utility enables a closed form solution which depends on the conditional expected return. With the CRRA utility, the loss in wealth resulting from estimation errors can be mitigated with the risk aversion parameter. In the decision rule, the risk aversion parameter α defines a fraction 1 of capital invested in the optimal portfolio. When α < 0, the control of deci1−α sion risk also reduces the impact of estimation error. Correspondingly, when α > 0, the over investment increases the effect of estimation error. Expected utility theory certainly captures some of the intuition for risk aversion over very large stakes. But the theory is manifestly not close to the right explanation for most risk attitudes, and does not explain the modest-scale risk aversion observed in practice. Rabin and Thaler (2001) think that the right explanation incorporates two concepts: loss aversion and mental accounting. Loss aversion is the tendency to feel the pain of a loss more acutely than the pleasure of an equal-sized gain. Loss aversion is incorporated in and Tversky and Kahneman’s prospect theory (1979), which models decision makers who react to changes in wealth, rather than levels, and are roughly twice as sensitive to perceived losses than to gains. Mental accounting, which refers to the tendency to assess risks in isolation rather than in a broader perspective. If investors focus on the long-term returns of stocks they would recognize how little risk there is, relative to bonds, and would be happy to hold stocks at a smaller equity premium. Instead, they consider short-term volatility, with frequent mental accounting losses, and demand a substantial equity premium as compensation. Rabin and Thaler argue that loss aversion and the tendency to isolate each risky choice must both be key components of a good descriptive theory of risk attitudes.

References Arrow, K (1965). The theory of risk aversion. In YJS Helsinki (Ed.), Aspects of the Theory of Risk Bearing, Chicago: Markham Publishing. Brandt, M and P Santa-Clara (2006). Dynamic portfolio selection by augmenting the asset space. Journal of Finance, 61, 2187–2217. Chopra, V and WT Ziemba (1993). The effect of errors in mean and co-variance estimates on optimal portfolio choice. Journal of Portfolio Management, 19, 6–11. Fama, EF and F French (1992). The cross-section of expected stock returns. Journal of Finance, 47(2), 427–466.

Section D. Risk Aversion and Static Portfolio Theory

315

Geyer, A and WT Ziemba (2008). The innovest Austrian pension fund financial planning model InnoALM. Operations Research, 56, 797–810. Grauer, RR and NH Hakansson (1995). Stein and CAPM estimators of the means in asset allocation. International Review of Financial Analysis, 4, 721–739. Hanoch, G and H Levy (1969). The efficiency analysis of choices involving risk. Review of Economic Studies, 36(3), 335–346. Jones, CS and J Shanken (2003). Mutual fund performance with learning across funds. Journal of Financial Economics, 78(3), 507–552. Kahneman, D and A Tversky (1979). Prospect theory: An analysis of decisions under risk. Econometrica, 47(2), 263–291. Kallberg, JG and WT Ziemba (1983). Comparison of alternative utility functions in portfolio selection problems. Management Science, 29, 1257–1276. Kelly, JR (1956). A new interpretation of the information rate. Bell System Technical Journal, 35, 917–926. Li, Y and WT Ziemba (1989). Characterizations of optimal portfolios by univariate and multivariate risk aversion. Management Science, 35, 259–269. Li, Y and WT Ziemba (1993). Univariate and multivariate measures of risk aversion and risk premiums. Annals of Operations Research, 45, 265–296. MacKinlay, AC and L Pastor (2000). Asset pricing models: Implications for expected returns and portfolio selection. The Review of Financial Studies, 13(4), 883–916. MacLean, LC, M Foster and WT Ziemba (2007). Covariance complexity and rates of return on assets. Journal of Banking and Finance, 31(11), 3503–3523. Markowitz, HM (1952). Portfolio selection. Journal of Finance, 7, 77–91. Markowitz, HM (1959). Portfolio Selection: Efficient Diversification of Investments. New York: John Wiley & Sons. Markowitz, HM (1987). Mean-Variance Analysis in Portfolio Choice and Capital Markets. Cambridge, MA: Basil Blackwell. Markowitz, HM and E Van Dijk (2006). Risk-return analysis. In SA Zenios and WT Ziemba (Eds.), Handbook of Asset and Liability Analysis, Volume 1. Amsterdam: Elsevier. Pastor, L and RF Stambaugh (1999). Costs of equity capital and model mispricing. Journal of Finance, 54, 67–121. Pratt, JW (1964). Risk aversion in the small and in the large. Econometrica, 32, 122–136. Rabin, M and RA Thaler (2001). Anomalies: Risk aversion. Journal of Economic Perspectives, 15(1), 219–232. Rosenberg, B, K Reid and R Lanstein (1985). Persuasive evidence of market inefficiency. Journal of Portfolio Management, 11(3), 9–16. Rosenberg, B (1974). Extra-market components of covariance in securities markets. Journal of Financial and Quantitative Analysis, 9(2), 263–274. Rubinstein, ME (1973). The fundamental theorem of parameter-preference security valuation. Journal of Financial and Quantitative Analysis, 8, 61–70. Tobin, J (1958). Liquidity preference as behavior towards risk. Review of Economic Studies, 25, 65–86. Vickson, RG (1975). Stochastic dominance for decreasing absolute risk aversion. Journal of Financial and Quantitative Analysis, 10(5), 799–811. Vickson, RG (1977). Stochastic orderings from partially known utility functions. Mathematics of Operations Research, 2(3), 244–252. Ziemba, WT, C Parkan and FJ Brooks-Hill (1974). Calculation of investment portfolios with risk free borrowing and lending. Management Science, 21, 209–222. Ziemba, WT (1975). Choosing investment portfolios when the returns have stable distributions. In WT Ziemba and RG Vickson (Eds.), Stochastic Optimization Models in Finance. San Diego: Academic Press.

This page intentionally left blank

Chapter 19 Econometrica, Vol. 32, No. 1-2 (January-April, 1964)

RISK AVERSION IN THE SMALL AND IN THE LARGEl BY JOHN

w. PRATT

This paper concerns utility functions for money. A measure of risk aversion in the small, the risk premium or insurance premium for an arbitrary risk, and a natural concept of decreasing risk aversion are discussed and related to one another. Risks are also considered as a proportion of total assets.

1.

SUMMARY AND INTRODUCTION

LET u(x) BE a utility function for money. The function r(x)= -u"(x)fu'(x) will be interpreted in various ways as a measure oflocal risk aversion (risk aversion in the small); neither u"(x) nor the curvature of the graph ofu is an appropriate measure. No simple measure of risk aversion in the large will be introduced. Global risks will, however, be considered, and it will be shown that one decision maker has greater local risk aversion r(x) than another at all x if and only if he is globally more risk-averse in the sense that, for every risk, his cash equivalent (the amount for which he would exchange the risk) is smaller than for the other decision maker. Equivalently, his risk premium (expected monetary value minus cash equivalent) is always larger, and he would be willing to pay more for insurance in any situation. From this it will be shown that a decision maker's local risk aversion r(x) is a decreasing function of x if and only if, for every risk, his cash equivalent is larger the larger his assets, and his risk premium and what he would be willing to pay for insurance are smaller. This condition, which many decision makers would subscribe to, involves the third derivative of u, as r' セP@ is equivalent to u'"u'?;,u" 2 • It is not satisfied by quadratic utilities in any region. All this means that some natural ways of thinking casually about utility functions may be misleading. Except for one family, convenient utility functions for which r(x) is decreasing are not so very easy to find. Help in this regard is given by some theorems showing that certain combinations of utility functions, in particular linear combinations with positive weights, have decreasing r(x) if all the functions in the combination have decreasing r(x). The related function r*(x)=xr(x) will be interpreted as a local measure of aversion to risks measured as a proportion of assets, and monotonicity of r*(x) will be proved to be equivalent to monotonicity of every risk's cash equivalent measured as a proportion of assets, and similarly for the risk premium and insurance. These results have both descriptive and normative implications. Utility functions for which r(x) is decreasing are logical candidates to use when trying to describe the behavior of people who, one feels, might generally pay less for insurance against 1 This research was supported by the National Science Foundation (grant NSF-G24035). Reproduction in whole or in part is permitted for any purpose of the United States Government. 122

317

318

JW Pratt RISK AVERSION

123

a given risk the greater their assets. And consideration of the yield and riskiness per investment dollar of investors' portfolios may suggest, at least in some contexts, description by utility functions for which r*(x) is first decreasing and then increasing. Normatively, it seems likely that many decision makers would feel they ought to pay less for insurance against a given risk the greater their assets. Such a decision maker will want to choose a utility function for which r(x) is decreasing, adding this condition to the others he must already consider (consistency and probably concavity) in forging a satisfactory utility from more or less malleable preliminary preferences. He may wish to add a further condition on r*(x). We do not assume or assert that utility may not change with time. Strictly speaking, we are concerned with utility at a specified time (when a decision must be made) for money at a (possibly later) specified time. Of course, our results pertain also to behavior at different times if utility does not change with time. For instance, a decision maker whose utility for total assets is unchanging and whose assets are increasing would be willing to pay less and less for insurance against a given risk as time progresses if his r(x) is a decreasing function of x. Notice that his actual expenditure for insurance might nevertheless increase if his risks are increasing along with his assets. The risk premium, cash equivalent, and insurance premium are defined andrelated to one another in Section 2. The local risk aversion function r(x) is introduced and interpreted in Sections 3 and 4. In Section 5, inequalities concerning global risks are obtained from inequalities between local risk aversion functions. Section 6 deals with constant risk aversion, and Section 7 demonstrates the equivalence of local and global definitions of decreasing (and increasing) risk aversion. Section 8 shows that certain operations preserve the property of decreasing risk aversion. Some examples are given in Section 9. Aversion to proportional risk is discussed in Sections 10 to 12. Section 13 concerns some related work of Kenneth J. Arrow. 2 Throughout this paper, the utility u(x) is regarded as a function of total assets rather than of changes which may result from a certain decision, so that x = 0 is equivalent to ruin, or perhaps to loss of all readily disposable assets. (This is essential only in connection with proportional risk aversion.) The symbol セ@ indicates that two functions are equivalent as utilities, that is, オ Q HクIセオ means there exist R HクI@ constants a and b (with b>O) such that u 1 (x)=a+buz(x) for all x. The utility functions discussed may, but need not, be bounded. It is assumed, however, that they are sufficiently regular to justify the proofs; generally it is enough that they be twice continuously differentiable with positive first derivative, which is already re2 The importance of the function r(x) was discovered independently by Kenneth J. Arrow and by Robert Schlaifer, in different contexts. The work presented here was, unfortunately, essentially completed before I learned of Arrow's related work. It is, however, a pleasure to acknowledge Schlaifer's stimulation and participation throughout, as well as that of John Bishop at certain points.

Chapter 19. Risk Aversion in the Small and in the Large

124

JOHN W. PRATT

quired for r(x) to be defined and continuous. A variable with a tilde over it, such as

z, is a random variable. The risks z considered may, but need not, have "objective" probability distributions. In formal statements, z refers only to risks which are not degenerate, that is, not constant with probability one, and interval refers only to an interval with more than one point. Also, increasing and decreasing mean nondecreasing and nonincreasing respectively; if we mean strictly increasing or decreasing we will say so. 2.

THE RISK PREMIUM

Consider a decision maker with assets x and utility function u. We shall be interested in the risk premium n such that he would be indifferent between receiving a risk z and receiving the non-random amount E(z) -n, that is, n less than the actuarial value E(z). If u is concave, then ョセ@ 0, but we don't require this. The risk premium depends on x and on the distribution of z, and will be denoted n(x,z). (It is not, as this notation might suggest, a function n(x,z) evaluated at a randomly selected value of z, which would be random.) By the properties of utility, (1)

u(x+E(z)-n(x,z))=E{u(x+z)}.

We shall consider only situations where E{u(x+z)} exists and is finite. Then n(x,z) exists and is uniquely defined by (1), since u(x+E(z)-n) is a strictly decreasing, continuous function of n ranging over all possible values of u. It follows immediately from (1) that, for any constant p., (2)

n(x,z)=n(x+ll, z-ll).

By choosing p.=E(z) (assuming it exists and is finite), we may thus reduce consideration to a risk z- p. which is actuarially neutral, that is, E(z- p.) = 0. Since the decision maker is indifferent between receiving the risk z and receiving for sure the amount na(x,z) = E(z) -n(x,z), this amount is sometimes called the cash equivalent or value of z. It is also the asking price for z, the smallest amount for which the decision maker would willingly sell z if he had it. It is given by (3a)

u(x+na we have (3c)

nix,z)= -na(x,z)=n(x,z)-E(z).

319

320

JW Pratt RISK AVERSION

125

If z is actuarially neutral, the risk premium and insurance premium coincide. The results of this paper will be stated in terms of the risk premium n, but could equally easily and meaningfully be stated in terms of the cash equivalent or insurance premium.

3.

LOCAL RISK AVERSION

To measure a decision maker's local aversion to risk, it is natural to consider his risk premium for a small, actuarially neutral risk z. We therefore consider n(x, z) for a risk z with E(z)=O and small variance a;; that is, we consider the behavior of n(x,z) as a;-+0. We assume the third absolute central moment of z is of smaller order than a;. (Ordinarily it is of order a;.) Expanding u around x on both sides of (1), we obtain under suitable regularity conditions 3 (4a)

u(x-n)=u(x)-nu'(x)+O(n 2 ) ,

(4b)

E{u(x +z)} =E{u(x) + zu'(x)+!z 2 u"(x)+O(z 3)} = u(x) +ta; u"(x) + o(a;) .

Setting these expressions equal, as required by (1), then gives

where (6)

r(x) = -

u"(x) d u'(x) = - dx log u'(x).

Thus the decision maker's risk premium for a small, actuarially neutral risk z is approximately r(x) times half the variance of z; that is, r(x) is twice the risk premium per unit of variance for infinitesimal risks. A sufficient regularity condition for (5) is that u have a third derivative which is continuous and bounded over the range of all z under discussion. The theorems to follow will not actually depend on (5), however. If z is not actuarially neutral, we have by (2), with p. = E(z), and (5):

Thus the risk premium for a risk z with arbitrary mean E(z) but small variance is approximately r(x + E(z)) times half the variance of z. It follows also that the risk premium will just equal and hence offset the actuarial value E(z) of a small risk (z); that is, the decision maker will be indifferent between having z and not having it when the actuarial value is approximately r(x) times half the variance of z. Thus r(x) 3 In expansions, O( ) means "terms of order at most" and o( ) means "terms of smaller order than."

Chapter 19. Risk Aversion in the Small and in the Large

126

JOHN W. PRATT

may also be interpreted as twice the actuarial value the decision maker requires per unit of variance for infinitesimal risks. Notice that it is the variance, not the standard deviation, that enters these formulas. To first order any (differentiable) utility is linear in small gambles. In this sense, these are second order formulas. Still another interpretation of r(x) arises in the special case z= ±h, that is, where the risk is to gain or lose a fixed amount h > 0. Such a risk is actuarially neutral if +h and -hare equally probable, so P(z=h)-P(z= -h) measures the probability premium ofz. Letp(x,h) be the probability premium such that the decision maker is indifferent between the status quo and a risk z = ± h with (8)

P(z=h)-P(z=-h)=p(x,h).

Then P(z=h)=t[I +p(x,h)], P(z= -h)=t[l-p(x,h)], and p(x,h) is defined by (9)

u(x)=E{u(x + z)} =t[l + p(x,h)] u(x+h)+![l- p(x,h)]u(x -h).

When u is expanded around x as before, (9) becomes (10)

u(x)=u(x)+hp(x,h)u'(x)+th 2 u"(x)+O(h 3 ) .

Solving for p(x,h), we find (11)

p(x,h)=thr(x)+O(h2 ) .

Thus for small h the decision maker is indifferent between the status quo and a risk of ±h with a probability premium of r(x) times th; that is, r(x) is twice the probability premium he requires per unit risked for small risks. In these ways we may interpret r(x) as a measure of the local risk aversion or local propensity to insure at the point x under the utility function u; -r(x) would measure locally liking for risk or propensity to gamble. Notice that we have not introduced any measure of risk aversion in the large. Aversion to ordinary (as opposed to infinitesimal) risks might be considered measured by n(x,z), but 1t is a much more complicated function than r. Despite the absence of any simple measure of risk aversion in the large, we shall see that comparisons of aversion to risk can be made simply in the large as well as in the small. By (6), integrating - r(x) gives log u'(x) + c; exponentiating and integrating again then gives e"u(x)+d. The constants of integration are immaterial because ・BオHクIK、セN@ (Note e">0.) Thus we may write (12)

u セ@ Je-fr ,

and we observe that the local risk aversion function r associated with any utility function u contains all essential information about u while eliminating everything arbitrary about u. However, decisions about ordinary (as opposed to "small") risks are determined by r only through u as given by (12), so it is not convenient entirely to eliminate u from consideration in favor of r.

321

322

JW Pratt RISK AVERSION

4.

127

CONCAVITY

The aversion to risk implied by a utility function u seems to be a form of concavity, and one might set out to measure concavity as representing aversion to risk. It is clear from the foregoing that for this purpose r(x)= -u"(x)ju'(x) can be considered a measure of the concavity of u at the point x. A case might perhaps be made for using instead some one-to-one function of r(x), but it should be noted that u"(x) or - u"(x) is not in itself a meaningful measure of concavity in utility theory, nor is the curvature (reciprocal of the signed radius of the tangent circle) u"(x)(l + [u'(x)F)- 3 12 • Multiplying u by a positive constant, for example, does not alter behavior but does alter u" and the curvature. A more striking and instructive example is provided by the function u(x)= -e-x. As x increases, this function approaches the asymptote u=O and looks graphically less and less concave and more and more like a horizontal straight line, in accordance with the fact that u'(x)=e-x and u"(x)= -e-x both approach 0. As a utility function, however, it does not change at all with the level of assets x, that is, the behavior implied by u(x) is the same for all x, since u(k+x)= M・j」クセオHIN@ In particular, the risk premium n(x,z) for any risk z and the probability premium p(x,h) for any h remain absolutely constant as x varies. Thus, regardless of the appearance of its graph, u(x) = -e-x is just as far from implying linear behavior at x = oo as at x = 0 or x = - oo. All this is duly reflected in r(x), which is constant: r(x)= -u"(x)ju'(x)= 1 for all x. One feature of u"(x) does have a meaning, namely its sign, which equals that of -r(x). A negative (positive) sign at x implies unwillingness (willingness) to accept small, actuarially neutral risks with assets x. Furthermore, a negative (positive) sign for all x implies strict concavity (convexity) and hence unwillingness (willingness) to accept any actuarially neutral risk with any assets. The absolute magnitude of u"(x) does not in itself have any meaning in utility theory, however.

5.

COMPARATIVE RISK AVERSION

Let u 1 and u2 be utility functions with local risk aversion functions r 1 and r 2 , respectively. If, at a point x, r 1 (x)>rz(x), then u 1 is locally more risk-averse than u 2 at the point x; that is, the corresponding risk premiums satisfy n 1 (x,z)> nz(x,z) for sufficiently small risks z, and the corresponding probability premiums satisfy p 1 (x,h)>pix,h) for sufficiently small h>O. The main point of the theorem we are about to prove is that the corresponding global properties also hold. For instance, if r 1 (x) > r 2 (x) for all x, that is, u 1 has greater local risk aversion than u 2 everywhere, then n 1 (x,z)>n 2 (x,z) for every risk z, so that u 1 is also globally more risk-averse in a natural sense. It is to be understood in this section that the probability distribution of z, which determines n 1 (x,z) and nz(x,z), is the same in each. We are comparing the risk

Chapter 19. Risk Aversion in the Small and in the Large

128

JOHN W. PRATT

premiums for the same probability distribution of risk but for two different utilities. This does not mean that when Theorem 1 is applied to two decision makers, they must have the same personal probability distributions, but only that the notation is imprecise. The theorem could be stated in terms of n 1 (x,z 1 ) and n 2 (x,z 2 ) where the distribution assigned to z1 by the first decision maker is the same as that assigned to z2 by the second decision maker. This would be less misleading, but also less convenient and less suggestive, especially for later use. More precise notation would be, for instance, n 1(x, F) and nix, F), where F is a cumulative distribution function. THEOREM 1: Let r;(x), n;(x,z), and p;(x) be the local risk aversion, risk premium, and probability premium corresponding to the utility function U;, i=l,2. Then the following conditions are equivalent, in either the strong form (indicated in brackets), or the weak form (with the bracketed material omitted). (a) r 1(x);;:;rix)for all x [and> for at least one x in every interval). (b) n 1 (x,z);;:; [> ]nix,z)for all x and z. (c) p 1(x,h);;:;[> )pix,h)for all x and all h>O. (d) u 1 (u; 1 (t)) is a [strictly] concave function oft.

(e)

オQ_IMセク@

s;: [ O. The same equivalences hold if" increasing" is substitutedfor "decreasing" throughout and/or attention is restricted throughout to an interval, that is, the requirement is added that x, x+z, x+h, and x-h all lie in a specified interval.

PROOF: This theorem follows upon application of Theorem 1 to u 1 (x)=u(x) and uix)=u(x+k) for arbitrary x and k. It is easily verified that (a') and hence also (b') and (c') are equivalent to (d') u'(u- 1(t)) is a [strictly] convex function oft. This corresponds to (d) of Theorem I. Corresponding to (e) of Theorem and (20)-(22) is (e') u'(x)u"'(x);;::;(u"(x))Z [and > for at least one x in every interval]. The equivalence of this to (a')-(c') follows from the fact that the sign of r'(x) is the same as that of(u"(x))Z-u'(x)u"'(x). Theorem 2 can be and originally was proved by way of(d') and (e'), essentially as Theorem I is proved in the present paper.

8. OPERATIONS WHICH PRESERVE DECREASING RISK AVERSION We have just seen that a utility function evinces decreasing risk aversion in a global sense if an only if its local risk aversion function r(x) is decreasing. Such a utility function seems of interest mainly if it is also risk-averse (concave, r;;::;O). Accordingly, we shall now formally define a utility function to be [strictly] decreasingly risk-averse if its local risk aversion function r is [strictly] decreasing and nonnegative. Then by Theorem 2, conditions (i) and (ii) of Section 7 are equivalent to the utility's being strictly decreasingly risk-averse. In this section we shall show that certain operations yield decreasingly riskaverse utility functions if applied to such functions. This facilitates proving that functions are decreasingly risk-averse and finding functions which have this property and also have reasonably simple formulas. In the proofs, r(x), r 1 (x), etc., are the local risk aversion functions belonging to u(x), u 1 (x), etc. THEOREM 3: Suppose a>O: u 1 (x)=u(ax+b) is [strictly] decreasingly risk-averse for ク P セク Q@ if and only ifu(x) is [strictly] decreasingly risk-averse for 。ク P K「セ@ クセ。 Q@ +b. PROOF: This follows directly from the easily verified formula: (26)

r 1 (x)=ar(ax+b).

THEOREM 4: If u 1 (x) is decreasingly risk-averse for ク P セク Q L@ and uz(x) is decreasingly risk-averse for オ Q Hク P Iセクオ Q Hク Q IL@ then u(x)=uz(u 1 (x)) is decreasingly

Chapter 19. Risk Aversion in the Small and in the Large

JOHN W. PRATT

132

risk-averse for x 0 セク Q L@ and strictly so unless one ofu 1 and u2 is linear from some x on and the other has constant risk aversion in some interval. PROOF: We have log u'(x) =log u;(ul_(x)) +log (27)

イHクI]

R Hオ

Q HクIオセKイ

オセHクIL@

and therefore

Q HクIN@

The functions r 2 (u 1 (x)), オセHクIL@ and r 1 (x) are ;;;;o and decreasing, and therefore so is r(x). Furthermore, オセHクI@ is strictly decreasing as long as r 1 (x)>0, so r(x) is strictly decreasing as long as r 1 (x) and r 2 (u 1(x)) are both >0. If one of them is 0 for some x, then it is 0 for all larger x, but if the other is strictly decreasing, then so is r. THEOREM 5: If u 1 , • . . , un are decreasingly risk-averse on an interval [x 0 , xtJ, and c 1 , ••• , c" are positive constants, then u = lセ@ C;U; is decreasingly risk-averse on [x 0 ,xtJ, and strictly so except on subintervals (if any) where all u; have equal and constant risk aversion. PROOF: The general statement follows from the case u=u 1 +u 2 • For this case (28)

(29)

オセG@

Kオセ@

r=- - - - = オセ@

Kオセ@

オセ@

---1'1

オセ@

+u;

+ - -オセ@ - 1 ' 2 ; オセ@

Kオセ@

r'= オセイ@

+u;r; オセ@

+u;

We have オセ^oL@ u;>O, イセoL@ and イ[セッN@ Therefore イGセoL@ and イセ@ = r; = 0. The conclusion follows.

and r'±oo

= 0,

(2.1)

wheref(y) is the density of y, given by

f(y)

=

1

r;c_

v27ray

exp

(

f.l-yf) ,

(yMセ@ 2cry

336

268

Y Li and WT Ziemba

Y. Li, W. T. Ziemba/ Univariate and multivariate measures

then Cov(x,g(ji))

= (Cov(x,ji))E[g'(ji)].

(2.2)

The proof of this lemma can be found in Stein [25], or Rubinstein [23, 24]. We now develop a new formula, called the "second-order covariance operator".

LEMMA 2 (SECOND-ORDER COVARIANCE OPERATOR)

Suppose that x, ji have a bivariate normal distribution and g : lR --+ lR is twice differentiable. If E[g'(ji)], E[g"(ji)] exist and lim (y- f.Lp)g(y)f(y)

y-+±oo

lim g'(y)f(y)

y-+±oo

= 0,

= 0,

(2.3) (2.4)

wheref(y) is the density of ji, given by

f(y)

1

= ../firup exp

(

-

(y- f.1.·)2) 2u( '

then Cov ((x- f.Lx) 2 ,g(ji))

= (Cov (x,ji)) 2E[g"(ji)].

(2.5)

The proof of this lemma is given in appendix A. Since the first-order and second-order covariance operators will be applied to u' and u" where u is a utility function, the following definition is given to ease the representations of several results.

REGULARITY CONDITIONS

A utility function u(x) satisfies the (single) regularity condition if u' (x) satisfies the condition for the first-order covariance operator. Moreover, u(x) satisfies the regularity conditions if in addition, u" (x) satisfies the first and second-order covariance operator conditions.

Chapter 20. Univariate and Multivariate Measures of Risk Aversion

Y. Li, W.T. Ziemba/ Univariate and multivariate measures

337

269

LEMMA3

Assuming that u(x) is twice continuously differentiable, then u(x) satisfies the (single) regularity condition if and only if2 lim オGHクIヲ・bセ@

=

x-+±oo

0.

(2.6)

Assuming that u(x) is continuously differentiable of order 4, then u(x) satisfies the regularity conditions if and only if in addition to (2.6), the following two conditions also hold: lim クオBHIヲ・bセ@

x-+±oo

= 0,

(2.7)

0,

(2.8)

and lim オ

x-+±oo

where B = 1/(2ai) and



HクIェ・bセ@

=

oi is the variance of a normal random variable x.

This lemma is a direct consequence of the definition of the regularity conditions and (2.1), (2.3) and (2.4). Thus the regularity conditions for u(x) will hold as long as u'(x), xu"(x) and does as x tends to infinity. Therefore, u 111 (x) do not approach infinity as fast as ・bセ@ the regularity conditions are the boundary conditions for the utility functions. They are satisfied by exponential, quadratic, logarithmic, power and many other utility functions. We now formally derive Rubinstein's measure of absolute risk aversion from the concept of risk premium in the two-risk situation. Suppose that a decision maker's preferences can be represented by a utility function u(x). He has a random initial wealth x and faces some other risky. He can insure against risky, but not x. Then the risk premium 7r(x, y) is the value satisfying E[u(x + .Y)]

= E[u(x + E.Y- 7r(x,y))].

(2.9)

When a risky is added to some non-random initial wealth x, a decision maker is risk averse if and only if 7r(x, y) > 0. However, if the risk y is added to some random initial wealth x, since y may be negatively correlated with x an4 then have a hedging effect on the initial random wealth, the corresponding risk premium 7r(x,y) may be negative. If x and yare bivariate normal random variables, the sign 2A

similar version of this condition was given by Chipman [3].

338

270

Y Li and WT Ziemba

Y. Li, W. T. Ziemba/ Univariate and multivariate measures

of the risk premium can be determined based on the Markowitz [17, 18] mean-variance criterion: THEOREM I

Suppose that .X, ji have a bivariate normal distribution, and 1r(.X, ji) is the risk premium for a decision maker with utility function u(x). If the decision maker is risk averse, that is, u" < 0, then

(i)

2 2 1r(x,ji) > 0 1'f O'x+.Y > crx;

(ii)

7r(.X,ji) = 0 if oi+y = oGセェ@

(iii)

7r(.X,ji) < 0 if oi+.Y < PGセN@

Proof

Without loss of generality, assume that E[ji] = 0. We only prove (i) since the proofs of (ii) and (iii) are analogous. If 1r(.X,ji)::::; 0, then

E[x + .Y] ::::; E[x- 1r]. Suppose that var (.X+ ji)

> var (.X),

so that var (.X+ ji) > var (.X- 1r). By [7, theorem 4],

E[u(x + ji)] < E[u(x- 1r)] for all u" < 0. But this contradicts the definition of 1r(x, ji) in (2.9).

D

This result shows that the sign of the risk premium for random wealth depends only on the difference between the variance of the initial random wealth accompanied by the additional random variation and the variance of the initial wealth alone. If the additional random wealth ji were to increase the total variance, a risk averse individual with random wealth .X plus a random variation ji would be indifferent between the status quo and a reduction in wealth by the amount of risk premium coupled with an elimination of the wealth variation due

Chapter 20. Univariate and Multivariate Measures of Risk Aversion

Y. Li, W. T. Ziemba/ Univariate and multivariate measures

339

271

to y. Conversely, if the additional random wealth y were to reduce the total variance, a risk averse individual would have to be compensated by an amount equal to the hedging premium, that is, the money value of the risk premium when the additional wealth were to be withdrawn. Although the sign of the risk premium can be determined by the comparison of variances, its magnitude would be expected to relate to the appropriate measure of a decision maker's risk aversion. We now derive an approximate expression for the risk premium. We still assume that the initial wealth and risk have a bivariate normal distribution and the utility function is well-behaved. For simplicity, we also assume that the risk has a zero mean. THEOREM2

Let x,y have a bivariate normal distribution such that E[y] = 0, and u(x) be an increasing utility function. Assume that u(x) satisfies the regularity conditions. Then the risk premium 1r(x, y) for the given x can be approximated by - -

7r(x,y) セ@

2)

2

1

where the approximation is of the order ッHMセKy@ utility u(x) = - exp (-,Bx),

7r(x,y)

E[u"(x)J

-2(o-.H.Y- O"x E[u'(x)] ,

ッMセIN@

In particular, for exponential

2 E[u"(x)] o-x) E[u'(x)] .

= MセHッKNy@

(2.10)

(2.11)

The proof of the theorem appears in appendix A. DEFINITION 1

Let x be a random variable and u(x) be an increasing utility function. Assume that u(x) is twice differentiable and E[u'(x)], and E[u"(x)] exist. Rubinstein's measure of absolute risk aversion of utility function u(x) at xis R(x)

=

E[u"(x)] E[u'(x)] .

Thus the risk aversion R(x) can then be interpreted as twice the risk premium a decision maker requires per unit of incremental variance for infinitesimal risk. If xis non-random and equal to the constant x, then (2.10) becomes

1r(x,y) セ@ セッMェイHクIL@

340

272

Y Li and WT Ziemba

Y. Li, W. T. Ziemba/ Univariate and multivariate measures

where the approximation is of order o(uj). Thus Rubinstein's measure of risk aversion is a generalization of the Arrow-Pratt measure. If x and ji are independent, (2.10) becomes

1r(x,ji) セ@ セオェrHクIL@ where the approximation is of order uj. This result is the case of a risk aversion measure with independent wealth and risk obtained by Kihlstrom et al. [13] and Ambarish and Kallberg [1]. For given bivariate normal random variables x and ji, a decision maker with larger Rubinstein's measure of risk aversion than another would require a higher risk premium in the absolute sense if the incremental variance is small. But if two individuals have constant risk aversion then this property holds for any bivariate normal random variables x and ji. COROLLARY l

Let x,ji be bivariate normal random variables and u;(x) -e-/3;x be exponential utility functions for i = 1, 2. Let R; (.X) and 11"; (x, ji) be Rubinstein's measures of absolute risk aversion and risk premiums for utility function u; (x) (i= 1,2), respectively. If

then

Equivalently,

(.X, ji) > 1rz (x, ji) > 0 if オセKヲ@

> オセ[@

if オセKヲ@

= オセ[@

1r, (.X, ji) < 1rz (x, ji) < o if オセKy@

< オセN@

(i)

7rt

(ii)

1r,(x,ji) = 1r2 (x,ji)

(iii)

0

This result follows immediately from theorem 2. For the exponential utility function, the Arrow-Pratt measure of absolute risk aversion is identical to Rubinstein's measure of absolute risk aversion. In general case, the two measures have the following relationship: THEOREM 3

A decision maker with an increasing utility function u(x) is risk averse, that is, r(x) > 0 for all x, if and only if R(x) > 0 for any random variable .X.

Chapter 20. Univariate and Multivariate Measures of Risk Aversion

Y. Li, W. T. Ziemba/ Univariate and multivariate measures

341

273

Proof R(x) > 0 for any random variable implies that R(x) > 0 for any degenerated random variable x x. But in this degenerated case, R(x) = r(x). Thus the sufficient condition is proved. The necessary condition is true since u is increasing, thus r(x) > 0 for any x implies the concavity of u, namely, u" セ@ 0. Hence, by the D definition of R(x), R(x) > 0 for any random variable x.

=

In the preceding analysis, the main results are based on the assumption that the initial random wealth and risk have a bivariate normal distribution. We now show that for the class of quadratic utility functions, all of the previous results hold for any random initial wealth and risk with an arbitrary bivariate distribution. Since the mean-variance criterion is also valid for the class of quadratic utility functions, we have the following result which is analogous to theorem 1: THEOREM4

Let U(x) = x- f3x 2 be a quadratic utility function on (-oo, 1/2{3) where f3 > 0, then for any random variables x and y on ( -oo, 1/2{3), the following properties hold: (i)

2 2 1r(x,y) > o 1.f ai+ji >ax;

(ii)

1r(x,y)

(iii)

1r(x,y) < o

=o

2. if clx+ji ax, .f 2 2 1 ai+ji 0. Let R(x) and 1r(x, y) be the risk aversion measure and risk premium of utility function u(x). Assume that x and y are in the suitable range such that R(x) > 0, and R(x- 1r) > 0, then 1r(x,y) セ@ AH。セKケM

。セIrHクL@

(2.12)

where the approximation is of the order o(oi+ji- oi). The proof appears in appendix A. As shown in corollary 1, under the assumption that two random variables have a bivariate normal distribution and the

342

274

Y Li and WT Ziemba

Y. Li, W. T. Ziemba/ Univariate and multivariate measures

decision makers have constant absolute risk aversion, one decision maker is more risk averse than the other if and only if the absolute value of the risk premium for the former is larger than that for the latter. Similarly, for the class of quadratic utility functions, the above result holds for random variables with arbitrary bivariate distributions. COROLLARY2

Let .X,ji be any two random variables and u1(x), u2 (x) be quadratic utility functions given by

Let Rlx) and 1r/x, y) be the risk aversion measures and risk premiums for utility function u;(x), i = 1, 2, respectively. Assume that R1 (.X)

> R2(i) > 0;

then

Equivalently,

(i)

1rJ (.X, ji) > 7!"2 (i,ji) > 0 if oBセKェゥ@

(ii)

1rJ(i,y)

(iii)

1r) (i,ji) < 7r2(i,ji) < 0 if oBセKェゥ@

=

7r2(i,ji)

0

if oBセKェゥ@

> PBセ[@ oBセ[@

=

< PBセN@

So far we have been concerned with risks and risk premiums in absolute terms. We now view them as the proportions of some fixed value w0 • Specifically, for two random variables .X and y, we consider w0i and w0 ji. The relative risk premium is defined as follows: DEFINITION 2

Let u(x) be an increasing utility function, and i, y be bivariate random variables. For a constant w0 , the relative risk premium 1r* (w0; .X, y) satisfies

E[u(w 0x + w0 ji)]

= E[u(w0i + w0 (Ey- 1r*))].

(2.13)

In portfolio selection problems, if w0 is the initial dollar investment, .X and y as the dollar returns on two risky assets for each dollar's investment, then the

Chapter 20. Univariate and Multivariate Measures of Risk Aversion

Y. Li, W.T. ZiembafUnivariate and multivariate measures

343

275

relative risk premium 1r* can be interpreted as the reduction in or compensation for the return on the second asset the investor would accept if that asset were to be replaced with a risk-free asset. DEFINITION 3

Let u(x) be an increasing utility function, and x be a random variable. For a constant w0 , Rubinstein's measure of relative risk aversion R*(w0 , x) is defined to be * _ E[u"(w0 x)J R (w0 , x) = -w0 E[u'(wo·x)] .

(2.14)

From (2.13) and (2.14), it follows that 1r*(l;x,y) = 1r(x,y)

R*(l,x) = R(x).

and

This implies that Rubinstein's measure of absolute risk aversion and absolute risk premium can be recovered from Rubinstein's measure of relative risk aversion and relative risk premium by taking the constant w0 as unity. The following theorem shows that Rubinstein's measure of relative risk aversion can also be justified by the associated relative risk premium in the local sense. THEOREM 6

Let u(x) be an increasing utility function, and x,y be bivariate random variables such that E[y] = 0. If either of the following conditions holds: (i) (ii)

x andy have a bivariate normal distribution, and u(x) satisfies the regularity conditions, u(x) is a quadratic utility function,

then for a constant w0 , the relative risk premium 1r* (w0 ; x, y) can be approximated by - -) セ@ 7r*( wo;x,y

where the approximation is of order ッHMセKy

2 21(D"x+yD"x2)R*( Wo,x-) ッMセIN@

Proof Based on the given conditions, from theorems 2 or 5, it follows that

(2.17)

344

Y Li and WT Ziemba

276

Y. Li, W. T. Ziemba/ Univariate and multivariate measures

or, -

7r ( WoX,

2 WoY-) -_ 2I Wo2( C1x+ji

Comparing (2.9) with (2.13) yields

that is, -1 7 r ( WoX,

wo

WoY-) .

(2.15)

Similarly, comparing definitions 1 with 3 yields (2.16)

Dividing the above equation by w0 , from (2.15) and (2.16), it follows that

D From (2.17), the interpretation of R* ( w0 , .X) is similar to that of R( x). If x is non-random (x = 1), from (2.14) and (2.15), then R*(w 0 , 1) = r*(w0 )

and 1r*(w0 ; 1,y)

= 1r*(w0 ,y),

where r*(x) and 1r*(x) are the Arrow-Pratt measure of relative risk aversion and risk premium. Thus Rubinstein's measure of relative risk aversion R* ( w0 , x) is a generalization of the Arrow-Pratt measure of relative risk aversion r* (x), and has the corresponding local properties of the latter. 3.

Multivariate measures of risk aversion

We first present the key tools for the derivation of these multivariate measures of risk aversion. The first-order covariance operator developed by Stein [25) and Rubinstein [23, 24] has been generalized to a multivariate version by Gassmann [6]. LEMMA 4 (FIRST-ORDER MULTIVARIATE COVARIANCE OPERATOR)

Let X, Y be multivariate random vectors on IR.m and IR.n, respectively, with joint distribution

Chapter 20. Univariate and Multivariate Measures of Risk Aversion

Y. Li, W.T. Ziemba( Univariate and multivariate measures

345

277

Let g : rNョセMKォ@ be di.fferentiable with Jacobian matrix lg of dimension k x n. Assume that Cov(X,g(Y)) and E[lg(Y)] exist. If lim g(y0 + ty)fy(Yo + ty) = 0

(3.1)

t--->±00

for ally, y 0 E R.n, wherefy is the probability density of Y, given by

then Cov (X,g(Y))

= Cov (X, Y)E[lg(Y)f

(3.2)

The proof can be found in Gassmann [6]. Now we develop a new formula which will be called "Second-Order Multivariate Covariance Operator". This operator is the multivariate version of secondorder covariance operator in the univariate case developed earlier in this section. LEMMA 5 (SECOND-ORDER MULTIVARIATE COVARIANCE OPERATOR)

Let X, Y be multivariate random vectors on R.n, with joint distribution

be twice continuously differentiable with the Jacobian matrix of Let H : rNョセMKク@ its ith row H; being J[H;(y)] of dimension n x n. If lim (Yo+ ty)H(Yo + ty)fy(Yo + ty)

t--->±00

lim J[H;(Yo

t->±oo

= 0,

+ ty)Jfy(Yo + ty) = 0,

(3.3) (3.4)

for i = 1, 2, ... , n, and y 0, y E R.n, where fy is the density of Y, then

E[(X

J-Lx)T H(Y)(X -J-Lx)J

= tr(:ExE[H(Y)]) + O(tr(Ej,y)).

(3.5)

The proof of this lemma is outlined in appendix B. Suppose that the preferences of a decision maker can be represented by a multi-attributed increasing concave utility function u(x) with xT = (x 1, ••• , xn)· We may interpret x as a wealth vector representing a decision maker's wealth measured at different points of time, or the wealth held in different assets at the same point of time.

346

Y Li and WT Ziemba

278

Y. Li, W. T. Ziemba/ Univariate and multivariate measures

Consider that a decision maker has two sources of risk in his wealth holdings. He has an initial wealth X which is a random vector. He also faces another risk Y which affects the wealth vector. These risks may be li:Orrelated. Assume that risk Y can be insured against but the initial wealth risk is uninsurable. We are interested in determining the amount of risk premium that a decision maker is willing to pay to make him indifferent between having the other risk with random wealth X+ Y and eliminating that risk by paying the premium 1r with remaining wealth X 1r. This model was proposed by Ambarish and Kallberg [1] and was developed from Ross's [22] partially insurable lottery and from the random initial wealth model of Kihlstrom et al. [13]. We now formally state the model's assumptions. A decision maker's preferences for wealth are represented by a von Neumann and Morgenstern utility function which is a mapping from the wealth space into the real line. The wealth variables are allowed to be negative. U :

Rnt-tlR.

(Al) The gradient of u at x,

'\7Tu(x)

(8u(x)j8x 1 , ... , 8u(x)j8xn),

satisfies

8u(x)j8xi > 0 for all x E IR.n and all i. (A2) The Hessian of u at x, H(x), is an n x n symmetric, negative semidefinite matrix with (i, j)th element lfu(x) / 8xi 8xj. (A3) u is continuously differentiable on Rn of order 4 and all of the relevant

integrals are finite. Let X, Y be two random vectors on Rn, and u(x) be a utility function satisfying (Al)-(A3). The risk premiums 1r are any constant vector on Rn such that E[u(X + Y)J

= E[u(X + EY- 1r)].

(3.6)

The left-hand side of (3.6) represents the expected utility of wealth when the decision maker has two types of risks; and the right-hand side represents the expected utility after one of the risks is eliminated. The following lemma is a result of Ambarish and Kallberg [1] with minor different assumptions. LEMMA6

Given that (Al)-(A3) are valid and E[Y] E['\lTu(X)]1r

= 0, then

= -E['\7T u(X) Y]- (1/2)E[YT H(X) Y] + o(trEx).

(3.7)

Chapter 20. Univariate and Multivariate Measures of Risk Aversion

347

279

Y. Li, W. T. Ziemba/ Univariate and multivariate measures

This result is derived from the Taylor expansion of both sides of (3.6). The proof appears in Ambarish and Kallberg [1]. For the derivation of the risk aversion measure with a correlated random wealth vector and a risk vector, it is assumed that (A4) X, Yare multivariate joint normal random vectors on lRn, with distribution

For simplicity and without loss of generality, assume that E[Y] = 0. MULTIVARIATE REGULARITY CONDITIONS

Given that (Al)-(A4) are valid, u satisfies the regularity condition if the gradient of u, 'Vu(x) in place of g satisfies condition (3.1); and u satisfies the regularity conditions if u satisfies the regularity condition and the Hessian of u, H(x) satisfies conditions (3.3) and (3.4) where all of the variables are replaced by x. (A5) u satisfies the regularity conditions. THEOREM 7

Given that (Al)-(A5) are valid, an approximation of order o(tr(Ex+Y- Ex)) is

(3.8) where (ET'Vu(X))- is a generalized inverse of E['Vu(X)]. The proof of this result is provided in appendix B. For any m x n matrix A, the n x m matrix A- is a generalized inverse if AA-A= A (21]. Thus the solutions for 1r are not unique. Hence, the risk premiums can be paid in different ways to induce the same expected utility. Let the ijth element of EX+Y- Ex be dij, and ui 8uj8xi, and uij = Hij. Written in summation form, (3.8) is then 1 n

n

I:: E[ui(X)]1ri セ@

I:: dijE[uij(X)], 2i,j=l

-

i=l

or,

セe{オゥHxI}Qイ@

n

セ@



1

セ@

n

(

[Eui(X)] セ、ゥェHMe{オxI}O@

n

)

.

(3.9)

348

Y Li and WT Ziemba

280

Y. Li, W.T. Ziemba/ Univariate and multivariate measures

Let R(X)

= [R!i(X)] = (

E[u!i(X)]) E[ui(X)] '

then (3.9) can be written as

(3.10) where diag (A) is a diagonal vector of the matrix A. Therefore, (3.11) is a typical approximation of the risk premium. The other approximations are located on a hyperphase passing through it and orthogonal to E[\i'u(X)]. DEFINITION 4

Given that (Al}-(A5) are valid. The absolute risk aversion matrix is R(X)

=(

E[u!i(X)]). E[ui(X)]

If the dimension of the wealth vector is one, then this matrix measure is the same as Rubinstein's measure and we have the same approximate relationship between risk aversion and risk premium as obtained in theorem 2. This matrix measure is thus a multivariate generalization of Rubinstein's absolute risk aversion measure. If the wealth vector X and risk vector Yare uncorrelated, from (3.11), we have 0 1r

セA@

diag(EyR(X)).

(3.12)

This relationship has been obtained by Ambarish and Kallberg [1]. In particular, if X is non-random, that is X= x, then

This is the matrix measure of risk aversion for non-random wealth developed by Duncan [4]. The approximate risk premium in (3.11) suggests that the multivariate formulation of risk aversion is important when either (Ex+Y- Ex) or R(X) is

Chapter 20. Univariate and Multivariate Measures of Risk Aversion

Y. Li, W. T. Ziemba/ Univariate and multivariate measures

349

281

not diagonal. In the .first case, the correlations between different dimensions of wealth would be unchanged if the risk premium were paid. In the second case, the utility function is additive, thus the risks in different dimensions of wealth are mutually independent. To further understand the meaning of the matrix measure of risk aversion, we introduce the directional risk premium and the directional matrix measure of risk aversion. We assume that the risk premiums can be paid only in one of the directions of the wealth vector. Then the risk premium in kth direction 1r(k) is the kth element of the risk premium vector 1r = (1rj) satisfying 1!'·

J

= { 1r(k) l 0,

if j = k; ifj # k

for any k 1, ... , K. From (3.8), the directional risk premiums can be approximated by (3.13) DEFINITION 5

The matrix measure of risk aversion in the kth direction is

fork= 1, ... ,n.

The directional matrix measures of risk aversion are positive definite since E[H(X)] is negative definite and E[uk(X)] > 0 for all of k. Comparing this definition with the definition of absolute risk aversion matrix yields (3.14)

where Diag HセI@ for a カ・」エッイセ@ means the matrix with diagonal elements ッヲセN@ Since the elements in the diagonal matrix are the marginal rates of substitution of utility in one direction for the other, the directional matrix measures of risk aversion are determined once the matrix measure of risk aversion R(X) and these marginal rates of substitution are known. On the other hand, since the kth row of R(X) is identical to that of R(k)(X), the matrix measure of risk aversion is a combination index of the set of directional matrix measures of risk aversion. The approximation of the risk premium in the kth direction is 7r(k). From (3.13), it follows that (3.15)

350

Y Li and WT Ziemba

282

Y. Li, W.T. ZiembafUnivariate and multivariate measures

THEOREMS

Given that (Al)-(A5) are valid,

(i)

ir(k)

> 0 if EX+Y- Ex is positive definite;

(ii)

ir(k)

=0

(iii)

ir(k)

< 0 if Ex+Y- Ex is negative definite,

for all k

if Ex+Y = Ex;

= 1,2, ... ,K.

The proof is given in appendix B. It is well known that if all individuals' preferences are represented by uni-

dimensional increasing concave utility functions, and the random variables have a joint normal distribution, the risks of the random variables can be completely represented by the variances. With the assumption that all individuals have multiattributed concave utility functions, and the random vectors have a joint multivariate normal distribution, the risks of the random vectors can analogously be represented by the covariance matrices. One random vector can be said to be more (less) risky than the other if the difference between the first covariance matrix and second covariance matrix is positive (negative) definite. The results in theorem 8 can be interpreted as follows: the directional risk premium is positive if the risk with two components of random wealth is higher than that with only one component; the directional risk premium is negative if the risk with two components of random wealth is lower than that with only one component. Therefore, the directional risk aversion matrices can yield the desirable results with multi-attributed utility functions and joint normally distributed wealth vectors. Appendix A: The univariate case

The proof of lemma 2 is based on:

LEMMA A

g : lR セ@

Let Let y be a normal random variable 'Yith mean P,ji and variance オセN@ lR be differentiable of order n. If E [g(' >(y)] exists for all i = 1, 2, ... , n and lim (y- J.Lyt-jg(j-I)(y)f(y)

y-+±oo

=0

for all}= 1,2, ... ,n,

Chapter 20. Univariate and Multivariate Measures of Risk Aversion

Y. Li, W.T. Ziemba/ Univariate and multivariate measures

351

283

wheref(y) is the density of y, then

To,g

E[g(y)J,

T1,g

= ojE[g'(y)], E[(y MセエケI

T2,g Tn,g

= (n-



ァHケI}@

= QWセe{ァHケIj@

QIWセtョMRLァ@

+ 17fE[g"(y)],

+ (n- 2)17fTn-3,g' + 17fTn-2,g"

for n 2: 3,

where Tn,g = E[(y MセエNyァHケI}@

Proof Sincef(y) is the density of the normal random variable y, 1

f(y) = vf21r !'(y)

=

.Y

17

HケセpLI@

exp

(

(y

/Ly)2) ,

2o}

f(y), y

QWセ@

f(y) = - _Y_j'(y). y -tLy

(A. I)

Then for n 2: 3,

l

+oo

-oo

MQWセ@

(y- IL.Ytg(y)J(y) dy

1-:

00

= MャWセHケ@

MセエケイャァHIjZ@

- j⦅セ@ QWセHa@

(y- ILyr-lg(y)J'(y) dy

f(y)[(n

l)(y MセエケRァHI@

+ (y -tLy)n-lg'(y)] dy)

+ B +C).

By assumption,

A= 0,

B = (n- l)Tn-2,g·

352

Y Li and WT Ziemba

284

Y. Li, W.T. Ziemba/ Univariate and multivariate measures

Substituting (A.l) into C yields

1:

00

c

-O'j

(y- J.Ljit- 2g'(y)f'(y) dy

-O'j((y- jNlェゥエM

-1:

00

オセHaャ@



ァGHケIヲゥAセ@

f(y)[(n- 2)(y- f.Lyt- 3g'(y)

+ (y- f.Lyt- 2g"(y)J dy)

+ Bl + Cl).

By assumption, AI = 0,

Bl = (n- 2)Tn-3,g''

Cl = Tn-2,g''·

Therefore, for n :2: 3,

Tn,g

= (n-

l)O'jTn-2,g + (n- 2)ujTn-3,g'

+ O'jTn-2,g''·

Hence,

= E[g(y)J, T1,g = ujE[g'(y)J,

To,g

T2,g = E[(y- f.Ly) 2g(y)J

= ujE[g(y)J + O'jE[g"(y)]. Proof of lemma 2

Cov ((.X- J.L.x) 2,g(y)) = E[(x- f.L.x) 2g(y)]- E[(x- J.L.x) 2]E[g(y)] =

E[(x- J.L.x) 2g(y)J- oGセe{ァHケIjN@

The conditional distribution of .X given

and variance

y is normal with mean

0

Chapter 20. Univariate and Multivariate Measures of Risk Aversion

Y. Li, W. T. Ziemba/ Univariate and multivariate measures

353

285

From the conditional mean,

ks. Alternatively considering u for g fixed the investor must receive more return for the same risk as k increases because - 2kw 2 dominates ( (e + l)w ). Hence for a given g ask decreases the investor becomes more risk averse and, as would be expected, he increases the proportion of the riskless asset in his portfolio.

The Exponential Utility Function The exponential utility function

u(w)

(19)

=

1 -e

-aw

,

is concave if and only if a :i;; 0 and it has constant absolute risk aversion equal to a at all wealth levels. Freund [7] has shown that maximizing E 1[1 - e-•!'"], where セLN@ N L) is equivalent to maximizing x - ax' L x/2. Letting x = (a, 1 - a)', セ@ = HセッL@ R )' and

a,

t

means that one may solve (5) by solving lmax,P.(a)

= 。セッK@

(1- a)R- a((l- a) 2 rrR 2)/2 -oo 0 and {j セ@ 0. The function u is defined only for w セ@ - {j/'Y and tends to - oo as w ----> - {j/'Y. As in the case of the power utility function it is appropriate to append a linear segment to u for w セ@ w0 where Wo > -{j/'Y. Then

(23)

u(w) = u1(w) = u2(w)

Since w

= 。セッ@

+ (1

loge (w - wo) ')'Wo

== 'Y {j + == log

([j

- a)R, w セ@

+

+ log

([j

+

')'Wo)

f'W)

w0 if and only if R セ@

ifw if w

(w 0 - 。セ P IOHQM

セ@ セ@

Wo Wo.

a) ==A. Thus

Chapter 22. Calculation of Investment Portfolios CALCULATION OF INVESTMENT PORTFOLIOS

387

221

08

05

QUARTERLY DATA l'o=0.93 w. =1.0)

0.4

0.2

00

0.1

0.3

0.5

0.4

0.7

-02

-0.4

VEARLV DATA ('b=0.95 Wo

Wo

=1.0)

= 1.0)

-0.6

-0.5

-0.8 6

FIGURE 5. Power Utility Function-Quarterly and Yearly Data: a* vs. {3.

FIGURE 6. Logarithmic Utility FunctionQuarterly and Yearly Data: a* vs. {3.

to calculate a* one must

max_.,., 0 and that the relevant expectations are finite. Computing the partial derivatives of the objective of (2) with respect to the portfolio proportions yields

aZ(x)jax;

eセサ@

= =

オGHキ

P セ[ス@

P ・クIキ

」ッカHキセ[@

LオGHキ

P セtクI@

+ wJ;E 0 and d 2p.j da 2 > 0, i.e., the indifference curves of portfolio mean and variance are convex to the a-axis. It can be shown, see e.g., Rubinstein [19], under usual two parameter capital market equilibrium assumptions that ap.jaa = aR.

392

JG Kallberg and WT Ziemba

1260

J. G. KALLBERG AND W. T. ZIEMBA

4.

Empirical Results

A major conclusion of this study is that "similar" RA values yield "similar" optimal portfolios, regardless of the functional forms of the utility functions concerned. The empirical results are outlined after defining these measures of similarity. An appropriate measure of the proximity of portfolios x 1 and x 2 is the percentage cash equivalent difference; see Dexter, Yu, and Ziemba [4]: u- 1 {Z(x 1) } - u- 1 {Z(x 2)} X 100

u- 1 { Z(x 1)}

(5)

where Z(xi) = E {u(exi)}, i = 1,2. Recall that u- 1{Z(xi)} is the cash equivalent value of the random return セtクゥ@ generated by portfolio xi. To formalize the concept of similar RA, let u be the base utility function to which comparisons of the other ui will be made. If f3 = セ@ and x* solves (2), the average risk a version of u is

(6) We say ui has the same risk aversion as u if its average where キMnHセtクJLRZIN@ risk aversion is the same as u's over the distribution of (u-optimal) final wealth, i.e.,

(7) The solution of (7) is the appropriate value of f3i to compare ui with u. To simplify the solution4 of (7), we let exponential utility (u 2) be the base utility function (since with constant absolute risk aversion, the right-hand side of (7) is simply {3 2 , the parameter of the exponential utility function). This choice also simplifies the calculation of the percentage error because

see Freund [7]. The expected utility 5 may now be computed explicitly using only the 4 Equation (7) was solved using a hybrid bisection-secant algorithm. Difficulties in solving this equation 6 so that motivated the exponential extension described in footnote 5. In all cases we used a tolerance of the {31 chosen were within ± of the true {31 • See [II] for further details. 5 0f the seven utility functions used in this study only the exponential is well behaved over the whole real line as required for normally distributed final wealth. The other utility functions for certain values of w may be undefined (e.g., log, special exponential, power and negative power), may have decreasing marginal utility (e.g., quadratic) or may be convex (e.g., arctan). A common method of modifying the domain utilizes a linear segment near the boundary of the problem area, see [25] for technique. For the data and utility functions used here this method (usually) leads to a jump discontinuity in the RA(w) function. This may lead to either instability of the optimal portfolio weights and/or difficulties in solving (7). Since RA = 0 at w = 0 for the arctan function the linear segment suffices since it also has RA = 0 at w = 0. For the log, special exponential, power and negative power we appended an exponential segment at the points w 0 = 0.01, 0.25, 0.8, and 0.6, respectively, to yield the extended function

w-

w-•

u(w)

=

-

{

u(w) u(wo)

+ K(l-

e-RX(w-w))

where

Ro = A-

-

u"(wo) u'(wo)

and

u'(w 0 ) K=--.

Ro A

Using this u(w) guarantees that u, u', u" and hence RA are continuous. We did not append any segment onto the quadratic utility since u' < 0 only above {3 1 ""0.5. However, the u' ""0 behavior near 0.5 yields flat behavior in this region, see e.g., Figure Ia.

Chapter 23. Comparison of Alternative Utility Functions

393

1261

ALTERNATIVE UTILITY FUNCTIONS IN PORTFOLIO SELECTION TABLE I

Security Means and Variances Mean I. Cunningham Drug Stores 2. National Cash Register 3. Metro-Goldwyn-Mayer 4. Gillette Co. 5. Household Finance Corp. 6. H. J. Heinz Co. 7. Anaconda Co. 8. Kaiser Al. & Chern. 9. Maytag Co. 10. Firestone Tire and Rubber

Monthly Variance

1.0194 1.0173 1.0162 1.0137 1.0128 1.0112 l.Olll 1.0095 1.0084 1.0064

0.0105 0.0050 0.0148 0.0044 0.0052 0.0049 0.0081 0.0073 0.0048 0.0036

Mean

Yearly Variance

1.2852 1.2549 1.1819 l.l694 l.l484 l.l608 1.1711 1.0861 l.l292 1.0626

0.3276 0.2152 0.1711 0.0666 0.0517 0.1413 0.1865 0.0324 0.1676 0.0155

TABLE2 Monthly Security Correlation Matrix

2 I 2 3 4 5 6 7 8 9 10

1.000

0.2843 1.0000

4 0.1313 0.1333 1.0000

0.3049 0.3859 0.2882 1.0000

Security Number 5 6 0.1739 -0.0101 0.1251 0.2063 1.0000

0.3457 0.2334 0.3243 0.2750 0.2847 1.0000

7 0.4556 0.1880 0.2974 0.4110 0.2757 0.3031 1.0000

0.3033 0.4155 0.2010 0.3557 0.2670 0.1259 0.4546 1.0000

9

10

0.4229 0.2680 0.1762 0.2088 0.2528 0.3178 0.3445 0.3315 1.0000

0.2822 0.1332 0.1272 0.2616 0.3167 0.2501 0.3020 0.1968 0.1898 1.0000

given means and variances. Furthermore, since

the calculation of the cash equivalent is simplified. The data were obtained from the CRSP tape and consisted of monthly observations of ten well-known U.S. securities. These were selected from a group of 60 to form a sample that provided ample diversification and smooth transitions in means and variances (so that no securities were obviously dominant or dominated). Ten was chosen as the number of securities because it appears to be the optimum number when one balances the criteria of ease of exposition, interpretation of results, computational considerations, and diversification properties. (Evans and Archer [5] found that about 10-12 securities provide ample diversification for individual investors.) The relatively stable period January 1965-December 1969 was used in an attempt to satisfy the stationarity assumption made. The price relatives were adjusted to reflect investment of dividends, stock splits, etc. The 60 observations were used to obtain maximum likelihood estimates of the means and covariances of monthly returns. Tables I and 2 summarize these statistics. 6 6 Quarterly and yearly means and variances estimates were obtained in the usual way after computing the corresponding compound rates of return. The variances were then combined with the monthly correlation coefficients to generate the quarterly and yearly covariances using the formula aij(k) = pija;(k)aj(k), where k =quarterly and yearly, and i, j = I, ... , 10 with i 7" j. Note that these are no longer MLE. See [10] for more details.

394 t:::J

0\

N

セ@

;J>

t"" t"" to

Corresponding Range Utility

Function

u(w)

Absolute Risk Aversion. RA

Theoretical Parameter

Parameter

Theoretical Range

Range

of RA with fixed

of RA with fixed w = I with Parameter

Range

Used

w=!

Range Used

trj

)0

Remarks u 1 begins with the highest mean return portfolio

Quadratic

w- /3w 2

2/3, 2/3 1w)

(I-

which is also a high variance portfolio. Then as {3 1 __,.0.50. RA __,. oo which yields the minimum 0-->w/2

0.....,.0.50

0-->

00

0->

00

variance portfolio. This is consistent with the theory that increasing RA yields more risk averse behavior, see Theorem I in Pratt [18].

u2 is quite similar to u 1 except that the final portfolio (when /3 2 = 10) has a low variance but

is not a minimum variance portfolio. There is a Exponential

Logarithmic

I - e- fl?"'

log( /3 3 + w)

/3,

I

(/3,+ w)

0--'>

00

0--'JoiO

-0.80--71.2

0-->

00

00

-->0

0---?10

5.0--70.45

gradual shift for increasing /3 2 out of the risky security I into the safer securities RセUN@ The largest change occurs in the 0 < /3 2 = RA < 3 range. When /32 = RA ;;. 3 and there is little change in behavior. As u3 ranges from -0.60 to 1.2 there is イ・ャ。セ@ tively little change in the optimal portfolio composition except for a gradual switch out of the safer securities 3, 5 into the riskier security I. Security 2 with its high mean and low variance always forms about half of the portfolio.

Cl

セ@

0

セ@

"セ@ N

to

;J>

JG Kallberg and WT Ziemba

,... 0

TABLE 3 Monthly Empirical Test Data

TABLE 3 (continued) Monthly Empirical Test Data Corresponding Range Utility

Function

Special

e/34/w

Exponential

Power

(w- w 0)Ps w0 =

Negative Power

0.75

- (w- w 0)-lh

w0

=

0.50

Absolute Risk Aversion, RA

Parameter Range

Theoretical Range of RA with fixed

w = 1 with Parameter

Range

Used

w=l

of RA with fixed

Range Used

2w+ {34 0.10---+4.1

max(0,2w) 0--> 00

セ@

2-->

2.1---+6.1

00

(I- /3,)

(w- w0 )

I+ [36 (w- w0 )

ッセエN@

0.05----* 1.0

0-->

00

0.10--72.1

4-->0

2-->

00

3.8 --+0

RNセV@

Remarks

Arctan

tan -l(w + /37)

Mキセ」ッ@

I +(w+/37) 2

PNQセT@

0-->0 max= 1.0

atfh=O

:;:: tn c -1

E

::;! '"Il

c z

g 0

z

"'z

identical to that of u 4 and indeed the optimal portfolio propol-tions are also nearly identical.

セ@

fh E [0.10, 4.1]

and always yields high variance portfolios. As

1.0--+0.38

セ@ セ@

u5 initially yields low variance portfolios when {3 5 = 0.05 and RA = 3.8. However, aS /3 5 ___, I.O, R A drops sharply to 0 and the corresponding optimal portfolios have high variances. There is a gradual shift out of the safer securities 5, 7, 9 into the high return (and also high variance) Security I. u 6 has an RA range of 2.2--+ 6.2 which is nearly

u7 has low risk aversion for all 2(w+/37)

セ@

There is relatively little variation in the optimal portfolio composition because RA ranges from 2.1 to 6.1 and hence never has low values. There is a gradual shift into the lower variance securities 4, 5 at the expense of the higher variance security 1 and security 2 (because of the diversification effects, in particular 2 and 5 are negatively correlated).

/h increases the portfolios become more and more risky since RA is decreasing.

t'"'

0

"'tn

セ@

0 z

Chapter 23. Comparison of Alternative Utility Functions

u(w)

Theoretical Parameter

N

a-,

395

""

396

JG Kallberg and WT Ziemba

1264

QUADRATIC MONTHLY

Vl

z

0

i= c.: 0 a. 0 c.: a.

5

·' 10

3

9

6

KMイセLNQS@ 05

35

25

-1

FIGURE

,g

Ia.

t.s

s(31

1.

1·2

Quadratic Monthly.

EXPONENTIAL MONTHLY

.a

Vl

セ@

.6

1-

セ@

a.

.s

0

c.: ·' a.

.2

FIGURE

lb.

Exponential Monthly.

,g

LOG MONTHLY .a

Vl

z 2

·6

1-

セ@

5

a. 0

a: a. .' .3

.2

-. 8

-. 6

-. 4

FIGURE

I c.

·'

Log Monthly.

.a

Chapter 23. Comparison of Alternative Utility Functions ALTERNATIVE UTILITY FUNCTIONS IN PORTFOLIO SELECTION .g

·•

SPECIAL EXP MONTHLY

·1

Ul .•

z

Q

:ro .sr----Q.

0

a:

·4

Q.

.3

.2

.1

.s

.g

1·3

FIGURE ld.

2-5

17

2.9

3 7

4 1

Special Exponential Monthly.

.7

POWER

MONTHLY

.6

.5

Vl

z

0

i=

a:

0 0.. 0

a:

0..

.2

·1

.LッZU]セN@ QセU][N@

T]UNZ[セ@

2:::5::::.;:35==;.

;;;,;;;;.6;:5;;;;;;;;;:;:7o:5=•ar5="'."\9r5--,,.o5

FIGURE le.

Power Monthly.

NEG. POWER MONTHLY

Vl

z

·4

0

f-

a:

.J

0 0.. 0

a:

·2

0..

·' FIGURE If.

Negative Power Monthly.

セ@ s

397

1265

398

JG Kallberg and WT Ziemba

1266

J. G. KALLBERG AND W. T. ZIEMBA

ARCTAN MONTHLY

Vl

セ@

i= セ@

3

0.. 0

a:

2

0..

., KLM⦅イUァNSセQWRYtT@

セ W@ FIGURE

lg.

Arctan Monthly.

The seven utility functions used in this study along with the ranges of the various parameter values are described in Table 3. This table also gives the corresponding ranges of the absolute risk aversion and describes the qualitative behavior of the portfolio composition for changing parameter values. For comparison, the theoretically largest ranges of the parameters and corresponding risk aversion are also given. Plots of the optimal proportions of the investor's initial wealth to be invested in each security over various parameter values using monthly data appear in Figures la-lg. (Quarterly and yearly plots, omitted here because of space considerations, may be found in [9].) The optimal portfolio weights were calculated at 21 discrete values and the plots were drawn accordingly. Note the similarity between Figures ld (special exponential) and If (negative power); this occurs because the RA ranges are virtually identical 2.1---* 6.1 and 2.2---* 6.2, respectively. Table 4 summarizes the solutions to the portfolio problems7 classified by RA = 0.5, I, 2, 4, and 10. For each value of RA the portfolio weights across various utility functions are very close and in all cases the percentage error is less than 10- 4 % (recorded as zero). Table 5 provides portfolio The portfolio weights using yearly data for the various utility functions and セN@ weights for different utility functions while not as close as with the monthly data are still reasonably close. They tend in the yearly as well as monthly data to be closer for smaller values of RA because at the higher level of risk aversion the resulting utility function equivalents have more "curvature" hence approximations are less accurate. The percentage error with the yearly data is still quite small and is always less than I%. Similar tables for quarterly data appear in [9], [ 10] and give results intermediate to the monthly and yearly values; the maximum percentage error is less than 0.2%. Of importance is the significance of particular RA values for the composition of the optimal portfolio. The results in Figures la-lg, and Tables 4 and 5, provide guidance regarding the significance of the (changes in) the magnitude of RA. For example, RA values greater than about 4 always yield very risk averse portfolios with low variance, and there is little change in the optimal composition even for large changes in RA. The range 2 < RA < 4 yields moderately risk averse portfolios with a modest degree of change in the optimal portfolio composition when RA changes. The range 0 < RA < 2 yields risky portfolios and there are dramatic changes in the optimal portfolio composition for even small changes in RA . 7 Additional calculations for different Kallberg [9].

RA

values, omitted here because of space considerations, appear in

セ@

TABLE4 Optimal Portfolio Weights for Alternative Utility Functions and RA Values: Monthly Data

Parameter

Exponential

Quadratic

Log

Power

Arctan

Exponential

Quadratic

Log

Power

0.5

0.165483

0.984259

0.879183

2.714659

1.0

0.247237

-0.013835

0.750988

I

0.618500

0.618500

0.621151

0.611083

0.626174

0.414406

0.418291

0.418190

0.410246

2

0.381500

0.381500

0.378849

0.388917

0.373826

0.506770

0.508365

0.508198

0.507673

0.078824

0.073344

0.073612

0.082082

3

セ@ セ@

;:; § t: ::! '"11

4 Security

セ@

5 6

B

7

Negative

Quadratic

Special Exponential

0.440442

8.156772

4.109086

I

0.179684

0.177566

0.169275

0.177531

0.118851

0.095302

0.079358

0.080020

2

0.438363

0.447755

0.440878

0.436866

0.342845

0.364688

0.336837

0.333334

3

0.086301

0.088783

0.089433

0.085550

0.062053

0.068848

0.053716

0.053053

4

0.051708

0.045061

0.054509

0.054076

0.120721

0.129129

0.129992

0.130616

5

0.243944

0.240835

0.245978

0.245978

0.264604

0.278670

0.259364

0.260044

0.044528

0.043953

0.062782

0.062214

6

I

Power

tr1

§ t: セ@

'11

c

z

B 0

7

z "'z

8 9

0.008197

0.009482 i

0.646398

0.019411

0.069754

0.071237

Mean

1.016303

1.016335

1.016260

1.016282

1.015080

1.015229

1.014559

1.014531

Variance

0.002509

0.002525

0.002487

0.002498

0.002113

0.002141

0.001994

0.001989

Expected Utility

0.982493

0.982493

0.942493

0.982493

0.999957

0.999957

0.999957

0.999957

0

0

0

0

0

0

0

"'

10

%Error

セ@

t:

-

-

1 Only those utility functions for which (7) has a solution (i.e., such that RA is attainable) are included for each value of J[A. ' Zeros and blanks indicate values less than w- 6 • 2 For comparison the riskiest portfolio (i.e. highest variance) has mean and variance of (1.016216, 0.014840) and the equal investment portfolio x = (0.1, ... , 0.1) has mean and variance of (1.012607, 0.002232).

セ@

0

z

Chapter 23. Comparison of Alternative Utility Functions

Utility Function

N

o--

\0

401

402

-

!j 0

TABLES Optimal Portfolio Weights for Alternative Utility Functions and RA Values: Yearly Data

Parameter

Exponential

Quadratic

Log

Power

Arctan

Exponential

Quadratic

Log

Power

0.5

0.148687

0.821859

0.907691

2.498551

1.0

0210483

-0.112980

0.765732

0.502147

0.480496

0.509709

0.342082

0.381592

0.287004

0.245520

9

0.497853

0.509913

0.490291

0.421457

0.454136

0.373936

0.339830

0.146545

0.146148

0.163999

0.141238

セ@

I

0.511442

0.518551

2

0.488558

0.481449

3

0.009590

4 Security

!-


JG Kallberg and WT Ziemba

Utility Function

Utility Function

Special

Parameter

I

Security

Negative Power

Exponential

Quadratic

Log

2.0

0.275157

-0.587446

0.350128

0.413450

0.213501

0.174457

0.208418

0.116236

0.149547

0.081275

0.120428

Exponential

Power

2

0.262884

0.300979

0.202592

0.240565

0.167586

0.205169

3

0.126734

0.133008

0.115598

0.129158

0.099687

0.113806

4

0.094320

0.048109

0.159015

0.116267

0.184367

0.157239

5

0.341605

0.309487

0.406559

0.364464

0.417025

0.403389

6

セ@

§

セ@

'"11

セ@

B セ@

7

"'z

8 9 0.050061

10 Mean

1.206489

1.214433

1.193087

1.201245

1.180277

1.193834

Variance

0.053081

0.061813

0.041833

0.048167

0.035125

0.042347

Expected Utility

0.900422

0.900265

0.899992

0.900356

0.898764

0.900039

0.0684

0.1868

0.0286

0.7158

%Error

-

セa@ I""'

セ@

0.1665

I

Chapter 23. Comparison of Alternative Utility Functions

セ@

TABLE 5 (continued) Optimal Portfolio Weights for Alternative Utility Functions and RA Values: Yearly Data

t::i

-...(

403

404

N -..! N

TABLE 5 (continued)

Optimal Portfolio Weights for Alternative Utility Functions and RA Values: Yearly Data Utility Function

Quadratic

Log

Special Exponential

Negative Power

4.0

0.351447

-0.832954

2.884400

1.443557

1

0.088239

0.082991

0.046975

0.021224

0.0476ll

Parameter

Exponential

10.0

Quadratic

0.423749

0.031553

Negative Power

4.935791 0.017099

2

0.169455

0.165982

0.11622

0.185274

0.112794

0.075474

0.050898

3

0.106894

0.106663

0.080160

0.104064

0.079600

0.056237

0.050898

0.041265

4

0.194026

0.198830

0.161247

0.048522

0.154474

O.lll908

0.094011

0.054%6

5

0.441385

0.445533

0.343318

0.44ll82

0.328958

0.237341

0.190958

0.135844

0.044526

0"" セ@ >

E セ@

6 7 8

0.087302

0.143023

0.184158

0.400185

0.470213

0.522144

9 0.252077

lO

0.199733

0.258232

Mean

l.l86170

1.185175

l.l51634

l.l58397

l.l49527

1.125199

1.108230

l.l01730

Variance

0.037743

0.037247

0.024382

0.027802

0.023756

0.017605

0.014485

0.013892

Expected Utility

0.988236

0.988236

0.987863

0.987589

0.987821

0.999969

0.999968

0.999%7

0.7030

0.7099

0.7827

0.2261

0.4728

%Error

-

0

-

----------- ------------

those utility functions for which (7) has a solution (i.e., such that RA is attainable) are included for each value of RA. Zeros and blanks indicate values less than 10- 6 . 2 For comparison the riskiest portfolio (i.e. highest variance) has mean and variance of (1.285196, 0.327557) and the equal investment portfolio x = (0.1, ... , 0.1) has mean and variance of (1.164958, 0.042719). 10nly

Cl

'

セ@

0

セ@ :-l

セ@

N

> ""

JG Kallberg and WT Ziemba

Security

Exponential

Chapter 23. Comparison of Alternative Utility Functions ALTERNATIVE UTILITY FUNCTIONS IN PORTFOLIO SELECTION

405

1273

To illustrate the relative "riskiness" of the various optimal portfolios, consider the monthly data with an exponential utility function. Then the variances at RA = (0.5, I, 2,4, 10) are (0.55,0.4!,0.31,0.24,0.20) times the variance at RA = 0 (i.e., the highest return portfolio). This is further illustrated in Figure 2 which shows the percentage of maximum variance as a function of RA for monthly, quarterly, and yearly data. The results also yield the following conclusions: I. The special exponential and negative power utility functions yield very risk averse portfolios. This is in contrast with the positive power utility function which yields highly risky portfolios and moderately risk averse portfolios for different values of f3 5 • 2. The arctangent utility function usually yields highly risky portfolios and has risk aversion equal to one or less. 3. The quadratic, exponential and logarithmic utility functions yield the largest range in variation of RA and yield the safest and riskiest portfolios for extreme values of /3. Extremely low values of RA are easily implementable for all three utility functions. However, extremely high values of RA are not implementable for the exponential or logarithmic functions because one must either have /3 2 セ@ oo or /3 3 セM w (which causes domain problems). Thus the quadratic utility function despite its two well-known limitations (u' < 0 for large w and rセ@ > 0) may well play a useful role as a computational surrogate for more plausible utility functions when the number of possible investment securities is large, say n セ@ 50. 4. With horizons of a year or less one can substitute easily derived surrogate utility functions that are mathematically convenient for more plausible but mathematically more complicated utility functions and feel confident that the errors produced in the calculation of the optimal portfolios are at most of the order of magnitude of the errors in the data. Moreover there is a fairly well-defined tradeoff of computational accuracy versus computing costs. For rough calculations one can use the measure w0RA (w) and obtain reasonable results for small variance problems (e.g., monthly data). For the most accurate results one uses Rubinstein's measure in (7). The most cost effective measure is w 0 RA. 5. An implication of the results for econometric research is that based only on observed portfolio data it will be extremely difficult to discriminate amongst alternative models since these models will be observationally practically equivalent with different parameter values. 1 0

.g

w z

.?

0

4

;::[

セ@

X 11.

'

0

yearly

FIGURE 2.

Percentage of Maximum Variance vs. RA.

406

JG Kallberg and WT Ziemba

1274

J. G. KALLBERG AND W. T. ZIEMBA

6. The results suggest that one can derive an analytically tractable utility function whose use will generate portfolio allocations nearly indistinguishable from those of a given risk averse utility function given a known distribution of returns. However if we substitute for u when the distribution is F(g) are the results robust if the distribution is really fHセI_@ Calculations in Kallberg and Ziemba [10] indicate that the maximum expected utility and optimal portfolio composition are relatively insensitive to errors in estimation of the variance-covariance matrix. However, errors in estimating the mean return vector do significantly change these quantities. 8

u

Appendix:

Optimality of Rubinstein's Risk Aversion Measure

THEOREM I. Suppose セ@ E R" has a multivariate normal distribution with finite means and variances and that u; is twice continuously differentiable with u; > 0 and u(' < 0 for i = I, 2. Consider the following two portfolio problems (Pl)

{ュ。クeLオ

Q Hキ { ュ。クeLオ

(P2)

Q セtクI@

I e'x =I, x > 0}, and

R Hキ

R セtクI@

I e'x =I, x

> 0},

where W; refers to investor i' s initial wealth. Suppose that x* is an optimal solution to (PI) and that キ Q e\オセHキ

Q セtクJI@

w2E 0,

and

0,

where i

=

0,

I, ... , n.

we have

Now cov(x, g,(y))

=

E(g'(y))cov(x,y),

whenever g is continuously differentiable, all expectations are finite and x andy have

8 This research was partially supported by National Research Council Grant No. 67-7147 and Canada Council Grant No. S75-l307·Rl. Without implicating them, we wish to thank M. J. Brennan, S. Ginsburg, R. C. Grinold, M. J. Gruber, D. Hausch, A. Kraus, M. Kusy, M. E. Rubinstein, G. Sick, and the referees for helpful discussions and comments on earlier drafts of this paper. Special thanks go to M. E. Rubinstein for formulating Theorem l for us and allowing us to publish this joint result here. Rubinstein devised a proof for the case when nonnegativity constraints are absent. The proof given here applies for the nonnegativity case. Cases when some or all of the assets may be sold short or long follow directly from this result.

Chapter 23. Comparison of Alternative Utility Functions

407

ALTERNATIVE UTILITY FUNCTIONS IN PORTFOLIO SELECTION

1275

a bivariate normal distribution. Thus cov( キ Q セ[L@

u;(w 1Vx))

= eセH@

オ[GHキ

Q セtクI」ッカHキ

Q セ[L@

キエeセHオ[G

=

キ Q セtクI@

Q vクIRZ」ッカHセ[GゥL@

(10) J

since

cov(u,Lv;) = Lcov(u,v;). Using (9) and (10) in (AI) yields A+ JL; = キ Q セ[eH@

u;( キ Q セtクI@

+ キゥeセH@

u!(w 1Vx)) 2: 」ッカHセ[@

Gセゥ@

)xi.

J

The Kuhn-Tucker conditions for (P2) are then

"A+ (i;

(A2)

=

キ R セ[eH@

u2(w2Vx)) + wiE 0,

=

0, and

where i

=

I, ... , n.

Now (AI) and (A2) are equivalent to:

(A+

/l;)/( w1E0 dwo

-

as

dR*(wo, Z*) 0, and R*(w0 , Z*) > 0, the result is immediate.

D

THEOREM 3. Under the hypotheses of Theorem 2, the mean return and variance of the optimal portfolio per dollar invested are increasing, constant, or decreasing functions ofinitial wealth as an investor's Rubinstein's measure ofrelative risk aversion is decreasing, constant, or increasing in his initial wealth. That is, (Var (Z*)) > 0

as

dR*(wo, Z*) d < 0; Wo

d (Var (Z*)) = 0 Wo

as

dR*(w0 , Z*) = 0; dwo

as

dR*(wo, Z*) > 0 dwo

(E[Z*]) > 0,

dd

_:!__

(E[Z*]) = 0,

d

/

(E[Z*J) < 0,

dd

(a)

/

(b) (c)

Wo

dwo Wo

Wo

Wo

(Var (Z*)) < 0

where w0 is the investor's initial wealth, Z * is the return per dollar invested and R * (w0 , Z *) is his Rubinstein's measure of relative risk aversion. · PROOF. Taking the derivative of E[Z * J with respect to w0 yields d E[Z*] dwo

dA* (E[X]- E[Y]). dwo

=-

Taking the derivative ofVar (Z*) with respect to w0 yields d -d Var (Z*)

=

Wo

dA* 2 - d (A* Var (X- Y)

+ Cov (X- Y, Y))

Wo

=

2 dA* E[X]- E[Y] dwo R*(wo, Z*)

using (2.5). Since E[X] > E[Y], and R*(w0 , Z*) > 0, by Theorem 2, the proof is complete. For the exponential utility function u(x) = -e-#x (fl > 0), R*(wo, Z*)

=

wofl.

=

f1 > 0.

D

Thus dR*(w 0 , Z*) d Wo

Therefore, Theorems 2 and 3 imply that if an investor has an exponential utility function,

Chapter 24. Characterizations of Optimal Portfolios CHARACTERIZATIONS OF OPTIMAL PORTFOLIOS

415

265

as his initial wealth increases, the optimal allocation of his portfolio with two risky investments have bivariate normally distributed returns to the investment with the higher mean return will decrease. Thus the mean return and the portfolio variance fall. For other utility functions, it is usually difficult to determine the increasing, constant or decreasing property of Rubinstein's measure of relative risk aversion in initial wealth. Fortunately, however, as Kallberg and Ziemba ( 1983) illustrated, for small variance problems (e.g., daily, monthly, or quarterly return data), Rubinstein's measure ofrelative risk aversion can be approximated by the Arrow-Pratt measure of relative risk aversion to obtain very accurate optimal portfolio weights and virtually identical expected utility within the limits of data estimation errors. 6 Thus in these cases, Rubinstein's measure of relative risk aversion can be expected to have the same increasing, constant or decreasing property in initial wealth as the Arrow-Pratt measure of relative risk aversion for any particular utility function. For example, as the logarithmic utility function u(x) = log (x + {3) has decreasing, constant or increasing Arrow-Pratt measure of relative risk aversion as f3 < 0, f3 = 0, or f3 > 0, respectively, Rubinstein's measure of relative risk aversion will also have these properties for the portfolio problems with investment horizons one quarter in the future or less. Therefore, for investors with the logarithmic utility functions, the wealth effect on optimal mixtures and risk-return characteristics of portfolios can be determined by Theorems 2 and 3, if the portfolios consist of two investments with bivariate normally distributed returns and the investment horizons are in a suitable range. 3. Risk Aversion Matrix and Optimal Portfolio

We start with a description of the portfolio model with a multi-attributed utility function. Assume that there are two risky investments with nominal returns X and Y and K factors FT = (Fh F 2 , ••• , FK) which affect the real value ofwealth. 7 The two returns and K factors are assumed to have a joint normal distribution. Consider that an investor is to allocate his initial wealth w0 between the two investments. We assume that in the existence of the uncertain factors, a risk-averse investor's preference for the real value of wealth can be represented by a multi-attributed increasing concave utility function u( w, F), where w represents the nominal value of the final wealth, and F is the vector of factors affecting the real value of wealth. Assume also that the investor's objective is to maximize the expected utility of the real value of his portfolio at the end of his investment horizon. Let a be the amount of dollar allocation of the initial wealth w0 to the investment with return X. Then his decision problem is max V(a) == E[u(a(X- Y)

+ w0 Y,

F)].

(3.1)

O.:o:;aswo

This model is an extension of the portfolio selection model in §2 and is developed from a generalized CAPM model in Losq and Chateau ( 1982). Suppose that the optimal allocation of the portfolio is a non-trivial allocation of investments. Differentiating ( 3.1) with respect to a yields the first-order condition V'(a) = E[uw(W*, F)(X- Y)] = 0

(3.2)

6 The approximation error depends nonlinearly on the variance of final wealth. When the variance of final wealth approaches zero, the average of the Arrow-Pratt measures converges to Rubinstein's measures. Although Kallberg and Ziemba (1983) have given empirical estimates of this error, theoretical approximations and/or bounds have not been developed. 7 For example, consider a portfolio problem with a mixture of investments in a mutual fund and a money market account with interest-rate risk, where the future inflation rate could be one of the factors.

416

Y Li and WT Ziemba

YUMING Ll AND WILLIAM T. ZIEMBA

266

=

where W* a* (X- Y) =Bu(w, F)jaw. Since

+ w0 Y is the terminal

value of the optimal portfolio, and u"'

Cov (A, B)= E[AB]- E[A]E[B] for random variables A and B, (3.2) can be written as Cov (uw(W*, F), X- Y)

=

-E[uw(W*, F)]E[X- Y].

(3.3)

We now present the key tool for the analysis in the multivariate case, which is the multivariate version of the Stein/Rubinstein covariance operator generalized by Gassmann (1987). LEMMA 4 (Multivariate Covariance Operator). Let X, Y be multivariate random vectors on R"' and R", respectively, with joint distribution

Let g: R" セMK@ Rk be differentiable with Jacobian matrix Jg of dimension k X n. Assume that Cov (X, g(Y)) and E[Jg(Y)] exist. If lim g(yo

+ ty).fv(Yo + ty)

=

(3.4)

0

t-±w

for all y, Yo E R", where fv is the probability density of Y, given by

JiY ( y) -_ ( 2 ·n-)" /2 1I Lv I 1/2 exp( -

1

2 ( y - !LY)

7

-1

Lv ( y - !LY ) ) '

then Cov (X, g(Y))

=

Cov (X, Y)E[Jg(Y)]T.

(3.5)

MULTIVARIATE REGULARITY CONDITION. Let u(x): R" セMK@ R be a multi-attributed utility function. Then u is said to satisfy a multivariate regularity condition if the conditions of multivariate covariance operator are satisfied with the gradient of u, \lu in place of g. LEMMA 5. Assume that the bivariate random variables X, Y, and the random vector F of dimension K have ajoint normal distribution. Also assume that u(w, F): RK+I セMK@

R 1 is an increasing concave utility function which is twice continuously differentiable and satisfies the multivariate regularity condition. If 0 < a* < w0 , then the optimal allocation of the portfolio a* satisfies

+ w0 Cov (Y, X-

Var (X- Y)a*

Y)

E[X -* Y] Rww(W , F) where (R"'"'' RwF1 , aversion 8 R(W*

'

••• ,

ォセャ@

f Cov

(Fk. X- Y) RwFk(W:, F)

(3.6)

Rww(W , F)

RwFx) is the first row of the matrix measure of absolute risk

F)=(- E[uu(W*, F)]) E[u;(W*, F)]

for

i,j = W, Fl' ... 'FK,

U; is the first-order partial derivative of u with respect to the ith dimension of u and uu is the second-order partial derivative of u with respect to the ith and jth dimensions.

8

See Li and Ziemba ( 1987) for the derivation and the properties of this measure.

Chapter 24. Characterizations of Optimal Portfolios

417

CHARACTERIZATIONS OF OPTIMAL PORTFOLIOS

267

PROOF. Applying the multivariate covariance operator to ( 3.3) and rearranging terms yields ( 3.6). The proof appears in Appendix. D Consider the portfolio problems for two investors with the same initial wealth w0 • Assume that u 0. Now (2) is a parametric concave program which would generally be difficult to solve, for all fJ > 0. It is easier to bypass the calculation of the J-t-d curve if the risk-free asset is not available (see Section IV). However, when the risk-free asset does exist, as we are assuming, it is convenient to consider points that are linear combinations of x 0 = 1 (total investment in the riskfree asset) and points that lie on cp({J). These combinations correspond to straight lines in J.l-d space as well because the mean is linear and the a-dispersion measure is positively homogeneous [i.e., f(U) = J...f(x) for all A. ;;:; OJ in the X;. Clearly the best points will lie on the line L that is a support to the concave function cp ({J) (proof below); see Fig. 1. Such heuristic arguments indicate that the J-t-d efficient surface is now L, and one gets the analog of the Tobin separation theorem for normal distributions which has the important implication that x;* JL.'J= 1 x/, i-# 0, is independent of u and of initial wealth. The result may be stated as the Separation Theorem (2')

max('x,

Let the efficiency problem be

s.t. f(x)

セ@

fJ,

e' x = w,

i ;;:; 0,

x 0 unconstrained.

In (2') f is the a-dispersion measure, w is the initial wealth, and explicit consideration of the risk-free asset (i = 0) is allowed. Suppose that asset i = 0 has no dispersion and that borrowing or lending any amount is possible at the fixed rate ( 0 . Assume thatfis convex and homogeneous of degree I and that u is concave and nondecreasing. 7 Samuelson [29] analyzes a problem similar to (2) in which one minimizes dispersion subject to the mean equalizing a given parameter. He notes that his problem is a convex program and he analyzes it using the Kuhn-Tucker conditions. His paper also contains some illustrative graphical results for some special cases. See also Fama [5] for a similar analysis, presented in the context of the Sharpe-Markowitz diagonal model, that is particularly concerned with diversification questions.

Chapter 25. Choosing Investment Portfolios

429

CHOOSING INVESTMENT PORTFOLIOS

a -dispersion

Fig. I

(a) Total separation: If x 0 * =1= 0, the relative proportions invested in the risky assets, namely x;*/L.j= 1 x/, i =I= 0, are independent of u and initial wealth w. (b) Partial separation: If x 0 * = 0, all investment is in the risky assets and L.}= 1 x/ = w and the x/ are independent of u. Proof (a) Suppose x 0 * =I= 0. Since x* solves (2') it must satisfy the Kuhn-Tucker conditions: (i) (ii) (iii)

ヲHセI@

セ@

セ@

fJ,

セ@

0,

セ[MスNHッヲOクI

(a) (b) セッ@

- J1

= 0,

ク[HセMaNッヲャIjL@

e'x = w, Jl セ@ 0, i = l, ... ,n. }.. セ@ 0, = o, i = 1, ... ,n.

It will suffice to show that for all u > 0 there exists a y =I= 0 such that x** = (yx 0 *,ux 1 *, ... ,ux.*) also solves the Kuhn-Tucker conditions (which are necessary and sufficient) for all P** セ@ fJu. Condition (i) is satisfied because ヲHセJI@

セ@ セJ@

セ@

p

=>

0

=>

= uf(x*) ヲHオセJI@ PGセJ@

セ@

+I X;*= i= I

ufJ

セ@

0,

n

Xo*

セ@

n

=>

YXo * + 0'

I

i=l

X;*

= 1

P**,

430

WT Ziemba W. T. ZIEMBA

Now ofjox; is homogeneous of degree 0, hence i=.

iff セN@ - Aof(x*) < セ@

_A of(ux*) :5 e

.,,

"' uX;

-

0, (iii) is equivalent to (ii(a)) which is satisfied for u.X* iff .X* satisfies (ii(a)). By Theorem 2 an optimal solution to (2') must solve (1); hence Li= 1 x;* = w and the x;* are independent of the u. The proof and statement of the separation theorem given here are similar in spirit to that given by Breen [2]. Breen considered an efficiency problem in which one minimizes 」クセ、ゥウー・イッョ@ given that expected return is a stipulated level as well as some alternative assumptions regarding the risk-free asset. For the analysis indicated in Fig. 1 to be valid it is necessary that the aQ dispersion measuref(x) {2:1= 1 s[クBス セ@ be convex and that cp be a concave function of {3. The function f is actually strictly convex as we now establish using

=

Theorem 3

(Minkowski's inequality) Mr(Y)

Suppose

={(1/n) it y{rr.

n

1

セ@

r > 1,

n


Mr(a+b).

Proof See, e.g., Hardy et a/. [13, p. 30].

Let P = {.X INxセ@ Lemma 1 f(x)

2

セ@

IX

0, e'x = 1}.

= {2:7= 1 S;xt} 11a

is a strictly convex function of .X on P if

> 1 and S; > 0, i = 1, ... , n.

Proof Let a; = AS; x/ and b; = (1- A) S; x/, i = 1, ... , n, where .X 1 =1=

and 0 < A < 1. Then by Minkowski's inequality

{o!n) it1 。[Bイセ@

+ {o!n)

it1 「[Bイセ@

>

{o!n) Jl H。[Kィエイセ@

x2 ,

Chapter 25. Choosing Investment Portfolios

431

CHOOSING INVESTMENT PORTFOLIOS Q Hence セヲHク I@ + (l-'A)f{x 2 ) > f{'Ax 1 + (l-'A)x 2 }. Thus f is strictly convex unless a and b are proportional. But this requires that there exist constants q 1 and q 2 not both zero such that q 1 a= q2 b or that

q 1 A.x/ =q 2 (1-A.)x/

=>

S; > 0)

(since

or that x/ and x/ are proportional since q 1 ' = q 1 ). and q 2 ' = q 2 (1- 'A) are both not zero. However, the condition that L:7= 1 x/ and L:7= 1 x/ = 1 means that x1 and x2 cannot be proportional unless x/ = x/ for all i, which is a contradiction. Remark f is convex but not strictly convex on M = is linear on every ray that contains the origin.

{xI x セ@

0} because f

Lemma 2 qJ is a concave function of f3 > f3L, where f3L > 0 is defined below, if 2 セ@ rx > 1 and S; セ@ 0, i = 1, ... , n.

Proof Let Kp ={xI e'x = 1, x セ@ 0, j{x) セ@ /3}. Let PL> 0 be the smallest f3 for which Kp # 0. Clearly PL セ@ f>, where ヲ^セ@ =min; S;, and Kp # 0 if and only if f3 セ@ PL· Choose /3 1 セ@ PL and /3 2 セ@ f3L· Since (2) has a linear objective function and a nonempty compact convex feasible region there is an optimal solution for all f3 セ@ PL· Let optimal solutions when f3 equals /3 1 and /3 2 be x1 and x2 , respectively. Consider

/3;. = x;. =

0

(1-A.)/31 +A.f3z,

セIN@

セ@

1,

(l-A.)x 1 + AX 2.

Now X;. is feasible when f3 = By the concavity of h(x) = セGクL@

/3;.

because the constraint set of (2) is convex. (i)

But h(x;)

since

x; is optimal,

i

=

(ii)

qJ (/3;)

= 1, 2, and (iii)

since

({J(/3;.)

denotes the maximum when f3=

qJ ({3 ;.) セ@

(1- A.) qJ (/31)

/3;..

Combining (i)-(iii) gives

+ AqJ (f3z).

432

WT Ziemba

W. T. ZIEMBA

One may find the slope of Land the point M (see Fig. 1) by maximizing

(3) g (xA)

セ[xMッ@ S ·"} 11,. , = サセ@ Li'=t L..•=l ;X,

By letting セ@

=

セM

0

"

s.t.

A

;;;;

,

P L@ i = 1, ... ,n, and utilizing g(x) =

{

セ@

n

セ[ク@

_

}

I{ セ@

n

= 1. ・Gセ@

e'x = \ 1/IZ

S;x;"'J

1,

.

We will assume that セGク@ > 0 for all "interesting" feasible x. This is a very minor assumption since the fact that the x ;;; 0 and e'x = 1 implies that there always exists an x such that セL@ x > 0 unless all セ@ セ@ 0 in which case it is optimal to invest entirely in the risk-free asset. It will now be shown that (3) has a unique solution. A differentiable function O(x): A-> R is said to be strictly pseudo-concave on A c Rn if for all x, .X E A, x #- .X (x-x)'VO(.X)

セ@

B(x) < O(.X)

=

0

[V denotes the gradient operator so VB= (88j8xt> ... ,aejaxn)']. Geometrically functions are strictly pseudo-concave if they strictly decrease in all directions that have a downward- or horizontal-pointing directional derivative. A normal distribution is such a function in R. Theorem 4 Suppose the differentiable functions '¥ and 1/J are defined on A, a convex subset of Rn, and that'¥ > 0 is concave and 1/J > 0 is strictly convex. Then () = '¥ 11/1 is strictly pseudo-concave on A. Proof VO = {(1/!V'¥- 'I'VI/1)/1/1 2 }. Let .X EA. Thus = {1/J(.X)V'I'(.X)-'I'(.X)VI/I(.X)}'( --) < O V()( X-)'( X --) X [I/J(x)] 2 X X = ,

X E

A,

X#-

X

= {1/!(.X)V'I'(.X)-'I'(.X)Vt/J(.X)}'(x-x)

セ@

0

1/J (.X)['¥ (x)- '¥ (x)J - ['¥ (x) Vt/J (x)]' (x- x) セ@

1/J(x)['l'(x)-'l'(x)J

+ 'l'(x)[I/J(x)-1/J(x)J
0 which implies that

0) ·

since 1/1 (x) > ( andlj!(x)>O

Chapter 25. Choosing Investment Portfolios CHOOSING INVESTMENT PORTFOLIOS

Lemma 3 Suppose/: A--+ R, where A is a convex subset of Rn, is differentiable at x, and there is a direction d such that Vf(x)'d > 0. Then a u > 0 exists such that for all r, u セ@ r > O,f(x + rd) > f(x). Proof See Zangwill ([32, p. 24]. Theorem 5 Suppose O(x): A--+ R, where A is a convex subset of Rn, is strictly pseudo-concave. Then the maximum of 8, if it exists, is attained at most at one point x E A. Proof Case I: 3x E A such that VO(x) = 0. Then by the strict pseudoconcavity of(), i.e., V()(x)'(y-x) セ@ 0 セ@ ()(y) < O(x) for ally E A, x is the unique maximizer. Case 2: セク@ E A such that VO(x) = 0. Suppose x maximizes e over A. Then Vy =1- x, V8(x)'(y-x) セ@ 0 セ@ ()(y) < ()(x) and x is clearly the unique maximum unless 3y E A such that V()(x)'(y-x) > 0. But by Lemma 3 that would imply that there exists a r > 0 such that O(x+r(y-x)) > 8(x), which contradicts the assumed optimality of x. (Note that the point [x+r(y- x)] = [ry+(l-r)x] E A, by the convexity of A.) Theorem 6 (3) has a unique solution. Proof The function g may be seen to be strictly pseudo-concave on P by letting \f'(.X) =''.X, which is positive and concave, and tf;(.X) = {L7=t S;xt} 11 a, which is positive and strictly convex. Hence by Theorem 5, it has at most one maximizer. But P is compact and g is continuous; hence g is maximized at a unique point x* E P.

Suppose that the fractional program (3) has been solved to obtain .X*. The problem then is to determine the optimal ratios of risky to nonrisky assets. The risky asset is

R

]セGNxJB@

fHセGクJ[@

t

セGクJL@

S;(.X;*Y,O,IX) =: F(R; R,SR,0,1X),

•=1

=

where セ@ ¢ 1 , ... , ¢., and the best combination of Rand the risk-free asset may be found by solving max |ヲGHaNI]erオ{セ

(4) M」ッ\NセA@

P KHQMaNIr}@

Problem (4) is a stochastic program with one random variable (R) and one decision variable (A.). Since u is concave in w it follows that \f' is concave in A.. Under the assumptions of Theorem 1, \f' (A.) and d\f' (A.)/dA. will be bounded.

433

434

WT Ziemba W. T. ZIEMBA

Now (5)

'P(.A.) = =

P KHQMNaIr}ヲ@

J_'Xloo オ{Naセ

t:

dR

オ{NaセッKHャMIサrs Q

。@

R.}]j(R.) dR.,

where the standardized stable variate., 1{

R-R =(SR)lfa セ@

F(R.; 0, l,O,cc).

The continuous derivative of 'I' is (6)

Joo go-R-(SR)lfa R.} -oo

d'P(.A.) = d).

X

、オ{NaセッKHャMIサrsヲ。@

R.}] J(R) dR..

dw Problems having the general form of (4) are generally easy to solve by combining a univariate numerical integration scheme with a search procedure that uses function or derivative evaluations (see Ziemba [35]). The difficulty here is that the density j(R.) is not known in closed form except for very special cases such as the normal (ri = 2) and Cauchy (ri = 1) distributions. However, Fama and Roll [7] have utilized series expansions ofj(R.), due to Bergstrom, to tabulate approximate values of j(R.) and F(R.). The tables consider \i = 1.0, 1.1, ... ,2.0 and R. = ±0, 0.05, ... , 1.0, 1.1, ... ,2.0, 2.2, ... ,4.0, 4.4, ... , 6.0, 7.0, 8.0, IO.O, 15.0, and 20.0 (a grid of 50 points). One may then utilize the tables to get a good approximation to (5) of the form (7)

m

Lp

'P(.A.) セ@ jセャ@





P KHQMNaIサrs オ{Naセ

where p Q ]pイサrN Q p Q ]QL@ Q ス^PLlZェセ symmetric). The approximation to (6) is (

8)

d'I'(.A.) セ@ d A1 -

ヲ。r⦅I}L@

and m=lOO (recall that

セ@ P.{fl -R-(S )lfaR.} 、オ{aセッKHャMNIサrsヲ。⦅jス}@

Qセ@

.L...

1

R

'>O

d

1

W

R

1s



One may then obtain an approximate solution .A.a to (5) via a golden section search using (7) or a bisecting search using (8) (see Zangwill [32] for details on these search methods.) The approximate optimal portfolio is then 8

Xo 0 = A0 キィッウ・セ[Z@

8

and

xt =

HャMa P

IセェL@

i = 1, ... ,n.

Since the ,; 1 are independent it is known (see Samuelson [28]) that x1* > 0 for each i ュゥョャZjGBセᄋ@

Chapter 25. Choosing Investment Portfolios CHOOSING INVESTMENT PORTFOLIOS

ID. The Dependent Case

=

A set of random variables r (r 1 , •.. ,r.) is said to be multivariate stable if every linear combination of r is univariate stable. The characteristic function of r is t/l,(t)

=t/l,(tl> ... , t.) =

Eit'r

=

s_: . .

J_""oo eit'r g(r) dr 1 ... dr.,

where g is the joint density of (r 10 ... , r.). If g(r) is symmetric, then the log characteristic function (see Ferguson [10]) is ln t/J,(t) = ib (t)- y(t), where the dispersion measure y(A.t 1 , ... , A.t.) = lA. Ia y (t 1 , ... , t.), i.e., positive homogeneous of degree a, y > 0, and the central tendency measure 0 if GiHセI]サA@ l:j= 1 HセGョ Q クIG QR ス Q BG@ is a convex function of x. The pertinent facts relating to the convexity of 'I' are summarized in Theorem 7 (a)

(Independent case)

Suppose m =nand thejjth element of

n1 is w1 > 0 (and all other elements are zero). Then 'I' is strictly convex on P.

(b) (Totally dependent case) Suppose m = 1 and 0 1 is positive definite. Then/is strictly convex on any convex subset of R". (c) (General case) Suppose n G; m G; 1, each n x n matrix is positive semidefinite, and for at least one j, x'0.1 .X > 0 if x セ@ 0. Thenfis convex on M. 9 Press [23] also proves that Ar+b is multivariate symmetric stable of form (9) if A is any m x n constant matrix and b is any n vector of constants.

Chapter 25. Choosing Investment Portfolios CHOOSING INVESTMENT PORTFOLIOS

Proof (a)

Let S1 > 0 be defined by w1 'P(x)

= {-t ェセONクGョ

Q クIG QR



= (IJ/l)Sj1 2 • Then Q@

..

=

jセ@

s1 x/,

which was shown to be strictly convex on P in Lemma I. (b)

Hence '¥ will be strictly convex if

0 and b/ =(ai

1 , ••• ,ain);

then

'PI (x) = ttl cj iセG@ bjr} .lfa 'P

Suppose that 1 is defined on Rn and that for some j, iセG@ bi =1- 01 ゥヲセ@ =1- 0. Hence 'P 1 is essentially positive and positively homogeneous on the convex cone Rn. Thus by Theorem 8, is convex if and only if it is quasi-convex. Now 'P 1 is quasi-convex if and only if N サセi@ サlセ]@ セ@ q} is 1 ci ャセG「ゥ。スO@ cJx'bil セ@ q'} is convex for all convex for all q セ@ 0, or equivalently, if サセil]エ@ q' qa セ@ 0. Since each ャセG「ゥ。@ is convex and the ci > 0, the sum iZセ]@ I cJx'bila is convex, hence the set N is convex.

'P 1

=

=

IV. Dropping the Risk-Free Asset Assumption

Let us reconsider problem (I)

Maximize Z(x)

= eセオHGクIL@

s.t.

e'x = I,

x

セ@

0.

Suppose エィ。セ]@ Hセ P L@ ... , セョy@ has a symmetric multivariate stable distribution G with characteristic exponent r.x (I < r.x セ@ 2), having the log characteristic function ャョゥOjセHエI@ = ゥセGエM y(t), where y(t) is positive homgeneous of degree rx. Suppose that a risk-free asset does not exist. Then one may (approximately) solve (1) as follows. 10 For any feasible x, say xk, セG@ x has the univariate symmetric stable distribution fHセGクォ[@ セGクォLケHIoイN@ Now (13) Z(xk)

=

f·J オHセGクォI@

、gHセI@

= jオ{セGKサケHクォIスャヲ。⦅r}N、L@ where the standardized variate

10 It may be noted that the following procedure is also applicable when the risk free asset does exist but the lending and borrowing rates differ. Let x 0 セ@ 0 and x1 セ@ 0 be the respectively. Then levels of borrowing and lending activities at constant rates セ P@ 。ョ、セB@ KセRx@ ᄋKセョx@ and e'x becomes -xo+Xl + ··· +x•. in (1) e'x becomes MセッxKQ@

440

WT Ziemba

W. T. ZIEMBA

The continuous partial derivatives of Z are (14)

oZ(xk)

= ヲサセゥ@

R}

+ セ{ケHクォI}ャM」ヲッ@

OX;

OX;

r:J.

x

、オ{セGクォKサケHIス



dw

ヲ」ク@

R] J(R)dR,

i

= 0, ... , n.

Now using the Fama-Roll tables, as in Section II, one may obtain a good approximation to (13) and (14) via (15) Z(xk) セ@

m

L pゥオ{セGクォKサケHwヲ」}L@ j=l

and

i = 0, ... ,n,

respectively, where Pi= Pr{R = RJ} > 0, L,j= 1 Pi = 1, and m = 100. One may then apply any standard nonlinear programming algorithm that uses function values and/or partial derivatives to solve (1) approximately, utilizing (15) and/or (16). If one uses an algorithm that utilizes only function values, such as the generalized programming algorithm (see Ziemba [33]), then for each evaluation of (IS) one merely performs m function evaluations of the form u(·) and adds them up with weights Pi. The evaluation of the portfolio is more complicated since for each i (i = 1, ... ,n) one must perform m function evaluations of the form { ·} du( · )Jdw and add these up with the weights Pi. One would suspect that it would be economically feasible to solve such approximate problems when there are say 40-60 investments, the grid m is say 20-50 points, and u, y, and k are reasonably convenient. It is possible, of course, to apply this direct solution approach even when the risk-free asset exists. However, the two-stage decomposition approach appears simpler because one must solve a fractional program in n variables plus a nonlinear program in one variable that has m terms. In the direct approach one must solve one nonlinear program inn variables having m terms that may fail to be concave or pseudo-concave (because u is concave and y is convex). Some numerical results are given by Ziemba [34].

Chapter 25. Choosing Investment Portfolios

441

CHOOSING INVESTMENT PORTFOLIOS

Appendix Proof of theorem 1

(a)

If

u(w) dF(w)i

セ@

IJu(w)J dF(w)

JL JwJP• 、Oセ^@ \セ^@

oo.

1

(iii) is a well-known property of stable distributions (see, e.g., Feller [9]). (ii) is by assumption, while (i) follows because the middle term is finite (see Loeve [16]). (b)

II

du(w)fdw dF(w)i

セ@

IJdu(w)fdwJ dF(w)

l R jキp セ^i@

R HゥセI@

oo.

(i)-(iii) follow for the same reasons as in (a) since (b) is a special case of (a). (c)

Now w

=セGクLN@

2:i'=o Sixt,O, IX) if and

FWx; セGクL@ _ w

={"n.

only if

_

キMセGク@

s,x, . .«} 1 /«,... F(w; 0, l,O,IX).

セᄋ]P@

Notice that F corresponding to the standardized variable w does not depend on x. Now

iオHセGクKウIキ、f[oLャx@

Ewu(w) = Iu(w)dF(w) =

=

where s(x) {2:1= 0 Si xt} 11«. Now if it is legitimate to differentiate under the integral sign aEu(w) =I 、オ{セGクKウHI@ ::l uxi

d

w] (l'.

.'( ) -) d (-. 1 0 ) ... + s, X w F w' 0' ' 'IX

w

= セゥ@ i、オ{セGク[ZHIキ}@

dF(w; 0, l,O,IX)

, I_du[e'x+s(x)w] _ +si(x) w dw dF(w;O,l,O,IX)

where s/(x)

1{

n

=- L IX

Sixt

}(1-«l/« xi-

1



i=O

Now under the assumptions セゥ@ and s/ are finite; hence aEu(w)jaxi is finite if the latter two integrals are finite, which they are using an argument as in (a) and the assumption on u'. Note that Ju'(w)wJ セ@ JwJfl if and only if Ju'(w)J セ@ JwJfl- 1 and that Ju'(w)J セ@ JwJfl- 1 implies that Ju'(w)J セ@ JwJfl for f3 > 1. Also

442

WT Ziemba W. T. ZIEMBA

one may differentiate under the integral sign because u is continuously differentiable and the absolute expected marginal utility is bounded (see Loeve [16]). (d) The proofs remain valid as long as the appropriate absolute integrals remain finite. We show here for case (a) that J\u(w)\ dF(w) < oo; similar arguments may be used to establish (b) and (c). Suppose v 1 > 0. Then without loss of generality we may take v 1 = 1 (and L 1 = 1). Now limlwl-oo(\u(w)\/\w\ 11) = 1 if and only if for all e such that 0 < e < I, 3N1 > 0 such that 'v' \w\ セ@ N 1 , l\u(w)\/\w\P'-1\ < e, or 1\u(w)\-\w\P'I < e \w\P' hence Ju(w)\ 32 . So the complexity of I: is Cm. Consider B3, the truncation estimator with number offactors m* < m. (i) If m \セGエィ・ョ@ RSL(Os) ::( 1. (ii) If m > セG@ 2 , such that RSL(ils) ::( 1 for 61 ;, then there exists a value 0 ::(

6f,

6f ::( ()

6f.

The truncation estimator is based on a simplification of the covariance structure. When the covariance is less complex, that is, the number of latent factors is small relative to the number of assets, the effect of truncation cannot make the additional risk greater than that for the mean rate of return. This is true for maximal complexity and therefore holds in general. If the number of factors is large, then in the most complex case the truncation estimator dominates if the idiosyncratic variance for each asset is close to the variance of the common factors. Because the maximum conditional complexity C111 is approaching the absolute maximum, the advantage of the covariance structure and the truncation estimator is small. Example 2. To illustrate the effects of complexity and truncation, consider an example with K = 8 assets. The covariance structure is determined by the matrix Am = m-1Em, where Em is defined by the first m columns of the matrix

E=

-1

-1

-1

-1

-1 -1

-1 -1

-1

-1

-1

-1 -1

-1

-1

1

-1

1

-1

-1

-1

1

-1

-1 -1

-1 -1

1 -1

-1 -1

-1

Chapter 26. Covariance Complexity and Rates of Return on Assets

455

L. C. MacLean et al. I Journal of Banking & Finance 31 ( 2007) 3503-3523

3513

The error variance is () 2 = I. The variance of each log-rate is 2 and tr I: = 16. The eigenvalues of I: are qJ = セ@ + I, with multiplicity m, and cP = I, with multiplicity K- m. This is the situation where conditional complexity (given (tr I:, m)) is maximal and the truncation error is greatest. In Table 1, the relative savings loss for the truncation estimator, with the range of values of m and m*, is presented. The cases in Table 1 present the most challenging structure. However, there is still an advantage in using the covariance in the estimation of the expected rate of return. As expected, including more factors (less truncation) improves the relative performance of the empirical Bayes (structural) estimator. The choice of a smaller number of factors (more truncation) gives results which are quite consistent across the models as m increases. 3.2. Monte Carlo study - NYSE and TSE

The analysis of theoretical risk has established the relative advantage of the truncation estimator which links the means to the covariance. To demonstrate the effect numerically, a Monte Carlo study was carried out using two data sets of actual asset returns: (i) end of month prices for 24 leading stocks from the Toronto Stock Exchange (TSE); (ii) end of month prices for 24leading stocks from the New York Stock Exchange (NYSE). For both markets, the data covers the years 1990-2002. The correlation structure for returns is different for the stocks from the two exchanges, so a performance comparison of estimators shows the significance of structure. The percent of variance accounted for by the top five eigenvalues for each correlation matrix is shown in Table 2 as is the covariance complexity for returns on each exchange. The NYSE stocks have a much more complex covariance. From the data on monthly closing prices for the set of 24 stocks on the Toronto Stock Exchange and the separate set of stocks on the New York Stock Exchange, the natural log of gross monthly rates of return were calculated. The mean vector and covariance matrix Table 1 Relative savings loss with truncation: maximum complexity case m*

m 2

3

4

5

6

7

0.9575

0.9219 0.9804

0.9007 0.8809 0.8581

0.8995 0.8769 0.8490 0.8154

0.9249 0.9061 0.8817 0.8419 0.8055

0.9252 0.9391 0.9214 0.8960 0.8572 0.7921

1.0 1.0 1.0 1.0 1.0 1.0 1.0

0.85

0.90

0.94

0.96

0.98

0.99

1.0

2 3 4 5 6 7

em

0.74

Table 2 Leading eigenvalues (% of variance)

NYSE TSE

q)]

tP2

q,,

¢4

RSL(B 8 ) ifO < /1 < 1.

The wealth loss depends upon the risk aversion at the time of decision. In the decision rule, the risk aversion parameter fi defines a fraction of capital invested in the optimal growth portfolio: Q セ[イ@ When f1 < 0, the control of decision risk also reduces the impact of estimation error. Correspondingly, when f3 > 0, the overinvestment increases the effect of estimation error. In general, the improvement in parameter estimation translates into better wealth performance. If the number of factors is unknown, then additional error from truncation will be included in the estimator and the investment decision. Consider the bias introduced by truncation

(39) If this bias is sufficiently small, then the relationship between wealth loss and estimation error is retained. Proposition 6. Consider assets with price dynamics defined by ( 1) and (2), and investment strategy defined by (34). Let OB be the truncation estimate with number offactors m* < m. If the bias for the truncation estimate satisfies , I o r ) B(B·) RSL(Bll) ifO < fJ < 1.

f3 < 0; (ii) RWL(X8 ) = RSL(08 ) if /1 = 0; (iii)

For the most complex case, where the unobserved factors are equally important, the ordering on estimation loss generates an ordering on wealth loss for alternative investment strategies when the number of factors in the price dynamics is truncated. 4.2. Application to N YSE

The dynamic investment process is now implemented with the data from the New York Stock Exchange. The approach is to forecast the prices for the next month using the truncation estimator (and for comparison with the Bayes-Stein estimator), then calculate the Kelly or growth optimal investment strategy with risk index fJ = 0. The amount set aside for investment in each month is held at $1000. So the strategy is constrained. The return on investment is from the actual return for the next month. The final wealth is the accumulation of returns on the monthly investments. Advantages of this methodology are: (i) the investment capital is the same for both estimation procedures; (ii) each month is the same except for the parameter estimates.

Chapter 26. Covariance Complexity and Rates of Return on Assets

461

L. C. MacLean eta/. I Journal of Banking & Finance 31 (2007) 3503-3523

3519

Table 4 Wealth comparison using altemative estimators Month 2 3 4 5 6 7 8 9 10 II

12 13 14 15 16 17 18 19 20 21 22 23 24

Bayes-Stein

Tmncation

Ratio

1000.00 2753.84 4734.16 6155.03 7072.46 8800.46 8057.76 9839.54 9915.73 12373.73 12945.05 14734.91 15114.73 15905.18 15753.94 16988.85 17817.41 16744.71 17483.44 19935.02 21354.56 25299.94 24760.45 28880.96

1000.00 3114.25 5755.60 7495.27 8635.16 10852.75 9693.90 12188.45 12495.84 16287.52 17198.59 19844.46 20335.51 21378.58 21355.97 23065.39 24258.14 22670.64 23516.54 26538.49 28376.71 33494.38 32790.56 37855.50

1.00 1.13 1.22 1.22 1.22 1.23 1.20 1.24 1.26 1.32 1.33 1.35 1.35 1.34 1.36 1.36 1.36 1.35 1.35 1.33 1.33 1.32 1.32 1.31

The segment of the data series from 1993 to 1996 is used. The forecast for rates of return for one month ahead are developed from the previous 24 months of prices, and they are used to determine the investment strategy. The return for the next month is calculated from the actual prices. This backcast is worked forward from January 1995 to December 1996, with the resulting accumulated capital shown in Table 4. The returns are very large, as is characteristic of that period, and although $1000 is set aside for investment, the actual amount in securities is much higher, since capital is borrowed to invest in high return securities. That is a property of the capital growth strategy- very risky, and is part of the rationale for caping the seed fund at $1000. The investment strategy is a vehicle for considering the estimation error, and the comparison of estimators rather than the size of returns is the focus. The important statistic is the relative return for the comparative estimators. The superiority of the truncation estimator, which was demonstrated in the simulations, yields superior strategies and greater wealth. The 24 month test interval is appropriate for comparing estimators since the return dynamics are consistent over that time. The time frame does not include the market collapse following the overvaluation of securities. 5. Conclusion

The estimation of the rates of return on assets is a critical ingredient to a successful investment strategy. An effective combination of modeling and data can result in significant improvement in capital accumulation. In this paper a Bayes dynamic pricing model

462

LC MacLean, ME Foster and WT Ziemba L. C. MacLean eta/. I Journal of" Banking & Finance 31 (2007) 3503-3523

3520

is the basis for a truncation estimator of the rate of return on assets. The key to the truncation estimator is the covariance between asset prices. The covariance is incorporated into the estimate of the mean. In that way, the common information in price movements contributes to improved estimation of individual estimates. The truncation estimator is compared to well known estimators - mean and HayesStein. If asset prices follow the Bayes model, then analytic results establish the superiority of the truncation estimator. From simulation results, the truncation estimator outperforms the alternatives in general. The savings in estimation error with the proposed estimator translate into better decisions and accumulated wealth. A back test on data from the New York Stock Exchange emphasizes the gains. The analytic formulation provides an assessment of risk aversion in the face of uncertain returns. In particular, the risk aversion index can be used to offset the loss from estimation error with the truncation estimator. Appendix A Proof of Proposition 1. With y = エセiZM[LkイャG@ V

ァセ@

d

= Bᆪゥセャ@

m

.

?

(·;'i.

> 0, and

MLセ[ZG@

d

';')-,and C

Lセ@

> 0. So

'

==

セM」@

Bᆪ[セA@ '/l

HセLエI@ x HLセN[I@

セク・、L@

InK

In HセL[MI@

-1

let Ym =my- Bᆪ[セ



. .

エェᄋ@

Consider

.

. .

Substltutmg for Ym and d1fferent1atmg,

.

< 0. Since the minimum

V = 0 occurs when

Yi = y, Cm is as stated. Proof of Proposition 2. Consider R(8, BB) = 「 R ᆪ[セB

セG@

< 0, so that R is strictly concave in (·y 1 , ••• , ]!

Q ゥォ 111 ).

Q@ + b2 )/(y1 + b2 ) 2 . tィ・ョセ^@ If R(8, BB)

';

=

0, and

.f(·y 1 , ••• , Y

111 ),

then

1"

m-1

m m ) .f ( m - 1 y' · · · ' m - 1 1'' 0 - .f (y' · · · ' y) < \l.f (';'' · · · ' ';')

=0, m-1

and the second statement is established. With "£;'_ 1y1 =my, consider the level sets

Ar={(l'J•···•

IL.m (y1 -y) 2 =r,t(j=mJ'}· ェセャ@

ェセャ@

Consider r, > r2, and (yJ, ... , ケセLI@ E aイセG@ (YT, ... , ケセLI@ EAr,. Since .f( y1 , ... , ]!11,) is strictly and monotonicity in variance follows. concave, then .f(Yl, ... , y)J < .f(YI, ... , ケセLI@ Since complexity increases as variance decreases, it follows that Bayes risk increases as complexity increases. Proof of Proposition 3. Consider the additional risk for the truncation estimator, 1 1 R+(8 8) = - 2 (L[1)trB1rB 1 +-trE(B- B)Sn(B- B( nK n

Chapter 26. Covariance Complexity and Rates of Return on Assets

463

L. C. MacLean et a/. I Journal of Banking & Finance 31 (2007) 3503-3523

3521

where B = LII"- 1 and B = LIS*- 1 • Also Sn = LGL' for G = diag(g 1 , ••• ,gK). With m and Ll=b 2 J known, then B=LII"- 1 =b 2 A 0, then m Q セ Q N@ 1 is at its maximum; and ir is not at its maximum, then mIセ Q L@ = 0. Stochastic benchmark goals can also be set by the user and are similarly penalized for underachievement. The benchmark target ii, is scenario dependenL ll is based on stochastic asset returns and fixed asset weights a, defining the benchmark portfolio

M.r:

M)r

N

t

Bt= キ[Zセl。Iサイ@ j=l i=l

The corresponding shortfall constraints are N

r;cw,,- P,, + s,J + Ni,B? B,.

1=

1•...•

r.

i=l

where

M,"

is the benchmmk target shortfall. These short-

falls are also penalized with a piecewise-linear convex risk measure.

If total wealth exceeds the target. a fraction y = I 0% of the exceeding an1ount is allocated to a reserve account and invested in the same way as other available funds. However. the wealth targets at future stages are adjusted. Additional nonnegative decision variables f5, arc introduced and the wealth target constraints become

i=l

t-1

W, + "L;yD,_j•

1 =I ..... T- I. where

L

T

v, =I

and

"L;w, =

T.

t=l

kE(\V.B)

Such concave objective functions with convex risk measures date to Kusy and Ziemba ( 1986). were used in the Russell- Yasuda model (Carino and Ziemba 1998), and are justified in an axiomatic sense in Rockafellar and Ziemba (2000). Nontechnical decision makers find the increasing penalty for target violations a good approach and easy to understand. In the implementation of the model presented in §4. the penalty function c,(M') corresponds to a quadratic utility function. Kallberg and Ziemba ( 1983) show. l(ll" normally distributed asset returns. that varying the average Arrow-Pratt absolute risk-aversion index RA traces out the whole spectrum of risk attitudes of all concave utility functions. The most aggressive behavior is log utility. which has R,1 = !/wealth. which is essentially zero. Typical 60-40 stock-bond pension funds have RA = 4. The Kallberg-Ziemba (1983) results indicate that for computational purposes, the quadratic utility function u( w) = w R 1f2w 2 will suffice and is easier to use in the optimization. The error in this approxin1ation is close to zero and well below the accuracy of the data. The parameter A in the objective corresponds to RA/2. which in the quadratic utility function is the weight

assigned to risk measured in tenns of variance. The objective function of the InnoALM model penalizes only wealth and benchmark target shortfalls. If the target growth is roughly equal to the average return of the portfolio, shortfalls measure only negative deviations from the mean,

whereas variance is based on positive and negative devi-

;V

L;CW,,- P,, + s,,)- i5, + M,"" =

Expectation is over T period scenarios 51 . The discount factors d, are related to the interest rater by d, = (l+r)-'. Usually. r is taken to be the three- or six-month treasury bill rate. The v, are weights for the wealth and benchmark shortfalls. and the w, are weights for the weighted sum of shortfalls at each stage normalized via

D1 =

0.

j=l

Because pension payments are based on wealth levels.

increasing these levels increases pension pay1nents. The reserves provide security for the pension plan's increase of pension payments at each future stage. The pension plan's objective function is to maximize the expected discounted value of terminal wealth in period T

ations. This implies that shortfalls account for only about half of the variance. Therefore. to obtain results in agreement with a quadratic utility function, we usc A = R A, rather than R 4 /2, in the objective function. To obtain a solution to the allocation problem ror general levels or total initial wealth w0• we use the rescaled parameter A= RAfw0 in the objective function. Using a quadratic function. the penalty function cr(Mk) is

c,(M')= "L;M7,(bi_ 1 +b). i=l

Chapter 28. The Innovest Austrian Pension Fund

495

Geyer and Ziemba: The Innovest Austrian Pension Fund FinanciaL PLanning Model JnnoALM oー・イ。エゥッョセ@

Research 56(4). pp. 797-810.

Figure 1.

セIRPX@

801

TNFORrviS

Scenario tree with a 2-2-3 node structure (12 scenarios).

Stage 1

Stage 2

Stage 3

(1=0)

(1=1)

(1=2)

Stage 4 (I

=3)

Scenario I Scenario 2 Scenario 3

Scenario 4

Scenario k

Scenario 12 Numhcr of nodes at t

n 3 = 12

Uncertainty is modelled using multiperiod discrete probability scenmios using statistical properties of the assets' returns. A scenario tree is defined by the number of stages 'md the number of arcs leaving a parlicular node. Figure 1 shows a tree with a 2-2-3 node structure for a three-period problem with four stages and introduces some dellnitions and termino1ogy. The tree a1ways starts with a sing1e node that corresponds to the present state (t = 0). Decisions are made al each node of the lree and depend on Jhe cunenl state, which reflects previous decisions and uncertain future paths. A single scenario s, is a trajectory that corresponds to a unique path leading from the single node at stage I (t = 0) lo a single node al t. Two scenarios s; and s;' are and differ in subseidentical until t - I (i.e., s;_, = ウ[セLI@ quent periods I, .... T. The scenario assigns specific values to a11 uncertain parameters along the trajectory, i.e., asset returns and benchmark targets for all periods. Given all T period scenmios 51 'md !heir respeclive probabililies. one has a complete description of the unce11ainty of the model. Allocations me based on oplimiLing ihe sJochasJic linear program with IBM's optimization solutions library using the stochastic extension library (OSLE version 3). IBM has ceased all sales of this product in 2004. While existing installations of OSLE can still be used, new implementions require alternative software such as the open source project COIN-OR (see http://www.coin-or.org). The library uses the Stochastic Mathen1atical Programming System (SMPS) input format for multistage stochastic programs (see King et al. 2005). The core-file contains information about the decisions variables, constraints. right-hand sides, and bounds. It contains all fixed coefficients and dummy entries for random clements. The stoch-file reflects the node structure of the scenario tree and contains all random elements-i.e., asset and benchmark returns-and probabilities. Nonanticipatory constraints arc imposed to guarantee that a decision made at a specific node is identical for all scenarios leaving that node, so the future cannot be

anticipated. This is implemented by specifying an appropriate sccnmio structure in the stoch input file. The time-file assigns decision variables and constraints to stages. The required stalements in the inpul files are automatically generated by the InnoALM system (see §4).

3. Scenario Generation and Statistical Inputs The uncc11ainry of the random return and other parameters in InnoALM is modeled using discrete probability scenarios. These scenarios are approximations of the true underlying probability distributions. The accuracy of the set of scenarios chosen and the probabilities of these scenarios in relation to reality contribute greatly to model success. However, the scenario approach generally leads to superior investment performance even if there arc errors in the estimations of both the actual scenario values and their probabilities. What the modeling e!Iorl allempls lo do is to cover well the nmge of possible future evolulion of the economic cnviromncnt. Decisions take into account all these possible outcomes, weighted by their likelihood. This generally leads to superior performance of multi period stochastic programming models compared with other approaches, such as mean-variance analysis, fixed mix, stochastic control, stochastic programming with decision rules, etc. Studies showing this superiority, both in and out of sample. include Kusy and Ziemba (1986), Carino and Turner (1998), Carino et al. (1994. 1998). Carino and Ziemba (1998). and Fleten et a!. (2002). Procedures for eslimating ihe joim distribution of fulure bond and stock returns have been discussed by Chen ct al. (1986), Keim and Stambaugh (1986), Person and Harvey (1993), Karolyi and Stultz (1996). and Bossaerts and Hill ion (1999). Procedures for estimating discrete scenmios rrom joint multivariate bond and stock forecasting models have been discussed by Mulvey (1996), Jamshidian and Zhu (1997), Carino et al. (1994, 1998), Carino and Ziemba (1998). Zenios (1999), Hpyland and Wallace (2001). Pflug (2001 ), and Roemisch and Heitsch (2003). The sccnm·ios in InnoALM m·c defined in terms of the dislribulion of asset returns 'md their Jirsl- and second-order moments. The latter can be prespecified by the user or estimated from the built-in database of historical returns. James-Stein estimates, which have frequently been suggested as the preferred approach. can be used lo eslimatc mean returns; sec. e.g .• Jorion ( 1985), Hensel and Turner (1998), and Grauer and Hakansson (1998). Sec also MacLean et a!. (2007) for an alternative approach using truncated estimators. For each asset the user can choose from the normal, the t-, or the historical (empirical) distribution. Empirical asset returns over short horizons (up to one month) lypically me nol normally distribuied bui have fal tails and arc skewed. Jackwcrth and Rubinstein (1997) show how much fatter the implied probability left tails of

496

A Geyer and WT Ziemba Geyer and Ziemba: The Iruwvest A.ustriun Pension Fund Financial

802

Operation8

Table 1.

r・セ。ョZィ@

Model InnoA.LM

56(4). pp

I[)

2008 INFORMS

Statistical properties of asset returns. Stocks Eur

Stocks Eur

1/70-9/00

1/86-9/00

Stocks US 1/70-9/00

10.6 16.1 -0.90 7.05 302.6

13.3 17.4 -1.43 8.43 277.3

10.7 19.0 -0.72 5.79 151.9

II. I

13.3 16.2 -0.10 2.28 3.9

11.0 20.1 -0.23 2.56 6.2

Stocks US 1/86-9/00

1/86-9/00

Bonds US 1/86-9/00

14.8 20.2 -1.04 7.09 155.6

6.5 3.7 -0.50 3.25 7.7

7.2 11.3 0.52 3.30 8.5

15.2 18.4 -0.28 2.45 4.2

6.5 4.8 -0.20 2.25 5.0

6.9 12.1 -0.42 2.26 8.7

Bonds Eur

Monthly returns

Mean(% p.a.) Std. dev. (% p.a.) Skewness Kurtosis Jarque-Hera test Annual returns

Mean(%) Std. dev. (%) Skewness Kuttosis Jarquc-Bcra test

17.2 -0.53 3.23 17.4

the S&P500 have become since the 1987 worldwide stock market crash because of investor fear of large declines. !-distributions model lat tails well (see Glasserman et al. 2000). The degrees-of-freedom parameter has to be set to a small value (e.g., five). However, both the normal and the !-distribution me symmetric distributions and might therefore underestimate the downside risk of an asset or portfolio. Skewed !-distributions arc an alternative to account for skewness and fat tails hut were not considered in this model. Pension fund planning models typically use rebalancing intervals that are much longer than one month. InnoALM can accommodate, e.g., weekly, monthly, annual, or longer rebalancing intervals. In the example in §4, annual and biannual periods me considered. Annual returns have distinctly different distributional properties than monthly returns (see Table l ). Given the difficulty associated with choosing an appropriate parametric distribution, we also use a nonparametric approach to generate random samples reflecting the shape of the historical return distribution. To simulate the historical distribution for a single asset, we compute stcmdaniit:ed annual returns Yr We use (overlapping) annual returns from monthly data rather than monthly returns because the planning intervals in the example presented in §4 arc in years, and thus the disttibution of annual returns is more appropriate than the distribution of monthly returns. A single element of the simulated historical return distribution is computed as follows. First, a random number u is drawn from a uniform distribution. This random number is treated as a probability 'md the corresponding percentile z is computed !rom the standardized returns. The percentile is a random draw !rom the ltistorical, standardized distribution with the property Pry, < zl = u. Multiplying z by a prcspccificd standard deviation and adding a prcspecificd mean yields the random return used at a particular node in the scenario tree (see below). Sampling from standardized rather than observed returns allows us to simulate historical distributions with means and standard deviations that might diller from the historically observed sample statistics.

This approach yields a random sample that matches the shape or the historical (fat tailed and/or skewed) distribution. The size or the random sample that can be generated by this approach is not limited by the number of available historical observations because any desired number of percentiles could be computed from historical returns. The approach cannot produce values that are more extreme than historically observed returns, however. State-dependent correlation matrices of InnoALM are a new feature and have not yet been used in pension planning or asset allocation models. lnnoALM uses three different correlation matrices and corresponding sets of standard deviations. The choice of a specific conelation matrix depends on the level of stock return volatility. We distinguish "extreme" (or "crash") periods, "highly volatile" periods, and "nonnal" periods. Each of the three periods or regimes is assigned a probability of occurrence pl (j = I, 2. 3). Harvey ( 1991), Karolyi and Stulz ( 1996). Solnik et al. (1996), and Das and Uppal (2004) study changing correlation structures over time. To estimate correlations 'md st,mdmd deviations for the three regimes, we use the regression approach suggested by Solnik et a!. (1996). Using monthly time series, we compute moving average (window length 36 months) estimates of correlations among all assets and standard deviations of U.S. equity returns. Correlations are regressed on U.S. stock return volatilities. The estimated regression equations are used to predict correlations for the three regimes (more details on this are presented in §4.1 ). Correlated random returns are simulated using the following procedure. For each asset i, we generate n 1 st 10'k '

sho11fal1

226.6

86.9

2.7 3.7 2.8 2.2 2.0 1.9 2.9

inspecting the model's wealth and risk implications across time and across scenarios.

4.3. Model Tests Because stare-dependent corrclalions have a significant impact on allocation decisions, we further investigate their nature and their implica1ions to r.cst r.hc model. While the

focus of the previous section was t.o compare V'-lrious stochastic assumptions and to highlight the model's benefits to lnnovcst, the purpose of the present section is to test the advantages of using mixing cmTClations in an out-ofsample context and a controlled experiment. Positive effects on the pension fund performance and its risk profile induced by the stochastic, multiperiod planning approach will be realized only if the portfolio is dynamically rebalanced, as implied by the optimal scenario tree. We first illustrate the decision rule implied by the model. We form qui ntiles of wealth and compute the average optimal weights a.."isigned to each quintile. Figure 3 shows the distribution of we ights for each of the five average levels of wealth for case TM at stage 2, which Figure 3.

Optimal TM weights conditional on quintiles of portfolio wealth at stage 2.

distribution over time for

WPセ][M

ャ@ -

600 500

\Vcalth target .5%-quamile Medi:m 95%-l]llan!ile

-5 セ@

400 セ@

セ@

セP P@ 200 97. 1

103.4

107.9

ll3.9

125.8

Average wealth in quintile at sta ge 2

Periods

• Bonds US

[] Bonds Europe

• Equities US

D Equities Europe

Chapter 28. The Innovest Austrian Pension Fund

501

Geyer and Ziemba: The lnnovest Austrian Pension Fund Financial Planning ModellnnoALM Operations Research 56(4). pp. 797-!Wl.

セIRPAs@

807

INFORMS

depends on the wealth level in a specific way. For example, if average attained wealth is 103.4, which is slightly below the target, a very cautious strategy is chosen. Bonds have the highest weight in this case (almost 50%). In this situation, the model implies that the risk of even stronger underachievement of the target is to be minimized. If attained wealth is far below the target (97. 1), the model implies more than 70% equities and a high share (I 0. 9%) of relatively risky U.S. bonds. With such strong underachievement, there is no room for a cautious strategy to attain the target level again. If average attained wealth is close to the target 107.5 putting more money into U.S. assets (72.4%), which are more risky than the corresponding European assets, is acceptable because risk does not play a big role in that situation. For wealth levels above the target, most of the portfolio is switched to European assets. which are safer than U.S. assets, thereby preserving the high levels of attained wealth. These results follow the objective of the fund, although individual pension beneficiaries might have different preferences. The decision rules implied by the optimal solution can be used to perform a test of the model using the following rebalancing strategy. Consider the 10-year period from January 1992 to January 2002. In the first month of this period, we assume that wealth is allocated according to the optimal solution for stage 1 (see Table 4). In each of the subsequent months, the portfolio is rebalanced as follows: we identify the current volatility regime (normal, highly volatile, or extreme) by comparing the U.S. stock return volatility estimated from the past 36 months of U.S. equity returns (sec §3, p. 13) to the 35th, 80th, and 95th percentiles of volatility as defined in §4. 1. The regime might have been identified incorrectly. Then, we search the scenario tree across all stages to find a node that corresponds to the current volatility regime (using the tags assigned to simulated return) and has the same or a similar level of wealth. Given the high short-run persistence of return covariance, it is plausible to assume that the current volatility and correlation regime prevails in the subsequent month ,md we use the optimal weights !rom that node to detennine the rebalancing decision. For the no-mixing cases NA, TA, and HA, the information ahout the current volatility regime cannot be used to identify optimal weights. In those cases, we usc the weights from a node with a level of wealth as close as possible to the current level of wealth. Table 6 presents summary statistics for the complete sample time period and the out-of-sample time period October 2000 to January 2002. The mixing correlation solutions assuming norrnal and /-distributions (cases NM and TM) provide a higher average return with lower standard deviation than the corresponding nonmixing cases (NA and TA). The advantage might be substantial, as indicated by the 14.9% average return of TM compared to 10.0% for TA. The /-statistic for this di!Ierence is 1.7 and is significant at the 5% level (one-sided test). Using the historical distribution and mixing correlations (HM) yields a lower average

return than no-mixing (HA). This result can be explained by a strong asymmetry in the pattern of conditional weights of HA (as shown in Figure 3 for TM) in favor of stocks, which contributes to the superior perl'mmance of HA compared to HM. In the constrained case TMC, the average return for the complete sample is in the same range as for the unconstrained cases. This is mainly due to relatively high weights assigned to U.S. bonds, which had high returns during the test period. The standard deviation of returns is much lower because the constraints imply a lower degree of rebalancing. To emphasize the difference between the cases TM and TA, Figure 4 compares the cumulated monthly returns obtained from the rebalancing strategy for the two cases as well as the results of two buy-and-hold strategies. One of the buy-and-hold strategies assumes that the portfolio weights on January 1992 arc fixed at the optimal TM weights throughout the test period, and the other uses the weights from single-period mean-variance analysis (see Table 4). Rebalancing on the hasis of the optimal TM scenario tree provides a substantial gain when compared to the buy-and-hold strategies or the performance using TA results, where rebalancing does not account for different correlation and volatility regimes. The results in Table 6 suggest that the TMC is the most attractive stralegy afler adjusting for risk (e.g., using a Sharpe ratio). However, such in- and out-of-sample comparisons depend on the observed asset returns and the test period. To isolate the potential benefits from considering state-dependent correlations, we performed the following controlled simulation experiment. Consider 10.000 10-year periods where simulated annual returns of the four assets arc assumed to have the statistical properties summarized in Table 2. The means and standard deviations of simulated returns across all 100,000 years are equal to those given in the last two rows of Table 2. One of the I 0 years is assumed to be an "extreme" year, two years correspond to "highly volalile" markels, and seven years are "normal" years using the corresponding correlation matrix and standard deviations from Table 2. Returns have normal, historical, or

Table 6.

Results of asset allocation strategies using the decision rule implied by the optimal scenario tree. Complete sample

0 I/92-0 I/02

NA NM HA HM TA

TM TMC

Out-of-sample

I0/00-0 I/02

Mean

Std. dcv.

Mean

Std. dcv.

11.6 13.1 12.6 11.8 10.0 14.9 12.4

16. I I5.5 16.5 16.5 16.0 15.9 8.5

-t7.1 -9.6 -15.7 -15.8 -14.6 -10.8 0.6

18.6 16.9 21.1 19.:l 18.9 17.6

9.9

502

A Geyer and WT Ziemba Geyer and Ziemba: The Innoresl Austrian Pension Fund Financial Planning ModellnnoALM

808

Operation8 Re-;ean:h 56{4). PP- 797- 810, © 2008 INFORMS

Figure 4.

Cumulative monthly returns for different strategies.

0.6 0.4

0.2

!-distributions. We compare rebalancing strategies that usc the implied decision rules of the optimal scenario tree, as explained in the in- a nd out-of-sample tests above. O nly the rebalancing strategies of the cases NM, HM, a nd TM

are appropriate in this simulated environment because the strategies are based on the optimal solution derived from a stochastic tree with mixing correlations. The strat.e gies for cases NA, HA, and TA arc based on the optimal solution assuming an average correlation matrix. For simplicity, we assume that the current volatility regime is known in the mixing strategies. Using this experimenl, we assess the value of scenariodependent correlation matrices by comparing NM to NA, HM to HA, and TM to TA (see Table 7). Case TMC is included to test the impact of constraints on asset weights. The average annual return over l 0,000 repetitions using TM weights is 10.0%, compared with 9.5 1% using the Table 7.

Average ann ual returns, expected terminal wealth, and the probability for wealth-target shortfalls (percentage) for di!Icrent rebalancing strategies. return(%)

Expected termi nal wealth

Prohahi lity of target shortfall

9.87 9.98 9.28 9.53 9.51 10.0 8.06

307.1 3 13.3 283.2 294.3 292.7 314.5 23 1.9

31.0 30.7 33.7 32.8 32.1 30.1 38.0

Average annual

NA NM HA

HM TA TM

TMC

TA weights. The mean di!Icrcnce has a !-statistic of 11.2, which indicates a highly significant advantage of using state-dependent correlations. The rebalancing strategy in case TM applies appropriate decision rules, while the strategy in case TA cannot respond adequately to diJierent regimes. The expected wealth at the end of year I0 is 3 14 .5 for TM, compared to only 292.7 for TA. The shortfall probability in case TM is also smaller. Similar differences can be observed by comparing the pairs NA-NM and HAHM, although the advantage of the mixing cases is not uniform across d istributions. For the constrained case TMC, we obtain an average return of only 8.06(k·, which indicates that the constraints imply insufficient rebalancing capacity. Thus, the relatively good performance of the TMC rebalancing strategy in the sample period 1992- 2002 might have been positively biased by the l'avorable conditions during that time. Rebalancing strategies always outperJ'orm buy-and-hold strategies (e.g., 10% compared to 9.2% in case TM). Summarizing, the main implications derived hom the model tests for practical applicatjons arc significant benefits of using scenario-dependent correlation matrices. The controlled experiments indicate that the advantages measured in terms of wealth and shm1fa11 probabilities are not sample dependent.

5. Conclusions and Final Remarks The model TnnoALM provides an easy-to-use tool to help Austrian pension funds' investment allocation committees

evaluate the effect of various policy choices in light of changing economic conditions and various goals, constraints, and liability commitments. The model includes

Chapter 28. The Innovest Austrian Pension Fund

503

Geyer and Ziemba: The lnnovest Austrian Pension Fund Financial Planning ModellnnoALAf

809

Operations Research 56(4). pp. 797-lHO, CD200.S INFORMS

features that reflect real investment practices. These include multiple scenarios, nonnormal distributions, and different volatility and correlation regimes. The model provides a systematic way to estimate in advance the likely results of particular policy changes and asset return realizations. This provides more confidence and justification to policy changes that might be controversial, such as a higher weight in equity and less in bonds than has traditionally been the case in Austria. The model is an ad vance on previous models and includes new features such as state-dependent correlation nultrices. Crucial to the success of the results are the scenario inputs and especia11y the mean return assumptions. The model has a num her of ways to estimate such scenarios. Given good inputs, the policy recommendations can improve current investment practice and provide greater confidence to the asset allocation process. The following quote by Konrad Kontriner (member of the board) and Wolfgang Herold (senior risk strategist) of Innovest emphasizes the practical importance of InnoALM: 'The llmoALM model has been in usc by Innovc::.t, an Austrian Siemens subsidiary, since its first draft versions in 2000. Meanwhile it has become the only comistcntly implemented and fully integrated proptietary tool for assessing pension allocation issues within Siemens AG worldwide. Apart from this, consulting projects for various European corporations and pensions funds outside of Siemens have been performed on the basis of the concepts of InnoALM. The key elements that make lnnoALM superior to other consulting models arc the flexibility to adopt individual constraints and target J'unctions in combination with the broad and deep array of results, which allows to investigate individual, path-dependent behavior of assets and liabilities as well as scenmio based and Monte-Carlo like risk assessment

of both sides. In light of recent changes in Austrian pension regulation the latter even gained additional importance, as the rather rigid a::.sct-based limits were relaxed for institutions that could prove sufficient risk management expettise for both assets and liabilities or the plan. Thus, the implementation of a scenario-based asset allocation model will lead to more flexible allocation restraints that will allow for more risk tolerance and will ultimately result in better long-term investment performance. Furthermore. some results of the model have been med by the Austrian regulatory authorities to assess the potential risk stemming from les::.-constraint pension plans."

Appendix. Pseudo-Code for Scenario Generation SET N % number of assets SET T % number of periods SET cp, hp, np% probabilities for crash, high and normal regimes SET m% T x Nmatrix of mean gross returns adjusted for period length

SET cs, hs, ns% T x N matrices of volatilities adjusted for period length SET cc, he, nc% Nx N correlations for crash, high and normal regimes FOR period= I to T nz = n_nodes [period] % total number of nodes in the current period COMPUTE nz x N matrix Z % anti the tic random numbers with mean 0 and std. dev 1; Z can have normal, t- or historical distribution DEFINE ic, ih, in% arrays with random pointers to rows of Z; ic, ih, and in have dimension CEIL(cp*nz), CEIL(hp*nz), and nz-CEIL(cp*nz)-CEIL(hp*nz); the joint index set is nonoverlapping; in each period a different set of random pointers is used COMPUTE mean 0 returns with required correlation and volatility in each regime: Y[ ic, . ] = (Z [ic, . ] *CHOL( cc)). *cs [period, .] Y[ih, .] = (Z [ih, .] *CHOL(hc)). *hs [period, .] Y[in, .] = (Z [in, .] *CHOL(nc)). *ns [period, .] COMPUTE R__period = m[period, .] returns for current period END FOR

+

Y % gross

Acknowledgments This research was supported by Jnnovest and the Natural Sciences and Engineering Research Council of Canada. Thanks to Melania Paunescu for computer help; Wolfgang Herold. Konrad Kontriner. and Johann Maurer for encouragement and support; Alan King and Alex Weissensteiner; the former area editor, Rick Rosenthal; the current area editor, Steve Graves; an anonymous associate editor, and the referees for helpful comments on earlier versions of this paper. This paper was presented at the Hermes Centre of Excellence in Computational Finance and Economics, the Conference on Asset and Liability Management: From Institutions to Households, Nicosia, Cyprus, May 2001; the International Conference on Financial Engineering, E-Commerce and Supply Chain, Athens, Greece. May 2001; the Chicago Fed/DePaul University Finance Seminm, June 200 I; the 9th International Conference on Stochastic Programming, Berlin, August 200 I; the IFIP/IIASA/GAMM Workshop on Dynamic Stochastic Optimization, IIASA, Laxenburg, Austria, March 2002; Southampton University and Imperial College, September 2002; the Euro Working Group on Financial Modeling, Cyprus and University of Bergamo, November 2002; the SAID Business School, Oxford University, Mmch 2003; the University of Zurich and the Mathematical Institute, Humboldt University, June 2003; the Operations Research Department of the ETH Zurich. July 2003; the Work-

504

A Geyer and WT Ziemba Geyer and Ziemba: The lnnovest I1ustrian Pension Fund Financial P!anninx Model Inno,1LM

810 shop on Asset Liability Management in Insurance in Vienna, September 2004; the University or Washington Statistics and Industrial Engineering Departments, June and November 2005; the Control Engineering Department, ETH, Zurich, October 2005; the 19th International Conterence on Mathematical Progranuning, Rio de Janeiro, August 2006; the International Workshop on Forecasting and Risk Control, Chinese Academy of Sciences, Beijing, December 2006; the Mathematical Institute, Oxrord University and the Department of Mathematics, University of Edinburgh, March 2007; Universities of Bergamo and Venice, May 2008; and Universities or Florence, Zurich, and Reading, June 2008.

References Bos. E. 1994. World population projections 1994-95. The World Bank. Washington, D.C.

Bossacrts. P., P. Hillion. 1999. Implementing statistical criteria to select return forecasting models: What do we learn? Rev. Financial S11td. 12 405-428. Campbell, J. Y., L. M. Viceira. 2002. Strategic Asset Allocation. Oxford University Press. Oxford, UK Carillo, D. R .. A. L. Turner. 1998. Multi period asset allocation with derivative assets . .T. M. Mulvey, W. T. Ziemba, eds. VVorldwide A.'i·se! and Liability Modelin;:;. Cambridge University Press, Cambridge, UK, 1l52-204. Carifio. D. R.. W. T. Ziemba. 1998. Fonnulation of jhe Russell- Yasuda Kasai financial planning model. Oper. Res. 46 433-449. Carino, D. R., D. H. Myers, W. T. Ziemba. 1998. Concepts. technical issues. and u:'.e:'. of jhe Ru:'.sell- Yasuda Kasai financial planning model. Oper. Res. 46 450--462. Carifio, D. R., T. Kent, D. H. Myers, C. Stacey, M. Sylvanus, A. L. Tumer. K. Watanabe, W. T. Ziemba. 1994. The Russell- Yasuda Kasal model: An asset/liability model for a Japanese insurance company using multistage stochastic programming. lntetfiuxs 24(1) 29-49. Chen. N .. R. Roll. S. A. Ross. 19l56. Economic forces ami the stock market. J. Hus. 59 383-403. Chopra, V., W. T. Ziemba. 1993. The effect of errors in mean and covmiance estimates on optimal pottfolio choice. 1. Portfolio Manaxement 19 6--11. Das. S. R., R. Uppal. 2004. Systemic risk and international portfolio choice . .!. Finance 59 2R09-2834. Dim:'.on, E.. P. Mar:'.h, M. Staunjon. 2002. Triumph !Of Years HセヲB@ Global Investment Returns. Prinedon Princeton. N.T. Dimson. E., P. Marsh, M. Staunton. 2006. Global Investment Returns Yearbook. AHN-Ambro. London. Douglass, J., 0. Wu, W. T. Ziemba. 2004. Stock ownership decisions in DC pension plans. 1. PorrjiJ!io Manaxemenr (Summer) 92-100. European Commission. 1997. Supplementary pensions in the single market. A Green Paper, COM 2R3, Bmssels, Belgium. l·•erson, W. E., C. R. Harvey. 1993. The risk and predictability of intemajional equity returns. Rev. h'nancial Stud. 6 527-566. Fleten, S.-S., K. Hoyland, S. W. Wallace. 2002. The performance of stochastic dynamic and flxcd mix portfolio model:'.. Lur. J. Oper. Res. 140 37-49. Glassennan. P., P. Heidelberg, P. Shahabuddin. 2000. Variance reduction techniques for value-at-ri:'.k with heavy-tailed risk factor:'.. J. A. Joines. R. R. Barton, K. Kang, P. A. Fishwick. cds. Proc. 2000 Winter Simulation Coi!f:. Orlando. FL, 604-609. Gondzio, J., R. Kouwenherg. 2001. High performance computing for asset-liability management. Opa Res. 49 l-:79-891. Grauer, R. R., N.H. Hakan:'.son. 1998. Stein and CAPM estimators of the means in asset allocation. ln!errwl. Rev. Financial Anal. 4 l:i-66.

Operation8 Research 56(4), pp. 797-810. ©2008 INFORMS

Harvey, C. R. 1991. The world price of covariance risk. J. Finance 46 111-157. Hensel, C. R.. A. L. Turner. 1998. Making superior asset allocation decisions: A practitioner's guide . .T. M. Mulvey, W. T. Ziemba, eds. Worldwide aセウ・イ@ and Liability Modelin;:;. Cambridge University Prc:'.s. Cambridge. UK, 62-83. Hpyland, K. S. W. Wallace. 2001. Generating scenario trees for multistage decision problems. Management Sci. 47 295-307. Jackwetth. J. C.. M. Rubinstein. 1997. Recovering probability distributions from option prices. J. Finance 51 1611-1631. Jam:'.hidian. F.. Y. Zhu. 1997. Scenario simulation: Theory and methodology. Finance and Stochastics 1 43--67. Jorion, P. 1985. lnjernational portfolio diversification with estimation risk. J. Bus. 58 259-27l5. Kallberg . .T. G., W. T. Ziemba. 1983. Comparison of alternative utility functions in portfolio selection problems. Manaxement Sci. 29 12:i7-1276. Karolyi. A., R. Stultz. 1996. Wh.Y do markets move together? An investigation of US-Japan stock return comovements . .!. Finance 51 951-986. Keim, D. B., R. F. Stambaugh. 1986. Predicting returns in the stock and bond market:'.. J. hnancial !:'conom. 17 357-390. King, A. J., S. E. Wright. G. R. Parija, R. Entriken. 2005. The IBM stochastic programming sy:'.tcm. S. W. Wallace. W. T. Ziemba, eds. Stochastic Series in Optinllzation, Applications HセヲB@ MPS SlAM. Philadelphia. Kusy. M. I.. W. T. Ziemba. 1986. A bmlk asset T as

T

P(A, B) A c;;; I, B c;;; 0.

varies from negative to zero to positive, implying the

effect of inforrnation offset on the SBE index? vVhen SI3K for

T

T =

> 0, eT is termed the T-lagged SI3E, and for

The population parameter

(-jT

(4)

T

0, 8° is termed the synchronous

< 0, eT is the ITI-leading SI3K

of SBE can only be obtained as a sample statistic via

the maxirni,-mtion of a sample-estimated value for fT (A, B). For this purpose, we use a sequence of annual time periods t

= 1, ... , T for the 10

N firms and in each time period t,

Chapter 32. DEA-Based Firm Strengths and Market Efficiency

the correlation rT(A,B) is estimated and it is denoted by セOtHaLbIN@

621

That is, the sample-

estimated correlation for period t is

(5) Then, the point estimate for

rT (A, B)

is the sample-average statistic

(6) Accordingly, the sample estimate for SBE in (4), denoted by (JT, is evaluated by solving the optimization model: W:=max{"'((A,B): aセiLboスN@

(7)

\Ve will utilize the statistical properties of (}T toward testing; the hypothesis that the SBE index 8T is not sufficiently large, for a given offsetting time period

T.

This is covered in

Section 3.2. In the next section, an algorithmic procedure for solving the model in (7) is considered.

3.1

Solution of SEE model

The optimization model in (7) is nonlinear, integer, and the objective function does not have a dosed-form expression. Therefore, the solution of this model is computationally tedious. In what follows, we employ a hybrid method consisting of a direct search technique that strives for local optimal solutions and a simulated annealing (SA) algorithm that attempts to improve local solutions toward global optimality. Our use of SA is also motivated by the application of DEA in predicting bank stock performance by Avkiran and Morita (2010). The direct search method employed here is from Edirisinghe and Zhang (2007), which involves a two-step heuri;;tic. pair of subsets

(A 0 ,

B0)

セwィ・ョ@

adapted to ;;olving the problem in (7), first, an initial

of the accounting ratios in Table 1 is randomly generated. In the

second step, and starting from this initial pair, the "((A, B) is improved by swapping each parameter in and out of the subsets A and B, until no further improvement is possible, sec the above reference for details. This process is generally seen to produce local maximum solutions as it is based on a local search (LS) procedure. Once this local search has terminated, at some subset pair (A, B), the solution method switches to the meta-heuristic termed Simulated Annealing (SA). SA is a technique to find a good solution to difficult combinatorial optimi11ation problerns by trying random variations of a current solution. A worse variation is accepted as the new solution with a probability that decreases as the computation proceeds. SA's major 11

622

C Edirisinghe, X Zhang and S-C Shyi

advantage over other methods is an ability to avoid becoming trapped at local optima. Although SA was originally developed in the 50's, it was Pincus (1970) who first noted the connection between SA algorithm and mathernatical optimi;:ation. Later, Kirkpatrick et al. (1983) proposed that SA forms the basis of an optimization technique for combinatorial (and other) problerns. Our implementation of SA proceeds by generating an input/output pair of subsets

(A', B'), as a neighbor to (A, B), by a 1- or 2-swap procedure: namely, 1 or 2 financial

B) are randomly chosen to be replaced by a random selec2 indices from I \ A (and from 0 \ B), respectively. This process generates pair of subsets (A, B). Then, the scalar ュ・エイゥ」セZ]@ Y(A,B)- y(A,i}) is

ratio indices from tion of 1 or yet another

A (and

from

evaluated. The local search procedure is then resumed to determine a new local maximum, starting from the pair

(A, B), with probability p :=

(1- p), the local search is not resurned, where

r;

exp Hセュ。クサN@

0}): or with probability

is a given parameter. In the latter event,

yet another new neighbor is randomly generated and the parameter

r;

is decreased at a

rate r so that r; +-- rr;, and it is checked whether the local search can be resumed. If this procedure fails to resume the LS after generating a certain number of consecutive neighbors, the LS procedure is re-initialized with a new random pair (A 0 ,B 0 ). After a predetermined maximum number of such initial random pairs have been exhausted, the maximum value obtained thus far for

3.2

Y(A,B)

is declared the estimated SBE index value,

rr.

Statistical tests of SBE

To check if the sample statistic

()T

of SBE is statistically significant, consider the following

hypothesis test for a pre-specified minimum positive SBE index value of Po (satisfying

O

2 years. Lagging efficiency is virtually non-exist ent , confirming the

widely-held belief that past performance of companies has only negligible predict ive power on market returns. By using 1-year periods, the effect of post-earnings-announcement drift (PEAD) is effectively filtered out. These observations a re uniformly valid in bot h t he U .S. 22

Chapter 32. DEA-Based Firm Strengths and Market Efficiency

633

and Japan. It is quite striking that B/M ratio emerges as an universally important metric insofar as forming future expectations of business strength. Accounting for broader market risk has only limited impact on the firm strength-based market efficiency. It also develops that factoring out B/M ratio does not altogether diminish SBE under

our DEA-bascd evaluation methodology. There arc several other accounting ratios, such as ROE, P /E, EPS, leverage, and asset オエゥャセZ。ッョ@

metrics, that serve quite prominently in

forming future expectations of business strength. This paper presents several key rnetrics that a practitioner may focus on, in addition to the well-known B/M or P /E ratios, such that future expectations of business strengths computed in a relative framework allow screening public firms for profitable investments.

References C. Abada, S.A. Thoreb, and J. Laffarga. Fundamental Analysis of Stocks by Two-stage DEA. Managerial and Decision Economics, 25:231-241, 2004. N.K. Avkiran and H. Morita. Predicting Japanese bank stock performance with a composite relative efficiency metric: a new investment tool. p。」セヲゥMbウョ@ Finance .Journal, 18:254271, 2010. R. Ball and P. Brown. An empirical evaluation of accounting income numbers . .Jou.rnal of Accounting Research, 6:159-178, 1968. R.D. Banker. Estimating most productive scale size using Data Envelopment Analysis. European .Journal of Operational Research, 17:35-44, 1984. R.D. Banker, A. Charnes, and vV.vV. Cooper. Some models for estimating technical and scale inefficiencies in Data Envelopment Analysis. Jlvfa:nagement Science, :30:1078 1092, 1984. V.L. Bernard and J.K. Thomas. Post-earnings-announcement drift: delayed price response or risk premium'? .Journal of Accounting Research, 27:1-36, 1989. J.Y. Campbell and R.J. Shiller. Stock prices, earnings, and expected dividends . .Journal of Finance, 43:661 676, 1988. L.K.C. Chan, N.Jcgadccsh, and J. Lakonishok. Momentum strategies . .Jomnal of Finance, 51:1681 1713, 1996. A. Charnes, vV.vV. Cooper, and E. Rhodes. Measuring the efficiency of decision-making units. European .Journal of Operational Research, 2:429-444, 1978. D. Dodd and D. Graham. Security Analysis. McGraw-Hill, NY, 19:34.

23

634

C Edirisinghe, X Zhang and S-C Shyi

N.C.P. Edirisinghe and X. Zhang. Generalized DEA model of fundamental analysis and its application to portfolio optimization . .Journal of Banking and Finance, 31:3311-3335, 2007. N.C.P. Edirisinghe and X. Zhang. Portfolio selection under DEA-based relative financial strength indicators: case of US industries . .Journal of the Operational Research Society, 59:842 856, 2008. N.C.P. Edirisinghe and X. Zhang. Input/output selection in DEA under expert information with application to financial markets. European .Journal of Operational Research, 207: 1669-1678, 2010. H. Falk and H. Levy. Market reaction to quarterly earnings' announcements: A stochastic dominance based test of market efficiency. Management Science, :35:425 446, 1989. E.F. Farna and K.R. French. Business conditions and expected returns on stocks and bonds . .Journal of Financial Economics, 25:23 49, 1999. E.F. Farna and K.R. French. Cornrnon risk factors in the returns on stocks and bonds. Journal of Financial Economics, 33:3-56, 1993. E.F. Fama and K.R. French. Size and book-to-market factors in earnings and stock returns. Jo·umal of Finance, 50:131-155, 1995. R. Frankel and C.M.C. Lee. Accounting valuation, market expectations, and cross-sectional stock returns. Jomnal of Acco·unt·ing and Economics, 25:283-319, 1998. R.A. Haugen. The race between value and growth. Journal of Investing, 6:23-31, 1997. M. Hirschey. Extreme Return Reversal in the Stock Market- Strong support for insightful fundamental analysis. The .Journal of Portfolio l'vfanagernent, 29:78-90, 2003. D.B. Keim and \V.T. Ziernba, editors. Security Market Imperfections in Worldwide Equity Markets. Cambridge University Press, Cambridge, UK, 2000. B.J. Kim and P.S. Lee. An analysis on the long-term performance of value investment strategy in Korea. aウゥ。Mp」セヲ@ .Journal of Financial Studies, 35:1-39, 2006. S. Kirkpatrick, Jr. C.D. Gerlatt, and M.P. Vecchi. Optimization by Simulated Annealing. Science, 220:671 680, 1983 .

.J. Lakonishok, A. Shleifer, and R.\V. Vishny. Contrarian Investment, Extrapolation, and Risk. Journal of Finance, 49:1541-1578, 1994 .

.J. Lintner. The Valuation of Risk Assets and the Selection of Risky Investments in Stock Portfolios and Capital I3udgc:ts. The Rev·iew of Ecorwrn·ic Statistics, 47:13-37, 1965. R.R. Mendenhall. Arbitrage Risk and Post-Earnings-Announcement Drift. Journal of Bv.siness, 77:875-894, 2004.

24

Chapter 32. DEA-Based Firm Strengths and Market Efficiency

J. Ohlson. Earnings, book values, and dividends in security valuation. Accounting Research, 11:661-687, 1995.

635

Contempomr·y

M. Pincus. A Monte Carlo Method for the Approximate Solution of Certain Types of Constrained Optimi:oation Problems. Operations Research, 18:1225-1228, 1970. G.D. Samaras, N.F. Matsatsinis, and C. Zopounidis. A multi-criteria DSS for stock evaluation using fundamental analysis. Em·opcan Journal of Operational Rcscar·ch, 187:13801401, 2008. IV. Sharpe. Capital Asset Prices: A Theory of Market Equilibrium under Conditions of Risk. Journal of Finance, 19:425-442, 1964. J.D. Stowe, T.R. Robinson, J.E Pinto, and D.\V. McLeavey. Analysis of equity investments: val'aation. United I3ook Press, Baltimore, MD, 2002. M.C. \Vong andY. Cheung. The practice of investrnent management in Hong Kong: market forecasting and stock selection. The Internat'ional Jonrnal of ]Management Science, 27: 451-465, 1999. Y. Zhang. Analyst responsiveness and the post-earnings-announcement drift. Journal of Accounting and Economics, 46:201-215, 2008. \V.T. Ziemba and S.L. Schwart7,. Invest Japan: The Structure, Performance and Opportu,nities of .Japan's Stock, Bond and Fund Markets. Probus Publishing Company, Chicago, 1991.

25

This page intentionally left blank

Chapter 33

The Kelly Strategy for Investing: Risk and Reward Leonard C. MacLcan*and William T. Zicmbat May 8, 2012 Abstract

A strategy which has been of long standing interest in investing is the Kelly strategy, where the expected logarithm of wealth is maximized. There are many attractive properties when this strategy is used over a long planning horizon. Notably, the asymptotic growth rate of wealth is optimal and the time to reach asymptotically large wealth targets is minimized. However, the risk aversion of the Kelly strategy is essentially zero and the strategy is very risky in the short term. Investors can lose most of their wealth with a string of bad outcomes. In this paper the Kelly strategy and the associated fractional strategies are considered. The advantages and disadvantages are described in general and in particular with application to a variety of market scenarios. The conclusion discusses the use of Kelly type strategies by great investors.

1

Introduction

In portfolio management an investor must determine how much capital to bet in risky assets at specified points in time, with a focus on the accumulation of capital to each time or at a planning horizon. A important investment strategy is the Kelly strategy, where the expected logarithm of final wealth is maximized. (Kelly, 1956.) Log utility dates to Daniel Bernoulli in 1738 who postulated that marginal utility was monotone increasing but declining with wealth and, specifically, was equal to the reciprocal of wealth, w which yields the utility of wealth u( w) = log( w). Prior to this it was assumed that decisions were made on an expected value or linear utility basis. This idea ushered in declining marginal utility (concavity) or risk aversion which is crucial in investment decision making. In his paper Bernoulli also discussed the St Petersburg paradox and how *School of Business Administration, Dalhousie University, Halifax, NS, Canada B3H 3J5 t Alumni Professor of Financial Modeling and Stochastic oーエゥュ^セ。ッョ@ (Emeritus), Sauder School of Business, University of British Columbia,Vancouver, BC, Canada V6T 1Z2, and Visiting Professor, ICMA Centre, University of Reading, Reading, UK, Visiting professor, University of Cyprus, Sabanci University, Istanbul, Luiss Carla Guido University, Rome and Korea Advanced Institute for Science and Technology.

1

637

638

LC MacLean and WT Ziemba

it might be analyzed using log(w). This problem concerns how much to pay for the following gamble: A fair coin with セ@ probability of heads is repeatedly tossed until heads occurs, ending the game. The investor pays c dollars and receives in return 2k-I with probability 2-k fork= 1, 2, ... should a head occur. Thus, after each succeeding loss, assuming a head does not appear, the bet is doubled to 2, 4, 8, ... etc. Clearly the expected value is セ@ + セ@ + セ@ + ... or infinity with linear utility. Bell and Cover (1980) argue that the St Petersburg gamble is attractive at any price c but the investor wants less of it as c --+ oo. The proportion of the investor's wealth invested in the St Petersburg gamble is always positive but decreases with increasing cost c . The rest of the wealth is in cash. Bernoulli offers two solutions since he felt that this gamble is worth a lot less than infinity. In the first solution, he arbitrarily sets a limit to the utility of very large payoffs. Specifically, any amount over 10 million is assumed to be equal to 224 . Under that bounded utility assumption, the expected value is 24 1 1 1 2(1)+4(2)+8(4)+ ... + ( 21) (2 24 )+ ( 21 )

25

(2 24 )+. .. = 12Hhe originall = 13.

When utility is log the expected value is 1

1

1

2zog1 + 4zog2 + + 8log4 +

... = log2 = 0.69315.

As Menger (1934) pointed out, the log, the square root and many other, but not all, concave utility functions eliminate the original St Petersburg paradox but it does not solve one where the payoffs grow faster than 2n. So if log is the utility function, one can create a new paradox by having the payoffs increase at least as fast as log reduces them so one still has an infinite sum for the expected utility. With exponentially growing payoffs one has 1

1

2zog(e) + 4zog(e

2

) + ... = oo.

The super St Petersburg paradox, in which even ElogX = oo is examined in Cover and Thomas (2006: p181, 182) where a satisfactory resolution is reached by looking at relative growth rates of wealth. To solve the St. Peterburg paradox with exponentially growing payoffs, a solution is simply to choose a utility function which, though unbounded, grows "sufficiently more" slowly than the inverse of the payoff fundion. Kelly (1956) is credited with using log utility in investing and repeated games. His analysis uses Bernouli trials. He established that log is the utility fiunction which m&'Cimizes the long run growth rate, and is myopic in the sense that period by period maximization is optimal. Latane (1959) introduced log utility as an investment criterion to finance independent of Kelly's work. Breiman (1961) established the basic mathematical properties of the expected log criterion: (i) 2

Chapter 33. The Kelly Strategy for Investing

639

wealth for the Kelly strategy overtakes almost surely that of any other essentially different strategy as the horizon bemmes infinitely distant; (ii) the strategy attains arbitrarily large wealth goals faster than any other strategy; (iii) with a fixed opportunity set a fixed Kelly strategy is optimal. In an economy with one log investor and all other investors with essentially different strategies, the log investor will eventually get all the economy's wealth. (Hens and Schenk-Hoppe, 2005.) The drawback of log, with its essentially セ・イ。@ Arrow-Pratt absolute risk aversion, is that in the short run it is the most risky utility function one would ever consider. Since there is essentially no risk aversion, the investments it suggests are very large and typically undiversified. Simulations show that log investors have much more final wealth most of the time than those using other strategies, but those investors can essentially go bankrupt a small pen:entage of the time, even facing very favorable choices. (Ziemba and Hausch, 1986.) One way to modify the growth-security profile is to use either ad hoc or scientifically computed fractional Kelly strategies that blend the log optimal portfolio with cash. For instance, a fractional Kelly strategy will keep accumulated capital above a specified wealth path with high probability given log normally distributed payoffs. This is equivalent to using a negative power utility function whose coefficient (analogous to a risk aversion index) is determined by the fraction and vice versa. Thus one moves the risk aversion away from zero to a higher level. This results in a smoother wealth path but has less growth. For non-lognormal payoff distributions, the fractional Kelly is an approximate solution to the optimal risk-return trade-off that can be inaccurate (MacLean, Thorp and Ziemba, 2010b). Davis and Lleo (2012) define a variation on the fractional Kelly strategy, where the blend is between the Kelly and an interternporal hedging portfolio. If returns are generated by a diffusion with Poisson jumps, then the modified fractional Kelly is optimal for the negative power utility function.

2

The Kelly Strategy

Consider a set of K investment opportunities whose outcomes are stochastic, that is they arc defined on the probability space (S1, B, P). Assume there is a payoff or return from investing in opportunity i at time t , given the outcome wE S1, defined by by r.,(t, w), i = 1, ... , K. For a unit of capital invested in asset i at time t the gross return is Ri(t,w)

=

1 +ri(t,w),i

=

l, ... K.

(1)

In the financial market, investments are made at points in time and the return on assets leads to the accumulation of capital for a bettor. In the analysis of investment strategies, the following structure is assumed: (a) All assets have limited liability. (b) There are no transactions costs, taxes, or problems with indivisibility of capital. (c) Capital can be borrowed or lent at a risk free interest rate r. 3

640

LC MacLean and WT Ziemba

(d) Short selling (borrowing against the asset return) is allowed. The returns on outcomes at time t from the K assets generate random vector R' (t) = (Ro( t), R 1 ( t), ... , RK (t) ), where R 0 ( t) = 1 + r. Suppose an investor has Wt units of capital at time t, with the proportions invested in each asset given by xi(t), i = 1, ... K, and X = (.1: 1 (t), ... , .TK(t))'. If the capital not invested is xi(t), and earns the risk free rate, then an investment strategy x 0 (t) = 1- lセャ@ at time t is the vector process

X(t)

=

(x 0 (t),X(t))'.

(2)

Given the investments WtX(t) at timet, the accumulated capital at time t

+1

lS

K

W(t

+ 1) =

WtR'(t)X(t)

=

Wt LRi(t)xi(t).

(3)

i=O

For a sequence ofT investment periods starting with capital w 0 , wealth at time Tis T

W(T)

= w0

IT R'(t)X(t).

(4)

t=l

Alternatively, wealth is

W(T)

セ@

w0

t H」イー{セ@

ln(R'(t)X(t))[)

The exponential form highlights the growth rate 1 T Gr(X) = T L ln(R'(t)X(t)),

7

(5)

(6)

t=l

with the strategy X= (X(1), ... , X(T)). If the distribution of accumulated capital (wealth) at the horiwn is the criterion for deciding on an investment strategy, then the rate of growth of capital becomes the determining factor when the horizon is distant. Consider then the average growth rate 1 T

EGr(X)

=

T L Eln(R'(t)X(t)).

(7)

i=l

The case usually discussed is when the incremental returns are serially independent. So the maximization of EGr(X) is max {Eln(R'(t)X(t)}, (8) separately for each t. If the returns distribution is the same for each t, a fixed strategy holds over time. The strategy which solves (8) is called the KELLY or optimal growth strategy. In a continuous time formulation with the prices on risky assets having a stationary lognormal distribution, the Kelly decision is determined from 4

Chapter 33. The Kelly Strategy for Investing

セ@

Qセ@

ma:c { (IL- r-e)' X+ r--

セス@

2x'6.X

641

(9)

,

with log returns having mean /Li, i = 1, ... , K. and covariance 6. = (5;J). Also, X' = ( x1, ... , :c K) are the investments in risky assets. For this continuous time problem, the Kelly strategy has the closed form

(10) The Kelly or log optimal portfolio is X*'= (x 0,X*'), where x 0 = 1- RZセ Q@ The continuous time formula can be viewed as an approximate solution to the discrete time betting problem. The Kelly strategy is a fixed mix. That is, the fraction of wealth bet in games is determined by X*, but rebalancing as wealth varies is required to maintain the fractions. The closed form for the Kelly styratcgy is useful, but it is based on the lognormality assumption. It is possible to move to a more general returns distribution by stratifying the financial market, with lognormality within a stratum. Assume that the financial market is separated into m. distinct strata or regimes, with the financial market switching regimes through a random process. Suppose the market is in regime kat timet, and let Yik(t),i = O, ... ,n be the log-prices in regime k. The price dynamics within a regime arc defined by the stochastic differential equations (ll) dYok (t) = Ttdt

dYk (t) = akdt

where Yk(t)

=

(

Y1k(t) ) • n'k '"

=

1

+ VNセ、zォL@ (

(12)

k = 1, ... , m

Ct1k )

,

6.k

=

(5ijk), dZk

=

Ynk(t) Ctnk where dZ;k, ·i = 1, ... , n are independent Brownian motions. So the risky asset prices in a regime are assumed to have a joint log-normal distribution. Consider that the regimes over time {81 , 8 2 , ... , Sn} follow a }.1arkov chain. The probability of switching from regime i at time t to regime j at time t + 1 is P[St+ 1 = jiSt = i] = PiJ· The transition between regimes is defined by the (stationary) transition matrix P

=

(

pーョセ[@

1 1

pーョセ[@

nn ) . Given an ini-

tial distribution over regimes p(O) = (p1(0), ... ,pn(O)), the subsequent regime distributions p(t) are determined by transitions. Then the conditional wealth at the end of period t if the regime is k, given the fixed investment strategy X ( t), is

Wk(t)

=

where Z£

w · exp { [X'(t)({tk- r-e) =

(Z1k, ... , Znk), ZJk セ@

+ r-- セxGHエIVNォ}@

N (0, 1). Let

5

+ xHエIGVNセ@

Zk}, (13)

642

LC MacLean and WT Ziemba

{xGHエIセォM

Rk(X(t)) = exp {

Te)

+ T- セxGHエIlャォ}@

+ X(t)' Llt zk}

(14)

be the return on the investment X ( t) in assets in period t. The rate of return is

ln(Rk(X(t)))

{クGHエIセォM

=

Te)

+ T- セxGHエIlャォ}@

+ X'(t)Llt Zk.

(15)

The parameters in the regime formulation are the transition matrix and the (mean, covariance) for log-returns within regimes. The distribution of logreturns conditional on the regime is normal, and the unconditional distribution is a mixture of normals. The mixing parameters are the transition probabilities. The mixture model is very suitable for returns distribution. However, the mixing parameters need to be correctly identified. The EM algorithm (Demptster et. al., 1977) has been successfully applied to this problem. If the rows in the transition matrix are the same then the mixture is defined by probabilities (P1,···,Pn). In this case the mixing parameters are the same for all time periods. The Kelly strategy is determined from rnax RZセ] Q@ Pk · E(ln(Rk(X)). This gives

X*=

whece X'

セ@

(

HセpォゥャI@

-

1HセpォヲャM

Te)),

· ) me the fraction" inv"ted in the n ri"ky """"'·

In the regimes case the Kelly strategy has a closed form, which is essentially the same as the Merton formula.

2.1

Performance: Risk and Return

The Kelly strategy is optimal for the asymptotic growth rate criterion, but there are other measures of performance which are of interest. In Table 1, some measures are defined, with a classification by wealth and time. In the definitions in the table, the notation T(W(X)::,. n) is the first passage time to the set [u, oo) and Wy is a Value at Risk (VaR) level. Return

Risk

py(X)

=

E {tn[W(T,X)]+ J

T/u(X)

=

E{T(W(X)::,. u)}

!T(X) Pl,u(X)

=

=

Pr[lnW(T, X) ::,. ln(wy )]

PT[T(W(X)::,. u) < T(W(X)

«:: 1)]

Table 1: Performance Measures The wealth and time dimensions of the stochastic process {W(t, X), t::,. 0} are alternative perspectives, with the time component emphasizing growth speed and wealth emphasing growth magnitude. The distributions for the random quantities are illustrated in Figure 1.

6

Chapter 33. The Kelly Strategy for Investing

Time

Figure 1: Wealth and Time dimensions of Wealth Process The Kelly strategy is defined by the expected value, but the risk characteristics are significant as \Nell. Risk is defined by the chance of falling short of targets. By t he wealth criterion, a value-at-risk 'WT is specified at the horizon T, so risk is defined by a VaR condition. In the case of passage time, the chance of achieving a desired wealth target before falling to an undesirable level is assessed. The Kelly strategy is optimal for the expected values rp and ry, but may not do well on risk measures 'Y and (3. These performance measures have been considered in several papers in the finance and probability literature (see, for example, Browne (1997), Dohi et al. (1995), Ethier and Tavare (1983), MacLean and Ziemba (1999), MacLean et al. (1992) , and Stutzer 2003). The evaluat ion of these measures is the basis for selecting an appropriate betting strategy, t hat is a strategy which satisfies our preferences. If the returns distributions are stationary and the strategy has constant investment proportions, a case which includes the Kelly strategy, then standard results on random walks (discrete time) and diffusions (continuous time) can be adapted to provide computational formulas for t he measures. For the evaluation of measures in the continuous time formulation of the log wealth process without regimes, we refer to Dohi et al. (1994) . Consider the following notation: 1. The total investment in risky assets is >.(X)

=セ

Q@ x., .

2. The risky asset s are combined into a single fund with instantaneous rate 7

643

644

LC MacLean and WT Ziemba

3. The mean and variance of the instantaneous rate of return on the fund are p,(.X) and CJ 2 (X), respectively. 4.

D(X)

=

+T

(JL(X)-r)A(X)

(a

2

("Y)2"' 2 (X)), where it is assumed D(X)

> 0.

5. セG^{}@\ is the cumulative normal distribution. Table 1 has formulas for the measures. I

Formula

Measure I

tpr(X)

(JL(X)- r),\(X) + r) T

D(X)·T-tn(;:;;;; ))

(

/T(X)

a(X)A(X)VT

dHセyNI@

7!u(X)

ln ( ャセオI@

1-(-";--) no

Bt,u(X)

2D X

u2(X)-'2(X) 2/J(X)

1-( T) u2(X).A2(X)

Table 2: Continuous time formulas In the case of financial market regimes, similar formulas apply. For example the one period problem with T = 1, TVo = 1, wy = w, n

tp(X)

=

LPk [(JLk(X)- r) ,\(X)+ r]

(16)

k=l n

!(X)= LPk Eu(W2(T))

for all

'U E uk,

with strict inequality for at least one

'U.

Another formulation for stochastic dominance follows from the distribution functions for wealth. Let W1 (T) and W 2 (T) have densities h and h, respectively. For a density .f, mnsider the nested integrations

(20) where Io = .f. Then W1 kth order dominates W2 if and only if I{' (w) -s: I { 2 ( w), for all w, with strict inequality for one w. It is convenient to write the nested integral in terms of loss functions of the wealth distribution. Consider wealth W, with distribution Fw. Let

(21)

P1 (a)= inf{7JIFw(7J) 2 a}, the ath percentile of the wealth distribution. For k ;-:> 2, define

w

Pk (a)=

!

Pk-da)

-oo

(Pk-1(a)- w)

k-1

dFw(w),

(22)

which is a normed measure of loss along the distribution for W. Then W1 kth order dominates W 2 if and only if ーセカ Q H。I@ -s: セカー R H。IL@ for all a, with strict inequality for one o.

10

Chapter 33. The Kelly Strategy for Investing

647

Although the utility based definition and the distribution based definition arc conceptually different, they are interchangable. Consider alternative terminal wealth variables WI (T) and W 2 (T). The following formulations of kth order stochastic dominance are equivalent: 1. Eu(WI) ;:;> Eu(W2) for u E Uk, with strict inequality for at least one u.

2.

I{,' (w) セ@ I1"2 (w),

for all w, with strict inequality for at least one w.

セ@ pr2 (a), for all a, with strict inequality for at least 3. Fork;:;> 2, ーセLカGH。I@ one a. Fork= 1, p;v'(a) ;:;> p;v2 (a), for all a, with strict inequality for at least one a.

The variations on the dominance relation provide insight into the conditions for the wealth process. The loss function definition of dominance is useful. It is clear that Plv(a),O セ。@ 1, is the r:umulative distribution, and ordering based on this measure expresses a preference for wealth. Since iセBGI@ (PI (a)- w) dF (w) = api (a) - iセ。I@

111

dF(w), the second order measure, p:f (a), 0 セ@

a セ@

1, is

w dF(w), and equivalent to the Lorenz curve (Lorenz 1905): L(a) = iセイ^I@ reflects risk aversion in that it is sensitive to the lower tail of the distribution. The measure p'f (a), 0 セ@ a セ@ 1, captures the aversion to variance. These characteristics are components of the preferences which investors might express with a utility function. In general, the ordering of wealth distributions at the horizon using stochastic dominance is not practical. Rather than use the full range of a values with the moments definition, it is more realistic to identify specific values and work with a lexicographic order. In particular, selecting two values of a to set up hi-criteria problems (pk( a 1), Pk ( a 2 )), has some appeal. Consider terminal wealth variables W 1 (T) and W 2 (T), and values n 1 < n 2. Then, fork;:;> 2, W1 (T) is kth order (a1, a2)-preferred to W2(T), denoted WI (T) »k W2(T), if (pr' (ai)Pr' (a2)) > (pr2 (ai),pkv2 (a 2 )). WI is pt order (ai,a 2 )-preferred to W 2 (T), if and only if (p;v1 (a I), ーセ QW@ ( a 2 )) > (p;v2 (a I), p;v2 ( a 2 ) ). The preference ordering is isotonic with respect to the dominance ordering: WI (T) セォ@ w2 (T) =? wl (T) »k w2 (T) (Ogryczak and Ruszczynski 2002). The (a 1 , a 2) values at which the Pk functions are compared has some convention. In Table 2, a pair of a values are used with the first, second and third order dominance relations to generate measures which can be used as a basis of comparing wealth distributions. I

Order First Second Third

-

I

PI (0.5) P2(l.O) P3(1.0)

PI (0.05) P2(0.05) P3(0.5)

Table 4: Stochastic Dominance Based Measures The intention in the table is to capture important aspects of wealth preferences in the choice of a values. So, the first order pair identifies the median wealth and 11

648

LC MacLean and WT Ziemba

the fifth percentile. The second order pair considers the mean and the lower 5% incomplete mean. The third captures the mean squared wealth and the lower semi-mean squared wealth.

3.1

Bi-criteria Problems

The problem of determining the optimal tradeoff of expectation and risk can be formalized by using a utility function over these two attributes similar to that in mean-variance analysis, sec e.g. Markowitz (1952, 1987). MacLean ct al (2004) do this another way by adding probability constraints on the drawdown then, using scenario analysis, an optimal fractionalized Kelly strategy can be determined. The examples in section 4 illustrate this tradeoff choice. Analogous to static mean-variance analysis, see Markowitz (1952,1987), a growthsecurity combination pair is inefficient if another pair in that combination has either a higher mean growth and no lower security level or a higher security level and no lower growth rate. A strategy is inefficient if its growth security combination is inefficient. Efficient growth-security combinations are those which are not inefficient. The efficient growth-security frontier is the set of all efficient growth- security pairs. An efficient tradeoff between growth and security occurs along the efficient frontier. The efficiency problems based on the performance criteria in Table 1 are given in Table 5. The constraint on security sets an acceptable risk a for not meeting the performance target. I

Model

Jvh Jvh Jvh

I

Performance Problem

Dominance Problem

1\!IaxE ャセ@ [W(T)c- 1]j A'Iax{;pr(X)br(X):;, 1- a} IVIin {'l)u(X)I6z,u(X):;, 1- a}

max {Poo(l.O)I&::; c} max {pl(0 ..5)1Pl(a):;, wr} max {PI(0.5)II'Jz.u(X):;, 1- a}

Table 5: Alternative Decision Models In addition to the growth-security problems, an expected utility problem 1Vh is defined. The power utility plays an important role in the Kelly and fractional Kelly strategies. The log is the power utility as p ---+ 0. The coefficient p is a risk aversion parameter, and the Arrow-Pratt relative risk aversion index is (1- p). If X*is the Kelly strategy for risky gambles, then Xf = ャセーxJL@ < 0, is the optimal solution to problem 1Vh in the continuous time case where returns are lognormally distributed. That formula gives an indication of the risk aversion property of fractional Kelly. That optimality property applies to the other problems. The subclass of fractional Kelly strategies is defined as x;={X/IX/=J·X*,o::;j::;l}.

(23)

The significance of the fractional Kelly strategies lies in their optimality for the problems in Table 5, assuming the geometric Brownian model for returns is correct.

12

Chapter 33. The Kelly Strategy for Investing

649

THEOREM

Let XIVrJ (t) be the optimal solution to growth problem J\!Ij, j = 1, 2, 3 defined in Table 5. Then X 111j ( t) E x;, that is, the solution is fractional Kelly. For proof see MacLean, Ziemba and Zhao (2005). In the continuous time formulation, the optimal investment strategies for the various problems have the same form. However, the actual fraction in each problem, which controls the allocation of capital to risky and risk-free instruments, depends on the decision model and parameters. The formulas for the fractions for different models are in Table 6. The notation pセ@ = (cj;- re)'X*(t) + r, and iJ 2 = X*' (t)l:,.X*(t) is used for the mean and variance of the rate of return on the Kelly strategy. Also at time t, is the minimum positive root of the equation rvyct+l - ·y + (1- rv) = 0 for c = ln(u)-ln(wt).

y;

1

I

Model

'

t

ln(wt)-ln(l)

Parameters

Fraction = ャセ」@

fi

H- --

((l, u), a)

M3

h _ t -

jj-T -2

ln(w,J-ln(l) ln( 1Vt) -ln(y* ·l)

Table 6: Investment Fractions The solutions displayed in the Table 6 are derived from the continuous time wealth equation, although the strategies are calculated at discrete decision points in time. The alternative problems in Table 5 can be based on the discrete time wealth equation, but the optimal solution is not necessarily fractional Kelly. However, the fractional Kelly solution may be near-optimal. If the feasible strategies for the discrete time problem are restricted to the class of fractional strategies, the solutions are effective (MacLean, Ziemba and Blazenko, 1992). That is, as the fraction changes, the growth (objective) and security (constraint) move in opposite directions, so that growth is monotone non-increasing in security. Specifically, for 0:::; .f :::; 1,

{ ;cPT(Xj)

セ@

0,

[セイHxェI@

S: 0}, { ;1Ju(Xj) 13

セ@

0,

;f3z.u(Xj) S: 0}.

(24)

650

LC MacLean and WT Ziemba

The implication of this monotonicity is that growth can be traded for security using the fraction allocated to the optimal growth portfolio. So the growth - security trade-off can be observed for various fractional Kelly strategies and suitable fractions (meeting investor preferences) can be determined.

4

Simulations

The evaluation of performance measures and the trade-off between risk and return with the fractional Kelly strategies has been developed analytically. he formulas and the contrast in risk and reward depend on parameters in the return distribution. This will be quantified with simulated trajectories from some well known cases (Bicksler and Thorp, 1973).

4.1

Bicksler - Thorp (1973) Case I - Uniform Returns

There is one risky asset R having mean return of + 12.5% , with the return uniformly distributed between 0.75 and 1.50 for each dollar invested. Assume we can lend or borrow capital at a risk free rate r = 0.0. Let ).. = the proportion of capital invested in the risky asset, where ).. ranges from 0.4 to 2.4 . So ).. = 2.4 means $1.4 is borrowed for each $1 of current wealth. The Kelly optimal growth investment in the risky asset for r = 0.0 is x = 2.8655. Tosee this consider the problem Jv!axx {E(ln(1 + r + x(R- r)}, where R is uniform on [a, b] and T =the risk free rate. We have the first order condition b R- T 1 ..,....-----(;-=---;-) X - - dR = 0,

l

a

1+r+x R-r

b-a

which reduces to

x(b- a)

=

(1

+ T)ln

(

1+r

+ x(b -

r) ) セ@ 1+r+x(a-r)

1

[1+r+x(b-r)]" ]・ゥセN@ 1+r+x(a-r)

In the case considered in the experiment , a = -0.25, b = 0.5, r = 0. The 1

°; J;;

equation becomes [ 11! 0 5xx = e 0 · 75 , with a solution x = 2.8655. The Kelly fractions for the different values of ).. are shown in Table 7. (The formula relating ).. and f for this experiment is given in MacLean, Thorp and Ziemba 2010b.) In their original simulation, Bicksler and Thorp use 10 and 20 yearly decision periods, and 50 simulated scenarios. \Ve use 40 yearly decision periods, with 3000 scenarios. Proportion: ).. Fraction: f Table 7: The Investment Proportions and Kelly Fractions for Case 1: MacLean, Thorp and Ziemba 2010b

14

Chapter 33. The Kelly Strategy for Investing

651

The numerical results from the simulation with T = 40 are in Table 8 and Figures 2 - 4. Although the Kelly investment is levered, the fractions in this case are less than 1. Fraction Statistic :\![ax Mean Min St. Dev. Skewness Kurtosis > 5 X 10 10-

>

5

X

10 2

> 10J > 104 > 105 u > 10

0.14k 34435.74 7045.27 728.45 4016.18 1.90 6.00 3000 3000 3000 2998 529 0

0.28k 743361.14 45675.75 425.57 60890.61 4.57 31.54 3000 3000 2999 2997 2524 293

0.56k 124068469.50 1538429.88 70.97 6047844.60 10.80 150.51 3000 2999 2997 2991 2851 2025

0.70k 1070576212.0 7877534.72 18.91 44547205.57 13.39 223.70 2999 2999 2991 2980 2847 2243

0.84k 7399787898 36387516.18 3.46 272356844.8 15.63 301.38 2998 2998 2976 2965 2803 2290

0

0

696

1165

1407

Table 8: Final Wealth Statistics by Kelly Fraction for Case !:MacLean, Thorp and Ziemba 2010b In this experiment the Kelly proportion is high, based on the attractiveness of the investment in stock. The largest fraction (0.838k) shows strong returns, although in the worst scenario most of the wealth is lost. The trajectories for the highest and lowest terminal wealth scenarios are displayed in Figure 6. The highest rate of growth is for the highest fraction, and correspondingly it has the largest wealth fallback.

15

652

LC MacLean and WT Ziemba Ln(Wealth)

25 Variable

- - 0.28k ---- 0.42k O.SGk ____.- 0.70k 0.84k

- -+- -

20

-+15

10

UlML⦅Nセ@

20

10

0

30

40 Time

(a) :\llaximum Ln(Wealth)

8 7 6

5 4 3

Variable - - 0.28k

2

---- 0.42k O.SGk

--+--

--4- 0.70k

1 P セ

-+ - 0.84k L⦅

M 0

LM

M 10

LM

ML

20

NMセ@

30

40 Time

(b) Minimum

Figure 2: Trajectories wit h Final Wealt h Extremes for Case I: MacLean, Thorp and Ziemba 2010b The distribution of terminal wealth in Figure 7 illust rates t he growt h of t he = 0.838k strategy. It intersects t he normal probability plot for ot her st rategies very early and increases its advantage. The linearity of t he plot s for all strat egies is evidence of the log-normality of final wealth. The inverse cumulat ive distribution plot indicates t hat t he chance of losses is small - the horizontal

f

16

Chapter 33. The Kelly Strategy for Investing

653

line indicates log of initial wealth. ln(Wealth)

25 Variable

- - 0.28k

- - 0.42k ---- 0.5Gk

20

- - 0.70k - - - 0.84k

15

10

5

0"--.----.----.----.----.----.-" 0.0

0.2

0.4

0.8

0 .6

1.0 CumPr

(a) Inverse Cumulative ln(Wealth)

25 Variable

- - 0.28k

20

- - 0.42k ---- O.SGk

- - 0.70k - - - 0.84k

15

10

5

d・NMLセ@

-4

-3

-2

-1

0

1

2

3

4 NScore

(b) Normal P lot

Figure 3: Final Ln(Wealth) Distributions by Fraction for Case 1: MacLean, Thorp and Ziemba 2010b As further evidence of the superiority of t he f = 0.838k strategy consider t he mean and standard deviation of log-wealth in Figure 8. The growth rate (mean

17

654

LC MacLean and WT Ziemba Ln(Wealth)

0.2

QTイMセ@

0.4

0.6

0.8

Stoev

Mean

3.0

13

2.5

12

2.0

11

1.5

1.0 QPL⦅Mセ@

0.2

0.4

0.6

0.8 fセ。」エゥッョ@

Figure 4: Mean-Std Trade-off for Case I: MacLean, Thorp and Ziemba 201Gb

ln(Wealth)) continues to increase since the fractional strategies are less then full Kelly. From the results of this experiment we can make the following statements. 1. The statistics describing end of horiwn (T = 40) wealth are again mono-

tone in the fraction of wealth invested in the Kelly portfolio. Specifically the maximum terminal wealth and the mean terminal wealth increase in the Kelly fraction. In contrast the minimum wealth decreases as the fraction increases and the standard deviation grows as the fraction increases. The growth and decay are much more pronounced than was the case in experiment 1. The minimum still remains above 0 since the fraction of Kelly is less than 1. There is a trade-off between wealth growth and risk, but the advantage of leveraged investment is clear. As illustrated with the cumulative distributions, the log-normality holds and the upside growth is more pronounced than the downside loss. Of course, the fractions are less than 1 so improved growth is expected. 2. The maximum and minimum final wealth trajectories clearly show the wealth growth - risk of various strategies. The mean-standard deviation trade-off favors the largest fraction, even though it is highly levered.

4.2

Bicksler - Thorp (1973) Case II - Equity Market Returns

In the second experiment there are two assets: US equities and US T-bills. According to Siegel (2002), during 1926-2001 US equities returned of 10.2%

18

Chapter 33. The Kelly Strategy for Investing

655

with a yearly standard deviation of 20.3%, and the mean return was 3.9% for short term government T-bills with zero standard deviation. We assume the choice is between these two assets in each period. The Kelly strategy is to invest a proportion of wealth x = 1.5288 in equities and sell short the T-bill at 1 - x = -0.5228 of current wealth. Consider the problem Jvfaxx {E(ln(1

+ r + x(R- r)},

where R is Gaussian with mean f-iR and standard deviation free rate. The solution is given by Merton (1990) as f-iR -r

r7R,

and r =the risk

X=---. !JR

The values in this experiment are fiR = 0.102, uR = 0.203, r = 0.039, so the Kelly strategy is x = 1.5288. With the short selling and levered strategies, there is a chance of substantial losses. For the simulations, the proportion: ).. of wealth invested in equities and the corresponding Kelly fraction f are provided in Table 9. (The formula relating A and f for this experiment is in the Appendix.) In their original simulation, Bicksler and Thorp used 10 and 20 yearly decision periods, and 50 simulated scenarios. We use 40 yearly decision periods, with 3000 scenarios. I

1

A

t

I

1

0.4

0.8

1.2

1.6

2.0

2.4

o.26

o.52

o.78

1.o5

1.31

1.57

Table 9: Kelly Fractions for Case II: MacLean, Thorp and Ziemba 2010b The results from the simulations with experiment 3 are contained in Table 10 and Figures 5 - 7. This experiment is based on actual market returns. The striking aspects of the statistics in Table 10 are the sizable gains and losses. For the the most aggressive strategy (1.57k), it is possible to lose 10,000 times the initial wealth. This assumes that the shortselling is permissable through to the horizon.

19

656

LC MacLean and WT Ziemba

Table 10: Final Wealth Statistics by Kelly Fraction for Case II: MacLean, Thorp and Ziemba 2010h Fraction Statistic Max Mean Min St. Dev. Skewness Kurtosis > 5 X 10 QPセ@

X QPセ@ 103 10" 100 " >10

> Ei > > >

0.26k 65842.09 12110.34 2367.92 6147.30 1.54 4.90 3000 3000 3000 3000 1698 0

0.52k 673058.45 30937.03 701.28 35980.17 4.88 51.85 3000 3000 3000 2996 2276 132

0.78k 5283234.28 76573.69 -4969.78 174683.09 13.01 305.66 2998 2998 2986 2954 2273 575

1.05k 33314627.67 182645.07 -133456.35 815091.13 25.92 950.96 2970 2955 2866 2779 2112 838

1.57k 769753090 895952.14 -102513723.8 15004915.61 45.45 2303.38 2184 2129 1960 1875 1375 751

0

0

9

116

270

The highest and lowest final wealth trajectories are presented in Figure 5. In the worst case, the trajectory is terminated to indicate the timing of vanishing wealth. There is quick bankruptcy for the aggressive strategies.

20

Chapter 33. The Kelly Strategy for Investing

Figure 5: Trajectories with Final Wealth Extremes for Case II: MacLean, Thorp and Ziemba 2010b The strong downside is further illustrated in the distribution of final wealth plot in Figure 6. The normal probability plots are almost linear on the upside (log-normality), but the downside is much more extreme than log-normal for

21

657

658

LC MacLean and WT Ziemba

all strategies except for 0.52k. Even the full Kelly is risky in this case. The inverse cumulative distribution shows a high probability of large losses with the most aggressive strategies. In constructing these plots the negative growth was incorporated with the formula growth= [signWr]ln(IWrl).

22

Chapter 33. The Kelly Strategy for Investing

659

Ln(Wealth)

20

10

0 Variable - - 0.52k - - 0.78k

-10

- - - - l.OSk

- - - 1.31k - - - 1.57k

-20 セMNLB@ 0.0

0.2

0.4

0.6

1.0

0 .8

CumPr

(a) In verse Cumulative Ln(Wealth)

20

10

0 V.:.riable

-- ----- -

-10

O.S2k 0.78k !.OSk 1.3lk

- - - 1.57k

MRPlLNセ@ -4

-3

-2

-1

0

1

2

3

4 NScore

(b) :\formal Plot

Figure 6: Final Ln(Wealth) Distributions by Fraction for Case II: MacLean, Thorp and Ziemba 2010b The mean-standard deviation trade-off in Figure 7 provides more evidence to the riskyness of the high proportion strategies. When the fraction exceeds the full Kelly, the drop-off in growth rat e is sharp, and t hat is matched by a sharp

23

660

LC MacLean and WT Ziemba Ln(Wealth)

0.50 Mean

0 .75

1.00

1.25

1.50

Stdev

8

10 7 6

9

5 8

4

3 7

2 1

6 0 0.50

0. 75

1.00

1.25

1.50 Fraction

Figure 7: Mean-Std Tradeoff for Case II: MacLean, Thorp and Ziemba 2010b increase in standard deviation. The results in experiment 3 lead to the following conclusions. 1. T he st atist ics describing t he end of the horizon (T = 40) wealt h are again monotone in the fraction of wealt h invested in t he Kelly portfolio. Specifically (i) t he maximum terminal wealth and t he mean terminal wealt h increase in the Kelly fraction; and (ii) t he minimum wealth decreases as the fraction increases and t he standard deviation grows as t he fraction increases. The growth and decay are pronounced and it is possible to have large losses. The fraction of the Kelly optimal growth strategy exceeds 1 in t he most levered strategies and this is very risky. T here is a trade-off between return and risk, but the mean for t he levered strategies is growing far less t han t he standard deviation. T he disadvantage of leveraged investment is clearly illustrated wit h t he cumulative distribut ion. T he log-normalit y of final wealt h does not hold for t he levered strategies.

2. The maximum and minimum final wealth t rajectories clearly show the return - risk of levered strategies. T he worst and best scenarios are the not same for all Kelly fractions. T he worst scenario for t he most levered strategy shows t he rapid decline in wealth. T he mean-standard deviation trade-off confirms the riskyness/folly of t he aggressive strategies.

5

Applications

The alternative capital growth problems wit h t he Kelly fraction determ ined by t he performance measures will be applied to a variety of investment and 24

Chapter 33. The Kelly Strategy for Investing

661

gambling problems in this section.

5.1

The Fundamental Problem of Dynamic Asset Allocation - 1\1!2

The methods of capital growth are now considered with the fundamental problem of investment in stocks, bonds, and cash over time. Portfolio performance is primarily determined by the strategic allocation to these assets over time (Hensel, Ezra and Ilkiw ,1991; Blake, Lehman and Timmerman, 1999). The data for the assets covers the period 1985 to 2002, with stock returns from the S&P500 and bond returns from the US Long Government Bond. In Table 11 are the means, variances and covariances of daily returns over the period of the data. The risk free rate was set at 5% per annum. I

Parameter Mean Variance Covariance

I

Stocks 0.00050 0.00062 0.000046

Bonds 0.00031 0.00035

I

Cash

I

1

0

Table 11: Return Parameters With those values as parameters in the price model, daily price trajectories for one year were generated. At 10 day intervals along a price trajectory, investment decisions were calculated based on the model lvh. That is, the median wealth was maximized, subject to a VaR constraint. The median wealth objective is equivalent to a mean log-wealth objective. The solutions come from the fractional Kelly formulas. The price history was used to estimate the price parameters. This price generation and investment process was repeated 1000 times. The expected wealth at the end of a one year planning horizon is shown in Table 12 for a range of values for parameters (pi, a) in the VaR constraint.

Pi

I a 0.01 0.02 0.03 0.04 0.05

0.95 1.1292 1.1509 1.1703 1.1895 1.2091

0.96 1.1053 1.1189 1.1310 1.1432 1.1557

0.97 1.0881 1.0953 1.1019 1.1084 1.1154

0.98 1.0764 1.0797 1.0826 1.0854 1.0884

0.99 1.0629 1.0649 1.0666 1.0681 1.0696

Table 12: Expected Terminal Wealth The VaR constraint has a significant effect on wealth. With the value at risk level set at 95% of initial wealth, the increase in wealth is good, and grows as the probability level drops. With a greater VaR level, the growth in capital is somewhat less. Of course, as the VaR requirements are tightened, the security is increased. So there is a tradeoff between growth and security. 25

662

LC MacLean and WT Ziemba

For comparison the problem 2\lh was implemented in discrete time, with years as the time units. The statistics on annual returns from the data are provided in Table 13. Parameter Mean Variance Correlation

I

I

Stocks 0.08750 0.1236 0.32

Bonds 0.0375 0.0597

I

Cash

I

0 0

Table 13: Annual Return Statistics A corresponding set of scenarios was created (sampling from a lognormal distribution for stocks and bonds), as displayed in Table 14. The sampling process is structured so that sample statistics are as dose as possible to the statistics in Table 11. (:'dacLean, Sanegre, Zhao and Ziemba, 2004).

I

Stocks 0.95 1.065 1.085 1.250

I

Bonds 1.015 1.100 0.965 1.070

I

Cash 1 1 1

1

I

Probability 0.25 0.25 0.25 0.25

I

Table 14: Return Scenarios The planning horizon was set at 3 years and the same scenarios were used each year. So there was 64 scenarios, each with a probability of 1/64. With this discrete time and discrete scenario formulation, problem IVI2 was solved with a = 0.01 and a variety of values for Starting wealth is $1. The results from solving the problems are shown in Table 15. (Details on this problem are in MacLean, Sanegre, Zhao and Ziemba, 2004). If the annual returns are compared to the results for the continuous time problem with the same a, the returns in this case are slightly lower. However, the continuous time application involved rebalancing every 10 days. In Table 15, it can be observed that the very strict VaR condition almost eliminates the possibility of growth. As well, the optimal strategy is not fractional, with the investment mix changing as the horizon approaches.

pr.

26

Chapter 33. The Kelly Strategy for Investing

Pi

stocks

0.950 0.970 0.990 0.995 0.999

1 1

0.456 0.270 0.270

Year 1 bonds 0 0 0.544 0.730 0.730

cash 0 0 0 0 0

stocks 0.492 0.333 0.270 0.219 0.008

Year 2 bonds 0.508 0.667 0.730 0.590 0.020

cash 0 0 0 0.191 0.972

stocks 0.492 0.333 0.270 0.218 0.008

663

Year 3 I cash bonds 0.508 0 0.667 0 0.730 0 0.192 0.590 0.972 0.020

Table 15: Investment Strategy and Rate of Return

5.2

Blackjack - 1\1!3

The game of Blackjack, or 21, evolved from several related card games in the 19th century. It became fashionable during World セw。イ@ I and now has enormous popularity, played by millions of people in casinos around the world. Billions of dollars are lost each year by people playing the game in Las Vegas alone. A small number of professionals and advanced amateurs, using various methods such as card counting, arc able to beat the game. Sec Janecek (1998) for a computer program that evaluates game statistics such as advantage depending on the card counting system used and casino rules. The object is to reach, or be close to, twenty-one with two or more cards. Scores above twenty-one arc said to bust or lose. Cards two to ten are worth their face value: Jacks, Queens and Kings are worth ten points, and Aces are worth one or eleven at the player's choice. The game is called blackjack because an ace and a ten-valued card was paid three for two and an additional bonus accrued if the two cards were the Ace of Spades and the Jack of Spades or Clubs. While this extra bonus has been dropped by current casinos, the name has stuck. Dealers normally play a fixed strategy of drawing cards until the total reaches seventeen or more, at which point they stop. A variation is when a soft seventeen (an ace with cards totaling six) is hit. It is better for the player if the dealer stands on soft seventeen. The house has an edge of 1-lOo/c against typical players. The strategy of mimicking the dealer loses about 8% because the player must hit first and busts about 28% of the time (0.28 2 "'=' 0.08). However, in Las Vegas the average player loses only about 1.5%. The edge for a successful card counter varies from about -5% to + 10% depending upon the favor ability of the deck. By wagering more in favorable situations and less or nothing when the deck is unfavorable, an average weighted edge is about 1-2%. An approximation to provide insight into the long-run behavior of a player's fortune is to assume that the game is a Bernoulli trial with a probability of success p = 0.51 and probability of loss 1-p = 0.49. Then !1 = 0, 1, K(O,p) = p with probability1r , and K(l,p) = -p with probability 1-Jr. The mean growth rate is Eln(l + K(w, p)) = 1rln(l + p) + (1-Jr )ln(l-p). The optimal fixed fraction strategy is P* = 21r-l if EK > 0; P* = 0 if EK :::.:; 0. This optimal strategy may be interpreted as the edge divided by the odds (1-1 in this case). In general, for two outcome win-or-lose situations where the size of the wager docs not

27

664

LC MacLean and WT Ziemba

influence the odds, the same edge divided by the odds formula holds. Hence, with a 2% edge, betting on a 10-1 shot, the opt imal wager is 0.2% of one's fortune. The growth rate of t he investor's fortune is shown in Figure 8. It is ・ エイゥ」。ャ@ around p* = 0.02. Security ュ ・ 。 セ オイ ・@ ゥセ@ also displayed in nearly セ ケ ュ Figure 8 in terms of the probability of doubling or quadrupling before halving. Since the growth rate and the security are both decreasing for p > P* , it follows that it is never advisable to wager more than P*· However, one may wish to trade off lower growth for more security using a fractional Kelly strategy. For example, a drop from p = 0.02 to 0.01 for a 0.5 fractional Kelly strategy, decreases t he growth rate by 25%, but increases t he chance of doubling before halving from 67% t o 89%. The optimal Kelly fraction for a required security of 80% is displayed: 0.013 = 0.66 x 0.02.

Proba bility

1.0

0.6

0.4

0.2

0.0 MNGBセ@K 0.0

Op imo l K•ll wager

0.01

0.02

0.03

Frac.tion of Wealth Wage red

Figure 8: Relative Growth versus Probability of Doubling before Halfing In t he Blackjack example, t he addit ional informat ion provided by t he security measure was important in reaching a final investment decision. In a flexible or adaptive decision environment, competing criteria would be balanced to achieve a satisfactory path of accumulated wealth. Professional Blackjack teams often use a fractional Kelly wagering strategy wit h t he fraction = 0.2 to 0.8. See Gottlieb (1985) for further discussion including t he use of adaptive strategies.

5.3

Horseracing -

}.;1 1

In a race with n horses, there are positive returns for wagers made on t he first t hree finishers. T he set of all outcomes with probability 1rijk is fl = (1, 2, 3), ... , (i ,j, k) , ... , (n- 2, n-1, n). If q; is t he probability that horse ·i wins, t hen Harville (1973) gave t he probability of an (i, j , k) fi nish as 1rij k = 28

Chapter 33. The Kelly Strategy for Investing

665

HQMアLIGイゥセ@ -qi). (For a bias correction to the Harville formulas see Hausch, Lo, and Ziemba 1994, 2008). An investor wagers the fractions Pil,Pi2,Pi3 of his fortune TVa on horse i to win, place, and show, respectively. One collects on a win bet when the horse is first, on a place bet when the horse is first or second, and on a show bet when the horse is first, second, or third. The order of finish does not matter for place and show bets. All bettors, wagering on a particular horse, share the net pool in proportion to the amount wagered, once the original amount of the winning bets are refunded and the winning horses share equally the resulting profits. A particular anomaly is the place and show wager. Let the player bets be (P;2,Pi3) for horse ito place and show respectively. Also (P;2, P;3) be the total place and show bets of other people and r2 = lセ]ャ@ P;2, r3 = lセ]ャ@ ri3· The return function on the bet p = { (p; 2, p 13 ), i = 1, ... , n} for a particular (i, j, k) outcome is Q pエRIMH[Kェク@ kHゥLェォIー]qpRKlセ Pi2 + Pj2 ) Pi2 + Pi2 Pj2 + Pj2 2

+ Q(P3 + lセ]ャ@

Pt3) - (Pi3 3

+ Pj3 + Pk3 + pijk3) X (

n

+Wo-

n

L

Pz2-

l=l,lioi,j,k

Pi:> p;;o

L

+ P;:o

n

Pt3 +r(

l=l,loli,j

L l=l,loli,j,k

+

Pj:> Pj:o

+ Pj:o

+ .

Pk:> ) Pik3 + Pk:o

n

Pz2

+

L

pl3).

l=l,loli,j

where Q = 1-the track take and r is a rebate fraction. This return function is developed in Hausch, Ziemba, and Rubinstein (1981). The return is net of transactions costs (track take) and includes the track rebate and the effect of bets on the odds. The optimal Kelly wager is determined from the problem

Max, {

セ@ jセLスGiッァ@

(K((ijk),p))]}

Subject to n

L(Pi2

+ P·i3)

K) is interested in hedging its risk. A dynamic hedging strategy for the writer of such an option is a dynamic investment policy, say {.6. 1 , t .=:: T}, which holds .6.r shares of the underlying stock at time t so as to ensure that the underwriter's position is riskless at aU times. It is also well known that this hedging, or replicating, strategy is given by 6.1 = c2 (t' sl ), where c2 (t' x) = ac ;ax. It is easy to see that the hedging strategy for the digital option is simply .6.r=Be

-r(T-I)

¢

(ln(Sd K)

+ (r- !a 2)(T- t)) セ@

av T

セ@

t

1 セᄋ@ S1av T- t

( _)

43

Observe that since !':. 1 is the nwnber of shari'.s of the underlying stock the investor holds at timet, the actual amount of money invested in the stock at timet is !':. 1 · S1 • A general treatment of options that discusses pricing and hedging of various options, including the digital and the derivation of (4.2) and (4.3) can be found in such basic texts as [12] and [2L]. A valuable source for more theoretical issues is [7]. To see the connection with our problem, consider an investor who at time t has sold this digital option for the Black-Scholes price of C (t, S1 ), and suppose the investor will then invest the proceeds in such a manner as to maximize the probability that he can pay off the claim of this option at time T, i.e., for all intents and purposes, the investor's 'wealth' at time t is C(t, Sr ), and the investor will then invest this wealth so as to maximize the probability that the terminal fortune from this strategy is equal to 8. Our previous results show that the optimal policy is at timet is given by J;" of (3 .5) with x = C(t, S1) and b = B, i.e., by ft (C(t, St); B).

Chapter 34. Reaching Goals by a Deadline Reaching goals by a deadline

691

559

But placing C(t, Stl of (4.2) into (3.5) with b = Band simplifying gives

(4.4)

where 6. 1 is given by (4.3). Thus, in this case, ft* is equivalent to the hedging strategy of the digital option. Moreover, if we specialize the representation of the optimal wealth process given in Corollary 3.2, i.e. of (3.7), to the single stock case with constant coefficients, we find that

x;

(4.5)

is the wealth of an investor at timet who started off with initial wealth Xo and is investing so as to maximize the probability of his terminal time T wealth being equal to B. Observe now that by (4.1) we have

from which we may infer that

When this is placed back into (4.5), it yields the following representation of the optimal wealth process in terms of the underlying stock price

Note that there is no explicit dependence on 11 in (4.6). If the investor's initial wealth is taken to be the Black-Scholes price of the digital option at time 0, i.e., take Xo = C(O, So), where C( ·, ·) is given by (4. 2), then it is seen that (4.6) reduces to

where C(t, St) is given by (4.2), i.e., the optimal wealth process under policy f/ is just the BJack-Scholes price for a digital option with payoff B! It is interesting to note that the Black-Scholes value (4.2) and its resulting hedging strategy (4.3) are both calculated and determined by the risk-neutral probability measure (under which 11 is replaced by r), while the optimal strategy for maximizing the probability of terminal wealth being greater than B was determined under the regular measure. The analysis above can be inverted to show the following rather interesting fact:

692

S Browne S.BROWNE

560

Proposition 4.1. Consider an inwstor, with initial wealth Xo, whose objective is to maximize the probability that tenninal wealth at time T exceed some fixed level B, and who has the following two investment opportunities-the risky stock S1 of (4.1) and a bond with constant return r _ Then investing according to the dynamically optimnl policy {//, 0 _: 0. The borrowing region is thus as follows: borrowing occurs when 1.1(r) < 1.1*(r), but not if 1.1(r) > v*(r). Equivalently, since 0, 8 > 0.

(Note that this includes logarithmic utility, when 8 = 1.) This power utility function has constant relative risk aversion 1/ ,li. The optimal policy for this case, call it Ur8 , 0 :::: t :::: T), is the vector (cf. [7], [ 15]) f/ (x) = J:rr; · x, where :rr; is the optimal growth policy discussed earlier. The utility maximizing investor invests more heavily in the risky stocks, relative to the

698

S Browne

566

S. BROWNE

probability maximizing investor, when f/(x) > f/(x; b), and vice versa. It is easily seen that this occurs for values (x, r) for which 4>(v) < 8 q:, (v) -

[T

8(s)'8(s) ds

and vice versa. Thus the dynamics of this comparison reduce essentially to that described above by the borrowing region, modified by the risk aversion parameter 8.

6. Proof of Theorem 3.1 In this section, we provide the proof of Theorem 3.1. We will first show that the function V satisfies the appropriate Hamilton-Jacobi-Bellman (HJB) equations of stochastic control theory and then employ a martingale argument to verify optimality. This will prove the Theorem as well as Corollary 3 .2. We then show how we obtained the candidate value function by extending the elegant argument of [ 10] to our case.

6.1. Verification of optimality Standard arguments in control theory (see e.g. [8, Example 2, p. 161]) show that the appropriate HJB optimality equation for V is sup{.A/ V(t, x)} = 0 I

(6.1)

subject to the boundary conditions

V(t, x; b)=

サセ@

forx

セ@

bR(t, T),

t::::

T

for x = 0, t :::: T

l[b!

(6.2)

fort= T.

The generator of (2.5) shows that the HJB optimality equation (6.1) is sup{Vt f

+ (f/(JL(t)- r(t)l) + r(t)x)Vx + 1f/I:(t)ft Vu} =

0.

(6.3)

Assuming that a classical solution to (6.3), say V, exists and satisfies Vx > 0, Vu < 0 for 0 < x < b R(t, T), we may then optimize with respect to / 1 in (6.3) to obtain the maximizer / 1*(x;

Vx

l

b)= -E- (t)(tt.(t)- r(r)l)Vu

l

セM。

1 Vx (t) O(t)-. Vu

(6.4)

When (6.4) is then placed back into (6.3) and the resulting equation simplified, we find that (6.3) is equivalent to the nonlinear partial differential equation V1

+ r(t)xVx

1

yl

- !B(t) 8(t)--' = 0,

Vu

fort < T, and 0 < x < b R(t, T)

(6.5)

subject to the (discontinuous) boundary condition (6.2). Recalling now the basic facts about the normal p.d.f. and c.d.f.:

d:;y) = ¢(y);

d¢(y) セMケNL@

"'( )·

y,

(6.6)

Chapter 34. Reaching Goals by a Deadline

699

567

Reaching goals by a deadline

it can be verified that for the function V (t, x; b) of (3.1 ), we have

Vxx =

-v"[ iT

8(s)'8(s)ds

I

1

V1 =-Vr[r(t)x+

2

1

b R(t, T)

B(t)'B(t)

j frr 8(s)'8(s)ds

[ 0 and that performance shortfall level l occurs if X{ = !Y1 for some t > 0. The active portfolio management problems we consider in the sequel are: (i) Maximizing the probability perfonnance goal u is reached before shortfall l occurs; (ii) Minimizing the expected time until the performance goal u is reached; (iii) Maximizing the expected time until shortfall l is reached; (iv) Maximizing the expected discounted reward obtained upon achieving goal u; (v) Minimizing the expected discounted penalty paid upon falling to shortfall Ievell. It is clear that for all these problems, the ratio of the wealth process to the benchmark is a sufficient statistic. Among other scenarios, these objectives are relevant to institutional money managers, whose performance is typically judged by the return on their managed portfolio relative to the return of a benchmark. Since X{ is a controlled geometric Brownian motion, and Y, is another geometric Brownian motion, it follows directly that the ratio process, zi, where z{ := x{ I Y,, is also a controlled geometric Brownian motion. Specifically, a direct application of Ito's formula gives

Chapter 35. Beating a Moving Target

717 281

Beating a moving target

Proposition 1 For X{, Y1 defined by (4) and (5), let z{ be defined by z{ := X{/Y1 • Then

dz{

=

z{

(r +t;' {t) dt + z{ (/a- b') dw

1• -

z{ f3 dW/k+l),

(7)

where the constant r and vector {t are defined by

r

:= r -a+ b' b + {3 2

and {t :=

'

r1 - ab.

f-1,-

(8)

ln the next section we provide a general theorem in stochastic optimal control for the process { z{, t ::;. 0} of (7) that covers all the problems described above as special cases. In a later section we consider the more standard problem of maximizing the expected discounted terminal utility of the ratio.

3 Optimal control and a verification theorem Most of the investment problems considered in this paper are all special cases of optimal control problems of the following (Dirichlet-type) form: For the process {Z{, t ::;> 0} given by (7), let

T{

:= inf{t

> 0 : z(

= x}

(9)

denote the first hitting time to the point x under a specific policy f = {j;, t ::;> 0}. For given numbers l, u, with l < Z0 < u, let Tl := min{,[, イLセス@ denote first escape time from the interval (l, u ), under this policy f. For a given nonnegative function ..\.(z) ::;> 0, a given real bounded continuous function g(z), and a function h(z) given for z = l,z = u, with h(u) < oo, let z/(z) be the reward function under policy f, defined by

vl(z)

=

eセ@ (.laTr g(Z{) exp { +h (

zセ Q I@

exp {-

-l

..\.(Z()ds} dt

lTr ..\.(Z{)ds})

(10)

with

v(z)

=

sup vl(z), /E:· 0 and W 77 < 0) solution to the nonlinear Dirichlet problem w R 2 (z) MKVコキ Wzz(z) Mセ

R HコIKMS

1 2







w R HコIKァM^Nキ@

=

0,

for l

< z < u (14)

with w(l)=h(l),

and w(u)=h(u),

(15)

and satisfies the following three conditions:

ZセcャI@

is bounded for all z in (l, u ); either (a) there exists an integrable random variable X such that for all t :;o. 0, w(Z() :;o. X; or (b) for every t :;o. 0, and every admissible policy f, we have

(i) (ii)

(iii)

[セI@

is locally Lipschitz continuous.

Then w(z) is the optimal value function, i.e., w(z) optimal control vector, _t;;, can then he written as

=

v(z), and moreover the

(17)

Chapter 35. Beating a Moving Target Beating a moving target

719 283

As highlighted earlier, the utility of Theorem 1 for our purposes is that for various choices of the functions g(·), h(·) and .A(-), it addresses all of the objective problems highlighted earlier. Moreover, it shows that for each of these problems, all we need do is solve the ordinary differential equation (14) and then take the appropriate derivatives to determine the optimal control by (17). The conditions (i) and (iii) are quite easy to check. Condition (ii) seems like the hard one, but for all cases considered here, it can be easily checked, as demonstrated below.

Remark 3.2: Observe that the last term in (17), (0"- 1)' b, is the vector portfolio strategy that minimizes the local variance of the ratio at every point. This can be seen by observing that z{ is distributionally equivalent to the controlled diffusion process dZ( = z{ [m(f;)dt + v(!;)dW1], where m(f) = r +( {1, v 2 (f) = ' ( ( O"' O", + (3 2 , and W is a standard Brownian motion in R 1 . It

b) ((

b)

follows that v 2 has minimizer fx = (0"- 1)' b, with minimal local variance v 2 (!:) = (3 2 . Observe that m(f*) = 5, where 5 is defined earlier in (13). The scalar in the first term in (17) is the inverse of the Arrow-Pratt measure of relative risk aversion of the optimal value function. For the sequel, we will consider several specific applications of Theorem 1. In the next section, we consider the probability maximizing problem. For all relevant parameter values, this problem has a solution. This stands in contrast to some of the other problems we consider later. Specifically, we also consider the problem of minimizing the expected time to beat the benchmark by a predetermined amount. This problem has a solution if and only if a pmiicularfavorability parameter is positive. If the favorability parameter is negative, then the same policy is optimal for the objective of maximizing the expected time until ruin, where ruin is defined as falling below the benchmark by a predetermined amount.

4 Maximizing the probability of beating the benchmark In this section we consider the case where the investor is trying to maximize the probability of beating the benchmark by some predetermined percentage, before going below it by some other predetermined percentage. This objective is relevant to the case of an institutional money manager who stands to receive a bonus upon hitting a performance measure, such as beating the return ofthe S&P 500 by a certain percentage, but whose job would be lost upon underperforming the benchmark by a predetermined percentage. To formalize this problem, let V * (z) denote the maximal possible probability of beating the benchmark before being beaten by it, when staliing from the state z; i.e., let Z0 = z, and let u,l be given constants with l < z < u, then

(rC rf).

< Theorem I now applies to this problem by taking V*(z) = ウオー Q pセ@ A= g = 0, and setting h(u) = 1 and h(l) = 0. Specifically, by Theorem 1, V*(z) must be the concave increasing solution to

720

S Browne

284

S. Browne

-

1 o V 2 (z) + r5zV-(z) +- n-z 2 V.-(z) = 0 Vzz(z) " 2/J セB@ '

' Y2-1

for l < z < u

(18)

with V*(l) = 0, vx(u) = l (take..\= g = 0 in (14)). The solution to the nonlinear Dirichlet problem of ( 18), subject to the boundary conditions V(l) = 0, V(u) = l is seen to be V(z) = (zc+l - lc+l) I (uc+l - lc+l)' where c is a root to the quadratic equation q(c) = 0, where q(·) is defined by

rP

?

.

q(c) := - c +oc-r. 2

(19)

This quadratic admits two real roots, c+, c-, with c

±

-r5 ±

j

rJ2 + 2{/32

(20)

= ----'---::-;;---_..:..:._

f32

Some direct manipulations establish further that c- < 0 < c+. For both roots, we have V2 > 0, however it is only for the negative root, c-, that we have Vzz < 0. Moreover, for this function it is readily verified that conditions (i), (ii) and (iii) hold. (Condition (iia) obviously holds since V*(z) is bounded above and below by I and 0). As such the optimal value function is seen to be zc-+1 -lc-+1

V * ( z ) -- uc

-1

-lc

+I '

for l < z < u .

(21)

Furthermore, the optimal investment policy for this objective, call it J;(z ), can be now obtained by substituting this value function (21) into ( 17) of Theorem 1. We summarize this in Corollary 1 For l < z < u, let V*(z) := supr Pz ( T[
0, then J;* = 7r*, for all t > 0, minimizes the expected time to beat the benchmark, while if e < 0, then J;* = 7r*, for all t > 0, maximizes the expected time until the investor is beaten by the benchmark. Thus if the investment opportunity is favorable relative to the benchmark, then the ordinary optimal growth policy (which is independent of any benchmark parameters) also maximizes growth relative to a benchmark. However if the investment opportunity is unfavorable relative to the benchmark, then the ordinary optimal growth policy still has some optimality properties in this setting, namely, it maximizes the expected time until ruin. We state this formally in the following corollary to Theorem 1.

(I{) with optimizerf"(z) = arg inlj eセ@ ( 7{} and let G*(z) := suprE= (r!") with optimizer j.(z) = arg sup1 Ez (r{).

Corollary 2 Let G*(z) := inlj E.

!JB > 0, then G*(z)=

セQョHANIL@

e

z

with j*(z)=7r*,jorall z l.

(27)

Proof Observe first that while Theorem 1 is stated in terms of a maximization problem, it obviously contains the minimization case, as we can apply Theorem 1

(!u) },

to G(z) := sup1 { -E= and then recognize that G* = -G. As such, Theorem 1 applied with g(z) = 1, >.(z) = 0, shows that both G* and G. must solve the ordinary differential equation

however we must have G*(u) = 0 and G*(l) = 0. Moreover, G* must be convex decreasing (since it is the solution to a minimization problem) while G* must be concave increasing. It is now easy to substitute the claimed values from

Chapter 35. Beating a Moving Target

723

287

Heating a moving target

(26) and (27) to verify that that in fact is the case, under the stated conditions on e. Furthermore, for both cases we have Gz/zGzz = -1, and as such (17) of Theorem 1 shows that the optimal control for both cases is then given by E- 1fl + (0"- 1)' b, which reduces to IT*. It remains to verify whether the conditions (i) (ii) and (iii) of Theorem I hold: it is clear that (i) and (iii) hold. Condition (iib) is seen to hold for this case since in both cases we have dG*(z )I dz = -I l(zB), and dG*(z )I dz = I I (z IBI) = -I I (z B). As such requirement (16) becomes

it e-2

£ (

(j;O"'

-b)'

(;;IT'

-b)+ ,62) ds < oo,

for every t

> 0.

But this requirement must hold by definition of admissibility.

Remark 5.1: Observe that

IT*

D

maximizes logarithmic utility for the investor. This

zf)] }

(xi)] }-

follows directly since sup1 { E [ln ( = sup1 { E [ln E [ln (Yr )]. While the connection between log-utility and the minimal expected time to beat the benchmark problem (or the maximal expected time to being beaten by it) is now obvious in light of the logarithmic value functions of (26) and (27), this was by no means obvious or intuitive apriori. To interpret the favorability condition properly, observe that if Z* denotes the ratio process under policy IT*, then Z,* is the geometric Brownian motion given by Z1* = Zo exp { Bt + (0"- 1 (p- rl))' W1• - jJW,(k+l)}

tE

[In (zt IZo)] =e. Thus, standard results on Brownian motion and so we have imply that for 8 > 0, Zt" --+ oo, a.s., while for 8 < 0, Z,* --+ 0, a.s. This in tum implies that < 00 for > 0, while 7/ < 00 for < 0, where denotes the first hitting time to x for the process Z*. Moreover, for any admissible policy {[,}, and for l < z < u, we have the following:

r:

Corollary 3 1(8

e

e

< 0, then infr Ez ( r{)

= oc, while ifB

r;

> 0, then sup1 Ez ( rf)

=

00.

e

This corollary states that for the unfavorable case, < 0, there is no admissible policy for which the expected time to beat the benchmark is finite, while for the favorable case, an admissible strategy can always be found whose expected time to ruin is infinite.

e,

Remark 5.2: The favorability parameter, helps in interpreting the portfolio policy fV of (22) obtained earlier for the probability maximizing objective. Specifically, some direct manipulations will verify that for c- of (20) being the negative root to the quadratic of (19), the following holds:

< -1 c- { > -1 = -1

if e > o if < o ifB=O.

e

(29)

724

S Browne

288

S. Brovme

Consider now the optimal policy for maximizing the probability of beating the benchmark, before being beaten by it,.f; of (22). Observe that (29) is equivalent to saying that > 1f* for e < 0, while < 1f* for e > 0 (vector inequalities here are to be interpreted componentwise). Thus in an unfavorable situation, the strategy that minimizes the probability of ruin is more aggressive than the strategy that maximizes the expected time to ruin. So in a sense, and in the picturesque language of Dubins and Savage (1975), we observe that for an unfavorable game, bold play maximizes the probability of success while timid play maximizes the expected playing time. Conversely, for a relatively favorable game, timid play will minimize the probability of ruin while a bolder strategy will minimize the expected time to achieve a goal.

fv

fv

6 Minimizing or maximizing expected discounted rewards In this section we consider the problem of either maximizing Eo minimizing E: ( e-A-1,), for a given discount rate ..\

(e-AT(,), or

> 0. The former objective is

relevant when there is some (positive) reward of achieving the upper goal, while the latter is relevant when there is some cost to hitting the lower goal. Thus the fonner is related to growth while the latter is related to survival. To address this formally, let P*(z)

=

sup1 E2 (e-AT;), with associated op-

timal control vector f*(z; ..\), and let F*(z)

=

infj Eo

(e-AT{), with associated

optimal control vector f*(z; ..\). From Theorem 1 (with g = 0) we see that for both cases, the value function must satisfY the nonlinear problem

セM

F_2(z) I 2 2 + i5zF-(z) + -8 z F--(z)- ..\F(z) 'Fzz(z) " 2' Mセ@

=

0

'

(30)

however the concavity/convexity properties and the boundary conditions ditler. Specifically, we require F* to be concave increasing with F*(u) = 1, while we require F* to be convex decreasing with F*(l) = 1. Solutions to (30) are of the form F(z) = Kzry+l, where K is a constant to be determined by the boundary condition, and 7) is a root to the cubic equation C(7); ..\) = 0, where C(rJ;A) := rh3 2 /2 + rl

[!P /2 + i5]

+r]

[5 Mセ

..\] -1.

(31)

For ..\ > 0, this cubic equation admits three distinct and real roots, call them 7]2, T)J, with

T}J,

TJI(..\) < -1, -1 < 7)2(..\) < 0,

7]3(..\) > 0.

By Theorem 1, for the maximization problem we require L* > 0, with < 0, which translates into the simultaneous requirements that 7) + 1 > 0, and 7)(7) + I) < 0. Thus, 7)2 (..\) is the appropriate root, and we then have

f P セ@

Chapter 35. Beating a Moving Target

725

289

Beating a moving target

F * (z)

=

(z )'72+! -

u

,

for z ::.; u.

It then follows from (22) that for this case we have the optimal control vector

Similarly, for the minimization problem, we require F* to be convex decreasing, which translates into the simultaneous requirements that TJ + 1 < 0 and rJ(rJ+ 1) > 0. Clearly, it is now r7 1(.\) which is the appropriate root here. As such, we have F*(z)

=

(y)

r11+1 '

for z :=, l,

with the associated optimal control vector

Observe that these optimal controls are again constant proportions, and for jl > 0 we have[* > f*. Observe further that we can write the cubic function C(T); .\) of (31) as C(T); .\) = (TJ +I) [q(rJ)- .A] +.\,where qC) is the quadratic defined earlier by (I 9). Thus as .\ -1- 0, C (T); 0) = ( TJ + I )q( TJ), which implies that as .\ 0 the three roots to the cubic equation must converge to the two quadratic roots c+, c- of (20), and the new root, -1. Since c- < 0 < c+, it must be that r73 (.\) converges to c+, however, it is by no means clear to which values the other 2 converge. In fact we will now show below that this depends on the sign of the favorabi1ity parameter, e. Recall first some facts about Laplace transforms of continuous nonnegative random variables: If H(.\) = E (e-.\ 7 ) , where 7 is a nonnegative valued random variable and .\ :=, 0, then H (0) := lim.\-1-0 H (.\) = !'( 7 < oo ). Thus if 7 is a defective random variable, then the defect is 1 - H(O). It is of course necessary that H(O) = 1 for £(7) < oo. Moreover, if £(7) < oo, then

+

() _ 1. 1-H(.\) £7-tm .\ . .\_j,.O

Consider now the negative root c-. As observed previously in (29), we have c- < (::,)- 1 if B > (S)O. This fact, combined with Corollary I, now gives the following relationship.

Proposition 2 As .\

+0, we have the following:

I. If 8 > 0, then T/2(.\) .J- T/2(0) = -1 , and T/1 (0) =c-. !!. If B < 0, then 7)1 (.\) t 7)1 (0) = -1, and 1)2(0) =c-.

726

S Browne

290

S. Brovme

7 Utility maximization So far, we have considered objectives related solely to the achievement of (arbitrary) goals. In this section we consider the case where the investor is interested in maximizing expected discounted terminal utility of the ratio at some fixed terminal timeT < oo. Specifically, for a given (concave increasing) utility function U (z ), let l/f(t, z) = sup1 E ( e->-(T -t) U (

4)).

For the special case U (z) = ln(z ), our previous results show that the ordinary log-optimal policy, 7r* is optimal once again, with l/f(t, z) = ln(z )+ (61- .\)(T- t). For the general case, an analysis similar to that of Theorem I will show that (under suitable regularity conditions), the optimal value function l/f must satisfY the nonlinear partial differential equation

subject to the boundary condition l/f(T,z) = U(z). (For the special case f3 = 0 this reduces to a problem treated in Merton (1971 ).) The optimal control function is then given by (33) For the special case where the utility function is of the form U (z) the explicit solution to the nonlinear Cauchy problem (32) is l/f(t, z)

=

z"+ 1 exp { Mセc@

(K; .\)(T-

t)}

=

z "+I,

(34)

where C(e;.\) is the cubic function defined earlier in (31). This utility function has constant relative risk aversion -K (for concavity, we would require that -1 < K < 0). Substitution of (34) into (33) shows that for this case the optimal control is again a constant proportional strategy, with

Comparison with earlier results will then show that just as there is a connection between maximizing logarithmic utility and the objective criteria of minimizing or maximizing the expected time to a goal, so too is there a connection between maximizing expected utility of terminal wealth for a power utility function, and the objective criteria of maximizing the probability of reaching a goal, or maximizing or minimizing the expected discounted reward of reaching certain goals. (In particular, by taking K = c-, T/1, J72.) This extends earlier connections obtained for the standard portfolio problem since it shows that power utility (Constant Relative Risk Aversion) functions relate to survival as well as growth objectives.

Chapter 35. Beating a Moving Target

727

Beating a moving target

291

8 Conclusions We have studied a variety of objective goal-related problems for the problem of outperforming a stochastic benchmark. The presence of a benchmark that need not be perfectly correlated with the investment opportunities allows a richer risk/reward framework than the standard model. We have identified a particular parameter whose sign determines whether we are in the favorable case or not. Regardless ofthe sign of this favorability parameter, we have determined the optimal policy for maximizing the probability of reaching a given performance goal before falling to a performance shortfall. For the favorable case, we have shown that the ordinary log-optimal policy is also the optimal policy for minimizing the expected time to the performance goal, while for the unfavorable case, the logoptimal policy maximizes the expected time to the performance shortfall. The discounted version of these problems has been solved as well. We have related all these goal problems to a more standard expected utility maximizing problem for power utility functions. Our results provide extended optimality properties of constant proportion policies.

A Appendix A.! Proof a_[ Theorem I

Observe first that for Markov control processes !Jf(t, z) E セ@ 1' 2 , the generator of the ratio process

{/t, t 2: 0}, and functions

z/

of (7) is

The Hamilton-Jacobi-Bellman optimality equation of dynamic programming (see Krylov (1980, Theorem 1.4.5) , or Fleming and Soner (1993, Sect. IV.5)) for maximizing z/(z) of(10) over control policiesj;, to be solved for a function v is supr{./t/v + g - >..v} = 0, subject to the Dirichlet boundary conditions v(l) = h(l) and v(u) = h(u ). Since 1/(z) is independent of time, the generator of (35) shows that this is equivalent to

mlx {

(r +/

[1) z

1/z

+ セ@ (

(f c/ - b)' (f cr' - b) + ,8

2)

z 2 1/zz + g -

AI/} 0. =

(36) Assuming now that (36) admits a classical solution with vx > 0 and Vxx < 0, we may then use standard calculus to optimize with respect to fin (36) to obtain the optimal control fi.mctionJ;;(x) of (17), with 1/ = w. When (17) is then substituted back into (36) and simplified, we obtain the nonlinear Dirichlet problem of (14) (with v = w). It remains only to verify that the policy 1,; is indeed optimal. Since the conditions required in the aforementioned results in Krylov (1980) and Fleming and Soner (1993) are not all necessarily met in this case, we will use the martingale

728

S Browne

292

S. Browne

optimality principle directly (cfRogers and Williams 1987, or Davis and Norman 1990), which entails finding an appropriate functional which is a uniformly integrable martingale under the (candidate) optimal policy, but a supermartingale under any other admissible policy, with respect to the filtration ·•Yt· To that end, let Af(s, t) := )..(Z[}dv, and define the process

J:

M(t, Z{) := e-"11 (0,t)w(Z{) +

t e-Jl'(O,s)g(Z{)ds,

for t :,;> 0,

(37)

./o

where w is the concave increasing solution to (14). Optimality off,; of(17) is then a direct consequence ofthe following lemma.

Lemma 1 For any admissible policy f, and M(t, ·)as defined in (37), we have

E (M

(t 1\ Tl,z(M

1 )):::;

M(O,Z0) = ·w(z),

for t :,;> 0

(38)

with equality holding if and only iff = J:, where J;; is the policy given in (1 7). Moreover, under policy J;;, the process {M (t 1\ Tl, Z/Mt) , t :.:> 0} is a uniformly integrable martingale. Proof Applying Ito's formula to M(t,Z{) of (37) using (7) shows that for O:::;s:::=;t:::=;Tf M(t,Z{)

=

M(s,Z()+

+

l

l

e-,V(s,v)Qlfv;Z[)dv

e-,l'(,,vlz(wz(Z() [ Hヲセ。M

b') dw,;-

LV、wセォKャIj@

(39)

where Qlf; z) denotes the function defined by

Qlf;z)

:=

(D ((fa'

-b)'

(fa' -b)+,6 )z wzz(z) 2

2

+z(r +/ {L)w 2 (z) + g(z)- )..(z)w(z).

(40)

Recognize now that for a fixed z, Q is a quadratic form in the vector f with Q1rlf; z) = Ewzz (z ). Since E is symmetric positive definite and w is concave in z, it is easily verified that we always have Qlf; z) :::; 0, and that the maximum is achieved at the value

with corresponding maximal value

Q lf * ;z )

= MセI@

w;(z) + i)zwz _ (z ) + -, 1 6 ?-z 2 W22 (Z ) + g(z)- A(Z , ) w(z ) Wzz(Z

2

=0

(41)

where the final equality follows from (14). Therefore the second term in the rhs of (39) is always less than or equal to 0. Moreover (39) shows that we have

Chapter 35. Beating a Moving Target

729

293

Beating a moving target

M (t A rl,z{IITr)- w(z)

-1

I liT(

e-;'/(O,v)Qlfv;Z{)dv (42)

> M(tllr 1,z{IIT1 ) -w(z),

(43)

Thus, by (ii) we see that the stochastic integral term in (39) is a local martingale that is in fact a supermartingale. (If (iia) holds, then the RHS of (43) is bounded below by an integrable random variable, which is sufficient to make the stochastic integral term a supermartingale. If (iib) holds, then the stochastic integral is a martingale directly, and hence it is also a supermartingale.) Hence, taking expectations on (39) therefore shows that

with the equality in (46) being achieved at the policy J,;. Thus we have established (38). Note that under the policy .fv*, the process Z* satisfies the stochastic differential equation

dZ;" (47)

where r* := rt:. By (iii) this equation admits a unique strong solution. Furthermore note that under the (optimal) policy, J,;, we have

M(t,Z/) =M(s,Z,*)

x

lH[ZセゥI@

-l' {-lv >.cz;)dp} exp

(cr- 111)' 、w[KjコキHzIセォMャ@

(48)

which by (i) above is seen to be a uniformly integrable martingale. This completes is indeed optimal. D the proof of the theorem and hence

J,;

730

S Browne 294

S. Browne

References 1. Black, F., Perold, A.F.: Theory of constant proportion portfolio insurance. J. Econ. Dynamics Control 16, 403-426 (1992) 2. Browne, S.: Optimal Investment Policies for a Firm with a Random Risk Process: Exponential Utility and Minimizing the Probability of Ruin. Math. Operat. Res. 20 937-958 (1995) 3. Browne, S. : Reaching Goals by a Deadline: Digital options and Continuous-Time Active Portfolio Management. Adv. Appl. Probab. (1996, forthcoming) 4. Browne, S.: Survival and Growth with a Fixed Liability: Optimal Portfolios in Continuous Time. Math. Opera!. Res. 22, 468-493 (1997) 5. Browne, S.: The Return on Investment from Proportional Portfolio Strategies. Adv. Appl. Probab. 30 (I), 216- 238 (1998) 6. Carpenter, .1 . N .: The Optimal Dynamic Investment Policy for a Fund Manager Compensated with an Incentive Fee. Preprint, New York University (1996) 7. Davis, M.H.A., Norman, A.. Portfolio Selection with Transactions Costs. Math. Operat. Res. 15, 676--713 (1990) 8. Dubins, L.E., Savage, L.J.: How to Gan1ble If You Must: Inequalities for Stochastic Processes. New York: Dover 1975 9. Duffie, D.: Dynamic Asset Pricing Theory, second edition. Princeton: Princeton University Press 1996 10. Dutta, P.: Bankruptcy and expected utility maximization. J. Econ. Dynamics ControllS, 539- 560 (1994) 11. Fleming, W.H ., Soner, H.M.: Controlled Markov Processes and Viscosity Solutions. Berlin Heidelberg New York: Springer 1993 12. lleath, D., Orey, S., Pestien, V., Sudderth, W.: Minimizing or Maximizing the Expected Time to Reach Zero. SIAM J. Control Optimizat. 25 (I), 195-205 ( 1987) 13. Krylov, N.Y.: Controlled Diffusion Processes. Berlin Heidelberg New York: Springer 1980 14. Majumdar, M., Radner, R .: Linear Models of Economic Survival under Production Uncertainty. Econ. Theory 1, 13-30 (1991) 15. Merton, R.: Optimum Consumption and Portfolio Rules in a Continuous Time Model. J. Econ. Theory 3, 373-413 (1971) 16. Merton, R.: Continuous Time Finance. Oxford: Blackwell 1990 17. Orey, S., Pestien, V., Sudderth, W.: Reaching Zero Rapidly. SIAM J. Control Optimizat. 25 (5), 1253-1265 (1987) 18. Perold, A.F., Sharpe, W.F.: Dynamic Strategies for Asset Allocation. Financial Analyst .J., Jan!Ft:b, 16- 27 (1988) 19. Pestien, V.C., Sudderth, W.D.: Continuous-Time Red and Black: How to control a di iTusion to a goal. Math. Opera!. Res. 10 (4), 599- 6 11 ( 1985) 20. Rogers, L.C.G., Willian1s, D. : Diffusions, Markov Processes, and Martingales, Vol. 2. New York: Wiley 1987 21. Sharpe, W.F., Alexander, G.F., Bailey, J.V.: Investments, 5th edn. Englewood Cliffs: Prentice Hall 1995

Chapter 36

1. Appl. Prob. 37, 126-147 1200()) Printed in Israel ©Applied Probability Trust 2000

STOCHASTIC DIFFERENTIAL PORTFOLIO GAMES SID BROWNE,* Columbia University and Goldman Sachs & Co.

Abstract We study stochastic dynamic investment games in continuous time between two investors (players) who have available two different, but possibly correlated, investment opportunities. There is a single payoff function which depends on both investors' wealth processes. One player chooses a dynamic portfolio strategy in order to maximize this expected payoff, while his opponent is simultaneously choosing a dynamic portfolio strategy so as to minimize the same quantity. This leads to a stochastic differential game with controlled drift and variance. For the most part, we consider games with payoffs that depend on the achievement of relative performance goals and/or shortfalls. We provide conditions under which a game with a general payoff function has an achievable value, and give an explicit representation for the value and resulting equilibrium portfolio strategies in that case. It is shown that non-perfect correlation is required to rule out trivial solutions. We then use this general result explicitly to solve a variety of specific games. For example, we solve a probability maximizing game, where each investor is trying to maximize the probability of beating the other's return by a given predetermined percentage. We also consider objectives related to the minimization or maximization of the expected time until one investor's return beats the other investor's return by a given percentage. Our results allow a new interpretation of the market price of risk in a Black-Scholes world. Games with discounting are also discussed, as are games of fixed duration related to utility maximization. Keywords: Stochastic differential games; portfolio theory; stochastic control; diffusions: martingales

AMS 1991 Subject Classification: Primary 93E05; 90A09 Secondary 93E20; 60G40; 60160

1. Introduction

This paper treats various versions of stochastic differential games as played between two 'small' investors, call them A and B. (The investors are called small in that their portfolio trading strategies do not affect the market prices of the underlying assets.) The games considered here are zero-sum, in that there is a single payoff function, with one investor trying to maximize this expected payoff, while simultaneously the other investor is trying to minimize the same quantity. There are two correlated risky investment opportunities, only one of which is available to each investor. The players compete by the choice of their individual dynamic portfolio trading strategy in the risky asset available to them and a risk-free asset that is freely available to both. There is complete revelation, or observation, in that A's strategy is instantaneously observed by B (without error) and vice versa. For the most part, the games we consider have discontinuous payoffs where Investor A wins if his fortune ever exceeds Investor B 's fortune by some predetermined amount, and similarly. Received 8 March 1999; revision received 28 June 1999. *Postal address: 402 Uris Hall, Graduate School of Business. Columbia University, New York, NY 10027. CSA. Email address: [email protected]

126

731

732

S Browne Stochastic differential portfolio games

127

Investor B wins the game if his fortune ever exceeds Investor P:s fortune by some (possibly other) predetermined amount. As we show later, we require non-perfect correlation between the investment opportunities so as to rule out trivial solutions to our games. Specifically, if the investment opportunities available to A and B are the same, then in any of our continuous-time stochastic differential games with perfect revelation, any move by Investor A can be immediately reacted to, and perfectly adjusted for, by Investor B, thus heading off any movement in the state variable. Thus, in our setting, the only interesting games are those where there is non-perfect correlation between the investment opportunities, allowing non-perfect adjustment and reaction between the players. Aside from the intrinsic probabilistic and game-theoretic interest, such a model is applicable in many economic settings. For example, our results have significant bearing on what is sometimes referred to as active portfolio management, where the objective of an individual investor is to beat the performance of a preselected benchmark portfolio (see e.g. Browne (1999)). While the chosen benchmark is most often a wealth process obtained from a known deterministic portfolio strategy (e.g. an index, such as the S&P 500), ourresults would provide a worst case and minimax analysis for how the benchmark would perform in a game-theoretic setting. These results could then be used in turn, for example, to set conservative capital requirements for a given preassigned maximally acceptable probability of underperformance relative to that benchmark. Another, perhaps more direct, example occurs in many trading firms, where each individual stock, or sector of stocks, is assigned to its own individual trader. Our model is then applicable to an analysis of the performance of these traders when a component of their compensation is determined by the achievement of relative goals, for example a bonus for the 'best' performer (the winner of the game), and/or a penalty, such as termination, for the worst performer (the loser). Similarly, our results, are of interest in a partial analysis of the competition played out between two fund managers, whose funds are invested in different markets and have different characteristics, who achieve rewards based on the relative performance of their funds. Finally, we also note that our results also allow new interpretations of the market price of risk of an asset in a Black-Scholes world, in that we show that the degree of advantage a player has over the other is determined solely by the market price of risk of his investment opportunity. An outline of the remainder of the paper, as well as a summary of our main results, are as follows. In the next section, we describe the formal model under consideration here. There are two correlated stocks as well as a risk-free asset called a bond. Each investor can invest freely in the risk-free asset but is allowed to invest in only one of the stocks, according to any admissible dynamic portfolio strategy. His opponent can also invest freely in the risk-free asset, but only in the other stock according to any admissible dynamic portfolio strategy. We then describe how the investors compete. The relevant state variable is the ratio of the two investors, and the game terminates when this ratio first exits an interval. In Section 3, we provide a general result in optimal control for a stochastic differential game with a general payoff function, in the context of our model. Specifically, we characterize conditions under which the value of this game will be the smooth solution to a particular nonlinear Dirichlet problem. The equilibrium, or competitively optimal, controls are then given by an explicit expression involving the derivatives of this value function. We then solve these Dirichlet problems explicitly for various specific examples in subsequent sections. The proof of Theorem 3.1 is presented in the final section of the paper. In Section 4 we consider the probability maximizing game, where Investor A is trying to maximize the probability of outperforming Investor B by a given percentage, before Investor B outperforms him by another given percentage. It turns out that a value for this game exists if

733

Chapter 36. Stochastic Differential Portfolio Games

!28

S. BROWNE

and only if a specific measure of advantage parameter, which is defined here as the ratio of the market price of risk for A's investment opportunity over the market price of risk forB's investment opportunity, takes values in a particular interval. This interval is determined solely by the instantaneous correlation between the investment opportunities. If this condition is met, then we give explicit solutions for the equilibrium portfolio strategies. Among other results, we show that the disadvantaged player has a relatively bolder strategy than the player who holds the advantage, as would be expected from the classical results of Dubins and Savage (1965) for single-player probability maximizing games. For the symmetric case, where no player holds the advantage, the equilibrium strategies reduce to the growth-optimal strategy. In Section 5 we consider games where the objective is to minimize the expected time to outperform the other player. There are two cases to consider, depending on which player has the advantage. In the symmetric case, the games do not have a finite value. In the non-symmetric case, the equilibrium portfolio strategies are the individual growth-optimal strategies, and a new connection is made with maximizing logarithmic utility. In Section 6 we consider games with discounting, where the objective of one player is to maximize the discounted reward achieved upon outperforming his opponent. For this game to have a value, we require a greater degree of advantage to exist than was required for the probability maximizing game. In Section 7 we consider fixed-duration utility-based games, where both investors obtain utility (or disutility) solely on the basis of their relative wealth, i.e. in terms of their ratio. The value for such games is then given (under appropriate conditions) as the solution to a particular non-linear Cauchy problem, and the saddlepoints, or competitively optimal control functions. are obtained in terms of the derivatives of this value function. An explicit solution is given for the case of power utility.

2. The portfolio model with competition The model under consideration here consists of three underlying processes: two correlated risky investment opportunities (e.g. stocks, or mutual funds) sOl and S(2), and a riskless asset B called a bond. The price processes for these assets will be denoted, respectively, by (sJ'>, s?J, B 1 , t 2: 0). While we allow both investors to invest freely in the risk-free asset. Investor A may trade only in the first stock, S(l l, and similarly, Investor B may trade only in the second stock, S(2). While there are only two correlated risky assets in our model. it is without any loss of generality since it is just a simple matter of algebra to generalize our results and analysis to a constant coefficients complete market model (see Duffie (1996)) with n risky stocks driven by n Brownian motions, for any arbitrary n > 2. In that case, we would split the n stocks into two groups, say with the first k stocks available to Investor A and the remaining 11- k stocks available to Investor B, with A being restricted from trading in B's group and vice versa for B. However, for notational and expositional ease, we consider just the (essentially equivalent) two-asset case. The probabilistic setting is as follows: we are given a filtered probability space (Q, :F, (J'i), P),

supportingtwocorrelatedBrownianmotions, wnl, wm, withE(wplw?l) =pt. (Specifically, J'i is the P-augmentation of the natural filtration :F;w := a{W[ 1l, w_?l: 0 セ@ s セ@ t}.) We will assume that the price process for each of the risky stocks follows a geometric Brownian motion, i.e. s:il satisfies the stochastic differential equation

734

S Browne 129

Stochastic differential portfolio games

(2.1) where Mi, i = 1, 2 are positive constants. The price of the risk-free asset is assumed to evolve according to dBr

= rBr dt

(2.2)

where r 2: 0. To avoid triviality, we assume Mi > r, fori = 1, 2. For the sequel, let the parameter e; denote the risk-adjusted excess return of stock S(il over the risk-free rate of return, fori = 1, 2. Specifically, Mi - r

e;=--, a;

for

i = 1, 2.

(2.3)

The parameter e; is also called the market price of risk for stock i, for i = 1, 2. Let f 1 denote the proportion of Investor P:s wealth invested in the risky stock sOl at time t under an investment policy f = {.fi, t 2: 0}, and, similarly, let g 1 denote the proportion of Investor B's wealth invested in the risky stock s< 2 l at time t under an investment policy g = {g1 , t 2: 0}. We assume that both {!1 , t 2: 0} and {g1 , t 2: 0} are suitable, admissible J=i-adapted control processes, i.e . .fi (or g 1 ) is a non-anticipative function that satisfies E J[ J? dt < oo (orE J[ dt < oo) for every T < oo. We place no other restrictions on for g, for example, we allow .fi (or g1) 2: 1, whereby the investor is leveraged and has borrowed to purchase the stock. (We also allow .fi (or g1 ) < 0, whereby the investor is selling the stock short; however, for Mi > r, for i = 1, 2, this never happens in any of the problems considered here.) For the sequel, we will let g, denote the set of admissible controls. Let x( denote the wealth of investor A at timet, if he follows policy f = {.fi, t 2: 0}, with Xo = x. Since any amount not invested in the risky stock is held in the bond, this process then evolves as

gl

(2.4)

upon substituting from (2.1) and (2.2) and using the definition (2.3). This is the wealth equation denote the wealth of investor B under first studied by Merton (1971). Similarly, if we let portfolio policy g = {g1 , t 2: 0}, then Yl evolves according to

Y/

(2.5)

where W/ 2) is another (standard) Brownian motion. To allow for complete generality, we allow w?l to be correlated with w?l, with correlation coefficient p, i.e. E(W/1) w?l) =pt.

735

Chapter 36. Stochastic Differential Portfolio Games S. BROWNE

130

2.1. Competition While there are many possible competitive objectives, here we are mainly interested in games with payoffs related to the achievement of relative performance goals and shortfalls. Specifically, for numbers /, u with !Yo < Xo < uYo, we say, in terms of objectives for Investor A, that (upper) performance goal u is reached if x{ = uY1g, for some t > 0 and that (lower) performance shortfall level/ occurs if x{ = lY/ for some t > 0. In general A wins if performance goal u is reached before performance shortfall level l is reached, while B wins if the converse happens. (Analogous objectives can obviously be stated in terms of Investor B with goal and shortfall reversed.) Some of the specific games we consider in the sequel, stated here from the point of view of Investor A, are: (i) maximizing the probability that performance goal u is reached before shortfall/ occurs (equivalently, maximizing the probability that A wins); (ii) minimizing the expected time until the performance goal u is reached; (iii) maximizing the expected time until shortfall/ is reached; (iv) maximizing the expected discounted reward obtained upon achieving goal u; (v) minimizing the expected discounted penalty paid upon falling to shortfall level/. In each case, Investor B 's objective is the converse. For all these games, the ratio of the two wealth processes is a sufficient statistic. In a later section, we also consider a fixed-duration utility-based version of the game where the ratio is also the pertinent state variable. Since X { is a diffusion process controlled by Investor A, and Y1g is another diffusion process controlled by Investor B, the ratio process, z!.g, where z{·R := x{ I Y/, is ajointlycontrolled diffusion process. Specifically, a direct application of Ito's formula gives Proposition 2.1. For the wealth processes X{, Yl defined hy (2.4) and (2.5), let defined by z{ :=X{ /Y/, Then

z{r:

be

(2.6)

where the function m(f, g) is defined by m(f, g)= m(f, g: a], az, e], &z, p)

=

falel- gaz&z

+ g 2ai- pawzfg

(2.7)

and where the parameters ei' i = 1' 2 are defined in (2.3 ). Alternatively, in integral form we have

z('g = Zoexp{{[m(fs. g,)-

セカ R HェL@

gs)]ds

+

L

!sal dWS(I)-

L

&a2dW,(2)} (2.8)

where the function v2 (f, g) is defined by v 2 (f, g)= v 2(f, g: a1, a2, p) =

J 2a 12 + g 2a}- 2fgawzp.

(2.9)

A consequence of this is that for Markovian control processes fr = f(Z{ 8 ) and g, = g(Z{ 8 ) (also referred to as pure strategies, see e.g. Friedman (1976)), the ratio process zf.g of (2.6) is a controlled Markov process whose generator, for arbitrary functions cp(t. z) E C? 1· 2 • is given by (2.10)

In the next section we provide a general theorem in stochastic optimal control for differential games associated with the process {Z/' 8 , t :::: 0} of(2.6) that covers all the games described

736

S Browne 131

Stochastic differential portfolio games

above as special cases. In a later section we consider the problem of maximizing the expected discounted utility of the ratio. More general results on stochastic differential games where the diffusion component of the process, as well as the drift, is controllable by both players are discussed in e.g. Fleming and Souganides (1989).

3. Value and equilibrium in a stochastic differential game For the process zi;g of (2.6), let

r{ 8

:= inf{t > 0:

z{ 8 =

x}

(3.1)

denote the first hitting time to the point x under the specific policies f {.fi, t 2:: 0} and g = {g 1 , t 2:: 0}. For given numbers l, u, with l < Zo < u, let rf,g := min{r/ 8 , r{ 8 } denote the first escape time from the interval (l, u), under the policies f, g. For a given non-negative function )..(z) 2:: 0, a given real bounded continuous function c(z), and a function h (z) given for z = l, z = u, with h (u) < oo, let vf,g (z) be the expected payoff function under the policy pair f, g, defined by vf,g (z)

= E2 (fo'f.gc(Z{ 8 ) exp

{-for

A(Z/' 8 ) ds} dt

+ ィHzZAセI@

exp{- fo,f.g),(Z{ 8 ) ds}). (3.2)

(Here and in the sequel, we use the notations PzO and E2 (-) as shorthand for P(- I Zo = z) and E(· I Zo = z).) The two investors compete in the following form: Investor A would like to choose a control function f in order to maximize vf,g (z), while simultaneously Investor B is trying to choose a control function g in order to minimize vf.g (z). We consider here only games with perfect revelation, or perfect observation, so that the players' choices are instantaneously revealed to their opponents. The game, or competition, terminates when the ratio process zf.g first exits the interval (l, u). Let .!!_(Z) = sup inf vf,g (z) and IJ(z) = inf sup vf,g (z) fE[j gEfj

gEfj fEfj

denote the lower and upper values of the game, respectively. If !'_(Z) = IJ(z) for every z, then the value of the game is given by v(z) := !'_(Z) = IJ(z). This value can be attained if a Nash equilibrium, equivalently a saddlepoint for the payoff vf,g (z), exists, i.e. if there exist two strategies, f* = {.!;*, t 2:: 0} and g* = {g:, t 2:: 0} such that for all z E (l, u), and all other admissible f and g vf,g* (z) ::'0 vf*,g* (z) ::'0 vf*,g (z).

(3.3)

If (3.3) holds, then v(z) = vf*,g* (z) (see for example Elliott (1976); Fleming and Souganides (1989); Maitra and Sudderth (1996)). The saddlepoint strategies f*, g* are referred to as the equilibrium, or competitively optimal, strategies. In the following theorem, we provide an explicit evaluation of the value of the game as the appropriate solution to a particular non-linear Dirichlet problem, as well as an evaluation of the competitively optimal strategies!,* and under suitable conditions. To enable the reader to proceed directly to the specific examples and applications in the subsequent sections, the proof of this theorem is presented in the final section of the paper. To state results more compactly, let us first introduce some notation and definitions:

g;,

737

Chapter 36. Stochastic Differential Portfolio Games 132

S. BROWNE

I. For an arbitrary function

1/J (z)

E

e 2 , let r denote the differential operator defined by (3.4)

2. For the sequel, we will say that an increasing strictly concave function 1/J(z) E e 2 (so 1/lzz < 0) is sufficiently fast-increasing on an interval (a, b) if the following condition holds: 21/lz(Z)

+ Zl/lcz (z)

> 0,

for all a < z < b.

(3.5)

(Observe that the fast-increasing condition (3.5) is equivalent to requiring that the ArrowPratt measure of relative risk-aversion for 1/J, defined as -zl/lzz/1/Jz, is less than 2.) 3. For the sequel, the parameter K will denote the ratio of the market prices of risk for the two risky assets. Specifically, for 0; as defined in (2.3) fori = I, 2, define the parameter K by (3.6)

We will see later that the parameter K is a measure of the degree of advantage one player has over the other. Investor A is said to have the advantage if K > I and Investor B has the advantage if K < I. In the symmetric case the two are neutral. Theorem 3.1. Suppose that Ill (z) : (l, u) r-+ Ht is a C'! 2 strictly concave, sufficiently fastincreasing (as in (3.5)) solution to the non-linear Dirichlet problem for l :::; z :::; u: 2 2 zlliz(z) 2 2 2 riJI(z) 02 [(I- K )IJI 2 (z)- (I+ K

-



2pK)(IIi 2 (z) + ziJ!cc(z))] + c(z)- A(Z)IIi(z)

=0 (3.7)

with lli(l)=h(l)

and

IJ!(u)=h(u).

(3.8)

Also suppose that for all admissible policies f and g, E2(rf·K) < oo for l < ; < u. and that Ill (z) satisfies the following conditions: (i) for all admissible policies f and g. and.for all t ::: 0, the .following moment condition

holds (3.9)

(ii) thefunctionziJ! 2 (z)H(z) is bounded on (l, u), where

(iii) the function zH(z)[!

+ H(z)]

is Lipschitz continuous on (l, u).

738

S Browne

Stochastic differential porifolio games

133

=

Then lll(z) is the value of the game described earlier; i.e. lll(z) = v(z) vf*.g* (z), and moreover this value is achieved at the saddlepoint control functions, or competitively-optimal portfolio strategies, fv* (z) and ァセ@ (z), given by (3.10)

(3.11)

Remark 3.1. The technical conditions (i), (ii) and (iii) above play a role at various points in the proof of Theorem 3.1, which is presented in the final section. Specifically, if condition (i) holds, then a stochastic integral term, which in general is only a continuous local martingale, is a martingale. If condition (ii) holds, then a particular martingale is uniformly integrable (and hence the martingale stopping theorem is valid), and condition (iii) ensures that the stochastic differential equation for the competitively optimal ratio process Z*·*, obtained when the controls of (3.10) and (3.11) are placed back into (2.6), admits a strong solution. (If (iii) and zm(fv*, ァセI@ are all Lipschitz continuous.) holds, then zfv*(z), コァセHI@ Remark 3.2. The parameters 8;/ai, i = 1, 2, in the optimal control functions, fv*(z) and ァセHコI@ of (3.10) and (3.11), are the individual growth-optimal portfolio strategies for the respective investors. Specifically, Investor A will choose f (z) = e, I GJ for all z. if he is interested in any or all of the following individual objectives, without any regard to actions by Investor B: maximizing logarithmic utility of wealth at a fixed terminal time; minimizing the expected time to reach any arbitrary fixed (obviously higher than initial) level of wealth; maximizing the growth rate of wealth, defined by sup 1 {liminfH 00 log(X()/t}. (See e.g. Merton (1990), Chapter 6 and Browne (1998) for reviews and further optimality results.) The obvious analogous results hold for Investor B. 3.1. The symmetric case

a,

= a2 =a, and so the market prices of In the (fully) symmetric case, /LJ = fL2 = fL and risk for the two stocks are the same, i.e. e, = セ@ say, and of course K = 1. For this case we observe that the determining ODE for the optimal value function of (3.7) reduces to

=()

zll1 2 (z) 2

-8 flll(z) (1- p)[lllz(Z)

+ zlllzz(z)] + c(z)- A.(z)IJI(z)

= 0,

(3.12)

and the associated equilibrium controls of (3.10) and (3.11) reduce to (3.13) (3.14)

Observe that the only difference between the players' strategies in (3.13) and (3.14) is in the treatment of the instantaneous correlation p.

739

Chapter 36. Stochastic Differential Portfolio Games 134

S. BROWNE

3.2. The complete, symmetric case The 'complete' case occurs when p 2 = I, in that there is then only one Brownian motion in the model. Without any loss of generality, let us consider only the case p = I. For the symmetric version of this case it is seen that the control functions of (3 .13) and (3 .14) reduce ァセ@ (z) = e1a, regardless of the particulars further to the growth-optimal proportion fv* (z) of the objective of the game and the value of Ill (z). However, when both players choose this policy, thefunctionsm(-, ·) of(2.7) and v 2 (·, ·) of(2.9) both reduce to zero, i.e. in this case we have

=

ュHセN@

セZ。L@

a,

e, e,

I)= カ R HセN@

セZ。N@

a.

l) =0

and as such we see from (2.6) that for the resulting ratio process we have dZ 1 = 0 for all t. As such, the state never changes, as any movement by a player will be immediately negated by his opponent. (This is never optimal if p 2 < 1.) The ODE of (3.12) reduces to the degenerate lll(z) = c(z)/A.(z), which need not be the value to the game. This degeneracy should be contrasted with the discrete-time complete case treated by Bell and Cover ( 1980), where a randomized version of the growth-optimal strategy is shown to be game-theoretic optimal for maximizing the probability of beating an opponent in a single play. Such a result obviously cannot hold in a continuous-time stochastic differential game with full revelation, since any randomization by a player will be immediately revealed to the other player, who can immediately (and exactly) adjust.

4. The probability maximizing game In this section, we consider the game where for two given numbers l < 1 < u, the objective of Investor A is to maximize the probability that he will outperform Investor B by u -I% before Investor B can outperform him by 1I l - I%. Similarly, Investor B wants to maximize the probability that he will outperform Investor A by 1I l - I% before Investor A can outperform him by u - I%. Put more simply: Investor A wants to maximize the probability of reaching u while Investor B is trying to maximize the probability of reaching /. Single-player games with related objectives have been studied previously in Pestien and Sudderth (1985. 1988 ). Mazumdar and Radner (1991) and Browne (1995, 1997, 1999). Let V (z) denote the value for this game-should it indeed exist: i.e. V ( z)

= sup m. fP z (1:1f.g f

セ@

>

f.g) Tu

= m. f sup P z (1:1ヲセァ@ セ@

>

J:g) Tu .

(4.1)

f

Theorem 3.1 applies to the probability maximizing game by taking A. = c = 0 in (3.7), and setting h(l) = 0 and h(u) = 1. Specifically, by Theorem 3.1, we find after simplification that V(z) must be the fast-increasing (in the sense of (3.5)) concave solution to for I < z < u

(4.2)

with V(/) = 0 and V(u) = 1. The solution to the non-linear Dirichlet problem of (4.2), subject to the boundary conditions Ill(/)= 0. lll(u) = I, is seen to be lll(z) = (zY -ZY)I(uY -IY), where the parameter y is defined by y

1 -K 2

= y (K, p) := -,------;;--.,,--1 + K 2 - 2pK.

(4.3)

740

S Browne 135

Stochastic differential portfolio games

Observe that for p 2 < 1, the denominator of (4.3) is positive for all K. As such, the sign of y depends on the sign of the numerator. Specifically, y < 0 if A has the advantage (i.e. if 81 > 82), while y > 0 if B has the advantage. Observe further that for the solution found above we have II!z > 0, regardless of the sign or magnitude of y, while Wzz < 0 only for y < I. Moreover, the required fast-increasing condition of (3.5), 2\llz + zWzz > 0, holds only for the case where -1 < y. Thus, we see that we require -1 < y < 1 for the game to have a value. It follows from (4.3) that this requirement is equivalent to the following two requirements on the parameters p and K: p 0 for i = I, 2, it follows that K > 0 and hence these conditions are trivially satisfied if p セ@ 0. Otherwise they are equivalent to p
0. The denominator of Cis always negative (since y 2 < 1), while the sign of the numerator of C depends on the sign of the quadratic QJ(K; p), where (4.10)

since the numerator of C can be written as Q1 (K; p)/K. For p < 0, Q1 (K; p) is trivially negative, and so C > 0. For p > 0, the two roots to the equation Q1 (K) = 0 are given by K-

=.!..o-R) p

and

K+

= .!..o +H). p

with Q1 (K) < 0 forK- < K < K+. Since we required K < I/ p, it is clear that we are only interestedinthesmallerroot,K-,andsoforK- < K < 1/p,itfollowsthatQJ(K) < 0. Moreover, a simple computation will show that K- < p, for p > 0, and since we in fact required K > p, we finally see that for all relevant K, we have Q1 (K) < 0, giving C > 0. Remark 4.1. The value function of (4.6) shows one manner in which the parameter K is a measure of advantage. Specifically, consider the probability maximizing game with l = I I u and Zo = l. Then it is natural to say that the player who has the higher probability of winning is the one with the advantage. Some direct manipulations will show that V ( l : y, u. I 1u) > if and only if y < 0, i.e. if and only if K > I. That is, Investor A has the advantage (a greater probability of winning) if his investment opportunity has the higher market price of risk.

!

Remark 4.2. Observe that the only structural difference in the investment policies of (3.1 0) and (3 .11) is in the treatment of the measure of advantage parameter K. Specifically, we see from (3.10) and (3.ll) that if A has the advantage, then the relative investment of B is greater, with the converse holding if B has the advantage. Thus a relatively 'bolder' strategy must be followed by the disadvantaged player, in particular on the order of the square of the measure of advantage parameter K. It is interesting to note that the determination of which player invests the larger absolute fraction of his wealth turns out to depend only on the instantaneous returns /Li, i = I. 2 and not the volatility parameters a;, i = l, 2. Specifically. after simplifying we observe that a2fh

11-2- r

implying that the player with the lower instantaneous return must invest more in his stock, in order to overcome the advantage of the other player. As can be seen, the volatility parameters, a1 and a2, do not play a role in determining which player invests a larger fraction of wealth. Remark 4.3. Observe further that since ヲセ@ and ァセ@ are constants, Proposition 2.1 implies that the optimal ratio process, Z*·*, is a geometric Brownian motion. Specifically, when we place the optimal controls of (4.7) and (4.8) into the functions m(f, g) of(2.7), and v 2 (f. g) of (2.9). we find that they reduce to (using the obvious identity 81 = 82K) elc

fhK 2

C)

m ( -, --

a1

a2

fhK 2

C)

= C 2812K(K- p) and V2(e1c -, --

a1

a2

= C2812(1

+ K2 -

2pK). (4.11)

742

S Browne

Stochastic differential portfolio games

137

From (2.8), we find that the optimal ratio process is the geometric Brownian motion (4.12) Observe that the constant min (4.11) is positive (since K > p), regardless of which player has the advantage, i.e. whether K > 1 or K < 1. However, the sign of E 1n(z;-*) depends on whether K > 1 or K < 1, with E1n(z;·*) > 0 if Investor A has the edge, and vice versa if Investor B has the edge. Remark 4.4. Proposition 2.1 exhibits the fact that for any admissible control functions f(z), g(z), the ratio process z!.g is a diffusion process with scale function given by sf,g (z) =

Jz exp{- セ@ セ@

[ m(f(y), g(y))] dy} y v2(j(y), g(y))

、セL@

for

I < z < u,

(4.13)

where m(f, g) and v 2 (f, g) are the functions defined in (2.7) and (2.9). As such for these given policies, the probability that Investor A wins the game can be written as

P (r f ,g < z u

T

sf,g(z)- sf,g(l) f ,g)- ---:o-----,---- Sf,g(u)-Sf,g(z)·

(4.14)

1

It follows from the single-player results of Pestien and Sudderth (1985, 1988) (see also Browne (1997), Remark 3.4) that for any given control function g(z), Investor A can maximize the probability in (4.14) by choosing the control policy that pointwise maximizes the ratio [zm(f, g)]/[z 2 v2(f, g)], which is equivalent to the pointwise maximizer of m(f, g)jv 2 (f, g). Similarly, for any given control policy j(z), Investor B can minimize the probability in (4.14) by choosing g to be the pointwiseminimizerofthe quantity m(f, g)jv 2 (f, g). Some computations will now accordingly show that the minimax value of the the function m(f, g)jv 2 (f, g) in fact occurs at the policies fv and gv of (4.7) and (4.8). See Nilakantan (1993) for some more general results along these lines. Remark 4.5. The value function of (4.6) can be used to set conservative capital requirements by setting it equal to a given preassigned probability of outperformance, say p, and then inverting for the required initial capital. Specifically, setting V (zo) = p and then solving for zo gives zo = (fY + p[uY -fY]) 11r. 4.1. The symmetric case For the symmetric case, we have e1 = e2 = e, a1 = a2 = a, and K = 1. For this case, so long as p 2 < 1, (4.3) becomes y (1, p) = 0. As such, by taking limits appropriately in (4.6) we observe that in the symmetric case the optimal value function reduces to

セ@

V(z; y, u, l) = V(z; 0, u, f)= lnG)

I

ln(T).

(4.15)

Moreover, for p 2 < 1, we see that C of (4.9) reduces to C(1, p, 0) = 1, and, as such, the = = competitively optimal controls of (4.7) and (4.8) reduce in the symmetric case to

e;a.

fv gv

Since the function in (4.15) satisfies the appropriate version of the Dirichlet problem of (3.12), we have the following.

743

Chapter 36. Stochastic Differential Portfolio Games 138

S. BROWNE

Corollary 4.1. ln the symmetric case, so long as p 2 < 1, the value of the game is given bv (4.15 ), and the competitively optimal policies for the probability maximizing problem are for each player to play the growth-optimal strategy,(} /CJ.

Observe that while the correlation parameter p does not play an explicit role here at all, in either the value function of (4.15) or the game-theoretic controls (} 1CJ, all of this holds only for p 2 < 1. Specifically, the limit in (4.15) is valid only for p 2 < I. This can be seen by observing thatfrom(4.3)wehave y(K, 1) = (l +K)/(1-K). As such, lim lim y(K, p) p-+

1 K-+ 1

= 0 f=

lim lim y(K, p) K-+

1 p-+ 1

= oo.

5. Expected time minimizing/maximizing games In this section we consider games where the objective is the minimization (maximization) of the expected time for one investor to outperform the other by a given percentage. The existence of a value for such games depends on which investor has the advantage, i.e. whether K > I or K < 1. Since the game is symmetric, in that one player's advantage is the other's disadvantage, we need only consider one game. Here we choose to study only the case where Investor A has the advantage (i.e. K > I) and as such is the minimizer (Investor A would be the maximizer if he were at a disadvantage with K < 1). Single-player games with minimal/maximal expected time objectives have been studied in Heath eta!. ( 1987) and Browne ( 1997, 1999). If Investor A has the advantage, in that K > 1, then he is trying to minimize the expected time to the performance goal u, while Investor B, in an effort to stop him, is trying to maximize the same expected time. Let G*(z) denote the value to this game, should it exist, i.e. G*(z) = infsupE 2 (r[ 8 ) = supinfE2 (r/"l. f

R

g

z < u.

for

(5.1)

f

As we show in the following theorem, the equilibrium portfolio policies turn out to be the individual growth-optimal portfolio policies. Theorem 5.1. Let G*(z) be the value of the game in (5.1) with associated optimal strategies J*(z) and g*(z). Then,for K > 1, G*(z) =

2

(}2

セ@ (K

-

1)

ln('!_),

with

Z

f*(z) =

!!.!_,

(}2

g*(z) = -

fJj

forall z S u. (5.2)

(J2

Proof While Theorem 3.1 is stated in terms of a maximization objective for Investor A and a minimization objective for Investor B, it can be applied to G*(z) of (5.1) by taking c(z) = -1, A = 0 and h(u) = 0. Specifically, G* (z) = -G(z) where G(z) = ウオーゥョヲサMeコHイェセ

f

X Iス@ g

As such, Theorem 3.1 applies directly to solution to

= infsup{-E 2 (r/ 8 )}, g f

for

z < u.

G, which in tum must be the fast-increasing concave for

::. < u. (5.3)

744

S Browne 139

Stochastic differential portfolio games

with G(u) = 0. It can be checked that the appropriate solution to (5.3) is indeed given by G(z) = -G*(z), where G*(z) is given in (5.2). (Observe that -G*(z) is sufficiently fastincreasing and concave only forK > 1.) It is easy to see that conditions (i), (ii) and (iii) of Theorem 3.1 hold for the appropriate value functions in the respective cases. In particular, condition (i) holds since for this case dG*(z)/ dz = -2[z8}(K 2 - 1)]- 1, and so (3.9) reduces to

cセスHャ@

セ@

K2)) 2

L

([f(Z[g)]2

+ [g(Z{'g)f) ds

< oo,

which must hold by the admissibility requirement on the policies f and g. As such, we may conclude that G* is the value of the game and substitute it into (3.1 0) and (3.11) to obtain the competitively optimal controls, which in this case reduce to the individual growth-optimal strategies.

5.1. Connections with logarithmic utility Observe that if we take logarithms in (2.8) and then take expectations, we get E(ln(Z{' 8 ))

= ln(Zo) + E {

[m Ap 2 .

] 1/2

747

Chapter 36. Stochastic Differential Portfolio Games S.BROWNE

142

7. Utility-based games So far, the objectives considered have related solely to the achievement of relative performance goals and shortfall levels, and the games considered allowed only one winner. In this section, we consider games of a fixed duration T, where both investors receive utility (or disutility) from the ratio of the wealth processes (i.e. from the relative performance of their respective wealths). Specifically, for given concave-increasing utility functions f3(z) and U (z), and for a given fixed terminal time T, let J f,g (t, z) be the expected payoff function under the policy pair f. g. defined by Jf,g(t, z) = Er,z

(iT

f3(Z[g) exp{-[A(Z['g) dv} ds

+ U(Z{gl exp{-JTA(Z{'g) ds }). (7.1)

(Here we use the notation Er,zO as shorthand forE(- I Zr = z).) Once again we assume that A is trying to maximize this quantity while B is trying to minimize it. Let J (t, z) denote the value of this game, should it exist, i.e. J(t, z)

= infsup Jf,g(t, z) = supinf Jf.g(t, z). g

f

f

(7.2)

g

and let h (t, z) and g J (t, z) denote the associated optimal strategies. Note that in this case we have time-dependence, which will lead to a non-linear Cauchy problem, as opposed to the Dirichlet problem of Theorem 3.1. An analysis similar to that of Theorem 3.1 and its proof (see next section) will show that, if T (t' z) : ro. T] X (0, 00) H- セr@ is a e 1,2 concave and sufficiently fast-increasing solution (in z) to the non-linear Cauchy problem: 11+

RQ

z1 2 セ・ゥ{HャM

K2 )Tz- (I

+ K2 -

2pK)(Tz

+ Zlzz)] + f3- At= 0

(7.3)

with I(T, z) = U(z), then subject to the appropriate regularity conditions (e.g. that l(z) satisfies conditions (i), (ii) and (iii) of Theorem 3.1), T(t, z) is the competitively optimal value function of the game in (7.2), i.e. T (t, z) = J (t, z), and in this case the competitively optimal control functions are given by (7.4)

*

gJ(t, z)

ez(Tz(t,z)) f'T(t, z) [(1- pK)('Y (t, z) + zl

= az

2

22

(t, z))- T 2 (t, z)].

(7.5)

(The proof of this result is in fact easier than that of its Dirichlet counterpart, Theorem 3.1, and so we leave it for the reader to fill in the missing details.) As an example of a utility-based game, consider the case where f3(z) = Jc(z) = 0, and where U (z) = za, for 0 < a < I. (The logarithmic case treated earlier would correspond to the limiting case of a = 0, since liiDa-+o(z" - I )/a= ln(z).) For this case we find the value function (7.6)

748

S Browne

143

Stochastic differential portfolio games

where q(a) is defined by

+

K2 - 2pK) 2 (l - K2 ) - a(l 2 2[(1 - p 2 )a 2 - 1]

q(a) ·- ae .-

The associated optimal strategies are again constant proportions, with *t

Mセ{HーェkャI。Q}@

h ( , z) -

O"J

(1

p 2) a 2 - 1

-

and

* 62 [ (l - pK)a - 1 ] g,(z) = (1 2) 2 1 . - p

a2

a

(7.7)

-

Comparison of the policies of (7.7) with those obtained previously for goal-based games will provide obvious further analogues between objective criteria and utility function (see Browne (1995, 1997, 1999) for other equivalences in single-player games).

8. Proofs To prove Theorem 3.1, we first exploit the the Hamilton-Jacobi-Bellman (HJB) equations of dynamic programming for single-player games (cf. Krylov (1980)) to obtain candidate value functions and equilibrium portfolio control strategies for the two-player games considered here. These controls and value functions will then be verified to be in fact competitively optimal via an extension of a fairly standard martingale argument. To proceed, observe that for any given policy function g(z) used by Investor B, the HJB optimality equation for Investor A for maximizing vf,g (z) of (3.2) over control policies {.fi} E g., to be solved for a function v*·g is (see e.g. Krylov (1980), Theorem 1.4.5): sup{.Af,g v*·g

+c -

}..v*·g}

= 0,

v*· 8 (l) = h(l), v*· 8 (u) = h(u),

(8.1)

f

where .Af.g is the generator given by (2.10). The infimum of this, v(z) = infg v*·g (z), is the upper-value function (see e.g. Fleming and Souganides (1989)). Similarly, for any given policy function f(z) used by Investor A, the HJB optimality equation for Investor B for minimizing vf,g (z) of (3.2) over control policies {g1 } E g., to be solved for a function vf·* is vf·*(l)

= h(l),

vf·*(u)

= h(u).

(8.2)

The supremum of this, .!!.(Z) = supf vf·*(z), is the lower-value function. Assuming now that (8.1) admits a classical solution with v:Zg < 0, we may use calculus to optimize with respect to fin (8.1) to obtain the maximizer (as a function of g) (8.3)

Similarly, if we assume that (8.2) admits a classical solution with 2zv{* minimizer will be given by

e(

2 g-( z: f) = -

a2

ZVzf,*

2zv[*

+ z2v{z*

)

+ j( z) pO"J- ( az

ZVzf,*

2zv[*

+ z2 v{z*

+ Z2 Vzzf,* ) . + z2v{z*

> 0, its

(8.4)

749

Chapter 36. Stochastic Differential Portfolio Games S. BROWNE

144

(Observe that B's second-order condition is the basis of the fast-increasing condition of (3.5).) The optimizers ](z : g) and g(z : f) of (8.3) and (8.4) are also referred to as the optimal reaction functions. a saddlepoint exists, and that hence the game must have an achievLet us assume now エィセ@ able value with v•·g = vf·• = v (see e.g. Elliott ( 1976), Maitra and Sudderth ( 1996), Fleming and Souganides (1989)). If this is the case, then we can find the saddlepoint by substituting g into (8.3) and j into (8.4) and solving the resulting linear equations. When we do this we obtain the optimal control functions (8.5) (8.6) where K is the measure of advantage parameter defined in (3.6), and where r is the differential operator of (3.4). When the control functions f*(z), g*(z) of (8.5) and (8.6) are then in tum substituted back into either (8.1) or (8.2), with v = v•·g* = vf* ·*, we obtain, after some manipulations. the non-linear Dirichlet problem of (3.7), with Ill = v. To complete the argument (i.e. to verify that v is indeed the value of the game, and is achieved by the policies f*, g* of (8.5) and (8.6)), we can now rely on the results of Fleming and Souganides (1989), who provide a quite general verification argument for stochastic differential games, of which the model treated here is a special case. Alternatively, we can construct a verification argument directly, similar to the standard martingale arguments in, for example, Fleming and Soner (1993). To carry out the latter program, define for any admissible policy pair f, g = l.fr. gr, t :::: 0}, the process fort :::: 0.

(8.7)

J;

A.(Z{K)ds. where Ill is the concave fast-increasing solution of (3.7), (3.8), and A{g := M (t : f, g) may be interpreted as a conditional (on :Fr) expectation of the gain if controls f, g are used up to time t, and the optimal controls thereafter. It can be shown that, under the conditions given in Theorem 3.1, M is a (uniformly integrable) martingale under the pair {f/. g!; t ::::OJ, but a supermartingale under the pair {!1 , g!; t :::: O}, for any admissible 1. and a submartingale under the pair {ft*. g1 ; t :::: OJ for any admissible {gr}, where f* and g* are the policies given in (8.5) and (8.6). The representation of the function \ll(z) of Theorem 3.1 as the value of the game. and the competitive optimality of the saddlepoint policies (j*. g*) of (8.5) and (8.6), will now follow as a consequence of the following lemma.

u;

Lemma 8.1. For any admissible policies f = (fr, t :::: OJ and g = {g 1 , t :::: 0}. with M (t f, g) as defined in (8.7), and f*, g* as defined in (8.5) and (8.6), we have Ez[M(t Ez[M(t

A ,f.g* : A

f, g*)] ::::: M(O. f, g*)

= \ll(z)

,f*,g: f*, g)] :C:: M(O: f*, g)= lll(z),

with equalities holding if and only iff

= f*

and g

=

g*

fort :::: 0 fort :C:: 0,

(8.8) (8.9)

750

S Browne

Stochastic differential portfolio games

145

Proof An application of Ito's formula to the process M(t : f, g) of (8.7) using (2.6) gives

M(t

1\ rf.g :

f,

g)=

M(O: f,

{t 1\Tf.g f g)+ Jo e-As" 8 Q(fs, gs

{tl\rf8

+ Jo

f e-A," 8

: Z{ 8 ) ds

Z{ 8 Wz(Z{ 8 )[fsaldW5(1)- g5 az dW5( 2)],

(8.10)

where the quadratic form Q(j, g : z) is defined by

Q(f, g: z) := f 2at(!z 21Jlzz(z))

+ g 2a}C!z 2Wzz(Z) + ziJ.lz(z))

+ (fa1(h - gazBz)ziJ.lz(z)+ c(z)- A.(z)IJ.l(z).

fgpawz(z 21Jlzz(z)

+ ziJ.lz(z))

Observe first that if assumption (i) holds, then the stochastic integral term in (8.10) is a finite-variance martingale, and hence uniformly integrable. Some direct computations will now show that for any given g, the max!E9- Q(j, g : z) is achieved at the control function J(z : g) of (8.3), and similarly, for any given f, the min8 Eg. Q(f, g : z) is achieved at the control function g(z : f) of (8.4). Therefore, the minimax value of Q is reached at the policies f* and g* of (8.5) and (8.6), and is equal to

Q(j*, g* : z)

= RQセ@

ziJ12

e}w -

K 2 )Wz

- (I

+ K2 -

2pK)(Wz

+ ZWzz)l + c(z)

- A.(z)IJ.l

=

0

(8.11) where the last equality follows from (3.7). As such, for other possible f, g, we have

Q(j, g*: z) :'0 sup Q(f, g*: z) = Q(j*, g*: z) f

f*

of (8.5), g* of (8.6), and for all

=0 = infQ(f*, g: z) :'0 Q(f*, g: z). g

(8.12) Therefore, from (8.10) we see that for all admissible control functions f, and g* of (8.6) we have

(The inequality in (8.13) following from (8.12).) Observe now that the stochastic integral term on the left-hand side of (8.13) is a continuous local martingale that is in fact a martingale by assumption (i) of Theorem 3.1. Hence, taking expectations on (8.13) directly gives the desired inequality of (8.8). Equality holds in both (8.13) and (8.8) for f = f*.

751

Chapter 36. Stochastic Differential Portfolio Games 146

S. BROWNE

Similarly, for all admissible control policies g, and f* of (8.5) we have f*"

1'!\T . exp{-A{*'

8}

z{*- 8 1Jl2 (Z{*· 8 ) [j*(Z{*' 8 )ai dw}IJ- g,a2dW}2J] {11\Tf*.g

= M(t

A

rf*,g: f*, g)- M(O, f*, g)- Jo

:=: M(t

!\

rf*,g: f*, g)- M(O, f*, g)-

exp(-A{*' 8 ) Q(f*, g5 : z{*· 8 )ds

1

11\Tf*,g

exp{-A{*' 8 ) {infQ(f, g:

0

z(· 8 ))ds

g

= M(t A rf*,g: f*, g)- M(O. f*, g),

(8.14)

where again, the inequality follows from (8.12). Now once again assumption (i) of Theorem 3.1 shows that the stochastic integral term in (8.14) is a continuous local martingale that is in fact a martingale. As such, inequality (8.9) is established by taking expectations in (8.14), with equality holding only if g = g*. Finally, observe that if condition (iii) of Theorem 3.1 holds, then both zf*(z) and zg*(z) as well as the function zm(f*(z), g*(z)), where m(f, g) is defined in (2.7), are all Lipschitz continuous, implying therefore that the drift and diffusion coefficients of the resulting competitively optimal ratio process, Z*·* := zf*.g*, are locally Lipschitz continuous; therefore the equation (2.6) with f = f*, g = g* admits a strong solution. Moreover, since Q(f*, g*; z) = 0, the process {M(t !\ rf*,g* : f*, g*), t :::_OJ is a (uniformly integrable) martingale under the conditions of Theorem 3.1. Thus we have shown that E2 [M(t

!\

rf.g*: f, g*)] ::0 Ez[M(t

!\

rf*,g*: f*, g*)] ::0 Ez[M(t

!\

rf*,g: f*, g)]. (8.15)

Since Ez[liffit_,. 00 inf M(t !\ rf,g : f, g)] = vf,g (z), we may now complete the proof by sending t---+ oo in (8.15) to obtain vf.g* (z) :=: vf*,g* (z) :=: vf*,g(z). The passage to the limit is justified for the left-hand side by Fatou's lemma (since everything is bounded from below), and for the other terms by using uniform integrability and the martingale stopping theorem (which is valid by the assumption that Ez(Tf.g) < oo for all admissible f and g). This completes the proof of Theorem 3.1. Acknowledgement

The author is grateful to Eugene Rozman for helpful discussions and computational assistance. References BELL, R. M. AND CoVER, T. M. (1980). Competitive optimality of logarithmic investment. Math. Operat. Res. 5. 161-166. BROWNE, S. (1995). Optimal investment policies for a firm with a random risk process: exponential utility and minimizing the probability of ruin. Math. Operat. Res. 20, 937-958. BROWNE, S. (1997). Survival and growth with a fixed liability: optimal portfolios in continuous time. Math. Ope rat. Res. 22, 468-493. BROWNE, S. (1998). The rate of return from proportional portfolio strategies. Adv. Appl. Pro b. 30, 216--238. BROWNE, S. ( 1999). Beating a moving target: optimal portfolio strategies for outperforming a stochastic benchmark. Finance and Stochastics 3, 275-294.

752

S Browne

Stochastic differential portfolio games

147

DUBINS, L. E. AND SAVAGE, L. J. (1965, repr. 1976). How to Gamble If You Must: Inequalities for Stochastic Processes. Dover, New York. DUFFIE, D. (1996). Dynamic Asset Pricing Theory, 2nd edn. Princeton University Press. ELLIOTT, R. (1976). The existence of value in stochastic differential games. SIAM J. Contr. and Opt. 14, 85-94. FLEMING, W. H. AND SaNER, H. M. (1993). Controlled Markov Processes and Viscosity Solutions. Springer, New York. FLEMING, W. H. AND SOUGANIDES, P. E. (1989). On the existence of value functions of two-player, zero-sum stochastic differential games. Indiana Univ. Math. J. 38, 293-314. FRIEDMAN, A. (1976). Stochastic Differential Equations and Applications, Vol 2. Academic Press, New York. HEATH, D., 0REY, S., PESTIEN, V. C. AND SUDDERTH, W. D. (1987). Minimizing or maximizing the expected time to reach zero. SIAM J. Contr. and Opt. 25, 195-205. KRYLOV, N. V. (1980). Controlled Diffusion Processes. Springer, New York. MAITRA, A. P. AND SUDDERTH, W. D. (1996). Discrete Gambling and Stochastic Games. Springer, New York. MAZUMDAR, M. AND RADNER, R. (1991 ). Linear models of economic survival under production uncertainty. Econ. Theory 1, 13-30. MERTON, R. (1971). Optimum consumption and portfolio rules in a continuous time model. J. Econ. Theory 3, 373413. MERTON, R. (1990). Continuous Time Finance. Blackwell, Oxford. NILAKANTAN, L. (1993). Continuous-Time Stochastic Games. Ph.D. dissertation, University of Minnesota, MN. 0REY, S., PESTIEN, V. C. AND SUDDERTH, W. D. (1987). Reaching zero rapidly. SlAM 1. Contr. and Opt. 25, 1253-1265. PEST! EN, V. C. AND SUDDERTH, W. D. (1985). Continuous-time red and black: how to control a diffusion to a goal. Math. Operat. Res. 10, 599-M I. PESTIEN, V. C. AND SUDDERTH, W. D. (1988). Continuous-time casino problems. Math. Operat. Res. 13,364-376.

Chapter 37 Fractional Kelly Strategies in Continuous Time: Recent Developments Mark Davis∗ and S´ebastien Lleo† Abstract The Kelly criterion and fractional Kelly strategies hold an important place in investment management theory and practice. Both the Kelly criterion and fractional Kelly strategies, e.g. invest a fraction f of one’s wealth in the Kelly portfolio and a proportion 1 − f in the risk-free asset, are optimal in the continuous time setting of the Merton [33] model. However, fractional Kelly strategies are no longer optimal when the basic assumptions of the Merton model, such as the lognormality of asset prices, are removed. In this chapter, we present an overview of some recent developments related to Kelly investment strategies in an incomplete market environment where asset prices are not lognormally distributed. We show how the definition of fractional Kelly strategies can be extended to guarantee optimality. The key idea is to get the definition of fractional Kelly strategies to coincide with the fund separation theorem related to the problem at hand. In these instances, fractional Kelly investment strategies appear as the natural solution for investors seeking to maximize the terminal power utility of their wealth.

1 Introduction The Kelly criterion and fractional Kelly strategies hold an important place in investment management theory and practice. The Kelly criterion maximizes the log-return on invested wealth and is therefore related to the seminal work of Bernoulli [9]. Early contributions to the theory and application of the Kelly criterion to gambling and investment include Kelly [20], Latan´e [23], Breiman [8], Thorp [43] or Markowitz [32]. The main reference is undeniably [30]. Readers interested in an historical account of the Kelly criterion and of its use at the gambling table and in the investment industry will refer to Poudstone [39]. From a practical investment management perspective, several of the most successful investors, including Keynes, Buffett and Gross have used Kelly-style strategies in their funds (see Ziemba [48], Thorp [44] and Ziemba and Ziemba [47] for details). The Kelly criterion has a number of good as well as bad properties, as discussed by MacLean, Thorp and Ziemba [27]. Its ‘good’ properties extend beyond practical ∗ Department of Mathematics, Imperial College London, London SW7 2AZ, England, Email: mark. [email protected] † Finance Department, Reims Management School, 59 rue Pierre Taittinger, 51100 Reims, France, Email: [email protected]

753

754

M Davis and S Lleo

asset management and into asset pricing theory, as the Kelly portfolio is the num´eraire portfolio associated with the physical probability measure. This observation forms the basis of the ‘benchmark approach to finance’ proposed by Platen [37] and Heath and Platen [38] (see also Long [26] and Becherer [3]). The ‘bad’ properties of the criterion are also well studied and understood. Samuelson, in particular, was a long time critique of the Kelly criterion (see [40], [42] and [41]). A main drawback of the Kelly criterion is that it is inherently a very risky investment. To address this shortcoming, MacLean, Ziemba and Blazenko [31] propose the following fractional Kelly strategy: invest a fraction f of one’s wealth in the Kelly portfolio and a proportion 1 − f in the risk-free asset. MacLean, Sanegre, Zhao and Ziemba [29] MacLean, Ziemba and Li [28] pursued further research in this direction. There are two key advantages to this definition: first, a fractional Kelly strategy is significantly less risky than the full Kelly portfolio, while maintaining a significant part of the upside. Second, fractional Kelly strategies are optimal in the continuous time setting of the Merton [33] model. In fact, fractional Kelly strategies correspond to the optimal investment of a power utility investor seeking to maximize the terminal utility of his/her wealth. Unfortunately, fractional Kelly strategies are no longer optimal when the basic assumptions of the Merton model, such as the lognormality of asset prices, are removed (see MacLean, Ziemba and Li [28]). In recent years, a number of attempts have been made to remedy this situation and extend the definition of fractional Kelly strategies to guarantee their optimality. In this chapter, we present an overview of some recent developments related to Kelly investment strategies in an incomplete market environment where asset prices are not lognormally distributed. In section 2, we introduce the Kelly portfolio and fractional Kelly strategies in the context of the Merton model. Next, we consider in Section 3 an Intertemporal Capital Asset Pricing Model (ICAPM) where the drift of the asset price dynamics are affine functions of some affine factors. In this ICAPM, traditionally defined fractional Kelly strategies are no longer optimal. We must therefore extend the definition of the fractional Kelly strategies along the lines of a fund separation theorem in order to guarantee optimality. In section 4, we present an extension to a benchmarked investor, that is an investor with the objective of outperforming a given investment benchmark, before considering the impact of partial observation on the underlying valuation factors in Section 6. Finally, we look at optimal investment strategies in a jump-diffusion setting, where asset prices follow jump-diffusion processes and the underlying factors are modelled as diffusion processes.

2 The Kelly Criterion Portfolio and Fractional Kelly Strategies in the Merton World 2.1 The Kelly Criterion Portfolio in the Merton Model We start by introducing some of the notation that we will need in the remainder of the chapter. Let (Ω, {Ft } , F , P) be the underlying probability space. On this

Chapter 37. Fractional Kelly Strategies in Continuous Time

755

space is defined an Rm -valued (Ft )-Brownian motion W (t) with components Wk (t), k = 1, . . . , N . Si (t) denotes the price at time t of the ith security, with i = 0, . . . , m. Let S0 denote the wealth invested in a money market account. The dynamics of the money market account is given by: dS0 (t) = rdt, S0 (t)

S0 (0) = s0

(2.1)

where r ∈ R+ is the risk-free rate. The dynamics of the m risky securities and n factors can be expressed as: N

 dSi (t) = µi dt + σik dWk (t), Si (t)

Si (0) = si ,

i = 1, . . . , m

(2.2)

k=1

where the market parameters µ = (µ1 , . . . , µm )′ represents the rate of return vector and the volatility Σ := [σij ], i = 1, . . . , m, j = 1, . . . , m is a m × m matrix. More synthetically, dS(t) = D(S(t))µdt + D(S(t))ΣdW (t)

(2.3)

where D(S(t)) denotes the diagonal matrix with S1 (t), . . . , SM (t) on the diagonal. We make the further assumption that: Assumption 2.1. The matrix Σ is positive definite. This assumption rules out simple arbitrage opportunities. Let Gt := σ((S(s), X(s)), 0 ≤ s ≤ t) be the sigma-field generated by the security, liability and factor processes up to time t. An investment strategy or control process is an Rm -valued process with the interpretation that hi (t) is the fraction of current portfolio value invested in the ith asset, i = 1, . . . , m. The fraction invested in the m money market account is then h0 (t) = 1 − i=1 hi (t).

Definition 2.2. An Rm -valued control process h(t) is in class A(T ) if the following conditions are satisfied: 1. h(t) is progressively measurable with respect to {B([0, t]) ⊗ Gt }t≥0 and is left continuous with right limit (c` adl` ag); T 2. P ( 0 |h(s)|2 ds < +∞) = 1, ∀T > 0; 3. the Dol´eans exponential χht , given by χht

  t   1 2 t ′ ′ ′ h(s) ΣdWs − γ := exp γ h(s) ΣΣ h(s)ds 2 0 0

(2.4)

is an exponential martingale, i.e. E[χhT ] = 1 Definition 2.3. We say that a control process h(t) is admissible if h(t) ∈ A(T ).

756

M Davis and S Lleo

Taking the budget equation into consideration, the wealth, V (t), of the asset in response to an investment strategy h ∈ H follows the dynamics dV (t) = rdt + h′ (t)(µ − r1)dt + h′ (t)ΣdWt V (t)

(2.5)

with initial endowment V (0) = v and where 1 ∈ Rm is the m-element unit column vector. Thus,  t  t 1 ln V (t) = ln v + h′ (s)ΣdWs r + h′ (s)(µ − r1) − h′ (s)ΣΣ′ h(s)ds + 2 0 0 (2.6) The objective of a Kelly investor with a fixed time horizon T is to maximize J(t, h; T, γ) = E[U (VT )] = E[ln VT ] A pointwise maximization of the criterion J yields the Kelly portfolio: h∗ = (ΣΣ′ )−1 (µ − r1) Substituting back in (2.6), we find that the wealth of a Kelly investor is   1 ′ ′ −1 ∗ V (T ) = v exp r + (µ − r1) (ΣΣ ) (µ − r1) (T − t) 2  + (µ − r1)′ (ΣΣ′ )−1 Σ(WT − Wt ) and the instantaneous growth rate of the strategy follows the dynamics   1 dRt = r + (µ − r1)′ (ΣΣ′ )−1 (µ − r1) dt + (µ − r1)′ (ΣΣ′ )−1 ΣdWt 2

(2.7)

(2.8)

2.2 Fractional Kelly Strategies To mitigate the risks inherent in a the Kelly investment strategy, MacLean, Ziemba and Blazenko [31] propose the following fractional Kelly strategy: invest a fraction f of one’s wealth in the Kelly portfolio and a proportion 1 − f in the risk-free asset. MacLean, Sanegre, Zhao and Ziemba [29] propose a methodology for computing the optimal fractional Kelly weights at discrete time intervals. In a continuous time setting where asset prices follow a geometric Brownian motion, they show that a fractional Kelly strategy is optimal with respect to Value at Risk and a Conditional Value at Risk criteria. MacLean, Ziemba and Li [28] further prove that fractional Kelly strategies are efficient when asset prices are lognormally distributed. This last result is actually a corollary to Merton’s Fund Separation theorem. In the Merton problem with no consumption and a power utility function, the objective of an investor is to maximize the expected utility of terminal wealth over a fixed time horizon T :  γ ln VT   γ V e J(t, h; T, γ) = E[U (VT )] = E T = E γ γ

Chapter 37. Fractional Kelly Strategies in Continuous Time

757

with risk aversion coefficient γ ∈ (−∞, 0) ∪ (0, 1). We define the value function Φ corresponding to the maximization of the auxiliary criterion function J(t, h; T, γ) as Φ(t) = sup J(t, h; T, γ)

(2.9)

h∈A

By Itˆo’s lemma, e

γ ln V (t)

   t g(h(s); γ)ds χht = v exp γ γ

(2.10)

0

where 1 g(h; γ) = − (1 − γ)h′ ΣΣ′ h + h′ (µ − r1) + r 2 and χht is defined in (2.4) We can solve the stochastic control problem associated with (2.9) by a change of measure argument (see exercise 8.18 in [35] or [21] in the context of risk sensiym tive control). Let Ph be the measure on (Ω, FT ) defined via the Radon-Nikod´ 1 derivative dPh := χhT dP

(2.11)

For h ∈ A(T ), Wth

= Wt − γ



t

Σ′ h(s)ds

0

is a standard Brownian motion under the measure Ph . Moreover, the control criterion under this new measure is

 T vγ h E exp γ g(h(s); γ)ds (2.12) I(t, h; T, γ) = γ t,x t where Eht [·] denotes the expectation taken with respect to the measure Ph at an initial time t. Under the measure Ph , the control problem can be solved through a pointwise maximisation of the auxiliary criterion function I(v, x; h; t, T ). The optimal control h∗ is simply the maximizer of the function g(x; h; t, T ) given by h∗ = which represents a position of

1 (ΣΣ′ )−1 (µ − r1) 1−γ

1 1−γ

in the Kelly criterion portfolio.

Remark 2.4. The change of measures simplifies the problem considerably: under the measure Ph we can solve the optimization through a pointwise maximization, as we did earlier on in the logarithmic utility case. 1 Lo` eve

[25] provides a thorough treatment of the Radon-Nikod´ ym theorem

758

M Davis and S Lleo

Substituting (2.13) into (2.10) and (2.4), we derive the value function Φ(t), or optimal utility of wealth,     1 vγ exp γ r + (µ − r1)(ΣΣ′ )−1 (µ − r1) (T − t) Φ(t) = γ 2(1 − γ) as well as an exact form for the exponential martingale χ∗t associated with the control h∗ :

γ ∗ (µ − r1)′ Σ−1 W (t) χt := exp 1−γ

2 1 γ − (µ − r1)′ (ΣΣ′ )−1 (µ − r1)t (2.13) 2 1−γ Therefore, fractional Kelly strategies appear as a consequence of a classical Fund Separation Theorem: Theorem 2.5 (Fund Separation Theorem). Any portfolio can be expressed as a linear combination of investments in the Kelly (log-utility) portfolio hK (t) = (ΣΣ′ )−1 (µ − r1)

(2.14)

and the risk-free rate. Moreover, if an investor has a risk sensitivity γ, the proportion 1 of the Kelly portfolio will equal 1−γ . A key limitation of fractional Kelly strategies is that they are only optimal within the Merton model, that is when asset prices are lognormally distributed (see also MacLean, Ziemba and Li [28]). This situation suggests that the definition of Fractional Kelly strategies could be broadened in order to guarantee optimality. We can take a first step in this direction by revisiting the ICAPM (see Merton[34]).

3 Incomplete Markets and Intertemporal Capital Asset Pricing Model 3.1 The Model Merton [34] proposed an Intertemporal Capital Asset Pricing Model (ICAPM) in which the drift rate of the asset prices depend on a number of Normally-distributed factors. Bielecki and Pliska (see in particular [5] and [6]), Kuroda and Nagai [21] as well as Davis and Lleo (see [11],[12] and [15]) further developed this idea in the context of risk-sensitive control. Let (Ω, {Ft }, F , P) be the underlying probability space. On this space is defined an RN -valued (Ft )-Brownian motion W (t) with components Wk (t), k = 1, . . . , N . Si (t) denotes the price at time t of the ith security, with i = 0, . . . , m, and Xj (t) denotes the level at time t of the jth factor, with j = 1, . . . , n. We also assume that the factors are observable.

Chapter 37. Fractional Kelly Strategies in Continuous Time

759

Let S0 denote the wealth invested in a money market account. The dynamics of the money market account is given by: dS0 (t) = (a0 + A′0 X(t))dt, S0 (t)

S0 (0) = s0

(3.1)

where a0 ∈ R is a scalar constant, A0 ∈ Rn is a n-element column vector and throughout the paper x′ denotes the transpose of the matrix or vector x. We further assume that the expected rates of return of the assets depend on n valuation factors X1 (t), . . . , Xn (t) which follow the dynamics given in equation (3.3) below Let N = n + m. The dynamics of the m risky securities and n factors are dS(t) = D(S(t))(a + AX(t))dt + D(S(t))ΣdW (t),

S(0) = s

(3.2)

and dX(t) = (b + BX(t))dt + ΛdW (t),

X(0) = x

(3.3)

where X(t) is the Rn -valued factor process with components Xj (t) and the market parameters a, A, b, B, Σ := [σij ], i = 1, . . . , m, j = 1, . . . , N , Λ := [Λij ], i = 1, . . . , n, j = 1, . . . , N are matrices of appropriate dimensions. Throughout the rest of the chapter, we assume the following: Assumption 3.1. The matrices ΣΣ′ and ΛΛ′ are positive definite. The wealth V (t) of the portfolio in response to an investment strategy h ∈ A(T ) is now factor-dependent with dynamics dV (t) ˆ = (a0 + A′0 X(t))dt + h′ (t)(ˆ a + AX(t))dt + h′ (t)ΣdWt V (t)

(3.4)

where a ˆ := a − a0 1, Aˆ := A − 1A′0 , and the initial endowment V (0) = v. The investor seeks to maximize Φ(t, x) = sup J(t, x, h; T, γ)

(3.5)

h∈A

where the expected utility of terminal wealth J(t, x, h; T, γ) is factor-dependent:  γ ln VT   γ e V J(t, x, h; T, γ) = E[U (VT )] = E T = E γ γ By Itˆo’s lemma, e

γ ln V (t)

   t g(Xs , h(s); γ)ds χht = v exp γ γ

(3.6)

0

where 1 ˆ + a0 + A′ x g(x, h; γ) = − (1 − γ)h′ ΣΣ′ h + h′ (ˆ a + Ax) 0 2 and the exponential martingale χht is still given by (2.4).

(3.7)

760

M Davis and S Lleo

Applying the change of measure argument, we obtain the control criterion under the measure Ph

 T vγ h exp γ g(Xs , h(s); γ)ds (3.8) I(t, x, h; T, γ) = E γ t,x t where Eht,x [·] denotes the expectation taken with respect to the measure Ph and with initial conditions (t, x). The dynamics of the state variable X(t) under the new measure is dX(t) = (b + BX(t) + γΛΣ′ h(t))dt + ΛdWth ,

t ∈ [0, T ]

(3.9)

The value function Φ for the auxiliary criterion function I(t, x; h; T, γ) is defined as Φ(t, x) = sup I(t, x; h; T, γ)

(3.10)

h∈A(T )

After solving the stochastic control problem (see [13] for an outline of the argument connecting ICAPM and risk-sensitive asset management, as well as [5] and [21] for details), we obtain the optimal investment policy h∗ (t) 1 −1 ˆ ˜ X(t))] (ΣΣ′ ) [ˆ a + AX(t) + γΣΛ′ DΦ(t, (3.11) h∗ (t) = 1−γ ˜ as well as a solution for the logarithmically transformed value function Φ 1 ˜ x) = 1 x′ Q(t)x + x′ q(t) + k(t) ln Φ(t, x) := Φ(t, (3.12) γ 2 where Q is a n × n symmetric non-negative matrix solving a Riccati equation, q is a n-element column vector solving a linear ODE and k is a scalar respectively satisfying (see [21] for details). As a result, 1 ˆ (ΣΣ′ )−1 [ˆ a + AX(t) + γΣΛ′ (Q(t)X(t) + q(t))] (3.13) h∗ (t) = 1−γ In the ICAPM, the classical Kelly strategy splitting the wealth of an investor in an allocation to the Kelly portfolio and an allocation to the money market account is no longer optimal. However, a new view of Fractional Kelly investing emerges as a consequence of the following ICAPM Fund Separation theorem: Theorem 3.2 (ICAPM Fund Separation Theorem). Any portfolio can be expressed as a linear combination of investments into two funds with respective risky asset allocations: ˆ a + AX(t)) hK (t) = (ΣΣ′ )−1 (ˆ hI (t) = −(ΣΣ′ )−1 ΣΛ′ (Q(t)X(t) + q(t))

(3.14)

and respective allocation to the money market account given by ′ ′ −1 ˆ hK (ˆ a + AX(t)) 0 (t) = 1 − 1 (ΣΣ ) hI0 (t) = 1 + 1′ (ΣΣ′ )−1 ΣΛ′ (Q(t)X(t) + q(t)) Moreover, if an investor has a risk aversion γ, then the respective weights of each γ 1 mutual fund in the investor’s portfolio equal 1−γ and − 1−γ , respectively.

Chapter 37. Fractional Kelly Strategies in Continuous Time

761

In the factor-based ICAPM, ˆ a + AX(t)] (ΣΣ′ )−1 [ˆ

(3.15)

represents the Kelly (log utility) portfolio and (ΣΣ′ )−1 ΣΛ′ (Q(t)X(t) + q(t))

(3.16)

is the ‘intertemporal hedging porfolio’ identified by Merton. The appropriate definition of Kelly strategies is not an investment in the Kelly portfolio supplemented by cash, but an investment in the Kelly portfolio and the intertemporal hedging portfolio. This new definition raises some questions as to the practicality of interemporal hedging portfolio as an investment option. When the asset price noise and the factor noise are uncorrelated, i.e. ΣΛ′ = 0, the intertemporal hedging portfolio vanishes and is replaced by an allocation to the money market asset. In this case, fractional Kelly strategies are optimal and we have the following corollary: Corollary 3.3 (ICAPM Fund Separation Theorem with Uncorrelated Noise). Any portfolio can be expressed as a linear combination of investments in the Kelly (log-utility) portfolio ˆ a + AX(t)) hK (t) = (ΣΣ′ )−1 (ˆ

(3.17)

and the money market asset. Moreover, if an investor has a risk sensitivity γ, the γ 1 and − 1−γ . proportion of the Kelly portfolio will equal 1−γ 3.2 Example: The Relevance of Valuation Factors In this section, we compare the investment strategies in the ICAPM and the Merton model in a simple setting with m = 1 risky asset and n = 1 factor. 3.2.1 Setting The single factor is a short term interest rate behaving according to the Vasicek model (see [45]): dX(t) = (b0 − b1 X(t))dt + ΛdW (t),

X(0) = x

(3.18)

or equivalently dX(t) = b1

 b0 − X(t) dt + ΛdW (t), b1

where Λ is a 2-element row vector and W (t) = (W1 (t), W2 (t))′ is a two-dimensional Brownian motion where W1 (t) is independent from W2 (t). The money market account pays the short term rate, and therfore dS0 (t) = X(t)dt, S0 (t)

S0 (0) = s0

(3.19)

762

M Davis and S Lleo

The dynamics of the stock is dS(t) = (a + AX(t))dt + ΣdW (t), S(t)

S(0) = s,

(3.20)

where Σ is a 2-element row vector. Typically, we would view W1 (t) as the noise associated with the short term interest rate so that Λ = (λ, 0). The Brownian motion W2 (t) would then capture the idiosyncratic noise in the share. In fact, we could define a new Brownian motion WS (t) as  WS (t) = ρW1 (t) + 1 − ρ2 W2 (t) where ρ is the correlation coefficient of WS (t) and W1 (t), and express the dynamics of the stock as dS(t) = (a + AX(t))dt + σS dWS (t), S(t)

S(0) = s,

Based on Theorem 3.2, the investor will allocate a fraction to the Kelly portfolio with risky asset allocation

1 1−γ

(3.21)

of his/her wealth

ˆ hK (t) = (ΣΣ′ )−1 (a + AX(t)) and take a short position amounting to a fraction intertemporal hedging portfolio

1 1−γ

of his/her wealth in the

hI (t) = −(ΣΣ′ )−1 ΣΛ′ (Q(t)X(t) + q(t)) We compare the ICAPM with a naive implementation of the Merton model. The Merton model ignores the valuation factor X(t). One possibility would be to set the risk-free rate r to the long-term average money market rate bb01 so that b0 dS0 (t) = rdt = dt, S0 (t) b1

S0 (0) = s0

(3.22)

Similarly, the drift of the stock could be set to a long-term average µ := a + A bb10 , so that dS(t) = µdt + σS dWS (t), S(t)

S(0) = s,

(3.23)

Based on Theorem 2.14 hK (t) = (ΣΣ′ )−1 (µ − r1) ¯ In equilibrium, that is when X(t) = E[X(t)] =: X(t), the composition of the Kelly portfolio in the ICAMP and in the classical Merton model will coincide ¯ ¯ = (ΣΣ′ )−1 (µ − r1) hK (t) = (ΣΣ′ )−1 (a + AˆX(t)) so that the only difference in terms of allocation lies in the intertemporal hedging portfolio.

Chapter 37. Fractional Kelly Strategies in Continuous Time

763

3.2.2 Numerical Example We consider an example with b = 0.04, B = −1 and Λ = (0.08, 0), a = 0.056, A = 1.1 and Σ = (0.20), meaning that the noise related to the asset is perfectly positively correlated with the noise from the factor. A perfect correlation between asset and factor noise is unrealistic. However, this condition will help us observe more clearly the behaviour of the intertemporal hedging portfolio. Indeed, the role of the intertemporal hedging portfolio is to use the covariance structure of the asset and factors in order to adjust the risk of the portfolio. If the assets and factor noise are uncorrelated, the intertemporal hedging portfolio has no use and the investor will invest his money in the money market account, as shown in Corollary 3.3. In equilibrium, the (full) Kelly portfolio is invested at 150% in the stock, regardless of whether we are considering the classical Merton model or the ICAPM. Figure 1 displays the allocation of the intertemporal hedging portfolio to the risky stock for various levels of risk aversion. The proportion of the intertemporal hedging portfolio short in the stock is modest: it reaches 1.20% for a risk aversion γ = −2 and stands at 0.56% for a risk aversion γ = −8. With 10 years left on the investment horizon, the short position is at its highest. It then declines at an accelerating rate as the horizon draws near, to finally reach zero at the end of the horizon. This is in sharp contrast with the myopic Kelly portfolio which remains fully invested regardless of the investment horizon. Finally, the short position is inversely related to the aversion: the higher the risk aversion, the smaller the short position. As a result of the difference in magnitude between the Kelly portfolio and the intertemporal hedging portfolio, the proportion of the investor’s total wealth

Figure 1. Short position of the Intertemporal Hedging Portfolio (IHP) in the stock as a function of time to investment horizon for various levels of risk aversion (γ = -RSC) in the case of perfectly correlated asset and factors.

764

M Davis and S Lleo

Figure 2. Proportion of the investor’s wealth in the stock as a function of time to investment horizon for various levels of risk aversion (γ = -RSC) in the case of perfectly correlated asset and factors.

Figure 3. Proportion of the Kelly portfolio invested in the stock as a function of the value of the underlying factor X(t) in the case of perfectly correlated asset and factors.

invested in the risky stock remains relatively constant through the investment horizon, as shown in Figure 2. As the value X(t) of the factor varies away from equilibrium conditions, the investment of the Kelly portfolio in the ICAPM will deviate from the 150% level of the Merton model. Figure 3 shows that in the ICAPM, the allocation of the

Chapter 37. Fractional Kelly Strategies in Continuous Time

765

Figure 4. Short position of the Intertemporal Hedging Portfolio (IHP) in the stock as a function of the value of the underlying factor X(t) for various levels of risk aversion (γ = -RSC) in the case of perfectly correlated asset and factors.

Kelly portfolio increases linearly with the factor level. The equation of the line is hK = 1.40 + 2.5 × x, implying that the Kelly portfolio is highly leveraged. The proportion of the intertemporal hedging portfolio invested in the stock increases with the factor level, as a result of the perfect correlation between the factor noise, although it still remains modest even with a 10-year investment horizon (see Figure 4). Finally, the overall asset allocation is significantly influenced by both the increase in leverage in the Kelly portfolio resulting from a rise in the factor level and the dilution of the Kelly portfolio associated with an increase in the risk aversion (see Figure 5).

4 Benchmarked Investment Management 4.1 The Model Davis and Lleo [11] and [12] study a benchmarked asset management using risksensitive control. Merton’s ICAPM with no consumption and a power utility function is closely related to risk-sensitive asset management (see [11] and [12]) and therefore to Linear Exponential-of-Quadratic Gaussian (LEQG) stochastic control. The results obtained in [11] and [12] readily extend to a benchmarked version of Merton’s ICAPM, up to some minor changes in the value function. The objective of a benchmarked investor is to outperform a given investment benchmark, such as the S&P 500 or the Salomon Smith Barney World Government Bond Index. Davis and Lleo [11] model the evolution of the benchmark level using

766

M Davis and S Lleo

Figure 5. Proportion of the investor’s wealth in the stock as a function of the value of the underlying factor X(t) for various levels of risk aversion (γ = -RSC) in the case of perfectly correlated asset and factors.

the SDE dL(t) = (c + C ′ X(t))dt + ς ′ dW (t), L(t)

L(0) = l

(4.1)

where c is a scalar constant, C is a n-element column vector and ς is a N -element column vector. The objective of the investor is to maximize the expected utility of terminal outperformance J(t, x, h; T, γ):  γ ln FT   γ FT e =E J(t, x, h; T, γ) = E[U (FT )] = E γ γ where F (t, x; h) is defined as the (log) excess return of the investor’s portfolio over the return of the benchmark, i.e. F (t, x, h) := ln

V (t, x, h) L(t, x, h)

By Itˆo’s lemma, the log of the excess return in response to a strategy h is  t  t v d ln L(s) d ln V (s) − F (t, x; h) = ln + l 0 0  t v ˆ (a0 + A′0 X(s) + h(s)′ (ˆ a + AX(s)))ds = ln + l 0   t 1 t ′ ′ h(s)′ ΣdW (s) h(s) ΣΣ h(s)ds + − 2 0 0   t  t 1 t ′ ′ (c + C X(s))ds + ς ′ dW (s) (4.2) ς ςds − − 2 0 0 0

Chapter 37. Fractional Kelly Strategies in Continuous Time

767

where v l Following an appropriate change of measure along the line described in the previous sections, the criterion function under the new measure can be expressed as

 T f0γ h E exp γ I(t, x, h; T, γ) = g(Xs , h(s); γ)ds γ t,x t F (0, x; h) = f0 := ln

where 1 ˆ (1 − γ)h′ ΣΣ′ h − a0 − A′0 x − h′ (ˆ a + Ax) 2 1 + γh′ Σς + (c + C ′ x) + (1 − γ)ς ′ ς 2 The value function Φ for the auxiliary criterion function I(t, x; h; T, γ) is defined as g(x, h; γ) =

Φ(t, x) = sup I(t, x; h; T, γ)

(4.3)

h∈A(T )

Solving the control problem, we find that the optimal investment policy h∗ (t) is given by h∗ =

1 ˆ + γΣΛ′ DΦ − γΣς) (ΣΣ′ )−1 (ˆ a + Ax 1−γ

˜ x) = and the logarithmically-transformed value function Φ(t, given by

1 γ

(4.4) ln Φ(t, x) is still

Φ(t, x) = x′ Q(t)x + x′ q(t) + k(t) where Q(t) solves a n-dimensional matrix Riccati equation and q(t) solves a n-dimensional linear ordinary differential equation. The following fund separation theorem defines benchmarked fractional Kelly strategies: Theorem 4.1 (Benchmarked Mutual Fund Theorem). Given a time t and a state vector X(t), any portfolio can be expressed as a linear combination of investments into two “mutual funds” with respective risky asset allocations: ˆ a + AX(t)) hK (t) = (ΣΣ′ )−1 (ˆ hC (t) = (ΣΣ′ )−1 [Σς − ΣΛ′ (q(t) + Q(t)X(t))]

(4.5)

and respective allocation to the money market account given by: ′ ′ −1 ˆ hK (ˆ a + AX(t)) 0 (t) = 1 − 1 (ΣΣ ) ′ ′ −1 [Σς − ΣΛ′ (q(t) + Q(t)X(t))] hC 0 (t) = 1 − 1 (ΣΣ )

Moreover, if an investor has a risk aversion γ, the respective weights of each mutual γ 1 fund in the investor’s portfolio are equal to 1−γ and − 1−γ . This result splits the benchmarked fractional Kelly strategy in three “funds”:

768

M Davis and S Lleo

1 to the Kelly portfolio. As γ → 0, this investors’s allocation 1. An allocation of 1−γ to the Kelly portfolio converges to 100% of his/her wealth, implying that a Kelly investor will ignore the benchmark and focus solely on growth maximization; 2. An allocation to the intertemporal hedging portfolio. Note that the intertemporal hedging portfolio in the benchmarked case is subtly different from its counterpart in the asset-only case. Indeed, the coefficient q solve a slightly different linear ODE which includes terms related to the benchmark dynamics; 3. An allocation to a benchmark-tracking portfolio. The allocation of this portfolio, (ΣΣ′ )−1 Σς, is in fact a projection of the benchmark risk on the subspace spanned by asset risk, that is an unbiased estimator of a linear relationship between asset risks and benchmark risk ς = Σ′ u;

4.2 Example: Replicable Benchmark We start with a similar setting as in Section 3.2, namely a one-factor model where the single factor is a short term interest rate behaving according to the Vasicek model: dX(t) = (b0 − b1 X(t))dt + ΛdW (t),

X(0) = x

(4.6)

where Λ is now a 3-element row vector and W (t) = (W1 (t), W2 (t), W3 (t))′ is a threedimensional Brownian motion. W1 (t), W2 (t) and W3 (t) are independent standard Brownian motions on (Ω, {Ft }, F , P). For numerical applications, we still take b = 0.04, B = −1 and Λ = (0.08, 0). The money market account pays the short term rate, and therfore dS0 (t) = X(t)dt, S0 (t)

S0 (0) = s0

(4.7)

The investor can buy either of two stocks S1 and S2 with respective time-t price S1 (t) and S2 (t). The dynamics of the price vector S(t) := (S1 (t), S2 (t)) is given by equation (3.2), that is dS(t) = D(S(t))(a + AX(t))dt + D(S(t))ΣdW (t),

S(0) = s,

In our example, we will take 

0.018 a= 0.064

 1.3 A= 0.9 and Σ=



0.108 0.0523 0 0.064 0 0.196

Hence, the drift of stock S1 is more sensitive to changes in the factor than the drift of stock S2 . To understand the relation between the diffusion of the stock price and

Chapter 37. Fractional Kelly Strategies in Continuous Time

769

the diffusion of the factor, we define two new Brownian motions:  1 WS (t) = ρ1 W1 (t) + 1 − ρ21 W2 (t)  WS2 (t) = ρ2 W1 (t) + 1 − ρ22 W3 (t)

The Brownian motions WS1 (t) and WS2 (t) represent the respective noise associated with stock S1 and S2 . They are correlated with the factor noise W1 (t), with correlation coefficient ρ1 = 0.9 and ρ2 = 0.2. Thus, stock S1 has a lower volatility and a higher correlation with the factor than stock S2 . The investor’s benchmark is an index with a 60% allocation to stock S1 and a 40% allocation to S2 . The benchmark is therefore easily replicable. Let wB = (60%, 40%)′ be the vector of benchmark weights, then the parameters of the benchmark’s ′ ′ ′ a, C = wB A and ς = wB Σ. SDE (4.1) are c = wB ¯ In equilibrium, that is when X(t) = X(t) = 0.04, the (full) Kelly portfolio is invested at 168.83% in stock S1 and 131.71% in stock S2 . The Kelly portfolio is myopic: its allocation does not change as the investment horizon changes. It only varies with the factor level. As expected, the allocation to the benchmark replicating portfolio equals (ΣΣ′ )−1 Σς = (60%, 40%) = wB irrespective of the time horizon or factor level: we can (and do) fully replicate the benchmark. Figure 6 displays the allocation of the intertemporal hedging portfolio to the two risky stock for various levels of risk aversion. The proportion of the intertemporal hedging portfolio short in either stock is still modest: it reaches 3.68% in stock S1 and 0.01% in stock S2 for a risk aversion γ = −2. The evolution of the allocation is similar to what we already observed in the asset only case: the short position is at its highest level when the investment horizon is at 10 years, and it then declines at an accelerating rate to finally reach zero at the end of the horizon. Figure 7 shows that the overall asset allocation is dominated by the Kelly portfolio at low risk-aversion levels and by the benchmark replicating portfolio at high levels of risk aversion. As a result, the proportion of the investor’s total wealth invested in the stocks remains relatively constant through the investment horizon. Keeping the time horizon fixed at ten years, and letting the factor level vary from 0 to 20%, we see that the allocation of the Kelly portfolio to the two stocks K evolves in opposite directions (see Figure 8). The asset allocation hK 1 and h2 to the two stocks are an affine function of the factor level x: hK 1 = 0.7960 + 22.3066x hK 1 = 1.5134 − 4.9071x

(4.8)

Not that the sign of the slope and of the intercept reflect the sign of the entries of vectors Aˆ and a ˆ. In our example, Aˆ1 = 0.3 while Aˆ2 = −0.1. The allocation to the benchmark replicating portfolio remains fixed at 60% in stock S1 and 40% in stock S2 . The proportion of the intertemporal hedging portfolio short stock S1 and S2 increases with the factor level (see Figure 8). The short position for the highly correlated stock S1 reaches 13.85% for a factor level of 20%

770

M Davis and S Lleo

(Short posiion in stock 1)

(a) (Short posiion in stock 2)

(b) Figure 6. Short position of the Intertemporal Hedging Portfolio (IHP) in the stock as a function of time to investment horizon.

and a risk aversion γ = −2. By comparison, the short position in less correlated stock S2 is only 0.36% for the same risk factor and risk aversion levels. Overall, the asset allocation is dominated by a combination of the Kelly portfolio and the benchmark replicating portfolio (see Figure 9). Within the range of factor values considered, the allocation to both stocks remains above the benchmark level. This implies that the investor will use leverage, even at fairly high risk aversion levels.

Chapter 37. Fractional Kelly Strategies in Continuous Time

771

(a) Allocation to stock 1

(b) Allocation to stock 2 Figure 7. Proportion of the investor’s wealth in the stock as a function of time to investment horizon for various levels of risk aversion (γ = -RSC).

5 ICAPM with Partial Observation Davis and Lleo [13] propose an extension to the case where the factor process X(t) is not directly observed and the asset allocation strategy ht must be adapted to the filtration FtS = σ{Si (u), 0 ≤ u ≤ t, i = 0, . . . , m} generated by the asset price processes alone. In the partial observation case, the pair of processes (X(t), Y (t)), where Yi (t) = log Si (t), takes the form of the ‘signal’ and ‘observation’ processes

772

M Davis and S Lleo

Figure 8. Proportion of the Kelly portfolio invested in the stocks as a function of the value of the underlying factor X(t).

in a Kalman filter system, and consequently the conditional distribution of X(t) is ˆ ˆ normal N (X(t), P (t)) where X(t) = E[X(t)|FtS ] satisfies the Kalman filter equation and P (t) is a deterministic matrix-valued function. By using this idea we can obtain ˆ and the dynamic an equivalent form of the problem in which X(t) is replaced by X(t) ˆ equation (3.3) by the Kalman filter. Optimal strategies take the form h(t, X(t)). This very old idea in stochastic control goes back at least to Wonham [46]. To simplify the presentation of the idea, we will assume that A0 , implying the short-term interest rate is constant. The more general case is outlined in Davis and Lleo [13]. In the partial obsrevation case we need to assume that the initial value X0 of the factor process is a normal random vector N (m0 , P0 ) with known mean m0 and covariance P0 , and is independent of the Brownian motion W . The observation process is the log price process Y (t) = ln St :   N  1 ′ σik dWk (t), dYi (t) = (a + AX(t))i − ΣΣii dt + 2 k=1

Yi (0) := yi (0) = ln Si (0)

(5.1)

The processes (X(t), Y (t)) and the filtering equations are standard: Proposition 5.1 (Kalman Filter). The conditional distribution of X(t) given ˆ FtS is N (X(t), P (t)), calculated as follows. (i) The innovations process U (t) ∈ Rm defined by ˆ dU (t) = (ΣΣ′ )−1/2 (dY (t) − AX(t)dt),

U (0) = 0

(5.2)

is a vector Brownian motion. ˆ (ii) X(t) is the unique solution of the SDE ˆ ˆ dX(t) = (b + B X(t))dt + (ΛΣ′ + P (t)A′ )(ΣΣ′ )−1/2 dU (t),

ˆ X(0) = m0 . (5.3)

Chapter 37. Fractional Kelly Strategies in Continuous Time

773

(Short posiion in stock 1)

(a) (Short posiion in stock 2)

(b) Figure 9. Short position of the Intertemporal Hedging Portfolio (IHP) in the stock as a function of the value of the underlying factor X(t).

(iii) P (t) is the unique non-negative definite symmetric solution of the matrix Riccati equation P˙ (t) = ΛΞΞ′ Λ′ − P (t)A′ (ΣΣ′ )−1 AP (t) + (B − ΛΣ′ (ΣΣ′ )−1 A)P (t) + P (t)(B ′ − A′ (ΣΣ′ )−1 ΣΛ′ ), where Ξ := I − Σ′ (Σ′ Σ)−1 Σ.

P (0) = P0

774

M Davis and S Lleo

(a) Allocation to stock 1

(b) Allocation to stock 2 Figure 10. Proportion of the investor’s wealth in the stock as a function of the value of the underlying factor X(t).

The Kalman filter has replaced our initial state process X(t) by an estimate ˆ X(t) with dynamics given in (5.3). To recover the asset price process, we use (5.1) together with (5.2) to obtain the dynamics of Y (t) N

 1 ′ ˆ [(ΣΣ′ )1/2 ]ik dUk (t) dYi (t) = [a + AX(t)] i dt − ΣΣii dt + 2 k=1

(5.4)

Chapter 37. Fractional Kelly Strategies in Continuous Time

775

and then apply Itˆ o to Si (t) = exp Yi (t) to get ˆ dS(t) = D(S(t))(a + AX(t))dt + D(S(t))(ΣΣ′ )1/2 dU (t) Si (0) = si ,

i = 1, . . . , m

(5.5)

Now we can solve the stochastic control problem with partial observation as in Section 3 simply by replacing the original asset price description (3.2) by (5.5), and the factor process description (3.3) by the Kalman filter equation (5.3) (see [13] for the full detail). The optimal investment policy h∗ (t) h∗ (t) =

1 ˆ + γ(ΛΣ′ + P (t)A′ )′ (Q(t)X(t) ˆ + q(t))] (ΣΣ′ )−1 [ˆ a + AX(t) 1−γ

(5.6)

where Q(t) satisfies a matrix Riccati equation and q(t) satisfies a vector linear ODE. The following fund separation theorem defines fractional Kelly strategies subject to partial observation: Corollary 5.2 (Fund Separation Theorem — Partial Observation). Any portfolio can be expressed as a linear combination of investments into two funds with respective risky asset allocations: ˆ a + AX(t)) hK (t) = (ΣΣ′ )−1 (ˆ ˆ + q(t)) hIP O (t) = −(ΣΣ′ )−1 (ΣΛ′ + AP (t))(Q(t)X(t)

(5.7)

where P (t) solves the matrix Riccati equation (5.4) and hIP O (t) can be viewed as a full allocation to two risky portfolios, an intertemporal portfolio hI , and a partial observation portfolio hP O : ˆ + q(t)) hI (t) = −(ΣΣ′ )−1 ΣΛ′ (Q(t)X(t) ˆ + q(t)) hP O (t) = −(ΣΣ′ )−1 AP (t)(Q(t)X(t)

(5.8)

The funds have respective allocation to the money market account given by ′ K hK 0 (t) = 1 − 1 h (t) O (t) = 1 + 1′ hIP O (t) hIP 0

Moreover, if an investor has a risk aversion γ, then the respective weights of each γ 1 mutual fund in the investor’s portfolio equal 1−γ and − 1−γ , respectively. This result splits the fractional Kelly strategy in three “funds”: 1 1−γ

to the Kelly portfolio subject to partial observation, that ˆ is with the Kalman estimate X(t) instead of the true value X(t); 2. An allocation to the intertemporal hedging portfolio subject to partial observation. Here again the intertemporal hedging portfolio in the partial observation case is subtly different from its counterpart in the full observation case; 3. An allocation of to a partial observation portfolio; 1. An allocation of

776

M Davis and S Lleo

6 Fractional Kelly Strategies in a Jump-Diffusion Setting Optimal investment and consumption problems in a jump diffusion setting have been an active area of research since Merton [33] introduced the possibility of jumps in asset prices. Investigations have tended to concentrate on two axes: the mathematical resolution of the investment problem, and the economic implications of the jumps. Mathematical research has generally focused on the resolution of the HJB partial integro-differential equation, to the detriment of an analysis of the optima asset allocation. Recent references include Øksendal and Sulem [36], Barles and Imbert [2] and Bouchard and Touzi [7] in the general context of jumpdiffusion control, as well as Davis and Lleo [15] [16] [14] in the context of risksensitive asset management. A notable exception, A¨ıt-Sahalia, Cacho-Diaz and Hurd [1] proposed a method based on orthogonal decompositions to solve jumpdiffusion asset allocation problems. In an infinite-horizon setting and under specific assumptions on the jump structure, they derive an analytical solution and a fund separation. Financial economics research has used jumps to model to model market events and systemic shocks. Although the emphasis is decidedly on investment strategies, this stream of investigation generally takes place in specific models and tends to rely on numerical analysis rather than a mathematical derivation. Liu, Longstaff and Pan [24] use the event risk model proposed by Duffie Pan and Singleton [17] to study the impact of a jump in asset prices and volatility on investment policies. They find that the joint risk of an increase in volatility and of a jump in asset prices lead to conservative investment strategies with limited leveraged and short positions. Das and Uppal [10] model asset prices as jump-diffusion processes to investigate the impact of systemic risk, embodied in an appropriately calibrated jump component, on asset allocations. They conclude that the impact of a systemic event has less to do with reduced diversification benefits than with the large losses leveraged investors experience. Somewhere in between, Davis and Lleo [13] consider two classes of jump-diffusion factor models. The first has affine drift, constant diffusion and jumps in asset prices only. The second class of models accommodates factor dependent drift and diffusion as well as jumps in factor levels. For clarity, we will only present the affine model with diffusion factors in this section. 6.1 Setting Let (Ω, {Ft }, F , P) be the underlying probability space. On this space is defined an RM -valued (Ft )-Brownian motion W (t) with components Wk (t), k = 1, . . . , M . Moreover, let (Z, BZ ) be a Borel space.2 Let p be an (Ft )-adapted σ-finite Poisson point process on Z whose underlying point functions are maps from a countable set

2Z

is a Polish space and BZ is the Borel σ-field

Chapter 37. Fractional Kelly Strategies in Continuous Time

777

Dp ⊂ (0, ∞) into Z. Define Zp := {U ∈ B(Z), E[Np (t, U )] < ∞ ∀ t}

(6.1)

where Np (dt, dz) is the Poisson random measure on (0, ∞) × Z induced by p. Our analysis focuses on stationary Poisson point processes of class (QL) with associated Poisson random measure Np (dt, dz). The class (QL) is defined in [19] (Definition II.3.1 p. 59) as Definition 6.1. An (Ft )-adapted point process p on (Ω, F , P) is said to be of class ˆp = (N ˆp (t, U )) such (QL) with respect to (Ft ) if it is σ-finite and there exists N that ˆp (t, U ) is a continuous (Ft )-adapted increasing process; (i.) for U ∈ Zp , t → N ˆp (t, U ) is a σ-finite measure on (Z, B(Z)); (ii.) for each t and a.a. ω ∈ Ω, U → N ˜p (t, U ) = Np (t, U ) − N ˆp (t, U ) is an (Ft )-martingale; (iii.) for U ∈ Zp , t → N ˆ p (t, U )} is called the compensator of the point process p. The random measure {N Because the Poisson point processes we consider are stationary, then their comˆp (t, U ) = ν(U )t where ν is the σ-finite characteristic pensators are of the form N measure of the Poisson point process p. Finally, for notational convenience, we fix throughout the paper a set Z0 ∈ BZ ¯p (dt, dz) as such that ν(Z\Z0 ) < ∞ and define the Poisson random measure N ¯p (dt, dz) N  ˆp (dt, dz) = Np (dt, dz) − ν(dz)dt =: N ˜p (dt, dz) if z ∈ Z0 Np (dt, dz) − N = Np (dt, dz) if z ∈ Z\Z0 The dynamics of the n factors X(t) and of the wealth invested in the money market account S0 (t) are the same as in the diffusionc ase: dX(t) = (b + BX(t))dt + ΛdW (t),

X(0) = x

(6.2)

where X(t) is the Rn -valued factor process with components Xj (t) and b ∈ Rn , B ∈ Rn×n and ΛRn×M , dS0 (t) = (a0 + A′0 X(t))dt, S0 (t)

S0 (0) = s0

(6.3)

where a0 ∈ R is a scalar constant, A0 ∈ Rn is a n-element column vector and throughout the paper x′ denotes the transpose of the matrix or vector x. The securities prices can exhibit jumps: let Si (t) denote the price at time t of the ith security, with i = 1, . . . , m. The dynamics of risky security i can be expressed as: N

 dSi (t) = (a + AX(t))i dt + σik dWk (t) + − Si (t ) k=1

Si (0) = si ,

i = 1, . . . , m



¯ p (dt, dz), ηi (z)N

Z

(6.4)

778

M Davis and S Lleo

where a ∈ Rm , A ∈ Rm×n , Σ := [σij ], i = 1, . . . , m, j = 1, . . . , M and γ(z) ∈ Rm satisfying Assumption 6.2: Assumption 6.2. η(z) ∈ Rm satisfies −1 ≤ ηimin ≤ ηi (z) ≤ ηimax < +∞,

i = 1, . . . , m

and −1 ≤ ηimin < 0 < ηimax < +∞,

i = 1, . . . , m

for i = 1, . . . , m. Furthermore, define S := supp(ν) ∈ BZ and ˜ := supp(ν ◦ γ −1 ) ∈ B(Rm ) S where supp(·) denotes the measure’s support, then we assume that ˜ is the smallest closed hypercube containing S. In addition, the vector-valued function γ(z) satisfies:  |η(z)|2 ν(dz) < ∞

m

min , γimax ] i=1 [γi

(6.5)

Z0

Note that Assumption 6.2 requires that each asset has, with positive probability, both upward and downward jump. The effect of this assumption is to bound the space of controls. Relation (6.5) is a standard condition. See Definition II.4.1 in Ikeda and Watanabe [19].3 Define the set J as ˜ J := {h ∈ Rm : −1 − h′ ψ < 0 ∀ψ ∈ S}

(6.6)

For a given z, the equation h′ γ(z) = −1 describes a hyperplane in Rm . J is a convex subset of Rm for all (t, x) ∈ [0, T ] × Rn . Let Gt := σ((S(s), X(s)), 0 ≤ s ≤ t) be the sigma-field generated by the security and factor processes up to time t. Definition 6.3. An Rm -valued control process h(t) is in class H if the following conditions are satisfied: 1. h(t) is progressively measurable with respect to {B([0, t]) ⊗ Gt }t≥0 and is c`adl` ag; T 2. P ( 0 |h(s)|2 ds < +∞) = 1, ∀T > 0; 3. h′ (t)γ(z) > −1, ∀t > 0, z ∈ Z, a.s. dν. 3 In

[19], FP and F2,loc are respectively given in equations II(3.2) and II(3.5) P

Chapter 37. Fractional Kelly Strategies in Continuous Time

779

Define the set K as K := {h ∈ H : h ∈ J

for a.a.t}

(6.7)

Definition 6.4. A control process h(t) is in class A(T ) if the following conditions are satisfied: 1. h ∈ H; 2. EχhT = 1 where χht , t ∈ (0, T ], is the Dol´eans exponential defined as   t  1 2 t ′ h h(s)′ ΣΣ′ (h(s)ds h(s) ΣdWs − γ χt := exp γ 2 0 0  t ˜ (ds, dz) ln(1 − G(z, h(s)))N + 0 Z   t {ln(1 − G(z, h(s))) + G(z, h(s))}ν(dz)ds , + 0

(6.8)

Z

and

G(z, h) = 1 − (1 + h′ η(z))γ

(6.9)

Definition 6.5. We say that a control process h(t) is admissible if h(t) ∈ A(T ). Taking this budget equation into consideration, the wealth, V (t) of the investor in response to an investment strategy h(t) ∈ H, follows the dynamics dV (t) ˆ = (a0 + A′0 X(t))dt + h′ (t)(ˆ a + AX(t))dt + h′ (t)ΣdWt V (t− )  ¯p (dt, dz) h′ (t)η(z)N +

(6.10)

Z

where a ˆ := a − a0 1, Aˆ := A − 1A′0 , 1 ∈ Rm denotes the m-element unit column vector and with initial endowment V (0) = v. 6.2 Control Problem The objective of an investor is to maximize the expected utility of terminal wealth over a fixed time horizon T :  γ ln VT   γ VT e =E J(t, h; T, γ) = E[U (VT )] = E γ γ By Itˆo,    t γ ln V (t) γ e g(Xs , h(s))ds χht (6.11) = v exp γ 0

where

1 ˆ a + Ax) g(x, h) = − (1 − γ)h′ ΣΣ′ h + a0 + A′0 x + h′ (ˆ 2    1 [(1 + h′ η(z))γ − 1] + h′ η(z)1Z0 (z) ν(dz) + γ Z

and the Dol´eans exponential χht is given by (6.8).

(6.12)

780

M Davis and S Lleo

Let Ph be the measure on (Ω, FT ) defined via the Radon-Nikod´ ym derivative dPh := χhT (6.13) dP For a change of measure to be possible, we must ensure that the following technical condition holds: G(z, h(s)) < 1 This condition is satisfied iff h′ (s)η(z) > −1

(6.14)

a.s. dν, which was already required for h(t) to be in class H (Condition 3 in Definition 6.3). Thus Ph is a probability measure for h ∈ A(T ). For h ∈ A(T ), Moreover,  t h Σ′ h(s)ds Wt = Wt − γ 0

is a standard Ph -Brownian motion and we define the Ph -compensated Poisson random measure by  t  t  t ˜ h (ds, dz) = N (ds, dz) − {1 − G(z, h(s))}ν(dz)ds N 0

0

Z0

=

0

0

Z0

 t

N (ds, dz) −

0

Z0

Z0

 t

{(1 + h′ η(z))γ }ν(dz)ds

Z0

As a result, X(s), 0 ≤ s ≤ t satisfies the SDE: dX(s) = f (X(s), h(s))ds + ΛdWsh

(6.15)

f (x, h) := b + Bx + γΛΣ′ h

(6.16)

where

Applying the change of measure argument, we obtain the criterion function under the measure Ph :

 T vγ h E exp γ g(Xs , h(s))ds (6.17) I(t, x, h; T ; γ) = γ t,x t The value function Φ for the auxiliary criterion function I(t, x, h; T ; γ) is defined as Φ(t, x) = inf I(t, x, h; T ; γ) h∈A

(6.18)

The corresponding HJB PIDE is 1 ∂Φ (t, x) + tr(ΛΛ′ D2 Φ(t, x)) + H(x, Φ, DΦ) ∂t 2 subject to terminal condition Φ(T, x) =

vγ γ

(6.19)

(6.20)

Chapter 37. Fractional Kelly Strategies in Continuous Time

781

where for r ∈ R, p ∈ Rn H(x, r, p) = inf {f (x, h)′ p − γg(x, h)r} h∈J

(6.21)

In this setting, the change of measure argument has reduced the problem to solving a stochastic control problem in the factor process, which has no jumps. As a result, we only need to solve a parabolic PDE, not a Partial Integro-Differential Equation (PIDE), which makes it easier to show that the HJB PDE admits a unique classical solution (see [15] for the full detail). To obtain the optimal control we will introduce the associated risk-sensitive ˜ x) := 1 ln Φ(t, x) with associated HJB PIDE value function Φ(t, γ ˜ ∂Φ ˜ D2 Φ) ˜ =0 (t, x) + sup Lh (x, DΦ, ∂t h∈J

(6.22)

γ 1 Lh (x, p, M ) = f (x, h)′ p + tr(ΛΛ′ M ) + p′ ΛΛ′ p − g(x, h) 2 2

(6.23)

where

and subject to the terminal condition 1 ˜ Φ(T, x) = ln v − ln γ, γ

x ∈ Rn .

(6.24)

The supremum in (6.22) can be expressed as γ 1 sup Lh (x, p, M ) = (b + Bx)′ p + tr(ΛΛ′ M ) + p′ ΛΛ′ p + a0 + A0 x 2 2 h∈Rm  1 ˆ (γ − 1)h′ ΣΣ′ h + γh′ ΣΛ′ p + h′ (ˆ + sup a + Ax) 2 h∈J   1 + (6.25) {[(1 + h′ η(z))γ − 1] − γh′ η(z)1Z0 (z)} γ Z Proposition 6.6. The supremum in (6.22)–(6.23) admits a unique Borel measurˆ x, p) for (t, x, p) ∈ [0, T ] × Rn × Rn . able maximizer h(t, Proof. See Section 3.3 in [15]. 6.3 Fractional Kelly Strategy The basic idea in a Fund Separation Theorem is that any admissible portfolio h(t) satisfying Definition 6.4 can be expressed as a linear combination of two (or more) portfolios. For example, one could find two portfolios A and B with respective asset allocations hA and hB such that h(t) = wA hA (t) + (1 − wA )hB (t) This decomposition is not unique. Focusing on optimal controls h∗ (t) only, we could 1 to guarantee a decomposition between the Kelly portfolio hK and take wA = 1−γ

782

M Davis and S Lleo

an intertemporal portfolio hIγ h∗ (t) =

1 γ hK (t) + hI (t) 1−γ γ−1 γ

The intertemporal hedging portfolio hIγ depends indirectly on the risk aversion level γ and as such is not universal. Theorem 6.7 (Fund Separation Theorem in a Jump-Diffusion Setting). The optimal asset allocation for an investor with a risk aversion γ can be expressed as a linear combination of an investment in the Kelly criterion (log-utility) portfolio and in an intertemporal hedging portfolio: h∗ (t) =

γ 1 hK (t) − hI (t) 1−γ 1−γ

where the risky allocation hK to the Kelly criterion (log-utility) portfolio solves the fixed point problem      η(z) ′ −1 ˆ (ˆ a + Ax) + (6.26) h = (ΣΣ ) − η(z)1Z0 (z) ν(dz) 1 + h′ η(z) Z and the risky allocation to the intertemporal hedging portfolio hIγ satisfies: hIγ (t) =

γ−1 ∗ 1 h (t) − hK (t) γ γ

The respective allocation to the money market account for each of the two funds are given by K hK 0 (t) = 1 − h (t)

hI0 (t) = 1 − hIγ (t) Proof. Consider the auxiliary function ℓ defined as ℓ(h; x, p) =

1 ˆ (1 − γ)h′ ΣΣ′ h − γh′ ΣΛ′ p − h′ (ˆ a + Ax) 2  1 − {[(1 + h′ η(z))γ − 1] − γh′ η(z)1Z0 (z)}ν(dz) γ Z

(6.27)

for h ∈ J , x ∈ Rn , p ∈ Rn .  A Taylor expansion of the integral term γ1 Z {[(1 + h′ η(t, z))γ − 1]}ν(dz) around h = 0 yields  1 {[(1 + h′ η(z))γ − 1]}ν(dz) γ Z   γ−1 ′ (γ − 1)(γ − 2) ′ (h η(z))2 + (h η(z))3 + · · · h′ η(z) + = 2 3! Z  (γ − 1)(γ − 2) . . . (γ − 1)(γ − k + 1) ′ k (h η(z)) + · · · ν(dz) + k!

Chapter 37. Fractional Kelly Strategies in Continuous Time

783

Taking the limit as γ → 0, we obtain:   1 lim {[(1 + h′ η(z))γ − 1]}ν(dz) = {ln[1 + h′ η(z)]}ν(dz) γ→0 γ Z Z

We now define the function ℓK (h; x, p) as the limit of ℓ as γ → 0: ℓK (h; x, p) = lim ℓ(h; x, p) γ→0

=

1 ′ ′ ˆ h ΣΣ h − h′ (ˆ a + Ax) 2  − {ln[1 + h′ η(z)] − h′ η(t, z)1Z0 (z)}ν(dz)

(6.28)

Z n

for h ∈ Rm , x ∈ Rn and p ∈ R . The Kelly allocation is the unique minimizer of (6.28). Applying the first order condition, we conclude that hk satisfies ∂ℓK =0 ∂h  

 η(z) − η(z)1Z0 (z) ν(dz) = 0 1 + h′ η(z) Z      η(z) ′ −1 ˆ − η(z)1Z0 (z) ν(dz) (ˆ a + Ax) + ⇔ h = (ΣΣ ) 1 + h′ η(z) Z ˆ − ⇔ ΣΣ h − (ˆ a + Ax) ′



It is difficult to get much more intuition regarding the behaviour of the intertemporal hedging portfolio due to its dependence on the risk-aversion γ, the asset price ˜ Practically, one can estimate the intertemporal jumps and the value function Φ. hedging portfolio through numerical methods such as a policy improvement scheme (see Bellman [4], Fleming and Richel [18] and Davis and Lleo [15]) or a finite difference method (see for example Kushner and Dupuis [22]). Davis and Lleo [13] show how to adapt a policy improvement scheme to compute numerically the allocation to the intertemporal hedging portfolio. Corollary 6.8 (Three Fund Separation Theorem in a Jump-Diffusion Setting). The optimal asset allocation for an investor with a risk aversion γ can be expressed as a linear combination of an investment in the Kelly criterion (log-utility) portfolio and in an intertemporal hedging portfolio: h∗ (t) =

γ 1 hK (t) − hI (t) 1−γ 1−γ

The risky allocation hK to the Kelly criterion (log-utility) portfolio solves the fixed point problem      η(z) ˆ + − η(z)1 a + AX) (z) ν(dz) (6.29) h = (ΣΣ′ )−1 (ˆ Z0 1 + h′ η(z) Z Alternatively, the risky allocation hK to the Kelly criterion (log-utility) portfolio can be expressed as a decomposition bewteen the standard Kelly allocation hKD and a

784

M Davis and S Lleo

jump-related portfolio hKJ : hK = hKD + hKJ

(6.30)

where the standard Kelly allocation hKD is given by ˆ hKD = (ˆ a + AX) and the jump-related portfolio hKJ solves the fixed point problem    η(z) − η(z)1 h = (ΣΣ′ )−1 (z) ν(dz) + hKS Z0 ′ η(z) 1 + h Z

(6.31)

(6.32)

The risky allocation to the intertemporal hedging portfolio hIγ can be expressed as ¯ I (t) hIγ (t) = hγ (t) + h that is an allocation between ¯ I ∈ S staisfying i. the fully risk-averse portfolio h ˜I hI (t) = −(ΣΣ′ )−1 ΣΛ′ DΦ

(6.33)

˜ and with S := {h ∈ J : h′ ψ ≥ 0 ∀ψ ∈ S} γ ii. the risk-aversion induced portfolio h . The respective allocation to the money market account for each of the two funds are given by K hK 0 (t) = 1 − h (t)

hI0 (t) = 1 − hIγ (t)

7 Conclusion In this chapter, we have presented an overview of some recent developments related to Kelly investment strategies. In particular, we showed how the definition of fractional Kelly strategies can be extended to guarantee optimality when asset prices are no longer lognormally distributed. The key idea is to get the definition of fractional Kelly strategies to correspond with the fund separation theorem related to the problem at hand. In these instances, fractional Kelly investment strategies appear as the natural solution for investors seeking to maximize the terminal power utility of their wealth.

References [1] Yacine A¨ıt-Sahalia, Julio Cacho-Diaz, and T. R. Hurd. Portfolio choice with jumps: a closed-form solution. The Annals of Applied Probability, 19(2): 556–584, 2009. [2] G. Barles and C. Imbert. Second-order elliptic integro-differential equations: Viscosity solutions’ theory revisited. Annales de l’Institut Henri Poincar´e, 25(3):567–585, 2008.

Chapter 37. Fractional Kelly Strategies in Continuous Time

785

[3] D. Becherer. The numeraire portfolio for unbounded semimartingales. Finance and Stochastics, 5(3):327–341, July 2001. [4] R. Bellman. Dynamic Programming. Princeton University Press, 1957. [5] T.R. Bielecki and S.R. Pliska. Risk-sensitive dynamic asset management. Applied Mathematics and Optimization, 39:337–360, 1999. [6] T.R. Bielecki and S.R. Pliska. Risk sensitive intertemporal CAPM. IEEE Transactions on Automatic Control, 49(3):420–432, March 2004. [7] B. Bouchard and N. Touzi. Weak dynamic programming principle for viscosity solutions. SIAM Journal on Control and Optimization, 49:948–962, 2011. [8] L. Breiman. Optimal gambling system for favorable games. In Proceedings of the 4th Berkeley Symposium on Mathematical Statistics and Probability, pages 63–68, 1961. [9] D. Bernoulli (Translated by L. Sommer). Exposition of a new theory on the measurement of risk. Econometrica, 22:22–36, 1954. [10] S.R. Das and R. Uppal. Systemic risk and international portfolio choice. Journal of Finance, 59(6):2809–2834, 12 2004. [11] M.H.A. Davis and S. Lleo. Risk-sensitive benchmarked asset management. Quantitative Finance, 8(4):415–426, June 2008. [12] M.H.A. Davis and S. Lleo. Fractional kelly strategies for benchmarked asset management. In L.C. MacLean, E.O. Thorp, and W.T. Ziemba, editors, The Kelly Capital Growth Investment Criterion: Theory and Practice. World Scientific, 2010. [13] M.H.A. Davis and S. Lleo. On the optimality of kelly strategies i: Theory. working paper, 2010. [14] M.H.A. Davis and S. Lleo. Risk-sensitive asset management and affine processes. In Masaaki Kijima, Chiaki Hara, Keiichi Tanaka, and Yukio Muromachi, editors, KIER-TMU International Workshop on Financial Engineering 2009, pages 1–43. World Scientific Publishing Co, 2010. [15] M.H.A. Davis and S. Lleo. Jump-diffusion risk-sensitive asset management i: Diffusion factor model. SIAM Journal on Financial Mathematics, 2:22–54, 2011. [16] M.H.A. Davis and S. Lleo. Risk-sensitive asset management in a jump-diffusion factor model. Working Paper, 2011. [17] D. Duffie, J. Pan, and K. Singleton. Transform analysis and asset pricing for affine jump-diffusions. Econometrica, 68(6):1343–1376, 2000. [18] W.H. Fleming and R.W. Rishel. Deterministic and Stochastic Optimal Control. Springer-Verlag, Berlin, 1975. [19] N. Ikeda and S. Watanabe. Stochastic Differential Equations and Diffusion Processes. North-Holland Publishing Company, 1981. [20] J. Kelly. A new interpretation of the information rate. Bell System Technical Journal, 35:917–926, 1956. [21] K. Kuroda and H. Nagai. Risk-sensitive portfolio optimization on infinite time horizon. Stochastics and Stochastics Reports, 73:309–331, 2002.

786

M Davis and S Lleo

[22] H.J. Kushner and P.G. Dupuis. Numerical Methods for Stochastic Control Problems in Continuous Time, volume 24 of Stochastic Modeling and Applied Probability. Springer-Verlag, 2 edition, 2001. [23] H. Latan´e. Criteria for choice among risky ventures. Journal of Political Economy, 67:144–155, 1959. [24] J. Liu, F. A. Longstaff, and J. Pan. Dynamic asset allocation with event risk. Journal of Finance, 68(1):231–259, 2 2003. [25] M. Loeve. Probability Theory. 2 edition, 1960. [26] J.B. Long. The numeraire portfolio. Journal of Financial Economics, 26:29–69, 1990. [27] L. MacLean, E. Thorp, and W.T. Ziemba. Long-term capital growth: The good and bad properties of the kelly criterion. Quantitative Finance, 10(7):681–687, September 2010. [28] L. MacLean, W.T. Ziemba, and Y. Li. Time to wealth goals in capital accumulation and the optimal trade-off of growth vesrus security. Quantitative Finance, 5(4):343–357, 2005. [29] L.C. MacLean, R. Sanegre, Y. Zao, and W.T. Ziemba. Capital growth with security. Journal of Economic Dynamics and Control, 28(5):937–954, 2004. [30] L.C. MacLean, E. Thorp, and W.T. Ziemba, editors. The Kelly Capital Growth Investment Criterion: Theory and Practice. World Scientific, 2010. [31] L.C. MacLean, W.T. Ziemba, and G. Blazenko. Growth versus security in dynamic investment analysis. Management Science, 38(11):1562–1585, November 1992. [32] H. Markowitz. Investment for the long run: New evidence for an old rule. Journal of Finance, 36(5):495–508, 1976. [33] R.C. Merton. Optimal consumption and portfolio rules in a continuous-time model. Journal of Economic Theory, 3:373–4113, 1971. [34] R.C. Merton. An intertemporal capital asset pricing model. Econometrica, 41(866–887), 1973. [35] B. Øksendal. Stochastic Differential Equations. Universitext. Springer-Verlag, 6 edition, 2003. [36] B. Øksendal and A. Sulem. Applied Stochastic Control of Jump Diffusions. Springer, 2005. [37] E. Platen. A benchmark approach to finance. Mathematical Finance, 16(1):131–151, January 2006. [38] E. Platen and D. Heath. A Benchmark Approach to Quantitative Finance. Springer Finance. Springer-Verlag, 2006. [39] W. Poundstone. Fortune’s Formula: The Untold Story of the Scientific System That Beat the Casinos and Wall Street. Hill and Wang, 2005. [40] P. Samuelson. Lifetime portfolio selection by dynamic stochastic programming. Review of Economics and Statistics, 51(3):239–246, 1969. [41] P. Samuelson. Why we should not make mean log of wealth big though years to act are long. Journal of Banking and Finance, 3:305–307, 1979.

Chapter 37. Fractional Kelly Strategies in Continuous Time

787

[42] P.A. Samuelson. The “fallacy” of maximizing the geometric mean in long sequences of investing or gambling. In Proceedings of the National Academy of Sciences of the United States of America, volume 68, pages 2493–2496, 1971. [43] E. Thorp. Portfolio choice and the kelly criterion. In Proceedings of the Business and Economics Section of the American Statistical Association, pages 215–224, 1971. [44] E. Thorp. The kelly criterion in blackjack, sports betting and the stock market. In S.A. Zenios and W.T. Ziemba, editors, Handbook of Asset and Liability Management, volume 1 of Handbook in Finance. North Holland, 2006. [45] O. Vasicek. An equilibrium characterization of the term structure. Journal of Financial Economics, 5:177–188, 1977. [46] W.M. Wonham. On the separation theorem of stochastic control. SIAM Journal of Control and Optimization, 6(2):312–326, 1968. [47] R. Ziemba and W.T. Ziemba. Scenarios for Risk Management and Global Investment Strategies. John Wiley & Sons, 2007. [48] W.T. Ziemba. The symmetric downside-risk sharpe ratio. Journal of Portfolio Management, 32(1):108–122, Fall 2005.

This page intentionally left blank

Chapter 38

Growth-optimal investments and numeraire portfolios under transaction costs Wael Bahsouna, Igor V. Evstigneevb and Michael I. Taksarc

Abstract The aim of this work is to extend the capital growth theory (Kelly, Latane, Breiman, Cover, Ziemba, Thorp, MacLean, Browne, Platen, Heath, and others) to asset market models with transaction costs. \Ve define a natural generalization of the notion of a numeraire portfolio proposed by Long and show how such portfolios can be used for constructing growth-optimal investment strategies. The analysis is based on the classical von Neumann-Gale model of economic dynamics, a stochastic version of which we use as a framework for the modeling of financial markets with frictions. Key words: capital growth theory, transaction costs, numeraire portfolios, random dynamical systems, convex multivalued operators, von Neumann-Gale model, rapid paths 2000 Mathematics 8-nbject Classification: 37H99, 37H15, 91B62, 91B28. JEL-Classification: C61, C62, 041, G10.

School of Mathematics, Loughborough University, Leicestershire, LEll 3TU, UK. Economics Department, University of :\1anchester, Oxford Road, Manchester M13 9PL, UK. E-mail: [email protected]. Phone: 0161-2754275. Fax: 0161-2754812. (Corresponding author.) c Mathematics Department, University of Missouri, Columbia, :\10 65211, USA. a

6

1

789

790

1

W Bahsoun, IV Evstigneev and MI Taksar

Introduction

How to invest in order to achieve the maximum growth rate of wealth in the long run? This question has been central to the literature on capital growth theory: Kelly [32], Latan6 [33], Brcirnan [6], Thorp [53], Ziemba and Vickson [54, 55, 56], Algoet and Cover [2], MacLean et al. [35, 36, 37], Hakansson and Ziemba [24], Browne [7, 8], Platen and Heath [44], and others 1 . For the most part, results available in this literature pertain to markets without transaction costs. Up to now, only some specialized models of markets with frictions have been analyzed in this field; sec e.g. Hausch and Ziemba [25], Taksar ct al. [52], Iyengar and Cover [26], Akian ct al. [1], and Iyengar [27]. The goal of the present work is to develop a capital growth theory within a general discrete-time framework taking into account proportional transaction costs. Our main tool in this study is one of the fundamental models in mathematical economics the von Neumann-Gale model of economic growth. The mathematical framework ofthc von Neumann-Gale model is a special class of multivalued dynamical systerns possessing certain properties of convexity and homogeneity. The original theory of such systems (von Neumann [57], Gale [23], Rockafellar [49], Makarov and Rubinov [39]) aimed basically at the modeling of economic dynamics. Initially, this theory was purely deterministic; it did not reflect the influence of random factors on economic growth. The importance of taking these factors into account was realized early on. First attempts of constructing stochastic analogues of the von Neumann-Gale model were undertaken in the 1970s by Dynkin [14, 15, 16], Radner [45, 46] and their research groups. However, the first attack on the problem left many questions unanswered. Studies in this direction faced serious mathematical difficulties. To overcome these difficulties, new mathematical techniques were required, that were developed only during the last decade-see [19, 21] and [5]. In a recent work of Dempster et al. [11], it has been observed that stochastic analogues of von Neumann-Gale dynamical systems provide a natural and convenient framework for the analysis of some fundamental problems in finance (asset pricing and hedging under transaction costs). This paper focuses on a different area of applications of such systems in finance. It demonstrates how methods and concepts developed in the context of von Neumann-Gale dynamics can be applied to the analysis of growth optimal investments under 1 The

state of the art in the field is reviewed in }vfacLean et al. [38].

2

Chapter 38. Growth-Optimal Investments and Numeraire Portfolios

791

transaction costs. A central notion related to von Neumann-Gale dynamical systems, that of a rapid path, plays a crucial role in this work. \Ve show that it yields a generali11ation of the concept of a rmmemire portfolio (Long [34]) suitable for the analysis of markets with transaction costs and trading constraints. We obtain results on the existence of asymptotically optimal trading strategies in markets with transaction costs by using results [5, 21] on the existence of rapid paths in von Neumann-Gale systems. The theory of von Neumann-Gale dynamica.l systems is one of the highlights of mathematical economics. The results we refer to combine advanced methods of ergodic theory, stochastic processes and functional analysis. In this paper, we concentrate only on the modeling issues and the applications in finance. The reader is referred to the literature cited for the proofs of the mathematical results employed in this work. The main goal of this article is to attract attention of theorists and practitioners working in quantitative finance to new powerful methods developed in the field. The paper is organized as follows. In Section 2 we describe the dynamic securities market model we deal with. Section 3 introduces the basic concepts and results related to the von Neumann-Gale dynamical systems. In Section 4 we apply these results to the analysis of capital growth under transaction costs. Section 5 discusses unsolved problems and directions for further research.

2

Dynamic securities market model

Let s 0 , s 1 , ... be a stochastic process with values in a measurable spaceS. The process ( ウエIゥ]セ@ models random factors influencing the market: the random element St represents the "state of the world" at date t = 0, 1, .... We denote by st := (so, s 1 , ... , St) the history of the process (st) up to date t. There are n assets traded in the market. A (contingent) portfolio of assets held by an investor at date t is represented by a vector

whose coordinates (portfolio positions) describe the holdings of assets i 1, 2, ... , n. The positions can be described either in terms of "physical units" of assets or in terms of their market values. A contingent portfolio Xt(st) depends generally on the whole history st of the process (st), which means that the investor can select his/her portfolio at date t based on information available by that date. In the applications which we will deal with (capital growth), the standard models, e.g. [2, 6, 32, 53], exclude short selling. Negative portfolio positions might lead to infinite negative values of logarithmic 3

792

W Bahsoun, IV Evstigneev and MI Taksar

functionals, playing a central role in the present context. Following this approach, we will assume that all contingent portfolios Xt (s 1) are represented by non-negative vector functions. All functions of st will be assumed to be measurable and those representing contingent portfolios essentially bounded. Any sequence of contingent portfolios x 0 ( s 0 ), x 1 ( s 1 ), x 2 ( s 2 ), ... will be called a trading strategy. Trading strategies describe possible scenarios of investors' actions at the financial market influenced by random factors. In the model, we arc given sets G 1 (st) r;::; セ@ x セウー・」ゥヲケョァ@ the self-financing (solvency) constraints. The main focus of the study is on self-financing trading strategies. A strategy x 0 ( s 0 ), x 1 ( s 1 ), x 2 ( s 2 ), ... is called self-financing if

(1) almost surely (a.s.) for all t セ@ 1. The inclusion (.rt_ 1 (st- 1 ),.rt(st)) E Gt(st) means that the portfolio Xt- 1 (s 1- 1 ) can be rebalanced to the portfolio Xt(s 1) at date t in the random situation st under transaction costs and trading constraints. The rebalancing of a portfolio excludes inflow of external funds, but it may take into account dividends paid by the assets. It is assumed that for each t セ@ 1, the set Gt ( st) is a closed convex cone depending measurably 2 on st. This assumption means that the model takes into account pTOpoTtional transaction costs. We give examples of the cones Gt(st) below. Example 1. No transaction costs. Let

be the vector of the market prices of assets i = 1, 2, ... , n at date t. Suppose that portfolio positions are measured in terms of the market values of assets. Define

W, ... ,

bn) A portfolio a = (a\ ... , an) can be rebalanced to a portfolio b = (without transaction costs) if and only if (a, b) E Gt( st). Example 2. PTOpoTtional transaction costs: single currency. Let Gt(st) be the set of those (a, b) E セ@ x セ@ for which

2 A closed set G( s) c:;; lRn is said to depend measurably on a parameter s if the distance to this set from each point in JRn is a measurable function of s.

4

Chapter 38. Growth-Optimal Investments and Numeraire Portfolios n

t

.

'""

-

t

L.)1- >-t,; (s )) ( i= l

i

793

qi ((s t-I)a ) i

qt-1 s

i

-b)+,

(3)

where r+ := rnax{r, 0} for the real number T. The transaction cost rates for buying and selling are given by the numbers >-tJst) 2: 0 and 1 > ^N セ [H@ st) 2: 0, respectively. A portfolio a = ( a 1 , ... , a") can ' be rebalanced to a' portfolio b = ( bl, ... , b") (with transaction costs) if and only if the pair of vectors (a, b) belongs to the cone Gt(st). Here, we again assume that the coordinates ai and bi of the portfolio vectors indicate the current market values of the asset holdings. The inequality in (3) expresses the fact that purchases of assets are made only at the expense of sales of other assets. The approach ba..'led on relations (3) is standard in the analysis of transaction costs; see e.g. Jouini and Kallal [28], Cvitanic and Karat:6as [10], and Pham and Tomi [43]. Example 3. PropoTtional transaction costs: several currencies. Consider an asset market where n currencies are traded. Suppose that for each t = 1, 2, .. . a matrix セl セェ@

(

st ) with セェ@

> 0 and セ Q[@

=

1

is given, specifying the exchange rates of the currencies i = 1, 2, .. ., n (including transaction costs). The number セェ@ ( st) shows how many units of currency i can be obtained by exchanging one unit of currency j . A portfolio a = (a 1 , ... , an) of currencies can be exchanged to a portfolio b = (b1 , ... , b" ) at date t in the random situation s 1 if and only if t here exists a nonnegative matrix (dfi) (exchange matrix) such that n

n

'""dji a -:/L t ' 0 :/: : bi /Blセエ@ '"" i ------

j= l

ij (

s t) dij ·t ·

j=l

Here, 、セ Q@ (i =l _j ) stands for the amount of currency j exchanged into currency i. The amount 、 セゥ@ of currency i is left unexchanged. The second inequality says that at time t the ith position of the portfolio cannot be greater than the sum RZ W セ Q@ セ セェ@ 、 セ j@ obtained as a result of the exchange. The model we deal with here is a version of the multicurrency models considered by Kabanov, Stricker and others (see e.g. [30], [31] and [29]) . In spite of some similarity, it cannot be included into the framework developed in the above papers. Note that in this example asset holdings are expressed in terms "physical units" of &'lsets (currencies). An important class of dynamic securities market models is formed by stationary models. They arc defined as follows. A model is called station-

5

794

W Bahsoun, IV Evstigneev and MI Taksar

ary if the stochastic process (St) is stationary:> and the given cones Gt (st) (specifying the solvency constraints) arc of the following form:

(4) where for each .st the set G(.c/) is a closed convex cone in irセ@ x irセ@ depending measurably on .st. Assumption (4) expresses the fact that the solvency constraints do not explicitly depend on time: their structure depends only on the current and previous states of the world-on the history .•/ of the underlying stochastic process. In the stationary context it is convenient to assume that St is defined for each t = 0, ±1, ±2, ... , and in this case the notation st refers to the infinite history 8t = ( ... , 8t_ 1 , 8t)· This convention will always apply when we shall deal with stationary models. If the stochastic process (.s 1) is stationary, then the models considered in Examples 1 and 2 are stationary if the asset returns Rt (8t) : = アセ@ (1:/) / qL 1(8t-l) and the transaction cost rates A(.s 1) and A+ (.s 1) do not explicitlvu depend エセGl@ t,1· on t:

Rt(.st)

=

R(.st), AtJ.st)

=

A;(.st).

The analogue of this assumption in the Example 3 is the condition that the exchange rates do not explicitly depend on t: エlセェ@ ( .st) = tLij ( .st). In the analysis of stationary models, we will consider a class of trading strategies called balanced. A strategy x 0 , x 1 , x 2 , ... is termed balanced if there exist a vector function x(s 0) E irセ@ and scalar function o:(.s 0) > 0 such that

(5) and lx(8°) I = 1. (We write 1·1 for the sum of the absolute values of the coordinates of a vector). According to (5), portfolios Xt(.s 1) grow with stationary proportions defined by the random vector process x(s 0 ),x(.s 1 ), ... and at a stationary rate o:(s 1 ), a(.s 2), .... The results of capital growth theory pertaining to stationary models (see Section 3) will be stated in terms of balanced trading strategies.

3

Von Neumann-Gale dynamical systems

Von Neumann-Gale dynamical systems are defined in terms of multivalued (set-valued) operators possessing properties of convexity and homogeneity. States of such systems are represented by elements of convex cones 3 Recall that a セエッ」ィ。ウゥ@ process ( st) ゥセ@ called stationary if for any m = 1, 2, ... and any measurable function ¢on the product of m copies of the spaceS x ... x S, the distribution of the random variable rPt := ¢( st+l, ... , st+m) does not depend on t.

6

Chapter 38. Growth-Optimal Investments and Numeraire Portfolios

795

Xt (t = 0, 1, ... ) in linear spaces. Possible one-step transitions from one state to another arc described in terms of given operators A 1 ( x), assigning to each x E Xt_ 1 a convex subset At(x) a and 0 :::; b' :::; b, then (a, b) E Gt (st) ("free disposal hypothesis"). All inequalities between vectors, strict and non-strict, are understood coordinatewise. Define

G;(st)

=

{(c, d):_;> 0: db- ca :S: 0 for all (a, b) E Gt(st)},

(8)

where ca and db denote the scalar products of the vectors. Let Pt denote the set of measurable vector functions p( .st) with values in セ@ such that 7

796

W Bahsoun, IV Evstigneev and MI Taksar

Elp(st) I < x. A dual path (dual trajectory) is a finite or infinite sequence P1(st),p2(st), ... such that PtE Pt (t ::::> 1) and

(9) for all t ::::> 1. We write Et(-) = E(·lst) for the conditional expectation given st. By virtue of (8) and (9), Et(Pt+IY) :s; PtX (a.s.) for any (x, y) E Zt. This inequality shows that for any path x 0 , x 1 , ... the sequence ofrandom variables p 1 x 0 ,p2 x 1 , ... is a supermartingale with respect to the given filtration in the underlying probability space generated by st. A dual path p 1 ,p 2 , ... is said to support a path x 0 , x 1 , ... if PtXt-l

=

(10)

1 (a.s.)

for all t ::::> 1. A trajectory is called rapid if there exists a dual trajectory supporting it. The term "rapid" is motivated by the fact that Et(Pt+lYt) PtYt-l

:s;

Et(Pt+lxt) PtXt-l

=

1 (a.s.)

for each path Yo, y 1 , ... with PtYt-l > 0 (see (9) and (10)). This means that the path x 0 , x 1 , ... maximizes the conditional expectation of the gmwth rate at each time t, the maximum being equal to 1. Growth rates are measured in terms of the random linear functions Pta of a E セM If states Xt of the von Neumann-Gale system represent portfolios whose positions arc expressed in terms of units of assets, then Pt can be interpreted as asset price vectors. If the ith coordinate クセ@ of the vector Xt stands for the market value of the ith position of the portfolio, then Pl may be regarded as a discount factor for the market price of the ith asset. Another motivation of the term "rapid path" lies in the fact that any rapid path is asymptotically optimal-see the next section.

4

Capital growth theory and von NeumannGale dynamical systems

From the point of view of capital growth, those investment strategies arc of primary interest for which investor's wealth grows at an asymptotically optimal rate. There are various approaches to the notion of asymptotic optimality. In the definition below, we follow essentially Algoet and Cover

[2].

8

Chapter 38. Growth-Optimal Investments and Numeraire Portfolios

797

Definition 1. Let x 0 , x 1 , ... be an investment strategy. It is called asymptotically optimal iffor any other investment strategy y 0 , y 1 , ... there exists a supermartingale セエ@ such that セ@

IYII

:::;

セエL@

t = 0, 1, ... (a.s.).

Recall that for a vector b = W, ... , bn) we write lbl = WI+ ... + lb''l. If b 2> 0, then I bI = b1 + ... + bn, and if the vector b represents a portfolio whose positions are measured in terms of the market values of assets, then lbl is the market value of this portfolio. Note that the above property remains valid if lbl is replaced by any function 1/Jt(st, b) (possibly random and depending on t) which satisfies

(11) where 0 < l < L are non-random constants. As an example of such a function, we can consider the liquidation value (or net a88et value) of the portfolio n

i=l

within the model defined by (3). This is the amount of money the investor gets if he/she decides to liquidate the portfolio (sell all the assets) at date t. Clearly condition ( 11) holds if the random variables 1- L|セゥ@ > 0 are uniformly bounded away from zero. · The strength of the above definition, which might seem not immediately intuitive, is illustrated by the following implications of asymptotic optimality. is a supermartingale, the As long as IYMix 1 1:::; E,1 , t = 0, 1, ... (a.s.), キィ・イセ@ following properties hold. (a) With probability one

IYtl

sup -1-.I < oo, t

Xt

1.e. for no strategy wealth can grow asymptotically faster than for To, T 1 , ... (a.s. ). (b) The strategy x 0 , x 1 , ... a.s. maximizes the exponential growth rate of wealth

1

lim sup -ln lxtl· t-->CXJ

t

(c) We have

IYtl

IYtl

sup E-l- 1< oo and sup E ln -1-.I < oo. t

Xt

t

9

Xt

798

W Bahsoun, IV Evstigneev and MI Taksar

Assertion (a) follows from a.s. convergence of non-negative supermartingales; (b) is immediate from (a); the first part of (c) holds because el is a non-negative supermartingale; the second part of (c) is obtained by using Jensen's inequality and the supermartingale property: E(lnet+ 1 ls 1) ::::; ln E(et+llst) ::::; lnet· This work aims at obtaining results on optimal growth in the model with transaction costs described in Section 2. The main results are concerned with the existence of asymptotically optimal strategies in the general (nonstationary) version of the model and the existence of asymptotically optimal balanced strategies in its stationary version. Our main tool for analyzing the questions of asymptotic optimality is the concept of a rapid path in the stochastic von Neumann-Gale system (see the previous section). Definition 2. A self-financing trading strategy x 0 , x 1 , ... is called rapid if it forms a rapid path in the underlying von Neumann-Gale dynamical system which defines the asset market model. When dealing with the dynamic securities market model defined in terms of a von Neumann-Gale dynamical system, we will use the terms "paths" and "self-financing trading strategies" interchangeably. In the context of the present model, rapid paths may be regarded as analogues of numemire portfolios (Long [34]). As we have noticed, the price system (or the system of discount factors) (Pt) involved in the definition of a rapid path is such that the value Pt+lXt of the portfolio Xt is always equal to one, while for any other feasible sequence (Yt) of contingent portfolios (selffinancing trading strategy), the values Pt+lYt form a supermartingale. In the classical case when transaction costs are absent, these conditions hold for the price vectors Pt+l := Atqt, where qt are the market prices and A.t 1 = qt.Tt is the market value of the numeraire portfolio Xt. The latter is defined so that the normalized prices qtf qtXt form a supermartingale. (Long [34] considered a model with unlimited short selling, and in that context one can speak of martingales rather than supermartingales.) The results are based on assumption (G.5) below. (G.5) There exists an integer l .2: 1 such that for every t .2: 0 and i = 1, ... , n there is a path Yt,i, ... , Yt+Z,i over the time interval [t, t + l] satisfying Yt,i =

e;, ... , Yt+Z,i .2: (e,

where e; = (0, 0, ... , 1, ... , 0) (the ith coordinate is 1 while the others are 0) and 1 is a strictly positive non-random constant. Proposition 1. If the constants Nit in condition (G.2) do not depend on t and assumption (G. 5) holds, then any rapid path is asymptotically optimal. Thus in order to prove the existence of asymptotically optimal strategies it is sufficient to establish the existence of infinite rapid paths. For a proof 10

Chapter 38. Growth-Optimal Investments and Numeraire Portfolios

799

of Proposition 1 see Evstigneev and Flam [17], Proposition 2.5. In specific dynamic securities market models, condition (G.5) holds typically with l = 1. Then it means a possibility of buying some fixed strictly positive amounts of all the assets by selling one unit of any asset i = 1, ... , n (or if portfolio positions are measured in terms of the market values of assets-by selling the amount of asset i worth a unit of cash). The main results of this paper are collected in the following theorem. Theorem 1. (i) Let x 0 ( s 0 ) be a vec:toT function in X 0 such that ce :::; x 0 (s 0 ) :::; Ce joT some constants 0 < c :::; C. Then theTe exists an infinite mpid path with initial state x 0 (s 0 ). (ii) If the model is stationaTy and (G.5) holds, then theTe exists a balanced mp'id path. (iii) If the constants 1\!It in condition (G.2) do not depend on t and assumption (G.5) holds, then the mpid paths whose existence is claimed in (i) and (ii) aTe asymptotically optimal. Assertion (iii) is immediate from Proposition 1. Statement (i) of the above theorem is proved in [5], where the existence of infinite rapid paths with the given initial state is established. The proof in [5] is conducted by passing to the limit from finite time horizons, for which the existence of rapid paths is obtained in [17]. The passage to the limit is based on a compactness principle involving Fatou's lemma in several dimensions (Schmeidler [51]). Assertion (ii) follows from the results of papers [20, 21], where not only the existence of a rapid path is proved, but also it is shown that there exists a balanced rapid path supported by a dual trajectory with the following special structure:

( P18

1)

(1

( t)

=ps),Pt8

=

p(.c/) ( 1 ),t=2,3, ... , (t- 1 ) ... as

as

(12)

where a( s 1 ) > 0 and p( s 1 ) '2' 0 are scalar and vector functions such that Elp(s 1 )1 < oo (balanced dual tmjectoTy). The triplet of functions a(·), p(·), x(·) involved in (5) and (12) is called a von Neumann equilibTium. It can be shown that if a(·), p( ·), x( ·) is a von Neumann equilibrium, then the balanced trajectory defined by (5) maximizes E ln a among all such trajectories. This means by definition that (5) is a von Neumann path. The existence of a von Neumann equilibrium established in [20, 21] is a deep result solving a problem that remained open for more than three decades. In the former of the two papers [20, 21], a version ofthe existence theorem for a von Neumann equilibrium is obtained which deals with an extended model defined in terms of mndomized paths. In the latter paper, the final result is derived by using the method of elimination of randomization (Dvoretzky, Wald and Wolfowitz

[13]). To use Theorem 1 in specific models, one has to verify assumptions (G.l)(G.5) (note that (G.3) is a consequence of (G.5) with l = 1). In Exam11

800

W Bahsoun, IV Evstigneev and MI Taksar

ple 1, these conditions follow from the assumption that the asset returns rセHウlI@ := アセHウlIェ⦅ are uniformly bounded and uniformly bounded Q HウlMャI@ away from zero. To obtain (G.l)-(G.5) in Example 3 it is sufficient to assume that the exchange rates jlセェ@ ( st) are uniformly bounded away from zero and infinity. In Example 2, all the conditions needed can be obtained if the above assumption regarding rセHウエI@ holds and the following requirement regarding the transaction costs is fulfilled: the random variables >..i.i (st) are uniformly bounded and the random variables 1- ^NセゥHウエI@ > 0 arc u'niformly bounded away from zero. In all the three cases, (G.2) holds with constants Mt independent oft.

5

Directions for further research

1. In the present paper, as well as in many studies on capital growth theory, admissible portfolios arc represented by non-negative vectors (short sales of assets are ruled out). It would be of interest to develop a more general model where short selling is allowed, possibly with some restrictions. One can assume that the set of admissible portfolio vectors at date t is some given cone Kt(.st) ;:2 irセN@ Important examples of Kt(.st) are cones defined in terms of margin requirements. Their informal definition (which can easily be formalized in one way or another) is as follows. A portfolio is admissible if and only if by liquidating its long positions, one can always obtain an amount sufficient to finance a margin lvf ( > 1) times the total short position. Under proportional transaction costs, this definition leads to polyhedral cones Kt (st). The case of polyhedral cones of admissible portfolios deserves special attention, in particular, because it might be more easily tractable than the general case. One might conjecture that for such models the theory of growth optimal investments can be deduced from that developed in the present paper. 2. It is a widespread misconception that capital growth theory is "nothing but the maximization of expected logarithmic utilities". This misconception has been partially reflected in the famous discussion between Shannon and Samuelson (sec, e.g., [9]), which has substantially influenced the whole history of capital growth theory. Of course there are some grounds for the above view: in the case of a frictionless market and exogenous asset returns, one can construct the growth optimal portfolio by the "myopic" maximization of the conditional expectation of the logarithm of the portfolio return (see [2]). I3ut this is not the case in more complex and realistic settings such as models with transaction costs or with endogenous asset prices, e.g. those pertaining to evolutionary finance [18]. The fundamental definition of a growth-optimal strategy is based on a robust, utility independent, notion of optimality al-

12

Chapter 38. Growth-Optimal Investments and Numeraire Portfolios

801

most surely. What functionals have to be maximized in order to sequentially generate an infinite rapid path, especially in a non-stationary setting, is an open question. Some progress in this direction has been made only in the deterministic von Neumann-Gale model, where efficient functionals (Rubinov [50]) have been constructed for that purpose. The extension of these results to the stochastic case is an interesting unsolved problem. 3. Results on the previous problem might be helpful for the design of methods for computing growth optimal investment strategies (rapid paths) under transaction costs. However, instead of constructing efficient functionals, which would in a sense copy the procedure of consecutive myopic optimization, one can attempt a different approach. One can try to construct approximations HクセカIサG]d@ for rapid paths over finite time horizons N = 1, 2, ... and then obtain a rapid path HクエIセ P@ as a limit of x{' as N---+ oo. As (x/;v)i":o one can select finite paths which maximi11e terminal logarithmic functionals of the form Eln lxNI· Pursuing this approach, we enter the fascinating area of turnpike theorems pioneered by Dorfman, Samuelson and Solow [12] (see [40, 42] for related deterministic results and [4] for their stochastic versions). Results of this kind express the general principle that all optimal paths defined over sufficiently large time horizons "have an inclination for leaning to essentially the same route--the turnpike". In our model, the role of the turnpike can be played by the von Neumann path, if the model is stationary, and any infinite rapid path, if it is non-stationary [3, 22]. Although the turnpike theory for stochastic von Neumann-Gale systems is well-developed [19], most of the results currently available in this field require certain conditions of strict convexity of the transition cones Gt. In financial applications, basic examples of the cones Gt are polyhedral and do not satisfy such conditions. Thus in order to realize this program one needs to develop stochastic turnpike theorems for the (linear) von Neumann model, which is a non-trivial task even in the deterministic case--see e.g. [41] and [39]. However, the introduction of randomness may paradoxically alleviate the problem, since the lack of strict convexity of the transition cones may be compensated by the strict concavity of the logarithmic functionals, which can be expected under appropriate assumptions of non-degeneracy of the underlying probability distributions (cf. [2], p. 882). 4. The analysis of this final problem might help to attack the two previous ones. Consider a Markov version ofthe stochastic von Neumann-Gale model, in which the process St is Markov and the transition cones Gt(.st) depend only on the current state St. In the theory of controlled Markov processes, there are various results on the sufficiency of Markov strategies [4, 16]. They show that among optimal strategies one can select at least one which is Markov in a properly defined sense. Do such theorems hold in the model at hand? 13

802

W Bahsoun, IV Evstigneev and MI Taksar

What is the proper definition of a Markov rapid path? In particular, should it involve only primal, or both primal and dual variables? "''hat methods could be employed for the proof of Markov sufficiency theorems for infinite rapid paths, where a direct application of dynamic programming is hampered (the time horizon does not have an "end")? These questions open an interesting area for further research and their resolution might fill a substantial gap in the theory of growth optimal investments under transaction costs. Acknowledgment. Financial support from the Swiss National Center of Competence in Research "Financial Valuation and Risk Management" (NCCR FINRISK), the NSF grant DMS-0505435, the State of Missouri Research Board, the University of Missouri-Columbia Research Council, Norwegian Financial Market Fund (project "Stochastic Dynamics of Financial Markets") and the Manchester School Visiting Fellowship Fund is gratefully acknowledged.

References [1] Akian, M., Sulem, A. and Taksar, M.I., Dynamic optimization of longternl growth rate for a mixed portfolio with transactions costs, Mathematica] Finance 11 (2001) 153-188. [2] Algoet, P.R. and Cover, T.M., Asymptotic optimality and asymptotic equipartition properties of log-optimum investment, Annals of Probability 16 (1988) 876-898. [3] Anoulova S.V., Evstigneev, LV. and Gundlach, V.M., Turnpike theorems for positive multivalued stochastic operators, Advances in Mathematical Economics 2 (2000) 1-20. [4] Arkin, V.I. and Evstigneev, LV., Stochastic Models of Control and Economic Dynamics, 1987, Academic Press, London. [5] Bahsoun, W., Evstignccv, LV. and Taksar, M.I., Rapid paths in von Neumann-Gale dynamical systems, Stochastics 80 (2008) 129-142. [6] Breiman, L., Optimal gambling systems for favorable games, Proceedings of the 4th Berkeley Symposium on Mathemat'ical Statistics and Probability, Vol. I, 1961, University of California Press, Berkeley, pp. 65-78. [7] Browne, S., Survival and growth with liability: Optimal portfolio strategies in continuous time, Mathematics of Operations Research 22 (1997) 468 493. 14

Chapter 38. Growth-Optimal Investments and Numeraire Portfolios

803

[8] I3rowne, S., Risk constrained dynamic active portfolio management, Management Science 46 (2000) 1188-1199. [9] Cover, T.M., Shannon and investment, IEEE Information Theory Society Newsletter, Summer 1998, Special Golden Jubilee Issue, pp. 10-11. [10] Cvitanic, J. and k。イエセウL@ I., Hedging and portfolio ッーエゥュセ。ョ@ under transaction costs: A martingale approach, Mathematical Finance 6 (1996) 133-165. [11] Dempster, M.A.H., Evstigneev, I.V. and Taksar, M.I., Asset pricing and hedging in financial markets with transaction costs: An approach based on the von Neumann-Gale model, Annals of Finance 2 (2006) 327-355. [12] Dorfman, R., Samuelson, P. and Solow, R., Linear Progmmm'ing and Economic Analysis, 1958, McGraw-Hill, New York. [13] Dvoretzky, A., Wald, A., and Wolfowitz, J., Elimination of randomゥセ。エッョ@ in certain problems of statistics and of the theory of games, Proceedings of the National Academy of Sciences of the USA 36 (1950) 256-260. [14] Dynkin, E.B., Some probability models for a developing economy, Soviet Mathematics Doklady 12 (1971) 1422-1425. [15] Dynkin, E.B., Stochastic concave dynamic programming, USSR Mathematics Sbornik 16 (1972) 501-515. [16] Dynkin, E.B. and Yushkevich, A.A., Controlled Markov Processes and their Applications, 1979, Springer, N.Y. [17] Evstigneev, I.V. and Flam, S.D., Rapid growth paths in rnultivalued dynamical systems generated by homogeneous convex stochastic operators, Set- Valued Analysis 6 (1998) 61 82. [18] Evstigneev, I.V., Hens, T. and Schenk-Hoppe, K.R., Evolutionary finance, in: T. Hens and K.R. Schenk-Hoppe (Eds.), Handbook of Financial Markets: Dynamics and Evolution, a volume in the Handbooks in Finance series, W.T. Ziemba, Ed., 2009, Elsevier, Amsterdam, pp. 507-566. [19] Evstigneev, I.V. and Schenk-Hoppe, K.R., The von Neumann-Gale growth model and its stochastic generalization, in: Dana, R.-A., LeVan, C., Mitra, T. and Nishimura, K. (Eds.), Handbook on Optimal Growth, Vol. I, Chapter 2, 2006, Springer, N.Y., pp. 337-383. 15

804

W Bahsoun, IV Evstigneev and MI Taksar

[20] Evstigneev, LV. and Schenk-Hoppe, K.R., Pure and randomized equilibria in the stochastic von Neumann-Gale model, JouTnal of Mathematical Economics 43 (2007) 871-887. [21] Evstigneev, LV. and Schenk-Hoppe, K.R., Stochastic equilibria in von Neumann-Gale dynamical systems, Tmnsactions of the AmeTican Mathematical Society 360 (2008) 3345-3364. [22] Evstigneev, LV. and Taksar, M.L, Rapid growth paths in convex-valued random dynamical systems, Stochastics and Dynamics 1 (2001) 493509. [23] Gale, D., A closed linear model of production, in: Kuhn, H.W. et al. (Eds.), Linear Inequalities and Related Systems, 1956, Princeton University Press, Princeton, pp. 285-303. [24] Hakansson, N.H. and Ziemba, vV.T., Capital growth theory, in: Jarrow, R., Maksimovic, A.V. and Ziemba, W.T. (Eds.), Handbooks in Opemtions ReseaTch and Management Science, Vol. 9, Finance, Chapter 3, 1995, Elsevier, Amsterdam, pp. 65-86. [25] Hausch, D.B. and Ziemba, \V.T., Transactions costs, market inefficiencies and entries in a racetrack betting model, Management Science 31 (1985) 381-394. [26] Iyengar, G. and Cover, T.M., Growth optimal investment in horse race markets with costs, IEEE Tmn.sactions on InfoTmation TheoTy 46 (2000) 2675-2683. [27] Iyengar, G., Universal investment in markets with transaction costs, Mathematical Finance 15 (2005) 359-371. [28] .Jouini, E. and Kallal, H., Martingales and arbitrage in securities markets with transaction costs, Joumal of Economic TheoTy 66 (1995) 178-197. [29] Kabanov, Yu.M., The arbitrage theory, in: Handbooks in Mathematical Finance: Option PTicing, InteTest Rates and Risk Management, .Touini, E., Cvitanic, J. and Musiela, M. (Eds.), 2001, Cambridge University Press, Cambridge, pp. 3-42. [30] Kabanov Yu.M., Hedging and liquidation under transaction costs m currency markets, Finance and Stochastics 3 (1999) 237-248.

16

Chapter 38. Growth-Optimal Investments and Numeraire Portfolios

805

[31] Kabanov, Yu.M. and Stricker, C., The Harrison-Pliska arbitrage pricing theorem under transaction costs, JouTrwl of l'vfathemat'ical Economics 35 (2001) 185-196. [32] Kelly, J.L., A new interpretation of information rate, Bell System Technical Journal 35 (1956) 917-926. [33] Latanc, H., Criteria for choice among risky ventures, Journal of Political Economy 67 (1959) 144-155. [34] Long, J.B., The numeraire portfolio, Journal of Financial Econornics 26 (1990) 29-69. [35] MacLean, L.C., Ziemba, vV.T. and Blazenko, G., Growth versus security in dynamic investment analysis, Management Science 38 (1992) 15621585. [36] MacLean, L.C., Sanegre, R., Zhao, Y. and Ziemba, W.T., Capital growth with security, Journal of Economic Dynamics and Control 28 (2004) 937-954. [37] MacLean, L.C., Ziemba, W.T. and Li, Y., Time to wealth goals in capital accumulation, Quantitative Finance 5 (2005) 343-355. [38] MacLean, L.C., Thorp, E.O. and Ziemba, W.T. (Eds.), The Kelly Capdal Gmwth Investment Crder·ion: TheoTy and Pmctice, 2011, \Vorld Scientific, Singapore. [39] Makarov, V.L. and Rubinov, A.M., Mathematical TheoTy of Economic Dynamics and Equilibria, 1977, Springer-Verlag, Berlin. [40] McKenzie, L.\V., Turnpikes, Amer'ican Economic Review Papers and Proceedings 88 (1998) 1-14. [41] Morishima, M., Equilibrium, Stability and Growth: A Multi-Sectoral Analysis, 1964, Oxford University Press, London. [42] Nikaido, H., Convex Structures and Economic Theory, 1968, Academic Press, New York. [43] Pham, H. and Touzi, N., The fundamental theorem of asset pncmg under cone constraints, Journal of Mathematical Economics 31 (1999) 265 279.

17

806

W Bahsoun, IV Evstigneev and MI Taksar

[44] Platen, E. and Heath, D., A Benchmark Approach to Quantitative Finance, 2006, Springer, N.Y. [45] Radner, R., Balanced stochastic growth at the maximum rate, in: Contributions to the von Neumann Growth Model (Conference Proceedings, Institute for Advanced Studies, Vienna, 1970), Zeitschrift fur Nationalokonomie, Suppl. No. 1, pp. 39-53. [46] Radner, R., Optimal steady-state behavior of an economy with stochastic production and resources, in: Day, R.H. and Robinson, S.M. (Eds.), Mathematical Topics in Economic Theory and Computation, 1972, SIAM, Philadelphia, pp. 99-112. [47] Radner, R., Optimal stationary consumption with stochastic production and resources, Jov.rnal of Economic Theory 6 (1973) 68-90. [48] Radner, R., Equilibrium under uncertainty, in: Arrow, K.J. and Intrilligator, M.D. (Eds.), Handbook of Mathematical Economics, 1982, North-Holland, Amsterdam, pp. 923 1006. [49] Rockafellar, R.T., l'vfonotone Processes of Convex and Concave Type, Memoirs of the American Mathematical Society, Volume 77, 1967, American Mathematical Society, Providence, RI. [50] Rubinov, A.M., Superlinear Multivalued Mappings and Their Applicat·ions to Problems of lVIathematical Economics, 1980, Nauka, Leningrad. (In Russian.) [51] Schmeidler, D., Fatou's lemma in several dimensions, Proceedings of the American Mathematical Society 24 (1970) 300-306. [52] Taksar, M.I., Klass, M.J. and Assaf, D., Diffusion model for optimal portfolio selection in the presence of brokerage fees, Mathematics of Operations Research 13 (1988) 277-294. [53] Thorp, E.O., Portfolio choice and the Kelly criterion, Proceedings of the Business and Economics Section of the American Statistical Association (1971)' 215-224. [54] Ziemba, W.T. and Vickson, R.G. (Eds.), Stochastic Optimization Models in Finance, 1975, Academic Press, N.Y. [55] Ziemba, W.T. and Vickson, R.G. (Eds.), Stochastic Optimization Models in Finance (2 ed.), 2006, World Scientific, Singapore. 18

Chapter 38. Growth-Optimal Investments and Numeraire Portfolios

807

[56] Ziemba, W.T. and Vickson, R.G., Models of optimal capital accumulation and portfolio selection and the capital growth criterion, in: The Kelly Capital Growth Investment Criterion: Theory and Practice, MacLean, L.C., Thorp, E.O. and Ziemba, W.T. (Eds.), 2011, World Scientific, Singapore, pp. 4 73-485. [57] Von Neumann, J., Ubcr cin okonomischcs Glcichungssystcm und cine Verallgemeinerung des Brouwerschen Fixpunktsatzes, in: Ergebnisse eines Mathematischen Kolloquiums, 8 (1937), pp. 1935-1936 (FranzDcutickc, Leipzig and Wicn), pp. 73-83. [Translated: A model of general economic equilibrium, Review of Economic Studies 13 (1945-1946) 1-9.]

PROBLEMS 1. Develop a generalization of the model considered in the paper allow-

ing for (restricted) short sales of assets. Extend the capital growth theory presented in the paper to this model, assuming that the set of admissible Consider the portfolio vectors at date tis given by some cone Kt(st) セ@ irセN@ special case of polyhedral cones Kt(s 1 ), including the main example of Kt(s 1) defined by margin requirements. Try to deduce the theory of growth optimal investments in the case of polyhedral Kt(.st) from that presented in the paper (where Kt(.st) = irセIN@ 2. What functionals Ft(Xt-l, .st, ·)have to be maximized in order to recursively generate an infinite rapid path (.rt) by the formula :rt E arg max Ft (Xt-l, st, ·)? Try to approach this question by using the idea of efficient functionals (Rubinov 1980) in the deterministic von Neumann-Gale model. To begin with, examine this problem in the case of a stationary or a Markov model. 3. Design methods for computing growth optimal investment strategies (rapid paths) based on the turnpike principle. To this end extend the turnpike theorems by Anoulova, Evstigneev and Gundlach (2000) to the case of not necessarily strictly convex, in particular polyhedral, transition cones Ct. Examine whether the assumption of strict convexity of Gt can be replaced by appropriate assumptions of non-degeneracy of the underlying probability distributions (similar to those guaranteeing the uniqueness of the growth optimal portfolio in Algoet and Cover 1988). 4. Develop a Markov version of the theory of growth-optimal investments under transaction costs. Assuming that the von Neumann-Gale investment model is Markov, i.e. the process St 1s Markov and the transition cones 19

808

W Bahsoun, IV Evstigneev and MI Taksar

Gt(st) depend only on St, show that among all infinite rapid paths with the given initial state there exists at least one possessing a (properly defined) Markov property. To obtain such a Markov s1t[ficiency theorem, try to employ the methods of stochastic control over an infinite time horizon that do not rely upon the Bellman principle of dynamic programming, which is directly applicable only in the finite horiwn case.

20

JOURNAL OF

F1nancial ELSEVIER

ECONOMICS

Journal of Financial Economics 67 (2003) 41-80

www.elsevier.com/locate/econbase

Chapter 39

A multivariate model of strategic asset allocation* John Y. Campbella,d,*, Yeung Lewis Chanb, Luis M. Viceirac,d,e "Department ()l Economics, Hanard University, Littauer Center 213, 1875 Cambridge Street, Cambridge, MA 02138, USA b School of Business and Management, Hong Kong University of Science and Technology, Kmvloon, Hong Kong c Harvard Business School, Harvard University, Boston, MA 02163, USA d National Bureau ol Economic Research, Cambridge, MA 02138, USA c Centre for Economic Policy Research, London EC1 V7 RR, UK Received I December 1999; accepted 3 January 2002

Abstract

We develop an approximate solution method for the optimal consumption and portfolio choice problem of an infinitely long-lived investor with Epstein-Zin utility who faces a set of asset returns described by a vector autoregression in returns and state variables. Empirical estimates in long-run annual and post-war quarterly U.S. data suggest that the predictability of stock returns greatly increases the optimal demand for stocks. The role of nominal bonds in long-term portfolios depends on the importance of real interest rate risk relative to other sources of risk. Long-term inflation-indexed bonds greatly increase the utility of conservative investors. セI@ 2002 Elsevier Science B.V. All rights reserved. JEL classification: G12 Keywords: lntertemporal hedging demand; Portfolio choice; Predictability; Strategic asset allocation

''Campbell acknowledges the financial support of the National Science Foundation, Chan the financial support of the Hong Kong RGC Competitive Earmarked Research Grant (HKUST 6065/0lH), and Viceira the financial support of the Division of Research of the Harvard Business School. We are grateful for helpful comments and suggestions by Ludger Hentschel, Anthony Lynch, an anonymous referee, and seminar participants at Harvard, the 1999 Intertemporal Asset Pricing Conference hosted by the Centre Tnteruniversitaire de Recherche en Analyse des Organizations (CYRANO) of Montreal, the 2000 WFA Meetings, and the Kellogg School of Management at Northwestern University. Josh White provided invaluable research assistance. *Corresponding author. Department of Economics, Littauer Center 213, Harvard University, 1875 Cambridge Street, Cambridge, MA 02138, USA. Fax: + 1-617-495-8570. E-mail address: [email protected] (J.Y. Campbell). 0304-405Xj02/$- see front matter Cl? 2002 Elsevier Science B.V. All rights reserved. PII: S 0 3 0 4 - 4 0 5 X ( 0 2 ) 0 0 2 3 I - 3

809

810 42

JY Campbell, YL Chan and LM Viceira J. Y. Campbell eta/. I Journal of Financial Economics 67 (2003) 41-80

1. Introduction

Academic research in finance has had a remarkable impact on many aspects of the financial services industry, from mutual fund management to securities pricing and issuance to corporate risk management. But academic financial economists have thus far provided surprisingly little guidance to financial planners who offer portfolio advice to long-term investors. The mean-variance analysis of Markowitz (1952) provides a basic paradigm and usefully emphasizes the effect of diversification on risk, but this model ignores several critically important factors. Most important, the analysis is static; it assumes that investors care only about risks to wealth one period ahead. In reality, however, many investors-individuals as well as institutions such as charitable foundations or universities-seek to finance a stream of consumption over a long lifetime. Financial economists have understood at least since the work of Samuelson (1969) and Merton (1969, 1971, 1973) that the solution to a multi period portfolio choice problem can be very different from the solution to a static portfolio choice problem. In particular, if investment opportunities vary over time, then long-term investors care about shocks to investment opportunities-the productivity of wealth-as well as shocks to wealth itself. They may wish to hedge their exposures to wealth productivity shocks, giving rise to intertemporal hedging demands for financial assets. Brennan et al. (1997) have coined the phrase "strategic asset allocation" to describe this farsighted response to time-varying investment opportunities. Unfortunately, Merton's intertemporal model is hard to solve in closed form. For many years solutions to the model were generally unavailable unless the investor had log utility of consumption with constant relative risk aversion equal to one, but this case is relatively uninteresting because it implies that Merton's model reduces to the static model. Rubinstein (1976a,b) obtains some important insights by adding a subsistence level to the log utility model, converting it into a model of declining relative risk aversion; in particular, he makes the point that long-term inflationindexed bonds, and not short-term bonds, are the riskless asset for long-term investors. 1 But these preferences are not standard and most economists have continued to assume constant relative risk aversion. The lack of closed-form solutions for optimal portfolios with constant relative risk aversion has limited the applicability of the Merton model; it has not become a usable empirical paradigm, has not displaced the Markowitz model, and has had little influence on financial planners and their clients. Recently, this situation has begun to change as a result of several related developments. First, computing power and numerical methods have advanced to the point at which realistic multiperiod portfolio choice problems can be solved 1 M odigliani and Sutch ( 1966) argue that honds are safe assets for long-term investors, and Stiglit7 ( 1970) builds a rigorous model to illustrate their point. Rubinstein ( 1981) also explores the intertemporal demand for long-term bonds in a three-period version of the Merton model under the assumption of exponential (constant absolute risk aversion) utility. Rubinstein (1976b), Lucas (1978), Breeden (1979), and Grossman and Shiller (1981) show how Merton's results could be interpreted in terms of consumption risk, an idea that has had major influence in macroeconomics.

Chapter 39. A Multivariate Model of Strategic Asset Allocation J. Y. Campbell et al. I Journal of Financial Economics 67 (2003) 41-80

811 43

numerically using discrete-state approximations. Balduzzi and Lynch (1999), Barberis (2000), Brennan et a!. (1997, 1999), Cocco et a!. (1998), and Lynch (200 1) are important examples of this style of work. Second, financial theorists have discovered some new closed-form solutions to the Merton model. In a continuoustime model with a constant riskless interest rate and a single risky asset whose expected return follows a mean-reverting (Ornstein-Uhlenbeck) process, for example, the model can be solved if long-lived investors have power utility defined over terminal wealth (Kim and Omberg, 1996), or if investors have power utility defined over consumption and the innovation to the expected asset return is perfectly correlated with the innovation to the unexpected return, making the asset market effectively complete (Wachter, 2002), or if the investor has Epstein and Zin (1989, 1991) utility with intertemporal elasticity of substitution equal to one (Campbell and Viceira, 1999; Schroder and Skiadas, 1999). Similar results are available in affine models of the term structure (Brennan and Xia, 2002; Campbell and Viceira, 2001; Liu, 1998; Wachter, 2000). Finally, approximate analytical solutions to the Merton model have been developed (Campbell and Viceira, 1999, 2001, 2002). These solutions are based on perturbations of the known exact solutions for intertemporal elasticity of substitution equal to one, so they are accurate provided that the intertemporal elasticity is not too far from one. They offer analytical insights into investor behavior in models that fall outside the still limited class that can be solved exactly. Despite this encouraging progress, it remains extremely hard to solve realistically complex cases of the Merton model. Discrete-state numerical algorithms become slow and unreliable in the presence of many assets and state variables, and approximate analytical methods seem to require a daunting quantity of algebra. Neither approach has been developed to the point at which one can specify a general vector autoregression (VAR) for asset returns and hope to solve the associated portfolio choice problem. The purpose of this paper is to remedy this situation by extending the approximate analytical approach of Campbell and Viceira (1999, 2001, 2002). Specifically, we show that if asset returns are described by a VAR, if the investor is infinitely longlived with Epstein-Zin utility, and if there are no borrowing or short-sales constraints on asset allocations, then the Campbell-Viceira approach implies a system of linear-quadratic equations for portfolio weights and consumption as functions of state variables. These equations are generally too cumbersome to solve analytically, but can be solved very rapidly by simple numerical methods. As the time interval of the model shrinks, the solutions become exact if the elasticity of intertemporal substitution equals one. They are accurate approximations for short time intervals and elasticities close to one. We apply our method to a VAR for short-term real interest rates, excess stock returns, and excess bond returns. We also include variables that have been identified as return predictors by past empirical research: the short-term interest rate (Fama and Schwert, 1977; Campbell, 1987; Glosten et a!., 1993); the dividend-price ratio (Campbell and Shiller, 1988; Fama and French, 1988); and the yield spread between long-term and short-term bonds (Shiller et al., 1983; Fama, 1984; Fama and French,

812

JY Campbell, YL Chan and LM Viceira

44

J. Y Campbell et al. I Journal of Financial Economics 67 ( 2003) 41 80

1989; Campbell and Shiller, 1991). In a variant of the basic approach we construct data on hypothetical inflation-indexed bond returns, following the approach of Campbell and Shiller (1996), and study the allocation to stocks, inflation-indexed bonds, nominal bonds, and bills. Two closely related papers are by Brennan et al. (1999) and Lynch (2001). Brennan et al. consider asset allocation among stocks, nominal bonds, bills, and interest-rate futures, using short- and long-term nominal interest rates and the dividend-price ratio as state variables. The investor is assumed to have power utility defined over wealth at a given horizon, and the stochastic optimization problem is solved using numerical dynamic programming imposing borrowing and short-sales constraints. Lynch considers asset allocation among portfolios of stocks sorted by size and book-to-market ratios, using the long-short yield spread and the dividendprice ratio as state variables, and assuming power utility defined over consumption. He solves the optimization problem with and without short-sales constraints, again using numerical dynamic programming. Our paper, by contrast, assumes recursive Epstein-Zin utility defined over an infinite stream of consumption and does not impose any portfolio constraints. The simplicity of our solution method allows us to consider an unrestricted VAR in which lagged returns are state variables along with the short-term nominal interest rate, dividend-price ratio, and yield spread. Our method also allows us to break intertemporal hedging demands into components associated with individual state variables. The organization of the paper is as follows. Section 2 explains our basic setup, and Section 3 describes our approximate solution method. Section 4 presents empirical results when stocks, nominal bonds, and bills are available. Section 5 considers portfolio allocation in the presence of inflation-indexed bonds. Section 6 concludes.

2. The model Our model is set in discrete time. We assume an infinitely long-lived investor with Epstein-Zin recursive preferences defined over a stream of consumption. This contrasts with papers such as Brennan et al. (1997, 1999), Kim and Omberg (1996), and Barberis (2000) that consider finite-horizon models with power utility defined over terminal wealth. We allow an arbitrary set of traded assets and state variables. Thus we do not make the assumption of Wachter (2000, 2002) that markets are complete, and we substantially extend the work of Campbell and Viceira (1999) in which there is a single risky asset with a single state variable. 2.1. Securities There are n assets available for investment. The investor allocates afterconsumption wealth among these assets. The real portfolio return Rp,t+I is given by n

Rp,t+I =

セ@

:X;,r(Ri,t+I- RI,t+I) ゥセR@

+ RI,t+I,

(I)

Chapter 39. A Multivariate Model of Strategic Asset Allocation J. Y Campbell et al. I Journal of Financial Economics 67 (2003) 41-80

813

45

where 'Xi,r is the portfolio weight on asset i. The first asset is a short-term instrument whose real return is Rt,t+l· Although we use the short-term return as a benchmark and measure other returns relative to it, we do not assume that this return is riskless. In practice, we use a nominal bill as the short-term asset; the nominal return on a nominal bill is riskless, but the real return is not because it is subject to short-term inflation risk. In most of our empirical analysis we consider two other assets: stocks and long-term nominal bonds. In Section 5, we also consider long-term inflationindexed bonds.

2.2. Dynamics of state variables We postulate that the dynamics of the relevant state variables are well captured by a first-order vector autoregressive process or VAR( 1). This type of dynamic specification has been used by Kandel and Stambaugh (1987), Campbell (1991, 1996), Rodrick (1992), and Barberis (2000), among others. In principle, the use of a V AR(l) is not restrictive since any vector autoregression can be rewritten as a V AR(l) through an expansion of the vector of state variables. For parsimony, however, in our empirical work we avoid additional lags that would require an expanded state vector with additional parameters to estimate. Specifically, we define l'l,t+ I

1'2,t+ I -

Xr+l

_ =

[ r3,t+l -

セGャLエK@

.

r n,t+l -

,

(2)

セGャLエK@

=

where ri,t+l log(Ri,t+t) for all i, and x 1+t is the vector of log excess returns. In our empirical application, r 1,t+l is the ex post real short rate, r 2 ,t+ 1 refers to the real stock return, and r 3,r+l is the export real return on nominal bonds. We allow the system to include other state variables st+ 1, such as the dividend-price ratio. Stacking r 1,t+ 1, x 1+ 1, and s1+t into an m x 1 vector z 1+ 1, we have rl,t+ll Zt+l =' [ Xt+l ·

(3)

St+l

We will call z 1+ 1 the state vector and we assume a first-order vector autoregression for z1+1: (4) where 0, W ∗ can be 0, with positive probability and log 0 is not defined (−∞). This limitation to f < 1 for periodic portfolio revisions is not necessary when the portfolio is continuously adjusted because the portfolio can be revised before total loss occurs. From eqn. (5),   f G(W ∗ ) ≡ log W ∗ = log (1 − f )erT + max(x − K, 0) C g(W ∗ ) ≡ E(G(W ∗ )) = E log W ∗  ∞ (log W ∗ )q(x)dx g(W ∗ ) =

(9) (10)

0

=



K

0

+



log{(1 − f )erT }q(x)dx ∞

K



log (1 − f )e

rT

 f + (x − K) q(x)dx C

where we replaced max(x − K, 0) in the last expression by x − K since they are equal over the range of integration, K ≤ x < ∞. We compute the corresponding values for stddev (G) = sqrt(Var G) using Var G = E(G2 ) − [E(G)]2

(11)

854

EO Thorp and S Mizusawa

Note that E(G) ≡ g(W ∗ ) has already been computed from equation (10). Similarly, we have  ∞ E(G2 ) = (log W ∗ )2 q(x)dx (12) 0

=



K

0

+

[log{(1 − f )erT }]2 q(x)dx





K



 2  f log (1 − f )erT + (x − K) q(x)dx C

g(W ∗ ) − r Sharpe(G) = stddev(G)

(12s)

Values of g = g(W ∗ ), v = stddev(G), and Sharpe (G) were computed from equations (10), (11), (12) and (12s) using Mathematica. They are displayed in Tables 1, 2 and 3. The tables show there are very different portfolios which have approximately the same g and σ(G) as the index. For example, Portfolio 1: K = 0.9, f = 0.2 has g = 0.0986, σ(G) = 0.2103, close to the index at K = 0.0, f = 1.0, g = 0.1000, σ(G) = 0.2000. Also interesting is Portfolio 2: K = 0.8, f = 0.2 which has g = 0.0918, σ(G) = 0.1580, giving up some growth for a risk reduction of about 1/5.

The Efficient Frontier To compute values of g, σ and Sharpe for portfolios combining the index and T-bills, with K = 0, f = 0.1 to 0.9, we modify equations (9), (10) and (12) by replacing f c max(x − K, 0) with f x, yielding G(W ∗ ) ≡ log W ∗ = log{(1 − f )erT + f x}  ∞ ∗ g(W ) = log{(1 − f )erT + f x}q(x)dx

(9a) (10a)

0

E(G2 ) =



∞ 0

[log{(1 − f )erT + f x}]2 q(x)dx

(12a)

where the first integral drops out for (10a) and (12a) because K = 0. The column f = 1 is the special case where no money is invested in T-bills. However, we can’t put it all in call options with K fixed and greater than zero because that gives a positive probability of total loss, which is forbidden by the policy max E log(W ∗ ). The sole admissible f = 1 investment is 100% in the index so we have, for the K = 0 rows only, g = 0.10, v = 0.20, Sharpe = 0.25, as shown.

K 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.9 2

f

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

0.0570 0.0577 0.0586 0.0596 0.0610 0.0629 0.0655 0.0692 0.0742 0.0801 0.0856 0.0884 0.0858 0.0753 0.0565 0.0322 0.0071 −0.0145 −0.0306 −0.0413 −0.0478

0.0635 0.0648 0.0664 0.0684 0.0708 0.0741 0.0785 0.0845 0.0918 0.0986 0.1014 0.0953 0.0760 0.0425 −0.0012 −0.0476 −0.0889 −0.1208 −0.1427 −0.1564 −0.1643

0.0695 0.0713 0.0735 0.0762 0.0795 0.0837 0.0893 0.0962 0.1034 0.1073 0.1019 0.0803 0.0390 −0.0200 −0.0877 −0.1529 −0.2069 −0.2464 −0.2724 −0.2881 −0.2971

0.0751 0.0773 0.0800 0.0832 0.0871 0.0919 0.0978 0.1045 0.1095 0.1068 0.0881 0.0458 −0.0219 −0.1082 −0.1998 −0.2825 −0.3479 −0.3939 −0.4233 −0.4408 −0.4505

0.0803 0.0828 0.0858 0.0894 0.0936 0.0986 0.1043 0.1094 0.1097 0.0964 0.0591 −0.0097 −0.1084 −0.2249 −0.3413 −0.4416 −0.5179 −0.5701 −0.6027 −0.6218 −0.6323

0.0850 0.0878 0.0910 0.0948 0.0990 0.1039 0.1086 0.1106 0.1032 0.0744 0.0116 −0.0910 −0.2272 −0.3783 −0.5219 −0.6409 −0.7285 −0.7870 −0.8229 −0.8436 −0.8548

0.0894 0.0923 0.0956 0.0994 0.1035 0.1077 0.1105 0.1076 0.0883 0.0368 −0.0614 −0.2092 −0.3928 −0.5860 −0.7616 −0.9020 −1.0024 −1.0679 −1.1074 −1.1298 −1.1418

0.0933 0.0963 0.0996 0.1032 0.1069 0.1099 0.1098 0.0990 0.0611 −0.0257 −0.1772 −0.3903 −0.6400 −0.8897 −1.1070 −1.2745 −1.3909 −1.4650 −1.5089 −1.5334 −1.5465

0.0968 0.0998 0.1030 0.1063 0.1093 0.1105 0.1057 0.0813 0.0097 −0.1421 −0.3905 −0.7188 −1.0810 −1.4237 −1.7076 −1.9174 −2.0582 −2.1455 −2.1961 −2.2238 −2.2384

0.1000

Chapter 40. Maximizing Capital Growth With Black Swan Protection

Table 1: Geometric Growth, Log Normal Model. Mean Yearly Geometric Growth: m = .1, Yearly Volatility: v = .2, Riskless Rate: r = .05, Option Life: T = 1 year Strike Price: K, Fraction of Capital invested: f

855

856

EO Thorp and S Mizusawa

Table 2: Standard Deviation of Geometric Growth, Log Normal Model. m = .1, v = .2, r = .05, T = 1, Strike: K, Fraction Invested: f K

f

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.9 2

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

0.0214 0.0236 0.0263 0.0298 0.0343 0.0403 0.0490 0.0620 0.0817 0.1105 0.1502 0.2006 0.2584 0.3156 0.3605 0.3825 0.3772 0.3480 0.3032 0.2518 0.2012

0.0423 0.0466 0.0519 0.0586 0.0674 0.0791 0.0959 0.1209 0.1580 0.2103 0.2778 0.3559 0.4343 0.4980 0.5327 0.5312 0.4962 0.4375 0.3675 0.2966 0.2317

0.0627 0.0691 0.0769 0.0868 0.0996 0.1169 0.1415 0.1781 0.2319 0.3057 0.3964 0.4944 0.5830 0.6440 0.6637 0.6396 0.5801 0.4992 0.4112 0.3268 0.2521

0.0829 0.0912 0.1015 0.1145 0.1313 0.1540 0.1866 0.2351 0.3060 0.4011 0.5135 0.6276 0.7219 0.7760 0.7788 0.7327 0.6510 0.5508 0.4475 0.3517 0.2689

0.1027 0.1131 0.1258 0.1418 0.1626 0.1910 0.2318 0.2932 0.3828 0.5006 0.6349 0.7638 0.8611 0.9055 0.8895 0.8208 0.7173 0.5986 0.4810 0.3746 0.2844

0.1224 0.1347 0.1498 0.1689 0.1939 0.2281 0.2780 0.3541 0.4654 0.6094 0.7678 0.9115 1.0096 1.0415 1.0040 0.9110 0.7845 0.6469 0.5147 0.3976 0.2999

0.1419 0.1561 0.1737 0.1960 0.2253 0.2658 0.3261 0.4202 0.5590 0.7358 0.9234 1.0834 1.1805 1.1957 1.1324 1.0110 0.8587 0.6999 0.5515 0.4228 0.3168

0.1613 0.1775 0.1977 0.2233 0.2571 0.3048 0.3779 0.4964 0.6743 0.8975 1.1249 1.3055 1.3990 1.3907 1.2929 1.1350 0.9501 0.7650 0.5967 0.4535 0.3375

0.1806 0.1990 0.2217 0.2508 0.2898 0.3459 0.4371 0.5963 0.8435 1.1494 1.4453 1.6584 1.7431 1.6942 1.5400 1.3245 1.0890 0.8635 0.6649 0.4999 0.3687

0.2000

0.9

1

Table 3: Sharpe of Geometric Growth, Log Normal Model. m = .1, v = .2, r = .05, T = 1, Strike: K, Fraction Invested: f K 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.9 2

f

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.3271 0.3263 0.3253 0.3241 0.3224 0.3201 0.3166 0.3099 0.2962 0.2722 0.2369 0.1916 0.1385 0.0800 0.0179 −0.0466 −0.1137 −0.1852 −0.2657 −0.3624 −0.4858

0.3194 0.3177 0.3157 0.3130 0.3096 0.3048 0.2976 0.2855 0.2644 0.2311 0.1850 0.1273 0.0600 −0.0150 −0.0962 −0.1837 −0.2799 −0.3903 −0.5243 −0.6958 −0.9250

0.3114 0.3088 0.3057 0.3016 0.2962 0.2887 0.2775 0.2595 0.2304 0.1876 0.1308 0.0614 −0.0189 −0.1086 −0.2074 −0.3172 −0.4429 −0.5937 −0.7839 −1.0348 −1.3765

0.3032 0.2997 0.2954 0.2897 0.2823 0.2719 0.2564 0.2320 0.1944 0.1416 0.0741 −0.0067 −0.0996 −0.2039 −0.3207 −0.4538 −0.6112 −0.8059 −1.0578 −1.3957 −1.8612

0.2948 0.2904 0.2848 0.2775 0.2679 0.2544 0.2342 0.2027 0.1558 0.0927 0.0143 −0.0782 −0.1840 −0.3036 −0.4399 −0.5989 −0.7917 −1.0358 −1.3571 −1.7933 −2.3990

0.2863 0.2808 0.2739 0.2649 0.2530 0.2361 0.2107 0.1713 0.1143 0.0400 −0.0500 −0.1547 −0.2745 −0.4112 −0.5696 −0.7584 −0.9923 −1.2938 −1.6961 −2.2473 −3.0170

0.2775 0.2709 0.2627 0.2519 0.2375 0.2169 0.1855 0.1371 0.0686 −0.0180 −0.1207 −0.2392 −0.3751 −0.5319 −0.7167 −0.9416 −1.2256 −1.5972 −2.0985 −2.7906 −3.7618

0.2685 0.2608 0.2511 0.2385 0.2214 0.1966 0.1582 0.0987 0.0165 −0.0843 −0.2020 −0.3373 −0.4932 −0.6757 −0.8950 −1.1670 −1.5166 −1.9805 −2.6126 −3.4913 −4.7304

0.2594 0.2500 0.2505 0.2393 0.2246 0.2045 0.1749 0.1273 0.0524 −0.0478 −0.1671 −0.3048 −0.4635 −0.6488 −0.8699 −1.1413 −1.4854 −1.9360 −2.5427 −3.3782 −4.5482 −6.2074

Chapter 40. Maximizing Capital Growth With Black Swan Protection

857

The column f = 0 is the special case where all the money is invested in T-bills. Since no options are purchased, K is irrelevant. The value of g is 0.05, the standard deviation is (conventionally) assumed to be zero and the Sharpe ratio is (0 − 0)/0, undefined. This column is omitted. A look at Tables 1, 2, and 3 shows that the geometric Sharpe ratio is greatest for a given f when K = 0, i.e. when we buy the index rather than call options. Also the Sharpe ratio decreases, for a given f , as we increase K, i.e. raise the strike price of the call options in the portfolio. Figure 1 plots g versus v for those points in the Table such that g ≥ 0.05 and v ≤ 0.50. Most of these points fall on, or very close to, a downward opening “parabola” approximating g = .05 + .35v − .5v 2 . The part left of, and including the peak, indicates the geometric mean/standard deviation efficient frontier. Maximum compound growth occurs at about g = 0.1106, v = 0.35(f = 0.6, K = 0.7). Note that the geometric efficient frontier is concave unlike the arithmetic efficient frontier, which with its constant Sharpe ratio, is a straight line. For instance, a straight line joining T-bills at g = 0.05, v = 0 with the index at g = 0.10, v = .20, crosses v = 0.10 at g = 0.075 whereas the efficient frontier value of g = 0.08 is about 0.5%/year higher The points from the column f = 0.1 start out on the efficient frontier but as K increases they fall away in their own lower “parabola,” which peaks at K = 1.1, g = 0.88, v = 0.20, Sharpe = 0.19. The points from f = 0.2 break away in a somewhat higher curve, peaking at K = 1.0, g = 0.10, v = 0.28, Sharpe = 0.19. The f = 0.3 points continue the pattern. For f ≥ 0.4 the points remain very close to the efficient frontier over the range of values in the plot. The shaded zone in Table 1 indicates the points in the Figure which appear to fall on the efficient frontier. From both Figure 1 and the Tables, it appears that we can produce any point on the geometric efficient frontier between T-bills (v = 0.00, g = 0.05) and the index (v = 0.20, g = 0.10) using a mix of only the index and treasury bills. Thus, for this part of the efficient frontier, using the metric of the geometric Sharpe ratio, call options on the index in the portfolio appear to have no advantage over using the index directly. However, the part of the efficient frontier between v = 0.20, g = 0.10 and v = .35, g = 0.11 comes only from call options plus T-bills. The (f, K) pairs which appear here, shaded in Table 2, are a subset of the shaded points in Table 1. Though using options for “Black Swan insurance” seems to confer no long run growth benefits, intuition suggests we look at short and intermediate payoff structures.

4. Simulations Because options protect against large losses, one wonders whether there are characteristics, such as less severe maximum drawdowns, that would make portfolios 1 or 2, for instance, “better” than the index. To explore this, we compare the maximum

858

0.12

0.11

f1 f2 f3 f4 f5 f6 f7 f8 f9 RisklessRate StockIndex

0.09

0.08

0.07

0.06

0.05

0.04 0

0.05

0.1

0.15

0.2

0.25

0.3

0.35

0.4

0.45

0.5

Standard Deviation Figure 1:

Geometric growth versus standard deviation for values in the Tables where g >= .05 and v 0

(17)

Chapter 40. Maximizing Capital Growth With Black Swan Protection

865

Again, the index pays no dividends and the portfolio is revised annually. Calls are European, i.e. only exercisable at expiration. Recall that the Black-Scholes formula for European calls results if we take the lognormal distribution and replace the expected growth rate by the riskless rate. If this mean-growth-rate-shifted lognormal distribution is w(x) then  ∞ −rt (x − K)w(x)dx (18) C =e x=K

However, because we have only specified a terminal distribution of index prices, and not one resulting from a process with known transition probabilities, we don’t have a “no arbitrage” model for call option pricing.1 Proceeding by analogy with equation (18), and choosing T = 1 (year), we priced the options in a risk-neutral setting using the fat-tailed density function of equation (17). Let u(x) equal p(x) after shifting its mean so that its expected growth rate equals the riskless rate, i.e.  ∞  ∞ u(x)dx = er u(x)dx = A

−∞

Note that when p(x) is mean-shifted to get u(x), the point where u(x) becomes non-zero is at some number A < 0, whereas for p(x) it was at 0. Now calculate C from  ∞ (x − K)u(x)dx (19) C = e−r x=K

Note that K > 0 > A so we don’t ever include the values of x where u(x) = 0. We obtained the density function u(x) by shifting the mean of p(x) to µ ˜ = exp(.05) = 1.05127, the one year wealth relative at our assumed riskless compound growth rate of 5%. The difference between the mean of u and µ ˜ is 1.12749 − 1.05127 = 0.07622 so we change −1.12562 in (17) to −1.2562 − 0.07622 = −1.04940 to get u(x) in equation (20). u(x) = 72.416(4 + 36.3275(−1.04940 + x)2 )−5/2

(20)

for x ≥ −0.07622 and u(x) = 0

for x < −0.07622.

Alternatively, we could have truncated the risk neutral t-distribution at zero, for µ ˜ and σ, finding it by the same process we used to find p(x). Table 4 compares the call prices we used for Extremistan with those for Mediocristan. The differences are small in magnitude across the entire range. However, the truncated t distribution leads to much greater maximum drawdowns than in the lognormal case as Figures 11 and 12 show. Figure 11 shows the 1 See Ekstrom, et al., Quantitative Finance Vol. II, No. 8, August 2011, page 1125, for a discussion of when we have no-arbitrage pricing or risk-neutral pricing models. Also see J. Huston McCulloch, “The Risk-Neutral Measure and Option Pricing Under Log-Stable Uncertainty,” http://www.econ. ohio-state.edu/jhm/papers/rnm.pdf.

866

EO Thorp and S Mizusawa Table 4: Call Prices Compared for Lognormal and Truncated Student t. K

Lognormal

.2 .4 .6 .8 1.0 1.2 1.4 1.6 1.8 2.0

.809 .619 .429 .245 .104 .0325 .00785 .00158 .000286 .0000479

Figure 11:

Figure 12:

Student t .809 .620 .433 .253 .105 .0299 .00861 .00315 .00142 .000748

Chapter 40. Maximizing Capital Growth With Black Swan Protection

867

simulated cumulative distribution of the maximum drawdown for the lognormal. The five curves represent fractions of 0.2, 0.4, 0.6, 0.8 and 1.0 in the index. As the fraction increases, the curves of course move progressively to the right. Figure 12 displays the same graphs for the truncated t distribution.

6. Geometric Growth, Standard Deviation and Sharpe Ratio Proceeding as before we compute g, σ and the Sharpe ratio from equations (9)–(12), (9a)–(12a), and (12s), replacing the lognormal g(x) by the truncated Student t u(x) throughout. The results, in Figure 13 and Tables 5–7, correspond to those for Mediocristan, Figure 1 and Tables 1–3. Figure 13 shows that for σ ≤ 0.10, g ≤ 0.08, the option and index portfolios lie close to a common curve. To prefer one over the other requires a different metric. In the range 0.10 < σ < 0.24, the option portfolios have a greater g for a given σ, or alternatively, less σ for a given g, with the exception of f = 0.1. Above σ = 0.24, the option portfolios have an efficient frontier peak at f = 0.4, K = 0.9, with g = 0.1168, σ = 0.3593. We illustrate some possible tradeoffs. It is helpful to use Figure 13 to compare the points for the pairs of portfolios in our examples. Figure 14 shows the MDD distribution for the option portfolio K = 1.1, f = 0.1, T = 32 years, where g = 0.0869, σ = 0.1926, and Sh = 0.1915, versus the index portfolio f = 1.0, which has g = 0.0968, σ = 0.2333, and Sh = 0.2005. Although the option portfolio has an annualized growth rate that is 1% less, and is well inside the efficient frontier, the reduction in MDD is enormous — a great comfort to the portfolio manager who wants to retain his clients and his job. The index portfolio has a 50% MDD with probability about 40% while the option portfolio rarely has an MDD this large. The investor Michael Korns has followed a similar strategy for more than a decade. Just as in the Mediocristan examples, the MDD distribution at shorter times typically favors an index portfolio for small MDD and a “comparable” option portfolio for large MDD, with the option MDD becoming more dominant as T increases. When K = 0.9, f = 0.2, T = 32 the option portfolio has g = 0.0999, σ = 0.1878, Sh = 0.2656 whereas the index portfolio with f = 1.0 is inside the efficient frontier at g = 0.0968, σ = 0.2333 and Sh = 0.2005. The dramatic reduction in “tail risk” is illustrated in Figure 15. The portfolio with the highest growth is K = 0.9, f = 0.4. Sitting at the peak of the efficient frontier, it yields g = 0.1168, σ = 0.3593, and Sh = 0.2274. Although g and Sh are much better than the index, the MDD graphs in Figure 16 show a “ride” so wild few investors are likely to stay with it.

868

0.1300

0.1200

0.1100 f1 f2 f3 0.0900

f4 f5 f6

0.0800

f7 0.0700

f8

0.0600

RisklessRate

f9 StkIndex 0.0500

0.0400 0.0000

0.0500

0.1000

0.1500

0.2000

0.2500

0.3000

0.3500

0.4000

0.4500

0.5000

Standard Deviation Figure 13:

Geometric growth versus standard deviation for values in the Tables where g >= .05 and v K 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.9 2

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.0570 0.0577 0.0585 0.0596 0.0610 0.0627 0.0652 0.0685 0.0733 0.0796 0.0861 0.0869 0.0741 0.0493 0.0219 −0.0008 −0.0174 −0.0289 −0.0366 −0.0418 −0.0454

0.0635 0.0648 0.0663 0.0683 0.0707 0.0738 0.0780 0.0836 0.0910 0.0999 0.1055 0.0950 0.0565 −0.0003 −0.0540 −0.0940 −0.1207 −0.1379 −0.1489 −0.1561 −0.1608

0.0695 0.0713 0.0734 0.0760 0.0793 0.0834 0.0886 0.0954 0.1039 0.1122 0.1118 0.0832 0.0135 −0.0763 −0.1543 −0.2089 −0.2436 −0.2652 −0.2786 −0.2872 −0.2928

0.0750 0.0772 0.0798 0.0829 0.0867 0.0913 0.0971 0.1042 0.1120 0.1168 0.1059 0.0534 −0.0522 −0.1768 −0.2786 −0.3467 −0.3887 −0.4140 −0.4296 −0.4393 −0.4456

0.0801 0.0826 0.0854 0.0888 0.0928 0.0976 0.1033 0.1096 0.1150 0.1135 0.0871 0.0043 −0.1429 −0.3052 −0.4317 −0.5134 −0.5624 −0.5914 −0.6089 −0.6197 −0.6267

0.0847 0.0873 0.0903 0.0938 0.0977 0.1022 0.1070 0.1113 0.1122 0.1007 0.0529 −0.0685 −0.2655 −0.4707 −0.6243 −0.7204 −0.7767 −0.8095 −0.8290 −0.8410 −0.8486

0.0887 0.0914 0.0943 0.0976 0.1011 0.1046 0.1076 0.1083 0.1019 0.0753 −0.0026 −0.1751 −0.4348 −0.6920 −0.8774 −0.9901 −1.0547 −1.0917 −1.1134 −1.1266 −1.1349

0.0922 0.0947 0.0974 0.1000 0.1025 0.1043 0.1040 0.0985 0.0802 0.0301 −0.0930 −0.3391 −0.6857 −1.0126 −1.2394 −1.3732 −1.4481 −1.4903 −1.5148 −1.5295 −1.5387

0.9

1

0.0951 0.0968 0.0971 0.0990 0.1004 0.1009 0.0991 0.0925 0.0756 0.0353 −0.0580 −0.2623 −0.6371 −1.1309 −1.5723 −1.8651 −2.0320 −2.1229 −2.1731 −2.2018 −2.2188 −2.2293

Table 6: Std. Deviation for Extremistan, r = .05, T = 1 year, using truncated Student T distribution with 4 degrees of freedom, Mean of 1.12749, Variance of 0.0518805. 7/14/2011. f–> K 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.9 2

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.0214 0.0236 0.0263 0.0297 0.0339 0.0395 0.0470 0.0577 0.0737 0.0988 0.1379 0.1926 0.2511 0.2924 0.3071 0.3004 0.2815 0.2579 0.2337 0.2108 0.1901

0.0425 0.0469 0.0522 0.0588 0.0670 0.0778 0.0924 0.1128 0.1426 0.1878 0.2547 0.3401 0.4178 0.4559 0.4505 0.4185 0.3765 0.3339 0.2949 0.2607 0.2312

0.0635 0.0700 0.0779 0.0877 0.1000 0.1160 0.1374 0.1671 0.2100 0.2733 0.3638 0.4723 0.5595 0.5874 0.5604 0.5057 0.4447 0.3874 0.3374 0.2950 0.2593

0.0845 0.0932 0.1038 0.1169 0.1333 0.1546 0.1830 0.2221 0.2781 0.3593 0.4719 0.6003 0.6926 0.7071 0.6576 0.5812 0.5028 0.4326 0.3731 0.3236 0.2827

0.1056 0.1168 0.1302 0.1467 0.1676 0.1945 0.2303 0.2796 0.3494 0.4494 0.5845 0.7318 0.8267 0.8249 0.7515 0.6531 0.5576 0.4748 0.4063 0.3502 0.3043

0.1272 0.1409 0.1575 0.1779 0.2036 0.2368 0.2810 0.3417 0.4271 0.5481 0.7081 0.8752 0.9707 0.9492 0.8491 0.7270 0.6134 0.5177 0.4398 0.3769 0.3261

0.1496 0.1662 0.1864 0.2113 0.2427 0.2834 0.3377 0.4121 0.5166 0.6631 0.8529 1.0427 1.1371 1.0909 0.9589 0.8093 0.6753 0.5650 0.4767 0.4063 0.3500

0.1731 0.1935 0.2182 0.2488 0.2877 0.3382 0.4058 0.4987 0.6291 0.8103 1.0405 1.2599 1.3510 1.2708 1.0967 0.9118 0.7519 0.6233 0.5221 0.4424 0.3792

0.9

1

0.1989 0.2333 0.2249 0.2567 0.2962 0.3467 0.4129 0.5022 0.6258 0.7996 1.0400 1.3384 1.6063 1.6897 1.5521 1.3098 1.0689 0.8687 0.7119 0.5908 0.4970 0.4235

870

EO Thorp and S Mizusawa

Table 7: Sharpe for Extremistan, r = .05, T = 1 year, using truncated Student T distribution with 4 degrees of freedom, Mean of 1.12749, Variance of 0.0518805. 7/14/2011. f–> K 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.9 2

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.3265 0.3255 0.3247 0.3240 0.3235 0.3231 0.3225 0.3208 0.3154 0.3000 0.2622 0.1915 0.0961 −0.0025 −0.0914 −0.1693 −0.2396 −0.3058 −0.3706 −0.4356 −0.5021

0.3173 0.3152 0.3132 0.3111 0.3089 0.3063 0.3030 0.2977 0.2877 0.2656 0.2180 0.1324 0.0156 −0.1103 −0.2308 −0.3439 −0.4533 −0.5626 −0.6744 −0.7905 −0.9118

0.3073 0.3040 0.3006 0.2969 0.2927 0.2877 0.2812 0.2720 0.2567 0.2274 0.1698 0.0703 −0.0653 −0.2151 −0.3645 −0.5119 −0.6603 −0.8135 −0.9738 −1.1430 −1.3217

0.2965 0.2919 0.2869 0.2814 0.2750 0.2673 0.2575 0.2439 0.2229 0.1860 0.1184 0.0056 −0.1476 −0.3207 −0.4996 −0.6825 −0.8724 −1.0727 −1.2854 −1.5120 −1.7530

0.2850 0.2788 0.2721 0.2645 0.2556 0.2450 0.2315 0.2132 0.1861 0.1413 0.0635 −0.0624 −0.2334 −0.4306 −0.6410 −0.8627 −1.0983 −1.3507 −1.6218 −1.9126 −2.2234

0.2726 0.2646 0.2559 0.2459 0.2344 0.2204 0.2028 0.1794 0.1457 0.0925 0.0041 −0.1354 −0.3250 −0.5485 −0.7941 −1.0598 −1.3477 −1.6602 −1.9986 −2.3637 −2.7554

0.2591 0.2489 0.2378 0.2252 0.2105 0.1928 0.1706 0.1415 0.1005 0.0382 −0.0617 −0.2159 −0.4264 −0.6802 −0.9671 −1.2852 −1.6358 −2.0206 −2.4405 −2.8956 −3.3857

0.2441 0.2312 0.2170 0.2011 0.1826 0.1605 0.1330 0.0973 0.0481 −0.0246 −0.1375 −0.3089 −0.5446 −0.8362 −1.1757 −1.5609 −1.9925 −2.4713 −2.9973 −3.5702 −4.1893

0.9

1

0.2266 0.2005 0.2093 0.1908 0.1702 0.1467 0.1188 0.0847 0.0409 −0.0184 −0.1039 −0.2333 −0.4277 −0.6989 −1.0453 −1.4622 −1.9477 −2.5015 −3.1229 −3.8111 −4.5647 −5.3825

Figure 14: Comparison of the cumulative distribution of Maximum Drawdowns for Options with K = 1.1 and f = .1 for T = 32 years (•) and Stock Index (fraction = 1.0) (×).

7. Caveats and Conclusions As there is no generally accepted fat-tailed distribution for describing equity index returns, we have made an arbitrary illustrative choice. Our truncated t-distribution is not inconsistent with the observed annual extreme returns over the last 86

Chapter 40. Maximizing Capital Growth With Black Swan Protection

871

Figure 15: Comparison of the cumulative distribution of Maximum Drawdowns for Options with K = .9 and f = .2 for T = 32 years (•) and Stock Index (fraction = 1.0) (×).

Figure 16: Comparison of the cumulative distribution of Maximum Drawdowns for Options with K = .9 and f = .4 for T = 32 years (•) and Stock Index (fraction = 1.0) (×).

calendar years. Some will argue that, even so, it is much too tame. We have also made the simplifying assumptions that options are only exercisable at expiration and that they can be priced by using the expected terminal value of the payoff, discounted at the riskless rate. Anyone who wanted to apply the methods given here would need to make their own set of choices for return distribution, option pricing model, type of option, portfolio revision period, parameters like µ, σ, and r, etc. Nonetheless, we would expect certain qualitative features of our results to persist more generally, such as: (1) Options portfolios can attain regions of the geometric mean-variance efficient frontier beyond the reach of the index portfolios. (2) If we regard one maximum drawdown distribution as better than another if its right tail dominates (i.e. less chance of extreme MDDs), then some options portfolios are both on the efficient frontier and have better MDDs than the index portfolio which is the closest in geometric mean, standard deviation, and Sharpe ratio.

872

EO Thorp and S Mizusawa

In any practical application we need to include transactions costs and the impact of taxes, either or both of which could offset the perceived advantages. One also might want to stagger option expiration dates, e.g. 1/4 per quarter, to smooth out costs, payoffs, and the need to replace — if American options are used — options which are exercised early.

Reference Bertocchi, M., Schwartz, S., and Ziemba W., Optimizing the Aging, Retirement and Pensions Dilemma. Wiley, N.Y. 2010.

Author Index Abada, C, 613 A¨ıt-Sahalia, Y, 814, 837 Akian, M, 790, 802 Algoet, PH, 790, 796, 802, 807 Allais, M, 101, 105, 113, 129, 132, 253, 255, 271, 477 Arditti, FD, 162 Arkin, VI, 802 Arrow, KJ, 31, 58, 87, 88, 100, 253, 318, 330, 312, 331, 391, 469, 806 Aumann, R, 87, 88, 191

Chan, YL, 809 Chipman, J, 265 Chernoff, H, 87 Chernov, M, 450 Chew, SH, 207–209, 216, 218, 222, 227 Chopra, V, 313, 365, 450, 457, 499 Cicchetti, CJ, 473, 476 Clark, HH, 274, 668 Clark, EV, 274, 668 Cocco, J, 811 Connor, G, 446 Coombs, CH, 113 Cover, TM, 638, 789, 790, 796, 802–804, 807 Cox, J, 59, 62 Crum, R, 270 Cvitani´c, J, 793, 803, 804

Bachelier, L, 31, 53 Bahsoun, W, 789, 802 Balduzzi, P, 811 Baltussen, G, 83, 177 Barberis, NC, 604, 811, 813, 845 Bar-Hillel, M, 266 Barrager, S, 183 Baumol, WJ, 166 Bell, RM, 638 Benartzi, S, 474 Bernouli, D, 269, 280, 282, 637 Bernoulli, N, 253 Bertocchi, M, 850 Black, F, 11, 31, 59, 849 Blazenko, G, 649, 805 Boyd, JH, 214, 217 Brandt, M, 311, 814, 837 Breiman, L, 638, 789, 790, 802 Brennan, MJ, 810–812, 840, 845 Brooks-Hill, FJ, 310, 375, 390 Brown, D, 276 Browne, S, 449, 605, 789, 790, 802, 803

Dammon, RM, 845 Dana, RA, 803, 804 Davidson, D, 87 Davis, M, 606, 607, 639 de Finetti, B, 475 Debreu, G, 58 Dekel, E, 207, 208, 218 Delbaen, F, 6 Dempster, MAH, 790, 803 Dexter, AS, 367, 373, 390, 392 Donaldson, JB, 215 Dorfman, R, 801, 803 Dr`eze, J, 181 Dubin, JA, 473, 476 Dvoretzky, A, 799, 803 Dynkin, EB, 790 Eaves, BC, 380 Eden, B, 183 Edirisinghe, C, 605, 611, 613 Edwards, W, 112, 253 Efron, B, 446, 453 Ellsberg, D, 125

Camerer, C, 248 Campbell, JY, 612, 811–823, 827, 832, 834, 837, 840, 845, 899 Cari˜ no, DR, 491, 494 Cass, D, 409 Chacko, G, 814

873

874

Author Index

Epstein, LG, 85, 207, 208, 209, 216, 222, 227, 230, 248, 815 Erakar, SE, 270 Evstigneev, IV, 789, 799, 802–804, 807 Fama, EF, 4, 311, 421–423, 428, 612, 613, 811, 821 Feller, W, 441 Fellner, W, 113, 125 Fischhoff, B, 270, 271, 274, 277 Fishburn, PC, 80, 87, 104, 113, 116, 137, 153, 270 Fl˚ am, SD, 799, 803 F¨ ollmer, H, 484 Foster, M, 314, 445 French, F, 4, 311, 612, 613, 811 Freund, RA, 376, 384, 392 Friedman, M, 82, 89, 130, 184 Frost, PA, 446 Galanter, E, 114 Gale, D, 606, 790, 804 Gassman, HI, 345, 416 Geyer, A, 313, 488, 491, 605 Giovannini, A, 815, 821 Glosten, LR, 811, 821 Goetzmann, WN, 827 Graham, D, 612 Grauer, RR, 446, 450 Gregory, R, 276 Grether, DM, 120 Grossman, SJ, 208, 822 Gundlach, VM, 802, 807 Hadar, J, 135, 153, 303 Hagen, O, 271 Hakansson, NH, 446, 450, 790, 804 Hall, RE, 208 Hammack, J, 276 Hanoch, G, 135, 139, 153, 280, 303, 311, 412 Hansen, L, 208, 228 Hansson, B, 113 Harrison, JM, 6, 32, 34, 55, 59, 63 Harvey, CR, 162 Hausch, DB, 8, 639, 666, 790, 804 Heath, D, 789, 790, 806 Hens, T, 449, 639, 803 Hershey, 270, 277 Herstein, IN, 88 Hodrick, RJ, 813, 821, 827

Intrilligator, MD, 806 Iyengar, G, 790, 804 Jacobs, BL, 5 Jensen, NE, 90 Johnsen, TH, 215 Jones, CS, 446, 449 Jorion, P, 447, 456, 485, 827 Jouini, E, 67, 793, 804 Kabanov, YM, 804 Kahneman, D, 82, 83, 99, 129–132, 134, 137, 140–142, 146, 147, 151, 153, 159, 164, 165, 167, 178, 248, 255, 261, 269, 270, 271, 274, 474, 477, 478 Kallal, H, 67, 793, 804 Kallberg, JG, 312, 313, 365, 370, 373, 409, 411, 415, 446, 494 Kallio, M, 7 Kandel, S, 473, 813 Kannai, Y, 191 Karatzas, I, 793, 803 Kelly, JL, 638, 789, 790, 805 Kihlstrom, RE, 220–222 Kim, TS, 612, 811, 827 King, AJ, 59, 67, 74–76 Kira, D, 389, 409 Knetsch, J, 276 Knispel, T, 484 Kochenberger, GA, 104, 116, 270 Koopmans, TC, 214, 215 Korajczyk, RA, 446 Kreps, DM, 6, 34, 39, 55, 59, 63, 181, 207, 208, 209, 217 Kuhn, HW, 804 Kunreuther, H, 277 Larsson, S, 103, 118 Latan`e, H, 638, 789, 790, 805 Laughhunn, DJ, 270 Levy, M, 129, 149, 299 Levy, H, 5, 82, 83, 135, 137, 139, 145, 153, 159, 162, 177, 178, 280, 282, 303, 311, 412 Li, Y, 333 Lichtenstein, S, 120, 270, 271, 274, 277 Lintner, J, 11, 149, 376, 378, 613 Liu, J, 811 Lleo, S, 606, 607, 639 Lo, AW, 492, 666 Lo´eve, M, 441, 442

Author Index

Loffler, G, 445 Long, JB, 789, 798, 805 Lucas Jr, RE, 214, 216, 217, 230 Luce, RD, 87, 90 Lynch, AW, 811, 845 MacCrimmon, KR, 103, 118 Machina, MJ, 208, 209, 219, 227, 248 MacKinlay, AC, 446–450, 456, 457 MacLean, LC, 284, 314, 445, 459, 488, 606, 637, 639, 643, 644, 649, 650, 789, 790, 805 Makarov, VL, 790, 805 Mangasarian, OL, 60, 379 Markowitz, HM, 4, 82, 104, 105, 112, 115, 117, 129, 130, 131, 142, 146, 149, 152, 167, 253, 283, 293, 338, 365, 373, 648, 484, 485, 810 Mehra, R, 208 Merton, RC, 31, 59, 458, 492, 810 Miller, L, 5 Milnor, J, 88 Mirman, LJ, 220, 221 Mizusawa, S, 608 Modigliani, F, 5, 181 Morgenstern, O, 80, 129, 271, 253 Morgerra, SD, 447, 448 Morris, C, 446, 453 Mossin, J, 181, 299 Mulvey, JM, 491 Murty, KG, 380 Musiela, M, 804 Naik, V, 67 Napp, C, 59, 60, 67, 70 Nelson, CR, 827 Nikaido, H, 805 Ohlson, JA, 284, 612 Parkan, C, 310, 375, 390 Pastor, L, 446–450, 456, 457 Payne, JW, 270 Pham, H, 793, 805 Platen, E, 789, 790, 806 Pliska, SR, 6, 32 Plott, CR, 120 Porteus, EL, 181, 207, 208, 209, 217 Post, T, 177 Pratt, JW, 87, 100, 253, 312, 391

875

Prescott, E, 208 Press, SJ, 421, 423, 435 Quiggin, J, 137, 139, 159 Rabin, M, 314, 469, 470, 472 Radner, R, 790, 806 Raiffa, H, 87, 89, 120 Ramsey, FP, 87 Rao, CR, 586 Rockafellar, RT, 484, 494, 790, 806 Rogers, LCG, 449 Roll, R, 421, 423 Ross, SA, 5, 58, 59, 61, 62, 283, 333 Rothschild, M, 135, 253, 284, 303 Roy, AD, 484 Rubin, H, 87 Rubinov, AM, 790, 801, 805, 807 Rubinstein, M, 162, 312, 334, 391, 810 Rudolf, M, 492 Russell, WR, 135, 153, 303 Samuelson, P, 247, 248, 284, 376, 390, 421, 422, 428, 438, 471, 472, 474, 800, 801, 803, 810 Sanegre, R, 805 Sarnat, M, 162 Savage, LJ, 80, 82, 88, 89, 130, 184, 253, 275 Savarino, JE, 446 Schachermayer, W, 6 Schenk-Hopp´e, KR, 449, 639, 803 Schlaifer, R, 87, 272 Schmeidler, D, 799 Schoemaker, PJH, 270, 277 Scholes, M, 31, 849 Schroder, M, 811 Schwartz, S, 5, 612 Schwert, GW, 811, 821 Segal, U, 248 Selden, L, 184, 208 Shanken, J, 446, 449 Shannon, 588, 800, 803 Sharpe, WF, 11, 149, 613 Shaw, J, 8 Shiller, RJ, 208, 612, 811, 812, 822, 837 Singleton, K, 208, 228 Slovic, P, 120, 270, 271, 274, 277 Solow, R, 801, 803 Sox, HC, 270, 274 Spence, M, 181

876

Author Index

Stambaugh, RF, 446, 473, 813, 827, 827 Stein, C, 391, 410, 447 Stiemke, E, 61 Stiglitz, JE, 135, 253, 284, 409 Stricker, C, 805 Stux, I, 184 Suppes, P, 87, 90 Swalm, RO, 131, 132 Taksar, MI, 789, 790, 802–804 Taleb, NN, 849 Thaler, RH, 274–277, 314, 467, 475 Thorp, EO, 8, 284, 608, 639, 644, 650, 654, 673, 789, 790, 805, 806 Tobin, J, 105, 139, 149, 283, 293, 299, 301, 391 Touzi, N, 793, 805 Treynor, JL, 11 Tucker, A, 7, 60 Tversky, A, 82, 83, 99, 129–132, 134, 140, 141, 146, 147, 151, 153, 159, 164, 165, 167, 178, 248, 261, 269, 270, 271, 274, 474, 477, 478 Van, L, 803 van Dam, C, 113 Viceira, LM, 809, 811, 814–820, 823, 827, 832, 834, 837, 840, 845 Vickson, RG, 365, 373, 790, 806, 807 Viscusi, WK, 139

Vliet, PV, 177 von Neumann, J, 80, 129, 790, 807 Wakker, PP, 83, 145, 177, 178 Wald, A, 799, 803 Wallace, SW, 491 Wachter, J, 811, 840 Weil, P, 815, 821 Whitt, W, 449 Williams, AC, 104 Wolfowitz, J, 799, 803 Xia, Y, 811, 840, 845 Yaari, ME, 137, 208 Yan, JA, 44 Yu, JNW, 367, 373, 390, 392 Yushkevich, AA, 803 Zeckhauser, R, 119, 181 Zhang, X, 61, 611, 613 Zhao, Y, 488, 805 Ziemba, WT, 5, 7, 8, 282, 284, 310, 312–314, 333, 365, 367, 370, 373, 375, 376, 389, 390, 392, 409, 411, 415, 434, 440, 445, 446, 450, 457, 459, 484, 488, 491, 492, 494, 499, 603, 605, 606, 612, 637, 639, 643, 644, 649, 650, 666, 668, 789, 790, 804–807 Zin, SE, 85, 207, 208, 209, 230, 248, 815

Subject Index asset

acceptability of insurance, 274 acceptance set, 505, 511, 570, 590 accumulation of wealth over time, 457 accuracy of prediction, 263 action space, 185 actuarial criterion, 272 favorable risk, 330 neutral, 320–321, 329 neutral risks, 322 premium principles, 543 additive conjoint measurement, 125 adjustment and anchoring, 265, 268 A-effective infimum, 590 affine similarity, 203 affinely equivalent, 200 aggregate consumption, 223, 226, 229 aggregation of independent, identical gambles, 471 aggregator, 214, 216–219, 230 Allais paradox, 208, 218, 254, 478 allocation of investment capital, 445 α-dispersion, 424 alternative choice models, 113 alternative options, 276 amalgamation of independent gambles, 247 ambiguity, 116 anchoring in the assessment of subjective probability distributions, 266 anomalies, 467 approximate risk-neutrality, 241 approximating induced preference, 199 arbitrage, 31, 58, 60 basic arbitrage theory, 61 model of capital asset pricing, 11 portfolio, 12 race track, 8 Arbitrage Pricing Theory (APT), 5 Archimedean axiom, 89, 549 aspiration, 122

class, 602 elasticity of the demand, 330 integration, 100 pricing, 4 pricing factor model, 579 returns, 229 returns and covariance complexity, 447 associative connections, 265 asymmetric mean absolute error, 579, 591–592 asymptotics capital requirements, 536 large portfolios, 538 atemporal distribution of consumption, 210–211 atemporal non-expected utility, 207–208, 219, 230 atemporal von Neumann–Morgenstern, 183, 187, 199 attitudes to changes, 115 availability, 264–268 aversion to ambiguity, 113 axiomatic analysis, 125 axioms for a binary relation, 187 balanced trading strategies, 794 BARRA, 5 base-rate frequency, 261 Baumol’s criterion, 166–168 Bayes estimate, 449–450, 453 model for price distributions, 446 risk, 261, 452 Bayes–Stein (Jorion) estimate, 445–456, 460 beating another portfolio, 704 “beat” the market, 602 benchmark beating, 704

877

878

Subject Index

investment management, 765 investor, 754 mutual fund theorem, 767 benchmark-tracking portfolio, 768 Bernoulli trails, 638 Bernoullian expected-utility theory, 89 biases effectiveness of a search set, 264 evaluation of conjunctive and disjunctive events, 265 imaginability, 264 retrievability of instances, 264 truncation estimate, 460 bi-criteria problems, 648 bidding, 120, 319 bilinear form, 113 bills, 844 binary option, 690 binary preference-indifference relation, 87 binary relation, 88, 90 Black Swan, 849 blackjack, 661 Black–Scholes formula, 7, 850–851 Boltzmann’s theorem, 587 bonds, 841 bounded horse lotteries, 93, 96 Brownian motion, 31 calculation of investment portfolios, 375 calibration problems, 247–478 Canadian lotto 6/49, 667 cancellation, 110, 125 canonical representation, 272 capital accumulation, 461 gains, 845 growth strategy, 461 growth theory, 789 capital asset pricing model (CAPM), 4, 11, 149, 584 equilibrium, 168 model, 229 relations, 585 cash equivalent, 317–319, 368–369 category-boundary effect, 272 Center for Research in Security Prices (CRSP), 368, 821 certainty equivalent, 129, 214, 216, 220, 523 approach, 130–132 framework, 141

functional, 218 tasks, 140 chance constraint, 565–566, 575 change of wealth, 109, 113, 129, 130, 133–134, 141, 149, 150, 155, 167, 248, 474 change of reference point, 122 characteristic exponent, 421, 435 characteristic function, 425 characterization of preferences, 130 Chebechev’s inequality, 166, 579, 585 Chew–Dekel class, 208, 220, 227 choices, values and frames, 269 Choquet integral, 527, 540, 559 closed convex hull, 505 coding outcomes, 110 cognitive biases, 267 coinsurance, 257 combined impact of imperfections, 67 common consequence effect, 255–256 common factor model, 448–449 common information, 462 common ratio effect, 255–256 comonotone, 565, 573 additivity, 547 deviation measure, 588 subadditivity, 547 compact distributions, 284 comparative statics, 230 compare risk attitudes, 247 complementary events, 118 complete guide to option pricing formulas, 850 complete markets, 32 compound events, 266 compound prospects, 124 concave, 557–559 distortion, 527, 540, 547 everywhere, 130 function, 114, 270 risk-averse preference function, 491 risk-averse utility function, 493 utilities, 296 utility-of-wealth, 467 concavity, 105, 322 extreme, 245 value, 275 concavity and monotonicity, 18 conditional complexity, 454 distribution, 449

Subject Index

expected return, 449 maximum complexity, 448 risk neutral probabilities, 62 value at risk, 518, 558, 571, 576 confidence interval, 266 conjugate function, 570 conjunctive events, 266 consequences, 269 conservatism, 262 consistent pricing rules, 54–55 consols, 837 constant proportion allocation strategies, 720 proportion portfolio strategies, 726, 740 consumption bundles, 186, 200 investment, 186 portfolio plan, 224 programs, 209, 211, 214 savings problem, 185, 193, 199–200 space, 213 wealth ratio, 815, 817–819, 821 contingent insurance, 106, 107 contraction mapping theorem, 217 contradictory data, 265 control criterion, 760 control for the downside risk, 686 convergence in quadratic mean, 28 convex entropic risk, 522, 534 indifference curves, 391 multivalued operators, 789 ordering, 586 penalty, 605 penalty costs, 493 principles of premium calculation, 544 programming, 574 convexity, 494, 505, 568, 574 correlation between innovations, 827 correlation matrices, 498 covariance, 448, 462 complexity, 445–448, 452, 455 matrix, 445, 450 returns, 446 structure of the innovations, 825–827 cross-sectionally correlated, 814 cumulative distribution, 152, 302, 643

879

distributions of two n-period risks, 305 distributions intersect, 290 probability, 167 probability distribution, 289 prospect theory, 150, 159, 160, 177 CVaR, 486, 571, 579, 593 deviation, 579, 581, 583–586, 589 Dantzig–Wolfe generalized programming algorithm, 376 data envelopment analysis, 611 dead-loss effect, 277 decision analysis, 183 isolation, 475 making, 269 theory, 271 uncertainty, 87 value, 277 weight, 99, 111, 116, 124, 125, 145, 273 decomposition of the covariance, 446 deductibles, 105 defined benefit plans, 492 defined contribution plans, 492 degree approximation, 205 calibration, 267 representativeness, 263 risk aversion, 472 demand for deductible insurance, 257 denumerable partition, 95 dependence of expected returns, 814 desirability of prospects, 116 detection of dominance, 111 deterministic consumption streams, 214, 219 deterministic target, 605 deviation measure, 533, 558, 567, 579–580, 582, 584–585, 592, 594 measure of semi-LR type, 580 measures of LR type, 580 penalties relative to expectation, 594 different investment horizons, 300 digital option, 689–693 dimension fund advisors, 4 diminishing marginal utility, 242

880

Subject Index

discounted expected value, 50 discrete arbitrage theory, 57 discrete-state approximations, 811 probability distribution, 296 time stochastic dynamic models, 603 disjunctive events, 266 distort small probabilities, 141 distribution functions, 281 distribution of rates of return, 447 distributivity axiom, 126 diversification, 149, 167 diversify, 156 dividend income, 845 yield, 837 price ratio, 813, 822, 826–827, 832, 836 dominance, 271 dominated alternatives, 120 dropping the risk-free asset assumption, 439 dual path, 796 duality theory, 62 Dutch book, 475, 476 dynamic programming, 183 dynamics of state variables, 813 editing operations, 111, 125 phase, 110 prospects, 111 efficiency criterion in the case of risk aversion, 288, 290, 300 frontier, 377, 562, 854 portfolios, 149 set, 155, 287, 288, 299, 302 set of stocks, 296 surface, 423 efficiency analysis of choices involving risk, 287 eigenvalues, 448, 450, 451, 454–455 elasticity intertemporal substitution, 814–819, 845 substitution, 216, 219, 220, 811 empirical asset returns, 495 Empirical Bayes estimate, 445–447, 450, 451, 456 empirical MV frontier, 164

empirical PSD-efficient set, 165 empirical PT-efficient set, 164 endowment effect, 276 entropy, 448 maximization, 586 Epstein–Zin utility, 605, 809, 811–812 equilibrium prices, 4 equity premium, 473 puzzle, 475 error, 578 means, 371 measure, 579, 590–592 projection, 591 variances and covariances, 371 Esscher principle, 545 essentially different strategy, 606 estimation expected rate of return, 445 mean rates of return, 450 Euler equation for consumption, 815 Eurobonds, 499 European bonds as a hedge for equities, 498 European call digital option, 692 evaluation of probabilities, 265 event, 88 evolutionary finance, 800 evolutionary stability, 449, 459 existence of a martingale measure, 8 expectation, 99, 122, 269 principle, 112–113 expected actuarial cost, 105 expected rates of return, 446 expected regret, 564, 567 expected shortfall, 518, 567, 572 expected time, 716 expected utility certainty equivalent, 219 ex-hypothesis, 477 extremely-concave, 242 generalization, 209 generalized analysis, 256 hypothesis, 472 maximization, 149, 181, 245, 254, 468–469, 477 model, 253 paradigm, 129 preference function, 208, 253 representation, 182, 183 theorem, 80, 90

Subject Index

theory, 99, 104, 106, 129, 141, 146, 177, 241, 242, 246, 471–476, 478 experience value, 277 exponential principle, 545 extreme periods, 496 extreme probabilities, 119 extremity-orientedness, 145, 147 factor, 451, 759, 761, 777 model, 12 process, 777 fair premium, 539 Farkas’ lemma of the alternative, 62 fat tails, 495, 608 feasible action correspondence, 185 Fed–Treasury accord, 821 final assets, 123 financial planning model InnoALM, 491 financial strength, 611 finitely additive probability measure, 90 first order autoregressive model, 604 first order stochastic dominance (FSD), 253, 257, 280, 558, 561, 565 first-degree stochastic dominance, 133–134, 152 first-order vector autoregressive process, 813 fixed mix, 602 formulation effects, 274 Fortune’s Formula, 280 fourfold pattern in behavior, 145 fractional Kelly strategy, 606, 645, 667, 753–756, 758, 760, 775, 781, 784 fractional program, 310 fractional program for Tobin separation, 379 framing, 273, 477 alternative, 276 effects, 278 experiment, 132 outcomes, 130, 141, 271, 274 free lunch, 37 Friedman–Savage utility function, 82 FSD-efficient set, 164 full certainty independence, 550 fund separation theorem, 753–754, 758, 782, 784 fundamental analysis, 611 fundamental problem of dynamic asset allocation, 661

881

fundamental theorem of asset pricing, 5, 7, 31–32, 35 gains and/or losses, 110–112, 122–123, 270, 274–275 gambler’s fallacy, 262, 267 Gauss programming language, 497 Gaussian normal, 295 Gaussian variables, 448 g-divergence, 520 general efficiency criterion, 300 general theory of subjective probabilities and expected utilities, 87 generalization of Tobin’s separation theorem, 423 generalized deviations, 532 generalized linear regression, 593 geometric growth, standard deviation and Sharpe ratio, 853 Girsanov’s formula, 52 global risk seeking, 177 globally evolutionary stable, 606 golden sections search, 385 growth optimal investments, 789 portfolio, 692 portfolio policies, 743 portfolio strategies, 738, 746 Hahn–Banach theorem, 6, 37 heuristic principles, 261 hierarchial linear model, 446 hierarchial model for returns, 447 highly volatile periods, 496 historical distributions, 498 HJB partial integro-differential equation, 776, 780 holding costs and benefits, 72 homogeneous of degree zero, 379 homoskedastic, 814 horse race lotteries, 88–91, 548, 664 hypothetical inflation-indexed bond, 821 hypothetical real consolbond, 822 hypothetical real perpetuities, 837 IBM OSL stochastic programming code, 491 IBM’s optimization solutions, 495, 497 ICAPM fund separation theorem, 760, 761 idiosyncratic variances, 449

882

Subject Index

illusion of validity, 263 illusory-correlation effect, 265 impact errors in means, 368 evidence, 262 periodic dividends and interest payments, 65 transaction costs, 66 imperfect markets, 66 implausibility of expected utility theory, 469 improved estimation, 462 income endowments, 182 incomplete computation, 265 incomplete market, 753–754, 758 inconsistent preferences, 107 independence axiom, 82, 89, 121, 208, 253–254, 531 imperfections, 69 indifference curves, 254, 256, 257 map, 199 sets, 191 surfaces, 196–197, 199 induced preference, 181–183, 185, 188, 190–193, 196–198, 200, 201, 205 infinite borrowing, 168 horizon temporal lotteries, 231 long-lived investor, 812 probability tree, 224 inflation dynamics, 821 indexed bonds, 837, 840, 841, 845 risk premium, 837 InnoALM, 488, 493, 502 insensitivity predictability, 263 prior probability, 261 sample size, 262 insufficient adjustment, 265 insurance policies, 476 premium, 318–320 integrated chance constraint, 564–566 internal consistency, 263, 267 intertemporal asset pricing model, 207 budget constraint, 815

capital asset pricing model, 754, 758, 760, 764, 765 consistency of preference, 209, 215 consumption-CAPM, 207 consumption lotteries, 207 elasticity of substitution, 828 hedging, 810, 812, 819 hedging demand, 832–834, 836–838, 841, 845 hedging opportunities, 844 hedging portfolio, 761, 768, 775, 782, 783 optimization problem, 227 optimizing behavior, 226 substitution, 230 intransitivities of choice, 111 intuitive predictions, 263 invariance, 271–273, 275, 278 inverse problem, 588 projection, 592 S-shaped distortion function, 177 stochastic dominance constraints, 561 investment decision making, 299 decision rules, 152 horizon, 299, 306 opportunity set, 818, 823 strategy, 458 isolation effect, 99, 107, 110, 124 joint normal distribution of returns, 366 judgment heuristics, 267 principles, 125 probability, 261, 266–267 uncertainty, 261 jump diffusion, 776 Kalman filter, 772, 774–775 Kelly criterion, 753–757 growth optimal investment strategy, 460 investment strategies, 753, 784 investor, 756 portfolio, 761 portfolio hK , 756, 760–763, 765, 768, 769, 775, 781

Subject Index

strategy, 459, 606, 761 classic, 760 fractional, 767 Knightian uncertainty, 508, 549 Kolmogorov inequalities, 586 KP class, 222 KP preferences, 228, 229 Kreps–Porteous (KP) and Chew–Dekel subclasses, 223 k-th order stochastic dominance, 559 Kuhn–Tucker conditions, 429 latent factors, 448 latent market factors, 456 law of small numbers, 262 law-invariant convex risk measures, 525 deviation measures, 580, 585–586 least squares regression, 593 Lemke’s algorithm, 380 level of adaptation, 277 life insurance contract, 50 lifetime wealth, 470 limitations of the mean-variance efficiency criterion, 293 limited liability, 17 linear complementarity problem, 310 exponential-of-quadratic Gausian stochastic control, 765 model, 446 objective function, 431 probabilities, 254, 472 probability weights, 477 programming, 376, 575 regression, 593 local propensity to insure, 321 log characteristic function, 425 log excess returns, 813 log return on the portfolio, 816 logarithmic utility function, 280, 310, 386–388, 405, 637, 646, 744 log-concave distribution, 588 log-linearize, 816 log-linearized Euler equation, 818 log-normal case, 166 log-normal return distribution, 164, 168, 284 log-optimal, 722 log-prices, 450 log-return, 447

883

long-term capital management, 492, 849 long-term inflation-indexed bonds, 813 long-term investors, 810, 844 long-term nominal bonds, 813, 822, 837, 844–845 Lorenz curve, 280, 647 loss growth rate, 459 preserving, 460 wealth, 457 loss aversion, 248, 270, 276, 314, 474, 478 parameter, 145 losses and costs, 276 lottery, 530 lotto games, 666 lower upper worst-case deviations, 580 lower partial moment, 563–564 lower range deviation, 587 lower range dominance, 580, 582, 589 lower semicontinuous positively homogenous convex functionals, 579 lower semideviation, 579 magnitude of the change, 113 manipulations of probabilities, 122 margin requirement, 70 marginal distribution of rates, 448 marginal utility, 245, 469 market complete, 58 efficiency, 611 equilibrium, 168, 585 return, 228 risk premium, 24 Markowitz mean-variance criterion, 293, 310 mean-variance (MV) model, 561 (µ, α) criterion, 294 MV-efficient set, 164 portfolio selection problem, 583 reverse S-shaped function, 139 stochastic dominance, 82, 129–134, 141, 146, 177 utility function, 82, 129, 134 martingale representation property, 32 valuation, 63 master fund, 584 maximal complexity, 454 maximize the expectation, 129, 458 maximizing logarithmic utility, 726, 738

884

Subject Index

maximizing the probability of beating a given goal, 724 reaching a given goal, 685, 726 reaching a goal in finite time, 688 maximum/minimum final wealth trajectories, 654 maximum conditional complexity, 452 maximum entropy principle, 580, 587 maximum likelihood, 445, 450 estimator, 447, 452, 453 maximum loss, 564, 569 MAXMINVAR, 527 MAXVAR, 527 mean absolute deviation, 424, 579–580, 586–587, 593 mean-deviation approach, 579 mean-dispersion efficiency curve, 427 mean hedging portfolio, 828 mean percentage cash equivalent loss, 313 mean rate return vector, 449 strategy, 459 mean-reverting (Ornstein–Uhlenbeck) process, 811 mean-risk, 564 models, 563 mean squared error matrix, 452 error of estimators, 447 mean-standard deviation approach, 377 mean value functional, 216, 226 mean value theorem, 324 mean-variance analysis, 149, 810 criterion, 287, 295, 296 diversification algorithm, 168 dominance, 139 efficiency, 151 model, 11 portfolio, 457 portfolio theory, 169 rule, 139, 152, 167 skewness, 162 median regression, 593 mental accounting, 269, 275–276, 314, 474–475 Merton’s fund separation theorem, 756 m-event partition, 96 minimal penalty function, 514 minimizing the expected time, 721

minimum variance portfolio, 158, 162, 840 Minkowski’s inequality, 430 miscalibrated expected-utility theory, 246 miscalibration, 469 chance, 262 regression, 263 mispricing, 446 mixed bets, 142 mixed CVaR deviation, 581, 583, 589, 592–593 mixed gambles, 177–178 mixed prospect, 124, 129, 133 mixed quantile regression, 594 mixing of correlations, 497 mixture set, 88, 90 model ambiguity, 533, 543 choice under uncertainty, 478 estimation errors, 446 preference, 277 modern decision theory, 267 modest-scale risk attitudes, 248, 469 modest-stakes gamble, 242 monetary risk measure, 511 money pump arguments, 475 money pumping, 476, 477 monotone non-decreasing concave functions, 281 utility functions U , 133 monotonicity, 568, 576 Monte Carlo simulations, 859 mortality tables, 51 multiattribute options, 274 multiperiod discrete probability scenarios, 491, 495 multiperiod dominance unrestricted utility function, 301 multiperiod efficiency criteria, 300 efficient set, 306 expected utility models, 230 investors, 299 portfolio choice, 810 stochastic programming, 491–493, 603 multivariate covariance operator, 416–417 multivariate model of strategic asset allocation, 809 mutual funds, 168

Subject Index

MV-efficient frontier, 162, 165, 168 set, 158, 160–161, 165 myopic allocations of nonlogarithmic investors, 838 component, 818 logarithmic investor, 838 loss-averse agents, 476 loss aversion, 474, 476–477 portfolio, 815, 832 narrow framing, 474 necessary and sufficient, 198 necessary and sufficient condition for (concavity) dominance, 293 necessary condition for dominance, 292 negative exponential utility function, 366 negative prospects, 104 Nikkei put warrant arbitrage, 8 no arbitrage condition, 5, 8, 32, 63 no free lunch, 7 nominal bill rate, 836 nondegenerate error measure, 591, 592 nonexpected utility preferences, 258 theory, 208, 248, 253 nonlinear programming algorithms, 376 nonlinearity of decision weights, 273 nonmonetary outcomes, 270 nonstochastic consumption programs, 220 normal distributions, 162, 168 normal periods, 496 normally distributed, 156 normative statistical theory, 263 normatively unacceptable consequences, 113 n-period efficient set, 303 n-period risks, 302 number of factors, 454 numeraire portfolios, 789, 791, 798 numerical dynamic programming, 812 numerical probability, 273 objective cumulative distribution, 130 objective frequency, 264 objective lotteries, 253 objective MV-efficient frontier, 164 objective probabilities, 158–159, 161, 168 observed preferences, 276 offset the loss from estimation error, 462

885

one fund theorem, 584 one-sided Chebyshev inequality, 585 open source project COIN-OR, 495 optimal consumption, 819 optimal consumption and portfolio choice problem, 809 optimal consumption-wealth ratio, 819 optimal criterion for F dominates G, 289 optimal efficiency criterion, 287, 293 optimal growth policy, 692, 704, 707 portfolio, 693 optimal log consumption-wealth ratio, 820 optimal portfolio allocations, 844 choice, 818, 837 rule, 817, 821 optimal proportions of the risky assets, 379 optimal ratios of risky to non-risky assets, 310, 378 optimal risk-sharing, 258 optimal selection rules, 287 optimal strategy, 458 optimality of Rubinstein’s risk aversion measure, 312, 391, 406 optimization codes, 72 optimized certainty equivalent, 521, 577 optimum proportional investment, 330 option pricing formulas, 850 ordinal transformation, 189 Orlicz hearts, 517 norm, 517 premium principle, 546 space, 517 outcome-convexity, 257–258 overestimation, 117 overinterpretation, 263 overweighting, 117–118, 121, 145, 269, 273 parallel hyperspaces, 196 parameter of loglinearization, 820 parameter uncertainty, 845 parametric concave program, 428 Pareto-optimal risk exchange, 546 partial integro-differential equation (PIDE), 780 partial observation, 754, 772, 775 partial separation, 429

886

partition, 91–92 payoff spaces, 185 penalties relative to expectation, 591 penalty function, 514 pension fund planning models, 496 pension payments, 494 perceived efficient frontier, 161 percentage of maximum variance, 405 performance comparison of estimators, 455 goal, 735 shortfall level, 735 phenomenon anchoring, 265 piecewise-linear function, 493 planned consumption growth, 819 point of tangency, 158, 162 pooled Canadian equity pension funds, 381 portfolio active management, 732 decisions, 457 long-term, 809 selection, 579 separation property, 225 theory, 578 weights, 493 positive homogeneity, 487, 505, 568, 576 positive linear transformation, 91–93 positive prospects, 104 positive risk premium, 325 positively homogeneous, 428, 437 posterior distribution, 449 means, 446 odds, 262 power framing, 277 utility function, 385, 387, 811 precautionary savings, 819 predictability accuracy, 263 asset returns, 844 biases, 264 errors, 445 excess stock and bond returns, 827 future securities returns, 445 preference class, 136 error, 203

Subject Index

optimality conditions, 814 skewness, 164 stability, 276 premium for a one year death insurance contract, 50 primitive preference, 185, 198–199 Princeton Newport Partners hedge fund, 673 principles of sampling, 267 prior distribution, 448–449 prior mean, 451 probability beating a stochastic benchmark, 608 conjunctive events, 266 distortion, 151, 177–178 distribution, 266 failure in complex systems, 266 insurance, 105–107, 121, 274 maximizing, 720 maximizing game, 732 measure Q, 31, 42, 88, 91, 505 outcomes, 124 outperforming, 732 premium, 321–323, 325–326 prior, 261–262, 267 rare events, 117 representations, 109 tree, 184, 210–211 weighting, 145–147 procedure of elicitation, 267 producer price index, 822 production tasks, 120 proper calibration, 266 properties of the Kelly optimal growth strategy, 645 proportional hazards premium principle, 547 proportional risk premium, 328–330 proportional transaction costs, 69 prospect theory, 82, 99–110, 125, 129–131, 133, 141, 145, 149–150, 270, 474 mean-variance analysis, 149 value functions, 133 violations, 145, 177 prospect stochastic dominance (PSD) efficient set, 151, 155, 158, 164, 165 inefficient, 161, 164 PSD and MSD are opposite, 136 SSD-efficient sets, 156 pseudo-certainty effect, 273 pseudo-concave, 379–380, 438

Subject Index

pseudo-concave objective function, 421 psychophysics, 277 approach, 269 chances, 272 PT and MV efficient sets, 150, 168 PT efficiency, 151 PT-efficient set, 168 purchase of insurance, 277 quadratic program, 310 quadratic utility, 294, 327, 317, 341, 382 quantile, 560, 565 quantile regression, 579, 591, 593 quasi-concave, 407 quasi-convexity, 439, 512 quasi-timing indifference, 223 Radon–Nikodym derivative, 34 Ramsey–Savage approach, 116 random dynamical systems, 789 random parameters, 446 random walk theory, 301 rank-dependent expected utility theory, 559 ranking of timeless wealth lotteries, 226 rapid path, 789, 791, 796 rates of return, 223 rational choice, 477 rational decision making, 80 rational expectations, 473 real consol bond, 840 real interest rates, 834 real perpetuity, 838, 840, 844 recursive intertemporal utility, 207, 219 recursive structure, 215 redundancy, 263 reference point, 110, 111, 113, 115, 122, 145, 271, 274 reference state, 275 reflection effect, 104 regression toward the mean, 263–264, 267 regular insurance, 121 regular prospect, 112 reject the S-shaped function, 139, 141 reject prospect theory, 140 relative entropy, 522 relative return for the comparative estimators, 461 relative savings loss (RSL), 453 relative wealth loss, 459 reliability, 263

887

Renaissance Medallion hedge fund, 674 replicating an option, 32 representations of outcomes, 109 representative agent assumption, 226 representative agent framework, 219 representativeness, 261, 263, 267, 268 representativeness heuristic, 261–262 resolution averse, 188, 198 neutral, 187 seeking, 188, 199 uncertainty, 222 retrievability of instances, 264 returns covariance matrix, 446 reversal of preferences, 108 reverse S-shaped, 82, 129, 131–132, 134–136, 138, 140, 146 risk attitude, 120, 207 choice, 269 comonotone, 528 component, 112 envelope, 582–583, 589 estimator, 451, 452 functional, 513 identifier, 582–583, 589 moderate-scale/large-scale, 471 monetary prospects, 467 neutral, 7, 52, 241, 469, 472, 688 neutral pricing, 49 optimal portfolios, 405 parameter, 457 premium, 312, 318–320, 322–323, 326, 329, 539 profiles, 581 tolerance, 366, 371 worst case, 564, 569 risk-acceptable regression, 594 risk aversion, 100, 105, 116, 121, 130, 136, 149–150, 167, 209, 216, 241–243, 253, 256–257, 269–270, 295, 312, 317, 325–329, 333–334, 467–468, 470, 474, 588, 814, 820, 832–833, 840–841 absolute, 16, 245 absurd, 470 Arrow-Pratt measure, 205, 333, 648 choice, 272 comparative, 219–222, 322 constant, 318, 325–326 constant proportional, 329

888

Subject Index

convex, 493–494, 512, 530, 567, 570, 573 declining, 221, 810 decreasing/increasing, 294, 311, 317–318, 325–327, 330 decreasing marginal utility, 130 efficient, 473, 878 effects of increasing, 828 efficient set, 306 extreme, 243 global, 177 hypothesis, 106 index, 462 in the small/in the large, 317 large stakes, 469 local, 312, 318, 320–323, 325 local proportional, 329–330 long-horizon investor, 840 measure, 317 modest-scale, 246, 474 multi-period dominance, 303 multivariate measures, 333 neutralizing, 247 over modest stakes, 247 parameter, 223, 459, 608 preserve decreasing, 326 proportional, 328 relative, 208, 457–458, 473, 810, 814 Rubinstein’s measure, 312, 334 strictly increasing, 327 utility functions, 564 riskless asset, 15 riskless component, 112 riskless rate of return, 49 risk measure averse, 570, 579, 588–592 coherent, 486, 512, 567, 570, 573, 588, 590 comonotone, 540 convex, 493–494, 512, 530, 567, 570, 573 CVaR, 579 CVaR-type, 589 deviation, 485 divergence, 520 downside, 563, 706 entropic, 520 Haezendonck, 521, 546 higher moment coherent, 576 polyhedral, 578 robustification, 533–534

L P (Ω) type, 589 semi-L P type, 576, 589 shortfall, 519, 521, 531 tail, 576 risk seeking dominance, 134 stochastic dominance, 177 risk-sensitive asset management, 760, 765 control, 758 robust large deviations, 536 loss probability, 536 preferences, 548 representation, 514, 526 law-invariant, 526 root mean squared error, 456 Rubinstein’s measure of risk aversion, 333, 391, 411, 413 S&P 500 index, 822 safety first principle, 585 salience, 264 sample covariance, 453, 454 sample size, 262, 267 sampling error, 453 sampling variability, 267 scenario-dependent correlations, 500 scenario tree, 495 search technique, 378 second order stochastic dominance, 133–134, 153, 177, 280, 558, 561 segregation, 110 semideviation, 564 semi-interquartile range, 424 semi-variance, 485, 563 separating hyperplane, 7, 62 sequential formulation, 108 Shannon differential entropy, 586 sharing of information, 446, 449, 457 Sharpe–Linter–Mossin capital asset pricing model, 167 Sharpe–Markowitz diagonal model, 422, 428 Sharpe ratio, 823, 832, 851–852 shift or reference point, 122, 123 shocks to investment opportunities, 810 short positions Pk+ , 70 shortfall constraints, 494 short-term interest rate, 811, 821, 833

Subject Index

Shrinkage (Stein) estimators, 446 Siemens pension fund, 491 sign dependence, 125 “similar” RA values yield “similar” optimal portfolios, 392 simple events, 266 simplification of prospects, 118 smaller risk premium, 325 small-scale insurance, 476, 477 solvability, 125 solving the approximate model, 817 space of temporal lotteries, 210 specific evidence, 262 spectral decomposition, 450 spurious casual explanations, 263 SSD-isotonic, 569, 576 S-shaped function, 131–132, 136, 140, 270 preferences, 154 utility function, 82, 135 value function, 115, 122, 129–130, 133, 138, 145, 168, 271 St. Petersburg paradox, 637 stable distributions, 421 standard lower and upper semideviations, 580 standard lower semideviation, 583–584, 586–587 state variable, 836 states of wealth, 272 stationarity, 302, 305 stationarity of preference, 215 stationary models, 793 stationary over time, 306 statistical properties of log returns, 498 Stein/Rubinstein, 334–335, 410 Stiemke’s theorem, 61, 63 stochastic benchmark goals, 494 portfolio, 603 stochastic cash flow stream, 58 stochastic differential games, 736 stochastic dominance constraints, 560, 573 investment, 141 methodology, 133 rules, 133–134, 150–151, 154, 528, 558–559, 564 stochastic optimization, 575, 812 stochastic programming, 555, 557

889

stochastic pure endowment economy, 229–230 stocks, 844 stocks, bonds and bills, 820 stop-loss order, 547 strategic asset allocation, 602, 810, 820 strategic asset allocation with inflation-indexed bonds, 837 stress tests, 514 strictly convex, 431 strictly pseudo-concave, 432 structure of relationships, 447 sub-additivity, 117, 118, 486–487, 505 subcertainty, 118, 121, 125 subdifferential, 570 subjective cumulative distribution, 129 subjective probability, 161, 253, 264, 267 distortion, 139 distributions, 266 transformation, 165 subjective risk free discounting factor, 77 subjective transformations, 151 subjective value, 111, 270 subjectively distort probabilities, 129, 158, 159, 167 subjectively distorted, 136 subproportionality, 118, 121 substitution, 118, 209 axiom, 102, 121, 188, 195, 197 consumption, 229 parameter, 220 risk aversion, and timing, 219 support for the reverse S-shaped value functions, 142 sure-thing principle, 89 symmetric distributions, 295 symmetric fair bets, 115 symmetric stable distribution, 310, 421 system of beliefs, 267 systematic errors, 265 systematic risk, 207, 228–229 systematic violations of the expected utility hypothesis, 254 tactical asset allocation, 602 tail conditional expectation, 572 tail VaR, 518, 572 temporal behavior of consumption, 208, 230 temporal CAPM, 207

890

Subject Index

temporal lottery, 181, 184–185, 197, 207, 210–211, 213, 224 temporal resolution, 209 temporal von Neumann–Morgenstern, 187, 189, 192–193, 196–198, 200–201, 205 theorems of the alternative, 7 theoretical wealth loss, 458 theory of insurance, 122 theory of investment, 253 third absolute central moment, 320 three fund separation theorem in a jump-diffusion setting, 783 timeless wealth lotteries, 218 time-separable power utility, 814 time-varying investment opportunities, 810 timing of resolution, 222–223 Tobin’s riskless asset, 376 Tobin’s separation theorem, 421, 427 total liquidation, 70 total separation, 429 total wealth, 133 trace, 452 trading strategy, 792 transaction costs, 789 transactions and trades, 274 transformation cumulative probability, 150, 159 normal distributions, 162 outcomes and probabilities, 110 value function, 122 transitivity, 80 translation invariance, 568, 576 truncation, 451 effect, 454 error, 451, 453 estimator, 451–456, 459, 461–462 TSD-efficient sets, 162 Tucker’s theorem of the alternative, 7, 57, 60 turn-of-the-year effect, 668 turnpike theorems, 801 two-period model, 188 two-parameter distributions, 294 two-period risk, 301 two-sided Chebyshev inequality, 585 two-stage lottery, 213 unbiased linear regression, 594

unbounded expected utility, 426 uncertain distribution of outcomes, 132 uncertainty aversion, 549 underestimation, 262 underweighted, 145, 273 univariate integration, 378, 426 unrestricted utility — the general efficiency criterion, 287 U.S. stock return volatilities, 496 utility, 319 dual, 559, 582 exponential, 384 maximization, 697, 726 recursive, 208, 214, 217, 222–223, 226 shortfall risk measure, 519 theory, 79, 557, 577 utility benefits of inflation indexation, 843 utility function, 116, 287, 317, 318, 321–323, 326, 328–330, 509, 548, 556–559 admissible, 288 concave, 138, 290 cubic, 383 intertemporal, 217–219, 220, 230 negative power, 405 nondecreasing, 133, 138, 300, 302–304 valuation of a contingent claim, 63 value at risk, 485, 518, 558, 564–566 value at risk average, 518, 572 value function, 99, 113, 115–116, 145, 820 value-weighted return, 821 VaR, 579 estimation, 822 state variables, 819 state vector, 817 system, 845 Vasicek model, 761, 768 vector autoregression (VAR), 809, 811, 836–838, 841 violation cumulative prospect theory, 145, 177 dominance, 119 volatility, 449, 557 volatility regimes, 497 von Neumann equilibrium, 799 von Neumann–Gale dynamical systems, 794

Subject Index

von Neumann–Gale model, 789 von Neumann–Morgenstern, 181, 287 axioms, 531 preferences, 243 utility, 181, 185, 208 utility function, 253 utility index, 221 von Neumann path, 799, 801 Wang’s premium principle, 547 weak certainty independence, 549 wealth accumulation, 299, 446

891

loss, 459 shortfall, 447 weighted contraction theorem, 217 weighting function, 116 worst conditional expectation, 572 worst-case mixed-CVaR deviation, 581, 589 Yan formulation, 7 yield spread, 811, 821, 822, 827, 833, 836 zero utility principle, 545