Edition 62 Volume 16 Jan - 2009 Magazine for students in Actuarial Sciences, Econometrics & Operations Research
Lying or Truth-telling: Why Does It Matter In Economics?
Jams Are Everywhere
How Physics May Help to Solve Our Traffic Problems
To DICE with Climate Change a Regression Analysis
How Less Data Can Be Better The Power to Chase Purchasing Power The Dutch Business Cycle:
Which Indicators Should We Monitor?
Dutch Ready for Eligibility Age Increase to 67 Years
((4 + 3)² + 4) / 1/4 = - - 11 + 6³ = - - -
http:// -
√576 + 11 x 3² = - - √196 = - -
--.---.---.--
Preface
Good Intentions J
anuary; a fresh start of the new year. This month will be the one in which most people want to make a difference, get rid of their bad habits and make the best of the year. Some will try to quit smoking, others want to lose weight and go to the gym more often. All are trying to work harder and achieve their goals. I could write down a whole list with my own good intentions, but unfortunately this edition of Aenorm already contained more pages than ever before, so little space remained. In this Aenorm you can find a few topics that relate to good intentions as well. Lying or truth telling was already introduced on our cover. Clearly it would be a good intention to stop lying and start being honest. But as it comes to lying or truth telling in economics it might be different. (Hint: go and read the article). Another topic is climate change. As we all know the climate has been changing a lot in the past years and if we do not adjust our lifestyle it will not get any better. So it would be a good idea to minimize pollution and keep our environment as clean as possible. One way to achieve this, is to use the car less often. This would result in less jams and might solve our traffic problems (p. 19). Hopefully you are now curious enough to go and read all articles. 2008 was a year in which we visited international conferences to become better-known as Aenorm all over the world. For me, representing the Public Relations of the magazine, I can only say we should continue this in 2009 and try to convince people to publish in Aenorm and subscribe to it all over the world! With the start of a new year the VSAE will get a new board. This also means that our lovely chief editor Lennart Dek will step aside to let the new president of our study association, miss Annelies Langelaar, show her skills as the new chief editor. I can only wish her all the best with Aenorm! On behalf of the committee I would also like to thank Lennart for all the effort he put in Aenorm in 2008. This magazine is his last masterpiece, so I hope you will enjoy reading it. Last but not least I would like to wish all of you a very happy new year in which you hopefully will achieve all your goals and make all your good intentions reality. It does not matter if this is obtaining your foundation course, bachelor or master degree or maybe even a promotion at your job, go for it! Shari Iskandar
AENORM
62
January 2009
1
Aenorm 62
Contest List
Lying or Truth-telling: Why Does it Matter in Economics? 4 Marta Serra-Garcia
How Less Data Can Be Better
9 Gijs Rennen
QIS4: The Treatment of Lapse Risk in a Life Portfolio
14
Rob Bruning Cover design: Michael Groen Aenorm has a circulation of 1900 copies for all students Actuarial Sciences and Econometrics & Operations Research at the University of Amsterdam and for all students in Econometrics at the VU University of Amsterdam. Aenorm is also distributed among all alumni of the VSAE. Aenorm is a joint publication of VSAE and Kraket. A free subsciption can be obtained at www.aenorm.eu. Insertion of an article does not mean that the opinion of the board of the VSAE, the board of Kraket or the redactional staff is verbalized. Nothing from this magazine can be duplicated without permission of VSAE or Kraket. No rights can be taken from the content of this magazine.
Jams Are Everywhere - How Physics may Help to Solve Our Traffic Problems 19 Andreas Schadschneider
Regional Real Exchange Rates and Phillips Curves in Monetary Unions: Evidence from the US and EMU 22 Jan Marc Berk and Job Swank
Hedging Prepayment Risk on Retail Mortgages
26
Dirk Veldhuizen
Determining a Logistics Network for Toyota’s Spare Parts Distribution 31
© 2009 VSAE/Kraket Kenneth Sörensen and Peter Goos
The Maximal Damage Paradigm in Antitrust Regulation: Is it Something New? 37 Harold Houba an
d Evgenia Motchenkova
Dutch Citizens Ready for Eligibility Age Increase from 65 to 67 Years 42 Karen van der Wiel
22
AENORM
62
January 2009
Optimal Diversity in Investments with Recombinant Innovation 46
Volume 16 Edition 62 January 2009 ISSN 1568-2188
Paolo Zeppini
To DICE with Climate Change
Chief editor: Lennart Dek
51
Editorial Board: Lennart Dek
Marleen de Ruiter
Container Logistics
Design: Carmen Cebrián
57
Lay-out: Taek Bijman
Iris Vis and Kees Jan Roodbergen
Editorial staff: Raymon Badloe Erik Beckers Daniëlla Brals Lennart Dek Jacco Meure Bas Meijers Chen Yeh
Longevity in the Netherlands: History and Projections for the Future 62 Henk Angerman and Tim Schulteis
The Jackknife and its Applications
Advertisers: Achmea All Options AON APG Delta Lloyd De Nederlandsche Bank Ernst & Young Hewitt KPMG Michael Page MiCompany ORTEC PricewaterhouseCoopers Shell SNS Reaal Towers Perrin Watson Wyatt Worldwide
66 Koen Jochmans
The Power to Chase Purchasing Power
70 Jan-Willem Wijckmans
The Dutch Business Cycle: Which Indicators Should We Monitor? 74
Information about advertising can be obtained from Tom Ruijter info@vsae.nl
Ard den Reijer
Puzzle 79
Editorial staff adresses: VSAE, Roetersstraat 11, 1018 WB Amsterdam, tel: 020-5254134
Facultive 80
Kraket, de Boelelaan 1105, 1081 HV Amsterdam, tel: 020-5986015 www.aenorm.eu
AENORM
62
January 2009
3
Econometrics
Lying or Truth-telling: Why Does it Matter in Economics? As buyers or sellers in different markets, we negotiate prices and quantities. In the process, we often communicate and share information with each other. Hence the question arises; does our communication lead to more efficient outcomes? Although we might intuitively expect so, under the standard theoretical framework in economics this assumption does not necessarily hold. Even if buyers have no reason to mislead sellers, the latter might believe they do and hence ignore their messages. In turn, this makes both truthful and untruthful statements equivalent for buyers in monetary terms and can deem communication irrelevant. However recent research suggests truthful and untruthful statements might not be equivalent in terms of utility, as often is assumed in the standard framework. Individuals exhibit an aversion to lying and are willing to forgo monetary payoffs for true statements. These findings shed a new light on the role of communication in economic interactions and support its potential in determining economic outcomes.
Marta Serra-Garcia is a PhD candidate in Economics at Tilburg University. In 2007 she completed her Master of Philosophy (MPhil.) in Economics at Tilburg University. Her interests are in the fields of behavioral and experimental economics. Her dissertation analyzes different mechanisms which affect cooperation in social dilemmas. One such mechanism on which she is currently focusing is communication.
Communication is present in many of our economic interactions. Despite its relevance, it is often difficult to predict the effect of communication on outcomes. Since communication is costless, discussions about intentions or an individual’s private information are not necessarily credible. In this article, we first discuss a strand of the literature, which focuses on individual’s preferences for truth-telling. Then we argue how these preferences can, under some circumstances, lead to more precise predictions regarding the role of communication in economic interactions. The literature on communication is quite wide. On the one hand, it studies communication in situations of symmetric information, in which individuals communicate their intended choices before a game (for example, in coordination games and social dilemmas). On the other hand, it examines the transmission of private information between informed and uninformed agents. We will concentrate on the latter strand
4
AENORM
62
January 2009
of the literature and work through an example. Imagine you want to buy a second-hand car. You go to a seller who has two apparently identical cars for sale. Suppose one car is labeled A while the other is labeled B. You know one car is of good quality and worth 100 to you, while the other is of bad quality and worth 0 to you. Unfortunately, you cannot tell which car is what quality, but the seller does know. Suppose the price of both cars is 50 and the seller earns a fixed wage of 50. Whatever car he sells you, he will earn 50 but you will only earn 50 if it is the high quality car and lose 50 otherwise. Table 1 summarizes the payoffs in this game, which we will call game 1. Payoff to seller
Payoff to buyer
High-quality car sold
50
50
Low-quality car sold
50
-50
Table 1: Payoffs in game 1
Imagine the seller tells you that ‘Car A is of better quality’. Would you believe him? He has no incentive to lie, since he earns the same amount of money whatever car he sells. Thus, you have reasons to believe him and buy car A. Now, imagine someone told you the seller is a compulsive liar and enjoys misleading buyers. Would you believe him now? If this is the case, you might doubt his statement and decide to ignore it. If you do, you will buy the ‘good’ car with a probability of a half, while with the same
of weet jij* een beter moment voor de beste beslissing van je leven? www.werkenbijpwc.nl
Assurance • Tax • Advisory
*connectedthinking ©2007 PricewaterhouseCoopers. Alle rechten voorbehouden.
AENORM
62
January 2009
5
Econometrics
probability, you might end up with the ‘bad’ car. Your expected payoff is then 0. If you are slightly risk averse, you might refuse to buy any car. Thus, in this case, the seller’s statements become irrelevant to your decision and also irrelevant in terms of monetary payoffs for him. Hence this simple example shows that we often do not have a precise prediction about the role of communication in economic interactions. In the first case, the high-quality car is sold and both seller and buyer are better off. In the second case, the transaction might never take place and efficiency gains might be forgone. Furthermore both outcomes are plausible within game 1.
ten exhibit an aversion to lying. Gneezy (2005) conducted an experimental study among university students in Israel. He found that many students are willing to forgo monetary payoffs and tell the truth. Interestingly, he also found that as the gains for the informed player (the seller, in our example) for lying increase, individuals lie more often. On the contrary, when the harm from lying to the uninformed player increases, individuals lie significantly less often. Other studies in the literature, as SánchezPagés and Vorsatz (2007), have found further evidence of lying aversion in similar situations. Interestingly, this experimental evidence has also been incorporated into theoretical models of
"Aversion to lying can help us to make sharper predictions" Things can get worse, if we slightly modify game 1. Suppose that the seller’s wage is equal to the profit made by selling the car. More precisely, the low-quality car is worth 0 to him while the high-quality one is worth 30. By selling the lowquality car, the seller earns 50, while selling the high quality one earns him 20. Table 2 shows the payoffs in this situation, called game 2. Payoff to seller
Payoff to buyer
High-quality car sold
20
50
Low-quality car sold
50
-50
Table 2: Payoffs in game 2
If the seller only cares about his own payoffs, what will he say? He will probably try to mislead you into buying the low-quality car. Given these incentives, you have no reason to believe his statements. Your expected payoff is then 0 and you might end up not buying a car at all. Thus, communication in game 2 is irrelevant to the transaction’s outcome. However, what happens if we assume individuals have a preference for truth telling? Well, under this assumption you will believe the seller and buy the high quality car in game 1. In game 2, you might also be willing to believe the seller, if his aversion to lying is stronger than the 30 monetary units he forgoes by telling the truth. Thus, individuals’ aversion to lying can help us make sharper predictions about individuals’ communication and its impact on outcomes. Recent studies have shown that individuals of-
6
AENORM
62
January 2009
communication. Demichelis and Weibull (2008) have developed a model in which they assume that individuals have lexicographic preferences for truth-telling, i.e. if a payoff can be achieved via a truthful and an untruthful message, the former is preferred. They study communication before a specific class of coordination games, in which two players want to coordinate which decision to make. Generally speaking, they then show that, under such preferences, the only evolutionarily stable equilibrium in these games is Pareto-efficient. This literature is very recent and many questions still remain unanswered. One of the most interesting questions consists of investigating the strength of individuals’ lying aversion. As we have seen in game 2, lying aversion can lead the seller to be truthful and the buyer to buy the car, only if the costs of lying are ‘high enough’. Further experimental studies could investigate the strength of lying aversion in different economic interactions, allowing us to have a better understanding of when it is relevant. This could lead to a better prediction power regarding the impact of communication in economic interactions. Overall, communication might not be irrelevant to economic outcomes and thus efficiency, if we consider individuals’ preferences for truthtelling. However, doing so poses further challenges. How should we model lying aversion? Could it be simply modeled as a utility cost, which individuals suffer when they do not tell the truth? This cost could also be dependent on the monetary payoffs obtained by lying, if the lie is thought to be trustworthy. To find the
Econometrics
answers to these questions, both empirical evidence, mainly stemming from experimental research and theoretical modeling will be needed. Furthermore it is the dialogue between them that will probably result in the most relevant answers. References Gneezy, U. (2005). Deception: The Role of Consequences, American Economic Review, 95(1), 384-394. SĂĄnchez-PagĂŠs, S. and Vorsatz, M. (2007). An experimental study of truth-telling in sender-receiver games, Games and Economic Behavior, 61, 86-112. Demichelis, S. and Weibull, J. W. (2008). Language, Meaning, and Games: A Model of Communication, Coordination, and Evolution, American Economic Review, 98(4), 12921311.
AENORM
62
January 2009
7
Verhoogd risico op een glansrijke carrière Michael Page Banking & Financial Services is toonaangevend op het gebied van werving & selectie en interim management bij banken, verzekeringsmaatschappijen, asset managers, pensioenfondsen en andere financiële partijen. Zij vervult financiële,
commerciële
en
specialistische
functies,
waar gewenst in samenwerking met Banking divisies in bijvoorbeeld Londen, New York, Parijs, Frankfurt, Singapore en Sydney. Door onze sector expertise en internationale wervingskracht bieden wij onze relaties meerwaarde. Zowel recent afgestudeerden als ervaren professionals kunnen wij in binnen- of buitenland hierdoor de best mogelijke stap in hun loopbaan bieden. Ook de eerste!
Een selectie uit onze actuele vacatures op www.michaelpage.nl
Econometristen Voor verschillende opdrachtgevers uit de top van het bankwezen en de pensioen- en verzekeringsbranche alsmede consultancy firms in Nederland zoeken wij ambitieuze Econometristen. Voor zowel startende als ervaren Econometristen zijn er diverse interessante mogelijkheden. Bij onze klanten bestaan goede doorgroei- en ontwikkelingsmogelijkheden en uitstekende arbeidsvoorwaarden. Geïnteresseerde kandidaten die openstaan voor een objectief advies over toekomstige werkgevers, worden uitgenodigd te reageren. Ref. 135611
Junior ALM Specialist
Naast de mogelijkheid om vrijblijvend kennis te maken met onze gespecialiseerde consultants kun je met betrekking tot de onderstaande vacatures gedetailleerde informatie opvragen via onze website www.michaelpage.nl. Je vindt de vacatures door op onze homepage het bijbehorende referentienummer in te toetsen. Vanaf onze website kun je direct reageren op de daar getoonde vacatures, daarnaast kun je jouw vragen, opmerkingen of cv direct sturen naar banking@michaelpage.nl.
ABN AMRO Hypotheken Groep is een van de grootste gespecialiseerde aanbieders van hypotheken. De AAHG-organisatie kenmerkt zich door een nuchtere en slagvaardige benadering, waarin management en directie gemakkelijk ‘aanspreekbaar’ zijn voor alle medewerkers. Het team Balansmanagement vervult een brugfunctie met de business. Het team heeft zijn verantwoordelijkheid sterk zien groeien en is daarom op zoek naar een Junior ALM Specialist. Je hebt een afgeronde kwantitatieve WO opleiding en weet inhoudelijk wat het belang is van balansbeheer en ALM-studies binnen een professionele financiële dienstverlener. Waardeoptimalisatie, financiering en financieringsconstructies, analyse en afdekking van renterisico en optimale pricing van hypotheken zijn onderdelen waar jij je in wilt ontwikkelen. Ref. 141845
Investments Analist
Financial Analyst
Onze opdrachtgever is een snelgroeiende vermogensbeheerder, onderdeel van één van de grootste internationale financiële spelers. De kernactiviteit van de organisatie is het leveren van inzicht en overzicht over het totale vermogen aan een aantal grootvermogende cliënten. Vanwege de groei in activiteiten zijn wij op zoek naar een Investments Analist voor de afdeling Investments. Als Investments Analist draag je bij aan het in kaart brengen en analyseren van ontwikkelingen in de financiële wereld en vertaal je deze richting zowel interne als externe cliënten. Deze functie geeft je de kans deel uit te maken van een hecht team met veel mogelijkheden tot ontwikkeling. Ref. 140104
EQT is a private equity funds that manages approximately € 12 billion in equity in 12 funds. The Amsterdam office is a small, professional fund administration office in Schiphol and is responsible for the management of the Dutch based funds and the holding company structure of the group. It provides support services to the Guernsey based funds and is overall responsible for the financial control of the group. For this office EQT has a vacancy for a Financial Analyst. You are responsible for actively maintaining the existing holding structures dividends, investments and ownership stakes (pro-actively) provide supporting personal information to the EQT organisation. Ref. 135378
Actuarissen
Investment Consultant
Voor verschillende consultancy firms alsmede de top van het bankwezen en de pensioen- en verzekeringsbranche zoeken wij ambitieuze Actuarissen. Recent afgestudeerden kunnen wij in binnen- of buitenland de best mogelijke eerste stap in hun loopbaan bieden. Graag komen wij in contact met kandidaten die hun wiskundig inzicht willen omzetten in praktische oplossingen en het leuk vinden om klantcontact te hebben. Bij onze klanten is het mogelijk om als specialist of als allround Actuaris door te groeien. Geïnteresseerden die openstaan voor een objectief advies over toekomstige werkgevers, worden uitgenodigd te reageren. Ref. 137278
Onze opdrachtgever is een wereldwijd opererende HRM-Consultancy en Outsourcings-organisatie. De kernactiviteit van de organisatie is het helpen van klanten met actuarieel advies, pensioenuitvoering en complete HRM-consultancy. Vanwege de groei op het gebied van Investment Services zijn wij op zoek naar een Investment Consultant voor de afdeling Investment Services. Als Investment Consultant adviseer je institutionele klanten bij het oplossen van vraagstukken op het gebied van beleggen en (extern) vermogensbeheer. Deze functie geeft je de kans deel uit te maken van een informeel team met veel mogelijkheden tot ontwikkeling. Ref. 141146
Specialist in Banking & Financial Services Recruitment 166 kantoren in 28 landen | www.michaelpage.nl
8
AENORM
62
January 2009
Econometrics
How Less Data Can Be Better During your study, you have probably been warned on many occasions that you can only draw valid conclusions or construct an accurate model if you have enough data. Therefore, the more data you have, the better your model is. However, what if you are in a situation where there is a relative abundance of data? Then the large size of the dataset might also have drawbacks. Fitting a model to your data can take quite some time or can even be impossible. Moreover what to do if you have more data from one part of the design space than from another? Will your model be accurate throughout the whole space or will it be more accurate in areas with more data at the cost of less accuracy in other areas with less data? In this article, we take a look at the effect of using a large non-uniform dataset to fit a Kriging model. We show that using a subset instead of the complete available dataset can have a number if benefits. Furthermore, we introduce and compare several new and current methods for selecting an adequate subset from the original dataset. In short, we will show that less data can be better.
Motivations Kriging is an interpolation technique, which finds its roots in geostatistics. Besides geostatistics, Kriging can also be applied in various other fields. For instance Sacks et al. (1989) applied Kriging in the field of deterministic simulation for the design and analysis of computer experiments. When building a model, intuition tells us that using more data will always result in better models. However, when the large given dataset is non-uniformly distributed over the whole design space, problems can occur. Furthermore these datasets can occur in several situations. The first situation we can think of is when we have a set of legacy data (Srivastava et al. 2004). Legacy datasets contain results of experiments, simulations or measurements performed in the past. As this data is not generated especially for fitting a global model, it is often not uniformly distributed over the whole design space. A second situation is when the data is the result of sequential optimization methods. These methods often generate more data points near the potential optima than in other regions (Booker et al. 1999). Hence if we want to use this data to fit a global metamodel, we will have to take into account that it contains clusters of points. These are just a few examples of situations where we can come across large non-uniform datasets. Using these sets can generate problems, which can often be partly or even completely resolved by using a more uniform subset. Here it is worthwhile noting that a subset is called uniform if the input data of the data points is “evenly spread” over the entire design
Gijs Rennen is a PhD student at the Deptartment of Econometrics and Operations Research at Tilburg University. His research is on design of computer experiments and multi-objective optimization. This article is an abridged version of Rennen (2008) which can be downloaded online for free. For comments about this article the author can be contacted via e-mail: g.rennen@uvt.nl.
space. Important motivations for using a uniform subset instead of the complete dataset are the following: • Increasing accuracy When fitting a model to a dataset, we generally try to optimize a function of the model’s accuracy at the points in the dataset. If this function treats all points equally, accuracy in regions of the design space rich in points can weigh heavier than accuracy in regions with fewer points. This will reduce the accuracy of the model measured over the complete design space. Thus using a more uniform dataset which would equalize the weights of the different regions in the design space could result in a better overall accuracy. • Time savings A second reason would be the reduction in necessary time to fit the Kriging model. This is certainly a significant issue as time-consumption is generally regarded as one of the main drawbacks of the Kriging method (Jin et al. 2001). • Avoiding numerical inaccuracies A common property of large non-uniform da-
AENORM
62
January 2009
9
Econometrics
tasets is that they can contain points which are positioned in close proximity of each other. This property can make the corresponding correlation matrix ill-conditioned (Davis and Morris 1997, Booker et al. 1999). Moreover solving a linear system with an ill-conditioned matrix could cause considerable numerical inaccuracies making the optimization of the Kriging model inaccurate. Hence removing certain points from the dataset can improve the numerical accuracy of the model. • Improving robustness Robustness of the Kriging model with respect to errors in the output data can also be negatively influenced when data points form clusters. Siem and Den Hertog (2007) show that points, which are at proximity of each other, could be assigned relatively large Kriging weights. Due to these large weights, errors in the output values at these points might have a large influence on the output value of the Kriging model at other points. Example To show that selecting a subset can significantly improve the above-mentioned aspects, we introduce the following simple artificial example. Here we try to approximate the six-hump camelback function (Dixon and Szegö, 1978): f ( x ) = 4 x12 −
21 4 1 6 x2 + x1 + x1 x2 − 4 x22 + 4 x24 , 10 3
with x1є [-2,2] and x2є [-1,1]. As our dataset, we take a maximum Latin Hypercube Design of 20 points (Van Dam et al. 2007) with four additional points lying close to an existing point as depicted in Figure 1. By adding these four points, we have created a cluster of points in the dataset. We test the effect of selecting a uniform subset by fitting Kriging models to the datasets first with the four additional points, then without them. To measure the effects of taking a subset
on the different aspects, we use a number of different performance measures. The definitions of these measures can be found in Rennen (2008), but for now it is central to know that lower is better for all measures. The results of the performance measures are given in Table 1. Without four additional points
With four additional points
RMSE
0.48
0.51
Maximum Error
2.03
2.46
Condition Number
76
1491564
Average Robustness
0.97
8.36
Maximum Robustness
1.57
100.66
Table 1: Performance of Kriging models fitted to the datasets with and without the four additional points.
Both the Root Mean Squared Error (RMSE) and the Maximum Error show that the first Kriging model’s accuracy, without the four points, is substantially better than that of the model with the four points. The second Kriging model focuses more on the region where these additional points lie and is, as a result, more accurate in this region. However, this additional accuracy is at the expense of accuracy in other regions, which deteriorates the overall accuracy. When we compare the condition numbers, we see a large difference. Thus the additional four points make the resulting Kriging model to be much more susceptible to numerical inaccuracies. Finally, the larger maximal and average robustness values indicate that this model is also less robust with respect to errors in the output data. For this example, we did not look at the time savings as these are negligible due to the small size of the dataset. Thus this simple example proves that removing some points can improve the quality of the Kriging model. Determining which points to remove is quite easy in this case, but becomes less straightforward when the number of points is considerably larger. Therefore, we introduce in the next section some current and new methods for selecting points from a dataset. Subset selection methods
Figure 1: Maximum Latin Hypercube Design of 20 points with 4 additional points.
10
AENORM
62
January 2009
Orthogonal Array Selection In the paper by Srivastava et al. (2004), Orthogonal Array Selection (OAS) is introduced. They apply it to the problem of selecting 500 or fewer points from a dataset containing 2490 points in 25-dimensional space. The selection process goes as follows: first a randomized orthogonal array is constructed. Second, for each point of the orthogonal array a ”nearest neigh-
Econometrics
bor” is determined, i.e. the data point closest to the orthogonal array point. All ”nearest neighbors” grouped together form the subset.
closest point. The point for which this distance is smallest is removed from the dataset. Sequential Selection In the four methods discussed above, the output values of the points are not used in the selection process. However as the selected points are used to determine a model of the output, it seems to be a good idea to explicitly use the output information in selecting the training points. We use the output values in the following way. First, we determine a small initial subset using one of the other methods. Then we fit a Kriging model to this subset and calculate the prediction error at the non-selected points. The idea is to add non-selected points with a large error to the subset. However, we should take into ac-
Fast Exchange Algorithm Lam et al. (2002) introduced the Fast Exchange Algorithm (FEX) in order to select a subset from a very large database containing characteristics of molecules. The algorithm aims to optimize the uniform cell coverage criterion in the following way. It starts with an initial subset and tries to improve it by exchanging points that are inside and outside the subset. This is done in two steps. The first step is to use a distribution of improvements to select a point to be added to the current subset. In the second step, the removable point is determined again using a distribution.
"Using uniform subsets, we indeed can find accurate Kriging models faster" Greedy MAXMIN Selection The greedy MAXMIN method originally comes from the dispersion problems’ field. As its name indicates, it seeks to maximize the minimal Euclidean distance between any two points in the subset. It starts with selecting the two points furthest away from each other. After that, it iteratively adds the point furthest away from the already chosen points.
count that if the Kriging model is inaccurate in a certain region, all points in that region will show a large error. Consequently, simply selecting for instance n1 points with the largest error might result in adding points that are clustered together in one or a couple of regions. To reduce this problem, we use two methods. The first method, which we call SS1, consists of adding one point at a time. This method completely solves the above-mentioned problem, but unfortunately it is quite time consuming to fit a new Kriging model after every added point. The second method determines the n2 > n1 worst points and then uses the greedy MAXMIN heuristic to select a uniform subset of n1 points. These points are then added to the training set. In our research, we have used n1= 10, n2= 40 and we refer to this method as SS1040.
Greedy DELETION Algorithm Besides MAXMIN Selection, we also use another simple greedy algorithm. This greedy method constructs a subset by iteratively removing one point of the pair of points with the smallest Euclidean distance between them. To decide which of the two points should be removed, we look for both points at the distance to its second Kriging
RAND
OAS
FEX
MAXMIN
DELETION
SS1
SS1040
RMSE
0.06
0.17
0.17
0.17
0.16
0.16
0.10
0.10
Maximum Error
0.76
1.62
1.60
1.56
1.54
1.53
0.96
0.94
Time Subset Selection
0.00
0.00
0.00
0.21
0.02
0.07
9.56
0.92
Time Model Fitting
15.09
0.10
0.08
0.10
0.05
0.05
0.08
0.08
Condition number
237859
1826
473
919
150
187
288
266
Average Robustness
6.33
1.11
0.94
0.98
0.82
0.84
0.67
0.71
Maximum Robustness
45.73
7.27
2.18
5.25
1.35
1.43
1.75
1.72
Table 2: Results obtained by taking subsets of 250 points from artificial datasets of 2000 points.
AENORM
62
January 2009
11
Econometrics
Numerical results We tested all five methods on artificial datasets of three different sizes and two levels of uniformity. Furthermore we performed tests on the HSCT dataset, which was also used by Srivastava et al. (2004). Table 2 shows results of one of the artificial datasets. The first column contains results of fitting a Kriging model to the original dataset and the second of using a randomly selected subset. For this dataset, the accuracy is reduced by taking a subset, but it is reduced the least for SS1 and SS1040. All the other performance measures improve by taking a subset and generally improve the most by using the SS1040, MAXMIN or DELETION algorithm. In general the tests show that by using uniform subsets, we can indeed find accurate Kriging models faster. Furthermore these Kriging models are more robust and less susceptible to numerical inaccuracies. When comparing the different methods for finding subsets, there is no overall winner. SS1040 generally performs well on accuracy, robustness and numerical accuracy. Compared to the other methods, SS1040 is relatively time consuming, but remains considerably faster than fitting a Kriging model to the complete dataset. The OAS method is considerably faster, but has a lower accuracy, robustness and numerical accuracy. Hence deciding which method is best for a practical application depends on how the different aspects are valued. More information This article is an abridged version of Rennen (2008), which can be downloaded for free online. References Booker, A.J., Dennis J.E., Frank, P.D., Serafini, D.B., Torczon, V., and Trosset, M.W. (1999). A rigorous framework for optimization of expensive functions by surrogates, Structural and Multidisciplinary Optimization, 17(1), 1–13. van Dam, E.R., Husslage, B.G.M., den Hertog, D. and Melissen, J.B.M. (2007). Maximin Latin hypercube designs in two dimensions, Operations Research, 55(1), 158–169. Davis, G.J. and Morris, M.D. (1997). Six factors which affect the condition number of matrices associated with Kriging, Mathematical Geology, 29, 669–683. Dixon, L.C.W. and Szegö, G.P. (1978). The global optimization problem: An introduction. In L.C.W. Dixon and G.P. Szegö (Eds.), Toward
12
AENORM
62
January 2009
Global Optimization, 2, 1–15, North-Holland. Jin, R., Chen, W. and Simpson, T. W. (2001). Comparative studies of metamodelling techniques under multiple modelling criteria, Structural and Multidisciplinary Optimization, 23, 1–13. Lam, R.L.H., Welch, W.J., and Young S.S. (2002). Uniform coverage designs for molecule selection, Technometrics, 44, 99–109. Rennen, G. (Published Online 2008). Subset Selection from Large Datasets for Kriging Modeling, Structural and Multidisciplinary Optimization. Sacks, J., Welch, W.J., Mitchell, T.J. and Wynn, H.P. (1989). Design and analysis of computer experiments, Statistical Science, 4, 409– 435. Siem, A.Y.D. and den Hertog, D. (2007). Kriging models that are robust with respect to simulation errors, CentER Discussion Paper 200768. Tilburg University. Srivastava, A., Hacker, K., Lewis, K. and Simpson, T.W. (2004). A method for using legacy data for metamodel-based design of large-scale systems, Structural and Multidisciplinary Optimization, 28, 145–155.
AENORM
62
January 2009
13
Actuarial Sciences
QIS4: The Treatment of Lapse Risk in a Life Portfolio One of the major challenges the European insurance industry faces is the introduction of Solvency II which is to be expected in 2012. Solvency II is a new risk-oriented framework for supervision. Within this framework rules are defined to determine the minimum amount of capital an insurance company has to possess in addition to its technical provisions. This summer a quantitative impact study (QIS4) has been performed by many European insurance companies. QIS4 was set up to test the new solvency-rules. In this article special attention is given to the treatment, within QIS4, of the risk connected to lapse and in particular to “pupping”. In case of pupping the policyholder continues the policy without further premium payment, thus creating a paid-up-policy (“pup”).
Rob Bruning1 studied actuarial science at the VU Amsterdam until 1984. He is now actuarial advisor at ASR Nederland and teacher in actuarial science at the University of Amsterdam (UVA). At ASR Nederland he is involved in the Solvency II project. During the early summer of 2008 he coordinated the Solvency II fieldstudy, QIS4, for all insurance companies within ASR Nederland.
Solvency II The new Solvency framework will be mandatory for all insurance companies within the European Union. Solvency II consists of three “pillars”, just like the solvency-regime for banks. The first pillar deals with the (quantitative) solvency requirements for insurance companies: what is the minimum amount of capital an insurance company has to keep on its balance sheet above the fair value provisions? Pillar 2 deals with the (internal) supervision and pillar 3 with transparency (disclosure). In the current Solvency I- regime the minimum required capital is mostly determined by simple rules. For example, the required capital can be 4% of the technical provision and 0,3% of the capital under risk. The provisions within Solvency I are based on assumptions of the historical price of the insurance, which for example include a fixed interest rate. Solvency II implies a radical break with these traditions:
1 Fair value approach on the whole balance sheet A fair value balance sheet is underlying the new solvency-framework. Also technical provisions are based on fair value. The available capital is derived from the balance sheet as the fair value of assets minus the fair value of liabilities. 2 Risk-orientation Every kind of risk an insurance company faces influences the level of required capital. Many of the risks that are recognised within Solvency II are measured by using shocks. For example: the capital charge for mortalityrisk is derived from a mortality shock scenario in which all assumed future mortality rates on best estimate basis are increased by 10%. The shocks are such that, theoretically, the probability to survive one year is 99,5%. An important aspect of Solvency II is the possibility that, under certain conditions, the insurance company is allowed to use an internal model. An internal model better fits the riskprofile of the business and can therefore lead to less capital requirement. QIS 4 During the early summer of 2008 the fourth quantitative impact study, QIS 4, was performed by many insurers. Participation was voluntary and the coordination was done by CEIOPS2. A main reason to conduct this field-test was to test the new solvency-rules. Furthermore, the test would help insurers to identify short-comings of
The author wishes to thank Drs. G.J.M. (Ger) Kock AAG, responsible for Insurance Risk & Value Management at ASR Nederland, for his comments on earlier versions of this article. The views expressed in this article are the personal view of the author. 2 Committee of European Insurance and Occupational Pension Supervisors; www.ceiops.eu 1
14
AENORM
62
January 2009
Actuarial Sciences
their models with respect to Solvency II. The results of QIS 4 for the Dutch companies are presented in a “Landenrapport”. One of the findings of this report is that the solvency margin (available capital above required capital) will increase under QIS4 for most life companies. In the following paragraph we will focus on the capital rules for a life-company. One of the remarks of the “Landenrapport” is that several companies found the mass lapse shock too conservative; an opinion that is not supported by DNB. Therefore we will especially evaluate the role of lapse in the solvency II-framework. Applying the new solvency-rules, an example The insurance portfolio We consider a specific portfolio of life-insurances, which all started 10 years ago. The insurances all provide a benefit in the case of death. All the insured were, at the start of the policies, 30-year old women. A premium has to be paid every year until the insured dies. This premium is based on the Dutch mortality table AG 2000-2005 for women and a fixed interest rate of 3%. The net premium is increased with 10% to cover expenses. Total insured capital is 100.000. The asset portfolio The assets consist of fixed interest government bonds, all with coupon rate of 4% and remaining time to maturity equal to 30 years. Total face value is 12.500. The balance sheet The balance sheet, shown in the next table, is composed according to Solvency I-principles and to QIS 4. We assume that within QIS 4 both the fair value of assets and liabilities is calculated with the yield-curve of year-end 2007 as provided by CEIOPS3. Balance Sheet Solvency I
QIS 4
10.776
10.776
Equity
1.158
8.243
Insurance Liabilities
9.618
329
Risk Margin
0
2.205
Total Liabilities
10.776
10.776
Total Assets Liabilities
The assets on the asset side are valued accor-
ding to fair value both under Solvency I and Solvency II. However, big differences occur on the liability-side. Under solvency I the insurance liability is still based on the mortality and interest assumption of the premium (AG 20002005 women, interest rate 3%)4. In QIS4 the liability is based on fair value and split in a “best estimate” (329) and a risk margin (2.205). The best estimate is the present value of the future insurance cash flow (premiums, benefits and expenses) on the basis of “best guess”, so without implicit prudency. In this example assumed future mortality rates are 90% of the rates used in the premium and lapse rates are 2% per year. Here, lapse means “pupping”: continuation of the policy without paying further premiums (creating a paid-up policy or “pup”). Of course in case of pupping the insured benefits are reduced. There is no surrender, which is usually not possible for this kind of insurances. The best estimate in this example is far below the Solvency I-provision, due to the fact that the level of the yield-curve at year-end 2007 lies far above the interest rate of 3%. The risk margin is an additional provision and can be seen as an explicit prudency-margin. According to QIS4-specifications the risk margin can be calculated as 6% of the net present value of the required capital with respect to insurance risks to be held in future year by year. The 6% is prescribed by CEIOPS5. The risk margin has a big impact on the liability and consequently on the available capital. The solvency position The solvency positions under both Solvency I and QIS 4 are shown in the next table: Solvency position Solvency I
QIS 4
Required
656
3.901
Available
1.158
8.243
Solvency Margin
502
4.342
Margin %%
77%
111%
Under QIS4 both the required capital and the available capital increase. Together this leads to a much better solvency margin under QIS4. The available capital is found on the balance-sheet presented before. How the required capital can be derived will be explained below. The required capital In Solvency I the required capital is determined as 4% of the technical provision and 0,3% of
See www.ceiops.eu In this example we assume that there is no profit-sharing. Otherwise the company might have created an extra shadow-accounting-provision according to IFRS 4. 5 This is the so-called “Cost-of-Capital method”. In this example some proxies have been used when determining the risk margin. 3 4
AENORM
62
January 2009
15
Je leert meer...
...als je niet voor de grootste kiest.
Wie graag goed wil leren zeilen, kan twee dingen onze nanciële, commerciële en IT-functies, maar doen. Je kunt aan boord stappen van een groot zeil- net zo goed voor onze traineeships waarin je schip en alles leren over een bepaald onderdeel, diverse functies bij verschillende afdelingen vervult. zoals de stand van het grootzeil of de fok. Of je kiest Waardoor je meer ervaring opdoet, meer leert en voor een iets kleiner bootje, waarop je al snel aan sneller groeit. SNS REAAL is met een balanstotaal van € 83 miljard en zo’n 7000 medehet roer staat en zelf de koers kunt bepalen. Starters werkers groot genoeg voor jouw ambities Zo werkt het ook met een startfunctie bij SNS REAAL, de innovatieve en snelgroeiende dienst- en klein genoeg voor een persoonlijk contact. verlener in bankieren en verzekeren. Waar je als Aan jou de keuze: laat je de koers van je carrière starter bij een hele grote organisatie vaak een vaste door anderen bepalen of sta je liever zelf aan plek krijgt met specieke werkzaamheden, kun je het roer? Kijk voor meer informatie over de je aan boord bij SNS REAAL in de volle breedte startfuncties en traineeships van SNS REAAL op van onze organisatie ontwikkelen. Dat geldt voor www.werkenbijsnsreaal.nl.
16
AENORM
62
January 2009
Actuarial Sciences
Actuarial Sciences
the capital at risk. The required capital under Solvency II is called the Solvency Capital Requirement (SCR) and comprises the elements shown in the next table: SCR Market (interest)
3.048
Life
1.745
Basic SCR
3.872
Operational risk
29
SCR
3.901
flation. But let us now focus on the lapse risk. The lapse risk The lapse risk is the maximum of three shock scenario’s. The result of the “lapse down”-shock (-568) is the change in fair value (assets – liabilities) when all future lapse rates reduce by 50%. The shock is negative which means that lower lapse rates are profitable for the insurer. In the “lapse up”-shock scenario the lapse rates are increased by 50%. The change in fair value is now 478, which implies that higher lapse rates are unfavourable to the insurer. The third lapse shock is a kind of lapse catastrophe sce-
"In a fair value environment the Lapse risk becomes explicitly visible" In this example the market risk is the dominating factor of the SCR. Market risks are risks related to assets and interest. In this simple example only interest rate risk plays a roll, i.e. the potential effect on the fair value (assets – liabilities) when the yield-curve changes6. In general, market-risk is the dominant risk for life-insurers within QIS4, with exception of unit-linked portfolio’s. In this example, the category of life risk consists of four sub-risks that contribute to the SCR: mortality, lapse, expense and catastrophe risk. Split up SCR Life Mortality
561
Longevity
0
Disability
0
lapse down
-568
lapse up
478
lapse mass
1.491
Lapse
1.491
Expense
242
Revision
0
CAT-risk
142
SCR Life
1.745
The lapse risk is, perhaps surprisingly, the main factor of the life-risk. Before we explore lapse risk further, we should mention that mortality risk is the risk of a sudden and permanent increase of mortality rates (by 10%), that catastrophe risk is the risk of a sudden increase of mortality rates for 1 year (1.5 per 1000 for every age) and expenses risk is the risk of higher expense than expected, for example due to in6
nario: the change in fair value in case of an immediate lapse of 30% of the portfolio. In this example the mass-lapse scenario is dominant. In the traditional world of Solvency I lapse was not recognised as a risk. In case of pupping the provision does not change in Solvency I (except for possible lapse-penalties). In case of surrender a surrender payment is compensated by release of provision. This example shows that pupping is a severe risk for the insurer in fair value terms (remember that in this example surrender does not exist). The fair value provision after pupping appears to be much higher than before pupping thus creating a loss for the insurer. Consequently, the mass lapse scenario of QIS4 leads to a (big) SCR. However, it should be noted that the volume of the lapse risk depends on the yield-curve. In case of a low yield-curve, pupping can be more favourable for the insurer then continuation of premium payment. Some comments on the mass lapse risk In a fair value environment the lapse risk becomes explicitly visible. For many life-portfolios the mass lapse shock in QIS 4 is the dominant factor within lapse risk. Some insurers have argued that a mass lapse of 30% is too high, but DNB does not agree with this opinion. Indeed a mass surrender is far less likely to occur than the disastrous situation that Banks can encounter; i.e. large numbers of clients transferring their money. Policy restrictions and fiscal rules can prevent policyholders from surrender. But lapse is not only surrender: the example shows that the option of “pupping” (continuation of the policy without further premium payment) is also a risk for the insurer. This risk arises
Details can be found in the QIS4 technical specifications. See www.ceiops.eu
AENORM
62
January 2009
17
Actuarial Sciences
especially in a high-interest world, where the fair-value provision is low and the provision is not enough to withstand lapse. Note that in the Netherlands the technical provision on the balance sheet currently has to meet the requirement that it is at least equal to the surrender value or the provision after pupping (surrenderfloor)7. The mass lapse shock of 30% within QIS4 must be seen in the context of the lack of such a surrenderfloor on the provision in the fair value balance sheet. In order to get an idea of the impact of lapse, the following table shows the solvency positions under QIS4 for the example described in the previous paragraph, under three alternative mass-shock scenarios: 0%, 30% and 100%.
References Bouwknegt, P. and Pelsser, A. (2003). The value of Non Enforcable Future Premiums in Life Insurance, Paper for AFIR Colloquium 2003, Maastricht. DNB (2008). QIS4 on Solvency II: Country Report for The Netherlands, (“Landenrapport”). Available at: www.dnb. nl/openboek/extern. Liedorp, F. and van Welie, D. (2008). QIS4 – Laatste halte voor Solvency II?, De Actuaris, 16(2), 42-45.
Solvency position QIS 4 Required
0%
30%
100%
3.414
3.901
6.621
Available
8.815
8.243
5.593
Solvency Margin
5.400
4.342
-1.028
Margin %%
158%
111%
-16%
A 30% mass lapse is relatively close to the 0% mass lapse alternative, because in the 0% alternative there is still a lapse risk of 478 due to the lapse-up shock. Note that the choice of mass lapse also affects the available capital, because the risk margin depends on the SCR for life risk. Conclusions QIS4 shows an increase of the solvency margin for most life companies. Lapse plays a dominant role within the life risk. When interest rates are high, the fair value provisions might be too low to withstand lapse. This can occur in case of surrender but also in case of pupping (continuation of the policy without any further premium payments). When measuring lapse risk within QIS 4 the insurer has to investigate which kind of lapse is the most unfavourable. The conclusion that a 30% mass lapse scenario within QIS4 is too conservative, as some insurers have stated, can be questioned. In a scenario of economic recession pupping poses a serious danger, perhaps even larger than surrender, because pupping is for the policyholder often easier to perform. Lapse, both surrender and pupping, can be a serious risk in a fair value world, because in that world there is no surrenderfloor on the technical provision anymore.
7
18
See: Staatsblad van het Koninkrijk der Nederlanden, jaargang 2006: Besluit prudentiële regels Wft, artikel 116.
AENORM
62
January 2009
ORM
Jams are Everywhere - How Physics may Help to Solve Our Traffic Problems Traffic jams are annoying and cause a lot of economic damages. Therefore it is surprising that despite all technological advances we still have to suffer from them. So naturally, the questions arise "Why can't traffic jams been avoided?" or "Can technology help us to make traffic more fluent?".
Surprisingly physicists have been studying such questions for a long time now, going back to the 50's of the last century. Beside the practical relevance of this topic, for a physicist traffic has much more to offer. In the early days traffic was mostly seen as an exotic fluid with some strange properties. Water, for instance, flows faster at bottlenecks, an effect you can easily test with a garden hose. However, this is very different for traffic; a jam will form at such a bottleneck. So, early theories of highway traffic were inspired by fluid-dynamical theories. These theories have the disadvantage of not distinguishing between individual particles or cars, but use densities as basic quantities In the early 90’s, a new class of models has been proposed for the description and simulation of traffic flow, the so-called cellular automata (CA). CA are a perfect tool for studying complex systems. A popular variant is the “Game of Life” which allows creating a simple “world” with moving, reproducing and dying entities. Cellular automata are ideal tools for modeling nonphysical or interdisciplinary problems. They are based on simple and intuitive rules which determine the dynamics of the “particles” instead of forcing action between them. The other main feature of CA is their discreteness: space is not treated as a continuum, but is divided into cells of a finite size. This approximation makes CA perfectly suited for computer simulations which can be done much more efficiently. In 1992, Kai Nagel and Michael Schreckenberg, at the time both affiliated with the Cologne University, have proposed the basic CA for traffic flow. The highway is divided into cells with a length of about 7.5 meter which corresponds to the typical space occupied by a car in a dense jam. Each cell can either be empty or occupied by exactly one car. The state of each car n (n=1,2,…,N) is characterized by its velocity vn
Andreas Schadschneider is professor in theoretical physics at Cologne University, Germany. He is working in the area of statistical physics, and has specialized on nonequilibrium systems. His main interest are interdisciplinary applications of physical methods, e.g. to highway traffic, pedestrian and evacuation dynamics, biological and economic systems.
which can take one of the vmax+1 allowed integer values v=0,1,…, vmax (see Fig. 1). The dynamics that determine the motion of cars consists of four simple and intuitive steps. First, all cars accelerate by a specified unit if they have not already reached the maximum velocity vmax. In the second step, drivers check whether they can drive safely at their current speed. If there is any danger of causing an accident, the velocity is reduced to a safe value. The third step is stochastic and takes into account that drivers are not driving perfectly, e.g. due to fluctuations in their behavior. In this step, the velocity of a car is reduced by one unit (vn → vn—1) with some probability p. This is sometimes called “dawdling”. In the final step each car moves forward the number of cells given by its current velocity. One such update cycle, which is performed synchronously for all cars, roughly corresponds to 1 second of real time. The simulation of one hour of real traffic thus requires 3600 update cycles. The Nagel-Schreckenberg model is able to explain many of the empirical observations made for highway traffic. Probably the most fascina-
Figure 1: A typical configuration in the NaSch model. The number in the upper right corner is the speed vn of the vehicle.
AENORM
62
January 2009
19
ORM
Figure 2: Experiment showing the occurance of phantom jams.
ting of these are the so-called “phantom traffic jams” or “jams out of nowhere” or, more scientifically, “spontaneous jams”. These are jams which occur without an obvious reason, like an accident or road construction. Most of you have probably encountered this phenomenon: You are standing in a traffic jam for some time and after it has evaporated you wonder what the reason of the jam was.
Figure 3: Schematic representation of the fundamental diagram. F denotes the free flow branch and J the jam line. The bubble represents the synchronized phase where the flow (or velocity) is not uniquely determined by the density.
few cars are on the highway, all can drive at the desired speed. Here the velocity is independent of density. Increasing the number of cars leads to interactions which reduce the average velocity, even up to creating jams. Densities of around 15% of the maximum already lead to a strong reduction of the average velocity on a highway.
"The main problem is still the human factor" In fact the occurrence of such jams has been studied in experiments. Fig. 2 shows a snapshot of such an experiment which has been performed for the German TV station ARD. About 25 drivers were asked to move on a circular course as fast as possible, but without causing an accident. After a few minutes of free flow, suddenly a jam was formed. The reason is that drivers overreact. If they approach the preceding car with too much speed, they have to brake in order to avoid an accident. Usually they brake stronger than necessary, i.e. they overreact, and thus force the drivers behind them to brake as well. If the traffic density is large enough this starts a chain reaction and, about 10 cars later, the first car has to stop completely and thus a jam is created. Exactly these overreactions, which occur when drivers start to lose their concentration, are captured in step 3 of the dynamics in the Nagel-Schreckenberg model. Apart from this phenomenon the model also reproduces more quantitative aspects of traffic flow, e.g. the so-called “fundamental diagram”. This aspect is, as already indicated by its name, very important in traffic engineering. It relates the average velocity of the vehicles with the traffic density. When the density is small and
20
AENORM
62
January 2009
Fig. 3 shows a schematic representation of the fundamental diagram according to the so-called three-phase theory. Instead of the average velocity it shows the density-dependence of the flow (or current). This quantity is easy to measure: one just has to count the number of cars passing a certain point within one hour. B. Kerner surprisingly discovered that the flow in a certain regime is not uniquely determined by the density, but can vary over a whole interval. For historical reasons this regime is called “synchronized phase” although “viscous traffic” may be a more appropriate description. There is still some debate about its precise origin, but it appears that the formation of platoons (clusters of cars moving with a small inter-vehicle distance) plays an important role. Although the model of Nagel and Schreckenberg captures the basic aspects of highway traffic, some aspects need to be improved, e.g. it cannot explain the synchronized phase. In real traffic it has been found that “anticipation” plays an important role. This means that drivers try to guess what the preceding driver is going to do. If (s)he has a clear road ahead then often headways are accepted which are much shorter than safe, e.g. headways which allow to avoid
ORM
an accident in case the preceding car starts to brake suddenly. The improved models based on the NagelSchreckenberg model nowadays perform so well that they can be used for traffic forecasting. The simplicity allows computers to perform simulations faster than real time: one hour of traffic on the German highway system can be simulated within seconds. Only this makes traffic forecasting possible. Indeed the predictions are so reliable that already an official forecast for the German state of Northrhine-Westphalia is available which is used by more than 200.000 persons. It can be found at www.autobahn.nrw. de and displays not only the current state of the highway system, based on measurements from inductive loops, but also 30 minute and 60 minute forecasts. So, what is the future of traffic management? As we have seen, reliable traffic forecasting is already possible today. But surprisingly it turns out that predicting traffic is in one respect much more difficult than weather forecasting. The difference is the human factor! The weather does not care if a forecast is shown on TV every evening. This will have no influence on the future development. For traffic, however, the situation is very different. Publication of a forecast will have a direct influence on the decisions of the drivers. Maybe they decide to take a different route, leave at a different time or even use public transport instead. So the public announcement of a traffic forecast immediately renders the forecast obsolete. Therefore the next big step in improving the reliability of these predictions comes from a better understanding of how drivers react to forecasts. First tests have already been performed by Schreckenberg in collaboration with Reinhard Selten, who received the Nobel Prize for economics (together with John Nash) for his work on game theory. Incorporating such results into the forecast should lead to a further improvement of its reliability.
We see that technology might indeed help to reduce the traffic problems encountered in everyday life. However, the main problem is still the “human factor”. Spontaneous jams occur because drivers are human and do not react perfectly in all situations. This could be improved by driver assistance systems, but one still has an acceptance problem because many drivers enjoy their freedom on the roads (especially in Germany, where there is no general speed limit on highways) and do not want to be controlled by “machines”. However, this might change in the near future if one realizes that such support systems allow for much more relaxed driving and reduce the probability of jams. References Chowdhury, D., Santen, L. and Schadschneider, A. (2000). Statistical Physics of Vehicular Traffic and Some Related Systems, Physics Reports, 329, 199. Chowdhury, D., Santen, L. and Schadschneider, A (2000). Simulation of vehicular traffic: a statistical physics perspective, Computing in Science and Engineering, 2, 80. Kerner, B.S. (2004). The Physics of Traffic, Springer. Knospe, W., Santen, L., Schadschneider, A., and Schreckenberg M. (2005). Optimization of highway networks and traffic forecasting, Physica, A346, 163. Schadschneider, A. (2006). Cellular automata models of highway traffic, Physica, A372, 142. A Java applet for the simulation of the NagelSchreckenberg model can be found at http:// www.thp.uni-koeln.de/~as.
Figure 4: Snapshot from the webpage www.autobahn.nrw.de which provides not only information about the current state of the highways in Northrhine-Westphalia, but also 30 and 60 minute forecasts.
AENORM
62
January 2009
21
Econometrics
Regional Real Exchange Rates and Phillips Curves in Monetary Unions: Evidence from the US and EMU Comparing inflation developments in the US and the euro area as a whole over the past quarter of a century, the most striking fact emanating from the data is that the differences are minor (see figure 1). The overall picture is one of steadily declining inflation rates on both sides of the Atlantic, sometimes interrupted by temporary hiccups due to external turmoil (late seventies) or strong domestic demand (late eighties and late nineties). There is also resemblance of inflation dynamics in the short run, with the euro area clearly lagging behind the US by one to two years. This correspondence in inflation patterns of the two economic blocks is remarkable for at least two reasons. First, it occurred in spite of huge swings in the exchange rate of the US dollar vis-à-vis European currencies during the period. Second, it occurred in spite of huge, although declining, inflation differentials between euro area countries, whereas regional inflation rates in the US moved more or less in tandem, as figure 1 shows. To ensure that the difference in inflation dispersion between the US and the euro area does not stem from a difference in aggregation, we have disaggregated the euro area into countries and into four large regions (comparable to the US). It is evident from figure 1 that the level of aggregation hardly matters. This paper deals with the second observation.
Jan Marc Berk is Head of Research and deputy director of the division Economics and Research of De Nederlandsche Bank. He is also a member of the Monetary Policy Committee of the ESCB, and of the board of the International Journal of Central Banking and SUERF (The European Money and Finance Forum). Job Swank is director of the division Economics and Research of De Nederlandsche Bank. He is also a member of the International Relations Committee of the ESCB, Deputy Crown member of the Economic Council of the Netherlands (SER) and part-time professor of Economic Policy, Faculty of Economics, Erasmus University Rotterdam.
Are regional inflation differentials in a monetary union a cause for concern? The answer is: it depends. If regional inflation differentials are the counterpart of price level convergence across the monetary union after a region or a subset of regions has been hit by an asymmetric supply shock, they are a welcome phenomenon and temporary by nature. If regional inflation dif-
ferentials are due to relative catching-up of regions (the so-called Balassa-Samuelson effect), they are part of a natural process and, in spite of their longer persistence, of no immediate concern to the central bank either. Things are different, however, if regional inflation differentials are the manifestation of rigidities in wage and price formation in certain regions, thereby potentially thwarting rather than effectuating price level convergence in the monetary union. As a consequence, monetary policy – typically aimed at controlling the average rate of inflation in the monetary union – may be suboptimal for some or all regions individually. This is why the European Central Bank (ECB) repeatedly stresses the need for structural reforms in labour and product markets in euro area member states. To see whether such appeals are justified in the present context, it is instructive to disentangle the causes of the regional inflation differentials depicted in figure 1, and to compare them between the two monetary unions. Our research1 aims to establish whether regional inflation differentials in the euro area and in
For details, see Berk, J.M. and Swank, J. (2007), ‘Regional real exchange rates and Phillips curves in monetary unions: evidence from the US and EMU’, DNB Working paper no 147, september 2007. Downloadable from www. dnb.nl 1
22
AENORM
62
January 2009
Econometrics
Figure 2: Real exchange rates Portugal and France vis-à-vis euro area: Actual and trend 1974-2004
Figure 1: Inflation dispersion in the US and the euro area 1977-2004
the US are persistent, or a reflection of an equilibrating mechanism in the monetary union. We take the results for the US, a well-established and fully integrated monetary union, as a benchmark against which regional price dynamics in the euro area can be compared to. We divide the US into 4 regions (Midwest, Northeast, South and West), and the euro area into the
mately reverts to a particular equilibrium path. Figure 2 illustrates this in the case of France and Portugal. The equilibrium path can exhibit a (deterministic) trend-like pattern due to the Balassa-Samuelson effect mentioned earlier. This is modelled and estimated as a closed system in that each regional real exchange rate has as numeraire a weighted average of all regional prices in the monetary union rather than the price of one particular region, implying that any Balassa-Samuelson effects cancel out in the aggregate. As we define this effect vis-à-vis the rest of the monetary union, some regions will have a negative trend and the sum of these effects (that is, for the entire monetary union) will be zero by definition. This condition implies that regional inflation processes in a monetary
"Regional real exchange rates adjust significantly faster to random shocks in the euro area than in the US" countries that are the founding members (that is, current member countries excluding Greece and Slovenia). Our sample period runs from 1977 until 2005, and therefore embeds both the ERM-episode and the first six years of EMU. In this way, we can test whether the transition from quasi-fixed to irrevocably fixed exchange rates has influeced the price formation proces in the euro area. The starting point of our analysis is the hypothesis that the relationship between regional prices in a monetary union may temporarily change due to transitory shocks, but is characterised by mean reversion and therefore ulti2
union are interdependent, which has implications for the econometric modelling of these processes. Our data allow us to accept the hypothesis of mean reversion of relative regional price levels – or regional real exchange rates – for both monetary unions.2 If regional real exchange rates are (trend) stationary, regional prices are cointegrated, granted that they have a unit root (all series in logs). We build on this property, using Granger’s representation theorem, to derive a set of regional Phillips curves of the hybrid New-Keynesian type. Regional inflation rates in this framework act to restore purchasing
Formally: we reject the null of no mean reversion.
AENORM
62
January 2009
23
Econometrics
power parity (PPP), barring Balassa-Samuelson effects, but also depend on explanatory variables typically included in the closed economy version of the hybrid Phillips curve: lagged inflation, expected future inflation and a proxy for real marginal cost. In fact, by exploiting the unit root properties of regional prices and regional real exchange rates, we obtain an open economy version of the hybrid Phillips curve that can be reconciled with a form derived from first principles. Put differently, we show that the open economy Phillips curve is reconcilable with PPP. We estimate such a model for both the US regions and the euro area countries, taking the abovementioned interdependencies of regions within the monetary union into account. We take an diagnostic view regarding the expectations formation mechanism, using both backward looking and forward-looking (ie rational expectations) processes. We furthermore investigate whether the price fomation process in the euro area has changed due to the irrevocably fixing of exchange rates in 1999. Our main finding is that the speed by which regional real exchange rates adjust to random shocks is, on average, significantly higher in the euro area than in the US, once BalassaSamuelson effects are taken into account. The average half life of such shocks is around two years in the euro area, against three years in the US, considerably shorter than the half lives reported in most studies of PPP. We also find that since the start of EMU, inflation rates of individual member states have taken over the role that nominal exchange rate adjustments used to play prior to EMU. This is quite surprising, as European labour and product markets are not known for their flexibility compared to the US. Our conjecture is that the near absence of labour mobility across euro area countries places a relatively great weight on real exchange rate adjustments after a country or a subset of countries has been hit by an asymmetric shock. Finally, while there is clear evidence of forwardlooking pricing behaviour in both monetary unions, inflation persistence is nevertheless substantial, especially in the US. All in all, although inflation differentials between euro area countries are larger than between regions in the US, this can not be said of the persistence of those differentials between euro area countries. Once Balassa-Samuelson effects are taken into account, it seems that regional inflation differentials in both the US and the euro area since 1999 are indicative of a economically healthy restoration of competitive forces within the respective monetary unions. The conclusion therefore seems warranted that these inflation differentials are not much of a cause for concern for the ECB!
24
AENORM
62
January 2009
Wat als zijn overboeking naar Hongkong halverwege de weg kwijtraakt? Een paar miljoen overmaken is zo gebeurd. Binnen enkele seconden is het aan de andere kant van de wereld. Hij twijfelt er niet aan dat zijn geld de juiste bestemming bereikt. Het gaat immers altijd goed. Maar wat als het toch van de weg af zou raken? Door hackers, fraude of een computerstoring? Daarom levert de Nederlandsche Bank (DNB) een bijdrage aan een zo soepel en veilig mogelijk betalingsverkeer. We onderhouden de betaalsystemen, grijpen in als problemen ontstaan en onderzoeken nieuwe betaalmogelijkheden. Het betalingsverkeer in goede banen leiden, is niet de enige taak van DNB. We houden ook toezicht op de financiële instellingen en dragen – als onderdeel van het Europese Stelsel van Centrale Banken – bij aan een solide monetair beleid. Zo maken we ons sterk voor de financiële stabiliteit van Nederland. Want vertrouwen in ons financiële stelsel is de voorwaarde voor welvaart en een gezonde economie. Wil jij daaraan meewerken? Kijk dan op www.werkenbijdnb.nl. | Economen | Juristen
Werken aan vertrouwen. AENORM
62
January 2009
25
Econometrics
Hedging Prepayment Risk on Retail Mortgages Hedging the interest rate risk of a retail mortgage portfolio is a difficult task for banks. Besides normal interest rate risk, there is risk caused by embedded options i.e. choices incorporated in the mortgage contract. In regular Dutch retail mortgages, the customer receives several embedded options. One of these options is the option to prepay the mortgage without incurring additional costs in case the customer moves to a new home. In this article we explain why there is risk when the customer receives this option. If the customer does not use the option, he will continue the current contract. We will discuss how the resulting interest rate risk can be hedged and how to determine the hedging price. However let us first discuss the mortgage’s financing.
Dirk Veldhuizen graduated in the master of Econometrics and OR and the master of Finance (honours track Quantitative Finance) at the Vrije Universiteit in September 2008. He now works at SNS Reaal as Risk Management Trainee. This article is based on his thesis for Quantitative Finance, written at SNS Reaal. Dr. Svetlana Borovkova (VU) and dr. Frans Boshuizen (SNS Reaal) supervised the thesis.
How to finance a mortgage and why is there prepayment risk? When a customer acquires a mortgage loan, he or she usually borrows money for a long period of time, say 10 years. Moreover he will have to pay a fixed interest rate for the entire period. In order to fund this mortgage the bank could issue covered bonds or attract retail savings, which is cash on ordinary savings accounts. Additionally in order to prevent interest rate risk, the bank needs to make sure that the mortgage and funding have the same maturity time i.e. they have the same duration. This will ensure that the market value of the assets and liabilities remains the same. Furthermore in certain circumstances the bank may be unable to find funding with the same maturity as the mortgage. In this case, the bank could hedge the resulting interest rate risk by entering into a regular interest rate swap. This enables the bank to interchange the interest rate on the acquired funding with the interest rate on funding with the same maturity as the mortgage. We show how this works for an index amortizing swap in Figure 1. Now let us discuss how the embedded option can yield a loss to the bank. First of all, suppose that the bank matched its mortgages and funding perfectly. Now, suppose that the yield
26
AENORM
62
January 2009
curve otherwise known as the term structure of interest rates, decreases for all interest rates. A consumer who moves to a new home can now prepay his mortgage for free and purchase a new mortgage loan for a lower interest rate. However the bank’s funding cannot be prepaid without compensating the counterparties for the interest rate decrease. Moreover the bonds’ value increase as the interest rates decrease; this is due to discounting. Thus the bank makes a loss when it repays the funding. Another option for the bank is keeping the relatively expensive funding, which eventually results in a loss over time compared to using new funding. Prepayment risk is defined as the interest rate risk due to early repayments or prepayments of mortgages. We will only focus on prepayments which are driven by the current interest rates on mortgages. We assume that non-interest rate driven prepayments are independent of interest rate movements. Hedging strategies A bank can set up a strategy to hedge prepayment risk. It needs to buy financial derivatives such that the cash of the prepaid mortgages can be lent to other parties while still yielding a sufficient interest rate. The bank needs a portfolio of receiver swaptions1 for this hedging strategy. It can also hedge its prepayment risk dynamically by checking every day, week or month what its current portfolio of mortgages and funding is. The mismatch can be hedged using regular swaps. The costs of this last strategy are unknown beforehand and it does not require any option premium. The challenge with the first hedging strategy is to determine the required amount of receiver swaptions. For example, when interest rates
Econometrics
Figure 1. Schematics of a hedge using an index amortizing payer swap (IAPS). Solid lines resemble the interest rate payments (i.r.p.). The dashed lines resemble the prepayments. The bank receives fixed interest rate payments on the mortgages, but borrows funding for a floating interest rate. This way the bank can re-use the cash of the prepaid mortgages for new loans for current market prices and prevent a loss on prepayments. The index amortizing swap with the prepayments, which equal the outcomes of the amortization function, exchanges exactly the right amount of fixed interest rate payments into floating interest rate payments to eliminate the interest rate risk of borrowing for a floating rate and lending for a fixed rate.
have been relatively high as in the fall of 2008, interest rate driven prepayments will be lower than expected. The receiver swaptions will expire worthless, because the fixed rate received on new swaps is higher, while the mortgage portfolio stays large. When interest rates fall for example in a few years, more customers than expected will be able to prepay their mortgage. Now the swaption portfolio is not large enough to cover all prepayments, which will result in a loss for the bank. Advanced hedging amortizing swaps
strategy
using
index
The use of index amortizing payer swaps, which is an over-the-counter contract, can be the answer to this challenge. The key feature of an index-amortizing swap is that the notional of the swap decreases or amortizes based on a function instead of a predetermined scheme. This function is usually based on a reference rate. We use the interest rate for a regular bond, which matures at the same time as our swapcontract, as reference rate. The bank chooses a function which yields a smaller amortization of the notional when the reference rate is high and a larger amortization when interest rates are low. This corresponds to observations of the prepayments on Dutch retail mortgages. For an index amortizing payer swap to be ef-
fective, the bank needs to pay a floating interest rate on the funding preferably one which resets every month. Using an index amortizing payer swap2 on the entire mortgage portfolio, the bank changes the floating interest rate payments it has to pay into the required fixed rate payments. Disregarding credit risk and assuming that the amortization function perfectly matches the customer behaviour, Figure 1 shows how the index amortizing payer swap works. The risks associated with an index amortizing swap are described in Galaif (1993). There are the usual risks like counterparty risk and liquidity risk, an index-amortizing swap being an over-the-counter product. However there is also model risk, which we will address later on. The product also has an interesting interest rate risk, which is asymmetric for the fixed rate payer and fixed rate receiver. Additionally the fixed rate payer has less interest rate risk than the fixed rate receiver. When the yield curve decreases and hence the reference rate decreases the fixed rate payer incurs a loss. This loss is relatively small because the amortization is high and the size of the contract decreases rapidly. Also, when the yield curve increases, the amortization is low and the fixed ratepayer makes a relatively large profit. As mentioned in the above, there is also model risk associated with index amortizing swaps.
A receiver swaption is an option to enter into an interest rate swap where you receive a fixed interest rate and pay a floating interest rate. The floating interest rate can be received on the lent cash. When the interest rates fall and you exercise the option, you can transform a relative low floating rate into a relatively high fixed rate. The net payoff of this swaption (when exercised) is the present value of the difference between the fixed interest rate received on the underlying swap of the swaption and the fixed rate which would be received when you enter a new at-the-money receiver swap. 2 When you enter a payer swap, you pay the fixed interest rate and receive the floating rate. When you enter a receiver swap, you pay the floating rate and receive the fixed rate. 7
AENORM
62
January 2009
27
IN DE LIFT BIJ
IK WORD HIER DIRECTEUR.
TRAINEES De tijd van traditioneel verzekeren is voorbij. ‘All-finance’ is de toekomst. En Delta Lloyd Groep wil hierin haar leidende marktpositie uitbouwen. Daarvoor hebben we mensen nodig. Heel goede mensen. Zo selecteren we ieder jaar een aantal afgestudeerde academici voor onze Trainee Programma’s die opleiden tot een leidinggevende functie binnen het concern. Ook hebben we een Business Course en het Young Talent Network waarin jonge, hoog opgeleide medewerkers elkaar inspireren tot bijzondere prestaties. Waarmee we maar willen zeggen: als je wilt, kun je bij Delta Lloyd Groep heel ver komen. Aan ons zal het niet liggen. Zet jezelf in de lift. Kijk op werken bij deltalloydgroep .nl
D E LTA L LO Y D G R O E P I S O N D E R A N D E R E D E LTA L LO Y D , O H R A E N A B N A M R O V E R Z E K E R I N G E N
28
AENORM
62
January 2009
Econometrics
Because the notional depends on a reference rate, the price of this product is path dependent. We need the entire history of the reference rate from the start of the swap in order to determine the current notional. One method to
same amount from the same counterparty only increases counterparty risk while it does not change the net cash flows of the swap. However for the pricing, it is convenient to pretend we exchange the notional.
"Another challenge in using an index amortizing swap as a hedge is to determine the correct amortization function" price this swap is simulation. For this we need an interest rate model, which may not capture the reality correctly. Fernard (1993) discusses a relatively simple example of how to price an index-amortizing swap. We will develop a general scheme for a more complex function. Another challenge when using an index- amortizing swap as a hedge is to determine the correct amortization function, i.e., the function which matches the prepayments of a portfolio of mortgages. This can prove to be rather difficult in practice. Having an inaccurate amortization function will hurt the effectiveness of the hedge. This is firm specific basis risk when using this product for a hedging strategy. We have seen before that the use of receiver swaptions yields the same type of risk.
Additionally when exchanging the notional, the value of a floating rate bond must equal the notional at the start of the contract. Setting the notional equal to 1, we have to find the fixed interest rate for which the expected present value of the interest payments, the prepayments and the remaining notional amount at the end of the contract, equals 1. It is important to note that we assume a risk-neutral world. Hence this translates into the following formula: E (rfixed ⋅ ∑ t =1 mn
Lt ⋅ dft m
+ ∑ t =1 (Lt −1 − Lt ) ⋅ df + Lmn ⋅ df ) = 1, mn −1
Where rfixed is the fixed interest rate which needs to be paid, m is the number of payments per year, n is the maturity of the contract in years, Lt is the notional amount at time t and dft is the discount factor for the tth interest payment. The notional amount Lt equals Lt= Lt-1— PRt-1, where PRt-1 is the prepayment in period t (we assume PR0=0 ), which is determined by our prepayment function. In order to determine the fixed rate of the index amortizing swap, we use the simulation scheme 1: Because the index-amortizing swap is path de-
Pricing an index amortizing swap In order to price an index amortizing swap, we determine the par fixed interest rate on the swap (the fixed interest rate such that the contract has no value to either parties). We thus need to equate the value of a bond, which pays a fixed interest rate and another, which pays a floating interest rate. Interest rate swaps exclude the exchange of the notional. Lending the notional to the counterparty and borrowing the
Step 0: Choose and calibrate a model for the yield curve. Step 1: Simulate the relevant part of the yield curve for all future dates in a risk-neutral world. Step 2: Construct the reference rate. Step 3: Calculate the prepayments per period. Step 4: Determine the present value of the interest payments, prepayments and repayment of the remaining notional. Step 5: Repeat step 1-4, N times, where N is the number of simulations we want to run. Step 6: Calculate the par swap rate by solving the following equation: i N nm L N nm i N ⋅ ∑ t 1 t ⋅ dfti ) += (∑ t 1 (Lit −1 − Lit ) ⋅ dfti + Linm ⋅ dfnm )= ∑ i 1 (rfixed = ∑ i 1= = m with i the index for the simulations and t the index for the time in a simulation. This yields: − ∑ = ∑ = − − ⋅ + ⋅
=
∑
= ∑ =
− ⋅
Scheme 1
AENORM
62
January 2009
29
Econometrics
pendent, we have to use this somewhat large formula instead of using the average fixed rate of all separate simulation runs. The resulting outcomes depend on the realizations of the yield curve. Greeks of an index amortizing swap Besides calculating the par fixed rate of an index amortizing payer swap, calculating the Greeks or sensitivity with respect to several parameters is a challenge as well. Glasserman (2004) describes three different solutions to calculating the Greeks in simulation. Two of those are semi analytical solutions. These solutions are however infeasible, because the path dependence of the swap creates very complex derivatives with respect to parameters (high dimensional with many terms). In the third solution, the Finite-Difference Approximation, we calculate the Greeks using brute force. First, we calculate the value of many different values for the parameters and then calculate the derivative by computing: =
+ ∆ − − ∆ ∆
where v(x) is the value of the index amortizing swap at parameters X and Δ is a small change in one of the parameters. Summary In this paper we first discussed the principles of an advanced hedging strategy for prepayment risk using index amortizing swaps and then we computed the price and Greeks of an indexamortizing swap. Moreover we argued that when using an index-amortizing swap we could avoid answering the question of how large the position in hedging instruments should be. This can make index amortizing swaps more accurate than a portfolio of regular receiver swaptions. However, we still need to estimate a prepayment function, which can be challenging in practice. References Boshuizen, F., Van der Vaart, A.W., Van Zanten, H., Banachewicz, K. and Zareba, P. (2006). Lecture notes course Stochastic Processes for Finance, Vrije Universiteit Amsterdam. Galaif, L. N. (1993). Index amortizing rate swaps, Quarterly Review Issue Winter, 63-70 Fernard, J. D. (1993). The pricing and hedging of index amortizing rate swaps, Quarterly Review Issue Winter, 71-74. Glasserman, P. (2004). Monte Carlo Methods
30
AENORM
62
January 2009
in Financial Engineering, New York : Springer Science and Business Media. Hull, J.C. (2006). Options, futures, and other derivatives 6th ed, Upper Saddle River: Pearson/Prentice Hall. Hull, J.C. and White, A. (1990). Pricing Interest Rate Derivate Securities, Review of Financial Studies, 3(4), 573-592.
ORM
ORM
Determining a Logistics Network for Toyota’s Spare Parts Distribution Like any car manufacturer Toyota produces and distributes spare parts for maintenance or repair of sold cars. Toyota’s customers in the spare parts market are dealers, that sell them to car repair shops, or use them themselves. To distribute the spare parts to the car dealers Toyota has set up a highly responsive distribution network that is able to deliver to any dealer within hours after an order has been placed. Peculiar about this network is that it is only partly owned by Toyota and involves a large number of so-called third-party logistics (3PL) providers. These are essentially transportation companies, that have some facilities for temporarily storing, sorting and packing the spare parts. Outsourcing part of the distribution makes Toyota’s network more flexible but also increases its complexity. The car manufacturer now faces decisions concerning which 3PL companies it should include in its network and which part of the distribution it should assign to each of the selected 3PL companies. We have developed a software tool to support these decisions. Toyota currently uses the tool to determine the optimal configuration of its network. Moreover, the software has proven useful in increasing Toyota’s negotiating power in its relationship with the transportation companies. This article describes Toyota’s distribution strategy and the development and working of our tool.
Toyota’s spare parts supply chain Figure 1 shows Toyota’s spare parts supply chain. During the day dealers place orders for spare parts. The orders are recorded at the main spare parts distribution center where they are consolidated, picked, and readied for transport. The consolidation stage ensures that different orders for the same spare part going to the same transport platform are grouped so that unnecessary moving about in the distribution center is avoided. All this activity takes place immediately after the deadline for placing orders has passed, around 6pm. After the spare parts have been loaded onto (large) trucks they leave for the various regional distribution centers, which are called transport platforms, and are essentially warehouses with some limited handling capabilities. Some handling equipment and personnel is necessary because the parts are unloaded, unpacked, packed into smaller quantities, and loaded onto smaller trucks, which ultimately deliver the spare parts to the different dealers. The process of setting up Toyota’s distribution channel is a difficult one that involves a lot of negotiating with the different potential transportation companies. When talking to these potential 3PL partners, Toyota’s supply chain managers attempt to assess the characteristics
Kenneth Sörensen and Peter Goos are professors in operational research and statistics at the Faculty of Applied Economics of the Universiteit Antwerpen. Together with Patrick Schittekat, who is a part-time researcher at the university and a consultant at ORTEC Belgium, they work on various challenging combinatorial optimization problem in supply chain management and design of experiments.
of each company in terms of handling quality, reliability and cost. The transport sector is one in which not all companies are equally reliable, to say the least, and filtering out the unrealistic proposals of some distribution companies is therefore of great importance. Prior to using our tool Toyota had experienced difficulties in the process of evaluating each transportation company and deciding which ones to work with. For companies it had already worked with, estimating reliability and handling quality could be done without too much difficulty. For new potential partners, however, Toyota found that assessing these characteristics could only be done within very large margins of error. Moreover, Toyota experienced a lot of problems when trying to determine the relative importance of each 3PL provider in its network, which limited its negotiating power to a large extent. Toyota had no idea of what would happen if it would
AENORM
62
January 2009
31
ORM
ORM
Car dealers order spare parts
Parts are loaded unto trucks
Parts are transported (FTL) to transport platforms
Parts are loaded unto smaller trucks
Parts are delivered to dealers in milk runs
Figure 1: Toyota's distribution process Automotive company
3PL partners
stop collaborating with some of its providers. In general, Toyota had a very limited view on potential alternative distribution networks that it could use.
Providing the Toyota supply chain expert with a set of — say — ten solutions all of which are very much alike would not make much sense, as it would not increase Toyota’s insight into the different alternatives. Generating structurally different solutions, however, gives Toyota a clear view on which 3PL transport platforms should at all cost be included in the transport network and which can be discarded at little or no additional cost. Our tool integrates a commercial vehicle routing solver that is both extremely powerful and versatile. The solver we chose was SHORTREC, developed by the company ORTEC. It allows us to model the vehicle routing problem in a very detailed fashion. Distances, for example, are calculated over an actual street map; different possible truck types, with different cost
"Our decision support system has greatly improved Toyota's negotiating power and responsiveness to changes in its supply chain" Decision support system requirements The decision support methodology we developed to solve these problems therefore had to be innovative in two ways. First, its goal could not just be to find the “best” solution. On the contrary, it had to offer Toyota’s supply chain experts with a lot of insight into the structure of potential alternative networks. With this knowledge about alternative networks, Toyota would suddenly find itself in a much stronger position for negotiating with the different (potential) 3PL partners. Second, to further increase their knowledge about the transportation network, Toyota needed a tool that could provide them with extremely accurate cost information on the different transportation options. It needed this information to ensure that a 3PL provider did not overcharge or (as sometimes happened) undercharge and then get into financial problems. Features of the decision support system The tool we developed does not generate a single “best” solution, but rather a set of structurally different solutions. In this way, a portfolio of transportation networks, all of which have a high quality (i.e. low cost) but which are all different from each other, is available to the decision makers at Toyota. The solutions in the portfolio are structurally different in that they use different sets of transport platforms.
32
AENORM
62
January 2009
structures, can be used; time windows (earliest and latest visiting times) at the dealers can be set; etc. Without going too much in detail, our method solves a so-called location–routing problem (Nagy and Salhi, 2007), that consists of two sub-problems: a facility location problem and a vehicle routing problem. Location problem: the selection of the 3PL providers The location problem is the problem of which transport platforms to select. This is done by using a technique called tabu search (Glover, 1989 and Glover, 1990). Essentially, this technique works by improving the solution one small step at a time. The tabu search method will for example try to close a transport platform or open one. If this small change increases the quality (decreases the cost) of the solution, the new solution is adopted as the current solution which the method then attempts to improve. This process of gradual improvement is called local search and is continued until no better solution can be found or a certain time limit has been reached. The main drawback of local search is that it often gets stuck in a so-called local optimum. A local optimum is a solution that cannot be improved by local search, i.e. no small change exists that improves the solution corresponding to the local optimum. Tabu search is especially designed to avoid getting
Wij bieden je
Dat wil niet zeggen dat je van Mars moet komen Als afgestudeerde wil je graag direct aan de slag. Bij ORTEC
Vanwege onze constante groei is ORTEC altijd op zoek
hoef je hier niet lang op te wachten. Je wordt direct op
naar enthousiaste studenten en afgestudeerden die de
projecten ingezet en krijgt veel eigen verantwoordelijkheid.
ruimte zoeken om zich te ontwikkelen en willen bijdragen
Bij ORTEC werken veel studenten. Sommigen schrijven bij
aan de volgende generatie optimalisatietechnologie.
ons een afstudeerscriptie, anderen werken enkele dagen per week als studentassistent.
Hiervoor denken we aan bèta’s in de studierichtingen: • Econometrie
Maar je staat er nooit alleen voor. Je kunt rekenen op
• Operationele Research
de expertise van je collega’s: stuk voor stuk experts
• Informatica
op het gebied van complexe optimalisatievraagstukken in
• Wiskunde
A0662
diverse logistieke en financiële sectoren. Hoogopgeleide, veelal jonge mensen die weten wat ze doen en jou naar een
Kijk voor vacatures en afstudeerplaatsen eens op
hoger niveau zullen brengen. Samen met je collega’s help je
www.ortec.com/atwork. Zit jouw ideale functie of afstu-
klanten gefundeerde beslissingen te nemen. Dit doe je met
deerplek er niet bij, stuur dan een open sollicitatie of
gebruik van wiskundige modellen en het toepassen van
scriptievoorstel naar recruitment@ortec.com.
simulatie- en optimalisatietechnieken.
EPROFESSIONALS IN PLANNING AENORM 62 January 2009 33
ORM
trapped in such a local optimum. In contrast with simple local search, this technique allows non-improving changes and uses memory structures to guide the (local) search out of a local optimum. One of those memory structures, the one that lends its name to the method, is called a tabu list. This list prohibits us from undoing the last few changes that were made, thereby preventing the search from ending up in an endless repetition of non-improving changes. The location decision, i.e. which transport platforms to use, determines the starting point for the vehicle routing problem. When a selection of transport platforms has been made, solving a (complicated) vehicle routing problem will tell us what the cost of using this specific set of 3PL providers is. These costs can be determined by calculating how much it would cost to deliver to all dealers using this specific set of transport platforms. Vehicle routing problem Vehicle routing is the branch of science that studies models and methods to determine the optimal schedule for a fleet of vehicles to visit a set of delivery points (Toth and Vigo, 2001). The simplest in this class of problems is the so-called capacitated vehicle routing problem or simply VRP. In this problem, a number of delivery points (called customers) have to be visited to deliver a homogeneous product. As an example, consider a set of trucks delivering petrol to people’s houses for their heating systems. Each customer has a certain demand (e.g. 3000 liters) and all vehicles have the same capacity e.g. 20.000 liters). The objective of this problem is to find the optimal schedule to visit all customers. This means we have to determine (1) which customers have to be visited by each of the available vehicles and (2)
(a)
Key Customer Depot
Figure 2: A vehicle routing problem and a possible solution
in which order each of the customers has to be visited, so that the total distance traveled by all vehicles is minimized (see figure 2). The VRP is surprisingly difficult to solve: if we insist on finding the best possible (or optimal) solution we can now – with present-day algorithms and computer power – reliably solve problems that involve about 150 customers, but this might take several hours of computing time on an average computer. If we want to solve larger problems our do it faster, we have to content ourselves with solutions that approximate the optimal solution. In real life, problems with more than 150 stops abound, of course, so in practice the usefulness of algorithms that try to find the optimal solution is limited. The problem that needs to be solved by our method is not only much larger than 150 customers, but also much more complex. It includes different starting and ending locations (the transport platforms), time windows at the transport platforms and the dealers, restrictions on how long a truck driver can drive and when he needs to take a break, and restrictions on which trucks can visit which companies (some
Figure 3: The best solution found by our method
34
AENORM
62
January 2009
(b)
ORM
are located in city centers and cannot be visited with huge trucks). The vehicle routing software that we use is therefore heuristic in nature, i.e. it does not guarantee that it can find the optimal solution, but tries to produce a satisfactory solution in reasonable time. Combined solution to location and routing problem Combining the methods for the location and the routing subproblems, our method generates a set of solutions that are structurally different from each other. Each of the generated solutions corresponds to a transport network that has a low cost. Toyota’s supply chain experts can then analyze each of the proposed networks in detail. One such solution (for the distribution of spare parts in Germany) is shown in figure 3. The results of our method have exceeded all expectations. The supply chain planners at Toyota have welcomed our tool and the insight it gives into the working of their supply chain with open arms. Each month, the tool is used to completely re-evaluate the entire spare parts supply chain and — if necessary — adjust it. The tool therefore provides a high level roadmap on which decisions can be based. Toyota found that its negotiating power and its responsiveness to all sorts of changes in the supply chain context have greatly improved. Acknowledgment The research described in this article was financially supported by the Research Foundation - Flanders (Fonds voor Wetenschappelijk Onderzoek - Vlaanderen). References Glover, F. (1989). Tabu Search–Part I, INFORMS Journal on Computing, 1(3), 190. Glover, F. (1990). Tabu Search-Part II, ORSA Journal on Computing, 2(1), 4–32. Nagy, G. and Salhi S. (2007). Location-routing: Issues, models and methods, European Journal of Operational Research, 177(2), 649–672. Toth, P. and Vigo D., editors (2001). The vehicle routing problem, SIAM Monographs on Discrete Mathematics and Applications. Society for Industrial and Applied Mathematics, Philadelphia, PA, USA.
AENORM
62
January 2009
35
Wat doe je? als je ambities als actuaris verder reiken dan de Nederlandse grenzen Achmea is in Nederland de grootste actuariële werkgever. Maar ons werkterrein beperkt zich niet tot Nederland. Ook over de grenzen heen zijn we actief. Met ons driejarig internationaal actuarieel traineeship leiden we actuarissen op om op internationaal niveau te werken. Na dit intensieve programma kun je als actuarieel professional aan de slag bij één van de Eureko onderdelen in Europa. Heb jij de ambitie om jezelf zowel inhoudelijk als persoonlijk te ontwikkelen in een internationale omgeving? Dan maken wij graag kennis met jou.
Internationaal actuarieel traineeship Tijdens het traineeship volg je op inhoudelijk vlak, naast een training on the job, een actuariële opleiding. Je start met een introductieperiode van een half jaar op één van onze actuariële afdelingen in Nederland. Gedurende deze periode leer je het actuariële vak beter kennen en bepaal je zelf je specialisatie. Tevens leer je zowel de internationale organisatie Eureko als de nationale organisatie Achmea beter kennen. Na deze periode vertrek je voor tweeënhalf jaar naar één van de Eureko onderdelen in Athene (Interamerican) of Dublin (Friends First), afhankelijk van jouw specialisatie. Na dit programma wordt je ingezet op kort- of langdurende opdrachten binnen Europa, afhankelijk van zowel jouw persoonlijke voorkeur als de vraag van de organisatie.
Wat we vragen Als internationaal actuarieel trainee beschik je over een afgeronde universitaire opleiding, bij voorkeur actuariële wetenschappen, econometrie of wiskunde. Je hebt maximaal twee jaar werkervaring bij een financiële dienstverlener. Daarnaast herken je jezelf in de volgende competenties: zeer analytisch, leergierig, zelfstandig handelend, bereid om internationaal te werken en communicatief zeer vaardig in de Engelse taal
Wat wij bieden Een unieke kans om jezelf in korte tijd snel te ontwikkelen in het actuariële vakgebied in een internationale omgeving. Vervolgens kun je jouw ambities waarmaken met de vele mogelijkheden die Eureko / Achmea te bieden heeft.
Achmea Achmea maakt deel uit van Eureko; een financiële dienstverlener met flinke ambities en ondernemingen in verschillende Europese landen. Zowel Eureko als Achmea hebben tot doel het creëren van waarde voor al onze stakeholders: klanten, distributiepartners, aandeelhouders en medewerkers. Om dat waar te kunnen maken hebben we medewerkers nodig die verder kijken dan hun eigen bureau. Die oog hebben voor wat er speelt. Maar vooral: mensen die zich inleven in onze klanten en dat weten te vertalen naar AV É R O A C H M E A
originele oplossingen.
CENTRA AL BEHEER ACHMEA
Meer weten?
FBTO
Voor meer informatie over het actuarieel traineeship kun je contact opnemen met Joan van Breukelen,
INTERPOLIS
recruiter, (06) 20 95 7231. We ontvangen je sollicitatie graag via www.werkenbijachmea.nl.
Z I LV E R E N K R U I S A C H M E A
Ontzorgen is een werkwoord 36
AENORM
62
January 2009
Econometrics
The Maximal Damage Paradigm in Antitrust Regulation: Is it Something New? It is well known that cartels are harmful for consumers. To counteract cartels, cartel formation is by law an economic crime with the antitrust authority (AA) as its crime fighter. Recently, Harrington (2004, 2005) studied a general model of cartel formation and its pricing based upon profit maximization. In this article, we discuss the novel approach in Houba et al. (2009), who take the maximal damage for consumers as the key criterion. Some developments of this approach are introduced and related to the literature.
Introduction Despite a large literature on enforcement against individual illegal behavior, the theory of regulation is still in its infancy when it comes to enforcing market competition. Illegal anti-competitive behavior is much more complicated since it typically is a concerted illegal action performed within an ongoing relationship over time, called a cartel. Any theory of regulation therefore requires a dynamic setting, for example an infinitely-repeated oligopoly model with grim trigger strategies. In this article, we discuss an innovative but unconventional approach in which we study the maximal-sustainable cartel price in a repeated oligopoly model, i.e., the largest cartel price for which the equilibrium conditions for sustainability hold. This differs from the standard approach in which the cartel maximizes profits. The main reason for doing so is that experimental economics establishes that economic agents often behave differently from standard microeconomic theory. Also, there is empirical evidence in support of Baumol’s (1958) hypothesis that managers of large corporations seek to maximize sales rather than profits. Sustainability of cartel behavior offers a more robust criterion that does not depend on the cartel’s objective. Then, the characterization of consumers’ maximal damage can be regarded as a worst-case scenario for consumers. It is therefore natural to apply this new approach to regulation and compare the main results with those obtained for a profit-maximizing cartel in Harrington (2004, 2005). This article is organized as follows. Section 2 introduces the maximal-sustainable cartel price in a benchmark model without regulation. In section 3 we analyze the impact of AA enforcement
Harold Houba is associate professor at the Department of Econometrics of the VU University. He obtained his Ph.D. from Tilburg University in 1994. Houba is affiliated with the VU since 1992. His specialization is bargaining theory with applications to labor and environmental economics. Evgenia Motchenkova is assistant professor at the Department of Economics of the VU University. She obtained her Ph.D. from Tilburg University in 2005. Motchenkova is affiliated with the VU since 2005. Her specialization is the economics of antitrust regulation, in particular leniency programs. Quan Wen is professor at the Department of Economics of the Vanderbilt University in Nashville. He obtained his Ph.D. from the University of Western Ontario in London, in 1991. Wen is affiliated with Vanderbilt since 2001. His specializations are bargaining theory, repeated games and their applications to economics.
on the maximal-sustainable cartel price. In section 4, we compare, by means of an example, our approach to the one in Harrington (2004, 2005). Section 5 concludes the analysis. The Benchmark Model Consider an oligopoly market where n≥2 symmetric firms compete in prices with either homogenous or heterogeneous products over infinitely many periods. All firms have a common discount factor δ є (0,1) per period. Since we deal with symmetric outcomes, we simplify (p,...,p) є \ n+ to p є \ + . We adopt the following notation:
AENORM
62
January 2009
37
Econometrics
- pN and pM denote the competitive (Nash) equilibrium price, respectively, the maximal collusive price. - π(p) is the profit function of an individual firm in any period. π(p) is continuous and strictly increasing for p є [pN,pM]. - πopt(p) is a firm’s profit from unilateral deviation against the cartel when all the other cartel members set their prices at p. πopt(p) is continuous and πopt(p)>π(p)>0 for p є [pN,pM]. - λ(p) is the degree of cartel stability and it is defined as opt N M ⎪⎧π(p) / π (p), for p ∈ (p , p ], λ( p) = ⎨ N 1, for p = p . ⎪⎩ λ(p)<λ(pN) ≡ 1 for all p є (pN,pM] is decreasing N for all p є (p ,pM]. The degree of cartel stability is a new concept. Standard intuition implies that a higher λ(p) means less incentives for cartel members to deviate and a stabler cartel. Furthermore, a higher cartel price implies a higher incentive for each cartel member to deviate. Since the function λ might be discontinuous at p=pN, as Example 2 illustrates, we introduce = λ lim+ λ(pN + ε ) ≤ 1 ε →0
and= λ λ( pM ).
The above oligopoly model without regulation is a standard infinitely-repeated game. Throughout this article, we focus on grimtrigger strategies to sustain cartel price p>pN in which any deviation leads to the repetition of the competitive (Nash) price in every period thereafter. The underlying rationale is that cartels are based upon trust and, by the reciprocal nature of humans, all trust is gone after someone cheats. The equilibrium concept is a subgame perfect equilibrium. In the absence of regulation, the necessary and sufficient condition to sustain p є (pN,pM] as a cartel price is δ 1 πopt (pN ) ≤ π( p) 1−δ 1−δ ⇔ δ ≥ 1 − λ(p).
πopt ( p) +
(1)
The socially worst outcome is the maximalsustainable cartel price that is defined pC = max p, N M p∈[ p , p ]
s.t.
(1).
(2)
Due to the monotonicity of π(p)/(1–δ), the maximal-sustainable cartel price pC also maximizes the cartel’s profit. A direct approach would solve (1) for p as a function of all parameters, which requires the inverse function λ-1(1–δ). Later, however, this approach is not applicable. Instead, we analyze the properties of the threshold level for δ as a function of p є (pN,pM] and, then, translate these properties into the maximal-sustainable
38
AENORM
62
January 2009
cartel price as a function of δ in the (δ,p)-plane. This indirect approach allows for an easier interpretation. Proposition 1 In the absence of regulation, the maximalsustainable cartel price pC is non-decreasing in δ є (0,1), and p C = p N, for δ є (0, 1 – λ ) , pC є [pN, pM), for δ є [1 – λ ; 1 – λ ), pC = pM, for δ є [1 – λ , 1), We conclude this section with a well-known example. Example 2 Consider a homogeneous Bertrand oligopoly model with linear demand 2–p and 0 marginal costs. Note that pN=0 and pM=1. For all p є (pN,pM], each of the n firms may deviate by slightly undercutting the others to obtain the full cartel profit, i.e., λ(p)=1 for all p є (pN,pM]. Consequently, λ = λ =1/n. Proposition 1 implies N p pC = M p
δ < − n δ ≥ − n
Antitrust Enforcement In this section, we examine the impact of regulation. Given p є [pN,pM], the probability that the AA investigates the market outcome in a period and finds the firms guilty of collusion is β(p) є [0,1), where β(p) is increasing in p and β(pN)=0. Upon being caught, violators will be fined by the amount k(p)π(p), where k(p) is increasing and continuous such that k(pN)=0 and k(p)>0 for all p є (pN,pM]. The function β(∙) reflects that a higher cartel price attracts suspicions and makes detection more likely. Any cartel takes this negative effect into account when deciding upon the price, see Harrington (2004, 2005). The AA is a passive player in this model, while firms are the active players. The detection probability β(∙) is limited by the resources of the authority, and the fine schedule k(∙) is limited by legislation. The OECD (2002) reports detection probabilities 1/7≤β(p)≤1/6 and penalty schemes 2≤k(p)≤3. These facts imply 2/7≤β(p) k(p)≤1/2, or an expected penalty roughly between 30% to 50% of the illegal cartel profits. Therefore, the AA may not be able to deter violations. Here, we assume 0<β(p)k(p)<1 for all p є (pN,pM] meaning any cartel is tempted to set its price above the competitive price. Another aspect is how cartel members react to detection. In some cases, being caught once is sufficient to deter cartel activity in the future. In other cases, the economic sector is notorious for cartel activities despite many convictions (meaning members pay the fines and continue
Econometrics
illegal business). γ є [0,1] is the cartel culture parameter that reflects the probability that the firms will behave competitively (i.e. stop illegal business) after each conviction. Notorious implies γ=0, while γ=1 means the sector becomes competitive after the first detection. All models in the literature assume either γ=0 or γ=1. Let V(p) be the present value of a cartel member’s expected profit if the cartel sets price p є [pN,pM] in every period. This value consists of the current illegal gains π(p), the expected fine β(p)k(p)π(p), the expected continuation payoff
where superscript R refers to regulation. Program (5) is a well-defined program since p є [pN,pM], which can be deducted as follows. Since the monotonicity properties of Λ(p) (increasing) and λ(p) (decreasing) are opposite, the intersection λ(p)=Λ (p) is unique and coincides with pR. Furthermore, any p≤pR can also be sustained by the cartel. So, the range of prices in (4) is a closed subinterval of [pN,pM]. Comparing (2) and (5), we observe that pN≤pR ≤pC≤pM meaning that regulation may reduce the maximal-sustainable cartel price in general. Similar to Proposition 1, we derive the
"Current regulation is ineffective" of a renewed cartel after detection β(p)(1–γ) δV(p), and the expected continuation payoff of not being detected (1–β(p))δV(p), from which we obtain V (p)
1 − β(p)k(p) π(p) π(p) < , 1 − δ[1 − γβ(p)] 1−δ
(3)
p ∈ (pN , pM ].
So, introduction of regulation reduces the cartel’s profitability. Does it also affect sustainability? The cartel has its own destabilizing forces working from within, because individual cartel members have an incentive to cheat on the cartel. Here, cartel members adopt modified grimtrigger strategies to sustain p>pN: 1 Firms continue to set a price p>0 with probability 1–γ after each conviction (and with probability γ set the competitive price pN ever after). 2 Any deviation by some cartel members leads to the competitive price pN in every period ever after. Then, the profit from a unilateral deviation is equal to the short term gain of πopt(p) in the current period, followed by the competitive equilibrium with π(pN)=0 forever after. Consequently, the necessary and sufficient condition to sustain cartel price p є (pN,pM] is V(p)≥πopt(p), or λ( p) ≥ Λ( p) ≡
1 − δ[1 − γβ( p)] 1 − β( p)k( p)
(4)
Under regulation the maximal-sustainable cartel price is given by pR = max p, N M p∈[ p , p ]
s.t.
(4).
(5)
thresholds on the discount factor δ. Doing so, (4) can be rewritten as δ ≥ Δ(p) ≡
1 − λ(p)[1 − β(p)k(p)] ≥ 1 − λ(p) (6) [1 − γβ(p)]
The function Δ(p) is continuous and increasing in p, k(p) and β(p). Proposition 3 Under regulation, the maximal-sustainable cartel price pR is non-decreasing in δ є (0,1) and decreasing in γ є [0,1]. Furthermore, we have p R = p N, pR є [pN, pR = pM,
),
for δ є (0, 1 – λ ) , for δ є [1 – λ , Δ(pM)), for δ є [Δ(pM), 1),
An overall increase in β(p) or k(p) shifts Δ(pM) and the entire curve to the right. Clearly, inequality (6) is more restrictive than (1) implying that introduction of regulation restricts the set of discount factors for which collusion can be sustained for every possible price p є (pN, ]. This implies that cartel stability is reduced compared to the benchmark case. Moreover, the fact that Δ(p) is increasing in p implies that the regulation is more effective against collusion on higher prices. When Δ(pM)≤δ<1, the regulation is not effective to deter the cartel from setting its monopoly price. It is interesting to investigate whether regulation can eradicate the monopoly price for all cartel cultures. Solving Δ(pM)<1 for γ yields γ < {λ(pM)[1 – k(pM)β(pM)]}/β(pM), where the right-hand side remains positive under 0<β(p)k(p)<1. Hence, industries that are notorious for cartel behavior cannot be eradica-
AENORM
62
January 2009
39
Econometrics
ted by regulation unless one is willing to adopt k(p)=k, where k<1. Then, (5) becomes regulation that fully takes away the illegal gains 1 1 − δ + γδβp (i.e. β(p)k(p)>1 for all p є (pN, ]). = p(δ, γ ) max p, s.t. ≥ . Another implication is related to the effect of p∈[0,1] n 1 − kβρ the degree of cartel stability λ(p) on sustainaNote that p=0 is feasible in the quadratic consbility of the monopoly price. Since ∂Δ/∂λ<0, traint if and only if δ≥1–1/n. The constraint can sectors where the degree of cartel stability is be rewritten as higher (λ(p) closer to 1) have less restrictive p ≤ [1–n(1–δ)]/[(nγδ+k)β], conditions for sustaining consumers´ worst priwhich is the solution to the problem if it is betce pM. This makes regulation less effective in ween 0 and 1. The right hand side is increasing these sectors. in δ. To summarize, we have The main message is a mixed blessing for regulation. On the one hand, proposition 3 identifies δ < − n non-empty sets of parameter values for which pR = regulation is effective in reducing the maximal − n − δ nγδ + k β − n ≤ δ < sustainable cartel price, i.e., pR<pC. On the other hand, as long as regulation obeys β(p) k(p)<1, there will remain a large non-empty Note that pR<pM=1 for all δ є (0,1) if and only if set of parameter values for which pR=pM, mea(nγ+k)β>1. Since βk<1, this condition can hold ning the regulation is ineffective. only when nγ is sufficiently large. For sectors We conclude this section with an example. with small numbers of firms and γ sufficienExample 4 Reconsider Example 2 and let β(p)=βp and
tly close to 0, the monopoly price will not be eradicated by regulation. Both possible cases, (nγ+k)β≤1, respectively, (nγ+k)β>1, are illustrated by figure 1, where the vertical dotted line at δ=1–1/n represents the discontinuous jump in pC of Example 2 from pN=0 to pM=1.
The profit-maximizing cartel price In this section, we compare our approach to Harrington (2004, 2005). He defines the endogenous cartel price as the profit-maximizing sustainable cartel price pπ: pπ = arg max V ( p), N M p∈[ p , p ]
s.t.
(4).
(7)
For explanatory reasons, we restrict attention to numerical values β(p)=p/2 and k(p)=3 and γ=2/3 in Example 4. Then, V p =
− p − pp ⋅ − δ + p n
Standard arguments imply V(p) fails both monotonicity and concavity on [pN,pM]=[0,1], but this function is single peaked on [pN,pM]. Checking the second-order conditions can be avoided, because V(p) is monotonically increasing from pN to its peak and monotonically decreasing from its peak to pM. So, application of the first-order conditions suffices. In contrast, (5) is a well-defined convex program. Harrington shows that, in general, (4) is nonbinding for δ sufficiently close to 1 and we may solve the first-order condition ∂V(p)/∂p=0. In our case, MAPLE returns ≡ − − δ + δ − δ + ∈ P δ
Figure 1
40
AENORM
62
January 2009
Econometrics
the profit-maximizing cartel price underestimates the maximal damage to consumers; a bias that might be huge. Since economic agents often behave differently from standard microeconomic theory, the criterion of sustainability of cartel behavior offers a more robust framework that does not depend on the cartel´s objective function. In this perspective, our approach is a worst-case scenario. References Baumol, W. (1958). On the theory of oligopoly, Economica, 25, 187-198.
Figure 2
for all δ є [0,1]. Taking also (4) into account implies that the profit-maximizing cartel price and pR. These two price pπ is the minimum of curves intersect at δ≈0.955. So, on the interval [0,0.955], the profit-maximizing cartel price pπ=pR, while for the interval (0.955,1] we have <pR and the equilibrium condition is nonpπ = ,pR}. binding. Figure 2 illustrates pπ =min{ This example offers important insights. Whenever constraint (4) in (7) is binding, the maximal-sustainable cartel price pR and the profit-maximizing cartel price pπ coincide and our approach is complementary to the analysis in Harrington (2004,2005). Otherwise, i.e., (4) in (7) is nonbinding, these two cartel prices systematically differ. As the figure shows, for δ close to 1, the profit-maximizing cartel price pπ is close to the competitive price pN and, therefore, seriously underestimates the potential maximal damage to consumers. Our approach offers a worst case scenario of sustainable cartel behavior.
Harrington, J. (2004). Cartel pricing dynamics in the presence of an antitrust authority, The Rand Journal of Economics, 35, 651-673. Harrington, J. (2005). Optimal cartel pricing in the presence of an antitrust authority, International Economic Review, 46, 145170. Houba, H., Motchenkova, E., and Wen, Q. (2009). Leniency programs in antitrust regulation with an endogenous detection probability, TI discussion paper, (in preparation). OECD (2002). Fighting hard-core cartels: Harm, effective sanctions and leniency programs. OECD Report 2002, OECD, Paris, France, http:// www.SourceOECD.org.
Conclusion In this article, we explore a general infinitelyrepeated oligopoly model for the analysis of violations of competition law under regulation. A novel concept is the maximal-sustainable cartel price that reflects consumers´ worst cartel price among those cartel prices that are sustainable, which endogenizes the cartel formation decision and its pricing strategy. This cartel price is related to the discount rate and the novel concepts of type and structure of the industry (λ) and cartel culture parameter (γ). Regulation is less effective in sectors where the degree of cartel stability is higher or where the sector´s cartel culture to continue business as usual is more prominent. Stylized facts from OECD countries imply that current regulation is ineffective. Finally, our approach is complementary to Harrington (2004, 2005) in case equilibrium conditions are binding. Otherwise,
AENORM
62
January 2009
41
Actuarial Sciences
Dutch Citizens Ready for Eligibility Age Increase from 65 to 67 Years A majority of the Dutch population expects that the government will increase the eligibility age for old age social security (AOW) some time in the coming decades from 65 to 67 years. If this is the case, the question rises whether the Dutch are also preparing for such an anticipated increase. To some extent this appears to be the case: individuals, who are relatively certain that the eligibility age will be increased, purchase private pension products more often.
Karen van der Wiel is currently a PhD student in the department of econometrics at Tilburg University, The Netherlands. She is associated to Netspar, a network that does research on pensions, aging and retirement and through this network the Dutch government institution that pays out old age social security benefits (SVB) finances her research. Karen received an Msc in economics from Erasmus University Rotterdam in 2005 (cum laude). In 2008, Karen spent a semester at University College London. Her research interests include applied microeconometrics, labor economics and experimental economics.
Measuring expectations Economists assume that future expectations play an important role in almost all long-term decisions. In the absence of reliable data, it was difficult to test this notion for a long time. However, in recent years it turned out that expectations can indeed be measured by asking respondents for the probability between 0 and 100 that a certain event will occur (Manski, 2004). An example of such a probability-question is asking individuals for the chance they will survive up to the age of 80 years. U.S. data shows that there is a remarkable correlation between the given answers and the actual survival rates of older individuals (Hurd and McGarry, 2002). Expectations can thus be measured using probability-questions. AOW expectations In May 2006, CentERdata in Tilburg therefore began measuring the monthly pension expectations of Dutch individuals in the so-called Pensionbarometer. Netspar, the `Network for Studies on Pensions, Aging and Retirementâ&#x20AC;&#x2122;, pays for this research in order to gain more insight into the confidence of the Dutch popula-
42
AENORM
62
January 2009
tion in the various elements of the Dutch pension system. Questions concerning the future of the old age social security system play a central role in the Pensionbarometer. Because of the political relevance, I will focus on the future expectations about the earliest possible age at which a pensioner may receive benefits in this article. It turns out that the average probability that people assign to an increase in the eligibility age from 65 to at least 67 years, is 51 (or 0.51). This average varies as the relevant time horizon differs. As one would expect, people believe that the probability of this policy change to be occurring sometime in the next twenty years (average probability is 54) is higher than it to be occurring sometime in the next ten years (average probability is 47). Figure 1 shows the development of the average expectation over the past two years (from May 2006 to October 2008). Pensionbarometer respondents were most optimistic about the future of the AOW eligibility at the end of 2006 (in the period of the national general elections) and were most pessimistic in June 2008 (around the publication of the report of the Bakker Commission). Savings behavior and AOW expectations Within the life-cycle/savings theory it is assumed that saving for old age is an important saving motive (Browning and Lusardi, 1996). The hypothesis is that most people no longer want and / or are able to work after reaching a certain age and that they will provide themselves with accumulated financial resources from that moment onwards. In the Netherlands, saving for old age is organized on three levels: 1) the government guarantees an AOW benefit for everyone; 2) most employers take care of an additional pension for their employees and 3)
Actuarial Sciences
from the Pensionbarometer with data from the DNB Household Savings Survey (DHS). This has lead to a dataset containing information on over 3,000 individuals. The dataset includes individualsâ&#x20AC;&#x2122; view on the future of the old age pension age and their saving behavior. Table 1 presents the crude percentages of respondents who own a private pension product (an annuity and/ or single premium policy). 39% of the total sample held such a third pillar pension product. This figure is higher than that of the entire Dutch population, since richer households are slightly over-represented in the DHS. In Table 1, all respondents of the DNB Household Savings Survey are divided into three groups: in the first column one finds individuals that are not convinced that the AOW-eligibility age will go up within 20 years (who provide a probability between 0 and 0.4), in the second column those that assign an average probability (between 0.4 and 0.6) and in the third column one finds individuals who are relatively pessimistic (who provide a probability between 0.6 and 1). The first row of Table 1 indeed shows that pessimistic respondents purchase more annuities and/ or single premium policies. The most pessimistic group is 11% more likely to have a private pension product. These results are the same for most of the subgroups. This implies the Dutch who anticipate a policy change do seem to prepare for it. This is not the case for every subgroup. Relatively young workers (under the age of 40) do not take their own expectations about the AOW into account when preparing for their retirement. The differences in participation rates for wealthier workers (with a gross income above 38,100 Euro) are also not significant between the expectations groups.
Probability that AOW-eligibility age will increase to 67 years
40
45
Probability 50
55
60
Average per month
0
50
Week from 1-1-2006
Within ten years
100
150
Within twenty years
Figure 1
each individual can also save for his or her own old age. Research from the seventies has demonstrated substitution effects between such public and individual savings systems. In the United States people saved less themselves after the government had installed old age social security (Feldstein, 1974). In recent years, the ageing of Western populations has forced many governments to discuss the future affordability of Pay-As-You-Go pension schemes. In some countries, such as Britain and Germany, this has already led to an increase in the social security eligibility age. The debate on how to keep the AOW sustainable is also ongoing in the Netherlands. Now suppose that the government would indeed raise the AOW eligibility age from 65 to 67 years. Given the interaction between public and private savings, it is evident that the Dutch should compensate the increase by saving more themselves. In reality however, it is still uncertain what will happen to the old age pension age in coming years, but the Dutch population does have certain expectations about its future. When future expectations are indeed important for longterm decisions, one would expect people who are relatively more pessimistic about the future of the AOW also save more for their own retirement by, for example, purchasing annuities and single premium insurance policies. If you expect to receive nothing from the state between your 65th and 67th birthday, and you do want to quit working before you reach the age of 67, youâ&#x20AC;&#x2122;ll need to save. For a policymaker it is particularly important to know whether people can deal with potential policy changes such as a higher AOW-eligibility age. If many families are unable to react properly to an anticipated increase in the old age pension age, it does not seem wise to decide to imlement such a change.
Regression Results Little can be concluded about the causality of the relationship between the AOW future expectations and savings behavior from the results in Table 1. If, for example, those with low educational attainment are all very optimistic about the future of the AOW-eligibility age and also coincidentally do not buy pension products, education is the cause of low pension scheme participation, not confidence in AOW. To (to a certain extent) rule out such confounding factors, I estimate a probit regression on the possession of a private pension product on the probability individuals assign to a two year higher eligibility age and on all other relevant personal characteristics (education level, age, a self-reported health status, having children, having a partner, sex and year of the sample). The results of this regression are shown in Table 2. This table displays the marginal effects of the subjective probability of an eligibility age increase on the probability of owning an annuity and/ or single premium policy. The marginal effect
Actual expectations and pension scheme participation CentERdata in Tilburg has been executing savings research commissioned by the Dutch Central Bank (DNB) for a long time. For the empirical analysis of this article I combined data
Perc. policy total
Prob. 0-0,4
Prob. 0,4-0,6
Prob. 0,6-1
Observations
32.85
38.75
43.02
3,228
Total observations
904
978
1,346
Perc. policy 30-39 years
33.48
30.16
32.84
Perc. policy 40-49 years
24,17
35.52
45.9262 AENORM
931 1,0972009 January
43
Actuarial Sciences
is 19% for the entire sample. This means that a person with average characteristics (average level of education, average age, average income, etc.) who is convinced that the AOW-age will go up is 19 percent more likely to own a private pension product than a similar person who is convinced that the AOW-age will remain unchanged. Table 2 shows that this marginal effect is quite high for all subgroups, except for persons under 40 and for people with a high income, like in Table 1. For the group between 40 and 49 years, the difference in pension product ownership between people who assign a probability of one and a probability of zero is even 42%.
Validity of Subjective Probabilities or Survival, The Economic Journal, 112(482), 966-985. Feldstein (1974). Social Security, Retirement, and Aggregate Accumulation, The Journal of Economy, 82(5), 905-926.
Induced Capital Political
Manski (2004). Measuring Expectations, Econometrica, 72(5), 1329-1376. Van der Wiel (2008), Preparing for policy changes: Social Security Expectations and Pension Scheme Participation, IZA Working Paper 3623.
Conclusion The Dutch population believes there is a good chance that within the next few years the earliest possible age to receive an AOW benefit, now 65 years, will be increased by the government. Whether this will actually happen remains to be seen. My research does show that those who are more convinced of a higher eligibility-age in the future, are preparing for such a policy change. Those who assign higher probabilities to an eligibility age of 67 years own more third pillar pension products, such as annuities and/ or single premium policies. An individual who is certain that the old age pension age increases, is 19% more likely to hold a third pillar product than a similar person who is convinced the AOW age will remain 65 years. Only the Dutch under the age of 40 and those with a gross annual income that is higher than 38,000 Euro do not pay so much attention to their own AOW future expectations. Perhaps one should not expect these groups are able to do so as retirement is still too far away for those under 40 and the AOW benefit is a minor part of the pension of the richer Dutchman. References Browning and Lusardi (1996). Household saving: Micro theories and micro facts, Journal of Economic Literature, 34(4), 1797-1855. Hurd and McGarry (2002). The Predictive
Prob. that AOW-eligibility age will be (at least) 67 within twenty years
Sample
Marginal effects
Standard errors
Total sample
19%**
0.06
30-39 years
4%
0.10
40-49 years
42% ***
0.11
50-59 years
16%
0.10
Income <25,150 Euro
21% **
0.08
Income 25,150 â&#x20AC;&#x201C; 38,100 Euro
30% **
0.10
Income > 38,100 Euro
5%
0.11
Table 2: Marginal effects regression analysis
44
AENORM
62
January 2009
Dubiël, trainee bij APG Asset Management Stephanie
‘Stel je voor… als trainee doe je straks zaken met New York en Hong Kong’
25.1517.08
‘APG Asset Management is een van de grootste pensioenvermogensbeheerders ter wereld. We beleggen zo’n 200 miljard euro aan vermogen. Als trainee krijg ik hier direct veel verantwoordelijkheid en alle kansen om mezelf te ontwikkelen. Ik draai volwaardig mee en heb dagelijks contact met collega’s in binnen- en buitenland. Ook krijg ik hier ‘training on the job’ en maak ik kennis met verschillende onderdelen van de organisatie. Zo kan ik ontdekken welk carrièrepad het beste bij mij past.’
APG Asset Management start twee keer per jaar een traineeprogramma. Daarvoor zoeken we toptrainees. APG beheert het pensioenvermogen van 2,6 miljoen Nederlanders. Wij hebben kantoren in Heerlen, Amsterdam, Hong Kong en New York. Wil je meer weten over het traineeship bij APG Asset Management? Ga dan naar www.investinyourcareer.nl en stel je voor…
AENORM
62
January 2009
45
Econometrics
Optimal Diversity in Investments with Recombinant Innovation When making decisions on investments in technological innovation, implicitly or explicitly choices are made about diversity of options, strategies or technologies. Such choices should ideally consider the benefits and costs of diversity and arrive at an optimal trade-off. One important benefit of diversity relates to the nature of innovation, which often results from combining two or more different and existing technologies or knowledge bases (Fleming, 2001; Ethiraj and Levinthal, 2004). For instance, a laptop computer in essence is a combination of a desktop computer and a battery; a laser is quantum mechanics integrated into an optical device, while an optical fibre for telecommunication is a laser applied to glass technology.
•
Paolo Zeppini is currently doing his PhD in economics at the University of Amsterdam, which he started after completing the two years master program (MPhil) in economics of Tinbergen Institute. Paolo graduated in Physics at the University of Florence, worked for two years in consultancy, did a master in quantitative finance at Bocconi University in Milano and worked for three years as a trader in an investment bank before coming to Amsterdam in 2006 to do research in economics.
This article presents a theoretical framework for the analysis of an investment decision problem where an innovative process is at work resulting from the interaction of two technologies. Such a process is called recombinant or modular innovation. The main idea is that in an investment decision problem where available options may recombine and give birth to an innovative option (technology), a certain degree of diversity of parent options can lead to higher benefits than specialization. The conceptual framework and the detailed theoretical analysis can be found in van den Bergh and Zeppini-Rossi (2008). A pilot model Consider a system of two investment options that can be combined to produce a third one. Let I denote cumulative investment in parent options. Investment I3 in the new option only occurs if it emerges, which happens with probability PE. The growth rates of parent options are proportional to investments, with shares α and 1−α. Let O1 and O2 represent the values of the cumulative investment in parent options and O3 the (expected) cumulative investment in the innovative option. The dynamics of the system is described by
46
AENORM
62
January 2009
= I1 = αI • 1 = (1) 2 I2 = (1−α)I • 3 = PE(O1,O2)I3 optimization problem is to find α that maxiThe mizes the final total benefits of parent and innovative options. The probability of emergence depends on two factors; the diversity of the parents and a scaling factor π which is the efficiency of the R&D process underlying recombinant innovation: PE(O1,O2) = πB(O1,O2)
(2)
Diversity is expressed as the balance B of parent options: B(O1,O2) = 4(O1O2)/(O1 + O2)2 Assuming that investment in parent options begins at time t=0, their value at time t is O1(t)=αIt and O2(t)=(1−α)It. Under this assumption the balance function is independent of time: B=4α(1−α). The probability of emergence is constant and only depends on α. The innovative option grows linearly: O3(t) = 4πI3α(1−α)t
(3)
The optimization problem of this investment decision is addressed considering the joint benefits of parents and innovative options: V(α;T) = O1(T;α)s + O2(T;α)s + O3(T;α)s
(4)
The returns to scale parameter s allows to model the trade-off between diversity and scale advantages of specialization. t=T is the time horizon. Normalizing benefits to its value in case of specialization V(α=0;T)=V(α=1;T)=IsTs one obtains
Econometrics
Figure 2: Benefits as a function of the investment share α under increasing returns to scale (s=0.5 for different values of innovation effectiveness π=0,0.2,0.4,0.6,0.8,1. Here I= 4I3.
Figure 1: Benefits as a function of the investment share α under increasing returns to scale (s=1.2) for different values of innovation effectiveness π=0,0.2,0.4,0.6,0.8,1. Here I= 4I3.
V α T = α s + − αs + C s α s − αs (5) I sT s where C=4πI3/I. Figure 1 reports the normalized benefits curves in a case of increasing returns to scale (s=1.2) for six different values of the factor π. There is a threshold value π such that for π<π the optimal decision is spe cialization, while for π>π diversity is optimal. Conversely, given a value π of the efficiency of recombinant innovation one may ask what the turning point of returns to scale is where diversity (α=1/2) becomes optimal. This threshold level solves the equation
V α ≡
_
1
s
⎛C ⎞ V (α =1 / 2) = _ [2 + ⎜ ⎟ ] =1 ⎝2⎠ 2s ~
(6)
If C=0 (for instance with π=0) we have =0. If C=1 (for instance with I=4I3 and π=1) we ≃ 1.2715. There is no closed form solution find as function of other parameters, but we can instead solve for C. For s>1 this solution is
= 2(2s−2)1/s
(7)
Proposition 1 For a given positive value of the recombination probability π, benefits from diversity are larger than benefits from specialization as far as I3/ I>1/π, for any value of returns to scale s. The reason is that the rate of growth of innovation is unbounded. Now assume the ratio of investments I3/I is given. For s=1 (constant re turns to scale) we have (1/2)s=1=1+C/4≥1, since C≥0. If a positive investment I3 is devoted to innovation, the following holds true:
Proposition 2 The threshold below which a diversified system is the optimal choice has the property that ≥1 and >1 iff π>0. Corollary 1 For all decreasing or constant returns a maximum value of total final benefits is realized for the allocation α=1/2, i.e. for maximum diversity. This situation is summarized in figure 2. The case of increasing returns to scale better represents real cases of technological innovation because of fixed costs and learning. Here there is a tradeoff between scale advantages and benefits from diversity. If the probability of recombinant innovation is insufficiently large, returns to scale may be too high for diversity to be the optimal choice. In figure 1 this holds for the bottom four curves. In general we have the following result, which completes Proposition 1: Corollary 2 Diversity α=1/2 can be optimal also with incre asing returns to scale ( >1) provided that the probability π of recombination is large enough. A general model Abandoning the hypothesis of zero initial values of parent options we introduce dynamics into the model. The general solution to system (1) is the following: O1(t) = O10 + I1t O2(t) = O20 + I2t t O3(t) = πI ∫ BOsO sds
(8)
AENORM
62
January 2009
47
Wie schrijft blijft? © 2007 KPMG Staffing & Facility Services B.V., een Nederlandse besloten vennootschap, is lid van het KPMG-netwerk van zelfstandige ondernemingen die verbonden zijn aan KPMG International, een Zwitserse coöperatie. Alle rechten voorbehouden.
Schrijf je scriptie of afstudeeropdracht bij KPMG. Eerlijk is eerlijk, niet iedere tekst is even goed. Maar vaak zit er iets slims of moois tussen. En heel soms iets onvergetelijks. Zo is dat op de wc-deur van een kroeg en zo is dat bij KPMG, waar studenten als jij een scriptie of afstudeeropdracht kunnen schrijven. Zo’n scriptie of afstudeeropdracht is een ideale manier om kennis te maken met KPMG. Misschien zelfs het begin van een prachtige carrière: geef je al schrijvend blijk van passie voor het vak, dan moet je maar eens serieus overwegen om te blijven. Meer weten? Kijk op www.kpmg.nl/stages.
A U D I T TA X A DV I S O R Y
48
AENORM
62
January 2009
Econometrics
In the long run (t>>Oi0/(αI), i=1,2) the parents’ initial values are negligible and the balance converges to a constant value:
B
4
(O10 + αIt )(O20 + (1 − α)It ) → 4α(1 − α) (9) (O10 + O20 + It )2
Proposition 3 In the long run the balance converges to the constant value B(α) = 4α(1−α), which is independent of initial values of parent options. O3 attains linear growth. In general the dynamics of the balance depends on the relative magnitude of ratios O10/O20 and α/(1−α). There is a necessary and sufficient condition for constant balance:
is a transitory phase in which the innovation “warms up” before becoming effective. This is a stylized fact of recombinant innovation. Relaxing the assumption of constant balance, we have a much more complicated expression for the new option. But in the long run (It>> O0) it reduces to the following: Oσ t ≃ α − αIt −
I It σ O − α − Oα I O
(15)
The logarithmic term adds negatively in expression (15), producing the expected convex time pattern of innovation, which tells about the diminishing marginal contribution of parent technologies.
"If investment is large enough, diversity will always become the optimal choice" Proposition 4 The balance is constant through time and equal to B(α) = 4α(1−α) iff O10/O20 = α/(1−α)
(10)
In the pilot model the probability of emergence is essentially the balance of parent options. Now a size factor is introduced to capture the positive effect that a larger cumulative size has on recombinant innovation, i.e. a kind of economies of scale effect in the innovation process. If the size effect is expressed by a factor S(O1,O2), the probability of emergence becomes: PE = πB(O1,O2)S(O1,O2)
(11)
where S(O1,O2) = 1 − e − σ (O1 + O2 )
(12)
Here ∂S/∂Oi=∂S/∂O=σ/eσO, with O=∑iOi. The parameter σ captures the sensitivity of PE to the size. Assume that condition (10) holds true: with constant balance B=4α(1−α) the rate of growth of innovation is
IPE t = IB − e − σ O + It
(13)
Its integral is
O t = πIα − αt +
e − σO − σIt − e σI
Optimization of diversity In the general model the optimization of final benefits is enriched by several dynamical effects. If a size factor is present in the probability of emergence, normalized benefits look:
s s s s s s (α;t) = α + (1−α) +C m(t) α (1−α)
(16)
where C=4πI3/I. A time dependent factor shows up, m(t)=1+(e−σIt−1)/σIt, with m′(t)>0, →0 m(t)=0 and →∞m(t)=1. Let’s define a function C(t)=Cm(t). Final benefits (16) are formally the same as in the pilot model (5), provided that C(t) is in place of C: although the size effect makes the system dynamic, still optimal diversity will be either α=0 and α=1 or α=1/2. This is better understood looking at figures 1 and 2. Given I, I3 and π, as time flows C(t) increases and the benefits curve goes from the lower curve π=0 (C=0) to the upper curve π=1 (C=1). If π is large enough, optimal diversity will shift at some time from α=0 and α=1 to α=1/2. There is a threshold level (t) where, for a given time horizon t, benefits with α=1/2 are the same as benefits from specialization (α=0 and α=1):
(14)
The second term tells about the size effect. •• • Here 3(t)>0 and 3(t)>0 ∀ t≥0: the innovative option has a convex time pattern. There
Figure 3: As time goes by, the region of returns to scale where diversity is optimal becomes larger.
AENORM
62
January 2009
49
Econometrics
(figure 4). In the long run symmetry is restored since the effect of initial values dissipates. Again, in the long run the system is well approximated by the pilot model. Conclusions and further research
Figure 4: Final benefits with positive initial values and no size effect. Here we have O10=1, O20=10, s=1.2, π=1 and I=4I3=1. The five time horizons are in units of 1/I.
C t v α = = + s t
st
=
(17)
Proposition 5 For a given time horizon t diversity (α=1/2) is optimal iff s< (t).
How does (t) behave? The larger t, the larger (t). The intuition is as follows. C(t) is increasing: time works in favour of recombinant innovation. As time flows, the region of returns to scale where diversity is optimal enlarges. The threshold (t) converges to the value of the pilot model (see figure 3). Even with π=1 diversity may never become the optimal solution if returns to scale are too high ( <s). But if investment I3 is large enough, diversity will always become the optimal choice. This is consistent with proposition 1: given returns to scale s, if one has infinite disposal of investment I3, threshold can always be made such that >s: at some time t one will see (t)>s. Concluding, the size factor introduces a dynamical scale effect into the system. The optimal solution may change through time, but it can only switch from α=0 and α=1 to α=1/2. This happens if and only if the probability of recombination is large enough. In the limit of infinite time (It>>O0) the size effect saturates ( →∞S(t)=1). If the time horizon is long enough the size factor can be discarded. The solution of the optimal diversity problem is approximated by the solution of the static pilot model. One final remark about the effect of initial values of parent options on optimization of diversity. The main consequence is a reduction of symmetry in the system (unless O10=O20). The symmetric allocation α=1/2 is not a solution to the above equation in general. Optimal diversity is represented by a function of time α*(t)
50
AENORM
62
January 2009
This study has proposed a problem of investment allocation where the decision maker faces a trade-off between scale advantages and benefits of diversity. Optimization of diversity comes down to finding an optimal trade-off between these two. Due to time constraints, high returns to scale and limited disposal of capital, specialization may result the best choice, despite recombinant innovation. But in principle if two technologies recombine diversity of investment can always be made the optimal decision. Several directions for future research can be identified. Investment in the innovative option can be endogenized, i.e. made part of the allocation decision. Extending the number of parent options allows for an examination of complexity aspects of recombinant innovation, as well as for assessing the marginal effect of new options. References Ethiraj, S. K. and Levinthal, D. (2004). Modularity and Innovation in complex systems. Management Sci., 50(2), 159–173. Fleming, L. (2001). Recombinant uncertainty in technological search. Management Sci., 47(1), 117–132. van den Bergh, J. C. J. M. and Zeppini-Rossi, P. (2008). Optimal diversity in investments with recombinant innovation, Tech. Rep. 2008 -091/1, Tinbergen Institute.
Econometrics
To DICE with Climate Change Global warming is the main contemporary environmental issue. It has even reached the G8 agenda at their meeting in Japan this summer. This clarifies that climate change, human beings, and especially the economy, are nowadays strictly connected. When emissions of greenhouse gases (GHGs) increase, so do the concentrations of GHGs in Earth’s atmosphere. This leads to an increase in radiative forcing which causes temperatures to rise and in turn is the source of many damages. That is why it has become important to link climate change to economic models. The DICE model (Dynamic Integrated model of Climate and the Economy), developed by W.D. Nordhaus, is one of the first models to link these two subjects. The aim of the DICE model is to connect the scientific causes and effects of global warming to the economics of emissions of GHGs (Nordhaus and Boyer, 1999).
The first DICE model has often been criticized for its overly simplified climate change module. In more recent versions Nordhaus (2007) altered DICE by including feedback links (processes that tend to in- or decrease a certain effect) in the climate change module. This paper examines the importance of feedback links for the recent DICE model, compared to the first, simplified DICE model; it does so by using a regression analysis. The enhanced greenhouse effect Earth’s surface bounces part of the sun transmitted short-wave radiation back into the atmosphere as short-wave radiation and it emits long-wave radiation (see figure 1). However some gases in the atmosphere absorb the longwave radiation, acting as a blanket. This is the natural greenhouse effect and is mainly (80%) caused by water vapour (Skinner, Porter and Parker, 2004). Clouds are formed by small water vapour particles attached to aerosols. They have two main functions. On the one hand they radiate short-wavelength radiation from the sun back into the atmosphere, reducing the energy balance. On the other hand they capture the long-wave radiation emitted by Earth and therefore enlarge the energy balance (Waterloo, Post and Horner, 2007), (Houghton, 2004, p.90-92). However, the natural greenhouse effect is not the source of the contemporary problem; it is the enhanced greenhouse effect that causes Earth to warm up extensively and hence contributes to the climate change (Chiras, 2006, p. 445). Contrary to its noteworthy influence on the natural greenhouse effect, water vapour does not play a significant role in the enhanced greenhouse effect, since it is not one of the GHGs emitted by humans (Houghton, 2004,
Marleen de Ruiter started her studies in econometrics at the UvA in 2003 but she switched to the Earth & Economics bachelor at the VU in 2006. She is expected to graduate in January 2009. Along with her studies, she is also a teaching assistant and researcher at the SPINlab-VU. This article is a summary of her bachelor thesis, which she wrote at the IVM under supervision of Dr. R. Dellink and Dr. M. van Drunen.
p. 16). When the emissions of GHGs increase due to human activities, the atmospheric concentration of the GHGs increases, the ‘blanket’ becomes thicker and the temperature rises (Skinner, Porter and Parker, 2004). However, the greenhouse effect becomes more complicated because of the several feedbacks which it induces (Houghton, 2004, p.90). Feedbacks tend to enlarge (positive feedback) or reduce (negative feedback) an impact (Skinner, Porter and Parker, 2004). Feedbacks that play an important role in the enhanced greenhouse effect are for instance the cloud-radiation feedback as described in the above and the water vapour feedback: when the atmosphere warms up because of the greenhouse effect, e.g. of water vapour, the amount of evaporation mainly coming from oceans raises. Moreover the relative humidity of the atmosphere increases causing more water vapour, which in turn tends to warm up the atmosphere (Waterloo, Post and Horner, 2007, p. 26). Important GHGs Table 1 shows the most common gases in the atmosphere as well as the ones that contribute to the enhanced greenhouse effect. According to the IPCC third assessment report (2001) the greenhouse gases, with an enhanced greenhouse effect mentioned here, all belong to the
AENORM
62
January 2009
51
Econometrics
formula Houghton (2004) uses to convert the atmospheric concentration of carbon dioxide (C) into its radiative forcing (R in W m–2) is given below where C0 is the pre-industrial CO2 concentration of 280 ppm. R=5.3*ln(C/C0)
Figure 1: The enhanced greenhouse effect. Source: IPCC (1990). Homepage (www.ipcc.ch), 10th of July.
Kyoto greenhouse basket. The gases not elaborated, because of their rare existence, but included into the Kyoto basket of greenhouse gases are: SF6, PFCs and HFCs. Aggregation of GHGs Scientists used to convert all the GHGs into carbon equivalent concentrations (often calculated in ppm or GtC2) in order to compare them with each other. The concept of Global Warming Potential has been used commonly. Houghton (2004) defines GWP as the ratio of the enhanced greenhouse effect of any gas compared to that of CO2. The index solves the problem of the different lifetimes of the GHGs (Houghton, 2004). Nowadays it is more common to convert all the concentrations of GHGs in the atmosphere into radiative forcing based on CO2 concentrations instead of carbon (Houghton, 2004). Houghton (2004, p.29) defines radiative forcing as “the change in average net radiation at the top of the troposphere, which occurs because of a change in the concentration of a GHG or some other change in the overall climate system”. When the radiative forcing balance between outgoing and incoming radiation is positive, the surface will warm up and vice versa Chiras, 2006). The
The constant in this formula (5.3) changes when we compute R for different gases. In its third assessment report, the IPCC (2001) gives a mathematical definition of GWP based on equivalent radiative forcing instead of equivalent concentrations. This gives the radiative forcing of every gas compared with that of CO2:
=
∫ ∫
⋅ ⋅
Where the ratio is computed “of the time-integrated radiative forcing from the instantaneous release of 1 kg of a trace substance relative to that of 1 kg of a reference gas” . TH is the time horizon for which the GWP as the radiative forcing of a certain GHG is computed (IPCC Working group I, 2001, chapter 6). The carbon cycle To fully determine climate change, restricted to the enhanced greenhouse effect, the carbon cycle needs to be defined. There are four main reservoirs through which carbon moves: atmosphere, land (biosphere), ocean (hydrosphere) and the geological reservoirs (lithosphere) as shown in figure 2. Humans tend to increase the amount of CO2 in numerous ways. When the amount of anthropogenic emitted CO2 increases (because of fossil burning and land-use change) the uptake of CO2 by oceans and land biota cannot entirely compensate it. Therefore, the carbon cycle is no longer balanced; hence the amount of at-
Gas
Mixing ratio or mole fraction expressed as fraction* or ppm1.
Relative contribution to the enhanced greenhouse effect (values date from 2004)
Nitrogen (N2)
0.78*
No contributor
Oxygen (O2)
0.21*
No contributor
Water vapour (H2O)
Variable (0-0.02*)
No contributor
Carbon dioxide (CO2)
370
60%
Methane (CH4)
1.8
15%
Nitrous dioxide (N2O)
0.3
5%
Chlorofluorocarbons (CFC’s)
0.001
12%
Ozone (O3)
Variable (0-0.1000)
8%
Table 1: Composition of gases in the atmosphere and their relative contribution to the enhanced greenhouse effect. Source: Mixing ratio’s in Earth’s atmosphere (Houghton, 2004, p. 16) and contribution to greenhouse effect (Skinner, Porter and Park, 2004, p. 514) 1 2
52
Parts per million. Gig tonnes of Carbon. 1GtC = 3.7Gt CO2
AENORM
62
January 2009
Econometrics
Figure 2: The main components of the natural carbon cycle; volumes and exchanges in billions of tonnes of carbon. Source: IPCC (2001) third assessment report; working group 1 – The scientific basis.
mospheric carbon increases (Skinner, Porter and Parker, 2004). The link between the carbon cycle and its effects upon radiative forcing is important. The IPCC (2007) defines how the climate responds to constant radiative forcing (equilibrium climate sensitivity) as the “equilibrium global average surface warming” when the concentration of CO2 is doubled. When the concentrations of GHGs build up in the atmosphere the radiative forcing increases causing the temperature of the atmosphere, and therefore the temperature of the shallow oceans, to rise as well. This warms the deeper oceans and leads to a positive feedback of warming the temperature of the atmosphere again (Houghton, 2004, p.93-95). The carbon cycle is influenced by several feedbacks as in the above-mentioned example. Different scenarios In order to describe DICE, one of the main characteristics of climate change models needs to be
defined; hence we divide climate change scenarios into three main groups. The first group consists of the ‘business-as-usual’ scenarios, also known as baseline or non-intervention scenarios. In these scenarios it is assumed that there won’t be major changes in current attitudes and priorities, except for the policies that have recently been agreed on and are soon to be implemented (European Environment Agency, 2008). The second main group consists of the mitigation scenarios. These are the opposite of the ‘business-as-usual’ scenarios; they are based on the assumption that there will be changes in the future, for example technological alterations or a changing attitude towards economic development, in order to reduce future global warming impacts (European Environment Agency, 2008), (Houghton, 2004). In the third scenario, the stringent scenario, emissions are rigidly controlled by the government (The encyclopaedia of the Earth, 2008).
Equation
Output
Optimal scenario
Base scenario
Stringent scenario
1
β1
(-0.023351, 0.002228)
(-0.040784, 0.003081)
(-0.015875, 0.001763)
β2
(0.603997, 0.037199)
(0.687701, 0.033681)
(0.583671, 0.044495)
R
2
β3 2 3
0.875639
0.912707
0.823092
(3.752063, 0.014432)
(3.787804, 0.011605)
(3.711871, 0.018494)
R2
0.987840
0.995803
0.978397
β4
(0.977864, 0.014752)
(0.953037, 0.011832)
(0.990999, 0.016899)
β5
(0.206136, 0.010147)
(0.223645, 0.008018)
(0.196290, 0.011682)
R2
0.998423
0.999687
0.994125
Table 2: Summary of the most important output of the three scenarios where the R and the (μ,σ) of the β’s are shown. 2
AENORM
62
January 2009
53
Econometrics
The DICE model
Figure 3: Changes in the climate change module of DICE: grey are the old parts and in black the recently added parts. Source: Nordhaus (1994 and 2008).
The aim of the DICE model is to connect the scientific causes and effects of global warming to the economics of emissions of greenhouse gases: economic activity (Q) causes an increase in emissions of GHGs (E) which increases the amount of C in the atmosphere (MAT), leading to an increase in radiative forcing (Forc) which causes a rise in atmospheric temperature (Tatm). This leads to an increase in radiative forcing which causes temperature to rise. Therefore it focuses on damages, rather than emission reduction, as a result of which it tries to find a balance between less economic investments today (loss) and less climate change in the future (gain). Time periods are divided into periods of ten years starting from 1995. The goal is to maximise the social-welfare given the constraints concerning climate change and utility, where the radiative forcing, carbon cycle and climate equations are globally aggregated. Simplified equations In order to examine the relevance of the feedback links, simplified equations are established based on the DICE-GAMS code and shown in box 1. (1) MAT(T+1)=(1-β1)∙MAT(T)+β2∙E(T) (2) FORC(T)= β3∙[log(MAT(T)+0.000001) /596.4]/log(2)+FORCOTH(T) (3) TATM(T+1)=(1-β4-LAM)∙TATM(T) +β5∙Forc(T) Box 1
A summary of the output of the regression analysis of the three scenarios is shown in table 2. What is striking about the adjusted R2 values is that for every single equation the base scenario has the highest and the stringent scenario has the lowest adjusted R2 value. This suggests that in a scenario where there are no climate change policies, the simpler model achieves al-
54
AENORM
62
January 2009
most the same output as the model that does include feedback links. However, at first one would expect that when there is more climate change, the feedback links would be stronger. An explanation for this could be that when there is more climate change, the relative effect of the feedback links, compared to the total effect, becomes smaller; e.g. the relationship between the two is not linear. Furthermore, the μ and σ-values of the βj’s are very similar to each other per scenario. When constructing confidence intervals for all equations and scenarios, the base scenario of each equation contains either the maximum or minimum value of all three confidence intervals and the stringent scenario contains exactly the opposite. This also implies that for every βj, the confidence intervals of the optimal scenario are between the confidence intervals of the three scenarios, where the ranges sometimes partly overlap each other. Conclusions In order to examine the relevance of the feedback links for the climate change module of DICE, three simplified equations, which didn’t include feedback links, have been established based on the equations described in the DICE 2007 GAMS code. When analysing the results, the adjusted R2 has been used as an important measure for the goodness of fit. It has been shown that for every j, βj has an adjusted R2>0.75. Therefore a re-estimation of the functional forms from the simplified equations based on the first DICE model approximates the most recent and more complex version of the functional form reasonably well. It should be noted however that this thesis is based on a number of boundaries and simplifications. Therefore, the subject of this thesis can be broadened in several ways. References Chiras, D.D. (2006). Environmental science, 7th edition, Massachusetts: Jones and Bartlett Publishers. Cunningham, W.P. and Cunningham, M.A. (2008). Environmental science: a global concern, 10th edition, New York: McGraw-Hill. Houghton, J. (2004). Global warming, Cambridge university press. IPCC third assessment report (2001). Working group 1, the scientific basis, Sweden: IPCC. IPCC (2007). Climate change 2007: synthesis rapport, Sweden: IPCC. Nordhaus, W.D. (1994). Managing the global
Econometrics
commons – The economics of climate change, MIT Press. Nordhaus, W. and Boyer, J. (1999). Roll the DICE again: the economics of global warming, MIT Press. Nordhaus, W.D. (2008). A Question of Balance: Economic Modeling of Global Warming, MIT Press. Skinner, B.J., Porter, S.C. and Park, J. (2004). Dynamic Earth – An introduction to physical geology, 5th edition, Hoboken: J. Wiley & Sons. Waterloo, M.J., Post, V.E.A. and Horner, K. (2007). Introduction to hydrology, syllabus, Vrije Universiteit press. European Environment Agency (2008). Homepage (http://www.eea.europa.eu), 17th of July. Encyclopaedia of the Earth (2008). Homepage (http://www.eoearth.org),11th of September. IPCC AR4 (2008). Homepage (http://www.ipcc. ch/pdf/assessment-report/ar4/syr/ar4_syr. pdf), 17th of July. US Climate change science program (2008). Homepage (http://www.climatescience.gov), 17th of July.
AENORM
62
January 2009
55
RUIMTE
voor uw ambities Risico’s raken uw ondernemersgeest en uw ambities. Aon adviseert u bij het inzichtelijk en beheersbaar maken van deze risico’s. Wij helpen u deze risico’s te beoordelen, te beheersen, te bewaken en te financieren. Aon staat voor de geïntegreerde inzet van hoogwaardige
expertise,
diensten
en
producten op het gebied van operationeel, financieel en personeel risicomanagement en verzekeringen. De focus van Aon is volledig gericht op het waarmaken van uw ambities.
In Nederland heeft Aon 12 vestigingen met 1.600 medewerkers. Het bedrijf maakt deel uit van Aon Corporation, Chicago, USA. Het wereldwijde Aon-netwerk omvat circa 500 kantoren in meer dan 120 landen en telt ruim 36.000 medewerkers. www.aon.nl.
4982aa
RIS IC O M A NA G E ME NT | E MPL O YEE B EN EF I T S | VER Z EKER I N GEN
56
AENORM
62
January 2009
ORM
Container Logistics Since the introduction of containers in the early 1960s, the containerised trade market has been growing rapidly. As a result, ships have grown in size to the current maximum of 13,000 TEU (twenty-foot-equivalent-unit; length of the smallest container is twenty feet). Ports and terminals should handle their operations efficiently to keep up with this growth and to provide the necessary capacity and customer service. Especially in Europe, where the competition among the ports in the Le Havre-Hamburg range is fierce. In other words, large amounts of containers have to be loaded, unloaded, stored, retrieved and transhipped in a short time span.
Terminal management needs to make many decisions in relation to all logistics processes to achieve the goal of minimizing berthing times of ships. Techniques from Operations Research can be used to formulate tools that assist in this decision making process. The complexity of the logistics processes ensures a continuous challenge for researchers to develop new mathematical techniques to obtain good and, whenever possible, optimal solutions. In this article, we first describe each of the logistics processes in more detail. Secondly, we summarise a novel approach for the storage and retrieval process as presented in the Operations Research paper of Vis and Roodbergen (2009). Logistics Processes at container terminals The process of unloading and loading a ship at a container terminal can be described as follows (see Figure 1): manned quay cranes unload containers from the shipâ&#x20AC;&#x2122;s hold and the deck. These containers are positioned on transportation vehicles, which travel between the ship and the seaside of the stack (i.e., storage area). When a transportation vehicle arrives at the stack, a crane (e.g., a straddle carrier) takes the container off the vehicle and stores it in the stack. After a certain storage period, the containers are retrieved from the stack and are transhipped to other modes of transportation, such as barges, trucks and trains. Transhipment to trucks and trains occurs at the landside of the stack. To load containers onto a ship, these processes need to be executed in the reverse order. Various decision problems need to be solved to obtain an efficient container terminal. Vis and De Koster (2003) present a classification of decision problems that arise at a container terminal. Examples of these decision problems include allocation of ships to berths, selection of transportation and storage systems, layout
Iris F.A. Vis is an Associate Professor of Logistics at the VU University Amsterdam. She holds an M.Sc. in Mathematics from the University of Leiden and a Ph.D. from the Erasmus University Rotterdam. She received the INFORMS Transportation Science Section Dissertation Award 2002. Her research interests are in design and optimization of container terminals, vehicle routing, and supply chain management Kees Jan Roodbergen is an Associate Professor of Logistics and Operations Management at Rotterdam School of Management, Erasmus University. He received his M.Sc. in Econometrics from the University of Groningen and his Ph.D. from the Erasmus University Rotterdam. His research interests include design and optimization of warehousing and cross docking environments.
Figure 1: Top: Loading and unloading a container ship at Ceres Paragon Amsterdam (Ceres Paragon Terminals) Bottom: Schematic drawing of logistics processes at a container terminal
AENORM
62
January 2009
57
ORM
Seaside Storage request with ith origin i i seaside id
Assigned storage location Retrieval request with destination landside
Storage location Pick-up and delivery point
Landside
Figure 2: Straddle carrier stores a container in the stack (Ceres Paragon Terminals)
of the terminal, dispatching of containers to vehicles and sequencing of storage and retrieval requests at the stack. In the remainder of this article we will focus on one decision problem in particular, namely sequencing requests for cranes operating in the stack. First, we will discuss stacking operations in more detail. Thereafter, we will describe an efficient algorithm that can determine optimal schedules. Stacking of containers Straddle carriers operate in a block of containers. Containers are stored for a certain period in such a block to await further transportation. To load the container on a ship or other modality, the container is retrieved from the stack. A block consists of a number of rows of containers with a pickup and delivery point (I/O point) at both ends of each row, namely one at the seaside and one at the landside. At a pickup and delivery point, the container is either taken from or placed on the ground. These pickup and delivery points are assumed to have infinite capacity. The transport of the container to and from this point is executed by another type of material handling equipment, for example, a multi trailer system. Clearly, the storage and transportation processes are decoupled. Vis and Roodbergen (2009) focus on the scheduling of storage and retrieval requests within the stack for a single straddle carrier (see Figure 2). A straddle carrier (SC) can travel over a single row of containers and transport a single container at the same time. At the head or tail of each row a SC can make the crossover to another row. A tour of a SC indicates the order in which a set of requests need to be handled. Each request has a pick-up and dropoff location. For a retrieval request, the pick-up location is simply the location where that container is currently stored and the drop-off location is the rowâ&#x20AC;&#x2122;s I/O-point at either the sea-side or the land-side of the stack, depending on the containerâ&#x20AC;&#x2122;s destination. For a storage request, the pick-up location is the I/O-point where the container has been delivered, which in turn depends on whether the container arrived by sea
58
AENORM
62
January 2009
Figure 3: Required movement to store and retrieve a container
or by land. The drop-off location is the assigned storage location in the same row. Figure 3 provides an illustration of the required movements to store or retrieve a container. As a result, a SC travels full from an origin to a destination location. Empty travel distances occur when a SC travels from the destination of a container to the origin of another container. Clearly, full travel distances are fixed. Therefore, it is required to schedule storage and retrieval requests for a SC in such a way that empty travel distances are minimized. In the remainder of this article we present a summary of the scheduling algorithm developed by Vis and Roodbergen (2009) to calculate an optimal tour for a SC to handle all requests. Figure 4a illustrates a straddle carrier operating in a stack with 6 rows, I/O-points at both ends of each row, the start and end point of the tour (i.e., depot) and several storage and retrieval requests with respectively their destination or origin that need to be handled. In finding an optimal tour we assume that the tour of a straddle carrier starts and ends at the same, given position. Furthermore, there are no restrictions on the sequence of storage and retrieval requests within a given row. However, once a row is enInput/output points (seaside) row
a11 1 s111
1 straddle carrier
1
a21
1
1 1 r11
a31
1
a41
1
a51
1
1
1 r31 1 2 s31
1 r51 6
1 1 s51
8
1 s41
8
1
3 1 r61 2 1 s61
8
7
1
1 a12
depot
Input/output points (landside) 1
(a)
0
1 a22 1
1 1 s62
2 r412
s212 1 r212
s112
a61
a2 3
1
v0
Retrieval request from seaside
Retrieval request from landside
Storage request from seaside
Storage request from landside
3
1 2 s41 1 a42
1
2 r61 1 a62
a52 1
1
(b)
Figure 4: (a) Example of a stack with 6 rows, I/O-points at both land- and seaside of each row and several storage and retrieval requests. (b) Directed network representation of (a). To keep the figure simple, we incorporated only arcs between two adjacent nodes.
ORM
tered, all requests in that row must be handled before the straddle carrier can continue to the next row. This assumption is quite reasonable in practice due to the large time needed to change rows. Note that this assumption restricts the number of times rows are entered and left; it does not put a direct restriction on the travel distances between rows, nor does it force a fixed sequence to visit the rows. If beneficial, a row may be initially skipped and only visited at a later stage. Vis and Roodbergen (2009) also show how this problem can be solved without these restrictions.
complete graph, based on this formulation, for only the first row is given in Figure 5b. We now have a directed network that conforms with the definition of the Rural Postman Problem (RPP, see e.g., Lawler et al. 1985). The more general Rural Postman Problem is known to be NP-hard (Lenstra and Rinnooy Kan, 1976). However, in our problem the nodes of the network are not freely positioned in Euclidian space, but restricted to a limited number of parallel lines (the rows of the stack). Connections between the rows can only occur at the head and tail of the rows. Furthermore, all required arcs in any row i either originate at or end in the head and tail
"This algorithm significantly outperforms common rules of thumb " Sequencing storage and retrieval requests The first step in the solution approach is to construct a network by representing each location which need to be visited to perform a storage or retrieval request as a node. Furthermore, the pick up and delivery points of each row and the depot are also represented by nodes. Between two nodes multiple directed arcs exist. Figure 4b represents the network for the example depicted in Figure 4a. The nodes s correspond to the locations, where containers must be stored. The nodes r correspond to the locations where containers must be picked up. The superscript indicates whether the storages (retrievals) originate from (end at) either input/output point I/O1 or I/O2. Node v0 corresponds to start point of the tour. Nodes ai1 and ai2 correspond to the locations of respectively I/O1 or I/O2 for each row i. Next, we need to define the arcs in the network. For any pair of nodes x,y in the same row we introduce two arcs: (x,y) and (y,x). Some of these arcs must be traversed, i.e., are required (see Figure 3). Furthermore, we introduce an unlimited number of copies of the arcs to make the connections between the rows. The length of an arc (x,y) is denoted as d(x,y) and simply equals the physical distance between the nodes x and y. Only directed arcs between two adjacent nodes have been drawn in Figure 4b to keep the figure 2 simple. For example, the arc( ai1 , s11 ) also exists with distance 9. Figure 5a shows row 1 of Figure 4b, with only the required arcs drawn. That is, the arc from a retrieval to the corresponding I/O-point, and the arcs from the appropriate I/O-point to the storages have been drawn. These arcs must be traversed, other arc traversals are optional, but some must be included to obtain a tour. A
of the row. We can exploit this special structure to solve this problem efficiently. The first step is to transform the current network formulation into a network formulation that classifies as a Steiner Travelling Salesman Problem (STSP). The Steiner Travelling Salesman Problem (STSP) looks for a shortest tour such that a given subset of the vertices is visited at least once and is only solvable in polynomial time for some situations (see e.g., Cornuéjols, Fonlupt and Naddef, 1985). To achieve this, we need to change the RPP network formulation such that we no longer have required arcs, but are still able to obtain a valid solution for the original problem. All nodes are also included in the STSP formulation (see Figure 5c). Most arc lengths d(x,y) are equal to their RPP counterparts with some exceptions. For example, before we can travel from 1 the retrieval location r11 to the storage location 2 s11, the straddle carrier first needs to deliver the container to I/O1. Thereafter, the straddle carrier needs to travel from I/O1 to the origin of the storage request, I/O2, and then with the 2 container to storage location s11 . We now have a network formulation that can serve as an input for our solution method. 1
1
r11 1
1
1 s11
1 8 8 8
9
(a)
a1
1
2 s11
1
1
1 r11
1 9
9
1 s11
12
13
8 10
11
1 1
a12
(b)
9 10 2 s11
11
1
10
7
10
9
2 s11
1 a12
a1
1
10
8
9
1
3
7
7 10
2 2
2
1 s11
a12
r11
2
a11
(c)
Figure 5: (a) Representation of row 1 of Figure 4b with only the required arcs drawn. (b) Graph of the RPP formulation with dashed lines representing the required arcs. (c) Graph of the reformulation as STSP. Required nodes are solid black. Arcs lengths that differ from their RPP counterparts are highlighted.
AENORM
62
January 2009
59
ORM
ai1
ai1
ai1
ai1
ai1
ai2
ai2
ai2
ai2
ai2
(iii)
(iv)
(v)
(i)
(ii)
Figure 6:Five ways to visit a row
Shortest tours in the network will be determined by means of dynamic programming. Ratliff and Rosenthal (1983) already used dynamic programming to solve the problem of routing order pickers in a warehouse. In their problem a shortest tour has to be found to retrieve a number of products from specified locations in a warehouse with a number of parallel aisles (partly comparable with the rows in a container stack). However, their method is only applicable for undirected networks without required arcs and does not incorporate storage requests. The objective of this dynamic programming algorithm is to construct a directed tour sub graph visiting all nodes at least once and all arcs at most once. The essence of a dynamic programming method consists of three components: the potential states, the possible transitions between states, and the costs involved in such a transition. By consistently adding arcs (transitions) to partial tours (states), the algorithm gradually builds towards a complete tour. There are two types of transitions in the algorithm. The first transition type consists of adding connections between two consecutive rows. The second type of transition consists of adding arcs within a single row. Given the fact that a row may be visited only once, only five ai1
1 ai+1
2
ai+1
2
ai
ai+1
2
2
ai
ai+1
1
ai+1
2
ai+1
2
ai 2
ai+1
2
ai+1
2
ai
ai
1
1
ai
ai
(0c)
ai+1
2
2
ai+1
1
2
1
ai+1
1
ai
2
ai2
ai+1
2
ai
2
ai+1
AENORM
62
a31
8
9
a41
a51
11
a61
10 5
1
12
7
2
17 a12
2 ai+1
â&#x17D;Łn/2â&#x17D;Ś a
1
ai+1
2
ai+1
January 2009
3
1
1
2
Figure 7: Possible transitions to make crossover to another row
60
a21
(0d)
(3a)
ai
2
4
ai+1
ai
1
ai+1
a11
1
ai
(2a)
1
ai
ai
ai
(1a)
ai
1
1
(0b)
1
1
ai+1
ai+1
(0a)
ai
1
ai1
useful arc configurations can lead to a shortest feasible tour in a single row (see Figure 6). These five arc configurations are, however, not always trivial to obtain. A separate optimization procedure is required to find a shortest sequence of jobs within a row for any given starting and ending point. The first step in this procedure is to formulate a bipartite network with a copy of each node s and r at both sides. The start point is included at the left side and the end point at the right side. The assignment problem can be solved by using an existing method such as the Hungarian method (e.g., Papadimitriou and Steiglitz, 1982). The resulting solution may have unconnected cycles. However, Vis and Roodbergen (2009) show that these can be patched such that the final result is optimal. The resulting five configurations can be used in the dynamic programming algorithm. The dynamic programming algorithm starts with an empty tour. Next, all possible routes within row 1 are determined and the costs are calculated according to the explanation provided above. We now either have one empty partial tour subgraph (if there is no work to be done in row 1) or four partial tour subgraphs. The next step is to make the connection to row 2. Intuitively, it is quite clear what these connections should look like (see Figure 7). For example, a partial tour subgraph with a south-to-south traversal (see Figure 6iv) in row 1 must be connected by two arcs at the south position. Any other option would either not result in a feasible tour or would not be optimal. Costs are counted in each step by simply adding the lengths of the arcs that are added. The next step is to de-
18
a22
6 16
a32
a42
15
14
a52
13
a62
v0
Figure 8: Optimal route for a straddle carrier to handle the requests as denoted in Figure 4a. The red numbers indicate the order in which the arcs are handled. The straddle carrier starts in row 1and finishes in row 2 with step 17 and returns in the depot with step 18.
ORM
termine all possible partial tour subgraphs by adding arcs in row 2. And so on. By applying this dynamic programming algorithm, a shortest directed multiple row storage and retrieval tour can be found in polynomial time. Figure 8 shows the solution for the example depicted in Figure 4a. Simulation studies demonstrate that this algorithm significantly outperforms common rules of thumb such as first-come-first served. This is just one example of a mathematical tool that can be used by the terminal management to assist in daily decision making. Many algorithms are available that can be directly implemented in computer systems at terminals. Both fast calculation times and good quality solutions demonstrate the usefulness of these methods to obtain direct savings in berth times of ships and logistics costs at terminals. References CornuĂŠjols, G., Fonlupt, J. and Naddef, D. (1985). The traveling salesman problem on a graph and some related integer polyhedra, Mathematical Programming, 33, 1-27. Lawler, E.L., Lenstra, J.K., Rinnooy Kan, A.G.H. and Shmoys, D.B. (1985). The Traveling Salesman Problem, a Guided Tour of Combinatorial Optimization, John Wiley & Sons, Chichester. Lenstra, J.K. and Rinnooy Kan, A.H.G. (1976). On general routing problems, Networks, 6, 273-280. Papadimitriou, C.H. and Steiglitz, K. (1982). Combinatorial Optimization, Algorithms and Complexity, Prentice-Hall, Inc., Englewood Cliffs. Ratliff, H.D. and Rosenthal, A.S. (1983). Orderpicking in a rectangular warehouse: a solvable case of the traveling salesman problem, Operations Research, 31(3), 507-521. Vis, I.F.A. and De Koster, R. (2003), Transshipment of containers at a container terminal: an overview, European Journal of Operational Research, 147, 1-16. Vis, I.F.A. and Roodbergen, K.J. (2009), Scheduling of container storage and retrieval, Operations Research, forthcoming. http:///www.ceresglobal.nl Terminals)
(Ceres
Paragon
http://www.irisvis.nl/container
AENORM
62
January 2009
61
Actuarial Sciences
Longevity in the Netherlands: History and Projections for the Future During recent years longevity (increasing life expectancy) has become an increasingly important issue for especially pension funds. Most pension funds have a Defined Benefit pension scheme. This means that the accrued pension rights imply a certain, guaranteed benefit per year. If, for example, life expectancy rises by one year, the pension fund must pay an extra year of benefit to, on average, all pensioners. Hence, the expected present value of the pension liabilities rises with increasing life expectancy.
Henk Angerman works at the ALM-department of the Algemene Pensioen Groep (APG). His daily occupations consist of both monitoring solvency positions of pension funds which are managed by APG and participating in isolated projects. An example of such a project is the so-called “langlevenproject”, which aims to make a forecast of the future life expectancy of Dutchmen in general and the Dutch pension participant in specific (these tend to live longer!).
Qualitative considerations
Tim Schulteis works at the PRO department (Product Research and Development) of the Algemene Pensioen Groep (APG). His main activity is conducting actuarial calculations in order to advice social partners, who determine the pension deal of a pension fund that is customer of APG. Furthermore, he is involved with the premium model of disability pensions and specific research projects, such as the survey to the future life expectancy in relation to the pension affirmation.
Currently pension funds already have the obligation (by regulation) to account for longevity in their premiums. A lot of research is done in the field of modelling life expectancy and probabilities of dying at a certain age. The aim is to produce a projection (prediction) for probabilities of death for different ages in different years. In classical actuarial mathematics, probabilities of death are assumed only to depend on age. A typical feature (by construction) of models for longevity is that the probability of death does not only depend on the age, but also on the year (e.g. the probability of death for a 28 year old in 2009 is different from the probability for a 28 year old in 2023). In the Netherlands these predictions of future life expectancy are mainly produced by the Dutch Statistical Office (CBS) and the Dutch Actuarial Association (AG). The Dutch Statistical Office uses so called “ex-
62
AENORM
pert opinion models” in which the opinions of experts on the future development of important causes of death are incorporated. In this publication, however, we will follow a different path and present the most commonly used statistical model. In this setup a certain functional dependence is assumed and the parameters of the model are fitted by methods known from econometrics.
62
January 2009
Before we present the model, we will try to give some intuition for the assumed functional form of the model by looking at some data. The first graph shows the natural logarithm of the probability of dying at a certain age for women in the period 2001-2006. First, we notice that these probabilities are fairly large for the zero year old and then decrease (showing the effect of infant mortality). After reaching a minimum at the age of about 10 years, the graph increases in a way that looks linear, implying an exponential growth of the probability of death with age: the probability of death increases by
log mortality rate women 2001-2006
Graph 1
Actuarial Sciences
log mortality rate
women
Graph 3
A first candidate could be of the form
Graph 2
year
approximately 10% per year. The second graph shows the dynamics of the probabilities of death for women at certain ages with time. Again the graph shows the natural logarithm of the probability of death (we will call this log-mortality in the remainder of this article). The graph shows a clear tendency of declining, but with different “speeds” of decline for different ages. The third graph shows the dynamics of life expectancy for men (thick line) and women (thin line) between 1850 and 2006. Here, life expec-
log m(x,t) = αx+κt+εx,t
(1)
Here, log m(x,t) denotes log-mortality. Hence, this model describes a situation, in which logmortality is given by an age-dependent factor (αx), a time-dependent trend (κt) and an error term (εx,t) that is assumed to be a white noise. A model of this form, however, does not capture the effect shown in the second graph. In the model the term κt does not depend on age and therefore the modelled dynamics of log mortality does not depend on age. We saw in the second graph that the decrease of log mortality with time is different for different ages.
"Existing models perform well from a statistical point of view, but may significantly underestimate future longevity" tancy is calculated as the expected value of the remaining life time, using the probabilities of death as they were measured in that year. In order to calculate the expected remaining life time for a 15 year old in, say, 1970 one needs future probabilities, since these persons are 16 in 1971, 17 in 1972 and so on. This has some major drawbacks. First, one needs projections for future mortality and secondly these projections must have a very long horizon (the cohort from the example turns 75 in 2030). Therefore, life expectancy in some year is always defined in terms of the probabilities belonging to that year. Model building We will use the intuition from the presented graphs to set up a model for the dynamics of mortality.
Therefore, our second candidate is a model of the form log m(x,t) = αx+βxκt+εx,t (2) This model enables us to capture the effect that model (1) missed. We can look at κt as a dependent trend and at βx as the “sensitivity” of the age x to this trend. We choose
αx =
1 T ∑ log m( x, t ) T t =1
(3)
where t=1,…,T is the observed time interval. If we look at log m(x,t) as being a matrix with 111 rows (corresponding to the ages 0,..,110) and T columns, notice that (2) assumes that the matrix A(x,t) defined as A(x,t) = log m(x,t) — αx
AENORM
62
January 2009
63
Actuarial Sciences
which is a matrix that depends on both x and t, can be, apart from the white noise εx,t, written as the product of a column vector βx, depending only on x, and a row vector κt, depending only on t.
Life expectancy women
CBS prediction 2006
Furthermore, we must impose some dynamics on κt. If we look at the fourth graph, we see that it might be a good attempt to model this with a first order autoregression κt = Θ+κt-1+ηt
(4)
where ηt is assumed to be a white noise. The model described by the equations (2), (3) and (4) is the so called Lee-Carter model for log mortality. Estimation of the parameters The estimation of the Lee-Carter model is done in consecutive steps. First, the mentioned decomposition A(x,t) = βx·κt is performed by so-called principal component analysis. Thereafter, the dynamics of κt are reestimated by techniques known from time series analysis to match the fitted and the observed data. After the parameter estimation is performed the Lee-Carter model can be used to compute predictions for future mortality. A complete description of the methods and the evaluation goes beyond the scope of this article. However, we mention the fact that an analysis of the residuals shows that for some countries a so-called cohort effect must be introduced (a term depending on t-x), whereas for Dutch data this extra variable is not necessary and the Lee-Carter model performs well. Concluding remarks We want to conclude our article by considering the dynamics of life expectancy in the world since 1880. First we present a graph in which the period 1880-2006 is divided into different time intervals for which the life expectancy of that Life expectancy
nd
ay rw
la Ice
n
pa
Ja
No d
lan
w
Ne
Graph 4
64
AENORM
62
a Ze
3 months per year
year
January 2009
year
Graph 5
specific country is depicted that had the highest life expectancy among all countries in that time interval. Notice that a linear regression would fit very well here and that the slope of this regression would be 0,25 (i.e. life expectancy increases by 3 months every year!). The fourth graph also shows life expectancy for the Netherlands for men and women. By collecting data on the amount of smokers it is possible to link this amount to some extent to the deviation from the “graph of best countries”. Since the percentage of smokers is decreasing, it is plausible that life expectancy will converge to life expectancy in Japan, or at least converge to the slope of 0,25. This would imply that predictions underestimate the effects of longevity. If parameters for a statistical model are estimated over a time period in which longevity effects were small for some reason (e.g. smoking) it is clear that projections based on these estimates will produce forecasts with “small” increase in longevity in the future. Hence, the Lee-Carter model performs well from a statistical point of view, but may not tell the whole truth and significantly underestimate future longevity in the case at hand. The fifth graph additionally shows the prediction made by the Dutch Statistical Office. From this it is clear that this prediction has a slope that is significantly lower than 0,25. Furthermore, life expectancy in the Netherlands in 2050 is predicted to be below current life expectancy in Japan. At the moment APG is working on models that take this effect into account.
B e p a a l d e o p t i ma l e b e l e g g i n g s m i x voo r een pens i oenfonds me t e e n v e r mo g e n v a n 4 . 3 m i l j a r d e u r o .
����� ����
����� ����� ���� �����
�������� �����
�����
�����
�����
�����
�����
���������
�����
�����������
Wat is haalbaar? En wat is verstandig? Hoeveel risico
investeringsstrategieën. We werken voor toonaan-
mag een pensioenfonds eigenlijk lopen? Je hebt
gevende bedrijven, waarmee we een hechte relatie
het wel over de oudedagvoorziening van honderd-
opbouwen om tot de beste oplossingen te komen.
duizenden mensen. Er moet voor hen hoe dan ook een
Onze manier van werken is open, gedreven en infor-
flinke taart overblijven. Bij Watson Wyatt kijken we
meel. We zijn op zoek naar startende en ervaren mede-
verder dan de cijfers. Want cijfers hebben betrekking
werkers, bij voorkeur met een opleiding Actuariaat,
op mensen. En op maatschappelijke ontwikkelingen.
Econometrie of (toegepaste) Wiskunde. Kijk voor meer
Dat maakt ons werk zo interessant en afwisselend.
informatie op werkenbijwatsonwyatt.nl.
Watson Wyatt adviseert ondernemingen en organisaties wereldwijd op het gebied van ‘mens en kapitaal’: pensioenen, beloningsstructuren, verzekeringen en
Watson Wyatt. Zet je aan het denken.
AENORM
62
January 2009
65
Econometrics
The Jackknife and its Applications The validity of econometric inference is almost invariably based on asymptotic theory and for good reason. Finite sample moments of estimators are generally very complex functions of the data, making their analysis extremely complicated. While relying on asymptotic approximations works well when the available sample is large enough, the small sample distribution of estimators might be poorly approximated if it is not. This potentially results in misleading point estimates and erroneous inference. The Jackknife is a nonparametric technique based on subsampling that aims at alleviating some of these issues. Its apparent simplicity and general applicability has provoked a long line of research in statistics but, until recently, its potential has been largely overlooked in the field of econometrics. I give a brief discussion of the origins of the Jackknife and provide some examples from the recent econometric literature where the Jackknife idea has been applied. Miller (1971) provides a good introduction to the Jackknife. Parr and Schucany (1980) contains a bibliography of statistics work pre 1980.
Origins of the Jackknife estimator
Koen Jochmans is a PhD student at the Department of Economics of the Katholieke Universiteit Leuven and a Pre-Doctoral Fellow of the EU RTN "Microdata: Methods and Practice" at CEMFI, Madrid. His main research interests lie in Microeconometrics and his work concerns identification, estimation, and inference in the presence of unobserved heterogeneity.
Quenouille (1949, 1956) noticed that, if B(θn) is a smooth function of the data, a linear combination of θn and estimators based on subsamples removes part of its bias, i.e. we obtain a bias-reduced estimator. The smoothness requirement is that, for some j ≥ 2, θ =
Preliminaries Let θ be the population parameter of interest and denote by θn an estimator based on a sample of size n. Assume θn to be consistent, i.e. as n → ∞, θn → θ. Although a variety of functionals could be of interest, for inference, we are most often concerned with the first two moments of θn. More precisely, we care about its finite sample bias and variance, which are defined as B(θn) = E[θn]-θ, = V (θn ) E[θnθ ] − E[θn ]E[θn ]'. ' n
(1) (2)
For example, when B(θn) is large, point estimates of θ will, on average, be located far from the truth. Additionally, confidence sets will have poor coverage, i.e., in repeated samples, they will contain θ far less often than their nominal condence would suggest.
+ + +
(3)
where B1,...,Bj are unknown constants and o(n-j) is a term that vanishes more quickely than n-j, as n → ∞. Assume for a moment that the data is i.i.d. Let θn-1,i denote an estimator from the subsample obtained by deleting the i-th observation from the full sample. Then B(θn-1,i) follows the same expansion as in (3) but with n replaced by n-1. Therefore − − θ − θ − =
+ −
(4)
and the delete-1 Jackknife estimator, = θ nJ
1 n ∑ [nθn − (n − 1)θn−1,i ], n i =1
(5)
will be unbiased up to first-order1. The term between brackets is referred to as a pseudovalue. The averaging over all possible pseudo-
Note that the expansion of the bias of the jackknifed estimator will generally have coefficients B2',...,Bj', say, that differ from those of θn. This is because the bias terms that are not removed by jackknifing are rescaled as the first-order bias of θn is estimated with error. Furthermore, jackknifing, at least in a cross-sectional setting, does not improve the mean squared error. 1
66
AENORM
62
January 2009
Econometrics
values is not required for bias reduction but is important for variance considerations. In the case of dependent data, (4) will generally not hold, but one may form a subsample of p < n consecutive observations where n = p+q and form an estimate of the bias as -p/q[θn-θp]. Tukey (1958) noticed Quenouille’s idea could be useful in the context of estimating V(θn) as well. He observed that pseudo-values could be treated as being approximately independent while having approximately the same variance as n1/2θn. Therefore the sample variation in the pseudo-values can be used to form the Jackknife variance estimator
ments grows. The Jackknife Instrumental Variables estimator (JIVE) (Angrist, Imbens and Krueger, 1999) uses Z-i and X-i, the matrices Z and X with the i-th observation deleted, to construct (11)
π − = − − −− −
for each i. As Z-i and X-i are independent of ξi and εi, this reduces the finite sample bias. JIVE then is
θ = ⊙ Π − − ⊙ Π −
(12)
"The potential of the Jackknife has been overlooked in the field of econometrics" n −1 n 1 n = VnJ (θ= [θ n −1,i − ∑ J 1 θ n −1, J ]2. ∑ i 1= n) n n
(6)
The idea of using pseudo-values for variance estimation bares similarity to the Bootstrap and, in fact, may be seen as an approximation to it (see e.g. Efron and Tibshirani, 1993). The delete-1 estimator is the most commonly encountered version of the Jackknife, but not the only one available. For example, Shao and Wu (1989) consider the delete-d Jackknife for non-smooth statistics such as the median, and Schucany, Gray and Owen (1971) propose an additional correction to θnJ t o achieve higher-order bias-reduction. Extensions to non-random sampling schemes and hybrid versions such as Jackknife-after-Bootstrap (Efron, 1992), have also been considered. Instrumental variables estimation Consider the classic endogeneous variables setup: Y = Xθ + ε, X = Zπ + ξ,
(7) (8)
where X may contain endogenous variables due to correlation between ε and ξ. Assuming X’Z to have full rank, the Two-Stage Least Squares estimator (based on i.i.d. data) is θn = [(Zπn)’X]-1[(Zπn)’Y], Zπn = Zπ + PZξ,
(9) (10)
where PZ = Z(Z’Z)-1Z’, and will be consistent if E[π’Z’ε] = 0. In finite samples however, Zπn and ε will be correlated, leading to bias in θn. The bias typically increases as the number of instru
with Πn-1 = (πn-1,1,..., πn-1,n)’. Whether JIVE actually leads to improved inference is still an unresolved issue. The interested reader is referred to Davidson and MacKinnon (2007) and the references therein. Nonlinear panel data models In applications with n x T panel data one often specifies a parametric model conditional on a fixed effect, representing time-constant unobserved heterogeneity amongst observations, and estimates θ by Maximum Likelihood (ML). Call this estimator θn,T. The ML estimator is, however, typically inconsistent for large and small T, the most appropriate asymptotics in most microeconometric panels. This is the incidental parameters problem (Neyman and Scott, 1948). While consistent for T → ∞, the limiting distribution of θn,T under double asymptotics is incorrectly centred, even if T grows as fast as n. Hahn and Newey (2004) noted that, for smooth models, the small T bias of θ∞,T = n →∞ θnT may be expanded in integer powers of T-1, i.e. Bθ∞T =
B B B + + jj + o j T T T T
(13)
Therefore, assuming the data to be i.i.d., the leading term in this expansion may be removed by applying the delete-1 Jackknife (as in (5) with n replaced by T) to obtain, say, θnPJT .Additionally, the removal of this term makes the asymptotic distribution of (nT)1/2( θnPJT -θ) centred around zero while retaining the efficiency of the ML estimator, even if T grows more slowly than n(T/ n3 → ∞ is required). As a consequence, finite sample inference is greatly improved if T is not too small. Additionally,
AENORM
62
January 2009
67
Kies je eigen weg
“ Don’t ask yourself if it’s a long road. Ask yourself if it’s a good journey.”
Sidney Poitier
Cijfermatige talenten op zoek naar een werkstudentschap Hewitt Associates is een wereldwijd opererende HRM-Consulting en Outsourcings-organisatie met zo'n 23.000 mensen in bijna veertig landen. In Nederland (350 collega's) helpen wij onze klanten met actuarieel advies, pensioenuitvoering en complete HRM-consultancy. Wij doen ons werk met passie, wat bij ons staat voor intellectuele uitdagingen, optimale kwaliteit en interessante klanten. Maar ook voor plezier in je werk, groei en een eigen koers. Wij zijn een bedrijf waarvan je mag verwachten dat het weet wat mensen beweegt in hun werk en wat ze in een carrière zoeken. Daarom vind je hier geen verhaal over targets en hoe wij tel kens weer weten die te bereiken. De weg erheen vinden wij veel belangrijker, omdat die het beste in mensen boven brengt. Bij Hewitt is dat een pad dat je in hoge mate zelf uitstippelt. En waar el ke bestemming een nieuw begin is. Aanbod. Hewitt is op zoek naar cijfermatige talenten, die tijdens hun studie 1 à 2 dagen per week praktijkervaring willen opdoen in het actuariële vakgebied. Bij ons vind je daarom ook de ruimte voor initiatief, een informele cultuur, continue uitdagingen en mogelijkheden om werken en studeren te combineren. Ook zijn er mogelijkheden om na je bachelor bij Hewitt in dienst te komen. Bij Hewitt krijg je de mogelijkheid om door te studeren tot actuaris. Niet alleen wordt deze studie volledig vergoed, ook krijg je jaarlijks extra studieverlof en kun je de studie in je eigen tempo doen! Wij willen graag in jou investeren, zodat je een goed beeld kan krijgen van het beroep actuaris en de organisatie. Bij ons mag je – sterker nog- moét je jezelf zijn, want pas dan haal je het beste uit jezelf en ben je in staat om je eigen koers te varen wanneer je bent afgestudeerd. Standplaats. Hewitt is gevestigd in Amsterdam, Eindhoven en Rotterdam. Op ieder kantoor is plaats voor 2 werkstudenten. Sollicitatieprocedure loopt gedurende het gehele jaar. Momenteel zijn wij voor kantoor Amsterdam op zoek naar een werkstudent. Interesse? Ben jij een vierdejaars student Wiskunde, Econometrie of Actuariële Wetenschappen en wil je naast studeren werkervaring opdoen binnen de actuariële dienstverlening, dan willen we je graag uitnodigen voor een kennismakingsgesprek om de mogelijkheden van werken en studeren met je te bespreken. Je kunt je curriculum vitae en brief met daarin een korte omschrijving van jezelf sturen naar Hewitt Associates B.V., Afdeling Human Resources, t.a.v. Laura Goeree, Postbus 12079, 1100 AB Amsterdam, of mail deze naar nlpz@hewitt.com. Voor meer informatie over het werkstudentschap, kan je contact opnemen met Laura Goeree, (Recruiter) 020-6609400.
Meer informatie over diverse functies en over werken bij Hewitt vind je op www.werkenbijhewitt.nl 68
AENORM
62
January 2009
Econometrics
Econometrics
as n is typically quite large, the mean squared error of θnPJT , E( θnPJT -θ)2 is much smaller than that of θn,T. Nonparametric estimation As a final application, imagine a two-step estimator, θn(πn) = θn,h, where πn is a kernel-based estimator with bandwidth h. This situation frequently arises in semi-parametric models with endogeneity. The convergence speed of θn,h depends on the rate at which h is allowed to shrink to zero. As a consequence, θn,h may not be root-n consistent and therefore n1/2(θn,h - θ) might not be centred at θh ≠ 0. This happens, for example, when undersmoothing is required (i.e. letting the bandwidth decrease only slowly with ). A common solution is to resort to higher-order kernels. An alternative approach has been given by Honore and Powell (2005), and is briefly described next. Assume that (θ∞,h - θ) is a sufficiently smooth function of h such that it may be expanded as θ∞h − θ =
B B B + + jj + o j h h h h
(14)
where j is such that the remainder term is o(n-1/2). Again a linear combination of estimators may be used to bias-reduce θn,h. Here the variability of the estimator as a function of h is exploited, not n. Let c1,...,ck be a chosen sequence of different constants. Then θn,cl h is an estimator based on bandwidth clh whose large n limit satisfies (14), with clh replacing h. Honoré and Powell’s (2005) bandwidth-driven Jackknife θnBJ,h then is obtained by forming a linear combination of such estimators which removes the first j bias terms from (14). Therefore, n1/2( θnBJ,h - θ) will be centred at zero and hence, θnBJ,h will be root-n consistent. References Angrist, J.D., Imbens, G.W. and Krueger, A.B. (1999). Jackknife instrumental variables estimation. Journal of Applied Econometrics, 14, 57-67.
Hahn, J. and Newey, W. K. (2004). Jackknife and analytical bias reduction for nonlinear panel models, Econometrica, 72, 12951319. Honoré, B. E. and Powell, J. L. (2005). Pairwise difference estimators for nonlinear models. In Andrews, D. W. and Stock, J. H., editors, Identification and Inference for Econometric Models, Essays in Honor of Thomas Rothenberg, 520-553, Cambridge University Press. Miller, R.G. (1971). The Jackknife - a review, Biometrika, 61,1-15. Neyman, J. and Scott, E. L. (1948). Consistent estimates based on partially consistent observations, Econometrica, 16, 1-32. Parr, W. C. and Schucany, W. R. (1980). The Jackknife: a bibliography, International Statistical Review, 48, 73-78. Quenouille, H. M. (1949). Approximate tests of correlation in time series. Journal of the Royal Statistical Society, Series B, 11, 6884. Quenouille, H. M. (1956). Notes on bias in estimation, Biometrika, 43, 353-360. Schucany, W. R., Gray, W. R. and Owen, D.B. (1971). On bias reduction in estimation, Journal of the American Statistical Associational, 66, 524-533. Shao, J. and Wu, C. (1989). A general theory for the Jackknife variance estimation, The Annals of Statistics, 17, 1176-1197. Tukey, J. W. (1958). Bias and condence in not-quite large samples, The Annals of Mathematical Statistics, 29, 614.
Davidson, R. and Mackinnon, J. G. (2007). Moments of IV and JIVE estimators. Econometrics Journal, 10, 541-553. Efron, B. (1992). Jackknife-after-Bootstrap standard errors and influence functions. Journal of the Royal Statistical Society, Series B, 54, 83-127. Efron, B. and Tibshirani, R. (1993). An Introduction to the Bootstrap, CRC Press.
AENORM
62
January 2009
69
Actuarial Sciences
The Power to Chase Purchasing Power One of the most important tasks of pension funds is to provide participants with a pension that is deprived of dilution due to inflation as much as possible. Within the context of a pension fund that compensates the inflation on a conditional basis, decreased purchasing power is not caused by inflation only, but rather by a combination of several factors. In this article, it is shown that an integral approach to the management of purchasing power risks can provide a highly improved outlook for pensioners without violating existing constraints
Jan-Willem Wijckmans studied econometrics in Groningen. Since 2003, he works at Cardano in Rotterdam as a risk management consultant. Based on quantitative research, he advises pension funds and insurance companies on sophisticated risk management solutions, with a focus on flexible tailored strategies. Cardano offers support for the entire risk management process, from research, advice, strategy implementation and control.
Inflation risks Ronald Reagan once said: “Inflation is as violent as a mugger, as frightening as an armed robber and as deadly as a hit man”. He was right. In Europe, economic policies all aim at keeping the level of price inflation at around 2%, a percentage deemed to be low and acceptable. However, over time even such a relative low inflation rate is capable of seriously diluting the amount of goods one can purchase. The greater the amount of time that goes by, the greater the dilution will be. The long-term character of pensions makes the inflation risk highly relevant in this area. A worker that builds up his pension over the course of 40 years has an average exposure to inflation of around 20 years. Over the course of this time period, he can lose around 30% of his purchasing power if inflation is on average 2%, but this can easily increase to over 50% if inflation averages at 4%. To put this figure into perspective: this means that due to inflation, someone who expects to receive a pension that is 80% of the average salary he has earned during his lifetime, effectively only receives 40% in real terms. Most pension funds therefore aim to compensate participants for inflation. Given the fact that the pension fund is a collective, the inflation risk for the pension fund as a whole is a lot smaller than for each individual separately. However, periods of high inflation will still con-
70
AENORM
62
January 2009
stitute a major risk factor for the fund. An increase in the expected inflation of 1% can cause the value of the expected real pension payouts to rise with 15% to 20%. To be able to actually make these additional payouts, the required return on the investments has to increase in line with inflation. Risk management of purchasing power The most common risk management tool that Dutch pension funds employ to mitigate this risk for the fund, is conditional indexation. Participants will receive indexation only when the nominal solvency position of the fund allows for it. If this is not the case, indexation is forfeited or postponed for better times. Although an individual is still better off under this system than on its own, indexation risks still exist in terms of purchasing power. If indexation is not received for several years in a row because of low solvency of the pension fund, the purchasing power of the participant can be highly impacted, especially if the low solvency situation coincides with high inflation. In order to minimize the probability and severity of purchasing power deficits, a pension fund can actively manage her indexation targets. For an average pension scheme, the most important drivers of purchasing power risks are: • Liability risk: as interest rates go down, the value of the pension liabilities that a pension fund accounts for will go up, decreasing the solvency; • Investment risk: as investments in for example equities and real estate (referred to as return assets) go down, the solvency also goes down; • Inflation risk: high inflations deplete the solvency buffers of a pension fund more rapidly. When solvency drops below the indexation boundary, high inflations accelerate the rate at which purchasing power decreases.
Actuarial Sciences
Probability of a drop in purchasing power by 20% or more 12% Reduced liability risk swaps 10%
Direction of efficieny loss
Reduced liability risk swaps/swaptions
8%
Basis
6% 4% 2%
Direction of efficieny gain
0% 0%
5%
10%
15% 20% Probability of solvency shortage
Figure 1
Ideally, all sources have to be targeted together in setting the investment guidelines for the fund. This is not a simple task, since all decisions have both a risk and return impact, on several important measures. A relative low equity allocation decreases the short term probability of a solvency shortage, but also impacts the possibilities for growth; simple hedging of all
portfolio is taken into account. The focus of the analysis is on the impact several policy changes can have on purchasing power. However, the nominal solvency risks are also important from a regulatory and continuation perspective. A solution that performs very well with respect to purchasing power, but introduces unacceptable high solvency risks is obviously not a practical and feasible solution. The analysis will therefore entail solution assessment on both solvency and purchasing power risks. The analysis is carried out on a 15 year horizon, although results are very much comparable for different time frames. Analysis From figure 1, the starting point (labelled ‘Basis’) shows that the probability of a solvency shortage (a funding ratio below 105%) is around 13% with a probability of a drop in purchasing power of over 20% amounting to 6%. For each
"Even an inflation rate of 2% can have serious impact over time."" nominal interest rate sensitivity decreases the immediate solvency risks, but also increases the exposure to high inflation rates; and finally, decreasing the exposure to inflation by converting the realised inflation to fixed inflation, can actually increase the probability of no indexation, because this increases the nominal funding ratio volatility. A good way to analyse the short and long term impact of different strategies is a scenario based Asset-Liability Model. The aim of such analysis is to define a practical and feasible strategy that minimizes relevant risks on different horizons. In the remainder of this article, the impact of various strategies on the funding ratio and purchasing power risks for pensioners will be discussed. Case study The analysis assumes an average-type Dutch pension fund in most policy respects: inflation compensation is granted only when the nominal funding ratio is greater than 120%, and the contribution rate is set to provide nominal service cost. The initial nominal funding ratio is set at 130%. In order to properly address the impact of investment risk, the percentage of equity investments in the initial asset mix is set to 50%, which is higher than that of the average Dutch pension fund. Also, no additional liability hedging apart from the existing bond
of the three mentioned risk drivers - liability risk, investment risk and inflation risk - the impact on purchasing power and solvency can be understood by varying the exposure of the fund towards one of the drivers individually. The liability risk exists, because the value of the liabilities is usually much more sensitive to changes in interest rate risk than the assets. To balance these effects, interest rate swaps can be used to align the asset sensitivity towards interest rates with the sensitivity of the liabilities. In figure 1, the allocation labelled ‘Reduced liability risk swaps’ shows the result. The probability of solvency shortage has decreased as expected, but purchasing power risks are actually higher. The main reason for this is the existing correlation between interest rates and inflation: high inflation rates are often accompanied by high interest rates. For a pension fund with lower interest rate sensitivities in the assets than in the liabilities, the higher expected payouts on the liabilities due to the high inflations can be ‘paid for’ through a lower value of the current liabilities. However, this is not the case when assets and liabilities are balanced in terms of interest rate sensitivity. So although the use of swaps lowers the current liability risk, it can actually increase the exposure in high inflation and high interest rates scenarios. A partial solution to this problem is not to use only swaps, but to incorporate swaptions (an option on a swap) in the liability hedge. With
AENORM
62
January 2009
71
Actuarial Sciences
Probability of a drop in purchasing power by 20% or more 12%
Direction of efficieny loss
10% Reduced liability risk swaps/swaptions
8%
Basis
6%
Reduced investment risk
4%
Combined
2%
Reduced inflation risk
Direction of efficieny gain
0% 0%
5%
10%
15% 20% Probability of solvency shortage
Figure 2
swaptions, one is protected for interest rates falls in a way similar to swaps, but one retains the upside in case of rising interest rates. It can be seen quite easily that this will lower the risk in the high inflation and high interest rate scenario described. The effect of using a combined strategy with both swaps and swaptions is shown with the allocation ‘Reduced liability risk swaps/swaptions’. Indeed the problem with increased purchasing power is lessened somewhat, although it cannot be entirely taken away. Next, we discuss the investment risk. This is lowered by physically moving away from return assets, in this case by as much as 20%. Although in most cases this can be done more efficiently by using options, for illustration purposes we have chosen to physically sell the return assets. Apparently (see allocation ‘Reduced investment risk’ in figure 2) this is a more attractive possibility than removing nominal interest rate risk, since it improves solvency risks, while leaving the purchasing power risks unchanged. It should be noted that again lowering the allocation to return assets will have less of an impact and purchasing power risk might even rise eventually. This is because return assets do not only have a higher risk than fixed income assets, but on average also bring a higher return needed to pay for inflation. A pension fund that already starts out with much less equity than the 50% chosen here will have less steering power with this instrument. However, it does illustrate the point about the impact of investment risk on both measures. Finally, we can reduce the exposure to inflation risk by using inflation swaps. These instruments allow a fund to receive the realised inflation that they need for participants in exchange for a fixed expected inflation rate, thus removing the uncertainty about future inflation levels. Figure 2 shows the results that reducing the inflation exposure serves a one-way purpose. The purchasing power risks are reduced substantially by over 60%. However, the price to pay is reflected in the solvency measure: a drastic increase in the probability of a solvency shortage. This phenomenon arises because inflation hedging is only useful when hedged for
72
AENORM
62
January 2009
all years to come. The value of all future inflation changes reflected in the inflation swap is however not shown on the liability side. This leads to increased volatility in the funding ratio with a higher risk of solvency shortage in those scenarios where inflation actually decreases. In those scenarios, the value of the inflation swap becomes negative, even though this is offset by the lower expected future inflation compensation on the liabilities. This result perfectly illustrates the risk management challenges when indexation is provided on a conditional basis. Now that the impact of the individual risk drivers has been properly assessed, it is time to show the possibilities when all three steering instruments are combined in a more tailored way. This is reflected by the ‘Combined’ allocation in Figure 2. No effort has been put in optimizing the result, since this is not the exercise currently of interest. The point of interest is that a sensible allocation of all risk drivers at hand can improve the outlook of a pension with fully retained purchasing power, without necessarily violating nominal constraints regarding the solvency. This is true in the relative simple case study shown here, but also in more complex pension cases, where other risk measures and possible return constraints have to be taken into account. Conclusion The indexation of pensions is a very important part of a pensioner’s income, especially when the inflation is high over a prolonged period. Within the set of policies of a pension fund that provides indexation on a conditional basis, three major drivers of risk that impact the purchasing power have been identified: liability risk, investment risk and inflation risk. It turns out that a standalone change in the exposure to one single risk driver may only have limited effects or may even lead to violation of conflicting constraints. However, a sensible optimization of all available exposures and risk management instruments can lead to both an improved outlook for the purchasing power of pensioners, as sustained policy results.
Do you want to take a look at all articles of Aenorm? Or are you searching for a particular subject?
www.Aenorm.eu Here you can find all articles ever published in Aenorm, use the database, subscribe and contact the editorial board.
AENORM
62
January 2009
73
Econometrics
The Dutch Business Cycle: Which Indicators Should We Monitor? With the credit crunch in the banking system and the stock market crash, it is just a matter of time until the unravelling financial turmoil is going to push the real economy into recession. What does that actually mean, being in a recession? And how can you measure and reasonably predict such a thing as a business cycle and its corresponding turning points? The business cycle indicator of the Durch central bank, De Nederlandsche Bank (DNB) extracts fluctuations from selected macroeconomic variables and represents them in a single index that describes the current stance of the business cycle. This article explains how the DNB business cycle indicator is constructed.
Ard den Reijer is currently employed at the central bank of Sweden, Sveriges Riksbank. He constructed the business cycle indicator while he was employed at the Dutch central bank, De Nederlandsche Bank1. The analysis is part of his PhD-thesis under supervision of prof. dr. L.H. Hoogduin (University of Amsterdam) and prof. dr. F.C. Palm (Maastricht University). See also his personal website; http://www.reijer.info
Various commercial, academic and government institutions use a business cycle indicator as an instrument to measure and predict business cycle developments and its turning points. An accurate assessment of the current and future state of the business cycle is valuable information in the decision process of policy makers and businesses. Most institutions, who regularly publish business cycle indicators, follow the approach of using leading and coincident indicators as developed at the National Bureau of Economic Research (NBER) in the US in the 1930s. Some institutions construct uniform business cycle indicators for several countries. Most notably, the American Conference Board and the Paris based Organisation of Economic Co-operation and Development (OECD) regularly publish their updated indicators for various countries. The OECD also runs an indicator that is specifically calibrated for the Dutch economy, as do the Netherlands Bureau for Economic Policy Analysis (CPB), Rabobank, Statistics Netherlands (CBS) and the universities of Groningen and Rotterdam. This article focuses on the business cycle indicator that is used by DNB and published on their website2.
The definition of a business cycle as formulated in the seminal contribution of Burns and Mitchell (1946) reads as follows: ”Business cycles are a type of fluctuation found in the aggregate economic activity of nations that organize their work mainly in business enterprises: a cycle consists of expansions occurring at about the same time in many economic activities, followed by similarly general recessions, contractions, and revivals which merge into the expansion phase of the next cycle; this sequence of changes is recurrent, but not periodic; in duration business cycles vary from more than one year to ten or twelve years; they are not divisible.” So, business cycles can broadly be defined as oscillating motions of economic activity, which are visible as patterns of fluctuations within macroeconomic variables such as output, production, interest rates, financial markets, consumption, unemployment and prices. The term ´cycle´ is misleading in the sense that it suggests a regular periodicity. Each single business cycle is unique and can be characterized by its depth, duration and diffusion. The depth of the cycle represents the cumulative rise (fall) of economic activity during the expansion (recession) phase. The duration of the cycle is the elapsed time between two peaks or troughs and varies from more than one year to ten or twelve years. The diffusion of the cycle represents the extent to which the business cycle is visible across different economic sectors and geographical areas. The DNB business cycle indicator presents the cyclical outlook in two single indices, as shown in figure 1. The coincident index of the DNB business cycle indicator summarizes the factual stance of the business cycle. The leading indi-
Views expressed are those of the individual author and do no necessarily reflect official positions of De Nederlandsche Bank. 2 See for the most recent update http://www.dnb.nl/en/z-onderzoek/auto88154.jsp 1
74
AENORM
62
January 2009
Econometrics
Grafiek1
DNB Business Cycle Indicator 2,0 1,5
Standardized deviation frrom trend
1,0 0,5 00 0,0 83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
0
1
2
3
4
5
6
7
8
9
-0,5 -1,0 -1,5 -2,0 -2,5 -3,0 Low Growth Phase
Composite Leading Indicator, May 2009
Composite Coincident Index, September 2008
Figure 1: DNB Business Cycle Indicator
mic time series variable, we can precisely extract the pre-specified business cycle fluctuations using the ideal band-pass filter defined as follows: Let 2 ≤ pl < pu < ∞ − ≥ = − = = = Then cyclet = ∑ = −∞ − ∞
Page 1
cator aims to replicate the coincident index and moreover, aims to project it up to six months ahead into the future. The indicator describes the business cycle as the deviation of economic activity from its trend level. The x-axis of the figure corresponds to the trend level and so, a positive (negative) index means that the level of economic activity lies above (below) its trend level. A downward (upward) sloping index means that the level of economic activity is growing slower (faster) than its potential growth, which is shown in figure 1 by the shaded areas. Following the peak in the business cycle, four subsequent phases can be distinguished as: i) slow-down; the cycle is positive, but downward sloping; ii) recession: the cycle is negative and downward sloping, which ends in a business cycle trough; iii) recovery: the cycle is negative, but upward sloping; iv) boom: the cycle is positive and upward sloping. The DNB business cycle indicator shows in Figure 1 that the Dutch economy is currently in the slow-down phase and will be entering the recession phase on short notice. Measuring the business cycle The DNB business cycle indicator is based on the notion of a business cycle as all the intrinsic cyclical motion visible in macroeconomic data consisting of waves within a specified frequency interval. This interval of business cycle frequencies corresponds with Burns and Mitchell’s (1946) taxonomy of business cycles as waves lasting longer than a pre-specified minimum duration and shorter than a pre-specified maximum duration. Business cycles at certain frequencies can be isolated by so called band-pass filters. This type of filters originates from electrical engineering and operates on a time series of observations for an economic variable like tuning a radio: some frequency ranges are eliminated and other frequency ranges, i.e. the business cycles, get passed through. In an ideal world in which we have an infinite amount of observations ∞=−∞ for an econo-
So, the cyclical fluctuation labelled cyclet consists of all the cyclical fluctuations incorporated in the time series variable xt whose waves are characterized by a duration longer than pl periods and shorter than pu periods. The DNB business cycle indicator, which is based on monthly observations, operationalizes 36 months, i.e. 3 years, as a lower bound and 132 months, i.e. 11 years, as an upper bound. Since we are not living in an ideal world, but in a real one with a limited amount of available observations, the econometric question is how to approximate the ideal band pass filter? Approximate band pass filters suffer from leakage, i.e. some frequency ranges that are supposed to be filtered out get passed through to some extent and compression, i.e. some frequency ranges that are supposed to pass through get dampened to some extent. Christiano and Fitzgerald (2003) (CF) propose an approximate band pass filter that is optimal in a weighted mean squared error sense conditional on a special characteristic that is typical for macroeconomic time series3. In the real world, we have a finite amount of T observations = . Then t, t=1,...,T with f=T-t+1, p=t-1 the coefficients of the CF-filter look like: = = − + − − = − ∑ − = − ∑ − Then =
∑
=−
−
Note that the filter coefficients converge to their ideal band pass filter equivalents in case we have a lot of observations at our disposal. So, → as p,f → ∞ Which indicators should we monitor? Real gross domestic product (GDP) as the aggregate of all economic activity constitutes an important statistic for business cycle measurement, the American and European business cycle dating committees4, however, monitor se-
More precisely, Christiano and Fitzgerald (2003) assume that the underlying economic time series variable follows a random walk process, i.e. that it is integrated to the first order: a so-called I(1) process. This type of process captures the trending behaviour that is typically observed in macroeconomic data. 4 See the websites of the American National Bureau of Economic Research (NBER) and the European Centre for Economic Policy Research (CEPR). 3
AENORM
62
January 2009
75
76
AENORM
62
January 2009
Econometrics
veral macroeconomic variables as the cyclical fluctuations of GDP do not always move synchronously with the ones of its underlying components. It is in general infeasible to identify one single variable, which covers a broad range of economic activity, represents the current stage of the business cycle and is available on a monthly basis. The coincident reference index is therefore constructed as a composite index of several synchronous indicators. The index constitutes a more robust standard of business cycle measurement since the idiosyncrasies of the individual series are averaged out. In addition to reflecting the cyclical properties of GDP, other important practical criteria for selecting the variables are a limited publication delay and minor data revisions. This means that new data releases are published shortly after the corresponding period has ended and that the initially published values are subject to only minor revisions during subsequent publications regarding the same period. As a result, the coincident index of the DNB business cycle indicator is based
relative to the share of services production. Despite this small share, the OECDâ&#x20AC;&#x2122;s indices are solely based on industrial production as its cyclical motion is considered a representative statistic for the business cycle. The staffing employment market is the segment of the labour market that is most sensitive to business cycle motions, because companies can adjust their use of staffing services immediately to changing market conditions. Cyclical motion in the economy is reflected in various macroeconomic time series in a dynamic sense as well. Depending on the timing behaviour of their cyclicality vis-Ă -vis the coincident index, variables can be classified as leading, coincident or lagging. Leading variables give by nature an early signal on the cyclical position. Economic plausibility requires that the leading indicators should be supported by economic theory either as possible causes of business cycles or as quickly reacting to positive or negative shocks. Since economic fluctuations originate from different sources, we combine a
"The Dutch economy is currently in the slowdown phase and will be entering the recession phase on short notice" on industrial production, household consumption and staffing employment. The volume index of consumption by households is closely related to retail sales, since it consists only of those expenditures for which households pay themselves. Industrial production constitutes approximately 25% of GDP and its share is declining Leading indicator variable
Lead time (in months)
IFO-indicator Germany
9
Expected business activity
9
Three month interest rate
26
Term structure of interest rates
30
Stock finished products
10
Order arrivals
15
Consumer Confidence
11
Stock Price Index (AEX)
7
Registered motor vehicles
9
Real house price
7
OECD leading indicator for the U.S.
18
Table 1 index
Composition of the leading
number of relevant leading indicators into a single composite index. The main selection criteria for the leading indicators are the lead time and the statistical conformity with the business cycle as represented by the composite coincident index. The leading indicator should show consistent timing by systematically anticipating peaks and troughs with a rather constant lead time and avoiding missed turning points or false signals. A missed turning point is a turning point in the coincident index, which the leading indicator fails to signal. A false signal occurs when the leading indicator shows a turning point that eventually does not materialize in the coincident index. The leading index of the DNB business cycle indicator consists of eleven selected leading indicator variables that together make up a balanced representation of near future cyclical developments. As shown in table 1, the composite leading indicator consists of three financial series, four business and consumer survey results and four real activity variables, of which two supply and two demand related. The three financial variables are the short term interest rate, the yield curve and the stock price index. Low short term interest rates reduce financing costs and will spur investment
AENORM
62
January 2009
77
Econometrics
demand. The AEX stock price index consists of the 25 most actively traded shares in the Netherlands and indicates the expected future corporate profitability and the underlying growth potential. An inverse yield curve is usually observed at the start of a recession period and acts therefore as a predictor. The four business and consumer survey results are expected business activity, the IFO-indicator of future expectations, domestic consumer confidence and the OECD´s leading indicator for the Unites States. The IFO-indicator represents the economic expectations of producers in Germany, the largest trading partner of the Netherlands. The OECD´s leading indicator for the U.S. reflects the short term outlook of the world´s largest economy, whose business cycle is known to be leading for the G7 countries. Finally, two supply related real activity variables are the order arrivals and the stock of finished products and two consumption related real activity variables are the real house price and the registrations of new cars. Housing wealth, and therefore real house price changes, are an important factor in the total wealth of consumers. The registration of new cars is a variable that quickly reacts to alterations in the business cycle. Summary Every month the composite coincident index shows the current state of the economy on the basis of industrial production, household consumption and employment. These three variables are important reference points for dating economic turning points. The composite leading index provides an estimate of the state of the economy in the near future. The leading index is based on three financial series, four business and consumer survey results and four real activity variables, of which two supply and two demand related. Referring to the four quadrant characterisation of business cycle phases, i.e. slow-down, recession, recovery and boom, the DNB business cycle indicator shows that the Dutch economy is currently in the slow-down phase and will be entering the recession phase on short notice. References Burns, A. and Mitchell, W. (1946). Measuring Business Cycles, National Bureau of Economic Research. Christiano, L. and Fitzgerald, T. (2003). The Band Pass Filter, International Economic Review, 44. Den Reijer, A.H.J., (2006). The Dutch Business Cycle: Which Indicators Should We Monitor?, DNB Working Paper no.100
78
AENORM
62
January 2009
Puzzle
Puzzle The puzzles in last edition were probably a bit too hard for most of the readers, since we received no right answers at all. Therefore, the puzzles in this edition are more straightforward. But first, the solutions to the puzzles of last edition. American election In the election puzzle, add the pluralities to the total vote and divide by the number of candidates ((5,219 + 22 + 30 + 73) / 4). The quotient will be the vote of the successful one, from which the votes of the others can be ascertained by subtraction. The counts were 1,336, 1,314, 1,306 and 1,263.
Illustrated above are Alice and the cat and a square containing the sentence “Was it a cat I saw?” multiple times. In how many different ways can you read this sentence in the above square? Start at any of the W’s, spell by moving to adjacent letters until you reach the C, then spell back out to the border again. You may move up and down, left and right. The boxer’s puzzle
A Chinese switch-word puzzle A possible solution to this puzzle is the word ‘interpreting’. The new puzzles: A popular game among young students is the boxer’s puzzle. The goal of this 2-player game is to get as many boxes as possible by drawing a straight line between two points. A possible state of this game is illustrated above. In every turn each player has to draw a straight line between two points (for example between A and B). When a player completes a square (there are for example lines between the points A, B, E and F), this box is won by this player and the same player has to draw another line. Now we are in the state of the illustration above and it’s your turn to draw a line. If you connect M to N your opponent could score four boxes in one run, then having the right to one more play, he would connect H and L which would win all the rest to him. What play would now be best and how many boxes will it win against the best possible play of your opponent?
Was it a cat I saw One of the most remarkable stories you may have read during your youth has to be Alice in Wonderland. During her journey through Wonderland, Alice comes along the Cheshire cat, which has a way of vanishing into thin air until nothing but its irresistible smile remains.
Solutions Solutions to the two puzzles above can be submitted up to December 1st. You can hand them in at the VSAE room; C6.06, mail them to info@ vsae.nl or send them to VSAE, for the attention of Aenorm puzzle 59, Roetersstraat 11, 1018 WB Amsterdam, Holland. Among the correct submissions, one book token will be won. Solutions can be both in English as in Dutch.
AENORM
62
January 2009
79
Facultive
VU University Amsterdam
University of Amsterdam The VSAE has hosted a number of projects in the last months of 2008. In November, the KBR (short foreign journey) brought 52 students in Actuarial Sciences, Econometrics and Operations Research to Prague. During our five day visit we saw the University and Czech National Bank, gazed at Prague’s castle and historic centre and enjoyed its nightlife. On the 17th of December, the VSAE organised the ´Actuariaatcongres´ - a congress on actuarial sciences -, which had Transparency of Insurances as its central theme. A number of speakers shared their view on the subject with an audience of 160 participants, consisting of both students and actuaries. A discussion panel concluded the congress, sending the participants home with numerous ideas about the role of actuaries in making insurances transparent. Since January is the exam month at the University of Amsterdam, the agenda of the VSAE includes only our monthly free drink and curling. From February onward a new board will try to bring the VSAE to even greater heights. Hopefully they succeed in doing so! Agenda 13 January Monthly free drink 22 January Curling
The past few months have been a great success for the study association Kraket. Many well organized activities, with a lot of participants. In October we organized a squash tournament, with many enthusiastic member of our association. In November Saen Options organized a Inhouse Day for our study association, with a trading game and a tour at the trading ground. On the 26th of November we went bowling with our fellow students of the VSAE, with over forty members of both Kraket and the VSAE. And last but not least we welcomed Saint Nicholas to our association. In a pleasant atmosphere the members of our association received presents and also a little poem. In the upcoming months we have planned some other activities. In January we will go ice-skating with our association, and the study trip to New York will also take place in January. In February we have planned an indoor soccer tournament with our association, and last but not least there is the LED (National Econometrics Day) in Groningen to look forward to. Agenda 7-24 January 2009 Study Trip 20 January 2009 Ice-Skating February 2009 Indoor Soccer Tournament 3 March 2009 LED 9 April 2009 Casedag Kraket
80
AENORM
62
January 2009
Excellent enough to make a difference?
(Aankomende) actuarissen – M/V Bij Ernst & Young werken accountants, belastingadviseurs en gespecialiseerde adviseurs in vrijwel elke richting van de financiële, juridische en notariële dienstverlening. Met ruim 4.600 medewerkers verspreid over 23 kantoren zijn we een van de grootste adviesorganisaties in Nederland. Er is bij ons altijd ruimte voor ambitieuze starters die het verschil willen maken. Ga jij de financieel directeur van een verzekeraar adviseren over een nieuwe premiestructuur of het toeslagenbeleid van een pensioenfonds toetsen? Of ben je meer geïnteresseerd in de modellen die beleggingsfondsen gebruiken bij de waardering van hun portfolio’s. Ernst & Young Actuarissen biedt de mogelijkheid om je breed te ontwikkelen op zowel inhoudelijk als persoonlijk vlak. Wij zijn onderdeel van een Europese organisatie waarin actuarissen uit alle financiële centra van Europa samenwerken. In de dynamische markt waarin wij werken hebben wij continu nieuwe adviseurs nodig. Voor onze kantoren in Amsterdam en Utrecht zijn wij op zoek naar (Aankomende) actuarissen. Ben je gedreven, leergierig, analytisch en adviseer je liever dan dat je wordt geadviseerd. Dan heb jij de instelling die wij zoeken. Wij bieden jou de mogelijkheid om na je studie als beginnend actuaris aan de slag te gaan. Ook kun je tijdens je studie voor een dag in de week aan de slag gaan als werkstudent of je afstudeerscriptie schrijven over een praktijk probleem.
Voor meer informatie kun je contact opnemen met Dieuwertje Huizer, 06-21252814 of dieuwertje.huizer@nl.ey.com. Solliciteren kan op www.ey.nl/carriere
RECRUITMENT
To some people, it’s just a job To others, it’s reaching your potential Consultancy opportunities at Towers Perrin Wil jij ook aan de slag in een uitdagende en prettige werkomgeving? Denk dan eens na over een toekomst bij Towers Perrin. Op dit moment zoeken we starters (econometristen en actuarissen) voor onze kantoren in Amsterdam, Rotterdam en Apeldoorn.
Of je nu internationale carrièrewensen hebt of het accent wilt leggen op een goede balans tussen privé en werk, bij Towers Perrin krijg je de ruimte. Als een van’s werelds grootste adviesorganisaties voor people, risk en financial management, vinden we het belangrijk dat mensen zich kunnen ontplooien op een manier die bij hen past. Dé Towers Perrin-consultant bestaat dan ook niet. Wat we wel gemeenschappelijk hebben is onze gedrevenheid en liefde voor het vak.
Wil je meer weten?
http://careers.towersperrin.com
Ga naar http://careers.towersperrin.com Voor vragen kun je terecht bij Philip Le Poole. Tel. 020 - 711 40 16 of mail: recruitment-nl@towersperrin.com