Journal of Risk Model Validation
ISSN:
1753-9579 (print)
1753-9587 (online)
Editor-in-chief: Steve Satchell
Need to know
- Discussion: what is a model?
- Analysis of the current regulation regarding Model Risk
- Discussion related to the root causes of Model Risk
- Proposition of corporate governance to address Model Risk
Abstract
One of the main concerns and regulatory topics that financial institutions have to deal with is model risk. Senior managers are increasingly considering model risk to be one of the most severe vulnerabilities to which a financial institution can be exposed. While the concept seems relatively simple, the definition of model risk (both traditional and regulatory), its origins (from dogma to misuse) and ways to manage it (from engineering conservatism into models to having a proper governance process) are not necessarily handled well by practitioners, academics and regulators. Having a clear definition of model failure and understanding its root causes allow us to adopt the appropriate management style to deal with potential issues that could lead to dramatic failures. In this paper, we conduct an analysis of model risk in an attempt to understand the main issues that lead to failures and the best way to address such issues.
Introduction
1 Introduction
“Ipse se nihil scire id unum sciat. [I know that I know nothing.] — Socrates”
Before discussing model risk in detail, it seems sensible to recall what a model is. According to the Oxford English Dictionary, a model is “a simplified description, especially a mathematical one, of a system or process, to assist calculations and predictions”.11 1 See http://www.lexico.com/en/definition/model. In this paper, when the term “model” is used, the definition that has just been given should be borne in mind.
Therefore, models are representations of reality and can only ever be an approximation. They reduce the universe to a smaller space, helping developers to understand patterns and the interactions between various elements (Hassani 2015). The definition of “model” itself implies various risks, as discussed in the following.
In this paper, we start with epistemology and go one step further into an analysis of the root causes of these risks (some of these are described in the remarkable books of Rosch and Scheule (2010), Morini (2011) and Tunaru (2015)). In addition, we integrate the outcomes of the latest pieces of regulation (ie, 2018) and put forward clear guidelines for governance to tackle all of the issues identified. Therefore, our work complements that in a recent, remarkable article by Bennett (2017).
1.1 Concept
To paraphrase a statement by pro-gun lobbyists: “guns don’t kill people, people kill people”. Despite the provocative nature of this sentence, we can draw a parallel between it and model risk: that is, models are only dangerous in people’s hands.22 2 Note that the developer is often the end user. A misplaced dependence or overreliance on a particular model – or the ego of the person conceiving it – may lead to dramatic failures.
The Federal Reserve (2011) and, more recently, the European Banking Authority (EBA) have officially addressed model risk, a topic that has been discussed by practitioners for a while but has only been dealt with partially. Indeed, in Article 1 of the EBA Regulatory Technical Standards (RTS) (European Banking Authority 2014a), the regulator includes model risk within the scope of operational risk, as can be seen in the following statement: “The competent authority shall verify that an institution has included legal risk, information and communication technology risks, as well as model risk, within the scope of operational risk with the exclusion of other kinds of risk.” The fact that model risk (Morini 2011) is included “within the scope of operational risk” is not surprising, considering that operational risk is defined as “the risk of direct or indirect loss resulting from inadequate or failed internal processes, people and systems or from external events”. Operational risk includes legal risks (as well as various boundary risks) but excludes reputational risk, and it is embedded in all banking products and activities (Basel Committee on Banking Supervision 2001b). Indeed, the failure of a model is included in the definition above, as model risk could arise from people, processes or inadequate systems or methodologies, as described in what follows.
Although authorities recognize model risk as a major problem, the way it is handled may differ from one regulator to the next. Indeed, while both the Federal Reserve (2011) and the EBA RTS (European Banking Authority 2014a) tend to cover a similar scope, the latter focuses on unexpected issues related to choosing the wrong model as well as unidentified limitations. Indeed, model risk arises when all of the characteristics, properties and assumptions of a model are unknown, not understood or voluntarily disregarded; in other words, when a model is used blindly. However, this approach is less devastating when compared with someone making an impulsive judgment call. We advocate that a scenario in which a simple model is used is still better than one in which no model is adopted. Therefore, a real and dramatic risk that any institution can face is misunderstanding the domain within which a model is valid. We will not hide behind a truism such as “all models are wrong but some are useful” (Box and Norman 1987). First, saying that a model is “wrong” is nothing more than a tautology, because a model is an approximation of reality: it is not reality itself. Second, all models are useful; even if a model is not applicable in a given scenario, it can still be useful in deducing that the phenomena to be captured are different from those for which the model has been designed. In other words, we can discover, using the process of elimination, that a model is simply not applicable in a given situation. Applying a model forcibly when we know or suspect that it is not applicable is wrong, because it introduces a bias in the perception of the pattern we are trying to highlight; this may prove misleading to users and the model’s outcomes may be fallacious.
Model risk is far more abstract than its description as a quantitative problem would suggest. It requires us to acknowledge that our quantitative methodologies, our understanding of the universe and our reduction of the gap between academic papers and their implementation in financial institutions are currently limited. This problem starts with mathematics itself. By analyzing the opposing views of foundationalists (mathematics has a permanent, secure foundation), absolutists (mathematical knowledge and truth are eternal, objective and absolute) and those espousing new schools of thought – such as fallibilists, humanists, relativists and social constructivists – we can observe that mathematics, as a representation of conceptual structures, is perpetually evolving. In addition, according to Paul Ernest (2004), mathematical knowledge is invented, not discovered. Although that statement may seem controversial, it is, in our opinion, sensible to consider that models are based on assumptions: depending on the underlying assumptions, any kind of model can be created, even if some might be absurd with respect to reality. Indeed, a clear distinction has to be made between theory and practice, that is, the theories of mathematics and their implementation (eg, a risk measure such as value-at-risk (VaR) or expected shortfall (ES) and its estimator). Consequently, even if a model is meaningless with respect to its implementation in reality, nothing prevents the theories of mathematics behind it from being written down. It is simply the responsibility of users and practitioners not to use such models if they are deemed inappropriate. In the meantime, we need to acknowledge the possibility that the theories of mathematics that have been created are representative of an absolute truth that has not yet materialized, has not been captured by data and, consequently, is not empirically verifiable. Going even further, Tegmark’s mathematical universe hypothesis (MUH) (Tegmark 2008) states that the physical universe is mathematics, ie, physical existence equals mathematical existence and vice versa.
1.2 Regulatory definition
Although the essence of this concept seems identical, regulators around the world have different ideas about how model risk should be managed or covered. For example, US and Polish regulators require model errors or uncertainty to be covered by financial institutions’ regulatory capital. The Federal Reserve defines model risk as follows (Federal Reserve 2011): “The use of models invariably presents model risk, which is the potential for adverse consequences from decisions based on incorrect or misused model outputs and reports. Model risk can lead to financial loss, poor business and strategic decision making, or damage to a bank’s reputation. Model risk occurs primarily for two reasons:
- •
The model may have fundamental errors and may produce inaccurate outputs when viewed against the design objective and intended business uses. The mathematical calculation and quantification exercise underlying any model generally involves application of theory, choice of sample design and numerical routines, selection of inputs and estimation, and implementation in information systems. Errors can occur at any point from design through implementation. In addition, shortcuts, simplifications, or approximations used to manage complicated problems could compromise the integrity and reliability of outputs from those calculations. Finally, the quality of model outputs depends on the quality of input data and assumptions, and errors in inputs or incorrect assumptions will lead to inaccurate outputs.
- •
The model may be used incorrectly or inappropriately. Even a fundamentally sound model producing accurate outputs consistent with the design objective of the model may exhibit high model risk if it is misapplied or misused. Models by their nature are simplifications of reality, and real-world events may prove those simplifications inappropriate. This is even more of a concern if a model is used outside the environment for which it was designed. Banks may do this intentionally as they apply existing models to new products or markets, or inadvertently as market conditions or customer behavior changes. Decision makers need to understand the limitations of a model to avoid using it in ways that are not consistent with the original intent. Limitations come in part from weaknesses in the model due to its various shortcomings, approximations, and uncertainties. Limitations are also a consequence of assumptions underlying a model that may restrict the scope to a limited set of specific circumstances and situations.”
Moreover, it states the following: “Another way in which banks may choose to be conservative is to hold an additional cushion of capital to protect against potential losses associated with model risk.” The European Banking Authority (2014a) is reducing the scope of model risk by acknowledging some kind of institutional “free will”, as follows: “As a specification of the paragraph 5(4), the following events, and the related losses, shall be excluded from the scope of operational risk: (a) events due to wrong selection of a model, made through a formalized corporate process; and (b) losses caused by a pricing model where the potential exposure to the model risk had been previously assessed.” Therefore, we can differentiate between expected model risk and unexpected model risk: the former should be managed and the latter should be offset, controlled, hedged or covered by a capital buffer. This differentiation is really interesting, as it recognizes the fact that the risk of a model failing is intrinsic to its use; this contrasts with US regulation, which tends to imply that a “good” model should not fail at all. The capture of a loss related to unexpected model failure in capital calculations should be an incentive for risk managers to understand the models they use, to test these abstractions and to ensure the models have been properly challenged and appropriately monitored. This piece of regulation implies that model risk cannot be fully offset. A model could fail for various reasons, and such risk is inherent in the use of any model. This acknowledgement seems to be a sensible way forward, as awareness is the first step toward sound risk management.
2 Root causes of failure
In this section, we discuss the root causes of model failure from both a philosophical and a practical point of view.
2.1 Dogmas
As described in Collins Dictionary,33 3 See http://www.collinsdictionary.com/dictionary/english/dogma. the term “dogma” comes from the Greek word “dokein”, which means opinion or belief. It is defined as follows:
- (1)
a religious doctrine or system of doctrines proclaimed by ecclesiastical authority as true; or
- (2)
a belief, principle, or doctrine or a code of beliefs, principles, or doctrines.
Besides, the first synonym provided is “blind faith”.
This definition has been extended to principles laid down by authorities whose opinions are considered to be incontrovertibly true, such as those of philosophical schools of thought and political parties. As a dogma is the basis and the cornerstone of an ideology, altering or failing to consider it would mechanically affect either a system’s paradigm or the ideology itself. Moreover, an ideology should always be considered a dangerous thing, as it is a utopic goal that can be used to justify the most vicious behavior exhibited to achieve it (Pinker 2012; Fiske and Rai 2014). Another famous definition of dogma was provided by Steve Jobs at the University of Stanford in 2005. He defined dogma as “living with the results of other people’s thinking” and, therefore, as a break from innovation and creativity.
The concept of dogma is intrinsically related to the concept of faith. We are not trying to be controversial or provocative when we say that faith may not be compatible with questioning, as it relies on a belief that is anchored deep inside one’s being. When a model becomes a dogma, people use it without questioning its legitimacy or soundness. Unfortunately, when most practitioners are blindly using the same models, their use quickly becomes so-called best practice. It is interesting to note that regulatory expectations (with respect to models) mechanically catalyze and spread dogmas. Relying on dogmas may even affect market microstructure itself (price formation, volatility, etc) (O’Hara 1998) and engender systemic risk.
The main problem with a dogma is that it does not accept questioning. It is profoundly anchored in our culture and the way we do and understand things, and it structures the way we think and approach problems. As an illustration, Euclidian axioms (Euclid 1956) may be considered dogmas as soon as they are accepted without questioning. Consequently, the knowledge that an alternative theory such as non-Euclidian geometry (Bonola 2007) exists leads to the use of Euclidian principles in inappropriate situations (eg, cosmology), which may lead to dramatic failures. As a case in point, we quote Nin (1972), who said: “When we blindly adopt a religion, a political system, a literary dogma, we become automatons.”
2.2 Paradigm shift
The proper definition of a paradigm shift, as stated by Thomas Kuhn (2012), is an alteration of the basic set of assumptions within “the ruling theory of science”. According to Kuhn, “[a] paradigm is what members of a scientific community, and they alone, share” (Kuhn 1979). It is noteworthy that Kuhn restricted his use of the word to “hard” sciences (ie, natural sciences as opposed to social sciences).
The main issue with a paradigm shift is that it may not be expected; consequently, a model may be rendered completely obsolete, leading to its use becoming significantly risky. It is interesting to note that self-fulfilling prophecies usually go hand in hand with paradigm shifts. Indeed, if everyone uses the same pricing strategy (eg, Black and Scholes for options, or the same discount factor for a range of other assets), then everyone engaged in the same business knows the others are using it too; as a result, prices tend to converge and products are evaluated in line with what the ruling pricing model says they should be. Consequently, when a paradigm shifts, the outcomes of the models that relied on the previous paradigm are not valid anymore with respect to the new paradigm. So, a paradigm shift may trigger the materialization of significant model risk. Research in many fields has been undertaken to detect paradigm shifts, as these may have a tremendous impact on our perceptions. To illustrate this, we can refer to medical science, a domain in which some research has proposed alternative ways of detecting cancers by predicting mutations (Ng et al 2012), or to seismology and its stress-forecasting approach, which advocates evaluating the likelihood of an earthquake as opposed to predicting it (Crampin et al 2008). A change in underlying paradigms may be considered a revolution.
While addressing paradigm shifts, it is necessary to discuss the difference between risk and uncertainty, as analyzed by Knight (1921). According to Knight, risk occurs when we do not know the outcome of a given situation but can accurately measure the odds. Uncertainty, on the other hand, occurs in situations where we cannot know all of the information we need to set accurate odds in the first place. The term acknowledges some fundamental degree of ignorance, a limit to knowledge and an essential unpredictability of future events. The concept of a black swan event, developed by Nassim Taleb (2007), is related to Knightian uncertainty.44 4 Please note that this theory is not unanimously accepted; here, it merely gives us the keys to understanding a particular situation. This statement is more or less valid for all of the theories discussed or presented in this paper. A black swan is an important and inherently unpredictable event that, once it has occurred, is rationalized with the benefit of hindsight. It is used to describe historical developments such as the widespread adoption of the personal computer, something entirely impossible to predict that nevertheless had world-changing effects. Another feature of the black swan theory is that appropriate preparation for such events is frequently hindered by the pretence of knowledge of all the risks. In other words, Knightian uncertainty is presumed not to exist in day-to-day affairs, often with disastrous consequences. These concepts are consistent with the idea of self-fulfilling prophecies going hand in hand with paradigm shifts, as stated in the previous paragraph.
While dealing with financial models, a paradigm shift that remains undetected may result in fallacious asset prices, risk measures or investment decisions. Some of the biggest operational risk losses due to model failures are because of a paradigm shift. However, as illustrated by Planck (1968), societies are reluctant to recognize paradigm shifts as they may not be fully understood and therefore may trigger a reaction of fear or rejection. According to Planck (1968): “A new scientific truth does not triumph by convincing its opponents and making them see the light, but rather because its opponents eventually die, and a new generation grows up that is familiar with it.”
The Gaussian copula used to evaluate collateralized debt obligations (CDOs) failing because of a change in market conditions is a typical example of an alteration of a basic set of assumptions, as it meant that dependencies between tranches were no longer linear but characterized by upper tails. Translated into a paradigm shift, the ruling theory – which supported that dependencies were linear and that nonlinearity was just an anomaly the market would correct – became that dependencies were nonlinear and linear dependencies were just a particular case. Another example is the zero weight that was associated with G10 debt in the risk-weighted asset metric used to calculate Basel ratios, even though corporate debt was nonzero. This led to strange bond acquisition strategies (Lenarčič et al 2016; Acharya et al 2016). In that case, the regulation itself engendered the paradigm shift.
2.3 Erroneous or strong assumptions
Erroneous assumptions or, more frequently, strong assumptions (even if models are very robust with respect to their assumptions) are the third potential issue that could lead to model failure. This is the issue we have seen with Gaussian yields, the market portfolio of the capital asset pricing model (CAPM) (Sharpe 1964; Roll 1977), the constant volatility in the Black–Scholes formula (Black and Scholes 1971) and the absence of arbitrage opportunity or market completeness (Flood 1991). Most assumptions are hardly sustainable and may not even have been proved right. However, the most destructive meta-hypothesis widely found in the infinite set model is the assumption of the stationarity of processes (Priestley 1983).
Let us recall, as a first step, the definition of a stationary process: this is a stochastic process whose joint probability distribution does not change over time. As a corollary, the mean and variance of this process, assuming they exist, neither follow a trend nor change over time. Obviously, if processes are stationary, they have usually only remained in this state for a short period of time.55 5 We acknowledge that the term “short” is relative; therefore, we define a time frame as that considered for the data used in the model. This is because by merely considering economic cycles, revolutions or human actions it is difficult to see this assumption as permanently reliable. For example, fraud data (eg, from credit card fraud) is not usually stationary, as every time fraud is observed a control is put in place, and therefore fraudsters’ behavior changes and adapts (Dal Pozzolo and Bontempi 2015). Consequently, past data is not necessarily informative in modeling, analyzing or predicting new fraudsters’ behavior. Unfortunately, the stationarity hypothesis (although it is not always inaccurate and therefore should not be discarded a priori) is the very assumption that leads practitioners to use constant volatility or, more generally, to use past data to predict the future.
Sometimes assumptions are made to ease the construction of a theoretical framework and to simplify calculations; however, is the resulting model usable in practice, or is it just a representation of academic progress? This paragraph is a rejoinder to the Ernest (2004) statement provided in the introduction: mathematics has not been discovered, it has been invented. Therefore, a mathematical model may not even be close to reality and may only be used to improve our understanding of some patterns. In some situations, what we learn from implementing a model is more important than its outcomes.
It is important to avoid misplaced reliance on a “one size fits all” kind of model. In this regard, we quote Maslow (1962), who said: “I suppose it is tempting, if the only tool you have is a hammer, to treat everything as if it were a nail.”
2.4 Domain of definition/application
The term “domain of definition” is most commonly used to describe the set of values for which a function (map, transformation, etc) is defined. For example, a mathematical expression that is defined for real values has a domain . The set of acceptable values that is returned by the function is called the range. In a larger sense, the domain is nothing more than a delimited area, that is, a zone in which a model can behave as expected or in the most appropriate manner.
Popper (1963, 1992) provided us with the best theorization of the limit of the domain of application. Models, behaviors, patterns and interactions are handled, understood and tested for “corroboration” and their outcomes discussed with respect to certain assumptions. This domain of application is a general concept, and, for most models, it is usually highly limited. A model is only valid under some particular conditions. A failure in the model may be caused by the fact that the model is just not applicable in a particular situation. This does not mean that the model is irrelevant; it just means that the model is not suitable in a given context. We can illustrate this point with an absurd metaphor: no one would use a 12 mm spanner on a 5 mm bolt, as it simply would not work. Similarly, distributed methodologies such as the loss distribution approach (Frachot et al 2001) in operational risk capital calculation are only reliable as long as the data is not autocorrelated (Guégan and Hassani 2013b), which happens quite often, but not always.
This section can be summarized by quoting Einstein (2011), who said: “Occurrences in this domain are beyond the reach of exact prediction because of the variety of factors in operation, not because of any lack of order in nature.”
2.5 Model misuse
This is a typical Basel II category 7 type of operational risk, which involves either a model not being used in the right way or its outcome not being properly understood. The misuse may be dramatic, leading to conduct risk or a massive misevaluation in the books. Basel II category 7 is defined as follows (Basel Committee on Banking Supervision 2001a, 2002): “Execution delivery and process management: losses from failed transaction processing or process management, from relations with trade counterparties and vendors.” A lack of understanding about how to properly use a particular model engenders model risk. This risk is potentially induced by a disconnect between the model developer – that is, the person who thoroughly understands what the model is supposed to do and how it is supposed to behave – and the model owner, who simply uses it. This type of risk may result from using wrong, inappropriate or scanty inputs; from mistakes in the way the model has been implemented; or due to an absence of rigour when using the model.
3 Model risk management
Once model risk has been defined, identified and understood, the failures it may lead to should be controlled, reported and mitigated.
3.1 Understanding a model failure to prevent it: the crash-test process
The first of the processes we suggest financial institutions implement is the crash-test process (Popper 1992), through which a model is fully challenged and rigorously tested. Originally, crash tests were a way to evaluate the resistance of a vehicle, along with its components, system and passengers, by driving it at a certain speed into a wall. Instead of a wall, we need to build specific situations or scenarios in which a model might fail to understand its environment and limitations.
The components of the crash-test process can be summarized as follows:
- (1)
random data to fail the assumptions, eg, using generalized Pareto distributed data in a model that is designed for a lognormal distribution;
- (2)
preposterous parameterization, eg, using extreme probabilities of default and losses given default in a credit risk model;
- (3)
variance of the results in the case of nonstationary scenarios, eg, using nonstationary data in a model to understand that model’s behavior; and
- (4)
aliasing and propagation error analysis, etc, eg, evaluating the diffusion of an initial error in a model including convolution or forecasting.
A proper example can be found in operational risk modeling (Guégan and Hassani 2012), as using a lognormal distribution on censored data may lead to a negative location parameter (commonly denoted ), which implies that most modeled losses are between 0 and 1. In some categories, however, that may not make any sense. This crash-test process would help us understand the limitations of a designed model, its sensitivity to assumptions, its domain of application and the impact a paradigm shift might have on it.
The crash-test process forces us to question the way models are built. Most of the time, due to short deadlines (either driven by endogenous or exogenous factors), the initial models implemented are simple and therefore may combine some of the issues presented in Section 2 (although, fortunately, this not always the case). Pearl and Mackenzie (2018) suggest a way of structuring the construction of models, starting from the knowledge (in a broad sense) we have of the issue to be dealt with, ie, information, past data, environment, etc. A deeper analysis of the information available might be particularly useful in curating a first set of assumptions that simplifies reality. Then, Pearl and Mackenzie suggest focusing on causality, ie, what connects one process to another. Note that it is not necessarily easy to test for causality. Here, if the past information consists of data, we refer the reader to Granger (1969) or Sugihara et al (2012), who propose statistical tests to evaluate the cause-and-effect relationship between two time series variables (correlation does not imply causation). Once causality has been addressed, a model can be structured to address the targeted issue. The model creation process consists of a loop, where, if an approach does not permit answering the scientific question, assumptions have to be challenged and causality reevaluated until a satisfactory approach is obtained.
The crash-test process assumes the model under scrutiny has not failed before and therefore no information has been recorded regarding a failure. However, if the model has failed before, a forensic analysis might be particularly informative;66 6 The Institute for Pure and Applied Mathematics at UCLA hosted a workshop on “Forensic Analysis of Financial Data” in May 2015. therefore, the possibility of conducting such an analysis should not be discarded.
3.2 Conservatism
Conservatism can be engineered at various levels of model construction (Guégan and Hassani 2015). Here, we suggest solutions at four stages, namely distribution selection and the capture of intrinsic dynamics, correlations and risk measures. Here, the term “dynamics” refers to an evolution of the data (and, consequently, of the model and its parameters over time), not to dynamical systems precisely (as in chaos theory or bifurcation theory), which might be highly unstable.
Regarding the distribution used, we have several choices for capturing skewness, leptokurtosis and extreme events (tails) – in other words, for capturing asymmetric shocks and fat tails – as follows:
- •
the generalized hyperbolic distribution (see Barndorff-Nielsen and Halgreen, 1977);
- •
the extreme value distribution that describes the limit distributions of normalized maxima (see Resnick, 1987);
- •
the generalized Pareto distribution, which appears as the limit distribution of scaled excesses over high thresholds (see Embrechts et al, 1997);
- •
the -stable distribution (see Nolan, 1997);
- •
the -and- distribution (see Hoaglin, 1985).
A knowledge of the dynamic components embedded in the input should promote the construction of a robust model. Analysis of the dynamics embedded within the data through a time series will permit better responsiveness to and conformance with reality. Both short-memory models, such as autoregressive (AR) processes and generalized autoregressive conditional heteroscedasticity (GARCH) models (Engle 1982; Bollerslev 1986), and long-term models, such as the Gegenbauer process (Guégan 2005), can be used.
The capture of dependencies can be achieved through copulas. Recall that a copula is a multivariate distribution function linking a large set of data through its standard uniform marginal distributions (Sklar 1959). In the literature, it has often been mentioned that using copulas is difficult when we have more than two risks: that is, apart from using elliptical copulas such as the Gaussian or the Student type (Gourier et al 2009). However, elliptic copulas fail to capture asymmetric and extreme shocks. The use of a Student copula with three degrees of freedom to capture the dependence between the largest losses mechanically implies a higher correlation between the smallest losses. One alternative is to use the Archimedean or extreme value copulas (Joe 1997), which are capable of capturing dependencies embedded in different parts of marginal distributions. However, the direct calibration of these copulas is not straightforward. To overcome the restrictions faced while employing these copulas, recent developments using either nested copulas (Joe 2005) or vine copulas (Brechmann et al 2012; Guégan and Maugis 2010; Guégan and Hassani 2013a) can be turned to.
In terms of risk measurement, although risks in banks were previously evaluated using standard deviation-based measures, the financial industry has moved toward quantile-based downside risk measures including VaR and ES. VaR measures the losses that may be expected for a given probability and corresponds to the quantile of a selected distribution. In addition, to measure the importance of an exposure beyond the VaR percentile, and to capture some diversification benefits, ES has been introduced. This has been widely dealt with in the literature: see, for example, Artzner et al (1999), Rockafellar and Uryasev (2000, 2002) and Delbaen (2000). The relationship between VaR and ES for some distributions can be found in Guégan and Hassani (2014).
Recently, several extensions have been analyzed, proposing new routes for risk measures. These are summarized below.
- •
A spectral risk measure (Acerbi and Tasche 2002), which can provide a convex combination of different ESs with appropriate weights.
- •
A distortion risk measure, which can be described as follows.
- –
When the distribution is shifted with a function close to the Gaussian one, as in Wang (2000) and Sereda et al (2010), the shift distribution remains unimodal. Thus, Guégan and Hassani (2014) proposed distorting the initial distribution with polynomials of odd degree in order to create several humps in the distribution. This helps to catch all of the information in the extremes of the distribution and introduces a new coherent risk measure computed under the distorted distribution.
- –
Guégan et al (2015) proposed a new measure combining the construction of a confidence interval around VaR with a spectral approach to provide a conservative envelope around the cdf, therefore focusing on the various indicators that can be used for risk management purposes.
- –
It is important to make a distinction between the model, that is, the theoretical framework, and its application to a particular problem involving the estimation of parameters, calibration, etc. A model may be appropriate, but model risk can still be generated by its parameterization.
Finally, solutions to a problem may engender larger issues. Therefore, it is particularly important to analyze the trade-off between precision and stability before implementing a solution.
3.3 Challenger–champion relationship
The challenger–champion relationship appears to be a sensible way of addressing model risk in general, as it avoids relying on a single approach to capture a pattern and provides some information on a model’s reliability (if the results tend to converge, then either both are right or – in the worst-case scenario – both are misleading). A model is considered the “champion” when it is used as the main solution (because it has been approved by regulators, for example). Such models have usually been industrialized, whereas challenger models are those that have been developed in parallel to champion ones, usually for validation purposes. At the very least, by having an alternative – by way of this process – we can avoid falling into dogmas and “best practice” traps. We acknowledge the fact that it is sometimes wrong to use a particular model, and this approach is a humble and pragmatic way of dealing with such issues. The challenger model should be built from zero and should rely, as much as possible, on alternative assumptions and/or methodology.
In addition, the challenger–champion relationship potentially enlarges the domain of a model’s validity and ensures the continuity of quantification in case one model fails. This is actually something that has been suggested in the latest pieces of regulation on stress testing by the Prudential Regulation Authority (2013) and the European Banking Authority (2014b).
3.4 Corporate architecture: developer–user relationship
Corporate architecture requires senior management to have a basic understanding of mathematical figures and models. Indeed, the Federal Reserve (2011) states that: “As part of their overall responsibilities, a bank’s board and senior management should establish a strong model risk management framework that fits into the broader risk management of the organization. That framework should be grounded in an understanding of model risk – not just for individual models but also in the aggregate. The framework should include standards for model development, implementation, use and validation.
While the board is ultimately responsible, it generally delegates to senior management the responsibility for executing and maintaining an effective model risk management framework. Duties of senior management include establishing adequate policies and procedures and ensuring compliance, assigning competent staff, overseeing model development and implementation, evaluating model results, ensuring effective challenge, reviewing validation and internal audit findings, and taking prompt remedial action when necessary.” Therefore, they should not be afraid of speaking out when they do not understand certain aspects of the business. Meanwhile, developers should be trained to communicate in the most effective way and should be able to translate technicalities into words a layperson will understand. For example, they should be able to explain their model in one page, one paragraph or one sentence. Here, another communication trade-off between simplification and accuracy must be handled carefully, as an oversimplified explanation could result in a loss of accuracy and therefore mislead the management team.
Mathematics, if considered as a branch of philosophy, is useful for representing concepts and simplifying them. It certainly does not complicate them. However, the symbolism in mathematics is quite complicated: it is like a foreign language and requires some practice before it becomes comprehensible to anyone. Therefore, model developers, model owners and model users (ie, the people who ultimate use them to make decisions) should be trained to develop the skills that they lack. This would help to prevent or at least attenuate the risk of a model being misused or misinterpreted.
Usually, we observe a disconnect between model developers and users. The developer is the supplier and the user is the customer. As for any other product, the model should be developed to satisfy customer demands, both in terms of what it does and in terms of visualization. Obviously, the developer should uphold the integrity of the outcomes, ie, models should be developed in such a manner that they capture and represent a situation as precisely as possible to ensure the actionability77 7 That is, the decisions taken that rely on those models. of their outcomes. As characterized by the user modeling field of research, the main goal is to customize and adapt systems to meet users’ specific needs.
Practically, a model should be adapted to its user, and while an ergonomic environment is probably the most important thing for users, it does not prevent them from misunderstanding a model. The users of a model should be able to understand and explain the main features of that model as well as its main assumptions and limitations. They should not necessarily be expected to explain it quantitatively; however, from a practical point of view, they should be able to anticipate when a model is going to fail, relying, for example, on the results of a crash-test program.
3.5 Model governance
In theory, model governance, as summarized below, is a fairly simple process that just needs to be followed rigorously. However, doing an inventory of models is a huge task, as the number of models used in a financial institution can be very large. Indeed, models are not limited to capital quantification or pricing; they are, in fact, used everywhere, from marketing to strategy and beyond. Most people do not even know they are using a model to make a decision because so many models have been integrated into their day-to-day work.
An appropriate governance may be summarized as follows.
- (1)
The models should be crash tested and the developers themselves should develop the challenges.
- (2)
The developers should ensure that, even if some issues cannot be dealt with, the model is at least conservative.
- (3)
All of the pieces developed should be properly documented. The key assumptions, sensitivities and limitations should be adequately detailed. The code should be fully explained, as each developer has their own way of programming.
- (4)
The models should be independently validated.
- (5)
The models should follow a proper approval process, performance monitoring should be conducted and their materiality should be appropriately reported.
- (6)
The models should be overseen and challenged by senior management.
- (7)
An appropriate change management process should be adopted when a model is deployed.
We believe that, in the very near future, once model risk has been properly apprehended (ie, once we have progressed from understanding it to implementing functional corporate governance), it will be possible to hedge residual risk and even insure it. However, we believe that it is too early for this, as the industry is (more or less) still in a period of inherent risk assessment.
4 Conclusion
This paper aims to address the risks associated with implementing a model. It analyzes the main root causes of failure and provides some solutions to manage these, ie, to control, hedge and mitigate them. Model risk is a part of the daily life of any practitioner and can never be fully offset. This paper is far from being exhaustive; for example, other authors may consider technical risk as part of model risk, and many other features may be added. However, it does summarize model risk as understood by a quant. Another aspect of model risk related to its accountability will have to be addressed by the authorities.88 8 This is the topic of a companion paper. Indeed, in many cases, and particularly within financial institutions, models have to be generally and explicitly approved by regulators before they can be used; as such, the people who should be held accountable in case of failure might not always be located within the financial institutions themselves.
As the term “model” is not only used in mathematics, we may conclude this paper by quoting Vincent Van Gogh (1882): “Do not quench your inspiration and your imagination; do not become the slave of your model.”
Declaration of interest
The opinions, ideas and approaches expressed or presented here are those of the authors and do not necessarily reflect Deloitte’s position. As a result, Deloitte cannot be held responsible for such views.
Acknowledgements
This work was achieved through studies conducted in the Laboratory of Excellence for Financial Regulation (Labex ReFi) supported by PRES HESAM under the reference ANR10LABX0095. It benefited from French government support managed by the National Research Agency (ANR) within the project Investissements d’Avenir Paris Nouveaux Mondes (Investments for the Future, Paris–New Worlds) under the reference ANR11IDEX000602.
References
- Acerbi, C., and Tasche, D. (2002). On the coherence of expected shortfall. Journal of Banking and Finance 26(7), 1487–1503 (https://doi.org/10.1016/S0378-4266(02)00283-2).
- Acharya, V. V., Gündüz, Y., and Johnson, T. C. (2016). Bank use of sovereign CDS in the eurozone crisis: hedging and risk incentives. Working Paper, Social Science Research Network (https://doi.org/10.2139/ssrn.2875323).
- Artzner, P., Delbaen, F., Eber, J.-M., and Heath, D. (1999). Coherent measures of risk. Mathematical Finance 9(3), 203–228 (https://doi.org/10.1111/1467-9965.00068).
- Barndorff-Nielsen, O., and Halgreen, C. (1977). Infinite divisibility of the hyperbolic and generalized inverse Gaussian distributions. Zeitschrift für Wahrscheinlichkeitstheorie und Verwandte Gebiete 38(4), 309–311 (https://doi.org/10.1007/BF00533162).
- Basel Committee on Banking Supervision (2001a). QIS 2 – operational risk loss data. Report, Bank for International Settlements.
- Basel Committee on Banking Supervision (2001b). On the regulatory treatment of operational risk. Working Paper, Bank for International Settlements.
- Basel Committee on Banking Supervision (2002). Operational risk data collection exercise. Report, Bank for International Settlements.
- Bennett, D. E. (2017). Governance and organizational requirements for effective model risk management. The Journal of Risk Model Validation 11(4), 97–116 (https://doi.org/10.21314/JRMV.2017.188).
- Black, F., and Scholes, M. (1971). The pricing of options and corporate liabilities. Journal of Political Economy 81(3), 637–654 (https://doi.org/10.1086/260062).
- Bollerslev, T. (1986). Generalized autoregressive conditional heteroskedasticity. Journal of Econometrics 31(3), 307–327 (https://doi.org/10.1016/0304-4076(86)90063-1).
- Bonola, R. (2007). Non-Euclidean Geometry. Cosimo Classics, London.
- Box, G., and Norman, R. (1987). Empirical Model-Building and Response Surfaces. Wiley.
- Brechmann, E., Czado, C., and Aas, K. (2012). Truncated regular vines in high dimensions with applications to financial data. Canadian Journal of Statistics 40(1), 68–85.
- Crampin, S., Gao, Y., and Peacock, S. (2008). Stress-forecasting (not predicting) earthquakes: a paradigm shift? Geology 36(5), 427–430 (https://doi.org/10.1130/G24643A.1).
- Dal Pozzolo, A., and Bontempi, G. (2015). Adaptive machine learning for credit card fraud detection. PhD Thesis, Université libre de Bruxelles, Faculté des Sciences – Informatique, Bruxelles.
- Delbaen, F. (2000). Coherent risk measures. Lecture Notes, Cattedra Galileiana at the Scuola Normale di Pisa (https://doi.org/10.1007/BF02809088).
- Einstein, A. (2011). Out of My Later Years: The Scientist, Philosopher, and Man Portrayed Through His Own Words. Philosophical Library.
- Embrechts, P., Kluppelberg, C., and Mikosh, T. (1997). Modelling Extremal Events: For Insurance and Finance. Springer (https://doi.org/10.1007/978-3-642-33483-2).
- Engle, R. (1982). Autoregressive conditional heteroscedasticity with estimates of the variance of United Kingdom inflation. Econometrica 50(4), 987–1007 (https://doi.org/10.2307/1912773).
- Ernest, P. (2004). The Philosophy of Mathematics Education. Taylor and Francis, London.
- Euclid (1956). The Thirteen Books of the Elements, Heath, T. (trans), Volume 1, Books 1–2. Dover, London.
- European Banking Authority (2014a). Draft regulatory technical standards on assessment methodologies for the advanced measurement approaches for operational risk under article 312 of regulation. Consultation Paper (EU) 575/2013, EBA.
- European Banking Authority (2014b). EU-wide stress test 2014. Methodological Note, EBA.
- Federal Reserve (2011). Supervisory guidance on model risk management. Supervision and Regulation Letters SR 11-7, Division of Banking Supervision and Regulation, Fed.
- Fiske, A., and Rai, T. (2014). Virtuous Violence – Hurting and Killing to Create, Sustain, End, and Honor Social Relationships. Cambridge University Press (https://doi.org/10.1017/CBO9781316104668).
- Flood, M. (1991). An introduction to complete markets. Review, March/April, Federal Reserve Bank of St. Louis (https://doi.org/10.20955/r.73.32-57).
- Frachot, A., Georges, P., and Roncalli, T. (2001). Loss distribution approach for operational risk. Working Paper, GRO, Crédit Lyonnais, Paris (https://doi.org/10.2139/ssrn.1032523).
- Gourier, E., Farkas, W., and Abbate, D. (2009). Operational risk quantification using extreme value theory and copulas: from theory to practice. The Journal of Operational Risk 4(3), 3–26 (https://doi.org/10.21314/JOP.2009.064).
- Granger, C. W. (1969). Investigating causal relations by econometric models and cross-spectral methods. Econometrica: Journal of the Econometric Society 37(3), 424–438 (https://doi.org/10.2307/1912791).
- Guégan, D. (2005). How can we define the concept of long memory? An econometric survey. Econometric Review 24(2), 113–149 (https://doi.org/10.1081/ETC-200067887).
- Guégan, D., and Hassani, B. (2012). Operational risk: a Basel II ++ step before Basel III. Journal of Risk Management in Financial Institutions 6(1), 37–53.
- Guégan, D., and Hassani, B. (2013a). Multivariate VaRs for operational risk capital computation: a vine structure approach. International Journal of Risk Assessment and Management 17(2), 148–170 (https://doi.org/10.1504/IJRAM.2013.057104).
- Guégan, D., and Hassani, B. (2013b). Using a time series approach to correct serial correlation in operational risk capital calculation. The Journal of Operational Risk 8(3), 31–56 (https://doi.org/10.21314/JOP.2013.126).
- Guégan, D., and Hassani, B. (2014). Distortion Risk Measures or the Transformation of Unimodal Distributions into Multimodal Functions. Springer (https://doi.org/10.1007/978-3-319-07524-2_2).
- Guégan, D., and Hassani, B. (2015). Stress testing engineering: the real risk measurement? In Future Perspectives in Risk Models and Finance, Bensoussan, A., Guégan, D., and Tapiro, C. (eds). Springer (https://doi.org/10.1007/978-3-319-07524-2_3).
- Guégan, D., and Maugis, P.-A. (2010). New prospects on vines. Insurance Markets and Companies: Analyses and Actuarial Computations 1(1), 15–22.
- Guégan, D., Hassani, B., and Li, K. (2015). The spectral stress VaR (SSVaR). Working Paper 2015.52, Sorbonne Center for Economics (https://doi.org/10.2139/ssrn.2622391).
- Hassani, B. (2015). Risk appetite in practice: Vulgaris Mathematica. Working Paper, Université Paris 1.
- Hoaglin, D. C. (1985). Summarizing shape numerically: the -and- distributions. In Exploring Data Tables, Trends, and Shapes, Hoaglin, D. C., Mosteller, F., and Tukey, J. W. (eds), pp. 461–513. Wiley (https://doi.org/10.1002/9781118150702.ch11).
- Joe, H. (1997). Multivariate Models and Dependence Concepts. Monographs on Statistics and Applied Probability. Chapman and Hall, London (https://doi.org/10.1201/b13150).
- Joe, H. (2005). Asymptotic efficiency of the two-stage estimation method for copula-based models. Journal of Multivariate Analysis 94, 401–419 (https://doi.org/10.1016/j.jmva.2004.06.003).
- Knight, F. H. (1921). Risk, Uncertainty and Profit. Hart, Schaffner and Marx, New York.
- Kuhn, T. (1979). The Essential Tension: Selected Studies in Scientific Tradition and Change. University Of Chicago Press.
- Kuhn, T. (2012). The Structure of Scientific Revolutions: 50th Anniversary Edition. University Of Chicago Press (https://doi.org/10.7208/chicago/9780226458144.001.0001).
- Lenarčič, A., Mevis, D., and Siklós, D. (2016). Tackling sovereign risk in European banks. Technical Report, European Stability Mechanism.
- Maslow, A. (1962). Toward a Psychology of Being. Van Nostrand, Princeton, NJ (https://doi.org/10.1037/10793-000).
- Morini, M. (2011). Understanding and Managing Model Risk. Wiley (https://doi.org/10.1002/9781118467312).
- Ng, S., Collisson, E., Sokolov, A., Goldstein, T., Gonzalez-Perez, A., Lopez-Bigas, N., Benz, C., Haussler, D., and Stuart, J. (2012). Paradigm-shift predicts the function of mutations in multiple cancers using pathway impact analysis. Applied and Translational Bioinformatics 28(18), 640–646 (https://doi.org/10.1093/bioinformatics/bts402).
- Nin, A. (1972). The Diary of Anaïs Nin, Volume 4: 1944–1947. Mariner Books, New York.
- Nolan, J. P. (1997). Numerical calculation of stable densities and distribution functions. Communications in Statistics: Stochastic Models 13(4), 759–774 (https://doi.org/10.1080/15326349708807450).
- O’Hara, M. (1998). Market Microstructure Theory. Wiley.
- Pearl, J., and Mackenzie, D. (2018). The Book of Why: The New Science of Cause and Effect. Basic Books, New York.
- Pinker, S. (2012). The Better Angels of Our Nature: A History of Violence and Humanity. Penguin, London.
- Planck, M. (1968). Scientific Autobiography and Other Papers. Philosophical Library, New York.
- Popper, K. (1963). Conjectures and refutations: the growth of scientific knowledge. Physics Today 16(11), 80 (https://doi.org/10.1063/1.3050617).
- Popper, K. (1992). Le réalisme et la science. Revue Philosophique de la France et de l’Étranger 182(1), 132–135.
- Priestley, M. B. (1983). Spectral Analysis and Time Series. Academic Press (https://doi.org/10.2307/1268567).
- Prudential Regulation Authority (2013). A framework for stress testing the UK banking system. Report, Bank of England.
- Resnick, S. (1987). Extreme Values, Regular Variation, and Point Processes. Springer Science and Business Media (https://doi.org/10.1007/978-0-387-75953-1).
- Rockafellar, R., and Uryasev, S. (2000). Optimization of conditional value-at-risk. The Journal of Risk 2(3), 21–41 (https://doi.org/10.21314/JOR.2000.038).
- Rockafellar, R., and Uryasev, S. (2002). Conditional value at risk for general loss distributions. Journal of Banking and Finance 26(7), 1443–1471 (https://doi.org/10.1016/S0378-4266(02)00271-6).
- Roll, R. (1977). A critique of the asset pricing theory’s tests part I: on past and potential testability of the theory. Journal of Financial Economics 4(2), 129–176 (https://doi.org/10.1016/0304-405X(77)90009-5).
- Rosch, D., and Scheule, H. (2010). Model Risk – Identification, Measurement and Management. Risk Books, London.
- Sereda, E., Bronshtein, E., Rachev, S., Fabozzi, F., Sun, W., and Stoyanov, S. (2010). Distortion Risk Measures in Portfolio Optimisation. Business and Economics Series, Volume 3. Springer (https://doi.org/10.1007/978-0-387-77439-8_25).
- Sharpe, W. (1964). Capital asset prices: a theory of market equilibrium under conditions of risk. Journal of Finance 19(3), 425–442 (https://doi.org/10.1111/j.1540-6261.1964.tb02865.x).
- Sklar, A. (1959). Fonctions de répartition en dimensions et leurs marges. Publications de l’Institut Statistique de l’Université de Paris 8, 229–231.
- Sugihara, G., May, R., Ye, H., Hsieh, C.-H., Deyle, E., Fogarty, M., and Munch, S. (2012). Detecting causality in complex ecosystems. Science 338(6106), 496–500 (https://doi.org/10.1126/science.1227079).
- Taleb, N. N. (2007). The Black Swan: The Impact of the Highly Improbable, Volume 2. Random House.
- Tegmark, M. (2008). The mathematical universe. Foundations of Physics 38(2), 101–150 (https://doi.org/10.1007/s10701-007-9186-9).
- Tunaru, R. (2015). Model Risk in Financial Markets: From Financial Engineering to Risk Management. World Scientific (https://doi.org/10.1142/9524).
- Van Gogh, V. (1882). Letter from Vincent Van Gogh to Theo Van Gogh. Letter, c. November 2/3, The Hague. URL: https://bit.ly/2b7nUFb.
- Wang, S. S. (2000). A class of distortion operators for pricing financial and insurance risks. Journal of Risk and Insurance 67(1), 15–36 (https://doi.org/10.2307/253675).
Only users who have a paid subscription or are part of a corporate subscription are able to print or copy content.
To access these options, along with all other subscription benefits, please contact info@risk.net or view our subscription options here: http://subscriptions.risk.net/subscribe
You are currently unable to print this content. Please contact info@risk.net to find out more.
You are currently unable to copy this content. Please contact info@risk.net to find out more.
Copyright Infopro Digital Limited. All rights reserved.
As outlined in our terms and conditions, https://www.infopro-digital.com/terms-and-conditions/subscriptions/ (point 2.4), printing is limited to a single copy.
If you would like to purchase additional rights please email info@risk.net
Copyright Infopro Digital Limited. All rights reserved.
You may share this content using our article tools. As outlined in our terms and conditions, https://www.infopro-digital.com/terms-and-conditions/subscriptions/ (clause 2.4), an Authorised User may only make one copy of the materials for their own personal use. You must also comply with the restrictions in clause 2.5.
If you would like to purchase additional rights please email info@risk.net