SlideShare una empresa de Scribd logo
1 de 23
Giulio Laudani #23                                                                                                                                     Cod 20163

                                             RISK MANAGEMENT
                  Interest risk _________________________________________________________________________________ 1
                     Methods to Compute Interest Risk: ______________________________________________________________________ 2

                     A proper Scheme to manage Internal Transfer Rate operation: _______________________________________________ 3

                  Credit risk __________________________________________________________________________________ 4
                     The Expected Loss Estimates: ___________________________________________________________________________ 4

                     The measure of the Unexpected Loss: ____________________________________________________________________ 7
                        Tools used to estimate UL: ________________________________________________________________________________________ 8
                        Comments on the UL tools employed: _______________________________________________________________________________ 9

                     Application and General comments on Credit Risk methods: _________________________________________________ 9

                  Market risk ________________________________________________________________________________ 10
                     Tools used to estimate market risk _____________________________________________________________________ 10

                     Different applications of the Market risk model: __________________________________________________________ 12

                  Measure of Volatility ________________________________________________________________________ 13

                  The general limits of the VAR approach are: _____________________________________________________ 13

                  Basel Committee Framework: _________________________________________________________________ 15
                     Basel I: ____________________________________________________________________________________________ 15

                     Basel committee II ___________________________________________________________________________________ 15

                  Insurance business __________________________________________________________________________ 17
                     Solvency II _________________________________________________________________________________________ 17
                        The formation process and directive structure: _______________________________________________________________________ 18
                        Some example: ________________________________________________________________________________________________ 19

                     Difference between Banks and Insurer __________________________________________________________________ 20

                     A new risk paradigm _________________________________________________________________________________ 21
                        Problems related to the crisis: ____________________________________________________________________________________ 21
                        All these problems have caused: __________________________________________________________________________________ 22
                        How the new RM activity looks like after the crisis: ____________________________________________________________________ 22


Interest risk
The source of this risk is due to the transformation of maturities and the mismatch between asset and liabilities. Those situations produce an
imbalance that will lead to:Refinancing risk whenever the maturity on asset is longer than liabilities one; Reinvesting risk when we are in the reverse
situation; Changing in the demand/offer side of the demand of liabilities and call loans, hence the risk linked to the elasticity of demand curve (it
won’t be study).

Those imbalances are shown in all items affected by any change in interest rate, hence we should consider not only the trading book, but the
broader one, that is the banking book, including any derivate whose value depends on market interest rate. The general rule on how treat this risk
are explained within theBasel Committee framework, whose general principles have been proposed to help/offer guideline to national authority,
which are in charge of the supervision of the bank institution, on how to best estimate and to best manage this risk. Here there is a list of the main
area described in the Basel Committee framework:
          Significant involvement of senior manager to overcome the traditional independency of risk management unit with the other operational
          division, this active rule of the senior management will grant uniformity on the criteria, objectiveness and proper procedure. This general
          provision is concretely declinated:
                     The board must be informed regularly and must approve the policy enhanced
                     It must ensure that the risk management has the competence to deal with
                                                                                                                                                              1
Giulio Laudani #23                                                                                                                                                          Cod 20163
                      All the monitoring activity must be put in place and there must operational limits
                      The bank must communicate to the supervisory and also to the public their own risk
            Create independent unit which has to help the senior management in the decision process, providing technical point of view and avoiding
            potential conflict of interest
            Bring the measurement of this risk at a consolidate level to adequately appraised and managed
            Integrate this risk measure into a day to day management of the bank, to steer the corporate policy in the way of how business is lead

Methods to Compute Interest Risk:
There exist several models to capture/measure this risk from the simplest to more advanced one, however each of them can provide a good
guidance to Risk manager since they analyze different angle of the same problem:

            Interest gap is an income based approach, it considers the effect of change in the market interest rate on the NI
                                                                      1
                  o     There are several definition of Gap starting from the simplest to the most complicated/accurate, all of them depends on the
                        time windows considered, hence it is a time dependent measure
                                The first is basically the difference between sensitive assets and liabilities, where by sensitive we mean all those asset
                                 which are going to be re-price/roll over in the time frame considered
                                                                                                      2
                                    The second one is named Maturity-Adjusted GAP ,which is a weighted average with weights the exposure of each
                                     instruments within its GAP. This procedure is quite heavy demanding on the computational side, hence it is simplified
                                                                                     3
                                     by using the concept of marginal GAP
                                    The standardized gap has been introduce to overcame the assumption of same effect for each category included in the
                                     balance sheet. It consists of computing the beta of each categories
                                                                                                                                                                                 4
                                 To take into account the change on demand for those items which haven’t an automatic indexing mechanism , we will
                                  use the average delay to allocate the instrument to the properly marginal Gap windows
                                 A similar approach can be used to compute the price and quantity interaction effect
                  o     The intrinsic limitation of this approach is the GAP is considering only the effect on the net income, hence it is ignoring the effect
                        on the market value of the fixed part of the balance sheet, whose value is effected by the way

            Duration gap is an equity based approach. It is more consistent with the newmarket to market accounting principle. We are going to use
            the durationwhich is a measure of the residual life of the instrument, and the modified durationwhich is the first derivative to the respect
            of interest of the cash flow equation, hence it can be used as a first approximation of the change in value of financial instruments
                  o The duration GAP is estimated bycomputing the MD of Assets and liabilities, where the second one is multiply with the ratio of
                      market value between liabilities and assets (financial leverage)


                  o     Limits of this approach are:The immunization will last a very short time, because the duration will change over time, the change
                        of interest rate will change the DM, it is a linear approximation, it is based on the assumption of uniform changes in interest
                        rate (basis risk), assuming parallel shift across maturities, and it is a costly strategy
                  o     Some of those limitations can be overcame by some wise modification of the general approach: the basis risk can be solved by
                        using a measure of sensitivity; the cost can be manage using derivatives instruments and the linear approximation can be reduce
                        by using the convexity parameter (second derivatives, distribution of cash flow around the duration)

            Cash-flow mapping (clumping) is the model that enable to consider asymmetrical shifts on the yield curve, furthermore it is the model
            suggested by the Basel committee. The method consists of dividing all the instruments in each of its cash flow,and of converting all the
            instruments into zero coupon, on whom we will use the appropriate zero coupon rate [maturity]


1
  The usage of the GAP brings to different macro definitions:
              Marginal GAP is the GAP computed for a specific window inside the horizon consider. It is related to the cumulative one by two relationship:
                    o      The sum of all the Marginal GAP is the cumulative one at time T
                    o      The difference of two cumulative gap with different time horizon gives us the marginal GAP between the two horizon
              Cumulative GAP is computed over the all-time horizon
2
  It aims to overcame the limit of the previously one, namely the over simplistic assumption of instantaneous change of interest rate for all the sensitive items
3
 The marginal GAP consists of aggregating items in time category on whom there will be computed the significant value and the maturity is simple the average of the beginning and ending
date
4
  The underling hp is that those items respond to the change with some delay
                                                                                                                                                                                           2
Giulio Laudani #23                                                                                                                            Cod 20163
                  o     In set up this model there are several choices that should be undertaken:
                                                                         5
                                  We need to estimate the vertices on the zero coupon curve, taking into accounttypical bank features: Change in short
                                   term interest rate are greater than longer rate, volatility decrease as maturities increase and bank’s cash flow are more
                                   concentrate in the short term
                                  We need to choose discrete interval since the “item by item” approach is too demanding, we are going to aggregate
                                                                   6
                                   them using their residual life or using MD to take into account the different sensitivity to change in rate
                  o     This method is suggested by Basel Committee to calculate a synthetic risk measure by applying this scheme:
                                  All the assets and liabilities are divided in 13 nodes applying the modified residual life
                                  For each node there is an associated risk coefficient, given by the product of DM and change in interest rate. It is
                                   multiplied by the Net position for each node
                                  It doesn’t allow bank to completely off set gaps, to account for asymmetrical movements
                                  This measure is compared with the capital requirement
                  o     The Basel methods has several limits:
                                  Use of Book value instead of market one as reference
                                  Presence of instruments which repay capital before maturity create a bias in the modified method residual life
                                   approach [duration drift]
                                  Presence of instruments without a fixed renegotiation date or linked to market interest rate
                                  This method is not adequate to consider derivatives
                                  There is no compensation allowed for different currency
                                  Basel committee doesn’t provide a unique solution to those problems, actually it leaves the national supervisory
                                   authority free to choose their models
                  o     The clumping method provides a procedure to broken real cash flow in two virtual one in effort to better match the nodes
                        vertexes, so that to improve the cash flow mapping results, but require an in-depth knowledge of all cash flow, that is why it can
                        be applied only to small portion of the bank’s portfolio
                                  The virtual cash flow must guarantee constant portfolio (same MV) value and same risk to ensure same change against
                                   interest rate movements
                                                   o                                                 is the first node
                                                   o                                where the rate and time is the node one
                                  A variation of the previously method is basing the clumping on the price volatility to take into account the risk
                                   correlation. The two virtual cash flow must sum up to the real volatility

A proper Scheme to manage Internal Transfer Rate operation:
The ITR is a series of virtual transaction between branches and the risk center unit. It aims to centralize all the decision regarding the risk taking
process, to better evaluate the profitability of each branch, to relive the branches from the need to care about the founding process and to transfer
all the risk to a central unit.

There are two possible scheme to be used: one is the Single ITR, which is the more easier to implement, but it is criticized due to the unique
(arbitrary, not related to market condition) rate which is used in the transactions, it handles only net flow, hence the branches are still affected by
part of the risk; the other one is the Multiple ITR, whichovercomes this limitation and allows to use multi interest rate for different maturities and
gross flows, hence each operation is taken into consideration

Here we are going to provide some examples of peculiar transaction and how ITR dealt with:
    Fixed rate transaction, the risk unit take all the risk, the branch is ensured by blocking is financing rate
    Floating rate transaction, the same of above plus a premium that repay the higher risk
    Not indexed at market rates poses two problems: there is no risk hedging instrument and there is a basis risk in fact if we use a derivative on
    similar underlying there could be an increase in the spread between the similar asset and the real one. The bank can decide who will bear the
    basis risk since neither the treasury or the branch are able to handle it
               To avoid arbitrage against Treasury, it could be possible to add a premium over the internal rates to give an incentive to choose the
               most favorable rate



5
The nodes used are usually consistent with the maturity available for hedging instruments
6
Note that the degree of risk doesn’t depend exclusively on residual life, hence we are simplifying the model in this case
                                                                                                                                                          3
Giulio Laudani #23                                                                                                                          Cod 20163
      Derivatives and special payoff, the option: the case of cap or floor or early repayment; price must be paid by the client with the spread charged
      or paid up-front. It can be beardby the branch or the treasury. The first possibility is simple using options

When we are building up an ITR system, we need to address all the desirable features, aiming to ensure a proper risk management system within our
institution. Hence, we need to ensure that any Change of profitability is due to only of credit risk, and the total bank’s profitability shouldn’t change;
                                                                                                                                7
we must protect our selffrom the interest rate risk and embedded options; we should use a multiple rate ITR and we should use gross flow
           8
methods ; we should use this system together with a cash mapping process

Credit risk
Credit risk represents the risk of the default of the counterpart (risk of insolvency) or deterioration in its creditworthiness (migration risk), hence it
will arise whenever the discount rate of future cash flow doesn’t reflect the risk of the transaction. Besides to those main sources of risk there are
other components: spread risk [whenever the market spread required for the same risk increase due to an increase in the risk aversion] and
recovery risk, country risk and pre-settlement risk [when a forward contract is cancelled due to the insolvency of the counterpart and the bank is
forced to replace it at unfavorable condition]. The positions accounted for determining the credit risk are the balance sheet items and the off
balance one (OTC transaction)

The real source of risk is its unexpectedcomponent in fact if the predictable one will be incorporate into the interest rate spread/premium and so
totally eliminated. This risk has two components Expected loss (EL) and the Unexpected loss (UL), our task is to estimate the first to properly price
the instrument and to measure the second to raise capital to cover this position (Basel Requirement). Estimate this risk and its impact is not an easy
task, in fact bank’s credit exposure are recorded at historical value and there is no secondary market to easily determine the market value of those
                                                                               9
position , hence we need to use internal asset pricing formula .

Here there is a set of rules to be implemented to properly set up adequate governance standards:
         Establishing appropriate credit risk environment, meaning that the board must approve and review credit risk strategies, pointing out the
         bank risk profile and profitability required. Senior managers must have the responsibility to adequately implement those strategies for all
         products
         Use sound credit granting process: establishing credit limits, new client policy or amendment as well as allowing re-new credit line with
         particular care
         Maintaining an appropriate credit control process
         Endure an adequate credit risk control, ensuring independency, IT instruments, recovery bad loan credit facilities and properly
         prudential level in assessing risk
         Supervisory should check that those requirement are met and posing limits on risk exposure

The Expected Loss Estimates:
The EL is function of three different components [referring only to insolvency risk] that must be estimated:

The PD estimation:
                                                                                                                         10
PD is the probability of default, it can be defined by a more subjective or objective definition and it can be assessed by backwarding [human base
decision or automated algorithms] or byforwarding [expectation on future development, market data] looking methods. It is estimated by applying
several models:

     Credit scoring models are statistical approach based on economic variables and financial indicators. They are used to forecast default and risk
      level of the borrow one by one or by discrete grades (this procedure is better since it reduces the error)
                 Linear discriminant analysis consistson classifying the data using different variables, to define a discriminatory value and to draw a
                 boundary line to clearly separate, as much as possible, the data between healthy and bad company.One of the possible variant is the
                 Altman’s Z-score which is a multivariate regression that already suggests some key variables.



7
 Close to those present in the market [even on the ask and bid side], facilitating the risk hedging
8
  Each transaction is considered one by one, no netting procedure ì, otherwise the branches will still retina some risk
9
 Note that to estimate default risk we can simply use book value, we need market one to estimate the effect of migration risk
10
   Depending on the decision the PD will be greater or lower, it depends on the criterion used
                                                                                                                                                        4
Giulio Laudani #23                                                                                                                                 Cod 20163
                           o      The score can be seen as a “weighted average” of the scores of a set of different variables, where unimportant variables are
                                  weighted close to zero, while important ones get high weights and counterproductive variables get negative coefficients.
                           o      Weights are chosen to discriminate as far as possible good firms’ scores from the bad firms’ ones, so mathematically we are
                                                                                             11
                                  aiming to maximize the distance between the two centroid with taking into account their variance [which is assumed to be
                                  the same over all the category, this hp can be ease]
                           o      To test the significance of this model we can use a parameter calledWilks’ Lambda which basically will compered the
                                  distance between the two centroids. It is the ratio between the variance of the healthy and abnormal scores over the total
                                  variation (similar to R^2)
                           o      From the score we can assess the PD by a formula based on the hp of normal distribution of independent variables. It
                                  depends on the score, on the cut-off level (which can be simple or more sophisticated one) and on the prior probability of
                                  default [ .


                           o   The cut off between company is computed using the past observations and can be simple the average between the distance
                               of the centroids (simple rules) or to ensure a given PD for the accepted company we can use a given formula; the problem is
                               that we can refuse good company that are in the grey/overlapping area.
                                        There is a possible correction term to take into consideration the average quality of the sample or to take into
                                         consideration the cost of I and II type of errors. It aims to minimize the first one by adding a new term which
                                         considers as an information the ratio between the I and II types of error
                         o These approaches suffer of several limits: the assumption on normal distribution, there are variant with Heteroskedasticity,
                               but they require lots of data and the so called “sample bias”
                         o To select the meaningful variables there are two possible approach: backward elimination or forward selection, keeping in
                               mind the ratio of the model
                     Regression models require defining a sample size, independent variables and how to estimate coefficient. There exist several variants:
                         o Simple which pose the problem of not limited dependent possible value
                         o Logit whish substitute the previously linear relationship with an exponential one
                         o Probituse a normal cumulative density function (smaller tail compared to logit)
                     Heuristic inductive models
                                                                                                         12
                         o Neural, it tries to minimize human learning approach (black-box features )
                         o Genetic based on the survivorship completion (only the best are able to transfer to future generation their genetic)
                     General limits of this family of models are:
                         o They don’t consider migration and qualitative issues
                         o The meaningfulness of the variable used is questionable, since they must be industrial specific (it is not the case) and needs
                               big sample to avoid bias caused by the presence of too healthy companies in the sample

Capital market models are similar to a VAR approach

               Use of spot or forward rate to estimate the default probability within one year and beyond. The key parameters are the spread, the PD,
               the recovery and the risk free rate. The PD parameter is computed using the corporate yield curve to maturity and the risk free one, we will
               use e^ form
                    o In case of maturity over one year we can use long term spot or forward spread, where d is the spread and k is the recovery rate
                                                                             spot case
                                                                             forward case
                     o     The spot will directly gave us the cumulative probability while the forward the marginal one, it is possible to obtain the
                           cumulative with the forward by the formula

                     o     While the marginal one is obtained with the spot by using the following formula:

                     o     Limits of this approach are:
                                     It is applicable only for listed company, that have listed bonds for all the relevant maturities, and it is also suffers of
                                      the market illusion


11
     It is the average value of historical company data for each sample category
12
     The procedure is obscure, and it is compared to a black box from where we gain result without knowing how it is working inside
                                                                                                                                                               5
Giulio Laudani #23                                                                                                                                                             Cod 20163
                                                                                                13
                                    Itassumes a risk neutral approach , hence we should consider/add a premium
                                    It relays on the expectation theory, however there exist liquidity premium
              Merton model is an example of structural model, meaning it is a model that recognize default as a consequence of some intrinsic
              company’s features, differently form the reduced model which simply recognize it as a possible event and reduce the problem to estimate
              its likelihood
                     o Characteristics are: it is based on the ideathat the shareholder position is similar to an option to default (when the value of the
                          assets are lower than liability one)
                                    It assumes that the value of the company follows a Brownian motion, with an increasing uncertainty
                                    At maturity it computes the frequency of the final result of the path simulated and check how many of them are below
                                     the threshold
                                    This probability is function of the asset volatility, nominal value of debt, starting company value and the debt maturity
                     o The contingent claim analysis, since those risks can be hedge using a put option the investment can be seen as a risk free
                          investment, where the put premium will be a proxy of the default probability. Thanks to this relationship we obtain the risk free
                          default probability, the value of the debt and the interest rate required by the bank
                     o Analyzing the spreads and default probability computed by applying the Merton model we notice how the riskier company
                          spread curve is negative sloped, while the safer one is reverse in behavior, this is a direct consequence of the survivorship bias
                                                     14
                          for the riskier company
                     o Limitations are:
                                    It assumed a unique zero coupon debt repayment
                                    It assume Brownian motion to describe the evolution of the price
                                    It doesn’t consider the case of early default, before maturity, meaning it doesn’t recognize the possibility that some
                                     evolution path may overcome the threshold before time T
                                    It assumes constant risk free, that can be easily solved
                                    It doesn’t consider migration risk
                                    Many of the variables need to be estimated since are not observable
                                    It is an arbitrage free approach, hence there should be the possibility to trade those asset, but it is not the case
              The KMV model is an attempt to overcome some of the Merton’s limits, namely the first problemand the estimation one
                     o Characteristic: it assesses the value of equity as equal to a call with same maturity of residual life of debt. Thanks to this idea it
                          can estimate the value of the company and the volatility, it use the price of the European call and the Ito’s formula
                          transformation to find the equation to estimate the volatility (2 unknown 2 equation)
                     o The KMV approach takes two steps
                                    Risk index is computed using the innovative concept of distance of default. Thanks to this innovation they can
                                     acknowledge the existence of short term and long term debt, hence the company may default only if the assets value
                                     drop below the short term debt



                                      This index is then converted into a probability using empirical law
                     o       Benefits are the speediness to adapt to change in the financial condition, it is stable compered to economic cycle and it allows to
                             assign specific EDF
                     o       Limits are that it can be used only for traded company (it can be solved by using comps) and it relays on the efficiency of the
                             market

EAD Estimation:

EAD is the exposure at default, it is generally assumed to be a deterministic number, but it may be stochastic if the borrower has the right to change
                                                                                 15
his size exposure. In this case we need to assess the drawn and un-drawn part . A Synonymous of EAD is recovery rate, it computed by applying
different formula referring to gross, actual and loading duration
                     o                               Gross method
                     o                                        Actual method, using for each flow its maturity to the occurrence of the default


13
 This model is assuming a risk neutral approach, meaning it starts with the equality between a risk free rate and the risky asset weighed by its survivorship probability
14
 Over time the pool of the weaker company will get stronger and stronger due the repayment of debt and the death of the worst among them
15
 The un-drawn part needs to be estimate considering the portion that will be drawn at default, to ease this task bank requires fees on this part that allows a more rational pricing (common in
Anglo-Saxon countries)
                                                                                                                                                                                             6
Giulio Laudani #23                                                                                                                           Cod 20163
               o                                           Loading duration, where “d” is the difference between RA duration and EAD duration
                                                      /TOT_PV(EAD) where “t” is the time elapsed from default
                             Where RA is the recovered amount, AD the administrative and legal cost, r is the appropriate discounter factor and d is
                              the duration of the process taking into account intermediate flow



LGD Estimation:

LGD is the expected loss rate in case of default, as PD it depends on the definition of default chosen by the bank, in fact the narrower the definition
the lower the LGD. Its value depends on those drivers:Characteristic of the borrower: industry sector, country  speediness of recovery procedure,
financial ratio; Characteristic of the exposure: presence of collateral or guarantee, priority level; Characteristic of the bank facilities to recover loan
and out of court settlements, and External factors, such as economic cycle [common with PD], interest rate level
           There are different procedures to estimate it, however all of them have to take into consideration all the indirect cost beard by the bank to
           recovery the proceed and the time needed:
                 o Market LGD is similar to the PD estimate using market rate, in fact knowing the other parameters we can infer the related LGD
                 o Work out LGD consists in building an historical database from where we could extract the LGD for a given type of borrower. This
                     method is the only applicable to bank’s loan which doesn’t have secondary market on which we can extract the market value
                     after credit events
                               We need to find about an appropriate discount factor of the recovery amount, which must be forward looking since
                                the procedure will be start in the future (after the beginning of the credit line)
                               We need to estimate the duration of the process itself, the book gives an example page 350. Note that this model is the
                                preferred one
           It is common in the risk management to find out some key variables which allow understanding the different recovery rate empirically
           experimented. This topic is important to understand the recovery risk, given the high volatility presented into database: It has been noted
           that the distribution is concentrated in the tails; The industry sector is a key element to explain differences; The priority level is
           important, but is not stable over time; and The presence of collateral is important
           Recovery risk is the risk to achieve a different LGD from the expected one, it is usually quite sizable and fluctuate over time due to a
           binomial distribution concentrated in the tails
                 o This risk arise from the uncertainty of those variables:
                               Amount to be recovered (EAD)
                               Administrative cost to be beard
                               Discount rate for future cash flow at future date
                               The duration of all the process
                 o It is calculated using those two formulas
                             Assuming a non-stochastic LGD 
                             Assuming a stochastic LGD

The link between LGD and PD must be taken into account, since they share common systematic factors. This relationship is negative, as empirical
evidence based on junk bond had proven. Doesn’t consider this risk may lead to an underestimation of risk. It is affected by:Chain effects: economic
downturn may reduce the value of assets; Financial asset and interest rates together with real estate one, and Industrial specific: the inventory may
lose more value for certain industry

The measure of the Unexpected Loss:
The UL can be simple define by the volatility around its mean (EL) or by some sort of CREDIT model based on VAR approach. This kind of exposure
can be effectively reduced by setting up a diversification policy.

We need to choose appropriate time horizon and confidence level: Time horizon is arbitrary chosen equal to 1 y, because both subjective and
objective criterions are not applicable due to the presence of illiquid market, of contract without explicit maturity and so on. However this decision
has a great operational fit:The bank use yearly budget, hence having yearly measure ease the budgeting work out; Usually 1 y it the correct time
elapse to raise new capital; The turnover of the asset is usually close to one year andFurthermore some model will account for long term risk
including migration risk in their computation. Confidence level may consider that the data cannot be explained by a zero mean normal distribution,


                                                                                                                                                         7
Giulio Laudani #23                                                                                                                           Cod 20163
since the mean is below zero and data are strongly asymmetrical.Hence the choice of Confidence level need to be credible, because it is less stable
and change a lot at different level

Tools used to estimate UL:

It is estimated using portfolio models that apply different definition of default (including or not migration) and level of loan correlation (explicitly or
implicitly modeled) as well as EL
           Creditmetrics is a multinomial approach it considers all borrowers’ change in rating as a credit events
                o The model basically consists on deriving the empirical distribution for any rating class movements, and byusing the respective
                     spread for each maturity to compute the expected value of the assets, and on doing the difference between this value and the
                     face value to compute the expected loss
                              We should compute the VAR percentile using the probability for each possible movement in the rating level to have a
                               measure of UL, since the distribution is not normal ( we cannot use the normal standard deviation time percentile
                               approach)
                o To overcome the Wilson’s critique regarding the accountability of the economic cycle it has been introduced the credit portfolio
                     view which is taking into consideration the economic variables
                              Note that if the time horizon used is a point in time there is no need to adjust by economic cycle, in this case there
                               would be a double counting
                o To use the model for portfolio of assets we need to estimate the correlation, by using those steps:
                              A modified version of the Merton model, where we find out in the normal distribution, all the thresholds for default
                               and change in rating. Using math integral we can compute all the probability of change marginal and cumulative for
                               each assets
                                         We pool together all these info with a bivariate normal (given the correlation) and the new distribution will
                                        give us all the joint probability
                                        Asset correlation, which is the one used in the precedent formula is computed using large building blocks
                                            o Correlations are first estimated among a large set of industries and countries (“risk factors”), For
                                                  each borrowers, a set of weights must be specified, expressing their sensitivity to different risk
                                                  factors and to idiosyncratic risk
                                            o Combining those weights and the risk factor correlations, an estimate of the pairwise correlation of
                                                  two firms can be obtained
                                        It is clever, because Asset value changes – unlike the distribution of losses – can be described by using normal
                                        distribution, it can then be described through a double normal distribution with correlation parameter ρ,
                                        hence from this distribution, the joint probabilities of the values of the two loans can be inferred, taking into
                                        account the transfer of risk
               o     Benefits are:
                               Uses objective and forward looking market dataInterest rate curves and stock indices correlations
                               Evaluates the portfolio market value
                               Takes into account migration risk
              o Limits are:
                               Needs a lot of data: forward rates, transition matrices
                               Assumes the bank is price-taker
                               Assumes stable transition matrices
                               Proxies correlations with stock indices
                               Maps counterparties to industries and countries in an arbitrary and discretionary way
          Credit risk + is based on an insurance approach using Poisson distribution to estimate probability of default
              o The idea behind this model is the equivalency of the bank and insurance risk, the major difference is the correlation among
                     bank’s clients. It aims to assess the risk portfolio, it doesn’t give us PD, its input are: PD and its volatility
              o We use a banding methodto link default probability with losses. It consistson creating category of similar expected losses and
                     computes the number of default for each band, each of those will be used in the Poisson distribution, It is advisable to use,
                     tosum up, the default number weighted with the ratio of j-esimo amount and band expected loss
                               The creation of category is made by dividing all exposures “Li” by L and round them up, getting standardized values
                                “vi”. The recovery rate is use to determine the exposure (used in the banking) in a deterministic way


                                                                                                                                                         8
Giulio Laudani #23                                                                                                                         Cod 20163
                            Each of those Poisson distribution, one for each bands, must be aggregate. However this first model assumes
                             independency between variables, but this can be overcame by assuming that the “number of default” for each band is
                             random (see below)
              o    The Poisson distribution assumed independent value and it works only for small PD. To estimate correlation we will run “n”
                   simulation within all the bands [assuming that the PD are stochastic] and then we made the weighted average conditional to all
                   the scenario occurrence and this new distribution will be the unconditional distribution, accounting for the asset correlation
              o    Benefits are:
                            PDs and exposures (book value) net of recovery are enough. The “correlated” version requires also sensitivities to the
                             economic cycle factors
                            An analytical solution exists, hence it is fast to be implemented
                            Possibility to obtain the distribution of losses without recurring to simulation techniques
              o    Limits are:
                            Only looks at default risk, so it Does not consider migration risk
                            Assumes constant exposures and doesnotconsiderrecoveryrisk
                            It isn’t a dynamic model. Meaning it cannot be used for changing portfolio without recomputed all the calculation

Comments on the UL tools employed:

Some comments on the major characteristic of those models:
        Default-mode versus multinomial one, where only credit risk + belongs to the first
        Future values vs. loss rate, meaning that the model can be based on the distribution of possible value or possible future loss, the first use
        as input the spread curve by maturity, while in the second the spread it is not necessary to be known. Creditmetrics is a typical market
        value model, while credit risk + loss one
        Conditional vs. un-Conditional, portfolio views belong to the first, however this distinction is useful only if the model works through the
        cycle
        Monte Carlo vs analytical solution
        Asset correlation vs. default correlation, it’s less important than other, in fact they are close to each other. Creditmetrics is belongs to
        asset correlation, while credit + to the second one.

Some major limits:
        Treatment of the recovery risk is not random (besides for credit +) and independent from PD
        Assumption of independence between exposure risk(usually treated as known) and default, while they are empirical positive related
        hence this assumption brings to an underestimation
        Assumption of independence between credit risk and market risk
        Impossibility to back testing due to yearly frequency, there is not enough data sample

Application and General comments on Credit Risk methods:
         Loan pricing by assessing EL and UL using a transparent process to properly consider the cost of capital absorbed. There is the problem to
         attribute the marginal benefits
              o                         ELR is the expected loss rate, it is the spread to compensate for the EL
              o                         VaR is a relative measure,       is the cost of equity
         Risk adjusted performance measurement is used to decide to undergo a specific investment
              o                                 where     is the lending rate applied, this number will be compared with the bank’s RAROC
         Setting limits on risk taking of the different business units, however it is crucial to properly define the appropriate level of aggregation for
         units and the VAR limits and frequency with which are checked as well as its involvement in the budgeting process
              o                                this formula will give the max loan amount that can be granted
         Optimizing the portfolio composition, this is limited due to bank’s loan characteristics (geo, no secondary market and limited rotation).
         However thank to recent derivatives development and secondary market it is now possible, the best solution should be to divide the Risk
         management optimization form the origination process




                                                                                                                                                       9
Giulio Laudani #23                                                                                                                          Cod 20163
Market risk
Market risk is usually identified as the risk inherent in the trading book (short term), but it should be extended even for those investments intended
to be kept in the financial statement for longer period. Nowadays, it has gained more importance due to the new accounting principle, the
securitization process and the growth of derivatives instruments. The key elements of this type of risk are:Exchange risk; Interest risk (different from
the previously one because it is related only to securities which have a secondary market and it will affect only a limited part of the balance sheet);
Equity risk; Volatility risk andCommodity risk

The traditional approacheswere:Nominal value which considers the risk as proportional to the nominal value. It has severe limitations such as: the
nominal value doesn’t reflect MV, it doesn’t capture the different degree of sensitivity to a change in the risk factor and it doesn’t consider volatility
and correlation; Sensitivity analysis is based on the usage of coefficient representing the risk for each securities category. However there still
drawback such as: you cannot aggregate those coefficients, it is an easy to communicate between division and senior since each position has a
unique measurement and it doesn’t consider the volatility and correlation (as nominal value method), meaning that without taking into
consideration the volatility of risk factor is basically not considering the real risk of the position

Tools used to estimate market risk
VAR models are characterized by: confidence level, maximum potential loss that a portfolio can suffer and a certain time horizon. This measure is
comparable between all securities class. These models aim to define the risk factors, the probability distribution of those risk and to summarize those
information in one risk parameter
     The easiest one is the parametricone: normal distribution of change in value in market factors, all possible information are summarized by Var-
     Cov matrix, the possible loss are correlated to the risk factor by a linear function and VAR is simply a multiple of standard deviation.
         o The most crucial hp are the one that ignore those problems:
                         Empirical data shows that the distribution of risk factor is skewed and with fatter tails
                         It works by simply reverting the normal distribution into density function from where we can compute the equivalent
                          percentile. [MV*σ*α*β]
                         The linear relationship is represented by the sensitivity coefficient which (for example) is the modified duration for bonds;
                          this is a deltanormal approach. Alternatively we can assumed that the prices are log normal distributed, hence there is a
                          one to one correspondence, this is an asset-normal approach
         o Typical issues are:
                         The confidence levelchoice represents the degree of risk aversion of the financial institution, or the target capital
                          requirement that allows to have on balance sheet the investors’ expected creditworthiness (investment on it) , in fact there
                          is a positive empirical relationship between those value
                         Time horizon is usually a short term measure (daily), the choice of the appropriate level is crucial since the higher the time
                          length the higher will be the VaR value
                                      The bank must take into account the liquidity level of its position, the size and a subjective idea on the time that
                                      the instrument will be in the trading book
                                      It is important to have enough data observation to ensure significance on the value. Hence for longer time
                                      horizon there could be problems to achieve this. Sometimes to overcome the problem it is possible to use the LRW
                                      hp to convert daily data into monthly or weekly
         o When we apply this method to portfolio we need to compute the correlation between asset, which is simply the matrix of Corr pre
               and post multiplied by the ordered securities VAR. This formula will grant diversification benefits, but will arise sub-additivity property
         o When we want to use multi risk factors we need to break down each security into elementary components which depend solely on
               one risk factor, and then aggregate them as a portfolio. The approach suggested is the mapping of risk position:
                         In the case of a foreign currency bond we have two risk components, we will proceed by computing each single VAR and
                          then sum them up using their correlation
                         Forward currency position: three positions, exchange rate spot and a spot investment/borrowing amount
                         Forward rate agreements consists of two components a debt spot position with a maturity prior to the investment spot one
                         Stockpositions we can consider each stock as a risk factor or use fictitious position toward relevant stock indexes, hence the
                          position is mapped to the relevant betas. It works only for well diversified portfolio without specific components
                         Bonds are instead mapped for each of their cash flow (clumping)
         o Limitations of this model are [note that this model is better if considering single risk profile]:
                         Risk factors change are assumed to be normal distribute

                                                                                                                                                      10
Giulio Laudani #23                                                                                                                       Cod 20163
                       Since the assumption of risk factor is to be explained by Brownian motion the volatility is assumed to be stable over time
                        and serial independence across risk factor, it is not empirically true
                       The relationship with loss is assumed to be linear
         o    Solutions provided are:
                       The possible solution to normal distribution hp is to use a mixture of Gaussian distribution (to solve the fait tale problem
                        not the empirical asymmetry) or t-student
                       The non-linearity can be solved by using a curve parameter, but this solution poses the problem of not normal distribution
                        of the underling changes since the curving coefficient is a chi-Square function. Furthermore by introducing a new parameter
                        we are increasing the model error, and, above all, there is the underling hp that the payoff are derivable, otherwise it is un-
                        applicable. This approach is less effective for joint shock. Another solution is using full valuation approach


Simulation modelallows to use different hp compared with the previously one:the risk factors may have different distributions, since the impact of
each risk factor change is value by full valuation it is possible to have a different relationship between loss and risk factor and it is a way to ensure
flexibility. Hence, This model are best used in case of nonlinear payoffs and to value extreme event [stress test]
            o Full valuation is a way of estimating price variation is not a simulation model, it consists on re-price the securities applying a price
                 formula (that must be define a priori)
            o All distribution are empirical, meaning based on the simulation. Here there is the difference between historical and Monte Carlo
                 where the fist use empirical distribution obtained by the observation while the second define a parametric distribution
            o Use of the Percentile logic, meaning that we will generate scenario by applying a given distribution and on those scenario ordered we
                 will compute the interesting percentile, hence the VaR will be the difference between the asset current value and its percentile value
            o There is great flexibility on defining the market risk change behavior
      Historical simulation transforms the historical data into future possible behavior, hence it uses the past distribution to predict future one
      without change
            o Merits of this model are:
                          It is simple to be understand and communicable
                          It doesn’t require hp about distribution or correlation (not Var-Cov required)
                          Non linearity required between risk factor and price change
            o Limits and possible solution are:
                          It assumes a stable and stationarity distribution given by past data
                        It may not applicable if the time series are limited or if there is the problem of the usual tradeoff between time length and
                           meaningfulness, however using ahybrid approach putting together the exponential weights and the simulation approach we
                           can use longer sample ensuring meaningfulness. Furthermore this hybrid approach allows the model not to have stable
                           distribution, note that is an approximation, since the current data have higher weights any past change in behavior is
                           smoothed
            o The methodology of bootstrapping and path generation are used to obtain bigger sample size
                          Bootstrapping is made to avoid loss of observation when we move from daily to weekly or monthly frequencies. It consist of
                           an extraction (allowing for reinterring) of observations, in this way we will built X paths (useful in case of exotic option
                           pricing when we need to know the path is important in defining the price) from whom we will extract the distribution. The
                           underling hp is that observation must be i.i.d
                          To try to overcame the hp of i.i.dwe can used something similar to the hybrid approach or there are two better proposals:
                                      Hull and White suggest adjusting data by weighting with current volatility, meaning by proportional move returns
                                      value. Heteroskedasticity can be incorporate by rescaling the time series with the available information at time T
                                      Filtered historical simulation is based on G-ARCH models which are used to filter data and to obtain residuals
                                      which are used to create scenario. Each residual is standardized (hoping that doing so they are i.i.d) and used in a
                                      path generation process where the first data is the filtered return, not the sample one, and it is multiply with he
                                      estimated conditional value of next period (predicted volatility by G-ARCH times previously residual)
      Monte Carlo simulation is based on the definition of an a priori parametric distribution, which should be consistent with the future data
      behavior. This model is more computing efficient since by increasing the parameters involvedthere is a proportional increase of variables
      number, but still demanding
            o Differently from historical simulation, it requires to compute or to assess the correlation between risk factor a priori (otherwise is like
                 assuming independency), but it allows to know the path evolution
            o To take into account the relationship we need a Corr matrix which is decomposed into two triangular matrices and used them to build
                 up scenarios
                                                                                                                                                   11
Giulio Laudani #23                                                                                                                    Cod 20163
        o Limits are the needs of an asset based estimation of Var-Cov to find out the joint distribution
   Stress testaims to estimate the effects connected with extreme events, meaning that the simulation is carry out following predominantly
   arbitrary and subjective manner just for few scenario
        o Use past shock and simulate them
        o Factor push analysis meaning use more standard deviation movements, note that there is the problem of great variation of
              insignificant risk factor are not significant
        o Derivatives policy group suggestion it is uni-dimensional (meaning each variable at time) process suggested on interest rate
              movement in the slope or shift as well as on other macro indicators
        o Multidimensional scenario can implement into two ways:
                        Simple where only some risk factor are let change while the others are kept constant
                        Predictive is the same of the previously one, but the others are changed according with their correlation to the moving one
        o This instrument should be a complement to the previously ones and must be followed up by practical actions to reduce risk, hence
              vulnerability. It allows to test liquidity risk

Different applications of the Market risk model:
   Unique riskmeasure for both horizontal and vertical communication between division
   Portfolio analysis is possible: there are several method to aggregate risk factors
   It is useful to determine the risk limit exposure both as nominal value, market value exposure (remembering that those measures will be affect
   by change in volatility) and maximum tolerable variation.
   It can be used to as a mean to risk adjust return by using RAROC [both ex-ante and ex-post], which is a profit/risk ratio, thanks to this
   instrument it is possible to monitor the units return, properly set incentive procedure




                                                                                                                                                12
Giulio Laudani #23                                                                                                                        Cod 20163
Measure of Volatility
Volatility estimation can be divided into two groups: historical with constant parameters and those with changing one and implied option. It is
important in the Var-Cov model, since it is a crucial parameter. For each of the subsequent model we need to define the time horizon/frequency
    The simplest is basically the computation of the Var-Cov matrix using an equally weighed average or an exponential one “Riskmetrix”using as
    data source historical dataset
           o Simple moving average model poses two problems:The sample size: the longer the higher the information, but it will be less reliable
               for prediction purpose, since it doesn’t reflect current situation;and Echo effect meaning that any shock will strongly affect the value
               both in the entrance and in the exit from the sample
           o Exponential moving average will overcame the previously problems, but it poses the problem of choosing the appropriate decay
               factor (which should depend on the data behavior, i.e. how fast they react to change, but it must be noted that it changes overtime)
               and the number of past observations must be evaluated, this problem is solved increasing the frequency so that to maximize the info
               content and minimize the error sampling
    G-ARCH model stands for: Heteroskedasticity, Conditional and Autoregressive. It uses the maximum likelihood criterion, hence it use market
    data to get a better estimate of the decay factor, and however it needs lots of datapoint.
                         The two common factors are:
                                     Past volatility  it indicates the rate of persistency. It is above 0,7
                                     Past prediction error square  it indicates the rapidity with which volatility adapt to new market shock. It lower
                                     than the first one
                         Benefits of this methodology are:
                                     It recognizes the serial correlation
                                     It gives adequately importance to new information
                                     The decay factor is directly determinate by market data
           o It use a normal distribution to describe the behavior of the prediction error, hence it is a poor method in case of skewness or
               leptokurtosis. The later problem is overcame by using t-student distribution while the first one (which come out of the fact that this
               model gives same importance to the sign of shocks, which is against empirical date)is solved by models that recognize this asymmetry,
               i.e. the negative effect has an higher impact:
                         IG-ARCH require that the sum of the coefficient is one, hence if the constant is set to zero is basically the exponential
                          moving average formula
                         EG-ARCH models the natural log of the variance (allowing the equation to give negative output), so that instead of square
                          value it use absolute value and real value (response to good and bad is consider)
                         AG-ARCH use the square, but it centered the data using a parameter which will amplify the negative effect an reduce the
                          positive one
           o It gives good measure for the immediately following period, but it is less informative for subsequent, in fact, since the parameters
               should guarantee mean reversing or better a converge series, it converge to the long term value
    Implied volatility is basically the volatility taken out form derivatives instruments, hence it represent the market expectation on that period, not
    really used in risk management
           o It may be affected by counterparty risk or by the liquidity level
           o It will change depending on the contract chosen (at/in/out the money), the maturity and the model used to price the option (it must
               be reliable and the instrument must be liquid to ensure efficiency)
           o The maturity and the time horizon for risk management system should coincide to ensure consistency
           o Computing covariance value is complex and we may not have data, moreover we need to check if the numbers are consistent
               between them, the matrix must be at least semi-positive. There are two way: the first by using derivatives with more than one
               underling such as quanto option, or using page 182

The general limits of the VAR approach are:

     It doesn’t define the behavior in case of extreme events higher than the thresholds (however it recognizes them); however the purpose of the
     VAR isn’t to make the bank completely safe from bankruptcy, but to reduce this risk to an acceptable level. Remember that higher capital
     requirement will lower the net income
          o However VAR fails to gives us a measure (probability) that allows to discriminate within excess lost between portfolio, it isn’t possible
               to understand which is higher


                                                                                                                                                    13
Giulio Laudani #23                                                                                                                     Cod 20163
   The VAR won’t take into consideration the customer’s relationship or any other qualitative indicator, however we can post that the bank is not
   force to use VAR as the only method to allow credit line, it may be the case that the management use other qualitative methods
   Some of the assumption behind VAR models are questionable, however if they are understood they can be overcame, furthermore VAR it is a
   good tool to allocate resource between units according to risk hence it gives more insight compared with other tools
   DifferentVAR approach can lead to different value, hence it can be seen as a weakness, however it must be noted that since the results highly
   depends the model there could be the cause that the underling hp of some of those approaches are not appropriate. Furthermore if there will
   be any bias it is uniform for all the securities, hence the relative risk allocation or perception is unchanged
   VAR is a too cyclical measure (it is different for each method), however the traders’ behavior follow this trendduring financial crisis
   VAR comes too late, meaning it is not able to predict crashes, however VAR purpose isn’t to predict crashes (which is impossible for any
   historical predictive method), but to generate consistent and uniform risk measure during normal conditions for day by day business regulation
   VAR has a sub additivity property, diversification allows to reduce aggregate VAR, however it is not always the case (such as for non-parametric
   models), hence it may strongly underestimate risk.
        o This last problem can be solved by a different measure proceed: Expected shortfall:This method allows to have an idea of the
              possibility to have excess lost compared with the VAR lost, it is the average of the excess lost beyond the VAR thresholds. It ensure
              sub additivity and it gives us a measure of how much money are needed to bail out the bank for the supervisory authority or the
              expected payment that an insurer have to pay against bearing this risk
   Using the Backtesting methodology we will be able to check the performance among different VaR models and to check the consistency with
   confidence level. The method is basically counting the exceptions to test independence among them and to test the confidence of interval:
   there are two test proposed:
        o The Kupiec Test: the null hp is that the model is good performing
                        We will compute the LR test between two binomial distribution, the numerator without conditioning (hence using the
                         theoretical percentile implemented), and the denominator conditional to the empirical observation. The test is distributed
                         according to a chi-square with one degree of freedom
                                                                    conditional coverage
                     Limits of the test are: it doesn’t account for serial dependence (meaning it doesn’t allow to value a model for its capability
                      to avoid time concentration of excess losses, hence it doesn’t value the quality of the model to react to change in the
                      market conditions); it requires lots of data, the power of the test is weak (and it decrease with the confidence level and
                      number of data)
       o    The Kristofferson Test: the test is willing to check that all the probability of exception and non-exception occurrence is not correlated
            to the previous occurrence
                     The LR test: it is the ratio between the Likelihood function assuming independency and the perfect correlation, it is possible
                      to joint test this LR test with the Kupiec one by summing up them, this new test will be distributed according to a chi-square
                      with two degree of freedom
       o    The paper of Saita and Sironi provides an empirical analysis using 7 equally weighed portfolio (equity, global) to test a Garch(1,1)
            model a Historical simulation test and an EW Moving average (lambda =0.94)at three confidence level
                     All the three models have failed to grant a consistent confidence level at 0,99 and 0,97; while at 0,95 the EWMA has been
                      more conservative and the other two still underestimate
                     The Kupiec test has well performed (only at 0,95) for some cases three for EWMA, one HS an two Garch
                     The extreme event have shown a higher vale than the theoretical one
                     The Kristofferson test has failed for basically all the cases, it has performed better at 0,99
                     Internal model based on those models have been more capital demanding the HS is the most conservative due to the abrupt
                      change
                     The analysis wasn’t able to define a winner, it was limited only for equity with daily horizon




                                                                                                                                                 14
Giulio Laudani #23                                                                                                                                                               Cod 20163
Basel Committee Framework:
This agreement is the result of the meeting of the G10 central bank governors to set up a common banking standard to ensure solvency of the
industry and avoid unfair completion based on different capital requirement by national supervisory authority, hence it aims to uniform the bank
capital requirement (it applies to consolidated account, hence it would foster the soundness of institution controlled by foreign banking group) and
to prevent future bank crisis (it would revert the bank trend to reduce capital). It is an ongoing work trying to achieve a best practice to ensure a
global stability to all the financial institution.

Basel I:
            Capital requirement computation is focus on the presence of three different tiers, which need to sum up to 8% i.e. the minimum
            requirement bound. This measure is net of goodwill and investment in non-consolidated bank or financial institution
                 o Tier 1is the most important and valued part of the capital. It consists of an upper tier: shares paid-up, disclosed reserves and
                      certain general provision (in UE the use of those items has been limited only for specific risk, and they have to respect: after tax,
                      immediately available and size, distribution and provisions must be shown separately) and the lower one: innovative instrument
                      which respect: un-redeemable, permanent, callable by issuer’s will (after 5 y), must be junior to all other instrument, absorb loss
                      without liquidation procedure and the remuneration can be deferred or if impossible forfeited
                 o Tier 2 is the first category of supplementary capital, it can be used at most for 50% of total capital. It consists on:
                                Undisclosed reserves meaning all those reserves which are created using revenue from off-balance sheet, however
                                 they must be solid as the disclosed one
                                Revaluation reserve (sizable in German and Japan), they are the difference between historical and current market
                                 value (they must come from a revaluation process) and can be used at 45% of their value (due to possible negative
                                 variation)
                                General provision, however they have been limited by the new accounting principle and must account for at most
                                 1,25% of the total capital
                                Hybrid capital instruments don’t need to put the bank under liquidation to absorb loss, remuneration can be waived or
                                 reduce and cannot be redeeming by credit (only with supervisory authorization). It must unsecured (as all this
                                 instruments) and fully paid up
                                Subordinated term debt, the main difference with the previously one is that to be liable of any loss there must be a
                                 liquidation procedure. The condition that those instruments have to meet are: maturity longer than 5 y, they must be
                                 “depreciated” at 20% each year and must be junior
                 o Tier 3 it cannot be accounted for more than 250% of tier 1 for market risk and it cannot exceed 50% of total capital requirement.
                                                            16
                      It includes short term subordinated debt and can be used to cover only market risk.
            Risk weights in the first Basel agreement are quite simple and “linear”. They are:
                 o 0% for cash, Government bond in the OCSE, claims on central bank
                 o 20% claim on central bank or country outside OCSE, claims on bank less than 1 y and multilateral development banks
                 o 50% loan secured by mortgage on residential
                 o 100% the others
                 o The OTC items pag. 555
            The major Limitsare:
                 o Focus on credit risk only, there is no consideration for currency risk or others (1996 solved)
                 o Poor differentiation of risk, there is no consideration for this issue on the weights, furthermore the class defined by the weight
                      allows for regulatory arbitrage, they aggregate too much
                 o Limited recognition of the link between maturity and credit risk as well risk mitigation instrument

Basel committee II
            Pillar 1 capital requirement has receipt all the Basel 1 criticism and it had tried to improve the overall capital system. It must be noted that
            the capital requirements are not fix, but can vary following the supervisory expectation both on the capital and the risk perceived by
            internal rating method (by multiplying them by a scalar factor)
                  o Risk weightsfor the normal one are related to the rating agency with some improvement, acknowledging the


16
 It must be at least 2 y, cannot be redeem, there is a lock-in if the capital ratio follow below the minimum plus 20%, and it must be reduce by loan loss forecast and security for trading one
                                                                                                                                                                                                  15
Giulio Laudani #23                                                                                                                   Cod 20163
           o      Internal rate is the possibility given to certain bank to use internal procedure to identify and estimate the main component of
                  risk, there exist two different versions:
                            Fundamental allows only to make estimation on PD, while the other EAD, LGD and maturity are chosen by the national
                             authority
                            Advancedgave the rights to use own estimations for all the risk components.It can be used if those condition are met:
                             7 rating classes, where the first must have a PD of 0,03% and there must be a default category
       Pillar 2 has been made to enforce the supervisory capability to control the bank system. This approach ensure a proactive prudential
       behavior based on those principles
             o Bank must set up a system of process and techniques aimed at establishing the overall capital adequacy
             o Supervisory authority must check those process and ensure the respect of the minimum capital requirement and promptly
                  intervene to avoid capital deterioration
       Pillar 3 has been done to reduce the opacity and to force the market to discipline unfair behavior and to promptly penalize bank which are
       taking more risk. The bank is so forced to publish info regarding: their economic and financial results, financial structure, and their risk
       management strategies, exposure to risks, accounting policy and corporate governance.




                                                                                                                                               16
Giulio Laudani #23                                                                                                                             Cod 20163
Insurance business

Solvency II
It will take 8 years to totally implement the new framework started in 2005. There have been 5 insurance conferences (QSI) which has seen a wide
participation. The main features are:
           This framework aims to enhance a more consistent standard across EU and ensure that capital requirements are more reflective of risks
           undertaken by insurers
                o Market consistency approach for valuing all assets and explicit pro-cyclical provisions
                                Asset and liabilities [non insurance one+ should be valued at price “arm’s length transaction” with reference to IFRS
                                 and no recognition of change in credit standing for financial liabilities
                                Asset and liabilities [insurance one] should be value with model (best estimate)
                                          Price needs to be the same irrespective of investment strategy
                                          Use of risk free rate adding a risk margin
                                          The pricing model consists on computing technical provisions based on
                                                 o Best estimate: PV of future cash flow, which are computed considering prudent and realistic
                                                       assumption, management and consumer possible future actions
                                                 o There is any more prudential provision (forbidden), all the guarantee, discretionary benefits and
                                                       option must be valued
                                                 o Reinsurance revocable must be valued separately and with allowance of credit risk
                                                 o The choice of which risk free rate poses some problems [EIOPA provides some guideline] such as:
                                                               Different currencies problem is solved using swap rate adjusted for credit risk and
                                                                providing by QIS5 spot rate curve for main currency
                                                               Different maturities need appropriate rate:
                                                                          Illiquidity asset: duration cap at 15 euro and 30 GPB and linearly reduce to 0 in
                                                                          5 years and rate adjusted
                                                                          The spot curve is calculated using a certain basket of corporate bonds
                                                                          The liquidity premium is cap at the max available on the market for similar cash
                                                                          flow without risk, and it should account for liabilities’ nature (it is provided by
                                                                          EU institution with the same frequency of interest rate)
                                                               The risk free rate are extrapolate using macroeconomic model to find out the
                                                                unconditional long term rate
                o Better recognition of diversification, risk mitigation and loss absorbing items
                o New supervisory approach, more proactive and EU coordinated among country and groups’ economic reality
                                Group complementary supervisory which have primary responsibility (can used internal model) the methods are
                                          Accounting consolidation (to eliminate double counting)
                                          Deduction and aggregation  recognition of diversification
                                Third countries relationship is articulated in three aspect
                                          The reinsurance supervision from equivalent third countries no difference
                                          Group capital requirement from equivalent third countries can be used their
                                          Group solvency from equivalent third countries, their authority is ok
                                          Transitional arrangements are put in place for important countries, that are not equivalent to EU regulation
           The reform aims to promote confidence and transparency among insurance
           It equally apply to reinsurers
           It wants to overcame the Solvency I limits:
                o Lack of harmonization
                o Inconsistency with new IFRS accounting principle
                o Capital requirement is not transparent and not adequate to risks, furthermore it was focus on back-looking aspect instead of
                      governance issue (good Risk management)
                o No recognition of economic reality of group (only plus requirement any reduction)




                                                                                                                                                         17
Giulio Laudani #23                                                                                                                          Cod 20163
The formation process and directive structure:

The formation process and authority involvement is basically EU lead by applying the Lamfalussy structure: L1 see the involvement of Parliament,
Ecofin e commission helped by special committee and L2 and L3 see the “joint committee” *ESMA, EIOPC, EBA+ check law. The directive structure
reflects Basel II/III’s pillar approach
          Pillar I risk calibration of financial requirement, hence it is focus on the quantitative issue
                 o The SCR is calculated in terms of potential loss of value at a confidence level of 99,5% over a 1 year time horizon considering all
                       quantifiable risks
                                 It should (internal model) account for even the liquidity premium risk
                                 Each risk is model individually with a modal approach using factor or scenario approach
                                 The individually risks are aggregate using corr matrix (provided by QIS5)
                                 It must be taken into account the absorbing capability of technical provisions (meaning future benefits: change in hp)
                                  and deferred tax
                                             Gross SCR are computed without considering benefits
                                             Net SC assumes bonuses to reduce/absorb loss
                                             The difference of the first two items will be FDB (future discretionary bonuses)
                                 Mitigation techniques and collateral and segregated assets are allowed under certain conditions (legally binding, actual
                                  transfer of risk…)
                                 Proportional recognition for techniques in force for less than 12 months are allowed under certain conditions( no risk
                                  for rollover and counterparty risk take into account)
                 o The MCR (minimum) is cap within the corridor of 25/45% of SCR
                                 Legal certainty, auditable and safety net protection
                                 It is based on percentages applied to combination of premium and technical provision
                 o Solvency II allow to choose which level of simplicity or sensitivity can be used to assess risk form the simplified method to
                       internal method
                                 The insurance must grants a sound framework to managing, controlling and measuring risk
                                 The authority can still request specific parameters for internal models
                 o The Own funds categorization:
                                 Basic Own funds: excess asset over liabilities, subordinated liabilities and adjustment (expected profits, net deferred
                                  tax and restricted reserves)
                                             They can be used for all tier capital requirement
                                 Ancillary Own fund (prior supervisory approval): off balance items that can be called up to absorb losses such as
                                  unpaid shares, letter of credit other legally binding documents
                                             It can be used to tier 2 and 3
                 o The capital eligible to meet SCR must be at least 50% tier 1, hybrid instrument can be at most 20% of tier 1 and tier 3 cannot
                       account for more than 15%
                 o The capital eligible to meet MCR must be 80% tier 1 no tier 3 available and no ancillary funds
                 o The criteria used to allocate OF in tiers are: subordination, loss absorbency, duration, freedom to redeem, absence of
                       encumbrances and of mandatory servicing costs
          Pillar II new supervisory relationship and setting up for governance
                 o Authorities are granted to take action to restore critical situation when SCR get closer to MCR, to ensure even in extreme
                       situation to preserve policyholders interest
                                 There is a convergence of supervisory standards
                                 Capital “add on-“ are authority demand to adjust risk assessing
                                 They can draft and implement measures, so that they can promote fairness act against breach of EU law and act in the
                                  emergency situations
                 o To take into account cyclical effects the authority can extend the recovery period, inherent rebalancing between SCR and
                       available capital, liquidity premium and equity symmetric dampener
                 o There are quantitative risk management standards, so that it will play a central role in the company
                                 Internal control system and internal audit and actuarial functions
                                 ORSA (own risk solvency assessment) specific risk profile, compliance with financial requirements, significance of
                                  deviation between the two
                 o There are new disclosure requirements bring market discipline to bear insurers

                                                                                                                                                      18
Giulio Laudani #23                                                                                                                       Cod 20163
         Pillar III opening up to market discipline provisions and reporting to authority
               o The disclosure to market must be at least annually and account for business performance, risk profile, system of governance,
                    capital management and valuation purposes
               o The Regular supervisor report to must be narrative with quantitative data, every 3 years but it can be asked to be provided
                    annually
               o Annual and quarterly quantitative template with a lag of 14 weeks and 5 weeks for the last one

The main implementing problems and open issues are: Liquidity premium, currency curve and how to allocate them to liabilities, There are too
many uncertainty and different interpretations, The QSI5 have recognize 92% of OF as tier 1, The group diversification has a low to free 20% of OF on
average which has created a huge difference with other non EU entities, Eliminating excessive complexity, reducing excessive volatility and reducing
penalization for long term business with guarantees (revising stresses for interest and spread risk at long duration); Improving calculation criteria 
contract boundaries and deferred tax and liabilities; Finding political compromises on political aspects  illiquidity premium, future profits,
diversification in risk margin and at group level; Designing appropriate transitional provision disclosure and reporting, internal model; Properly
manage the new task of supervisory authority (new tools, culture and power); and The impact of the new discipline will be huge, there is a change on
the overall business with extensive impact on governance, strategy, expectation of authority and new competitive scenario

Some example:

         Example of the usage of the aggregation of modules for SCR calculation purpose based on the market risk case, which is divided into
         several categories
              o Market investment must have a minimum rating of BBB as an overall rule
              o The equity risk module
                             Speculative are treated depending on the market on where there are listed and on CEIOPS Advice
                                          Duration based approach (22%) for certain lines of business
                                          Symmetric adjustments are made by EIOPA on a benchmark of listed company and are sufficiently public. The
                                          adjustment is cap between -/+ 10% and it is computed using a formula that take in consideration the current
                                          level and the weighted average of daily data over a window of 36 month
                             Strategic participations are stressed at 22%
              o The interest rate module
                             All asset sensible to interest should be stressed, as well as insurance liabilities in their PV change
                             There is no stress test on volatility of interest rates
              o Spread risk module consists in any change in the credit spread over the risk free rate
                             It is calculated on separate items (Bonds, Credit derivatives and ABS on loan)
                                          Derivatives which are used for risk mitigation are excluded
                                          Bonds and ABS are value using a table using rating, a conversion factor (larger for the last)
                             Instruments of EU with AAA or AA rating are excluded, and the other have lower capital requirement as well as
                              corporate AAA rating
              o Currency risk module
                             It can be calculated on net or gross position
                             It consider a max variation of 25%
                             No diversification benefits, but the EU currencies are advantaged with lower stress level
              o Property risk module, it the risk related to real estate
                             25% for all property (direct) investment in company in real estate is subject to equity module
                             No geographical diversification
              o Concentration risk module, it aims to consider the risk related to lack of sufficient diversification or to exposure to large default
                   risk (single issuer)
                             It apply to all the category of assets, considering only the exposures in excess of certain threshold
                             It depends on the exposure and rating of the counterpart
                             It is excluded participation internal the group, when the risk is beard by policyholders or when it is already considered
                              in the default counterparty module
                             It doesn’t consider geographical concentration
                             EU with AAA and AA no capital requirement
              o Liquidity risk module is new and it allows for change in illiquidity premium
                             Negative correlation with spread risk (-50%)
                                                                                                                                                   19
Riassunto risk
Riassunto risk
Riassunto risk
Riassunto risk

Más contenido relacionado

Destacado

1. understanding the financial markets
1. understanding the financial markets1. understanding the financial markets
1. understanding the financial markets
realsharetraders
 
Community participaion community participation1
Community participaion community participation1Community participaion community participation1
Community participaion community participation1
Subramaniam Jeyachandran
 
CDS and Loss function: CDO Tranches
CDS and Loss function: CDO TranchesCDS and Loss function: CDO Tranches
CDS and Loss function: CDO Tranches
Bank Industry
 
Final fathimanew[1]
Final fathimanew[1]Final fathimanew[1]
Final fathimanew[1]
Rohina Banu
 
Market abuse directive and market manipulation
Market abuse directive and market manipulationMarket abuse directive and market manipulation
Market abuse directive and market manipulation
Bank Industry
 

Destacado (20)

Schema econ
Schema econSchema econ
Schema econ
 
Schema anf
Schema anfSchema anf
Schema anf
 
1. understanding the financial markets
1. understanding the financial markets1. understanding the financial markets
1. understanding the financial markets
 
Schema fr
Schema frSchema fr
Schema fr
 
Schema cf
Schema cfSchema cf
Schema cf
 
Community participaion community participation1
Community participaion community participation1Community participaion community participation1
Community participaion community participation1
 
IELTS Vocabulary 2
IELTS Vocabulary 2IELTS Vocabulary 2
IELTS Vocabulary 2
 
Margrabe option
Margrabe optionMargrabe option
Margrabe option
 
CDS and Loss function: CDO Tranches
CDS and Loss function: CDO TranchesCDS and Loss function: CDO Tranches
CDS and Loss function: CDO Tranches
 
Final fathimanew[1]
Final fathimanew[1]Final fathimanew[1]
Final fathimanew[1]
 
Innovation & Creativity - Innovation Hotspots & The Six Thinking Hats
Innovation & Creativity - Innovation Hotspots & The Six Thinking HatsInnovation & Creativity - Innovation Hotspots & The Six Thinking Hats
Innovation & Creativity - Innovation Hotspots & The Six Thinking Hats
 
Video Games Market
Video Games MarketVideo Games Market
Video Games Market
 
Bailout.consulting.brevini
Bailout.consulting.breviniBailout.consulting.brevini
Bailout.consulting.brevini
 
network storage
network storagenetwork storage
network storage
 
Market abuse directive and market manipulation
Market abuse directive and market manipulationMarket abuse directive and market manipulation
Market abuse directive and market manipulation
 
Business Ideas for Housewives
Business Ideas for HousewivesBusiness Ideas for Housewives
Business Ideas for Housewives
 
IELTS Vocabulary 1
IELTS Vocabulary 1IELTS Vocabulary 1
IELTS Vocabulary 1
 
2011년 KTH H3 컨퍼런스 Track B, 세션4 "Advanced Git" by A.J
2011년 KTH H3 컨퍼런스 Track B, 세션4 "Advanced Git" by A.J2011년 KTH H3 컨퍼런스 Track B, 세션4 "Advanced Git" by A.J
2011년 KTH H3 컨퍼런스 Track B, 세션4 "Advanced Git" by A.J
 
Animal Bioacoustics by Steve Sens
Animal Bioacoustics by Steve SensAnimal Bioacoustics by Steve Sens
Animal Bioacoustics by Steve Sens
 
Making Joyful Noise With Your Computer For Fun and Profit
Making Joyful Noise With Your Computer For Fun and ProfitMaking Joyful Noise With Your Computer For Fun and Profit
Making Joyful Noise With Your Computer For Fun and Profit
 

Similar a Riassunto risk

Stochastic modelling of the loss given default (LGD) for non-defaulted assets
Stochastic modelling of the loss given default (LGD) for non-defaulted assetsStochastic modelling of the loss given default (LGD) for non-defaulted assets
Stochastic modelling of the loss given default (LGD) for non-defaulted assets
GRATeam
 

Similar a Riassunto risk (20)

SIMM_ISDA_WPv7GMS_SEPT16
SIMM_ISDA_WPv7GMS_SEPT16SIMM_ISDA_WPv7GMS_SEPT16
SIMM_ISDA_WPv7GMS_SEPT16
 
Methodologies of Margin Setting for Exchanges
Methodologies of Margin Setting for ExchangesMethodologies of Margin Setting for Exchanges
Methodologies of Margin Setting for Exchanges
 
Article cem into sa-ccr
Article   cem into sa-ccrArticle   cem into sa-ccr
Article cem into sa-ccr
 
Counterparty Credit RISK | Evolution of standardised approach
Counterparty Credit RISK | Evolution of standardised approachCounterparty Credit RISK | Evolution of standardised approach
Counterparty Credit RISK | Evolution of standardised approach
 
Blog 2017 01_irrbb - basel irrbb guidelines
Blog 2017 01_irrbb - basel irrbb guidelinesBlog 2017 01_irrbb - basel irrbb guidelines
Blog 2017 01_irrbb - basel irrbb guidelines
 
MH_V3_samples_ch2
MH_V3_samples_ch2MH_V3_samples_ch2
MH_V3_samples_ch2
 
EAD Parameter : A stochastic way to model the Credit Conversion Factor
EAD Parameter : A stochastic way to model the Credit Conversion FactorEAD Parameter : A stochastic way to model the Credit Conversion Factor
EAD Parameter : A stochastic way to model the Credit Conversion Factor
 
Blog 2016 13 - impairment modeling in retail - many moving parts
Blog 2016 13 - impairment modeling in retail - many moving partsBlog 2016 13 - impairment modeling in retail - many moving parts
Blog 2016 13 - impairment modeling in retail - many moving parts
 
UNEP Finance Initiative (UNEP-FI) - Jan 09 - UBS Global Warming Greenhouse I...
UNEP Finance Initiative (UNEP-FI)  - Jan 09 - UBS Global Warming Greenhouse I...UNEP Finance Initiative (UNEP-FI)  - Jan 09 - UBS Global Warming Greenhouse I...
UNEP Finance Initiative (UNEP-FI) - Jan 09 - UBS Global Warming Greenhouse I...
 
Early warning system_ white paper
Early warning system_ white paperEarly warning system_ white paper
Early warning system_ white paper
 
Mrotek Cullinane Feature Article Iasa Interpreter Summer 2010
Mrotek Cullinane Feature Article Iasa Interpreter Summer 2010Mrotek Cullinane Feature Article Iasa Interpreter Summer 2010
Mrotek Cullinane Feature Article Iasa Interpreter Summer 2010
 
FRTB Technology Overview
FRTB Technology OverviewFRTB Technology Overview
FRTB Technology Overview
 
The Interest Rate Risk on the Banking Book
The Interest Rate Risk on the Banking BookThe Interest Rate Risk on the Banking Book
The Interest Rate Risk on the Banking Book
 
How to manage Interest Rate Risk in the Banking Book considering the monetary...
How to manage Interest Rate Risk in the Banking Book considering the monetary...How to manage Interest Rate Risk in the Banking Book considering the monetary...
How to manage Interest Rate Risk in the Banking Book considering the monetary...
 
Bank Capital
Bank CapitalBank Capital
Bank Capital
 
Rwa density what lies behind this underrated financial ratio
Rwa density what lies behind this underrated financial ratioRwa density what lies behind this underrated financial ratio
Rwa density what lies behind this underrated financial ratio
 
Stochastic modelling of the loss given default (LGD) for non-defaulted assets
Stochastic modelling of the loss given default (LGD) for non-defaulted assetsStochastic modelling of the loss given default (LGD) for non-defaulted assets
Stochastic modelling of the loss given default (LGD) for non-defaulted assets
 
CCAR and stress-testing segmentation insights
CCAR and stress-testing segmentation insightsCCAR and stress-testing segmentation insights
CCAR and stress-testing segmentation insights
 
Looking back in time for benchmarking - developing countries
Looking back in time for benchmarking - developing countriesLooking back in time for benchmarking - developing countries
Looking back in time for benchmarking - developing countries
 
Implementation of a Credit Risk Management Platform for a Large Insurer Based...
Implementation of a Credit Risk Management Platform for a Large Insurer Based...Implementation of a Credit Risk Management Platform for a Large Insurer Based...
Implementation of a Credit Risk Management Platform for a Large Insurer Based...
 

Último

Economics Presentation-2.pdf xxjshshsjsjsjwjw
Economics Presentation-2.pdf xxjshshsjsjsjwjwEconomics Presentation-2.pdf xxjshshsjsjsjwjw
Economics Presentation-2.pdf xxjshshsjsjsjwjw
mordockmatt25
 
QATAR Pills for Abortion -+971*55*85*39*980-in Dubai. Abu Dhabi.
QATAR Pills for Abortion -+971*55*85*39*980-in Dubai. Abu Dhabi.QATAR Pills for Abortion -+971*55*85*39*980-in Dubai. Abu Dhabi.
QATAR Pills for Abortion -+971*55*85*39*980-in Dubai. Abu Dhabi.
hyt3577
 
+97470301568>>buy weed in qatar,buy thc oil in qatar doha>>buy cannabis oil i...
+97470301568>>buy weed in qatar,buy thc oil in qatar doha>>buy cannabis oil i...+97470301568>>buy weed in qatar,buy thc oil in qatar doha>>buy cannabis oil i...
+97470301568>>buy weed in qatar,buy thc oil in qatar doha>>buy cannabis oil i...
Health
 
+971565801893>>SAFE ORIGINAL ABORTION PILLS FOR SALE IN DUBAI,RAK CITY,ABUDHA...
+971565801893>>SAFE ORIGINAL ABORTION PILLS FOR SALE IN DUBAI,RAK CITY,ABUDHA...+971565801893>>SAFE ORIGINAL ABORTION PILLS FOR SALE IN DUBAI,RAK CITY,ABUDHA...
+971565801893>>SAFE ORIGINAL ABORTION PILLS FOR SALE IN DUBAI,RAK CITY,ABUDHA...
Health
 
abortion pills in Jeddah Saudi Arabia (+919707899604)cytotec pills in Riyadh
abortion pills in Jeddah Saudi Arabia (+919707899604)cytotec pills in Riyadhabortion pills in Jeddah Saudi Arabia (+919707899604)cytotec pills in Riyadh
abortion pills in Jeddah Saudi Arabia (+919707899604)cytotec pills in Riyadh
samsungultra782445
 
Law of Demand.pptxnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnn
Law of Demand.pptxnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnLaw of Demand.pptxnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnn
Law of Demand.pptxnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnn
TintoTom3
 

Último (20)

Economics Presentation-2.pdf xxjshshsjsjsjwjw
Economics Presentation-2.pdf xxjshshsjsjsjwjwEconomics Presentation-2.pdf xxjshshsjsjsjwjw
Economics Presentation-2.pdf xxjshshsjsjsjwjw
 
QATAR Pills for Abortion -+971*55*85*39*980-in Dubai. Abu Dhabi.
QATAR Pills for Abortion -+971*55*85*39*980-in Dubai. Abu Dhabi.QATAR Pills for Abortion -+971*55*85*39*980-in Dubai. Abu Dhabi.
QATAR Pills for Abortion -+971*55*85*39*980-in Dubai. Abu Dhabi.
 
劳伦森大学毕业证
劳伦森大学毕业证劳伦森大学毕业证
劳伦森大学毕业证
 
Test bank for advanced assessment interpreting findings and formulating diffe...
Test bank for advanced assessment interpreting findings and formulating diffe...Test bank for advanced assessment interpreting findings and formulating diffe...
Test bank for advanced assessment interpreting findings and formulating diffe...
 
Stock Market Brief Deck (Under Pressure).pdf
Stock Market Brief Deck (Under Pressure).pdfStock Market Brief Deck (Under Pressure).pdf
Stock Market Brief Deck (Under Pressure).pdf
 
uk-no 1 kala ilam expert specialist in uk and qatar kala ilam expert speciali...
uk-no 1 kala ilam expert specialist in uk and qatar kala ilam expert speciali...uk-no 1 kala ilam expert specialist in uk and qatar kala ilam expert speciali...
uk-no 1 kala ilam expert specialist in uk and qatar kala ilam expert speciali...
 
cost-volume-profit analysis.ppt(managerial accounting).pptx
cost-volume-profit analysis.ppt(managerial accounting).pptxcost-volume-profit analysis.ppt(managerial accounting).pptx
cost-volume-profit analysis.ppt(managerial accounting).pptx
 
Shrambal_Distributors_Newsletter_May-2024.pdf
Shrambal_Distributors_Newsletter_May-2024.pdfShrambal_Distributors_Newsletter_May-2024.pdf
Shrambal_Distributors_Newsletter_May-2024.pdf
 
7 tips trading Deriv Accumulator Options
7 tips trading Deriv Accumulator Options7 tips trading Deriv Accumulator Options
7 tips trading Deriv Accumulator Options
 
Business Principles, Tools, and Techniques in Participating in Various Types...
Business Principles, Tools, and Techniques  in Participating in Various Types...Business Principles, Tools, and Techniques  in Participating in Various Types...
Business Principles, Tools, and Techniques in Participating in Various Types...
 
Famous Kala Jadu, Black magic expert in Faisalabad and Kala ilam specialist i...
Famous Kala Jadu, Black magic expert in Faisalabad and Kala ilam specialist i...Famous Kala Jadu, Black magic expert in Faisalabad and Kala ilam specialist i...
Famous Kala Jadu, Black magic expert in Faisalabad and Kala ilam specialist i...
 
+97470301568>>buy weed in qatar,buy thc oil in qatar doha>>buy cannabis oil i...
+97470301568>>buy weed in qatar,buy thc oil in qatar doha>>buy cannabis oil i...+97470301568>>buy weed in qatar,buy thc oil in qatar doha>>buy cannabis oil i...
+97470301568>>buy weed in qatar,buy thc oil in qatar doha>>buy cannabis oil i...
 
+971565801893>>SAFE ORIGINAL ABORTION PILLS FOR SALE IN DUBAI,RAK CITY,ABUDHA...
+971565801893>>SAFE ORIGINAL ABORTION PILLS FOR SALE IN DUBAI,RAK CITY,ABUDHA...+971565801893>>SAFE ORIGINAL ABORTION PILLS FOR SALE IN DUBAI,RAK CITY,ABUDHA...
+971565801893>>SAFE ORIGINAL ABORTION PILLS FOR SALE IN DUBAI,RAK CITY,ABUDHA...
 
abortion pills in Jeddah Saudi Arabia (+919707899604)cytotec pills in Riyadh
abortion pills in Jeddah Saudi Arabia (+919707899604)cytotec pills in Riyadhabortion pills in Jeddah Saudi Arabia (+919707899604)cytotec pills in Riyadh
abortion pills in Jeddah Saudi Arabia (+919707899604)cytotec pills in Riyadh
 
Law of Demand.pptxnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnn
Law of Demand.pptxnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnLaw of Demand.pptxnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnn
Law of Demand.pptxnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnn
 
W.D. Gann Theory Complete Information.pdf
W.D. Gann Theory Complete Information.pdfW.D. Gann Theory Complete Information.pdf
W.D. Gann Theory Complete Information.pdf
 
Technology industry / Finnish economic outlook
Technology industry / Finnish economic outlookTechnology industry / Finnish economic outlook
Technology industry / Finnish economic outlook
 
Toronto dominion bank investor presentation.pdf
Toronto dominion bank investor presentation.pdfToronto dominion bank investor presentation.pdf
Toronto dominion bank investor presentation.pdf
 
Call Girls Howrah ( 8250092165 ) Cheap rates call girls | Get low budget
Call Girls Howrah ( 8250092165 ) Cheap rates call girls | Get low budgetCall Girls Howrah ( 8250092165 ) Cheap rates call girls | Get low budget
Call Girls Howrah ( 8250092165 ) Cheap rates call girls | Get low budget
 
Black magic specialist in Canada (Kala ilam specialist in UK) Bangali Amil ba...
Black magic specialist in Canada (Kala ilam specialist in UK) Bangali Amil ba...Black magic specialist in Canada (Kala ilam specialist in UK) Bangali Amil ba...
Black magic specialist in Canada (Kala ilam specialist in UK) Bangali Amil ba...
 

Riassunto risk

  • 1. Giulio Laudani #23 Cod 20163 RISK MANAGEMENT Interest risk _________________________________________________________________________________ 1 Methods to Compute Interest Risk: ______________________________________________________________________ 2 A proper Scheme to manage Internal Transfer Rate operation: _______________________________________________ 3 Credit risk __________________________________________________________________________________ 4 The Expected Loss Estimates: ___________________________________________________________________________ 4 The measure of the Unexpected Loss: ____________________________________________________________________ 7 Tools used to estimate UL: ________________________________________________________________________________________ 8 Comments on the UL tools employed: _______________________________________________________________________________ 9 Application and General comments on Credit Risk methods: _________________________________________________ 9 Market risk ________________________________________________________________________________ 10 Tools used to estimate market risk _____________________________________________________________________ 10 Different applications of the Market risk model: __________________________________________________________ 12 Measure of Volatility ________________________________________________________________________ 13 The general limits of the VAR approach are: _____________________________________________________ 13 Basel Committee Framework: _________________________________________________________________ 15 Basel I: ____________________________________________________________________________________________ 15 Basel committee II ___________________________________________________________________________________ 15 Insurance business __________________________________________________________________________ 17 Solvency II _________________________________________________________________________________________ 17 The formation process and directive structure: _______________________________________________________________________ 18 Some example: ________________________________________________________________________________________________ 19 Difference between Banks and Insurer __________________________________________________________________ 20 A new risk paradigm _________________________________________________________________________________ 21 Problems related to the crisis: ____________________________________________________________________________________ 21 All these problems have caused: __________________________________________________________________________________ 22 How the new RM activity looks like after the crisis: ____________________________________________________________________ 22 Interest risk The source of this risk is due to the transformation of maturities and the mismatch between asset and liabilities. Those situations produce an imbalance that will lead to:Refinancing risk whenever the maturity on asset is longer than liabilities one; Reinvesting risk when we are in the reverse situation; Changing in the demand/offer side of the demand of liabilities and call loans, hence the risk linked to the elasticity of demand curve (it won’t be study). Those imbalances are shown in all items affected by any change in interest rate, hence we should consider not only the trading book, but the broader one, that is the banking book, including any derivate whose value depends on market interest rate. The general rule on how treat this risk are explained within theBasel Committee framework, whose general principles have been proposed to help/offer guideline to national authority, which are in charge of the supervision of the bank institution, on how to best estimate and to best manage this risk. Here there is a list of the main area described in the Basel Committee framework: Significant involvement of senior manager to overcome the traditional independency of risk management unit with the other operational division, this active rule of the senior management will grant uniformity on the criteria, objectiveness and proper procedure. This general provision is concretely declinated: The board must be informed regularly and must approve the policy enhanced It must ensure that the risk management has the competence to deal with 1
  • 2. Giulio Laudani #23 Cod 20163 All the monitoring activity must be put in place and there must operational limits The bank must communicate to the supervisory and also to the public their own risk Create independent unit which has to help the senior management in the decision process, providing technical point of view and avoiding potential conflict of interest Bring the measurement of this risk at a consolidate level to adequately appraised and managed Integrate this risk measure into a day to day management of the bank, to steer the corporate policy in the way of how business is lead Methods to Compute Interest Risk: There exist several models to capture/measure this risk from the simplest to more advanced one, however each of them can provide a good guidance to Risk manager since they analyze different angle of the same problem: Interest gap is an income based approach, it considers the effect of change in the market interest rate on the NI 1 o There are several definition of Gap starting from the simplest to the most complicated/accurate, all of them depends on the time windows considered, hence it is a time dependent measure  The first is basically the difference between sensitive assets and liabilities, where by sensitive we mean all those asset which are going to be re-price/roll over in the time frame considered 2  The second one is named Maturity-Adjusted GAP ,which is a weighted average with weights the exposure of each instruments within its GAP. This procedure is quite heavy demanding on the computational side, hence it is simplified 3 by using the concept of marginal GAP  The standardized gap has been introduce to overcame the assumption of same effect for each category included in the balance sheet. It consists of computing the beta of each categories 4  To take into account the change on demand for those items which haven’t an automatic indexing mechanism , we will use the average delay to allocate the instrument to the properly marginal Gap windows  A similar approach can be used to compute the price and quantity interaction effect o The intrinsic limitation of this approach is the GAP is considering only the effect on the net income, hence it is ignoring the effect on the market value of the fixed part of the balance sheet, whose value is effected by the way Duration gap is an equity based approach. It is more consistent with the newmarket to market accounting principle. We are going to use the durationwhich is a measure of the residual life of the instrument, and the modified durationwhich is the first derivative to the respect of interest of the cash flow equation, hence it can be used as a first approximation of the change in value of financial instruments o The duration GAP is estimated bycomputing the MD of Assets and liabilities, where the second one is multiply with the ratio of market value between liabilities and assets (financial leverage) o Limits of this approach are:The immunization will last a very short time, because the duration will change over time, the change of interest rate will change the DM, it is a linear approximation, it is based on the assumption of uniform changes in interest rate (basis risk), assuming parallel shift across maturities, and it is a costly strategy o Some of those limitations can be overcame by some wise modification of the general approach: the basis risk can be solved by using a measure of sensitivity; the cost can be manage using derivatives instruments and the linear approximation can be reduce by using the convexity parameter (second derivatives, distribution of cash flow around the duration) Cash-flow mapping (clumping) is the model that enable to consider asymmetrical shifts on the yield curve, furthermore it is the model suggested by the Basel committee. The method consists of dividing all the instruments in each of its cash flow,and of converting all the instruments into zero coupon, on whom we will use the appropriate zero coupon rate [maturity] 1 The usage of the GAP brings to different macro definitions: Marginal GAP is the GAP computed for a specific window inside the horizon consider. It is related to the cumulative one by two relationship: o The sum of all the Marginal GAP is the cumulative one at time T o The difference of two cumulative gap with different time horizon gives us the marginal GAP between the two horizon Cumulative GAP is computed over the all-time horizon 2 It aims to overcame the limit of the previously one, namely the over simplistic assumption of instantaneous change of interest rate for all the sensitive items 3 The marginal GAP consists of aggregating items in time category on whom there will be computed the significant value and the maturity is simple the average of the beginning and ending date 4 The underling hp is that those items respond to the change with some delay 2
  • 3. Giulio Laudani #23 Cod 20163 o In set up this model there are several choices that should be undertaken: 5  We need to estimate the vertices on the zero coupon curve, taking into accounttypical bank features: Change in short term interest rate are greater than longer rate, volatility decrease as maturities increase and bank’s cash flow are more concentrate in the short term  We need to choose discrete interval since the “item by item” approach is too demanding, we are going to aggregate 6 them using their residual life or using MD to take into account the different sensitivity to change in rate o This method is suggested by Basel Committee to calculate a synthetic risk measure by applying this scheme:  All the assets and liabilities are divided in 13 nodes applying the modified residual life  For each node there is an associated risk coefficient, given by the product of DM and change in interest rate. It is multiplied by the Net position for each node  It doesn’t allow bank to completely off set gaps, to account for asymmetrical movements  This measure is compared with the capital requirement o The Basel methods has several limits:  Use of Book value instead of market one as reference  Presence of instruments which repay capital before maturity create a bias in the modified method residual life approach [duration drift]  Presence of instruments without a fixed renegotiation date or linked to market interest rate  This method is not adequate to consider derivatives  There is no compensation allowed for different currency  Basel committee doesn’t provide a unique solution to those problems, actually it leaves the national supervisory authority free to choose their models o The clumping method provides a procedure to broken real cash flow in two virtual one in effort to better match the nodes vertexes, so that to improve the cash flow mapping results, but require an in-depth knowledge of all cash flow, that is why it can be applied only to small portion of the bank’s portfolio  The virtual cash flow must guarantee constant portfolio (same MV) value and same risk to ensure same change against interest rate movements o  is the first node o  where the rate and time is the node one  A variation of the previously method is basing the clumping on the price volatility to take into account the risk correlation. The two virtual cash flow must sum up to the real volatility A proper Scheme to manage Internal Transfer Rate operation: The ITR is a series of virtual transaction between branches and the risk center unit. It aims to centralize all the decision regarding the risk taking process, to better evaluate the profitability of each branch, to relive the branches from the need to care about the founding process and to transfer all the risk to a central unit. There are two possible scheme to be used: one is the Single ITR, which is the more easier to implement, but it is criticized due to the unique (arbitrary, not related to market condition) rate which is used in the transactions, it handles only net flow, hence the branches are still affected by part of the risk; the other one is the Multiple ITR, whichovercomes this limitation and allows to use multi interest rate for different maturities and gross flows, hence each operation is taken into consideration Here we are going to provide some examples of peculiar transaction and how ITR dealt with: Fixed rate transaction, the risk unit take all the risk, the branch is ensured by blocking is financing rate Floating rate transaction, the same of above plus a premium that repay the higher risk Not indexed at market rates poses two problems: there is no risk hedging instrument and there is a basis risk in fact if we use a derivative on similar underlying there could be an increase in the spread between the similar asset and the real one. The bank can decide who will bear the basis risk since neither the treasury or the branch are able to handle it To avoid arbitrage against Treasury, it could be possible to add a premium over the internal rates to give an incentive to choose the most favorable rate 5 The nodes used are usually consistent with the maturity available for hedging instruments 6 Note that the degree of risk doesn’t depend exclusively on residual life, hence we are simplifying the model in this case 3
  • 4. Giulio Laudani #23 Cod 20163 Derivatives and special payoff, the option: the case of cap or floor or early repayment; price must be paid by the client with the spread charged or paid up-front. It can be beardby the branch or the treasury. The first possibility is simple using options When we are building up an ITR system, we need to address all the desirable features, aiming to ensure a proper risk management system within our institution. Hence, we need to ensure that any Change of profitability is due to only of credit risk, and the total bank’s profitability shouldn’t change; 7 we must protect our selffrom the interest rate risk and embedded options; we should use a multiple rate ITR and we should use gross flow 8 methods ; we should use this system together with a cash mapping process Credit risk Credit risk represents the risk of the default of the counterpart (risk of insolvency) or deterioration in its creditworthiness (migration risk), hence it will arise whenever the discount rate of future cash flow doesn’t reflect the risk of the transaction. Besides to those main sources of risk there are other components: spread risk [whenever the market spread required for the same risk increase due to an increase in the risk aversion] and recovery risk, country risk and pre-settlement risk [when a forward contract is cancelled due to the insolvency of the counterpart and the bank is forced to replace it at unfavorable condition]. The positions accounted for determining the credit risk are the balance sheet items and the off balance one (OTC transaction) The real source of risk is its unexpectedcomponent in fact if the predictable one will be incorporate into the interest rate spread/premium and so totally eliminated. This risk has two components Expected loss (EL) and the Unexpected loss (UL), our task is to estimate the first to properly price the instrument and to measure the second to raise capital to cover this position (Basel Requirement). Estimate this risk and its impact is not an easy task, in fact bank’s credit exposure are recorded at historical value and there is no secondary market to easily determine the market value of those 9 position , hence we need to use internal asset pricing formula . Here there is a set of rules to be implemented to properly set up adequate governance standards: Establishing appropriate credit risk environment, meaning that the board must approve and review credit risk strategies, pointing out the bank risk profile and profitability required. Senior managers must have the responsibility to adequately implement those strategies for all products Use sound credit granting process: establishing credit limits, new client policy or amendment as well as allowing re-new credit line with particular care Maintaining an appropriate credit control process Endure an adequate credit risk control, ensuring independency, IT instruments, recovery bad loan credit facilities and properly prudential level in assessing risk Supervisory should check that those requirement are met and posing limits on risk exposure The Expected Loss Estimates: The EL is function of three different components [referring only to insolvency risk] that must be estimated: The PD estimation: 10 PD is the probability of default, it can be defined by a more subjective or objective definition and it can be assessed by backwarding [human base decision or automated algorithms] or byforwarding [expectation on future development, market data] looking methods. It is estimated by applying several models:  Credit scoring models are statistical approach based on economic variables and financial indicators. They are used to forecast default and risk level of the borrow one by one or by discrete grades (this procedure is better since it reduces the error) Linear discriminant analysis consistson classifying the data using different variables, to define a discriminatory value and to draw a boundary line to clearly separate, as much as possible, the data between healthy and bad company.One of the possible variant is the Altman’s Z-score which is a multivariate regression that already suggests some key variables. 7 Close to those present in the market [even on the ask and bid side], facilitating the risk hedging 8 Each transaction is considered one by one, no netting procedure ì, otherwise the branches will still retina some risk 9 Note that to estimate default risk we can simply use book value, we need market one to estimate the effect of migration risk 10 Depending on the decision the PD will be greater or lower, it depends on the criterion used 4
  • 5. Giulio Laudani #23 Cod 20163 o The score can be seen as a “weighted average” of the scores of a set of different variables, where unimportant variables are weighted close to zero, while important ones get high weights and counterproductive variables get negative coefficients. o Weights are chosen to discriminate as far as possible good firms’ scores from the bad firms’ ones, so mathematically we are 11 aiming to maximize the distance between the two centroid with taking into account their variance [which is assumed to be the same over all the category, this hp can be ease] o To test the significance of this model we can use a parameter calledWilks’ Lambda which basically will compered the distance between the two centroids. It is the ratio between the variance of the healthy and abnormal scores over the total variation (similar to R^2) o From the score we can assess the PD by a formula based on the hp of normal distribution of independent variables. It depends on the score, on the cut-off level (which can be simple or more sophisticated one) and on the prior probability of default [ . o The cut off between company is computed using the past observations and can be simple the average between the distance of the centroids (simple rules) or to ensure a given PD for the accepted company we can use a given formula; the problem is that we can refuse good company that are in the grey/overlapping area.  There is a possible correction term to take into consideration the average quality of the sample or to take into consideration the cost of I and II type of errors. It aims to minimize the first one by adding a new term which considers as an information the ratio between the I and II types of error o These approaches suffer of several limits: the assumption on normal distribution, there are variant with Heteroskedasticity, but they require lots of data and the so called “sample bias” o To select the meaningful variables there are two possible approach: backward elimination or forward selection, keeping in mind the ratio of the model Regression models require defining a sample size, independent variables and how to estimate coefficient. There exist several variants: o Simple which pose the problem of not limited dependent possible value o Logit whish substitute the previously linear relationship with an exponential one o Probituse a normal cumulative density function (smaller tail compared to logit) Heuristic inductive models 12 o Neural, it tries to minimize human learning approach (black-box features ) o Genetic based on the survivorship completion (only the best are able to transfer to future generation their genetic) General limits of this family of models are: o They don’t consider migration and qualitative issues o The meaningfulness of the variable used is questionable, since they must be industrial specific (it is not the case) and needs big sample to avoid bias caused by the presence of too healthy companies in the sample Capital market models are similar to a VAR approach Use of spot or forward rate to estimate the default probability within one year and beyond. The key parameters are the spread, the PD, the recovery and the risk free rate. The PD parameter is computed using the corporate yield curve to maturity and the risk free one, we will use e^ form o In case of maturity over one year we can use long term spot or forward spread, where d is the spread and k is the recovery rate  spot case  forward case o The spot will directly gave us the cumulative probability while the forward the marginal one, it is possible to obtain the cumulative with the forward by the formula o While the marginal one is obtained with the spot by using the following formula: o Limits of this approach are:  It is applicable only for listed company, that have listed bonds for all the relevant maturities, and it is also suffers of the market illusion 11 It is the average value of historical company data for each sample category 12 The procedure is obscure, and it is compared to a black box from where we gain result without knowing how it is working inside 5
  • 6. Giulio Laudani #23 Cod 20163 13  Itassumes a risk neutral approach , hence we should consider/add a premium  It relays on the expectation theory, however there exist liquidity premium Merton model is an example of structural model, meaning it is a model that recognize default as a consequence of some intrinsic company’s features, differently form the reduced model which simply recognize it as a possible event and reduce the problem to estimate its likelihood o Characteristics are: it is based on the ideathat the shareholder position is similar to an option to default (when the value of the assets are lower than liability one)  It assumes that the value of the company follows a Brownian motion, with an increasing uncertainty  At maturity it computes the frequency of the final result of the path simulated and check how many of them are below the threshold  This probability is function of the asset volatility, nominal value of debt, starting company value and the debt maturity o The contingent claim analysis, since those risks can be hedge using a put option the investment can be seen as a risk free investment, where the put premium will be a proxy of the default probability. Thanks to this relationship we obtain the risk free default probability, the value of the debt and the interest rate required by the bank o Analyzing the spreads and default probability computed by applying the Merton model we notice how the riskier company spread curve is negative sloped, while the safer one is reverse in behavior, this is a direct consequence of the survivorship bias 14 for the riskier company o Limitations are:  It assumed a unique zero coupon debt repayment  It assume Brownian motion to describe the evolution of the price  It doesn’t consider the case of early default, before maturity, meaning it doesn’t recognize the possibility that some evolution path may overcome the threshold before time T  It assumes constant risk free, that can be easily solved  It doesn’t consider migration risk  Many of the variables need to be estimated since are not observable  It is an arbitrage free approach, hence there should be the possibility to trade those asset, but it is not the case The KMV model is an attempt to overcome some of the Merton’s limits, namely the first problemand the estimation one o Characteristic: it assesses the value of equity as equal to a call with same maturity of residual life of debt. Thanks to this idea it can estimate the value of the company and the volatility, it use the price of the European call and the Ito’s formula transformation to find the equation to estimate the volatility (2 unknown 2 equation) o The KMV approach takes two steps  Risk index is computed using the innovative concept of distance of default. Thanks to this innovation they can acknowledge the existence of short term and long term debt, hence the company may default only if the assets value drop below the short term debt  This index is then converted into a probability using empirical law o Benefits are the speediness to adapt to change in the financial condition, it is stable compered to economic cycle and it allows to assign specific EDF o Limits are that it can be used only for traded company (it can be solved by using comps) and it relays on the efficiency of the market EAD Estimation: EAD is the exposure at default, it is generally assumed to be a deterministic number, but it may be stochastic if the borrower has the right to change 15 his size exposure. In this case we need to assess the drawn and un-drawn part . A Synonymous of EAD is recovery rate, it computed by applying different formula referring to gross, actual and loading duration o  Gross method o  Actual method, using for each flow its maturity to the occurrence of the default 13 This model is assuming a risk neutral approach, meaning it starts with the equality between a risk free rate and the risky asset weighed by its survivorship probability 14 Over time the pool of the weaker company will get stronger and stronger due the repayment of debt and the death of the worst among them 15 The un-drawn part needs to be estimate considering the portion that will be drawn at default, to ease this task bank requires fees on this part that allows a more rational pricing (common in Anglo-Saxon countries) 6
  • 7. Giulio Laudani #23 Cod 20163 o  Loading duration, where “d” is the difference between RA duration and EAD duration  /TOT_PV(EAD) where “t” is the time elapsed from default  Where RA is the recovered amount, AD the administrative and legal cost, r is the appropriate discounter factor and d is the duration of the process taking into account intermediate flow LGD Estimation: LGD is the expected loss rate in case of default, as PD it depends on the definition of default chosen by the bank, in fact the narrower the definition the lower the LGD. Its value depends on those drivers:Characteristic of the borrower: industry sector, country  speediness of recovery procedure, financial ratio; Characteristic of the exposure: presence of collateral or guarantee, priority level; Characteristic of the bank facilities to recover loan and out of court settlements, and External factors, such as economic cycle [common with PD], interest rate level There are different procedures to estimate it, however all of them have to take into consideration all the indirect cost beard by the bank to recovery the proceed and the time needed: o Market LGD is similar to the PD estimate using market rate, in fact knowing the other parameters we can infer the related LGD o Work out LGD consists in building an historical database from where we could extract the LGD for a given type of borrower. This method is the only applicable to bank’s loan which doesn’t have secondary market on which we can extract the market value after credit events  We need to find about an appropriate discount factor of the recovery amount, which must be forward looking since the procedure will be start in the future (after the beginning of the credit line)  We need to estimate the duration of the process itself, the book gives an example page 350. Note that this model is the preferred one It is common in the risk management to find out some key variables which allow understanding the different recovery rate empirically experimented. This topic is important to understand the recovery risk, given the high volatility presented into database: It has been noted that the distribution is concentrated in the tails; The industry sector is a key element to explain differences; The priority level is important, but is not stable over time; and The presence of collateral is important Recovery risk is the risk to achieve a different LGD from the expected one, it is usually quite sizable and fluctuate over time due to a binomial distribution concentrated in the tails o This risk arise from the uncertainty of those variables:  Amount to be recovered (EAD)  Administrative cost to be beard  Discount rate for future cash flow at future date  The duration of all the process o It is calculated using those two formulas  Assuming a non-stochastic LGD   Assuming a stochastic LGD The link between LGD and PD must be taken into account, since they share common systematic factors. This relationship is negative, as empirical evidence based on junk bond had proven. Doesn’t consider this risk may lead to an underestimation of risk. It is affected by:Chain effects: economic downturn may reduce the value of assets; Financial asset and interest rates together with real estate one, and Industrial specific: the inventory may lose more value for certain industry The measure of the Unexpected Loss: The UL can be simple define by the volatility around its mean (EL) or by some sort of CREDIT model based on VAR approach. This kind of exposure can be effectively reduced by setting up a diversification policy. We need to choose appropriate time horizon and confidence level: Time horizon is arbitrary chosen equal to 1 y, because both subjective and objective criterions are not applicable due to the presence of illiquid market, of contract without explicit maturity and so on. However this decision has a great operational fit:The bank use yearly budget, hence having yearly measure ease the budgeting work out; Usually 1 y it the correct time elapse to raise new capital; The turnover of the asset is usually close to one year andFurthermore some model will account for long term risk including migration risk in their computation. Confidence level may consider that the data cannot be explained by a zero mean normal distribution, 7
  • 8. Giulio Laudani #23 Cod 20163 since the mean is below zero and data are strongly asymmetrical.Hence the choice of Confidence level need to be credible, because it is less stable and change a lot at different level Tools used to estimate UL: It is estimated using portfolio models that apply different definition of default (including or not migration) and level of loan correlation (explicitly or implicitly modeled) as well as EL Creditmetrics is a multinomial approach it considers all borrowers’ change in rating as a credit events o The model basically consists on deriving the empirical distribution for any rating class movements, and byusing the respective spread for each maturity to compute the expected value of the assets, and on doing the difference between this value and the face value to compute the expected loss  We should compute the VAR percentile using the probability for each possible movement in the rating level to have a measure of UL, since the distribution is not normal ( we cannot use the normal standard deviation time percentile approach) o To overcome the Wilson’s critique regarding the accountability of the economic cycle it has been introduced the credit portfolio view which is taking into consideration the economic variables  Note that if the time horizon used is a point in time there is no need to adjust by economic cycle, in this case there would be a double counting o To use the model for portfolio of assets we need to estimate the correlation, by using those steps:  A modified version of the Merton model, where we find out in the normal distribution, all the thresholds for default and change in rating. Using math integral we can compute all the probability of change marginal and cumulative for each assets We pool together all these info with a bivariate normal (given the correlation) and the new distribution will give us all the joint probability Asset correlation, which is the one used in the precedent formula is computed using large building blocks o Correlations are first estimated among a large set of industries and countries (“risk factors”), For each borrowers, a set of weights must be specified, expressing their sensitivity to different risk factors and to idiosyncratic risk o Combining those weights and the risk factor correlations, an estimate of the pairwise correlation of two firms can be obtained It is clever, because Asset value changes – unlike the distribution of losses – can be described by using normal distribution, it can then be described through a double normal distribution with correlation parameter ρ, hence from this distribution, the joint probabilities of the values of the two loans can be inferred, taking into account the transfer of risk o Benefits are:  Uses objective and forward looking market dataInterest rate curves and stock indices correlations  Evaluates the portfolio market value  Takes into account migration risk o Limits are:  Needs a lot of data: forward rates, transition matrices  Assumes the bank is price-taker  Assumes stable transition matrices  Proxies correlations with stock indices  Maps counterparties to industries and countries in an arbitrary and discretionary way Credit risk + is based on an insurance approach using Poisson distribution to estimate probability of default o The idea behind this model is the equivalency of the bank and insurance risk, the major difference is the correlation among bank’s clients. It aims to assess the risk portfolio, it doesn’t give us PD, its input are: PD and its volatility o We use a banding methodto link default probability with losses. It consistson creating category of similar expected losses and computes the number of default for each band, each of those will be used in the Poisson distribution, It is advisable to use, tosum up, the default number weighted with the ratio of j-esimo amount and band expected loss  The creation of category is made by dividing all exposures “Li” by L and round them up, getting standardized values “vi”. The recovery rate is use to determine the exposure (used in the banking) in a deterministic way 8
  • 9. Giulio Laudani #23 Cod 20163  Each of those Poisson distribution, one for each bands, must be aggregate. However this first model assumes independency between variables, but this can be overcame by assuming that the “number of default” for each band is random (see below) o The Poisson distribution assumed independent value and it works only for small PD. To estimate correlation we will run “n” simulation within all the bands [assuming that the PD are stochastic] and then we made the weighted average conditional to all the scenario occurrence and this new distribution will be the unconditional distribution, accounting for the asset correlation o Benefits are:  PDs and exposures (book value) net of recovery are enough. The “correlated” version requires also sensitivities to the economic cycle factors  An analytical solution exists, hence it is fast to be implemented  Possibility to obtain the distribution of losses without recurring to simulation techniques o Limits are:  Only looks at default risk, so it Does not consider migration risk  Assumes constant exposures and doesnotconsiderrecoveryrisk  It isn’t a dynamic model. Meaning it cannot be used for changing portfolio without recomputed all the calculation Comments on the UL tools employed: Some comments on the major characteristic of those models: Default-mode versus multinomial one, where only credit risk + belongs to the first Future values vs. loss rate, meaning that the model can be based on the distribution of possible value or possible future loss, the first use as input the spread curve by maturity, while in the second the spread it is not necessary to be known. Creditmetrics is a typical market value model, while credit risk + loss one Conditional vs. un-Conditional, portfolio views belong to the first, however this distinction is useful only if the model works through the cycle Monte Carlo vs analytical solution Asset correlation vs. default correlation, it’s less important than other, in fact they are close to each other. Creditmetrics is belongs to asset correlation, while credit + to the second one. Some major limits: Treatment of the recovery risk is not random (besides for credit +) and independent from PD Assumption of independence between exposure risk(usually treated as known) and default, while they are empirical positive related hence this assumption brings to an underestimation Assumption of independence between credit risk and market risk Impossibility to back testing due to yearly frequency, there is not enough data sample Application and General comments on Credit Risk methods: Loan pricing by assessing EL and UL using a transparent process to properly consider the cost of capital absorbed. There is the problem to attribute the marginal benefits o  ELR is the expected loss rate, it is the spread to compensate for the EL o  VaR is a relative measure, is the cost of equity Risk adjusted performance measurement is used to decide to undergo a specific investment o  where is the lending rate applied, this number will be compared with the bank’s RAROC Setting limits on risk taking of the different business units, however it is crucial to properly define the appropriate level of aggregation for units and the VAR limits and frequency with which are checked as well as its involvement in the budgeting process o  this formula will give the max loan amount that can be granted Optimizing the portfolio composition, this is limited due to bank’s loan characteristics (geo, no secondary market and limited rotation). However thank to recent derivatives development and secondary market it is now possible, the best solution should be to divide the Risk management optimization form the origination process 9
  • 10. Giulio Laudani #23 Cod 20163 Market risk Market risk is usually identified as the risk inherent in the trading book (short term), but it should be extended even for those investments intended to be kept in the financial statement for longer period. Nowadays, it has gained more importance due to the new accounting principle, the securitization process and the growth of derivatives instruments. The key elements of this type of risk are:Exchange risk; Interest risk (different from the previously one because it is related only to securities which have a secondary market and it will affect only a limited part of the balance sheet); Equity risk; Volatility risk andCommodity risk The traditional approacheswere:Nominal value which considers the risk as proportional to the nominal value. It has severe limitations such as: the nominal value doesn’t reflect MV, it doesn’t capture the different degree of sensitivity to a change in the risk factor and it doesn’t consider volatility and correlation; Sensitivity analysis is based on the usage of coefficient representing the risk for each securities category. However there still drawback such as: you cannot aggregate those coefficients, it is an easy to communicate between division and senior since each position has a unique measurement and it doesn’t consider the volatility and correlation (as nominal value method), meaning that without taking into consideration the volatility of risk factor is basically not considering the real risk of the position Tools used to estimate market risk VAR models are characterized by: confidence level, maximum potential loss that a portfolio can suffer and a certain time horizon. This measure is comparable between all securities class. These models aim to define the risk factors, the probability distribution of those risk and to summarize those information in one risk parameter The easiest one is the parametricone: normal distribution of change in value in market factors, all possible information are summarized by Var- Cov matrix, the possible loss are correlated to the risk factor by a linear function and VAR is simply a multiple of standard deviation. o The most crucial hp are the one that ignore those problems:  Empirical data shows that the distribution of risk factor is skewed and with fatter tails  It works by simply reverting the normal distribution into density function from where we can compute the equivalent percentile. [MV*σ*α*β]  The linear relationship is represented by the sensitivity coefficient which (for example) is the modified duration for bonds; this is a deltanormal approach. Alternatively we can assumed that the prices are log normal distributed, hence there is a one to one correspondence, this is an asset-normal approach o Typical issues are:  The confidence levelchoice represents the degree of risk aversion of the financial institution, or the target capital requirement that allows to have on balance sheet the investors’ expected creditworthiness (investment on it) , in fact there is a positive empirical relationship between those value  Time horizon is usually a short term measure (daily), the choice of the appropriate level is crucial since the higher the time length the higher will be the VaR value The bank must take into account the liquidity level of its position, the size and a subjective idea on the time that the instrument will be in the trading book It is important to have enough data observation to ensure significance on the value. Hence for longer time horizon there could be problems to achieve this. Sometimes to overcome the problem it is possible to use the LRW hp to convert daily data into monthly or weekly o When we apply this method to portfolio we need to compute the correlation between asset, which is simply the matrix of Corr pre and post multiplied by the ordered securities VAR. This formula will grant diversification benefits, but will arise sub-additivity property o When we want to use multi risk factors we need to break down each security into elementary components which depend solely on one risk factor, and then aggregate them as a portfolio. The approach suggested is the mapping of risk position:  In the case of a foreign currency bond we have two risk components, we will proceed by computing each single VAR and then sum them up using their correlation  Forward currency position: three positions, exchange rate spot and a spot investment/borrowing amount  Forward rate agreements consists of two components a debt spot position with a maturity prior to the investment spot one  Stockpositions we can consider each stock as a risk factor or use fictitious position toward relevant stock indexes, hence the position is mapped to the relevant betas. It works only for well diversified portfolio without specific components  Bonds are instead mapped for each of their cash flow (clumping) o Limitations of this model are [note that this model is better if considering single risk profile]:  Risk factors change are assumed to be normal distribute 10
  • 11. Giulio Laudani #23 Cod 20163  Since the assumption of risk factor is to be explained by Brownian motion the volatility is assumed to be stable over time and serial independence across risk factor, it is not empirically true  The relationship with loss is assumed to be linear o Solutions provided are:  The possible solution to normal distribution hp is to use a mixture of Gaussian distribution (to solve the fait tale problem not the empirical asymmetry) or t-student  The non-linearity can be solved by using a curve parameter, but this solution poses the problem of not normal distribution of the underling changes since the curving coefficient is a chi-Square function. Furthermore by introducing a new parameter we are increasing the model error, and, above all, there is the underling hp that the payoff are derivable, otherwise it is un- applicable. This approach is less effective for joint shock. Another solution is using full valuation approach Simulation modelallows to use different hp compared with the previously one:the risk factors may have different distributions, since the impact of each risk factor change is value by full valuation it is possible to have a different relationship between loss and risk factor and it is a way to ensure flexibility. Hence, This model are best used in case of nonlinear payoffs and to value extreme event [stress test] o Full valuation is a way of estimating price variation is not a simulation model, it consists on re-price the securities applying a price formula (that must be define a priori) o All distribution are empirical, meaning based on the simulation. Here there is the difference between historical and Monte Carlo where the fist use empirical distribution obtained by the observation while the second define a parametric distribution o Use of the Percentile logic, meaning that we will generate scenario by applying a given distribution and on those scenario ordered we will compute the interesting percentile, hence the VaR will be the difference between the asset current value and its percentile value o There is great flexibility on defining the market risk change behavior Historical simulation transforms the historical data into future possible behavior, hence it uses the past distribution to predict future one without change o Merits of this model are:  It is simple to be understand and communicable  It doesn’t require hp about distribution or correlation (not Var-Cov required)  Non linearity required between risk factor and price change o Limits and possible solution are:  It assumes a stable and stationarity distribution given by past data  It may not applicable if the time series are limited or if there is the problem of the usual tradeoff between time length and meaningfulness, however using ahybrid approach putting together the exponential weights and the simulation approach we can use longer sample ensuring meaningfulness. Furthermore this hybrid approach allows the model not to have stable distribution, note that is an approximation, since the current data have higher weights any past change in behavior is smoothed o The methodology of bootstrapping and path generation are used to obtain bigger sample size  Bootstrapping is made to avoid loss of observation when we move from daily to weekly or monthly frequencies. It consist of an extraction (allowing for reinterring) of observations, in this way we will built X paths (useful in case of exotic option pricing when we need to know the path is important in defining the price) from whom we will extract the distribution. The underling hp is that observation must be i.i.d  To try to overcame the hp of i.i.dwe can used something similar to the hybrid approach or there are two better proposals: Hull and White suggest adjusting data by weighting with current volatility, meaning by proportional move returns value. Heteroskedasticity can be incorporate by rescaling the time series with the available information at time T Filtered historical simulation is based on G-ARCH models which are used to filter data and to obtain residuals which are used to create scenario. Each residual is standardized (hoping that doing so they are i.i.d) and used in a path generation process where the first data is the filtered return, not the sample one, and it is multiply with he estimated conditional value of next period (predicted volatility by G-ARCH times previously residual) Monte Carlo simulation is based on the definition of an a priori parametric distribution, which should be consistent with the future data behavior. This model is more computing efficient since by increasing the parameters involvedthere is a proportional increase of variables number, but still demanding o Differently from historical simulation, it requires to compute or to assess the correlation between risk factor a priori (otherwise is like assuming independency), but it allows to know the path evolution o To take into account the relationship we need a Corr matrix which is decomposed into two triangular matrices and used them to build up scenarios 11
  • 12. Giulio Laudani #23 Cod 20163 o Limits are the needs of an asset based estimation of Var-Cov to find out the joint distribution Stress testaims to estimate the effects connected with extreme events, meaning that the simulation is carry out following predominantly arbitrary and subjective manner just for few scenario o Use past shock and simulate them o Factor push analysis meaning use more standard deviation movements, note that there is the problem of great variation of insignificant risk factor are not significant o Derivatives policy group suggestion it is uni-dimensional (meaning each variable at time) process suggested on interest rate movement in the slope or shift as well as on other macro indicators o Multidimensional scenario can implement into two ways:  Simple where only some risk factor are let change while the others are kept constant  Predictive is the same of the previously one, but the others are changed according with their correlation to the moving one o This instrument should be a complement to the previously ones and must be followed up by practical actions to reduce risk, hence vulnerability. It allows to test liquidity risk Different applications of the Market risk model: Unique riskmeasure for both horizontal and vertical communication between division Portfolio analysis is possible: there are several method to aggregate risk factors It is useful to determine the risk limit exposure both as nominal value, market value exposure (remembering that those measures will be affect by change in volatility) and maximum tolerable variation. It can be used to as a mean to risk adjust return by using RAROC [both ex-ante and ex-post], which is a profit/risk ratio, thanks to this instrument it is possible to monitor the units return, properly set incentive procedure 12
  • 13. Giulio Laudani #23 Cod 20163 Measure of Volatility Volatility estimation can be divided into two groups: historical with constant parameters and those with changing one and implied option. It is important in the Var-Cov model, since it is a crucial parameter. For each of the subsequent model we need to define the time horizon/frequency The simplest is basically the computation of the Var-Cov matrix using an equally weighed average or an exponential one “Riskmetrix”using as data source historical dataset o Simple moving average model poses two problems:The sample size: the longer the higher the information, but it will be less reliable for prediction purpose, since it doesn’t reflect current situation;and Echo effect meaning that any shock will strongly affect the value both in the entrance and in the exit from the sample o Exponential moving average will overcame the previously problems, but it poses the problem of choosing the appropriate decay factor (which should depend on the data behavior, i.e. how fast they react to change, but it must be noted that it changes overtime) and the number of past observations must be evaluated, this problem is solved increasing the frequency so that to maximize the info content and minimize the error sampling G-ARCH model stands for: Heteroskedasticity, Conditional and Autoregressive. It uses the maximum likelihood criterion, hence it use market data to get a better estimate of the decay factor, and however it needs lots of datapoint.  The two common factors are: Past volatility  it indicates the rate of persistency. It is above 0,7 Past prediction error square  it indicates the rapidity with which volatility adapt to new market shock. It lower than the first one  Benefits of this methodology are: It recognizes the serial correlation It gives adequately importance to new information The decay factor is directly determinate by market data o It use a normal distribution to describe the behavior of the prediction error, hence it is a poor method in case of skewness or leptokurtosis. The later problem is overcame by using t-student distribution while the first one (which come out of the fact that this model gives same importance to the sign of shocks, which is against empirical date)is solved by models that recognize this asymmetry, i.e. the negative effect has an higher impact:  IG-ARCH require that the sum of the coefficient is one, hence if the constant is set to zero is basically the exponential moving average formula  EG-ARCH models the natural log of the variance (allowing the equation to give negative output), so that instead of square value it use absolute value and real value (response to good and bad is consider)  AG-ARCH use the square, but it centered the data using a parameter which will amplify the negative effect an reduce the positive one o It gives good measure for the immediately following period, but it is less informative for subsequent, in fact, since the parameters should guarantee mean reversing or better a converge series, it converge to the long term value Implied volatility is basically the volatility taken out form derivatives instruments, hence it represent the market expectation on that period, not really used in risk management o It may be affected by counterparty risk or by the liquidity level o It will change depending on the contract chosen (at/in/out the money), the maturity and the model used to price the option (it must be reliable and the instrument must be liquid to ensure efficiency) o The maturity and the time horizon for risk management system should coincide to ensure consistency o Computing covariance value is complex and we may not have data, moreover we need to check if the numbers are consistent between them, the matrix must be at least semi-positive. There are two way: the first by using derivatives with more than one underling such as quanto option, or using page 182 The general limits of the VAR approach are: It doesn’t define the behavior in case of extreme events higher than the thresholds (however it recognizes them); however the purpose of the VAR isn’t to make the bank completely safe from bankruptcy, but to reduce this risk to an acceptable level. Remember that higher capital requirement will lower the net income o However VAR fails to gives us a measure (probability) that allows to discriminate within excess lost between portfolio, it isn’t possible to understand which is higher 13
  • 14. Giulio Laudani #23 Cod 20163 The VAR won’t take into consideration the customer’s relationship or any other qualitative indicator, however we can post that the bank is not force to use VAR as the only method to allow credit line, it may be the case that the management use other qualitative methods Some of the assumption behind VAR models are questionable, however if they are understood they can be overcame, furthermore VAR it is a good tool to allocate resource between units according to risk hence it gives more insight compared with other tools DifferentVAR approach can lead to different value, hence it can be seen as a weakness, however it must be noted that since the results highly depends the model there could be the cause that the underling hp of some of those approaches are not appropriate. Furthermore if there will be any bias it is uniform for all the securities, hence the relative risk allocation or perception is unchanged VAR is a too cyclical measure (it is different for each method), however the traders’ behavior follow this trendduring financial crisis VAR comes too late, meaning it is not able to predict crashes, however VAR purpose isn’t to predict crashes (which is impossible for any historical predictive method), but to generate consistent and uniform risk measure during normal conditions for day by day business regulation VAR has a sub additivity property, diversification allows to reduce aggregate VAR, however it is not always the case (such as for non-parametric models), hence it may strongly underestimate risk. o This last problem can be solved by a different measure proceed: Expected shortfall:This method allows to have an idea of the possibility to have excess lost compared with the VAR lost, it is the average of the excess lost beyond the VAR thresholds. It ensure sub additivity and it gives us a measure of how much money are needed to bail out the bank for the supervisory authority or the expected payment that an insurer have to pay against bearing this risk Using the Backtesting methodology we will be able to check the performance among different VaR models and to check the consistency with confidence level. The method is basically counting the exceptions to test independence among them and to test the confidence of interval: there are two test proposed: o The Kupiec Test: the null hp is that the model is good performing  We will compute the LR test between two binomial distribution, the numerator without conditioning (hence using the theoretical percentile implemented), and the denominator conditional to the empirical observation. The test is distributed according to a chi-square with one degree of freedom  conditional coverage  Limits of the test are: it doesn’t account for serial dependence (meaning it doesn’t allow to value a model for its capability to avoid time concentration of excess losses, hence it doesn’t value the quality of the model to react to change in the market conditions); it requires lots of data, the power of the test is weak (and it decrease with the confidence level and number of data) o The Kristofferson Test: the test is willing to check that all the probability of exception and non-exception occurrence is not correlated to the previous occurrence  The LR test: it is the ratio between the Likelihood function assuming independency and the perfect correlation, it is possible to joint test this LR test with the Kupiec one by summing up them, this new test will be distributed according to a chi-square with two degree of freedom o The paper of Saita and Sironi provides an empirical analysis using 7 equally weighed portfolio (equity, global) to test a Garch(1,1) model a Historical simulation test and an EW Moving average (lambda =0.94)at three confidence level  All the three models have failed to grant a consistent confidence level at 0,99 and 0,97; while at 0,95 the EWMA has been more conservative and the other two still underestimate  The Kupiec test has well performed (only at 0,95) for some cases three for EWMA, one HS an two Garch  The extreme event have shown a higher vale than the theoretical one  The Kristofferson test has failed for basically all the cases, it has performed better at 0,99  Internal model based on those models have been more capital demanding the HS is the most conservative due to the abrupt change  The analysis wasn’t able to define a winner, it was limited only for equity with daily horizon 14
  • 15. Giulio Laudani #23 Cod 20163 Basel Committee Framework: This agreement is the result of the meeting of the G10 central bank governors to set up a common banking standard to ensure solvency of the industry and avoid unfair completion based on different capital requirement by national supervisory authority, hence it aims to uniform the bank capital requirement (it applies to consolidated account, hence it would foster the soundness of institution controlled by foreign banking group) and to prevent future bank crisis (it would revert the bank trend to reduce capital). It is an ongoing work trying to achieve a best practice to ensure a global stability to all the financial institution. Basel I: Capital requirement computation is focus on the presence of three different tiers, which need to sum up to 8% i.e. the minimum requirement bound. This measure is net of goodwill and investment in non-consolidated bank or financial institution o Tier 1is the most important and valued part of the capital. It consists of an upper tier: shares paid-up, disclosed reserves and certain general provision (in UE the use of those items has been limited only for specific risk, and they have to respect: after tax, immediately available and size, distribution and provisions must be shown separately) and the lower one: innovative instrument which respect: un-redeemable, permanent, callable by issuer’s will (after 5 y), must be junior to all other instrument, absorb loss without liquidation procedure and the remuneration can be deferred or if impossible forfeited o Tier 2 is the first category of supplementary capital, it can be used at most for 50% of total capital. It consists on:  Undisclosed reserves meaning all those reserves which are created using revenue from off-balance sheet, however they must be solid as the disclosed one  Revaluation reserve (sizable in German and Japan), they are the difference between historical and current market value (they must come from a revaluation process) and can be used at 45% of their value (due to possible negative variation)  General provision, however they have been limited by the new accounting principle and must account for at most 1,25% of the total capital  Hybrid capital instruments don’t need to put the bank under liquidation to absorb loss, remuneration can be waived or reduce and cannot be redeeming by credit (only with supervisory authorization). It must unsecured (as all this instruments) and fully paid up  Subordinated term debt, the main difference with the previously one is that to be liable of any loss there must be a liquidation procedure. The condition that those instruments have to meet are: maturity longer than 5 y, they must be “depreciated” at 20% each year and must be junior o Tier 3 it cannot be accounted for more than 250% of tier 1 for market risk and it cannot exceed 50% of total capital requirement. 16 It includes short term subordinated debt and can be used to cover only market risk. Risk weights in the first Basel agreement are quite simple and “linear”. They are: o 0% for cash, Government bond in the OCSE, claims on central bank o 20% claim on central bank or country outside OCSE, claims on bank less than 1 y and multilateral development banks o 50% loan secured by mortgage on residential o 100% the others o The OTC items pag. 555 The major Limitsare: o Focus on credit risk only, there is no consideration for currency risk or others (1996 solved) o Poor differentiation of risk, there is no consideration for this issue on the weights, furthermore the class defined by the weight allows for regulatory arbitrage, they aggregate too much o Limited recognition of the link between maturity and credit risk as well risk mitigation instrument Basel committee II Pillar 1 capital requirement has receipt all the Basel 1 criticism and it had tried to improve the overall capital system. It must be noted that the capital requirements are not fix, but can vary following the supervisory expectation both on the capital and the risk perceived by internal rating method (by multiplying them by a scalar factor) o Risk weightsfor the normal one are related to the rating agency with some improvement, acknowledging the 16 It must be at least 2 y, cannot be redeem, there is a lock-in if the capital ratio follow below the minimum plus 20%, and it must be reduce by loan loss forecast and security for trading one 15
  • 16. Giulio Laudani #23 Cod 20163 o Internal rate is the possibility given to certain bank to use internal procedure to identify and estimate the main component of risk, there exist two different versions:  Fundamental allows only to make estimation on PD, while the other EAD, LGD and maturity are chosen by the national authority  Advancedgave the rights to use own estimations for all the risk components.It can be used if those condition are met: 7 rating classes, where the first must have a PD of 0,03% and there must be a default category Pillar 2 has been made to enforce the supervisory capability to control the bank system. This approach ensure a proactive prudential behavior based on those principles o Bank must set up a system of process and techniques aimed at establishing the overall capital adequacy o Supervisory authority must check those process and ensure the respect of the minimum capital requirement and promptly intervene to avoid capital deterioration Pillar 3 has been done to reduce the opacity and to force the market to discipline unfair behavior and to promptly penalize bank which are taking more risk. The bank is so forced to publish info regarding: their economic and financial results, financial structure, and their risk management strategies, exposure to risks, accounting policy and corporate governance. 16
  • 17. Giulio Laudani #23 Cod 20163 Insurance business Solvency II It will take 8 years to totally implement the new framework started in 2005. There have been 5 insurance conferences (QSI) which has seen a wide participation. The main features are: This framework aims to enhance a more consistent standard across EU and ensure that capital requirements are more reflective of risks undertaken by insurers o Market consistency approach for valuing all assets and explicit pro-cyclical provisions  Asset and liabilities [non insurance one+ should be valued at price “arm’s length transaction” with reference to IFRS and no recognition of change in credit standing for financial liabilities  Asset and liabilities [insurance one] should be value with model (best estimate) Price needs to be the same irrespective of investment strategy Use of risk free rate adding a risk margin The pricing model consists on computing technical provisions based on o Best estimate: PV of future cash flow, which are computed considering prudent and realistic assumption, management and consumer possible future actions o There is any more prudential provision (forbidden), all the guarantee, discretionary benefits and option must be valued o Reinsurance revocable must be valued separately and with allowance of credit risk o The choice of which risk free rate poses some problems [EIOPA provides some guideline] such as:  Different currencies problem is solved using swap rate adjusted for credit risk and providing by QIS5 spot rate curve for main currency  Different maturities need appropriate rate: Illiquidity asset: duration cap at 15 euro and 30 GPB and linearly reduce to 0 in 5 years and rate adjusted The spot curve is calculated using a certain basket of corporate bonds The liquidity premium is cap at the max available on the market for similar cash flow without risk, and it should account for liabilities’ nature (it is provided by EU institution with the same frequency of interest rate)  The risk free rate are extrapolate using macroeconomic model to find out the unconditional long term rate o Better recognition of diversification, risk mitigation and loss absorbing items o New supervisory approach, more proactive and EU coordinated among country and groups’ economic reality  Group complementary supervisory which have primary responsibility (can used internal model) the methods are Accounting consolidation (to eliminate double counting) Deduction and aggregation  recognition of diversification  Third countries relationship is articulated in three aspect The reinsurance supervision from equivalent third countries no difference Group capital requirement from equivalent third countries can be used their Group solvency from equivalent third countries, their authority is ok Transitional arrangements are put in place for important countries, that are not equivalent to EU regulation The reform aims to promote confidence and transparency among insurance It equally apply to reinsurers It wants to overcame the Solvency I limits: o Lack of harmonization o Inconsistency with new IFRS accounting principle o Capital requirement is not transparent and not adequate to risks, furthermore it was focus on back-looking aspect instead of governance issue (good Risk management) o No recognition of economic reality of group (only plus requirement any reduction) 17
  • 18. Giulio Laudani #23 Cod 20163 The formation process and directive structure: The formation process and authority involvement is basically EU lead by applying the Lamfalussy structure: L1 see the involvement of Parliament, Ecofin e commission helped by special committee and L2 and L3 see the “joint committee” *ESMA, EIOPC, EBA+ check law. The directive structure reflects Basel II/III’s pillar approach Pillar I risk calibration of financial requirement, hence it is focus on the quantitative issue o The SCR is calculated in terms of potential loss of value at a confidence level of 99,5% over a 1 year time horizon considering all quantifiable risks  It should (internal model) account for even the liquidity premium risk  Each risk is model individually with a modal approach using factor or scenario approach  The individually risks are aggregate using corr matrix (provided by QIS5)  It must be taken into account the absorbing capability of technical provisions (meaning future benefits: change in hp) and deferred tax Gross SCR are computed without considering benefits Net SC assumes bonuses to reduce/absorb loss The difference of the first two items will be FDB (future discretionary bonuses)  Mitigation techniques and collateral and segregated assets are allowed under certain conditions (legally binding, actual transfer of risk…)  Proportional recognition for techniques in force for less than 12 months are allowed under certain conditions( no risk for rollover and counterparty risk take into account) o The MCR (minimum) is cap within the corridor of 25/45% of SCR  Legal certainty, auditable and safety net protection  It is based on percentages applied to combination of premium and technical provision o Solvency II allow to choose which level of simplicity or sensitivity can be used to assess risk form the simplified method to internal method  The insurance must grants a sound framework to managing, controlling and measuring risk  The authority can still request specific parameters for internal models o The Own funds categorization:  Basic Own funds: excess asset over liabilities, subordinated liabilities and adjustment (expected profits, net deferred tax and restricted reserves) They can be used for all tier capital requirement  Ancillary Own fund (prior supervisory approval): off balance items that can be called up to absorb losses such as unpaid shares, letter of credit other legally binding documents It can be used to tier 2 and 3 o The capital eligible to meet SCR must be at least 50% tier 1, hybrid instrument can be at most 20% of tier 1 and tier 3 cannot account for more than 15% o The capital eligible to meet MCR must be 80% tier 1 no tier 3 available and no ancillary funds o The criteria used to allocate OF in tiers are: subordination, loss absorbency, duration, freedom to redeem, absence of encumbrances and of mandatory servicing costs Pillar II new supervisory relationship and setting up for governance o Authorities are granted to take action to restore critical situation when SCR get closer to MCR, to ensure even in extreme situation to preserve policyholders interest  There is a convergence of supervisory standards  Capital “add on-“ are authority demand to adjust risk assessing  They can draft and implement measures, so that they can promote fairness act against breach of EU law and act in the emergency situations o To take into account cyclical effects the authority can extend the recovery period, inherent rebalancing between SCR and available capital, liquidity premium and equity symmetric dampener o There are quantitative risk management standards, so that it will play a central role in the company  Internal control system and internal audit and actuarial functions  ORSA (own risk solvency assessment) specific risk profile, compliance with financial requirements, significance of deviation between the two o There are new disclosure requirements bring market discipline to bear insurers 18
  • 19. Giulio Laudani #23 Cod 20163 Pillar III opening up to market discipline provisions and reporting to authority o The disclosure to market must be at least annually and account for business performance, risk profile, system of governance, capital management and valuation purposes o The Regular supervisor report to must be narrative with quantitative data, every 3 years but it can be asked to be provided annually o Annual and quarterly quantitative template with a lag of 14 weeks and 5 weeks for the last one The main implementing problems and open issues are: Liquidity premium, currency curve and how to allocate them to liabilities, There are too many uncertainty and different interpretations, The QSI5 have recognize 92% of OF as tier 1, The group diversification has a low to free 20% of OF on average which has created a huge difference with other non EU entities, Eliminating excessive complexity, reducing excessive volatility and reducing penalization for long term business with guarantees (revising stresses for interest and spread risk at long duration); Improving calculation criteria  contract boundaries and deferred tax and liabilities; Finding political compromises on political aspects  illiquidity premium, future profits, diversification in risk margin and at group level; Designing appropriate transitional provision disclosure and reporting, internal model; Properly manage the new task of supervisory authority (new tools, culture and power); and The impact of the new discipline will be huge, there is a change on the overall business with extensive impact on governance, strategy, expectation of authority and new competitive scenario Some example: Example of the usage of the aggregation of modules for SCR calculation purpose based on the market risk case, which is divided into several categories o Market investment must have a minimum rating of BBB as an overall rule o The equity risk module  Speculative are treated depending on the market on where there are listed and on CEIOPS Advice Duration based approach (22%) for certain lines of business Symmetric adjustments are made by EIOPA on a benchmark of listed company and are sufficiently public. The adjustment is cap between -/+ 10% and it is computed using a formula that take in consideration the current level and the weighted average of daily data over a window of 36 month  Strategic participations are stressed at 22% o The interest rate module  All asset sensible to interest should be stressed, as well as insurance liabilities in their PV change  There is no stress test on volatility of interest rates o Spread risk module consists in any change in the credit spread over the risk free rate  It is calculated on separate items (Bonds, Credit derivatives and ABS on loan) Derivatives which are used for risk mitigation are excluded Bonds and ABS are value using a table using rating, a conversion factor (larger for the last)  Instruments of EU with AAA or AA rating are excluded, and the other have lower capital requirement as well as corporate AAA rating o Currency risk module  It can be calculated on net or gross position  It consider a max variation of 25%  No diversification benefits, but the EU currencies are advantaged with lower stress level o Property risk module, it the risk related to real estate  25% for all property (direct) investment in company in real estate is subject to equity module  No geographical diversification o Concentration risk module, it aims to consider the risk related to lack of sufficient diversification or to exposure to large default risk (single issuer)  It apply to all the category of assets, considering only the exposures in excess of certain threshold  It depends on the exposure and rating of the counterpart  It is excluded participation internal the group, when the risk is beard by policyholders or when it is already considered in the default counterparty module  It doesn’t consider geographical concentration  EU with AAA and AA no capital requirement o Liquidity risk module is new and it allows for change in illiquidity premium  Negative correlation with spread risk (-50%) 19