Topic 1. Operational loss data and incidents
Topic 2. Regulatory Requirements for Operational Loss Data
Topic 3. Data Collection Process
Topic 4. Comprehensive Data
Topic 5. Incident Dates
Topic 6. Boundary Events
Topic 7. Data Quality Requirements
Topic 8. Operational Risk Data
Topic 9. Risk and Control Self-Assessments (RCSAs)
Topic 10. RCSAs: Impact scales
Topic 11. RCSAs: Likelihood Assessment scales
Topic 12. RCSAs: Heatmaps
Importance of Historical Data: Risk management and measurement heavily utilize historical losses and events.
Operational Risk Management (ORM) Framework: The ORM framework is structured in four concentric circles:
Incident/loss database
Assessment through RCSAs
Monitoring through KRIs
Takeaways from major loss events or high-risk exposures
Benefits of Incident Information
Internal incident information aids in scenario identification and assessment, and helps quantify regulatory capital reserves.
Information on incidents at other banks provides general awareness and allows for necessary precautions or risk mitigation techniques.
Thorough and reliable incident data helps minimize Pillar 2 extra capital requirements.
Basel Committee on Banking Supervision (BCBS) Criteria: The BCBS outlines numerous criteria for data quality and collection. Key requirements include:
At least 10 years of data
Minimum collection threshold of EUR20,000
Internal losses classified into Basel event-type categories
Reporting of occurrence and recovery dates
Other Criteria: Regulatory criteria also cover governance issues and the exclusion of losses due to credit and market risk.
Documented Procedures: Banks must have "documented procedures and processes for the identification, collection, and treatment of internal loss data".
Clarity and Standardization: Clarity on what is reported is the first step to ensure sufficient data quality.
Items to consider include dates, incident types, thresholds, and loss quantification.
A standardized approach to gathering and reporting information is crucial.
Challenges: Operational incidents are often challenging to describe, and quantification can be subjective.
Free Text Usage: The use of free text should be carefully controlled to provide only supplementary detail, not the primary source of loss information.
Data Fields for Reporting: Common data fields used in reporting include:
Place of occurrence
Date of occurrence
Event type and Cause type
Control failures
Impact type (loss, gain, near miss) ,
Gross loss, Recovery and Net loss
Material Activities and Exposures: Internal loss data must "capture all material activities and exposures from all appropriate subsystems and geographic locations".
Loss Reporting Threshold: While the BCBS does not clarify "material activities and exposures," the minimum loss reporting threshold is EUR20,000. Many banks use a lower threshold (e.g., EUR/USD/GBP 1,000-10,000).
Direct vs. Indirect Losses:
Regulations focus on recognizing financial losses directly stemming from operational incidents (e.g., compensatory payments, regulatory penalties, fund losses due to errors).
Less direct losses, such as lost customers due to reputation decline, are generally not technically considered by regulations.
Including less direct costs is good management practice for reflecting the full cost of operational risk and making better decisions.
Grouped Losses: Regulations require grouped losses with a common cause to be recognized. For example, an IT failure causing multiple related losses should be combined and reported as a grouped loss.
Required Reporting: Loss events must be reported with gross loss amounts, net of recoveries, and reference dates. A summary of sources and causes should also be provided.
Four Relevant Dates
Occurrence: When the event first happened
Discovery: When the event was identified
Reporting: When the event was reported
Accounting: When the event was settled and entered into the bank's general ledger
Time Lags
The time between occurrence and discovery can be several years for certain incidents.
The settlement date can be years after occurrence, especially for incidents with regulatory and legal implications.
Occurrence to discovery interval indicates visibility of issues.
Discovery to reporting interval measures reporting timeliness.
Caution with Old Data: Banks must exercise caution when analyzing recently "paid out" losses that occurred many years ago in a different business environment, as they may not be appropriate for future projections.
Q1. A bank employee has been manipulating suspense (transitory) accounts for a prolonged time, and this event was determined to be a significant operational incident. Which of the following yime intervals in terms of the operational incident date is most likely to be the longest?
A. Date of reporting to date of accounting.
B. Date of discovery to date of reporting.
C. Date of occurrence to date of discovery.
D. Date of occurrence to date of reporting.
Explanation: D is correct.
The dates of an operational incident (in order) are as follows: occurrence, discovery, reporting, and accounting. In the case of a prolonged and undiscovered manipulation of suspense accounts, the period between the date of occurrence and the date of discovery is long. However, because the date of reporting happens after
the date of discovery, then the time interval between the date of occurrence and the date of reporting would be the longest.
Definition: Boundary events are those occurring in a risk class different from their cause. This can include market and/or credit losses caused by human error (operational in nature).
BCBS Classification
Boundary events are considered credit risk and not attributed back to their actual cause. Operational losses in a credit risk context are often included in Risk-Weighted Assets (RWAs) and therefore do not need to be included in operational losses if already in RWAs.
Boundary events are not considered market risk. Therefore, operational losses in a market risk context are included in operational losses.
Independent Review: The BCBS requires banks to "have processes to independently review the comprehensiveness and accuracy of loss data".
Corroboration: Corroboration from multiple sources, including the general ledger (GL), is required to ensure completeness of significant losses.
General Ledger Limitations:
Captures only direct, easier-to-quantify items, often omitting intangible/harder-to-quantify indirect items.
May not record items that indirectly show up as lower revenues, but should be recorded as expenses or losses on the income statement.
Other Data Sources:
IT logs can match IT issues to recorded operational incidents (e.g., P1/P2 IT issues equivalent to operational risk incidents).
Customer complaints, unfavorable press/media reports, and reserves for upcoming lawsuits help mitigate underreporting.
Addressing Underreporting: Underreporting, especially in smaller banks, can be due to uncertainty, unwillingness to pursue control failures, and avoidance of bureaucratic processes. Mitigation methods include gentle reminders, using risk metrics in managerial performance assessments, and internal audit involvement.
Distinct Characteristics: Operational risk data differs vastly from market and credit risk data.
Idiosyncratic and Uncorrelated: It is idiosyncratic and largely uncorrelated to general markets, posing challenges for measurement and modeling.
Fat Tails: The presence of wide (fat) tails in their returns and frequency distributions indicates a greater occurrence of extreme events.
Interrelated Nature: The interrelated nature of operational risk creates interpretive difficulties in linking causes and effects.
Purpose: RCSAs involve assessing the probability and severity of operational risk.Types of Risk Assessment:
Inherent Risk: Assessment without considering current controls.
Residual Risk: Remaining risk after considering control effectiveness.
Frequency: Typically performed annually, but quarterly for very significant operational risks.
Process: Usually involves discussions among participants about inherent risks, controls, and control effectiveness. Questionnaires may be used at larger banks.
Control Types:
Preventive: Reduce the chance of risk materializing.
Corrective: Minimize negative effects if a risk materializes.
Objective: To understand inherent risks, internal controls, and residual risks, and determine if additional mitigation is needed.
Nature: Primarily nonquantitative and requires significant judgment; not meant to quantify risk exposures.
Backtesting: Some banks backtest RCSA results against historical incidents, often revealing underestimates of probability but overestimates of severity.
Alternatives: Risk and Control Assessment (RCA) and Residual Risk Self-Assessment (RRSA).
Disadvantages: Subjectivity, behavioral biases, limited data, and inconsistency between business units. Standardized risk descriptions and definitions are needed for consistency.
Output: Documented and sent to unit managers for approval, outlining significant residual risks and potential consequences of control failures.
Risk Mitigation: Helps identify if risk is within appetite; if not, new controls, enhanced existing controls, exposure reduction (e.g., reduced transaction volumes), or insurance may be implemented.
Q2. Which of the following statements regarding risk assessment is most accurate?
A. Risk and control self-assessments (RCSAs) are a good mixture of qualitative and quantative aspects.
B. The main purpose of risk assessment is to prioritize risk management and risk mitigation responses.
C. When backtesting the results of risk assessment against past incident experience, there is a tendency to overestimate the likelihood of an operational risk event.
D. When backtesting the results of risk assessment against past incident experience, there is a tendency to underestimate the impact of an operational risk event.
Explanation: B is correct.
RCSAs are qualitative and largely judgment based. When backtesting the results of risk assessment against past incident experience, there is a tendency to underestimate the likelihood and overestimate the impact of an operational risk event.
Definition: Impact refers to financial, regulatory, customer, and reputation aspects; in resilience, it can include continuity of service.
Ratings: Typically consist of four or five ratings (e.g., low, medium, high, very high, extreme).
Measurement:
Financial Impacts: Usually based on a relative measure like percentage of revenue or operating income to accommodate different sizes.
Customer Impacts: Usually based on a percentage of the customer base.
Regulatory and Reputation Impacts: Typically nonquantitative, though some metrics may be used.
Measurement: Stated as a probability or frequency of occurrence.
Time Horizon: RCSAs have a maximum time horizon of 1 year.
Probability Example: A 1-in-25-year event signifies a 4% chance of occurring within the year, not necessarily occurring once over a 25-year period.
Historical Data Limitations: History is a starting point but can be a poor indicator of the future, especially in volatile environments or with emerging risks (e.g., technological advancements, cyberattacks).
Rating Categories: A five-point scale might include: remote, unlikely, possible, likely, highly likely.
Definition: The RCSA matrix, or heatmap, combines likelihood and impact.
Color-Coded Intensity: Combinations are assigned colors (e.g., green, yellow, amber, red) corresponding to risk intensity.
Green: Permissible risk exposure, no further steps needed (e.g., possible likelihood, low impact; or remote likelihood, high impact).
Yellow: Permissible but near limit, monitoring and mitigation steps may be needed (e.g., likely likelihood, low impact; or unlikely likelihood, high impact).
Amber: Exceeds permissible limit, risk must be lowered or escalated for approval (e.g., likely likelihood, medium impact; or remote likelihood, extreme impact).
Red: Far exceeds permissible limit, risk must immediately be lowered (e.g., likely likelihood, high impact; or unlikely likelihood, extreme impact).
Qualitative Nature: Ratings are qualitative and should not be compared mathematically. For example, a "remote chance/major impact" risk is not mathematically equivalent to a "possible chance/low impact" risk, even if numerical ratings might appear so.
Topic 1. Key Risk Indicators (KRIs)
Topic 2. Key Performance Indicators (KPIs)
Topic 3. Key Control Indicators (KCIs)
Topic 4. Fault Tree Analysis (FTA)
Topic 5. Factor Analysis of Information Risk (FAIR)
Topic 6. Swiss Cheese Model
Topic 7. Root-Cause Analysis
Definition: KRIs indicate a bank's exposure level at a specific point in time.
Preventive KRIs: These show a rise or fall in the intensity of a cause of a risk.
They indicate a rise or fall of the impact or likelihood of the risk.
Examples of KRIs for Likelihood:
Increase in the number of transactions processed per employee (risk of error).
Increase in sales level needed to earn a bonus (risk of fraud).
Examples of KRIs for Impact:
Increase in the sensitivity of data maintained on a server (greater impact if data leakage were to occur).
Increase in transaction limits for traders (greater impact if unauthorized trading were to occur).
Risk Appetite and Priorities: KRIs account for the bank's risk appetite, and the chosen metrics reflect the bank's priorities in terms of its goals and the most important risks to manage.
Thresholds: Thresholds in KRIs reflect the level of control desired by management for risk mitigation.
Definition: KPIs provide measurements on how effectively the bank operates.
Examples of KPIs:
Number of customer complaints.
Error rates on customer transactions.
Average downtime of IT systems.
Definition: KCIs provide measurements on how effectively the bank's controls are operating.
Examples of KCIs: Number of business continuity plans not reviewed or updated before the set due dates.
Errors remaining after two sets of independent and qualified reviews.
Number of general ledger data entry errors after application of edit checks.
Significant Overlap: KRIs, KPIs, and KCIs often overlap significantly, and a single metric can have elements of all three. This is particularly true when analyzing a deficiency from a required benchmark level.
Control Failures: Key control failures are an automatic source of risk and can embody elements of KRIs, KPIs, and KCIs.
Example Scenario: Scenario: Control failure with incorrect transaction processing despite multiple checks.
KCI: Relates to the inaccurate transaction processing.
KRI: Relates to increased risk of litigation by customers resulting from the errors.
KPI: Relates to weak back-office abilities.
Q3. One of the bank's compliance managers notices that there are two business continuity plans that have not been reviewed, and it is now past their review due date. The incomplete reviews of the business continuity plans are best described as a:
A. key assessment indicator.
B. key control indicator.
C. key performance indicator.
D. key risk indicator.
Explanation: B is correct.
KCIs provide measurements on how effectively the bank’s controls are operating and, in that regard, the fact that the business continuity plans are now past due their review date means that the control of the due date is not operating effectively. KRIs will indicate the bank’s exposure level at a specific point in time. KPIs provide measurements on how effectively the bank operates.
Nature: FTA is a deductive failure analysis method where fault trees break down failure scenarios into external and internal conditions necessary for a major event to occur.
Application in Banks: Increasingly used in banks where a chain of failures leads to a major loss.
Conditions: Conditions can happen simultaneously ("AND conditions") or alternately ("OR conditions").
Joint Probability (Independent Events): Assuming all events are independent, the joint probability of occurrence is the multiplicative product of their separate probabilities.
Example: Four independent controls each with a 5% chance of failing, the probability of all failing simultaneously is 0.054=0.00000625=0.000625%. This represents the minimum probability of the scenario.
Real-World Independence: In reality, failures are unlikely to be completely independent, as controls are often designed together.
Conditional Probabilities: Conditional or Bayesian probabilities are needed for a more realistic sense of occurrence, where probabilities are updated with new information.
Example: Probability of transaction error increases given an IT service interruption.
Exposure Consideration: It's crucial to consider exposure (e.g., number of employees) in addition to low failure rates when assessing likelihood.
Risk Factor Probabilities: Causal analysis requires estimating risk factor probabilities from third-party and empirical information.
Probabilities for weather-related events might come from insurance companies.
Baseline probabilities for scenarios can be derived from peer experiences.
Worst-case scenarios from RCSA sessions might also be considered.
Q4. In the context of a potential nonmalicious data leakage incident by a bank employee, the following information is provided:
Using fault tree analysis (FTA) and assuming that the conditions just listed are fully independent, the likelihood of the risk of data leakage materializing for a given time period is closest to:
A. 0.0001485 %.
B. 0.0004365 %.
C. 0.0048015%.
D. 0.0253935 %.
Explanation: A is correct.
The equation is 0.99 (phishing email received) × 0.05 (firewall failure) × 0.10 (employee failure) × 0.03 (activity detection failure) × 0.01 (exit of information) = 0.000001485 = 0.0001485%.
Concept: In factor models like FAIR, risks are broken down into individual factors.
Three Steps of FAIR Model:
Determine risk factors and how they interrelate.
Measure each factor.
Computationally combine all factors.
Output: The output is a loss distribution for a given scenario.
Scenario Attributes: Scenarios have attributes such as: Asset at risk.
Threat.
Threat type.
Losses occurring if risk occurs.
Estimates and Distributions: Estimates are made for the frequency of losses and likely loss amounts, all expressed as distributions (not single points). These distributions serve as inputs for Monte Carlo simulations.
Monte Carlo Output: The output from Monte Carlo simulations is the distribution of the simulated losses.
Analogy: The model posits that defensive layers are like slices of Swiss cheese with continually shifting holes.
Risk Materialization: A bad outcome typically only occurs when holes in multiple layers momentarily align, allowing a "trajectory of accident opportunity" that brings hazards into damaging contact with victims.
Control Layering: Proper control layering is essential to prevent the alignment of these "holes" and avoid a "perfect storm".
Control Characteristics:
Control failure rates should not be correlated.
Compensating controls should effectively nullify any weaknesses.
Periodic reviews are required to ensure controls remain independent and reliable.
In an operational risk context, root-cause analysis is performed by the first line of defense for significant incidents and near misses.
The second line of defense then adds to and/or queries the analysis
Risk event placed centrally with causes/preventive controls on left side and impacts/detective/corrective controls on right side
Topic 1. Loss Distribution Approach (LDA)
Topic 2. Extreme Value Theory (EVT)
Topic 3. Internal and External Loss Data
Topic 4. Capital Modeling
Topic 5. Operational Resilience
Q5. A quantitative analyst is in the process of combining internal and external loss data for model calibration purposes. The analyst has established specific criteria for the inclusion or exclusion of
losses in the dataset. The method used by the analyst is best described as:
A. cut-off mix.
B. filtering.
C. scaling.
D. rule based.
Explanation: B is correct.
Filtering involves setting specific rules for the inclusion or exclusion of losses in the dataset to be used for modeling to avoid potential manipulation of results.
With cut-off mix, at a specific loss threshold when there are few(er) internal losses, then external losses would be included in the model to have sufficient data points to generate a reliable distribution. With scaling, there are size adjustments for losses to be consistent with the bank's size or inflation adjustments when a large number of years are involved. In this context, there is no such thing as a rule-based method.