• Shuffle
    Toggle On
    Toggle Off
  • Alphabetize
    Toggle On
    Toggle Off
  • Front First
    Toggle On
    Toggle Off
  • Both Sides
    Toggle On
    Toggle Off
  • Read
    Toggle On
    Toggle Off
Reading...
Front

Card Range To Study

through

image

Play button

image

Play button

image

Progress

1/14

Click to flip

Use LEFT and RIGHT arrow keys to navigate between flashcards;

Use UP and DOWN arrow keys to flip the card;

H to show hint;

A reads text to speech;

14 Cards in this Set

  • Front
  • Back
Why NCCI moved from 17 limits to 5 for ELFs
1. ELFs at any pair of excess limits are highly correlated across classes.
2. Limits below $100,000 were heavily represented in the prior 17 limits.
3. THey wanted to cover the range of limits commonly used for retrospective rating.
General excess ratio formula
R(L) = ExpectedLossesexcessofL / TotalExpectedLosses
= 1- LossEliminationRatio(L)
Normalized excess ratio function for injury type i
S(r) = E[max(X_i/Mu_i - r,0)] = integral(t-r)*g_i(t)*dt
Injury type weighted excess ratios
Rj(L) = sum( w_i,j * S_i (L/Mu_i,j) )

Rc(L) = sum( w_i,c *S_8 (L/Mu_i,c)
Credibility weighted class excess ratio vectors
R_c final = z*R_c + (1-z)*R_j
z= min(n/(n+k)*1.5,1)
n = # of claims in class
k = average # of claims per class
Other credibility options considered
1. Using the median instead of the average for k.
2. Excluding Medical Only claims from the analysis.
3. Including only serious claims in the analysis.
4. Requiring a minimum # of claims for classes used in the calculation of k.
5. Various square root rules, such as z=sqrt(n/384, which corresponds to a 95% chance of n being within 10% of its expected value.
Measuring the distance between vectors
NCCI used the usual L-squared or Euclidean distance, which they would have calculated for every pair of classes C of the R_c final vectors, measured as:
||x-y||2 = sqrt(sum(xi-yi)^2)
The NCCI also considered using the L-1 distance of ||x-y||1 = sum |xi - yi|. This would have the advantage in that it minimizes the relative error in estimating excess premium. THe relative error in estimating premium for class c with limit L is PLR * |R_j(L) - R_c(L)|. Since the analysis was not sensitive to the distance measure, they chose to use the traditional squared distance measure.
Why the NCCI didn't use standardization
1. The resultant hazard groups using standardization didn't differ much from not using it.
2. Excess ratios at different limits have a similar unit of measure, which is dollars of exces loss per dollar of total loss. Standardization would have eliminated this common denominator.
3. All excess ratios are between 0 and 1, while standardization could have led to results outside this range.
4. There is a greater range of excess ratios at lower limits, and this is a good thing since it is based on actual data (compared to excess ratios at higher limits being based more on fitted loss distributions). Standardization would have given this real data less weight.
Weighted k-means algorithm
1. Decide on the number of clusters (potential hazard groups) k to target.
2. Start with an arbitrary initial assignment of classes into k cluster (denoted as HG below).
3. Compute the centroid of each cluster i for the k clusters as: R_i = sum(w_c*R_C) / sum( w_c)
4. For each class, find the closest centroid usign the L^2 distance, and assign the class to that cluster. If any classes have been re-assigned in this step, go back to step 3. Continue until no classes are re-assigned.
Desirable optimality properties for k-means
K-means maximizes the equivalent of R-squared from linear regression, which means maximizing the percentage of total variation explained by the hazard groups. This is equivalent to stating that k-means minimizes the within variance and maximizes the between variance, which means the hazard groups will be homogeneous and well separated.
Test statistic used to decide # of hazard groups
1. The Calinski and Harabasz statistic of Trace(B) / (k-1) / Trace(W) / (n-k), where n is # of classes, k is # of hazard groups. Higher values of this statistic indicate a better # of clusters. The test is also known as the Pseudo-F test since it resembles the F-test of regression analysis.
2. The Cubic Clustering Criterion (CCC) statistic, which compares the amount of variance explained by a given set of clusters to that expected when clusters are formed at random based on the multidimensional uniform distribution. Again, a high value of this statistic indicates better performance.
Why the NCCI didn't rely on the CCC test statistic showing 9 groups.
1. Milligan and Cooper found the Calinski and Harabasz statistic outperformed the CCC statistic.
2. The CCC statistic deserves less weight when correlation is present, which was the case.
3. The selection of # of hazard groups ought to be driven by the large classes where most of the experience was concentrated. Using these highly or fully credible classes showed 7 as the optimal number.
4. There was a crossover in the excess ratios between hazard groups when using 9 groups, which is something that isn't appealing in practice.
Underwriter considerations
1. Similarity between class codes that were in different groups.
2. Degree of exposure to automobile accidents in a given class.
3. Extent heavy machinery is used in a given class.
3 key ideas of hazard group remapping.
1. Computing excess ratios by class
2. Sorting classes based on excess ratios.
3. Cluster analysis