Adjusting VAR to correct sample volatility bias

David Frank proposes an adjustment to sample variance for the computation of value-at-risk



In this article, we propose a simple adjustment that can be used when computing value-at-risk (RiskMetrics Group 1996) if a sample standard deviation of returns rather than the true standard deviation is employed. Using the sample standard deviation as if it were the true standard deviation (the usual procedure) results in VAR estimates that are biased downwards. We propose a correction for this bias by adjusting the VAR level itself, where the adjustment depends on

Only users who have a paid subscription or are part of a corporate subscription are able to print or copy content.

To access these options, along with all other subscription benefits, please contact or view our subscription options here:

You are currently unable to copy this content. Please contact to find out more.

Sorry, our subscription options are not loading right now

Please try again later. Get in touch with our customer services team if this issue persists.

New to View our subscription options

If you already have an account, please sign in here.

You need to sign in to use this feature. If you don’t have a account, please register for a trial.

Sign in
You are currently on corporate access.

To use this feature you will need an individual account. If you have one already please sign in.

Sign in.

Alternatively you can request an individual account here