Skip to main content
← Back to C Definitions

Critical values

What Is Critical Values?

Critical values are threshold points in statistical hypothesis testing that determine whether to reject a null hypothesis. They define the boundaries of the rejection region (or critical region) within a probability distribution of a test statistic. In the realm of inferential statistics, these values are crucial for making informed decision making about population parameters based on sample data. If a calculated test statistic falls beyond the critical value(s), it indicates that the observed result is sufficiently extreme to warrant rejecting the null hypothesis at a chosen significance level.

History and Origin

The concept of critical values is intrinsically linked to the development of hypothesis testing in the early 20th century, primarily through the foundational work of statisticians like Ronald Fisher, Jerzy Neyman, and Egon Pearson. While early forms of statistical reasoning for making inferences from data existed in the 1700s, it was Fisher who formalized the idea of a null hypothesis and the use of p-values in the 1920s and 1930s. Fisher's approach focused on determining the strength of evidence against a null hypothesis. A famous anecdote often cited in the history of hypothesis testing is the "Lady Tasting Tea" experiment, where Ronald Fisher devised a method to test whether a woman could genuinely distinguish between cups of tea prepared by adding milk before or after the tea. This experiment, described in his 1935 book The Design of Experiments, laid the groundwork for formalizing the process of rejecting a null hypothesis based on observed data.7

Subsequently, Jerzy Neyman and Egon Pearson introduced a more formal framework for hypothesis testing, emphasizing the need for two competing hypotheses (a null and an alternative hypothesis) and the explicit consideration of Type I error and Type II error rates. Their work, developed in the 1930s, provided the basis for setting fixed significance levels (alpha) and using critical values to delineate rejection regions, a practice that became central to modern statistical inference.

Key Takeaways

Formula and Calculation

While there isn't a single "formula" for critical values in the traditional sense, they are determined by three key factors:

  1. The chosen Significance level ($\alpha$): This is the probability of making a Type I error, typically set at 0.05 (5%), 0.01 (1%), or 0.10 (10%).
  2. The degrees of freedom (df): This value, often related to the sample size, dictates the specific shape of the sampling distribution. For example, in a t-test with one sample, the degrees of freedom are ( n - 1 ), where ( n ) is the sample size.
  3. The type of statistical test: Different tests (e.g., z-test, t-test, chi-square test, F-test) correspond to different probability distributions (Normal distribution, T-distribution, Chi-square distribution, F-distribution), each with its own set of critical values.

Critical values are typically found by consulting statistical tables (like a t-distribution table or a standard normal Z-score table) or using statistical software. For instance, for a two-tailed t-test with a significance level of (\alpha = 0.05) and (df = 20), one would look up the value in the t-table corresponding to these parameters. The table would provide the positive critical value, and due to symmetry, the negative of this value would also be a critical value, defining the two tails of the rejection region.6,5

Interpreting the Critical Values

The interpretation of critical values revolves around comparing the calculated test statistic from sample data against these thresholds.

  • For a two-tailed test: If the absolute value of the test statistic is greater than the positive critical value, or less than the negative critical value, the result falls into the rejection region. This means the observed data are considered sufficiently unlikely under the null hypothesis to reject it. For example, in a test where critical values are (\pm 1.96), a test statistic of (2.05) would lead to the rejection of the null hypothesis.
  • For a one-tailed test:
    • Right-tailed: If the test statistic is greater than the positive critical value, the null hypothesis is rejected.
    • Left-tailed: If the test statistic is less than the negative critical value, the null hypothesis is rejected.

If the test statistic falls within the bounds of the critical values (i.e., in the "acceptance region" or "fail-to-reject region"), there is not enough evidence to reject the null hypothesis at the specified significance level. It is important to note that failing to reject the null hypothesis does not mean it is true, but rather that the data do not provide sufficient evidence to conclude it is false.

Hypothetical Example

Consider an investment firm that wants to assess if a new trading algorithm generates an average daily return significantly different from zero. The historical average daily return for similar algorithms has been essentially zero.

  1. Formulate Hypotheses:
  2. Choose Significance Level: The firm sets a significance level ((\alpha)) of 0.05 for a two-tailed test.
  3. Collect Data and Calculate Test Statistic: The algorithm is run for 30 trading days ((n = 30)), yielding a sample mean return and a standard deviation. A t-test is appropriate for this scenario with unknown population standard deviation. Let's assume the calculated test statistic (t-value) is (2.10).
  4. Determine Critical Values: With degrees of freedom (df = n - 1 = 30 - 1 = 29), and (\alpha = 0.05) for a two-tailed t-distribution test, consulting a t-distribution table reveals critical values of approximately (\pm 2.045).
  5. Make a Decision: Since the calculated test statistic ((2.10)) is greater than the positive critical value ((2.045)), it falls into the rejection region. The firm would reject the null hypothesis. This suggests that there is statistically significant evidence to conclude that the new algorithm's average daily return is indeed different from zero.

Practical Applications

Critical values are fundamental in various financial and economic contexts, enabling data-driven decision making.

  • Investment Analysis: Portfolio managers use critical values in hypothesis testing to evaluate if an investment strategy's returns are significantly different from a benchmark, or if a stock's price movements are statistically independent of market trends. For instance, testing if a fund's alpha is statistically greater than zero.
  • Risk Management: In assessing the effectiveness of a risk model, critical values help determine if observed losses exceed expected thresholds by a statistically significant margin, indicating potential model failure or a need for recalibration.
  • Economic Research: Economists employ critical values to test hypotheses about macroeconomic indicators, such as whether a policy change has had a statistically significant impact on inflation or unemployment rates.
  • Market Efficiency Studies: Researchers use critical values to test hypotheses related to market efficiency, for example, whether past stock prices can predict future prices beyond what would be expected by chance.
  • Regulatory Compliance: Financial regulators may use statistical tests with defined critical values to monitor for unusual trading patterns that could indicate market manipulation or other illicit activities. The NIST Engineering Statistics Handbook: Critical Values of the Student's-t Distribution provides comprehensive guidance on the application of statistical methods, including the use of critical values, in various engineering and scientific fields, which are transferable to quantitative finance.4

Limitations and Criticisms

Despite their widespread use, the application and interpretation of critical values and the broader framework of hypothesis testing face several limitations and criticisms.

One significant concern is the binary nature of the reject/fail-to-reject decision. Relying solely on whether a test statistic crosses a critical value can oversimplify complex findings and potentially mask the actual size or practical importance of an effect. A statistically significant result (one that crosses the critical value) does not automatically imply practical significance. Conversely, a result that fails to cross the critical value might still indicate a real, albeit small, effect that simply lacked the statistical power to be detected at the chosen significance level, potentially leading to a Type II error.

Moreover, the arbitrary nature of common significance levels (e.g., (\alpha = 0.05)) has been a point of contention. This threshold, while conventional, can lead to rigid interpretations, where a p-value of 0.049 is deemed "significant" while a p-value of 0.051 is not, despite minimal practical difference. The emphasis on these thresholds can encourage practices like "p-hacking" or "cherry-picking" data, where researchers might manipulate analyses until a statistically significant result is achieved. The American Statistical Association Statement on Statistical Significance and P-Values (which are closely related to critical values) in a seminal 2016 statement, emphasizing that statistical significance should not be the sole basis for scientific or financial conclusions.3

Another criticism highlights that critical values and p-values do not provide the probability that the null hypothesis is true or the probability that findings are reproducible. They only indicate the compatibility of the data with a specified statistical model under the null hypothesis. Over-reliance on critical values without considering other factors like effect sizes, confidence intervals, and the broader context of the research or financial problem can lead to misleading conclusions.

Critical Values vs. P-value

Critical values and p-values are both integral to hypothesis testing, serving as distinct but complementary tools for making statistical inferences. The fundamental difference lies in their approach to the decision-making process.

  • Critical Value Approach: This method involves pre-determining one or more critical values based on the chosen significance level and the probability distribution of the test statistic. The calculated test statistic is then compared directly to these critical values. If the test statistic falls into the rejection region defined by the critical values, the null hypothesis is rejected. This approach is akin to setting a predefined boundary and checking if the observed evidence crosses it.

  • P-value Approach: The p-value (or probability value) is the probability of observing a test statistic as extreme as, or more extreme than, the one calculated from the sample data, assuming the null hypothesis is true. Instead of comparing the test statistic to a critical value, the p-value is compared directly to the chosen significance level ((\alpha)). If the p-value is less than or equal to (\alpha), the null hypothesis is rejected. This approach provides a measure of the strength of evidence against the null hypothesis, offering more granularity than a simple binary decision based on critical values alone.

While they represent different ways of arriving at a conclusion, both methods will lead to the same statistical decision given the same data, hypotheses, and significance level. Confusion often arises because both are used to evaluate statistical significance, but they provide different perspectives on the data's relationship to the null hypothesis. The p-value quantifies the probability of the observed data under the null hypothesis, while critical values delineate the specific cut-off points for rejection.

FAQs

What does a critical value tell you?

A critical value tells you the threshold that a test statistic must exceed (or fall below, depending on the test) in order for the result to be considered statistically significant at a given significance level. It helps you decide whether there is enough evidence to reject the null hypothesis.

How do you find a critical value?

Critical values are found using statistical tables (like a T-distribution table or Normal distribution table) or statistical software. You need to know your chosen significance level ((\alpha)), whether your test is one-tailed or two-tailed, and the degrees of freedom associated with your test.2

Are critical values always positive?

Not always. In a two-tailed test, there are typically two critical values: a positive one and a negative one, defining rejection regions in both tails of the distribution. For example, if the critical values are (\pm 1.96), then both (1.96) and (-1.96) are critical values. In one-tailed tests, you will have only one critical value, which can be positive or negative depending on whether it's a right-tailed or left-tailed test.

What happens if your test statistic is beyond the critical value?

If your test statistic falls beyond the critical value (e.g., greater than a positive critical value in a right-tailed test, or less than a negative critical value in a left-tailed test, or outside the range of both critical values in a two-tailed test), it means the result is statistically significant. This leads to the rejection of the null hypothesis, suggesting that the observed data are unlikely to have occurred if the null hypothesis were true.

Can critical values be used for confidence intervals?

Yes, critical values are also used in the construction of confidence intervals. For a given confidence level (e.g., 95%), the critical value helps determine the margin of error, which in turn defines the upper and lower bounds of the interval. A 95% confidence interval corresponds to a 0.05 significance level in a two-tailed hypothesis testing.1

AI Financial Advisor

Get personalized investment advice

  • AI-powered portfolio analysis
  • Smart rebalancing recommendations
  • Risk assessment & management
  • Tax-efficient strategies

Used by 30,000+ investors