Composite Hypothesis Test

Hypothesis tests > Composite Hypothesis Test

What is a Composite Hypothesis Test?

A composite hypothesis test contains more than one parameter and more than one model. In a simple hypothesis test, the probability density functions for both the null hypothesis (H 0 ) and alternate hypothesis (H 1 ) are known. In academic and hypothetical situations, the simple hypothesis test works for most cases. However, in real life it’s much more challenging to specify all of the pdfs for a particular situation.

Approaches to Composite Hypothesis Testing

composite hypothesis test

  • Bayesian approach : the unknown parameter is assigned a prior PDF.
  • Generalized likelihood ratio test approach : the unknown parameter is estimated and placed into a likelihood ratio test.

Composite Null Hypothesis

In real life, null hypotheses are usually composite unless the problem is very simple. An example of a composite hypothesis, which has multiple possible values, is: H 0 : μ ≥ 100

Ghobadzadeh, A. et. al. Separating Function Estimation Tests: A New Perspective on Binary Composite Hypothesis Testing. Retrieved August 19, 2019 from: http://post.queensu.ca/~gazor/T-SP-13550-2012.pdf Lindsey, J. Parametric Statistical Inference . Retrieved August 19, 2019 from: https://books.google.com/books?id=YnsQ-NnMxJ8C Nowak, R. (2010). Lecture 10: Composite Hypothesis Testing. Retrieved August 19, 2019 from: http://nowak.ece.wisc.edu/ece830/ece830_fall11_lecture10.pdf Lecture 29 : Bayesian Composite Hypothesis Testing. Retrieved August 19, 2019 from: https://nptel.ac.in/courses/117103018/module8/lec29/1.html

Leave a Comment

You must be logged in to post a comment.

Definition: Simple and composite hypothesis

Definition: Let $H$ be a statistical hypothesis . Then,

$H$ is called a simple hypothesis, if it completely specifies the population distribution; in this case, the sampling distribution of the test statistic is a function of sample size alone.

$H$ is called a composite hypothesis, if it does not completely specify the population distribution; for example, the hypothesis may only specify one parameter of the distribution and leave others unspecified.

  • Wikipedia (2021): "Exclusion of the null hypothesis" ; in: Wikipedia, the free encyclopedia , retrieved on 2021-03-19 ; URL: https://en.wikipedia.org/wiki/Exclusion_of_the_null_hypothesis#Terminology .

eMathZone

Simple Hypothesis and Composite Hypothesis

A simple hypothesis is one in which all parameters of the distribution are specified. For example, the heights of college students are normally distributed with $${\sigma ^2} = 4$$, and the hypothesis that its mean $$\mu $$ is, say, $$62”$$; that is, $${H_o}:\mu = 62$$. So we have stated a simple hypothesis, as the mean and variance together specify a normal distribution completely. A simple hypothesis, in general, states that $$\theta = {\theta _o}$$ where $${\theta _o}$$ is the specified value of a parameter $$\theta $$, ($$\theta $$ may represent $$\mu ,p,{\mu _1} – {\mu _2}$$ etc).

A hypothesis which is not simple (i.e. in which not all of the parameters are specified) is called a composite hypothesis. For instance, if we hypothesize that $${H_o}:\mu > 62$$ (and $${\sigma ^2} = 4$$) or$${H_o}:\mu = 62$$ and $${\sigma ^2} < 4$$, the hypothesis becomes a composite hypothesis because we cannot know the exact distribution of the population in either case. Obviously, the parameters $$\mu > 62”$$ and$${\sigma ^2} < 4$$ have more than one value and no specified values are being assigned. The general form of a composite hypothesis is $$\theta \leqslant {\theta _o}$$ or $$\theta \geqslant {\theta _o}$$; that is, the parameter $$\theta $$ does not exceed or does not fall short of a specified value $${\theta _o}$$. The concept of simple and composite hypotheses applies to both the null hypothesis and alternative hypothesis.

Hypotheses may also be classified as exact and inexact. A hypothesis is said to be an exact hypothesis if it selects a unique value for the parameter, such as $${H_o}:\mu = 62$$ or $$p > 0.5$$. A hypothesis is called an inexact hypothesis when it indicates more than one possible value for the parameter, such as $${H_o}:\mu \ne 62$$ or $${H_o}:p = 62$$. A simple hypothesis must be exact while an exact hypothesis is not necessarily a simple hypothesis. An inexact hypothesis is a composite hypothesis.

One Comment

Etini August 5 @ 7:28 pm

How can i design a sequential test for the shape parameters of the beta distribution

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

What does "Composite Hypothesis" mean?

Definition of Composite Hypothesis in the context of A/B testing (online controlled experiments).

What is a Composite Hypothesis?

In hypothesis testing a composite hypothesis is a hypothesis which covers a set of values from the parameter space. For example, if the entire parameter space covers -∞ to +∞ a composite hypothesis could be μ ≤ 0. It could be any other number as well, such 1, 2 or 3,1245. The alternative hypothesis is always a composite hypothesis : either one-sided hypothesis if the null is composite or a two-sided one if the null is a point null. The "composite" part means that such a hypothesis is the union of many simple point hypotheses.

In a Null Hypothesis Statistical Test only the null hypothesis can be a point hypothesis. Also, a composite hypothesis usually spans from -∞ to zero or some value of practical significance or from such a value to +∞.

Testing a composite null is what is most often of interest in an A/B testing scenario as we are usually interested in detecting and estimating effects in only one direction: either an increase in conversion rate or average revenue per user, or a decrease in unsubscribe events would be of interest and not its opposite. In fact, running a test so long as to detect a statistically significant negative outcome can result in significant business harm.

Like this glossary entry? For an in-depth and comprehensive reading on A/B testing stats, check out the book "Statistical Methods in Online A/B Testing" by the author of this glossary, Georgi Georgiev.

Articles on Composite Hypothesis

One-tailed vs Two-tailed Tests of Significance in A/B Testing blog.analytics-toolkit.com

Related A/B Testing terms

Purchase Statistical Methods in Online A/B Testing

Statistical Methods in Online A/B Testing

Take your A/B testing program to the next level with the most comprehensive book on user testing statistics in e-commerce.

Glossary index by letter

Select a letter to see all A/B testing terms starting with that letter or visit the Glossary homepage to see all.

  • Privacy Policy

Research Method

Home » What is a Hypothesis – Types, Examples and Writing Guide

What is a Hypothesis – Types, Examples and Writing Guide

Table of Contents

What is a Hypothesis

Definition:

Hypothesis is an educated guess or proposed explanation for a phenomenon, based on some initial observations or data. It is a tentative statement that can be tested and potentially proven or disproven through further investigation and experimentation.

Hypothesis is often used in scientific research to guide the design of experiments and the collection and analysis of data. It is an essential element of the scientific method, as it allows researchers to make predictions about the outcome of their experiments and to test those predictions to determine their accuracy.

Types of Hypothesis

Types of Hypothesis are as follows:

Research Hypothesis

A research hypothesis is a statement that predicts a relationship between variables. It is usually formulated as a specific statement that can be tested through research, and it is often used in scientific research to guide the design of experiments.

Null Hypothesis

The null hypothesis is a statement that assumes there is no significant difference or relationship between variables. It is often used as a starting point for testing the research hypothesis, and if the results of the study reject the null hypothesis, it suggests that there is a significant difference or relationship between variables.

Alternative Hypothesis

An alternative hypothesis is a statement that assumes there is a significant difference or relationship between variables. It is often used as an alternative to the null hypothesis and is tested against the null hypothesis to determine which statement is more accurate.

Directional Hypothesis

A directional hypothesis is a statement that predicts the direction of the relationship between variables. For example, a researcher might predict that increasing the amount of exercise will result in a decrease in body weight.

Non-directional Hypothesis

A non-directional hypothesis is a statement that predicts the relationship between variables but does not specify the direction. For example, a researcher might predict that there is a relationship between the amount of exercise and body weight, but they do not specify whether increasing or decreasing exercise will affect body weight.

Statistical Hypothesis

A statistical hypothesis is a statement that assumes a particular statistical model or distribution for the data. It is often used in statistical analysis to test the significance of a particular result.

Composite Hypothesis

A composite hypothesis is a statement that assumes more than one condition or outcome. It can be divided into several sub-hypotheses, each of which represents a different possible outcome.

Empirical Hypothesis

An empirical hypothesis is a statement that is based on observed phenomena or data. It is often used in scientific research to develop theories or models that explain the observed phenomena.

Simple Hypothesis

A simple hypothesis is a statement that assumes only one outcome or condition. It is often used in scientific research to test a single variable or factor.

Complex Hypothesis

A complex hypothesis is a statement that assumes multiple outcomes or conditions. It is often used in scientific research to test the effects of multiple variables or factors on a particular outcome.

Applications of Hypothesis

Hypotheses are used in various fields to guide research and make predictions about the outcomes of experiments or observations. Here are some examples of how hypotheses are applied in different fields:

  • Science : In scientific research, hypotheses are used to test the validity of theories and models that explain natural phenomena. For example, a hypothesis might be formulated to test the effects of a particular variable on a natural system, such as the effects of climate change on an ecosystem.
  • Medicine : In medical research, hypotheses are used to test the effectiveness of treatments and therapies for specific conditions. For example, a hypothesis might be formulated to test the effects of a new drug on a particular disease.
  • Psychology : In psychology, hypotheses are used to test theories and models of human behavior and cognition. For example, a hypothesis might be formulated to test the effects of a particular stimulus on the brain or behavior.
  • Sociology : In sociology, hypotheses are used to test theories and models of social phenomena, such as the effects of social structures or institutions on human behavior. For example, a hypothesis might be formulated to test the effects of income inequality on crime rates.
  • Business : In business research, hypotheses are used to test the validity of theories and models that explain business phenomena, such as consumer behavior or market trends. For example, a hypothesis might be formulated to test the effects of a new marketing campaign on consumer buying behavior.
  • Engineering : In engineering, hypotheses are used to test the effectiveness of new technologies or designs. For example, a hypothesis might be formulated to test the efficiency of a new solar panel design.

How to write a Hypothesis

Here are the steps to follow when writing a hypothesis:

Identify the Research Question

The first step is to identify the research question that you want to answer through your study. This question should be clear, specific, and focused. It should be something that can be investigated empirically and that has some relevance or significance in the field.

Conduct a Literature Review

Before writing your hypothesis, it’s essential to conduct a thorough literature review to understand what is already known about the topic. This will help you to identify the research gap and formulate a hypothesis that builds on existing knowledge.

Determine the Variables

The next step is to identify the variables involved in the research question. A variable is any characteristic or factor that can vary or change. There are two types of variables: independent and dependent. The independent variable is the one that is manipulated or changed by the researcher, while the dependent variable is the one that is measured or observed as a result of the independent variable.

Formulate the Hypothesis

Based on the research question and the variables involved, you can now formulate your hypothesis. A hypothesis should be a clear and concise statement that predicts the relationship between the variables. It should be testable through empirical research and based on existing theory or evidence.

Write the Null Hypothesis

The null hypothesis is the opposite of the alternative hypothesis, which is the hypothesis that you are testing. The null hypothesis states that there is no significant difference or relationship between the variables. It is important to write the null hypothesis because it allows you to compare your results with what would be expected by chance.

Refine the Hypothesis

After formulating the hypothesis, it’s important to refine it and make it more precise. This may involve clarifying the variables, specifying the direction of the relationship, or making the hypothesis more testable.

Examples of Hypothesis

Here are a few examples of hypotheses in different fields:

  • Psychology : “Increased exposure to violent video games leads to increased aggressive behavior in adolescents.”
  • Biology : “Higher levels of carbon dioxide in the atmosphere will lead to increased plant growth.”
  • Sociology : “Individuals who grow up in households with higher socioeconomic status will have higher levels of education and income as adults.”
  • Education : “Implementing a new teaching method will result in higher student achievement scores.”
  • Marketing : “Customers who receive a personalized email will be more likely to make a purchase than those who receive a generic email.”
  • Physics : “An increase in temperature will cause an increase in the volume of a gas, assuming all other variables remain constant.”
  • Medicine : “Consuming a diet high in saturated fats will increase the risk of developing heart disease.”

Purpose of Hypothesis

The purpose of a hypothesis is to provide a testable explanation for an observed phenomenon or a prediction of a future outcome based on existing knowledge or theories. A hypothesis is an essential part of the scientific method and helps to guide the research process by providing a clear focus for investigation. It enables scientists to design experiments or studies to gather evidence and data that can support or refute the proposed explanation or prediction.

The formulation of a hypothesis is based on existing knowledge, observations, and theories, and it should be specific, testable, and falsifiable. A specific hypothesis helps to define the research question, which is important in the research process as it guides the selection of an appropriate research design and methodology. Testability of the hypothesis means that it can be proven or disproven through empirical data collection and analysis. Falsifiability means that the hypothesis should be formulated in such a way that it can be proven wrong if it is incorrect.

In addition to guiding the research process, the testing of hypotheses can lead to new discoveries and advancements in scientific knowledge. When a hypothesis is supported by the data, it can be used to develop new theories or models to explain the observed phenomenon. When a hypothesis is not supported by the data, it can help to refine existing theories or prompt the development of new hypotheses to explain the phenomenon.

When to use Hypothesis

Here are some common situations in which hypotheses are used:

  • In scientific research , hypotheses are used to guide the design of experiments and to help researchers make predictions about the outcomes of those experiments.
  • In social science research , hypotheses are used to test theories about human behavior, social relationships, and other phenomena.
  • I n business , hypotheses can be used to guide decisions about marketing, product development, and other areas. For example, a hypothesis might be that a new product will sell well in a particular market, and this hypothesis can be tested through market research.

Characteristics of Hypothesis

Here are some common characteristics of a hypothesis:

  • Testable : A hypothesis must be able to be tested through observation or experimentation. This means that it must be possible to collect data that will either support or refute the hypothesis.
  • Falsifiable : A hypothesis must be able to be proven false if it is not supported by the data. If a hypothesis cannot be falsified, then it is not a scientific hypothesis.
  • Clear and concise : A hypothesis should be stated in a clear and concise manner so that it can be easily understood and tested.
  • Based on existing knowledge : A hypothesis should be based on existing knowledge and research in the field. It should not be based on personal beliefs or opinions.
  • Specific : A hypothesis should be specific in terms of the variables being tested and the predicted outcome. This will help to ensure that the research is focused and well-designed.
  • Tentative: A hypothesis is a tentative statement or assumption that requires further testing and evidence to be confirmed or refuted. It is not a final conclusion or assertion.
  • Relevant : A hypothesis should be relevant to the research question or problem being studied. It should address a gap in knowledge or provide a new perspective on the issue.

Advantages of Hypothesis

Hypotheses have several advantages in scientific research and experimentation:

  • Guides research: A hypothesis provides a clear and specific direction for research. It helps to focus the research question, select appropriate methods and variables, and interpret the results.
  • Predictive powe r: A hypothesis makes predictions about the outcome of research, which can be tested through experimentation. This allows researchers to evaluate the validity of the hypothesis and make new discoveries.
  • Facilitates communication: A hypothesis provides a common language and framework for scientists to communicate with one another about their research. This helps to facilitate the exchange of ideas and promotes collaboration.
  • Efficient use of resources: A hypothesis helps researchers to use their time, resources, and funding efficiently by directing them towards specific research questions and methods that are most likely to yield results.
  • Provides a basis for further research: A hypothesis that is supported by data provides a basis for further research and exploration. It can lead to new hypotheses, theories, and discoveries.
  • Increases objectivity: A hypothesis can help to increase objectivity in research by providing a clear and specific framework for testing and interpreting results. This can reduce bias and increase the reliability of research findings.

Limitations of Hypothesis

Some Limitations of the Hypothesis are as follows:

  • Limited to observable phenomena: Hypotheses are limited to observable phenomena and cannot account for unobservable or intangible factors. This means that some research questions may not be amenable to hypothesis testing.
  • May be inaccurate or incomplete: Hypotheses are based on existing knowledge and research, which may be incomplete or inaccurate. This can lead to flawed hypotheses and erroneous conclusions.
  • May be biased: Hypotheses may be biased by the researcher’s own beliefs, values, or assumptions. This can lead to selective interpretation of data and a lack of objectivity in research.
  • Cannot prove causation: A hypothesis can only show a correlation between variables, but it cannot prove causation. This requires further experimentation and analysis.
  • Limited to specific contexts: Hypotheses are limited to specific contexts and may not be generalizable to other situations or populations. This means that results may not be applicable in other contexts or may require further testing.
  • May be affected by chance : Hypotheses may be affected by chance or random variation, which can obscure or distort the true relationship between variables.

About the author

' src=

Muhammad Hassan

Researcher, Academic Writer, Web developer

You may also like

Research Methods

Research Methods – Types, Examples and Guide

Ethical Considerations

Ethical Considerations – Types, Examples and...

Research Results

Research Results Section – Writing Guide and...

Tables in Research Paper

Tables in Research Paper – Types, Creating Guide...

Informed Consent in Research

Informed Consent in Research – Types, Templates...

Research Paper

Research Paper – Structure, Examples and Writing...

Tutorial Playlist

Statistics tutorial, everything you need to know about the probability density function in statistics, the best guide to understand central limit theorem, an in-depth guide to measures of central tendency : mean, median and mode, the ultimate guide to understand conditional probability.

A Comprehensive Look at Percentile in Statistics

The Best Guide to Understand Bayes Theorem

Everything you need to know about the normal distribution, an in-depth explanation of cumulative distribution function, a complete guide to chi-square test, what is hypothesis testing in statistics types and examples, understanding the fundamentals of arithmetic and geometric progression, the definitive guide to understand spearman’s rank correlation, mean squared error: overview, examples, concepts and more, all you need to know about the empirical rule in statistics, the complete guide to skewness and kurtosis, a holistic look at bernoulli distribution.

All You Need to Know About Bias in Statistics

A Complete Guide to Get a Grasp of Time Series Analysis

The Key Differences Between Z-Test Vs. T-Test

The Complete Guide to Understand Pearson's Correlation

A complete guide on the types of statistical studies, everything you need to know about poisson distribution, your best guide to understand correlation vs. regression, the most comprehensive guide for beginners on what is correlation, hypothesis testing in statistics - types | examples.

Lesson 10 of 24 By Avijeet Biswal

What Is Hypothesis Testing in Statistics? Types and Examples

Table of Contents

In today’s data-driven world, decisions are based on data all the time. Hypothesis plays a crucial role in that process, whether it may be making business decisions, in the health sector, academia, or in quality improvement. Without hypothesis & hypothesis tests, you risk drawing the wrong conclusions and making bad decisions. In this tutorial, you will look at Hypothesis Testing in Statistics.

The Ultimate Ticket to Top Data Science Job Roles

The Ultimate Ticket to Top Data Science Job Roles

What Is Hypothesis Testing in Statistics?

Hypothesis Testing is a type of statistical analysis in which you put your assumptions about a population parameter to the test. It is used to estimate the relationship between 2 statistical variables.

Let's discuss few examples of statistical hypothesis from real-life - 

  • A teacher assumes that 60% of his college's students come from lower-middle-class families.
  • A doctor believes that 3D (Diet, Dose, and Discipline) is 90% effective for diabetic patients.

Now that you know about hypothesis testing, look at the two types of hypothesis testing in statistics.

Hypothesis Testing Formula

Z = ( x̅ – μ0 ) / (σ /√n)

  • Here, x̅ is the sample mean,
  • μ0 is the population mean,
  • σ is the standard deviation,
  • n is the sample size.

How Hypothesis Testing Works?

An analyst performs hypothesis testing on a statistical sample to present evidence of the plausibility of the null hypothesis. Measurements and analyses are conducted on a random sample of the population to test a theory. Analysts use a random population sample to test two hypotheses: the null and alternative hypotheses.

The null hypothesis is typically an equality hypothesis between population parameters; for example, a null hypothesis may claim that the population means return equals zero. The alternate hypothesis is essentially the inverse of the null hypothesis (e.g., the population means the return is not equal to zero). As a result, they are mutually exclusive, and only one can be correct. One of the two possibilities, however, will always be correct.

Your Dream Career is Just Around The Corner!

Your Dream Career is Just Around The Corner!

Null Hypothesis and Alternative Hypothesis

The Null Hypothesis is the assumption that the event will not occur. A null hypothesis has no bearing on the study's outcome unless it is rejected.

H0 is the symbol for it, and it is pronounced H-naught.

The Alternate Hypothesis is the logical opposite of the null hypothesis. The acceptance of the alternative hypothesis follows the rejection of the null hypothesis. H1 is the symbol for it.

Let's understand this with an example.

A sanitizer manufacturer claims that its product kills 95 percent of germs on average. 

To put this company's claim to the test, create a null and alternate hypothesis.

H0 (Null Hypothesis): Average = 95%.

Alternative Hypothesis (H1): The average is less than 95%.

Another straightforward example to understand this concept is determining whether or not a coin is fair and balanced. The null hypothesis states that the probability of a show of heads is equal to the likelihood of a show of tails. In contrast, the alternate theory states that the probability of a show of heads and tails would be very different.

Become a Data Scientist with Hands-on Training!

Become a Data Scientist with Hands-on Training!

Hypothesis Testing Calculation With Examples

Let's consider a hypothesis test for the average height of women in the United States. Suppose our null hypothesis is that the average height is 5'4". We gather a sample of 100 women and determine that their average height is 5'5". The standard deviation of population is 2.

To calculate the z-score, we would use the following formula:

z = ( x̅ – μ0 ) / (σ /√n)

z = (5'5" - 5'4") / (2" / √100)

z = 0.5 / (0.045)

We will reject the null hypothesis as the z-score of 11.11 is very large and conclude that there is evidence to suggest that the average height of women in the US is greater than 5'4".

Steps in Hypothesis Testing

Hypothesis testing is a statistical method to determine if there is enough evidence in a sample of data to infer that a certain condition is true for the entire population. Here’s a breakdown of the typical steps involved in hypothesis testing:

Formulate Hypotheses

  • Null Hypothesis (H0): This hypothesis states that there is no effect or difference, and it is the hypothesis you attempt to reject with your test.
  • Alternative Hypothesis (H1 or Ha): This hypothesis is what you might believe to be true or hope to prove true. It is usually considered the opposite of the null hypothesis.

Choose the Significance Level (α)

The significance level, often denoted by alpha (α), is the probability of rejecting the null hypothesis when it is true. Common choices for α are 0.05 (5%), 0.01 (1%), and 0.10 (10%).

Select the Appropriate Test

Choose a statistical test based on the type of data and the hypothesis. Common tests include t-tests, chi-square tests, ANOVA, and regression analysis. The selection depends on data type, distribution, sample size, and whether the hypothesis is one-tailed or two-tailed.

Collect Data

Gather the data that will be analyzed in the test. This data should be representative of the population to infer conclusions accurately.

Calculate the Test Statistic

Based on the collected data and the chosen test, calculate a test statistic that reflects how much the observed data deviates from the null hypothesis.

Determine the p-value

The p-value is the probability of observing test results at least as extreme as the results observed, assuming the null hypothesis is correct. It helps determine the strength of the evidence against the null hypothesis.

Make a Decision

Compare the p-value to the chosen significance level:

  • If the p-value ≤ α: Reject the null hypothesis, suggesting sufficient evidence in the data supports the alternative hypothesis.
  • If the p-value > α: Do not reject the null hypothesis, suggesting insufficient evidence to support the alternative hypothesis.

Report the Results

Present the findings from the hypothesis test, including the test statistic, p-value, and the conclusion about the hypotheses.

Perform Post-hoc Analysis (if necessary)

Depending on the results and the study design, further analysis may be needed to explore the data more deeply or to address multiple comparisons if several hypotheses were tested simultaneously.

Types of Hypothesis Testing

To determine whether a discovery or relationship is statistically significant, hypothesis testing uses a z-test. It usually checks to see if two means are the same (the null hypothesis). Only when the population standard deviation is known and the sample size is 30 data points or more, can a z-test be applied.

A statistical test called a t-test is employed to compare the means of two groups. To determine whether two groups differ or if a procedure or treatment affects the population of interest, it is frequently used in hypothesis testing.

Chi-Square 

You utilize a Chi-square test for hypothesis testing concerning whether your data is as predicted. To determine if the expected and observed results are well-fitted, the Chi-square test analyzes the differences between categorical variables from a random sample. The test's fundamental premise is that the observed values in your data should be compared to the predicted values that would be present if the null hypothesis were true.

Hypothesis Testing and Confidence Intervals

Both confidence intervals and hypothesis tests are inferential techniques that depend on approximating the sample distribution. Data from a sample is used to estimate a population parameter using confidence intervals. Data from a sample is used in hypothesis testing to examine a given hypothesis. We must have a postulated parameter to conduct hypothesis testing.

Bootstrap distributions and randomization distributions are created using comparable simulation techniques. The observed sample statistic is the focal point of a bootstrap distribution, whereas the null hypothesis value is the focal point of a randomization distribution.

A variety of feasible population parameter estimates are included in confidence ranges. In this lesson, we created just two-tailed confidence intervals. There is a direct connection between these two-tail confidence intervals and these two-tail hypothesis tests. The results of a two-tailed hypothesis test and two-tailed confidence intervals typically provide the same results. In other words, a hypothesis test at the 0.05 level will virtually always fail to reject the null hypothesis if the 95% confidence interval contains the predicted value. A hypothesis test at the 0.05 level will nearly certainly reject the null hypothesis if the 95% confidence interval does not include the hypothesized parameter.

Become a Data Scientist through hands-on learning with hackathons, masterclasses, webinars, and Ask-Me-Anything! Start learning now!

Simple and Composite Hypothesis Testing

Depending on the population distribution, you can classify the statistical hypothesis into two types.

Simple Hypothesis: A simple hypothesis specifies an exact value for the parameter.

Composite Hypothesis: A composite hypothesis specifies a range of values.

A company is claiming that their average sales for this quarter are 1000 units. This is an example of a simple hypothesis.

Suppose the company claims that the sales are in the range of 900 to 1000 units. Then this is a case of a composite hypothesis.

One-Tailed and Two-Tailed Hypothesis Testing

The One-Tailed test, also called a directional test, considers a critical region of data that would result in the null hypothesis being rejected if the test sample falls into it, inevitably meaning the acceptance of the alternate hypothesis.

In a one-tailed test, the critical distribution area is one-sided, meaning the test sample is either greater or lesser than a specific value.

In two tails, the test sample is checked to be greater or less than a range of values in a Two-Tailed test, implying that the critical distribution area is two-sided.

If the sample falls within this range, the alternate hypothesis will be accepted, and the null hypothesis will be rejected.

Become a Data Scientist With Real-World Experience

Become a Data Scientist With Real-World Experience

Right Tailed Hypothesis Testing

If the larger than (>) sign appears in your hypothesis statement, you are using a right-tailed test, also known as an upper test. Or, to put it another way, the disparity is to the right. For instance, you can contrast the battery life before and after a change in production. Your hypothesis statements can be the following if you want to know if the battery life is longer than the original (let's say 90 hours):

  • The null hypothesis is (H0 <= 90) or less change.
  • A possibility is that battery life has risen (H1) > 90.

The crucial point in this situation is that the alternate hypothesis (H1), not the null hypothesis, decides whether you get a right-tailed test.

Left Tailed Hypothesis Testing

Alternative hypotheses that assert the true value of a parameter is lower than the null hypothesis are tested with a left-tailed test; they are indicated by the asterisk "<".

Suppose H0: mean = 50 and H1: mean not equal to 50

According to the H1, the mean can be greater than or less than 50. This is an example of a Two-tailed test.

In a similar manner, if H0: mean >=50, then H1: mean <50

Here the mean is less than 50. It is called a One-tailed test.

Type 1 and Type 2 Error

A hypothesis test can result in two types of errors.

Type 1 Error: A Type-I error occurs when sample results reject the null hypothesis despite being true.

Type 2 Error: A Type-II error occurs when the null hypothesis is not rejected when it is false, unlike a Type-I error.

Suppose a teacher evaluates the examination paper to decide whether a student passes or fails.

H0: Student has passed

H1: Student has failed

Type I error will be the teacher failing the student [rejects H0] although the student scored the passing marks [H0 was true]. 

Type II error will be the case where the teacher passes the student [do not reject H0] although the student did not score the passing marks [H1 is true].

Our Data Scientist Master's Program covers core topics such as R, Python, Machine Learning, Tableau, Hadoop, and Spark. Get started on your journey today!

Limitations of Hypothesis Testing

Hypothesis testing has some limitations that researchers should be aware of:

  • It cannot prove or establish the truth: Hypothesis testing provides evidence to support or reject a hypothesis, but it cannot confirm the absolute truth of the research question.
  • Results are sample-specific: Hypothesis testing is based on analyzing a sample from a population, and the conclusions drawn are specific to that particular sample.
  • Possible errors: During hypothesis testing, there is a chance of committing type I error (rejecting a true null hypothesis) or type II error (failing to reject a false null hypothesis).
  • Assumptions and requirements: Different tests have specific assumptions and requirements that must be met to accurately interpret results.

Learn All The Tricks Of The BI Trade

Learn All The Tricks Of The BI Trade

After reading this tutorial, you would have a much better understanding of hypothesis testing, one of the most important concepts in the field of Data Science . The majority of hypotheses are based on speculation about observed behavior, natural phenomena, or established theories.

If you are interested in statistics of data science and skills needed for such a career, you ought to explore the Post Graduate Program in Data Science.

If you have any questions regarding this ‘Hypothesis Testing In Statistics’ tutorial, do share them in the comment section. Our subject matter expert will respond to your queries. Happy learning!

1. What is hypothesis testing in statistics with example?

Hypothesis testing is a statistical method used to determine if there is enough evidence in a sample data to draw conclusions about a population. It involves formulating two competing hypotheses, the null hypothesis (H0) and the alternative hypothesis (Ha), and then collecting data to assess the evidence. An example: testing if a new drug improves patient recovery (Ha) compared to the standard treatment (H0) based on collected patient data.

2. What is H0 and H1 in statistics?

In statistics, H0​ and H1​ represent the null and alternative hypotheses. The null hypothesis, H0​, is the default assumption that no effect or difference exists between groups or conditions. The alternative hypothesis, H1​, is the competing claim suggesting an effect or a difference. Statistical tests determine whether to reject the null hypothesis in favor of the alternative hypothesis based on the data.

3. What is a simple hypothesis with an example?

A simple hypothesis is a specific statement predicting a single relationship between two variables. It posits a direct and uncomplicated outcome. For example, a simple hypothesis might state, "Increased sunlight exposure increases the growth rate of sunflowers." Here, the hypothesis suggests a direct relationship between the amount of sunlight (independent variable) and the growth rate of sunflowers (dependent variable), with no additional variables considered.

4. What are the 3 major types of hypothesis?

The three major types of hypotheses are:

  • Null Hypothesis (H0): Represents the default assumption, stating that there is no significant effect or relationship in the data.
  • Alternative Hypothesis (Ha): Contradicts the null hypothesis and proposes a specific effect or relationship that researchers want to investigate.
  • Nondirectional Hypothesis: An alternative hypothesis that doesn't specify the direction of the effect, leaving it open for both positive and negative possibilities.

Find our PL-300 Microsoft Power BI Certification Training Online Classroom training classes in top cities:

NameDatePlace
7 Sep -22 Sep 2024,
Weekend batch
Your City
21 Sep -6 Oct 2024,
Weekend batch
Your City
12 Oct -27 Oct 2024,
Weekend batch
Your City

About the Author

Avijeet Biswal

Avijeet is a Senior Research Analyst at Simplilearn. Passionate about Data Analytics, Machine Learning, and Deep Learning, Avijeet is also interested in politics, cricket, and football.

Recommended Resources

The Key Differences Between Z-Test Vs. T-Test

Free eBook: Top Programming Languages For A Data Scientist

Normality Test in Minitab: Minitab with Statistics

Normality Test in Minitab: Minitab with Statistics

A Comprehensive Look at Percentile in Statistics

Machine Learning Career Guide: A Playbook to Becoming a Machine Learning Engineer

  • PMP, PMI, PMBOK, CAPM, PgMP, PfMP, ACP, PBA, RMP, SP, and OPM3 are registered marks of the Project Management Institute, Inc.

Lesson 27: Likelihood Ratio Tests

In this lesson, we'll learn how to apply a method for developing a hypothesis test for situations in which both the null and alternative hypotheses are composite. That's not completely accurate. The method, called the likelihood ratio test , can be used even when the hypotheses are simple, but it is most commonly used when the alternative hypothesis is composite. Throughout the lesson, we'll continue to assume that we know the the functional form of the probability density (or mass) function, but we don't know the value of one (or more) of its parameters. That is, we might know that the data come from a normal distrbution, but we don't know the mean or variance of the distribution, and hence the interest in performing a hypothesis test about the unknown parameter(s).

27.1 - A Definition and Simple Example

The title of this page is a little risky, as there are few simple examples when it comes to likelihood ratio testing! But, we'll work to make the example as simple as possible, namely by assuming again, unrealistically, that we know the population variance, but not the population mean. Before we state the definition of a likelihood ratio test, and then investigate our simple, but unrealistic, example, we first need to define some notation that we'll use throughout the lesson.

We'll assume that the probability density (or mass) function of X is \(f(x;\theta)\) where \(\theta\) represents one or more unknown parameters. Then:

  • Let \(\Omega\) (greek letter "omega") denote the total possible parameter space of \(\theta\), that is, the set of all possible values of \(\theta\) as specified in totality in the null and alternative hypotheses.
  • Let \(H_0 : \theta \in \omega\) denote the null hypothesis where \(\omega\) (greek letter "omega") is a subset of the parameter space \(\Omega\).
  • Let \(H_A : \theta \in \omega'\) denote the alternative hypothesis where \(\omega '\) is the complement of \(\omega\) with respect to the parameter space \(\Omega\).

Let's make sure we are clear about that phrase "where \(\omega '\) is the complement of \(\omega\) with respect to the parameter space \(\Omega\)."

Example 27-1

If the total parameter space of the mean \(\mu\) is \(\Omega = {\mu: −∞ < \mu < ∞}\) and the null hypothesis is specified as \(H_0: \mu = 3\), how should we specify the alternative hypothesis so that the alternative parameter space is the complement of the null parameter space?

If the null parameter space is \(\Omega = {\mu: \mu = 3}\), then the alternative parameter space is everything that is in \(\Omega = {\mu: −∞ < \mu < ∞}\) that is not in \(\Omega\). That is, the alternative parameter space is \(\Omega ' = {\mu: \mu ≠ 3}\). And, so the alternative hypothesis is:

\(H_A : \mu \ne 3\)

In this case, we'd be interested in deriving a two-tailed test.

Example 27-2

If the alternative hypothesis is \(H_A: \mu > 3\), how should we (technically) specify the null hypothesis so that the null parameter space is the complement of the alternative parameter space?

If the alternative parameter space is (\omega ' = {\mu: \mu > 3}\), then the null parameter space is \(\omega = {\mu: \mu ≤ 3}\). And, so the null hypothesis is:

\(H_0 : \mu \le 3\)

Now, the reality is that some authors do specify the null hypothesis as such, even when they mean \(H_0: \mu = 3\). Ours don't, and so we won't. (That's why I put that "technically" in parentheses up above.) At any rate, in this case, we'd be interested in deriving a one-tailed test.

Definition. Let:

\(L(\hat{\omega})\) denote the maximum of the likelihood function with respect to \(\theta\) when \(\theta\) is in the null parameter space \(\omega\).

\(L(\hat{\Omega})\) denote the maximum of the likelihood function with respect to \(\theta\) when \(\theta\) is in the entire parameter space \(\Omega\).

Then, the likelihood ratio is the quotient:

\(\lambda = \dfrac{L(\hat{\omega})}{L(\hat{\Omega})}\)

And, to test the null hypothesis \(H_0 : \theta \in \omega\) against the alternative hypothesis \(H_A : \theta \in \omega'\), the critical region for the likelihood ratio test is the set of sample points for which:

\(\lambda = \dfrac{L(\hat{\omega})}{L(\hat{\Omega})} \le k\)

where \(0 < k < 1\), and k is selected so that the test has a desired significance level \(\alpha\).

Example 27-3

Honeyspoon

A food processing company packages honey in small glass jars. Each jar is supposed to contain 10 fluid ounces of the sweet and gooey good stuff. Previous experience suggests that the volume X , the volume in fluid ounces of a randomly selected jar of the company's honey is normally distributed with a known variance of 2. Derive the likelihood ratio test for testing, at a significance level of \(\alpha = 0.05\), the null hypothesis \(H_0: \mu = 10\) against the alternative hypothesis H_A: \mu ≠ 10\).

Because we are interested in testing the null hypothesis \(H_0: \mu = 10\) against the alternative hypothesis \(H_A: \mu ≠ 10\) for a normal mean, our total parameter space is:

\(\Omega =\left \{\mu : -\infty < \mu < \infty \right \}\)

and our null parameter space is:

\(\omega = \left \{10\right \}\)

Now, to find the likelihood ratio, as defined above, we first need to find \(L(\hat{\omega})\). Well, when the null hypothesis \(H_0: \mu = 10\) is true, the mean \(\mu\) can take on only one value, namely, \(\mu = 10\). Therefore:

\(L(\hat{\omega}) = L(10)\)

We also need to find \(L(\hat{\Omega})\) in order to define the likelihood ratio. To find it, we must find the value of \(\mu\)   that maximizes \(L(\mu)\) . Well, we did that back when we studied maximum likelihood as a method of estimation. We showed that \(\hat{\mu} = \bar{x}\) is the maximum likelihood estimate of \(\mu\) . Therefore:

\(L(\hat{\Omega}) = L(\bar{x})\)

Now, putting it all together to form the likelihood ratio, we get:

which simplifies to:

Now, let's step aside for a minute and focus just on the summation in the numerator. If we "add 0" in a special way to the quantity in parentheses:

we can show that the summation can be written as:

\(\sum_{i=1}^{n}(x_i - 10)^2 = \sum_{i=1}^{n}(x_i - \bar{x})^2 + n(\bar{x} -10)^2 \)

Therefore, the likelihood ratio becomes:

which greatly simplifies to:

\(\lambda = exp \left [-\dfrac{n}{4}(\bar{x}-10)^2 \right ]\)

Now, the likelihood ratio test tells us to reject the null hypothesis when the likelihood ratio \(\lambda\) is small, that is, when:

\(\lambda = exp\left[-\dfrac{n}{4}(\bar{x}-10)^2 \right] \le k\)

where k is chosen to ensure that, in this case, \(\alpha = 0.05\). Well, by taking the natural log of both sides of the inequality, we can show that \(\lambda ≤ k\) is equivalent to:

\( -\dfrac{n}{4}(\bar{x}-10)^2 \le \text{ln} k \)

which, by multiplying through by −4/ n , is equivalent to:

\((\bar{x}-10)^2 \ge -\dfrac{4}{n} \text{ln} k \)

which is equivalent to:

\(\dfrac{|\bar{X}-10|}{\sigma / \sqrt{n}} \ge \dfrac{\sqrt{-(4/n)\text{ln} k}}{\sigma / \sqrt{n}} =k* \)

Aha! We should recognize that quantity on the left-side of the inequality! We know that:

\(Z = \dfrac{\bar{X}-10}{\sigma / \sqrt{n}} \)

follows a standard normal distribution when \(H_0: \mu = 10\). Therefore we can determine the appropriate \(k^*\) by using the standard normal table. We have shown that the likelihood ratio test tells us to reject the null hypothesis \(H_0: \mu = 10\) in favor of the alternative hypothesis \(H_A: \mu ≠ 10\) for all sample means for which the following holds:

\(\dfrac{|\bar{X}-10|}{ \sqrt{2} / \sqrt{n}} \ge z_{0.025} = 1.96 \)

Doing so will ensure that our probability of committing a Type I error is set to \(\alpha = 0.05\), as desired.

27.2 - The T-Test For One Mean

Well, geez, now why would we be revisiting the t -test for a mean \(\mu\) when we have already studied it back in the hypothesis testing section? Well, the answer, it turns out, is that, as we'll soon see, the t -test for a mean \(\mu\) is the likelihood ratio test! Let's take a look!

Example 27-4

sunset

Suppose that a random sample \(X_1 , X_2 , \dots , X_n\) arises from a normal population with unknown mean \(\mu\) and unknown variance \(\sigma^2\). (Yes, back to the realistic situation, in which we don't know the population variance either.) Find the size \(\alpha\) likelihood ratio test for testing the null hypothesis \(H_0: \mu = \mu_0\) against the two-sided alternative hypothesis \(H_A: \mu ≠ \mu_0\) .

Our unrestricted parameter space is:

\( \Omega = \left\{ (\mu, \sigma^2) : -\infty < \mu < \infty, 0 < \sigma^2 < \infty \right\} \)

Under the null hypothesis, the mean \(\mu\) is the only parameter that is restricted. Therefore, our parameter space under the null hypothesis is:

\( \omega = \left\{(\mu, \sigma^2) : \mu =\mu_0, 0 < \sigma^2 < \infty \right\}\)

Now, first consider the case where the mean and variance are unrestricted. We showed back when we studied maximum likelihood estimation that the maximum likelihood estimates of \(\mu\) and \(\sigma^2\) are, respectively:

\(\hat{\mu} = \bar{x} \text{ and } \hat{\sigma}^2 = \dfrac{1}{n}\sum_{i=1}^{n}(x_i - \bar{x})^2 \)

Therefore, the maximum of the likelihood function for the unrestricted parameter space is:

\( L(\hat{\Omega})= \left[\dfrac{ne^{-1}}{2\pi \Sigma (x_i - \bar{x})^2} \right]^{n/2} \)

Now, under the null parameter space, the maximum likelihood estimates of \(\mu\)   and \(\sigma^2\) are, respectively:

\( \hat{\mu} = \mu_0 \text{ and } \hat{\sigma}^2 = \dfrac{1}{n}\sum_{i=1}^{n}(x_i - \mu_0)^2 \)

Therefore, the likelihood under the null hypothesis is:

\( L(\hat{\omega})= \left[\dfrac{ne^{-1}}{2\pi \Sigma (x_i - \mu_0)^2} \right]^{n/2} \)

And now taking the ratio of the two likelihoods, we get:

which reduces to:

\( \lambda = \left[ \dfrac{\sum_{i=1}^{n}(x_i - \bar{x})^2}{\sum_{i=1}^{n}(x_i - \mu_0)^2} \right] ^{n/2}\)

Focusing only on the denominator for a minute, let's do that trick again of "adding 0" in just the right away. Adding 0 to the quantity in the parentheses, we get:

\( \sum_{i=1}^{n}(x_i - \mu_0)^2 = \sum_{i=1}^{n}(x_i - \bar{x})^2 +n(\bar{x} - \mu_0)^2 \)

Then, our likelihood ratio \(\lambda\) becomes:

\( \lambda = \left[ \dfrac{\sum_{i=1}^{n}(x_i - \bar{x})^2}{\sum_{i=1}^{n}(x_i - \mu_0)^2} \right] ^{n/2} = \left[ \dfrac{\sum_{i=1}^{n}(x_i - \bar{x})^2}{ \sum_{i=1}^{n}(x_i - \bar{x})^2 +n(\bar{x} - \mu_0)^2} \right] ^{n/2} \)

which, upon dividing through numerator and denominator by \( \sum_{i=1}^{n}(x_i - \bar{x})^2 \) simplifies to:

Therefore, the likelihood ratio test's critical region, which is given by the inequality \(\lambda ≤ k\), is equivalent to:

which with some minor algebraic manipulation can be shown to be equivalent to:

So, in a nutshell, we've shown that the likelihood ratio test tells us that for this situation we should reject the null hypothesis \(H_0: \mu= \mu_0\) in favor of the alternative hypothesis \(H_A: \mu ≠ \mu_0\)   if:

\( \dfrac{(\bar{x}-\mu_0)^2 }{s^2 / n} \ge k^{*} \)

Well, okay, so I started out this page claiming that the t -test for a mean \(\mu\) is the likelihood ratio test. Is it? Well, the above critical region is equivalent to rejecting the null hypothesis if:

\( \dfrac{|\bar{x}-\mu_0| }{s / \sqrt{n}} \ge k^{**} \)

Does that look familiar? We previously learned that if \(X_1, X_2, \dots, X_n\) are normally distributed with mean \(\mu\) and variance \(\sigma^2\), then:

\( T = \dfrac{\bar{X}-\mu}{S / \sqrt{n}} \)

follows a T distribution with n − 1 degrees of freedom. So, this tells us that we should use the T distribution to choose \(k^{**}\) . That is, set:

\(k^{**} = t_{\alpha /2, n-1}\)

and we have our size \(\alpha\) t -test that ensures the probability of committing a Type I error is \(\alpha\).

It turns out... we didn't know it at the time... but every hypothesis test that we derived in the hypothesis testing section is a likelihood ratio test. Back then, we derived each test using distributional results of the relevant statistic(s), but we could have alternatively, and perhaps just as easily, derived the tests using the likelihood ratio testing method.

Biostatistics

  • Data Science
  • Programming
  • Social Science
  • Certificates
  • Undergraduate
  • For Businesses
  • FAQs and Knowledge Base

Test Yourself

  • Instructors

Composite Hypothesis

Composite Hypothesis:

A statistical hypothesis which does not completely specify the distribution of a random variable is referred to as a composite hypothesis.

Browse Other Glossary Entries

Planning on taking an introductory statistics course, but not sure if you need to start at the beginning? Review the course description for each of our introductory statistics courses and estimate which best matches your level, then take the self test for that course. If you get all or almost all the questions correct, move on and take the next test.

Data Analytics

Considering becoming adata scientist, customer analyst or our data science certificate program?

Advanced Statistics Quiz

Looking at statistics for graduate programs or to enhance your foundational knowledge?

Regression Quiz

Entering the biostatistics field? Test your skill here.

Stay Informed

Read up on our latest blogs

Learn about our certificate programs

Find the right course for you

We'd love to answer your questions

Our mentors and academic advisors are standing by to help guide you towards the courses or program that makes the most sense for you and your goals.

300 W Main St STE 301, Charlottesville, VA 22903

(434) 973-7673

[email protected]

By submitting your information, you agree to receive email communications from Statistics.com. All information submitted is subject to our privacy policy . You may opt out of receiving communications at any time.

What is the difference between Simple and Composite Hypothesis.

what is composite hypothesis

What is a hypothesis?

A h ypothesis is an educated guess about how something works. In the scientific method, a hypothesis is an idea that can be tested. If the hypothesis is correct, then the experiment will support the hypothesis. If the hypothesis is incorrect, the experiment will not support the hypothesis.

A hypothesis is simple if it specifies the population completely, i.e., it specifies the population distribution uniquely, while a composite hypothesis leads to two or more possibilities.

Before diving further into their differences, let’s first define a few terms that are handy in understanding the concept of a hypothesis.

Let’s dive in;

Difference between hypothesis and theory.

A hypothesis is a proposed explanation for a phenomenon. A scientific theory is a well-substantiated explanation for an aspect of the natural world supported by a vast body of evidence. Theories are generally much broader in scope than hypotheses and are often not as specific.

The objective of statistics is to make inferences about a population based on information contained in the sample.

There are two major areas of statistical inference, namely;

  • Estimation of parameter
  • Hypothesis testing

We will develop general methods for testing hypotheses and then apply them to common problems.

Statistical hypothesis

A statistical hypothesis is a testable statement about a population parameter. The statement is based on an assumption about the population parameter. This assumption is usually made about the population parameters based on past research or experience. The statistical hypothesis is used to make predictions about future events. These predictions are based on the assumption that the population parameters will remain the same.

A statistical hypothesis is about a population parameter, usually denoted by some symbol, such as μ or θ.

Statistical hypothesis testing is a method of statistical inference. There are two types of statistical hypothesis tests:

  • A point null hypothesis specifies that a population parameter (such as the mean) equals a specific value. For example, the null hypothesis could be that μ=0.
  • A composite null hypothesis specifies that a population parameter is less than, greater than, or not equal to a specific value. For example, the null hypothesis could be that μ≠0.

The alternative hypothesis is the hypothesis that is being tested against the null hypothesis. The alternative hypothesis could be that μ>0 or μ<0.

A statistical hypothesis test determines whether or not to reject the null hypothesis. The null hypothesis is rejected if the test statistic is greater than or less than the critical value.

Hypothesis Testing

A hypothesis is a statement or claims about how two variables are related. Hypothesis testing is a statistical procedure used to assess whether the null hypothesis—a statement that there is no difference between two groups or no association between two variables—can be rejected based on sample data. There are four steps in hypothesis testing:

  • State the null and alternative hypotheses.
  • Select a significance level.
  • Calculate the test statistic.
  • Interpret the results.

The first step is to state the null and alternative hypotheses. The null hypothesis is that the two variables have no difference or association. The alternative hypothesis is the statement that there is a difference or an association between two variables.

The second step is to select a significance level. The significance level is the probability of rejecting the null hypothesis when it is true. The most common significance levels are 0.05 and 0.01.

The third step is to calculate the test statistic. The test statistic measures the difference between the null and alternative hypotheses. There are many different test statistics, and the choice of test statistic depends on the data type and hypothesis test.

The fourth and final step is to interpret the results. The results of a hypothesis test are either significant or not significant. A significant result means that the null hypothesis can be rejected. A non-significant result means that the null hypothesis cannot be rejected.

hypothesis - simple and composite

Null Hypothesis vs Alternative Hypothesis

In statistics, a null hypothesis is a statement one seeks to disprove, reject or nullify. Most commonly, it is a statement that the phenomenon being studied produces no effect or makes no difference. For example, if one were testing the efficacy of a new drug, the null hypothesis would be that the drug does not affect the treated condition.

The null hypothesis is usually denoted H0, and the alternate hypothesis is denoted H1. If the null hypothesis is rejected in favor of the alternative hypothesis, it is said to be “statistically significant.” The null hypothesis is often assumed to be true until it can be proved otherwise.

Many different types of tests can be used to test a null hypothesis. The most common is the Student’s t-test, which compares the means of two groups. If the t-test is significant, there is a statistically significant difference between the two groups.

Other tests that can be used to test the null hypothesis include the chi-square, Fisher’s exact, and Wilcoxon rank-sum tests.

The alternative hypothesis is the hypothesis that is being tested in a statistical test. This is the hypothesis that is the opposite of the null hypothesis. We are trying to find evidence for the alternative hypothesis in a test.

Simple and Composite Hypothesis

Simple hypothesis.

Hypotheses can be composite or simple, and both are useful depending on the research question and the available evidence.

A simple hypothesis is a straightforward statement that proposes a relationship between two variables. It is a clear, concise statement that is easy to test and evaluate. A simple hypothesis is often used in experimental research where the researcher wants to test the effect of one variable on another.

Examples of hypothesis :

An example of a simple hypothesis is “students who study more will get better grades.” This hypothesis proposes a direct relationship between the amount of time a student spends studying and their academic performance. This hypothesis is testable by comparing the grades of students who study more with those who study less.

Another example of a simple hypothesis is “increased exposure to sunlight will result in higher vitamin D levels.” This hypothesis proposes a direct relationship between sunlight exposure and vitamin D levels. This hypothesis is testable by measuring the vitamin D levels of individuals with varying levels of sunlight exposure.

Simple hypotheses are advantageous because they are easy to test and evaluate. They also allow researchers to focus on a specific research question and avoid unnecessary complexity. Simple hypotheses are particularly useful in experimental research where researchers manipulate one variable to observe its effect on another.

However, simple hypotheses also have limitations. They may oversimplify complex phenomena, and their results may not generalize to a larger population. The available evidence may also limit simple hypotheses, and additional research may be necessary to understand the relationship between variables fully.

In essence, a simple hypothesis is a straightforward statement that proposes a relationship between two variables. Simple hypotheses are useful in experimental research and allow researchers to focus on a specific research question. However, simple hypotheses also have limitations and should be evaluated in the context of the available evidence and research question.

Composite Hypothesis

A composite hypothesis, on the other hand, proposes multiple relationships between two or more variables. For example, a composite hypothesis might state that “there is a significant difference between the average heights of men and women, and there is also a significant difference between the average heights of people from different continents.”

Composite hypothesis testing is a statistical technique used to determine the probability of an event or phenomenon based on observed data. This technique is often used in scientific research, quality control, and decision-making processes where the outcome of a particular experiment or test is uncertain.

A composite hypothesis is an alternative hypothesis encompassing a range of possible outcomes. It is defined as a hypothesis with more than one parameter value. For example, if we are testing the hypothesis that the mean of a population is greater than a certain value, we could define the composite hypothesis as follows:

H1: μ > μ0, where μ is the population means, and μ0 is the hypothesized value of the mean.

The composite hypothesis, in this case, includes all values of μ greater than μ0. This means we are not specifying a specific value of μ, but rather a range of possible values.

Composite hypothesis testing involves evaluating the probability of observing a particular result under the null hypothesis and then comparing it to the probability of observing the same result under the composite hypothesis. The result is considered significant if the probability of observing it under the composite hypothesis is sufficiently low.

We use statistical tests such as the t-test, F-test, or chi-square test to test a composite hypothesis. Given the null hypothesis and the observed data, these tests allow us to calculate the probability of observing a particular result.

In conclusion, composite hypothesis testing is a valuable statistical technique used to determine the probability of an event or phenomenon based on observed data. It allows us to test hypotheses that encompass a range of possible outcomes and is an essential tool for scientific research, quality control , and decision-making processes.

Understanding composite hypothesis testing is essential for anyone working in these fields and can help ensure that decisions are made based on solid statistical evidence.

what is composite hypothesis

Post navigation

Previous post.

How to create a normal distribution in R with examples

How to Use A Log Transformation in R To Rescale Your Data

5 Best Practices for Database Partitioning in Cloud Environments

5 Best Practices for Database Partitioning in Cloud Environments

Horizontal and Vertical Partitioning in Databases

Horizontal and Vertical Partitioning in Databases

Data Cleaning with Proven Strategies

Beginner’s Guide to Tidying Up Your Datasets: Data Cleaning 101 with Proven Strategies

U.S. flag

An official website of the United States government

The .gov means it’s official. Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

The site is secure. The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

  • Publications
  • Account settings

Preview improvements coming to the PMC website in October 2024. Learn More or Try it out now .

  • Advanced Search
  • Journal List
  • Health Serv Res
  • v.41(3 Pt 1); 2006 Jun

When to Combine Hypotheses and Adjust for Multiple Tests

Associated data.

To provide guidelines for identifying composite hypotheses and addressing the probability of false rejection for multiple hypotheses.

Data Sources and Study Setting

Examples from the literature in health services research are used to motivate the discussion of composite hypothesis tests and multiple hypotheses.

This article is a didactic presentation.

Principal Findings

It is not rare to find mistaken inferences in health services research because of inattention to appropriate hypothesis generation and multiple hypotheses testing. Guidelines are presented to help researchers identify composite hypotheses and set significance levels to account for multiple tests.

Conclusions

It is important for the quality of scholarship that inferences are valid: properly identifying composite hypotheses and accounting for multiple tests provides some assurance in this regard.

Recent issues of Health Services Research ( HSR ), the Journal of Health Economics , and Medical Care each contain articles that lack attention to the requirements of multiple hypotheses. The problems with multiple hypotheses are well known and often addressed in textbooks on research methods under the topics of joint tests (e.g., Greene 2003 ; Kennedy 2003 ) and significance level adjustment (e.g., Kleinbaum et al. 1998 ; Rothman and Greenland 1998 ; Portney and Watkins 2000 ; Myers and Well 2003 ; Stock and Watson 2003 ); yet, a look at applied journals in health services research quickly reveals that attention to the issue is not universal.

This paper has two goals: to remind researchers of issues regarding multiple hypotheses and to provide a few helpful guidelines. I first discuss when to combine hypotheses into a composite for a joint test; I then discuss the adjustment of test criterion for sets of hypotheses. Although often treated in statistics as two solutions to the same problem ( Johnson and Wichern 1992 ), here I treat them as separate tasks with distinct motivations.

In this paper I focus on Neyman–Pearson testing using Fisher's p -value as the interpretational quantity. Classically, a test compares an observed value of a statistic with a specified region of the statistic's range; if the value falls in the region, the data are considered not likely to have been generated given the hypothesis is true, and the hypothesis is rejected. However, it is common practice to instead compare a p -value to a significance level, rejecting the hypothesis if the p -value is smaller than the significance level. Because most tests are based on tail areas of distributions, this is a distinction without a difference for the purpose of this paper, and so I will use the p -value and significance-level terms in this discussion.

Of greater import is the requirement that hypotheses are stated a priori. A test is based on the prior assertion that if a given hypothesis is true, the data generating process will produce a value of the selected statistic that falls into the rejection region with probability equal to the corresponding significance level, which typically corresponds to a p -value smaller than the significance level. Setting hypotheses a priori is important in order to avoid a combinatorial explosion of error. For example, in a multiple regression model the a posteriori interpretation of regression coefficients in the absence of prior hypotheses does not account for the fact that the pattern of coefficients may be generated by chance. The important distinction is between the a priori hypothesis “the coefficient estimates for these particular variables in the data will be significant” and the a posteriori observation that “the coefficient estimates for these particular variables are significant.” In the first case, even if variables other than those identified in the hypothesis do not have statistically significant coefficients, the hypothesis is rejected nonetheless. In the second case, the observation applies to any set of variables that happen to have “statistically significant” coefficients. Hence, it is the probability that any set of variables have resultant “significant” statistics that drives the a posteriori case. As the investigator will interpret any number of significant coefficients that happen to result, the probability of significant results, given that no relationships actually exist, is the probability of getting any pattern of significance across the set of explanatory variables. This is different from a specific a priori case in which the pattern is preestablished by the explicit hypotheses. See the literatures on False Discovery Rate (e.g., Benjamini and Hochberg 1995 ; Benjamini and Liu 1999 ; Yekutieli and Benjamini 1999 ; Kwong, Holland, and Cheung 2002 ; Sarkar 2004 ; Ghosh, Chen, and Raghunathan 2005 ) and Empirical Bayes ( Efron et al. 2001 ; Cox and Wong 2004 ) for methods appropriate for a posteriori investigation.

WHEN TO COMBINE HYPOTHESES

What is achieved by testing an individual a priori hypothesis in the presence of multiple hypotheses? The answer to this question provides guidance for determining when a composite hypothesis (i.e., a composition of single hypotheses) is warranted. The significance level used for an individual test is the marginal probability of falsely rejecting the hypothesis: the probability of falsely rejecting the hypothesis regardless of whether the remaining hypotheses are rejected (see online-only appendix for details). The implied indifference to the status of the remaining hypotheses, however, is indefensible if the conclusions require a specific result from other hypotheses. This point underlies a guiding principle:

Guideline 1 : A joint test of a composite hypothesis ought to be used if an inference or conclusion requires multiple hypotheses to be simultaneously true .

The guideline is motivated by the logic of the inference or conclusion and is independent of significance levels. Examples from the literature can prove helpful in understanding the application of this guideline. Because it is unnecessarily pejorative to reference specific studies, the following discussion will only identify the nature of the problem in selected articles but not the articles themselves (the editors of HSR and the reviewers of this paper were provided the explicit references), but the general character of the examples ought to be familiar to most researchers.

Example 1—Polynomials

Two recent articles in the Journal of Health Economics each regressed a dependent variable on, among other variables, a second order polynomial—a practice used to capture nonlinear relationships. The null hypothesis for each coefficient of the polynomial was rejected according to its individual t -statistic. It was concluded that the explanatory variable had a parabolic relationship with the dependent variable, suggesting the authors rejected the hypotheses that both coefficients were simultaneously zero: the joint hypothesis regarding both coefficients is the relevant one. This is different from a researcher testing second-order nonlinearity (as opposed to testing the parabolic shape); in this case an individual test of the coefficient on the second-order term (i.e., the coefficient on the squared variable) is appropriate because the value of the first order term is not influential in this judgment of nonlinearity.

Example 2—Set of Dummy Variables

A recent article in Medical Care categorized a count variable into three size-groups and used a corresponding set of dummy variables to represent the two largest (the smallest group being the reference category); based on the individual significance of the two dummy variables they rejected the hypothesis that both coefficients were zero and concluded that the dependent variables was related to being larger on the underlying concept. In this conclusion, they collapsed two categories into a single statement about being larger on the underlying variable. Yet, if the authors meant that both categories are larger than the reference group, then it is a test of both coefficients being simultaneously zero that is relevant. A similar example using dummy variables is if we have an a priori hypothesis that the utilization of emergency services is not greater for blacks than whites, and another a priori hypothesis stating that utilization is not greater for Native Americans than whites. We may be justified in testing each coefficient if our interest in each minority group is independent of the other. However, a claim that “blacks and Native Americans both do not differ from whites in their utilization” makes sense only if both coefficients are simultaneously zero. Again, a joint test is indicated.

Example 3—Combined Effects across Independent Variables

Recent articles in HSR and the Journal of Health Economics , developed and tested a priori hypotheses regarding individual model parameters. So far, so good; but it was then inferred that the expected value of the dependent variable would differ between groups defined by different profiles of the explanatory variables. Here again the conclusion requires rejecting that the coefficients are simultaneously zero. For example, suppose we reject the hypothesis that age does not differentiate health care utilization and we reject the hypothesis that wealth does not differentiate health care utilization. These individual hypothesis tests do not warrant claims regarding wealthy elderly, poor youth, or other combinations. The coefficients for the age and wealth variables must both be nonzero, if such claims are to be made.

Example 4—Combined Effects across Dependent Variables

Recent articles in Medical Care , HSR , and the Journal of Health Economics each included analyses in which the same set independent variables were regressed on a number of dependent variables. Individual independent variables were considered regarding their influence across the various dependent variables. If an independent variable is considered to be simultaneously related to a number of dependent variables, then a joint test of a composite hypothesis is warranted. For example, suppose I wish to test a proposition that after controlling for age, health care utilization does not differ by sex. Suppose I use a two-part model (one part models the probability of any utilization, the other part models positive utilization given some utilization). 1 In this case I have two dependent variables (an indicator of any utilization and another variable measuring how much utilization gives positive utilization). If my proposition is correct then the coefficients on sex across both models should be simultaneously zero: a joint test is appropriate. If instead I test the two sex coefficients separately, I will implicitly be testing the hypotheses that (1) sex does not differentiate any utilization whether or not it differentiates positive utilization and (2) sex does not differentiate positive utilization whether or not it differentiates any utilization , which statistically does not address the original proposition. One might suppose if the dependent variables were conditionally independent from each other the joint test would provide similar results as the two individual hypotheses, not so. The type 1 error rate when using the individual tests is too large, unless the corresponding significance levels are divided by the number of hypotheses (see the section on adjusting for multiple tests below), in which case this type of adjustment is sufficient for independent tests.

Alternatively, suppose I wish to consider the effects of using nurse educators regarding diabetes care on health outcomes (e.g., A1c levels) and on patients' satisfaction with their health care organization, but my interest in these effects are independent of each other. In this case I am interested in two separate hypotheses, say for example (1) there is no effect on health outcomes regardless of the effect on satisfaction and (2) there is no effect on satisfaction regardless of any effect on outcomes. So long as I do not interpret these as a test that both effects are simultaneously zero, I can legitimately consider each hypothesis separately. But if each individual test does not reject the null, I should not infer that both effects are zero in the population (even with appropriate power) as this would require a joint test.

The preceding examples are in terms of individual model parameters. Guideline 1, however, applies to any set of hypotheses regardless of their complexity. In general, if a researcher desires to test a theory with multiple implications that must simultaneously hold for the theory to survive the test, then the failure of a single implication (as an independent hypothesis) defeats the theory. A joint hypothesis test is indicated.

The following guideline presents another heuristic to distinguish the need for joint versus separate tests.

Guideline 2 : If a conclusion would follow from a single hypothesis fully developed, tested, and reported in isolation from other hypotheses, then a single hypothesis test is warranted .

Guideline 2 asks whether a paper written about a given inference or conclusion would be coherent if based solely on the result of a single hypothesis. If so, then a single hypothesis test is warranted; if not, then consideration should be given to the possibility of a composite hypothesis. One could not support a claim that wealthy elderly use more services than poor youth, based solely on the hypothesis relating wealth and utilization, information regarding age is required.

Unfortunately, joint tests have a limitation that must be kept in mind, particularly when the hypothesis being tested is not the hypothesis of interest (which is often the case with null hypotheses). Rejecting a joint test of a composite hypothesis does not tell us which specific alternative case is warranted. Remember that a joint test of N hypotheses has 2 N −1 possible alternatives (in terms of the patterns of possible true and false hypotheses); for example, a joint test of two hypotheses (say, h 1 and h 2 ) has three possible alternatives ( h 1 true and h 2 false; h 1 false and h 2 true; and both h 1 and h 2 false); a joint test of five hypotheses has 31 possible alternatives. If your interest is in a specific alternative (e.g., all hypotheses are false, which is common and is the case in many of the examples discussed above), the rejection of a joint test does not provide unique support.

To answer the question of why the joint hypothesis was rejected, it can be useful to switch from the testing paradigm to a classic p -value paradigm by inspecting the relative “level of evidence” the data provides regarding each alternative case. Here p -values are used to suggest individual components of the composite hypothesis that are relatively not well supported by the data, providing a starting point for further theory development. In this exercise, the individual p -values of the component hypotheses are compared relative to each other; they are not interpreted in terms of significance. For example, if a two component composite null hypothesis is rejected but the individual p -values are .45 and .15, the “nonsignificance” of the p -values is irrelevant, it is the observation that one of the p -values is greatly smaller than others that provides a hint regarding why the joint hypothesis was rejected. This is theory and model building, not testing; hence, analyzing the joint test by inspecting the marginal p -values associated with its individual components is warranted as a useful heuristic—but admittedly not very satisfactory.

Alternatively, because identifying reasons for failure of a joint hypothesis is an a posteriori exercise, one could apply the methods of False Discovery Rate ( Benjamini and Hochberg 1995 ; Benjamini and Liu 1999 ; Yekutieli and Benjamini 1999 ; Kwong, Holland, and Cheung 2002 ; Sarkar 2004 ; Ghosh, Chen, and Raghunathan 2005 ) or Empirical Bayes Factors ( Efron et al. 2001 ; Cox and Wong 2004 ) to identify reasons for failure of the joint test (i.e., the individual hypotheses that are more likely to be false).

WHEN TO ADJUST SIGNIFICANCE LEVELS

In this section I use the phrase “significance level” to mean the criterion used in a test (commonly termed the “ α level”); I use the phrase “probability of false rejection,” denoted by pfr , to refer to the probability of falsely rejecting one or more hypotheses. A significance level is an operational part of a test (denoting the probability associated with the test's rejection region) whereas a probability of false rejection is a theoretical result of a test or grouping of tests. I use the modifier “acceptable” in conjunction with pfr to mean a probability of false rejection deemed the largest tolerable risk. I use the modifier “implied” in conjunction with pfr to mean the probability of false rejection resulting from the application of a test or group of tests. An acceptable pfr is subjective and set by the researcher, whereas an implied pfr is objective and calculated by the researcher. Suppose I wish to test three hypotheses, and I consider a 0.1 or less probability of falsely rejecting at least one of the hypotheses as acceptable across the tests; the acceptable pfr is 0.1. If I set the significance level for each test to 0.05 (thereby determining to reject a hypothesis if the p -value of its corresponding statistic is less than .05), the probability of false rejection is 1–(1 − 0.05) 3 =0.143; this is the implied pfr of the analysis associated with the hypothesis testing strategy. In this case, my strategy has an implied pfr value (0.143) that exceeds my acceptable pfr value (0.1); by this accounting, my strategy is unacceptable in terms of the risk of falsely rejecting hypotheses.

The preceding section on joint hypothesis tests presents guidance for identifying appropriate individual and composite hypotheses. Once a set of hypotheses is identified for testing, significance levels for each test must be set; or more generally, the rejection regions of the statistic must be selected. This task requires setting acceptable pfr 's; that is determining the acceptable risk of rejecting a true hypothesis.

A pfr can be associated with any group of tests. Typically no more than three levels are considered: individual hypotheses, mutually exclusive families of hypotheses, and the full analysis-wide set of hypotheses. Although common, it is not required to use the same acceptable pfr for each test. Some hypotheses may have stronger prior evidence or different goals than others, warranting different test-specific acceptable pfr 's. A family of hypotheses is a subset of the hypotheses in the analysis. They are grouped as a family explicitly because the researcher wishes to control the probability of false rejection among those particular hypotheses. For example, a researcher may be investigating two specific health outcomes and have a set of hypotheses for each; the hypotheses associated with each outcome may be considered a family, and the researcher may desire that the pfr for each family be constrained to some level. An acceptable analysis-wide pfr reflects the researcher's willingness to argue their study remains useful in the face of criticisms such as “Given your hypotheses are correct, the probability of reporting one or more false rejections is P ” or “Given your hypotheses are correct, the expected number of false rejections is N .”

The usefulness of setting pfr 's depends on one's perspective. From one view, we might contend that the information content of investigating 10 hypotheses should not change depending on whether we pursue a single study comprising all 10 hypotheses or we pursue 10 studies each containing one of the hypotheses; yet if we apply an analysis-wide pfr to the study with 10 hypotheses, we expect to falsely reject fewer hypotheses than we expect if we tested each hypothesis in a separate study. 2 If the hypotheses are independent such that the 10 independent repetitions of the data generating process do not in themselves accrue a benefit, there is merit to this observation and we might suppose that an analysis-wide pfr is no more warranted than a program-wide pfr (i.e., across multiple studies).

Our judgment might change, however, if we take a different view of the problem. Suppose I designed a study comprising 10 hypotheses that has an implied pfr corresponding to an expected false rejection of 9 of the 10 hypotheses. Should I pursue this study of 10 hypotheses for which I expect to falsely reject 90 percent of them if my theory is correct? Moreover, is it likely the study would be funded? I suggest the answers are both no. What if instead we expect 80 percent false rejections, or 40 percent, or 10 percent? The question naturally arises, what implied pfr is sufficient to warrant pursuit of the study? To answer that question is to provide an acceptable pfr . Once an acceptable pfr is established it seems prudent to check whether the design can achieve it, and if it cannot, to make adjustments. Clearly, this motivation applies to any family of hypotheses within a study as well. From this perspective, the use of pfr 's in the design of a single study is warranted.

This is not to suggest that researchers ought to always concern themselves with analysis-wide pfr 's in their most expansive sense; only that such considerations can be warranted. Research is often more complex than the preceding discussion implies. For example, it is common practice to report a table of descriptive statistics and nuisance parameters (e.g., parameters on control variables) as background along side the core hypotheses of a study. A researcher may legitimately decide that controlling the Type 1 error across these statistics is unimportant and focus on an acceptable pfr only for the family of core hypotheses. In this case, however, a more elegant approach is to provide interval estimates for the descriptive statistics and nuisance parameters without the pretense of “testing,” particularly as a priori hypotheses about background descriptive statistics are not often developed, thereby precluding them from the present consideration.

In setting acceptable pfr 's, the researcher should keep in mind that the probability of mistakenly rejecting hypotheses increases with the number of hypothesis tests. For example, if a researcher has settled on 10 tests for their analysis, the probability of mistakenly rejecting one or more of the 10 hypotheses at a significance level of 0.05 is approximately 0.4. Is it acceptable to engage a set of tests when the probability of falsely rejecting one or more of them is 40 percent? The answer is a matter of judgment depending on the level of risk a researcher, reviewer, editor, or reader is willing to take regarding the reported findings.

Being arbitrary, the designation of an acceptable pfr is not likely to garner universal support. Authors in some research fields, with the goal of minimizing false reports in the literature, recommend adjusting the significance levels for tests to restrict the overall analysis-wide error rate to 0.05 ( Maxwell and Delaney 2000 ). However, when there are numerous tests, this rule can dramatically shrink the tests' significance levels and either require a considerably larger sample or severely diminish power. A recent article in Health Services Research reported results of 74 tests using a significance level of 0.05: there is a 98 percent chance of one or more false rejections across the analysis, a 10 percent chance of six or more, and a 5 percent chance of between seven and eight or more. The expected number of false rejections in the analysis is approximately four. The analysis-wide pfr can be restricted to less than 0.05 as Maxwell and Delaney (2000) suggest by setting the significance levels to 0.00068 (the process of adjustment is explained below). This recommended significance level is two orders of magnitude smaller than 0.05. If an analysis-wide pfr of 0.98 (associated with the significance levels of 0.05) is deemed unacceptably high and an analysis-wide pfr of 0.05 (associated with the significance levels of 0.00068) is deemed too strict, the researchers may settle on a reasoned intermediate value. For example, to obtain a just-better-than-even odds against a false rejection across the full analysis of 74 tests (e.g., setting the pfr to 0.499), the significance levels would have been adjusted to 0.0093. Alternatively, the researchers might desire to control the expected number of false rejections across the full analysis, which can be calculated as the sum of the individual significance levels. For example, setting significance levels to 0.01351 provides an expectation of one false rejection among the 74 tests rather than the expected four associated with the original 0.05 significance levels. The adjusted significance levels in this example are less than the original significance level of 0.05, and they vary in their magnitude (and therefore power to discern effects) depending on their underlying reasoning.

Whatever rational determines the acceptable pfr' s for the analysis, the significance levels must be set to assure these pfr 's are not exceeded at any level for which they are set. Guideline 3 presents one procedure to accomplish this task.

Guideline 3 : A five-step procedure for setting significance levels . Step 1 . Determine the set of hypotheses to be tested (applying Guidelines 1 and 2 to identify any joint hypotheses), assign an acceptable pfr to each hypothesis, and set the significance levels equal to these pfr' s. Step 2 . Determine families of hypotheses, if any, within which the probability of false rejection is to be controlled, and assign each family an acceptable pfr . Step 3 . Assign an acceptable analysis-wide pfr if desired. Step 4 . For each family, compare the implied family pfr with the acceptable family pfr . If the implied pfr is greater than the acceptable pfr , adjust the significance levels (see the following discussion on adjustment) so the implied pfr based on the adjusted significance levels is no greater than the acceptable pfr . Step 5 . If an analysis-wide acceptable pfr is set, calculate the analysis-wide pfr implied by the significance levels from Step 4 . If the implied pfr exceeds the acceptable analysis-wide pfr , then adjust the test-specific significance levels such that the implied pfr does not exceed the acceptable pfr .

By this procedure the resulting significance levels will assure that the acceptable pfr at each level (hypothesis, family, and analysis) is not exceeded. The resulting significance levels are governed by the strictest pfr 's. Ignoring a level is implicitly setting its pfr to the sum of the associated significance levels.

Steps 4 and 5 of Guideline 3 require the adjustment of significance levels. One approach to making such adjustments is the ad hoc reassessment of the acceptable hypothesis-specific pfr 's such that they are smaller. By this approach, the researcher reconsiders her acceptable pfr for each hypothesis and recalculates the comparisons with the higher level pfr 's. Of course, the outside observer could rightfully wonder how well reasoned these decisions were to begin with if they are so conveniently modified. A common alternative is to leave all pfr 's as they were originally designated and use a Bonferroni-type adjustment (or other available adjustment method). To preserve the relative importance indicated by the relative magnitudes among the acceptable pfr 's, a researcher transforms the current significance levels into normalized weights and sets the new significance levels as the weight multiplied by the higher-level pfr . For example, if a family of three hypotheses has significance levels of 0.05, 0.025, and 0.01, the implied family-level pfr is 0.083. If the acceptable family pfr is 0.05, then the implied pfr is greater than the acceptable pfr and adjustment is indicated. Weights are constructed from the significance levels as w 1 = 0.05/(0.05+0.025+0.01), w 2 = 0.025/(0.05+0.025+0.01), and w 3 = 0.01/(0.05+0.025+0.01). The adjusted significance levels are then calculated as 0.029= w 1 × 0.05, 0.015= w 2 × 0.05, and 0.006= w 3 × 0.05, which have an implied family-level pfr = 0.049 meeting our requirement that it not exceed the acceptable pfr of 0.05.

In the preceding example the adjusted significance levels implied a pfr that was less than the acceptable pfr (i.e., 0.049<0.05). A Bonferroni-type adjustment assures that the implied pfr is less than or equal to the acceptable pfr , consequently it is conservative and may unnecessarily diminish power by setting overly strict significance levels. An adjustment with better power, while not exceeding the acceptable pfr , can be attained by inflating the Bonferroni adjusted significance levels by a constant factor until the implied pfr is equal to the acceptable pfr . Although perhaps trivial in the present example, inflating each Bonferroni-adjusted significance level by a factor of 1.014 yields significance levels with an implied family-level pfr of 0.05, exactly that of the acceptable pfr .

Adjusting significance levels by reweighting produces the distribution of a higher-level pfr according to the relative importance implied by the initial significance levels. The final adjusted significance levels are a redistribution of the strictest pfr ; therefore, adjusted significance levels no longer represent the acceptable hypothesis-level pfr 's. However, the implied hypothesis-level pfr 's will be less than the acceptable hypothesis-level pfr 's thereby meeting the requirement that the probability of false rejection is satisfactory at all levels. Because, when the inflation factor of the preceding paragraph is not used, the adjusted significance levels sum to the strictest pfr , this pfr can be interpreted as the expected number of false rejections. If the inflation factor is used, then of course the adjusted significance levels will be larger and the expected number of rejections will be larger than the pfr .

Sample size and power calculations should be based on the final adjusted significance levels. If the corresponding sample size is infeasible or the power unacceptable, then reconsideration of the study design, sampling strategy or estimators is warranted. If no additional efficiency is forthcoming, then a re-evaluation of the project goals may save the proposed study. For example, it may be that changing the study goals from guiding policy decisions to furthering theory will warrant greater leniency in the acceptable pfr 's.

CONCLUSIONS

This paper is not intended as a tutorial on the statistical procedures for joint tests and significance level adjustment: there is considerable information in the statistics and research methods literatures regarding these details. F -statistics and χ 2 statistics are commonly available for testing sets of hypotheses expressed as functions of jointly estimated model parameters, including both single and multiple equation models (see, e.g., Greene 2003 ; Kennedy 2003 ). More generally, there are joint tests available for sets of hypotheses generated from separately estimated models (see the routine Seemingly Unrelated Estimation, based on White's sandwich variance estimator, in STATA 2003 ); for example, hypotheses comparing functions of parameters from a linear regression, a logistic regression, and a Poisson model can be jointly tested. If these tests are not applicable, tests based on bootstrapped data sets can often be successfully used ( Efron and Tibshirani 1993 ). Regarding basic descriptions of Bonferroni adjustment, see Johnson and Wichern (1992) , Harris (2001) , and Portney and Watkins (2000) , among others.

The preceding section on when to adjust significance levels implies such adjustments are warranted. This is not a universally accepted view; indeed, the use of adjustments for multiple tests has been the focus of considerable debate (see e.g., Rothman 1990 ; Saville 1990 ; 1995 , 1998 ; Goodman 1998 ; Thompson 1998a , b ). When reviewing this debate, or considering the merit of multiple testing adjustment, the distinction between a priori hypotheses and a posteriori observations is important, a distinction carefully drawn by Thompson (1998a) .

Although didactic in nature, I do not presume the ideas presented here are new to the majority of health services researchers; however, reading the journals in our field suggests that we may sometimes forget to apply what we know. The two goals of this paper were to remind researchers to consider their approach to multiple hypotheses and to provide some guidelines. Whether researchers use these guidelines or others, it is important for the quality of scholarship that we draw valid inferences from the evidence we consider: properly identifying composite hypotheses and accounting for multiple tests provides some assurance in this regard.

Supplementary Material

The following supplementary material for this article is available online:

When to Combine Hypothesis and Adjust for Multiple Tests.

1 Appreciation to an anonymous reviewer for suggesting the two-part model as an example.

2 Appreciation to an anonymous reviewer for pointing out the 10 hypotheses/10 studies example.

  • Benjamini Y, Hochberg Y. ““Controlling the False Discovery Rate: A Practical and Powerful Approach to Multiple Testing”.” Journal of the Royal Statistical Society B. 1995; 57 (1):289–300. [ Google Scholar ]
  • Benjamini Y, Liu W. ““A Step-Down Multiple Hypotheses Testing Procedure That Controls the False Discovery Rate under Independence”.” Journal of Statistical Planning and Inference. 1999; 82 (1-2):163–70. [ Google Scholar ]
  • Cox DR, Wong MY. ““A Simple Procedure for the Selection of Significant Effects”.” Journal of the Royal Statistical Society B. 2004; 66 (2):395–400. [ Google Scholar ]
  • Efron B, Tibshirani RJ. An Introduction to the Bootstrap. New York: Chapman & Hall; 1993. [ Google Scholar ]
  • Efron B, Tibshirani R, Storey JD, Tusher V. ““Empirical Bayes Analysis of a Microarray Experiment”.” Journal of the American Statistical Association. 2001; 96 (456):1151–60. [ Google Scholar ]
  • Ghosh D, Chen W, Raghunathan T. ““The False Discovery Rate: A Variable Selection Perspective”.” Journal of Statistical Planning and Inference. 2005 in press. Available tat //www.sciencedirect.com . [ Google Scholar ]
  • Goodman SN. ““Multiple Comparisons, Explained”.” American Journal of Epidemiology. 1998; 147 (9):807–12. [ PubMed ] [ Google Scholar ]
  • Greene WH. Econometric Analysis. Upper Saddle River, NJ: Prentice Hall; 2003. [ Google Scholar ]
  • Harris RJ. A Primer of Multivariate Statistics. Mahwah, NJ: Lawrence Erlbaum Associates; 2001. [ Google Scholar ]
  • Johnson RA, Wichern DW. Applied Multivariate Statistical Analysis. Englewood Cliffs, NJ: Prentice-Hall; 1992. [ Google Scholar ]
  • Kennedy P. A Guide to Econometrics. Cambridge, MA: MIT Press; 2003. [ Google Scholar ]
  • Kleinbaum DG, Kupper LL, Muller KE, Nizan A. Applied Regression Analysis and Other Multivariate Methods. Boston: Duxbury Press; 1998. [ Google Scholar ]
  • Kwong KS, Holland B, Cheung SH. ““A Modified Benjamini-Hochberg Multiple Comparisons Procedure for Controlling the False Discovery Rate”.” Journal of Statistical Planning and Inference. 2002; 104 (2):351–62. [ Google Scholar ]
  • Maxwell SE, Delaney HD. Designing Experiments and Analyzing Data. A Model Comparison Perspective. Mahwah, NJ: Lawrence Erlbaum Associates; 2000. [ Google Scholar ]
  • Myers JL, Well AD. Research Design and Statistical Analysis. Mahwah, NJ: Lawrence Erlbaum Associates; 2003. [ Google Scholar ]
  • Portney LG, Watkins MP. Foundations of Clinical Research: Applications to Practice. Upper Saddle River, NJ: Prentice Hall Health; 2000. [ Google Scholar ]
  • Rothman KJ. ““No Adjustments Are Needed for Multiple Comparisons”.” Epidemiology. 1990; 1 (1):43–6. [ PubMed ] [ Google Scholar ]
  • Rothman KJ, Greenland S. Modern Epidemiology. New York: Lippincott Williams & Wilkins; 1998. [ Google Scholar ]
  • Sarkar SK. ““FDR-Controlling Stepwise Procedures and Their False Negatives Rates”.” Journal of Statistical Planning and Inference. 2004; 125 (1-2):119–37. [ Google Scholar ]
  • Saville DJ. ““Multiple Comparison Procedures: The Practical Solution”.” American Statistician. 1990; 44 (2):174–80. [ Google Scholar ]
  • Savitz DA, Olshan AF. ““Multiple Comparisons and Related Issues in the Interpretation of Epidemiologic Data”.” American Journal of Epidemiology. 1995; 142 :904–8. [ PubMed ] [ Google Scholar ]
  • Savitz DA, Olshan AF. ““Describing Data Requires No Adjustment for Multiple Comparisons: A Reply from Savitz and Olshan”.” American Journal of Epidemiology. 1998; 147 (9):813–4. [ PubMed ] [ Google Scholar ]
  • STATA. In STATA 8 Reference S-Z. College Station, TX: Stata Press; 2003. ““Seemingly Unrelated Estimation—Suest”.” pp. 126–47. [ Google Scholar ]
  • Stock JH, Watson MW. Introduction to Econometrics. Boston: Addison Wesley; 2003. [ Google Scholar ]
  • Thompson JR. ““Invited Commentary Re: Multiple Comparisons and Related Issues in the Interpretation of Epidemiologic Data”.” American Journal of Epidemiology. 1998a; 147 (9):801–6. [ PubMed ] [ Google Scholar ]
  • Thompson JR. ““A Response to “Describing Data Requires No Adjustment for Multiple Comparisons””.” American Journal of Epidemiology. 1998b; 147 (9):915. [ PubMed ] [ Google Scholar ]
  • Yekutieli D, Benjamini Y. ““Resampling-Based False Discovery Rate Controlling Multiple Test Procedures for Correlated Test Statistics”.” Journal of Statistical Planning and Inference. 1999; 82 (1-2):171–96. [ Google Scholar ]

Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

  • View all journals
  • Explore content
  • About the journal
  • Publish with us
  • Sign up for alerts
  • Published: 23 August 2024

Ventral attention network connectivity is linked to cortical maturation and cognitive ability in childhood

  • Hao-Ming Dong   ORCID: orcid.org/0000-0002-0176-2969 1 ,
  • Xi-Han Zhang   ORCID: orcid.org/0000-0003-2630-0063 1 ,
  • Loïc Labache   ORCID: orcid.org/0000-0002-5733-0743 1 ,
  • Shaoshi Zhang 2 ,
  • Leon Qi Rong Ooi   ORCID: orcid.org/0000-0002-3546-4580 2 ,
  • B. T. Thomas Yeo   ORCID: orcid.org/0000-0002-0119-3276 2 , 3 , 4 ,
  • Daniel S. Margulies   ORCID: orcid.org/0000-0002-8880-9204 5 ,
  • Avram J. Holmes   ORCID: orcid.org/0000-0001-6583-803X 6   na1 &
  • Xi-Nian Zuo   ORCID: orcid.org/0000-0001-9110-585X 7 , 8 , 9   na1  

Nature Neuroscience ( 2024 ) Cite this article

10 Altmetric

Metrics details

  • Cognitive neuroscience

The human brain experiences functional changes through childhood and adolescence, shifting from an organizational framework anchored within sensorimotor and visual regions into one that is balanced through interactions with later-maturing aspects of association cortex. Here, we link this profile of functional reorganization to the development of ventral attention network connectivity across independent datasets. We demonstrate that maturational changes in cortical organization link preferentially to within-network connectivity and heightened degree centrality in the ventral attention network, whereas connectivity within network-linked vertices predicts cognitive ability. This connectivity is associated closely with maturational refinement of cortical organization. Children with low ventral attention network connectivity exhibit adolescent-like topographical profiles, suggesting that attentional systems may be relevant in understanding how brain functions are refined across development. These data suggest a role for attention networks in supporting age-dependent shifts in cortical organization and cognition across childhood and adolescence.

This is a preview of subscription content, access via your institution

Access options

Access Nature and 54 other Nature Portfolio journals

Get Nature+, our best-value online-access subscription

24,99 € / 30 days

cancel any time

Subscribe to this journal

Receive 12 print issues and online access

195,33 € per year

only 16,28 € per issue

Buy this article

  • Purchase on SpringerLink
  • Instant access to full article PDF

Prices may be subject to local taxes which are calculated during checkout

what is composite hypothesis

Similar content being viewed by others

what is composite hypothesis

The development of structural covariance networks during the transition from childhood to adolescence

what is composite hypothesis

Functional connectivity development along the sensorimotor-association axis enhances the cortical hierarchy

what is composite hypothesis

A shifting role of thalamocortical connectivity in the emergence of cortical functional organization

Data availability.

Data from the CCNP dataset used here are available at CCNP—Lifespan Brain-Mind Development Data Community at Science Data Bank ( https://ccnp.scidb.cn/en ) including both anonymized neuroimaging data ( https://doi.org/10.57760/sciencedb.07860 ) and unthresholded whole-brain connectivity matrices grouped by relevant ages (children and adolescents) ( https://doi.org/10.11922/sciencedb.00886 ). The raw CCNP data are available from the website upon reasonable request. The ABCD data used in this report came from the Annual Release v.2.0 ( https://doi.org/10.15154/1503209 ) of the ABCD BIDS Community Collection (ABCC; NDA Collection 3165). Source data are provided with this paper.

Code availability

Code is available via GitHub: (1) preprocessing CCNP datasets ( https://github.com/zuoxinian/CCS ); (2) preprocessing ABCD datasets ( https://github.com/ThomasYeoLab/ABCD_scripts ); (3) FC gradient analysis ( https://github.com/NeuroanatomyAndConnectivity/gradient_analysis ); and (4) Gradient maturation analysis ( https://github.com/HolmesLab/GradientMaturation ).

Casey, B. J., Heller, A. S., Gee, D. G. & Cohen, A. O. Development of the emotional brain. Neurosci. Lett. 693 , 29–34 (2019).

Article   CAS   PubMed   Google Scholar  

Casey, B. J., Getz, S. & Galvan, A. The adolescent brain. Dev. Rev. 28 , 62–77 (2008).

Article   CAS   PubMed   PubMed Central   Google Scholar  

Luna, B. et al. Maturation of widely distributed brain function subserves cognitive development. Neuroimage 13 , 786–793 (2001).

Kang, H. J. et al. Spatio-temporal transcriptome of the human brain. Nature 478 , 483–489 (2011).

Huttenlocher, P. R. & Dabholkar, A. S. Regional differences in synaptogenesis in human cerebral cortex. J. Comp. Neurol. 387 , 167–178 (1997).

Paquola, C. et al. Shifts in myeloarchitecture characterise adolescent development of cortical gradients. eLife 8 , e50482 (2019).

Zilles, K., Palomero-Gallagher, N. & Amunts, K. Development of cortical folding during evolution and ontogeny. Trends Neurosci. 36 , 275–284 (2013).

Reardon, P. K. et al. Normative brain size variation and brain shape diversity in humans. Science 360 , 1222–1227 (2018).

Bethlehem, R. A. I. et al. Brain charts for the human lifespan. Nature 604 , 525–533 (2022).

Margulies, D. S. et al. Situating the default-mode network along a principal gradient of macroscale cortical organization. Proc. Natl Acad. Sci. USA 113 , 12574–12579 (2016).

Yeo, B. T. et al. The organization of the human cerebral cortex estimated by intrinsic functional connectivity. J. Neurophysiol. 106 , 1125–1165 (2011).

Article   PubMed   Google Scholar  

Gao, W. et al. Temporal and spatial evolution of brain network topology during the first two years of life. PLoS ONE 6 , e25278 (2011).

Fair, D. A. et al. Functional brain networks develop from a ‘Local to Distributed’ organization. PLoS Comput. Biol. 5 , e1000381 (2009).

Article   PubMed   PubMed Central   Google Scholar  

Dong, H. M., Margulies, D. S., Zuo, X. N. & Holmes, A. J. Shifting gradients of macroscale cortical organization mark the transition from childhood to adolescence. Proc. Natl Acad. Sci. USA 118 , e2024448118 (2021).

Somerville, L. H., Hare, T. & Casey, B. J. Frontostriatal maturation predicts cognitive control failure to appetitive cues in adolescents. J. Cogn. Neurosci. 23 , 2123–2134 (2011).

Tottenham, N. & Sheridan, M. A. A review of adversity, the amygdala and the hippocampus: a consideration of developmental timing. Front. Hum. Neurosci. 3 , 68 (2009).

PubMed   Google Scholar  

Sydnor, V. J. et al. Neurodevelopment of the association cortices: patterns, mechanisms, and implications for psychopathology. Neuron 109 , 2820–2846 (2021).

Fair, D. A. et al. Development of distinct control networks through segregation and integration. Proc. Natl Acad. Sci. USA 104 , 13507–13512 (2007).

Betzel, R. F. et al. Changes in structural and functional connectivity among resting-state networks across the human lifespan. Neuroimage 102 , 345–357 (2014).

Betzel, R. F. et al. Generative models of the human connectome. Neuroimage 124 , 1054–1064 (2016).

Zuo, X. N. et al. Human connectomics across the life span. Trends Cogn. Sci. 21 , 32–45 (2017).

Tooley, U. A., Bassett, D. S. & Mackey, A. P. Functional brain network community structure in childhood: Unfinished territories and fuzzy boundaries. Neuroimage 247 , 118843 (2022).

Mesulam, M. M. From sensation to cognition. Brain 121 , 1013–1052 (1998).

Sepulcre, J., Sabuncu, M. R., Yeo, T. B., Liu, H. & Johnson, K. A. Stepwise connectivity of the modal cortex reveals the multimodal organization of the human brain. J. Neurosci. 32 , 10649–10661 (2012).

Power, J. D. et al. Functional network organization of the human brain. Neuron 72 , 665–678 (2011).

Dosenbach, N. U. et al. Distinct brain networks for adaptive and stable task control in humans. Proc. Natl Acad. Sci. USA 104 , 11073–11078 (2007).

Seeley, W. W. et al. Dissociable intrinsic connectivity networks for salience processing and executive control. J. Neurosci. 27 , 2349–2356 (2007).

Dosenbach, N. U., Fair, D. A., Cohen, A. L., Schlaggar, B. L. & Petersen, S. E. A dual-networks architecture of top-down control. Trends Cogn. Sci. 12 , 99–105 (2008).

Dosenbach, N. U. et al. A core system for the implementation of task sets. Neuron 50 , 799–812 (2006).

Labache, L., Ge, T., Yeo, B. T. T. & Holmes, A. J. Language network lateralization is reflected throughout the macroscale functional organization of cortex. Nat. Commun. 14 , 3405 (2023).

Liu, S. et al. Chinese color nest project: an accelerated longitudinal brain-mind cohort. Dev. Cogn. Neurosci. 52 , 101020 (2021).

Alexander-Bloch, A. F. et al. On testing for spatial correspondence between maps of human brain structure and function. Neuroimage 178 , 540–551 (2018).

Schaefer, A. et al. Local-global parcellation of the human cerebral cortex from intrinsic functional connectivity MRI. Cereb. Cortex 28 , 3095–3114 (2018).

Gogtay, N. et al. Dynamic mapping of human cortical development during childhood through early adulthood. Proc. Natl Acad. Sci. USA 101 , 8174–8179 (2004).

Langs, G. et al. Identifying shared brain networks in individuals by decoupling functional and anatomical variability. Cereb. Cortex 26 , 4004–4014 (2016).

Coifman, R. R. & Lafon, S. Diffusion maps. Appl. Comput. Harmon. A 21 , 5–30 (2006).

Article   Google Scholar  

Fan, X. R. et al. A longitudinal resource for population neuroscience of school-age children and adolescents in China. Sci. Data 10 , 545 (2023).

Casey, B. J. et al. The Adolescent Brain Cognitive Development (ABCD) study: imaging acquisition across 21 sites. Dev. Cogn. Neurosci. 32 , 43–54 (2018).

Luciana, M. et al. Adolescent neurocognitive development and impacts of substance use: overview of the adolescent brain cognitive development (ABCD) baseline neurocognition battery. Dev. Cogn. Neurosci. 32 , 67–79 (2018).

Ricard, J. A. et al. Confronting racially exclusionary practices in the acquisition and analyses of neuroimaging data. Nat. Neurosci. 26 , 4–11 (2023).

Li, J. et al. Cross-ethnicity/race generalization failure of behavioral prediction from resting-state functional connectivity. Sci. Adv. 8 , eabj1812 (2022).

Fox, M. D. et al. The human brain is intrinsically organized into dynamic, anticorrelated functional networks. Proc. Natl Acad. Sci. USA 102 , 9673–9678 (2005).

Corbetta, M. & Shulman, G. L. Control of goal-directed and stimulus-driven attention in the brain. Nat. Rev. Neurosci. 3 , 201–215 (2002).

Gordon, E. M. et al. Precision functional mapping of individual human brains. Neuron 95 , 791–807 e797 (2017).

Sridharan, D., Levitin, D. J. & Menon, V. A critical role for the right fronto-insular cortex in switching between central-executive and default-mode networks. Proc. Natl Acad. Sci. USA 105 , 12569–12574 (2008).

Farrant, K. & Uddin, L. Q. Asymmetric development of dorsal and ventral attention networks in the human brain. Dev. Cogn. Neurosci. 12 , 165–174 (2015).

Casey, B. J., Giedd, J. N. & Thomas, K. M. Structural and functional brain development and its relation to cognitive development. Biol. Psychol. 54 , 241–257 (2000).

Molloy, M. F. et al. Effect of extremely preterm birth on adolescent brain network organization. Brain Connect. 13 , 394–409 (2023).

Huizinga, M., Dolan, C. V. & van der Molen, M. W. Age-related change in executive function: developmental trends and a latent variable analysis. Neuropsychologia 44 , 2017–2036 (2006).

Luna, B., Garver, K. E., Urban, T. A., Lazar, N. A. & Sweeney, J. A. Maturation of cognitive processes from late childhood to adulthood. Child Dev. 75 , 1357–1372 (2004).

Gordon, E. M. et al. A somato-cognitive action network alternates with effector regions in motor cortex. Nature 617 , 351–359 (2023).

Pfisterer, U. & Khodosevich, K. Neuronal survival in the brain: neuron type-specific mechanisms. Cell Death Dis. 8 , e2643 (2017).

Bullmore, E. & Sporns, O. The economy of brain network organization. Nat. Rev. Neurosci. 13 , 336–349 (2012).

Gee, D. G. et al. Early developmental emergence of human amygdala-prefrontal connectivity after maternal deprivation. Proc. Natl Acad. Sci. USA 110 , 15638–15643 (2013).

Dong, H. M. et al. Charting brain growth in tandem with brain templates at school age. Sci. Bull. 65 , 1924–1934 (2020).

Yang, N. et al. Chinese color nest project: growing up in China (in Chinese). Chin. Sci. Bull. 62 , 3008–3022 (2017).

Auchter, A. M. et al. A description of the ABCD organizational structure and communication framework. Dev Cogn Neurosci 32 , 8–15 (2018).

Clark, D. B. et al. Biomedical ethics and clinical oversight in multisite observational neuroimaging studies with children and adolescents: the ABCD experience. Dev. Cogn. Neurosci. 32 , 143–154 (2018).

Manjon, J. V. & Coupe, P. volBrain: an online MRI brain volumetry system. Front. Neuroinform. 10 , 30 (2016).

Xu, T. et al. A connectome computation system for discovery science of brain. Sci. Bull. 60 , 86–95 (2015).

Xing, X. X. et al. Connectome computation system: 2015–2021 updates. Sci. Bull. 67 , 448–451 (2022).

Friston, K. J. et al. Statistical parametric maps in functional imaging: a general linear approach. Hum. Brain Mapp. 2 , 189–210 (1994).

Jenkinson, M., Beckmann, C. F., Behrens, T. E., Woolrich, M. W. & Smith, S. M. FSL. Neuroimage 62 , 782–790 (2012).

Cox, R. W. AFNI: software for analysis and visualization of functional magnetic resonance neuroimages. Comput. Biomed. Res. 29 , 162–173 (1996).

Fischl, B. FreeSurfer. Neuroimage 62 , 774–781 (2012).

Pruim, R. H. R. et al. ICA-AROMA: a robust ICA-based strategy for removing motion artifacts from fMRI data. Neuroimage 112 , 267–277 (2015).

Hagler, D. J. Jr. et al. Image processing and analysis methods for the adolescent brain cognitive development study. Neuroimage 202 , 116091 (2019).

Chen, J. et al. Shared and unique brain network features predict cognitive, personality, and mental health scores in the ABCD study. Nat. Commun. 13 , 2217 (2022).

Download references

Acknowledgements

This work was supported by the STI 2030—the major projects of the Brain Science and Brain-Inspired Intelligence Technology (2021ZD0200500 to X.-N.Z.), the National Institute of Mental Health (grants R01MH120080 and R01MH123245 to A.J.H.), the Major Fund for International Collaboration of National Natural Science Foundation of China (81220108014 to X.-N.Z.) and the National Basic Science Data Center ‘Interdisciplinary Brain Database for In vivo Population Imaging’ (ID-BRAIN to X.-N.Z.). B.T.T.Y. is supported by the NUS Yong Loo Lin School of Medicine (NUHSRO/2020/124/TMR/LOA), the Singapore National Medical Research Council (NMRC) LCG (OFLCG19May-0035), NMRC CTG-IIT (CTGIIT23jan-0001), NMRC STaR (STaR20nov-0003), Singapore Ministry of Health (MOH) Centre Grant (CG21APR1009), the Temasek Foundation (TF2223-IMH-01) and the United States NIH (R01MH120080 and R01MH133334). Any opinions, findings and conclusions or recommendations expressed in this material are those of the authors and do not reflect the views of the Singapore NMRC, MOH or Temasek Foundation.

Author information

These authors contributed equally: Avram J. Holmes, Xi-Nian Zuo.

Authors and Affiliations

Department of Psychology, Yale University, New Haven, CT, USA

Hao-Ming Dong, Xi-Han Zhang & Loïc Labache

Centre for Sleep and Cognition and Centre for Translational Magnetic Resonance Research, Yong Loo Lin School of Medicine, Singapore, National University of Singapore, Singapore, Singapore

Shaoshi Zhang, Leon Qi Rong Ooi & B. T. Thomas Yeo

Department of Electrical and Computer Engineering, National University of Singapore, Singapore, Singapore

B. T. Thomas Yeo

N.1 Institute for Health and Institute for Digital Medicine, National University of Singapore, Singapore, Singapore

Centre National de la Recherche Scientifique, Frontlab, Institut du Cerveau et de la Moelle Epinière, Paris, France

Daniel S. Margulies

Department of Psychiatry, Brain Health Institute, Rutgers University, Piscataway, NJ, USA

Avram J. Holmes

State Key Laboratory of Cognitive Neuroscience and Learning, Beijing Normal University, Beijing, China

Xi-Nian Zuo

National Basic Science Data Center, Beijing, China

Developmental Population Neuroscience Research Center, IDG/McGovern Institute for Brain Research, Beijing Normal University, Beijing, China

You can also search for this author in PubMed   Google Scholar

Contributions

H.-M.D., A.J.H. and X.-N.Z. designed the research. A.J.H. and X.-N.Z. supervised the research. H.-M.D. and A.J.H. conducted analyses and made figures. X.-H.Z., L.L., S.Z., L.Q.R.O. and B.T.T.Y. conducted validation analyses based on the ABCD dataset. H.-M.D., A.J.H. and X.-N.Z. wrote the initial draft. X.-H.Z., L.L., S.Z., L.Q.R.O., B.T.T.Y. and D.S.M. edited the paper.

Corresponding authors

Correspondence to Hao-Ming Dong , Avram J. Holmes or Xi-Nian Zuo .

Ethics declarations

Competing interests.

The authors declare no competing interests.

Peer review

Peer review information.

Nature Neuroscience thanks Brenden Tervo-Clemmens and the other, anonymous, reviewer(s) for their contribution to the peer review of this work.

Additional information

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Extended data

Extended data fig. 1 the transition from unimodal to transmodal organization revealed by the increasing in percentiles of functional connectome..

The threshold of connectivity matrices was adjusted in children and adolescent group and then redrive the gradients. The results revealed a marked transition in functional connectivity strength from childhood to adolescence. With the 95% threshold retaining the strongest connections, a unimodal organization was evident in both children and adolescents as their primary gradients. However, as additional weaker connections are included in the functional connectome at a 90% threshold, the primary gradients diverged for children and adolescents, revealing a unimodal and transmodal organization, respectively. Yet, with an 85% threshold incorporating even more weaker connections, the primary gradients converged into a transmodal organization for both children and adolescents.

Source data

Extended data fig. 2 gradient maps in low ventral attention connectivity groups derived from longitudinal data..

A set of child participants (n=22) were identified from the low ventral attention group who were also subsequently scanned in their adolescence. Surface maps exhibit a stable adolescent-like gradient architecture in both childhood and adolescence. Their first gradient in childhood is highly correlated (r=0.9429, p<0.01, two-sided spin test) with the first gradient that in their adolescence. A consistent group profile that was also evident when considering their second gradients in both childhood and adolescence (r=0.9353, p<0.01, two-sided spin test).

Extended Data Fig. 3 Gradient maps in high ventral attention connectivity groups derived from longitudinal data.

A set of child participants (n=21) were identified from the high ventral attention group who were also subsequently scanned in their adolescence. Surface maps exhibit a developmentally normative pattern of gradient reversals from their childhood to adolescence. Their first gradient in childhood was highly correlated with their second gradient in adolescence (absolute r=0.9793, p<0.01, two-sided spin test), while their second gradient in childhood were highly correlated with the first gradient in their adolescence (absolute r=0.9748, p<0.01, two-sided spin test).

Extended Data Fig. 4 Gradient maps with functional networks dropped off separately.

Virtual lesion analyses were performed for all the networks respectively. It is revealed that in children group, the drop off of visual, somato/motor, ventral attention and frontoparietal networks generating transmodal organization in the first gradient, while the drop off of dorsal attention, limbic and default networks conserve the unimodal organization in the first gradient. Readers should interpret these maps with caution as functional networks each contain distinct numbers of vertices along the cortical sheet. Accordingly, the direct examination across the canonical networks, is likely biased by their relative sizes.

Supplementary information

Supplementary information.

Supplementary Discussion and Tables 1–17.

Reporting Summary

Source data figs. 1–5 and extended data figs. 1–4.

Degree centrality values, Euclidean values and network-level mean and standard gradient values after dropping off ventral attention network and the null distribution generated by the permutation test. Config file for plotting Chord diagram. Gradient values of High/Low ventral attention groups in the CCNP dataset. Gradient values of high/low ventral attention groups in the ABCD dataset. Gradient values of adolescents and children groups with different percentiles of functional connectome. Gradient values in low ventral attention connectivity groups derived from longitudinal data. Gradient values in high ventral attention connectivity groups derived from longitudinal data. Gradient maps with functional networks dropped off separately.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Cite this article.

Dong, HM., Zhang, XH., Labache, L. et al. Ventral attention network connectivity is linked to cortical maturation and cognitive ability in childhood. Nat Neurosci (2024). https://doi.org/10.1038/s41593-024-01736-x

Download citation

Received : 19 May 2023

Accepted : 18 July 2024

Published : 23 August 2024

DOI : https://doi.org/10.1038/s41593-024-01736-x

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

Quick links

  • Explore articles by subject
  • Guide to authors
  • Editorial policies

Sign up for the Nature Briefing newsletter — what matters in science, free to your inbox daily.

what is composite hypothesis

IMAGES

  1. Practice deriving detectors for composite hypothesis

    what is composite hypothesis

  2. Simple and Composite Statistical Hypothesis definitions

    what is composite hypothesis

  3. Lecture 4. The Multinomial Distribution (II)

    what is composite hypothesis

  4. Unit 3

    what is composite hypothesis

  5. Simple and composite hypothesis ch 16 lec 2

    what is composite hypothesis

  6. 1192012 155942 f023_=_statistical_inference

    what is composite hypothesis

COMMENTS

  1. Composite Hypothesis Test

    What is a composite hypothesis test? Approaches to testing using multiple parameters, explained in simple terms. Composite null hypothesis explained.

  2. PDF Lecture 10: Composite Hypothesis Testing

    Lecture 10: Composite Hypothesis Testing In many real world problems, it is di cult to precisely specify probability distributions. Our models for data may involve unknown parameters or other characteristics. Here are a few motivating examples.

  3. PDF Composite Hypotheses

    The first two options given above are called one-sided tests. The third is called a two-sided test, Rejection and failure to reject the null hypothesis, critical regions, C, and type I and type II errors have the same meaning for a composite hypotheses as it does with a simple hypothesis.

  4. Simple and Composite Hypothesis

    This lecture explains simple and composite hypotheses. Other videos ‪@DrHarishGarg‬ How to write H0 and H1: • How to Write Null and Alternative Hyp... Simple and Composite Hypothesis ...

  5. Composite Hypothesis

    Composite Hypothesis. In subject area: Mathematics. Classically, composite hypotheses are used to determine if a point null is statistically distinguishable from the best alternative, or to determine if the best supported alternative lies on a specified side of the point null. From: Philosophy of Statistics, 2011.

  6. Simple and composite hypothesis

    H H is called a simple hypothesis, if it completely specifies the population distribution; in this case, the sampling distribution of the test statistic is a function of sample size alone. H H is called a composite hypothesis, if it does not completely specify the population distribution; for example, the hypothesis may only specify one parameter of the distribution and leave others unspecified.

  7. Simple Hypothesis and Composite Hypothesis

    The concept of simple and composite hypotheses applies to both the null hypothesis and alternative hypothesis. Hypotheses may also be classified as exact and inexact.

  8. What does "Composite Hypothesis" mean?

    The "composite" part means that such a hypothesis is the union of many simple point hypotheses. In a Null Hypothesis Statistical Test only the null hypothesis can be a point hypothesis. Also, a composite hypothesis usually spans from -∞ to zero or some value of practical significance or from such a value to +∞.

  9. PDF Lecture 7

    7.1 Composite null and alternative hypotheses This week we will discuss various hypothesis testing problems involving a composite null hypothesis and a compositive alternative hypothesis. To motivate the discussion, consider the following examples:

  10. PDF Topic 16: Composite Hypotheses

    Rejection and failure to reject the null hypothesis, critical regions, C, and type I and type II errors have the same meaning for a composite hypotheses as it does with a simple hypothesis.

  11. PDF ECE531 Lecture 5: Composite Hypothesis Testing

    Recall that simple hypothesis testing problems have one state per hypothesis. Composite hypothesis testing problems have at least one hypothesis containing more than one state.

  12. PDF Topic 18: Composite Hypotheses

    The first two options given above are called one-sided tests. The third is called a two-sided test, Rejection and failure to reject the null hypothesis, critical regions, C, and type I and type II errors have the same meaning for a composite hypotheses as it does with a simple hypothesis.

  13. What is a Hypothesis

    Composite Hypothesis A composite hypothesis is a statement that assumes more than one condition or outcome. It can be divided into several sub-hypotheses, each of which represents a different possible outcome.

  14. PDF Practical Statistics

    A different Likelihood ratio for composite hypothesis testing On composite hypotheses, where both null and alternate hypothesis map to values of μ, we can define an alternative likelihood-ratio test statistics that has better properties

  15. Lesson 27: Likelihood Ratio Tests

    In this lesson, we'll learn how to apply a method for developing a hypothesis test for situations in which both the null and alternative hypotheses are composite. That's not completely accurate. The method, called the likelihood ratio test, can be used even when the hypotheses are simple, but it is most commonly used when the alternative ...

  16. Hypothesis Testing in Statistics

    Explore hypothesis testing, a fundamental method in data analysis. Understand how to use it to draw accurate conclusions and make informed decisions.

  17. Lesson 27: Likelihood Ratio Tests

    In this lesson, we'll learn how to apply a method for developing a hypothesis test for situations in which both the null and alternative hypotheses are composite. That's not completely accurate. The method, called the likelihood ratio test, can be used even when the hypotheses are simple, but it is most commonly used when the alternative hypothesis is composite. Throughout the lesson, we'll ...

  18. Composite Hypothesis

    Composite Hypothesis: A statistical hypothesis which does not completely specify the distribution of a random variable is referred to as a composite hypothesis.

  19. What is the difference between Simple and Composite Hypothesis

    A composite hypothesis, on the other hand, proposes multiple relationships between two or more variables. For example, a composite hypothesis might state that "there is a significant difference between the average heights of men and women, and there is also a significant difference between the average heights of people from different ...

  20. PDF Section 11 Goodness-of-fit for composite hypothese

    The test accepts the hypothesis that the data is normal. Notice, however, that something is different. Matlab grouped the data into 6 intervals, so chi-squared test from previous lecture should have r − 1 = 6 − 1 = 5 degrees of freedom, but we have 'df: 3'! The difference is that now our hypothesis is not that the data comes from a particular given distribution but that the data comes ...

  21. PDF Asymptotic Tests of Composite Hypotheses

    A composite hypothesis does not point to a unique probability measure to be used in the hypothesis testing, and this makes it more challenging to test composite hypotheses than simple hypotheses.

  22. When to Combine Hypotheses and Adjust for Multiple Tests

    To provide guidelines for identifying composite hypotheses and addressing the probability of false rejection for multiple hypotheses.Examples from the literature in health services research are used to motivate the discussion of composite hypothesis tests ...

  23. Full article: Finite-Sample Two-Group Composite Hypothesis Testing via

    In the problem of composite hypothesis testing, identifying the potential uniformly most powerful (UMP) unbiased test is of great interest. Beyond typical hypothesis settings with exponential famil...

  24. Ventral attention network connectivity is linked to cortical ...

    Understanding brain development and systems linked to behavioral change is a key goal in population neuroscience. The authors show the ventral attention network is key for brain development and ...