You have selected the station identification number for Sherbrooke. To get general information on finding station ID's, click the following link to the tutorial:
Oliver, John E. Climatology: Selected Applications. p 7.
There are 18 runs in the Sherbrooke data from 1920 to 1970. The total number of elements that make up the sample is 50 (each yearly mean minimum temperature constitutes one element). According to the table, at a .10 significance limit there should be at least 22 runs. We can therefore conclude, with 90% confidence, that this data is not homogeneous. Is this inhomogeneity caused by a large-scale climatic change or by an inconsistancy in the area surrounding the observing station? To answer this question, we analyze the mean minimum temperature at another station only a few miles away.
Locate Dataset and Variable | dataset. |
Select Temporal Domain and Station | in the Time text box. in the ISTA Station text box. The station ID 7018000 is for Shawinigan. To get more information on finding station ID's, click the following link to the tutorial: |
Compute Yearly Mean Minimum Temperature | |
View Yearly Mean Minimum Temperature Time Series | Based on visual inspection, these data appear to be more homogeneous than the data that taken at Sherbrooke. There isn't a distinct upward trend in the minimum temperatures, as there was in the Sherbrooke data. |
Subtract Median From Dataset | The median should be -0.6845208 degrees Celsius. Take note of this value. link. The above command subtracts the median (-0.6845208° Celsius) from each value in the dataset. |
Analyze Homogeneity of Data | link. A table will appear with Time in one column and (Min Temp - -0.6845208) in the other column. Shawinigan is only located a few miles northwest of Sherbrooke across the St. Lawrence River, yet the minimum temperature at Sherbrooke exhibited a noticeable upward trend over the time period while the minimum temperature at Shawinigan did not. Therefore, we can conclude that the inhomogeneity at Sherbrooke is not the result of large-scale climatic change. Instead, from 1920 to 1970, Sherbrooke had been heavily affected by human development. The increased density and height of buildings surrounding the observing station in Sherbrooke caused a small heat island, which in turn created an inhomogeneity in the data. Shawinigan, on the other hand, was not affected by development and in turn, did not experience a gradual warming over the period. |
Posted on 24th February 2022 by Tarik Suljic
My previous blog explored how measures of central tendency (MCT) are used in the majority of clinical research papers, and that reporting MCTs means very little, if anything, in the absence of other secondary data. In this two-part blog, I will discuss the basics of measures of variability and explain how we can use these to evaluate data homogeneity further.
The first part will cover named variability measures (variability measures bearing units with them) and the second part will cover unnamed variability measures (coefficient of variation and z-score).
In this blog, you will learn about:
Measures of variability are statistical tools that assess data homogeneity thus providing us with information on the quality of the sample middle.
Generally, causes of data variations can be classified into three groups:
Mistakes that encompass these errors, if not taken into consideration, are likely to result in distorted validity and systematic errors, therefore resulting in bias. The validity can be defined as the degree to which measurement and/or study represent true inferences. For a study sample, internal validity may be true; however, the external validity that applies to a certain population, may not be.
For example, if a study of the effectiveness of drug A vs drug B for chronic myeloid leukaemia (which occurs mostly in older people) recruits a high proportion of younger patients, the results that arise from this study may be true for the study sample (internal validity), but the results may not apply to the greater population (poor external validity).
One tool we can use to assess the homogeneity of a study sample are measures of variability. We can split them into two categories: (1) those that have units– measures of variability in a narrow sense and (2) those that do not – coefficients. The examples below are all from the first category.
The easiest of all variability measures is range. It represents a span between the two extreme numbers of a set.
For instance, if a set has entries of 15, 17, 19, 20 and 21, the range would be from the minimum number, 15, and the maximum number, 21. Hence the range is calculated as 21-15=6. Values in a range include the units; for example, kilograms (kg) if we are describing the weight of young children.
A range of 6kg in a study of young children is quite large. If it was an intervention research study, this may introduce bias as these children are likely to demonstrate different physiology. If, however, the research study was about a diagnostic radiographic test, then it would be encouraged to have diverse baseline differences because we need to know that a certain device can detect a certain disease in individuals of different ages.
The important thing is that the range of any dataset is taken into context of the study which the data comes from.
In the simplest terms, standard deviation tells you how spread out the data is. Standard deviation (SD) is a linear measure of how far each observed value is from the mean , and describe it not only as a raw value but as being ‘n-standard-deviations away from the mean.’
Let’s first imagine that we have a set A={5, 9, 2, 7, 1}. The formula for SD is:
Where σ is SD, x i is a set constituent, μ is the mean, and N is the number of set constituents.
First, we need to calculate the mean by adding up all constituents of the set and then dividing by the total number of set constituents:
Next, we calculate the square difference of each set constituent and then add them all up before dividing with the total number of set constituents, and finally square-rooting it.
The square difference is calculated by subtracting each set constituent from the mean.
Considering that we obtained (one) SD of approximately ±3 units, we can now draw a numbered line including all the data, and break it down in equal portions, according to the standard deviation.
We said that SD is a linear measure of spread from the mean. Accordingly, the mean of our dataset is 4.8 units, and the SD is ±3 units. So, anything 3 units above the mean, which is 7.8 units, is 1SD away from the mean. Likewise, anything 3 units below the mean is said to be minus 1SD away from the mean. Similarly, if something is 6 units away from the mean, in either direction, it is said to be ±2-standard-deviations away from the mean.
The greater the standard deviation, the greater the variability of the data.
In a total population of students who are being graded 1-5 in several subjects, we expect the majority of them to stick around the (hypothetical) mean, which is 3. Students whose grade point average (GPA) turns out to be either 1 or 5 can be considered as exceptional results. This is true for a normal distribution in which ~68% of sample data is found within 1SD from the mean, ~95.5% within 2SD. Anything above or below 2SD may be considered extreme. Therefore, standard deviation, as a measure of variability, helps us understand where each of our constituents falls under the curve of normal distribution, also known as Gaussian or bell curve.
Variance, like SD, describes how distant set constituents are from the mean. However, the key difference is that variance is an average of squared differences from the sample mean, thus informing us about the average degree to which each dataset constituent differs from the mean.
The formula for variance is as follows:
In the previous section, the SD was ±2.96 units. Should we want to obtain the variance, we just square it.
The figure above is a graphical representation of how to calculate population variance, which we will talk through.
The variance tells us how variable the data is, as a squared value. This is less widely used than standard deviation (which is the linear equivalent of variance) as the standard deviation is in the same units as the mean.
According to the American Lung Association , an estimated prevalence of lung cancer for 2020 in the USA was around 541,000. Should there be research investigating efficacy of a therapy in lung cancer patients in the USA, it would not be possible for all 541,000 patients to be enrolled in the research. Instead, only a representative number of the total number of patients would be enrolled. This representative subset of the lung cancer population is called a ‘sample’.
The formulae above for both standard deviation and variance were written as for a population , and not for a sample , although a substantial amount of research is actually conducted on samples representing population and not on a population itself. For this reason, (bio)statisticians use adjusted formulae to calculate standard deviation and variance of a sample. The key differences between these two formulae are listed in the table below:
If we take two studies which recruit people of different ages (in years):
A | 65, 44, 52, 18, 46, 73, 25, 34, 46, 49, 53, 61 |
B | 47, 51, 44, 39, 47, 50, 43, 49, 51, 40, 39, 52 |
Looking at the table below, Group A has a more diverse, variable dataset than Group B. Therefore, the results that come out of Group A have a greater chance that they will apply to all of the members of the dataset (internal validity).
We are looking for a study of middle-aged participants. The age range in Group B (39-52) is lower than in Group A (18-73). The results of Group B have greater chances of applying to the majority of patients suffering from this disease than the results of Group A (external validity), as well as to the patients from the study sample (internal validity).
Furthermore, from the table, we can see that the measures of central tendency appear to be very similar between Group A and Group B. Yet, the variance and standard deviation are by far and large greater in Group A compared to Group B, indicating that the dataset has quite a large variability in its constituents. Hence, we can say that Group B has a more homogenous dataset than Group A.
You can continue your learning in my next blog , which discusses the coefficient of variation and z-score.
Leave a reply cancel reply.
Your email address will not be published. Required fields are marked *
Save my name, email, and website in this browser for the next time I comment.
You will receive our monthly newsletter and free access to Trip Premium.
Using measures of variability to inspect homogeneity of a sample: part 2.
Measures of variability are statistical tools that help us assess data variability by informing us about the quality of a dataset mean. This second of two blogs on the topic will look at coefficient of variation and the z-score.
A beginner’s guide to standard deviation and standard error: what are they, how are they different and how do you calculate them?
How can you tell if a variable is nominal, ordinal, or numerical? Why does it even matter? Determining the appropriate variable type used in a study is essential to determining the correct statistical method to use when obtaining your results. It is important not to take the variables out of context because more often than not, the same variable that can be ordinal can also be numerical, depending on how the data was recorded and analyzed. This post will give you a specific example that may help you better grasp this concept.
The goodness-of-fit test can be used to decide whether a population fits a given distribution, but it will not suffice to decide whether two populations follow the same unknown distribution. A different test, called the test for homogeneity , can be used to draw a conclusion about whether two populations have the same distribution. To calculate the test statistic for a test for homogeneity, follow the same procedure as with the test of independence.
The expected value for each cell needs to be at least five for you to use this test.
Hypotheses H 0 : The distributions of the two populations are the same. H a : The distributions of the two populations are not the same.
Test Statistic Use a χ 2 χ 2 test statistic. It is computed in the same way as the test for independence.
Degrees of freedom ( df ) df = number of columns – 1
Requirements All values in the table must be greater than or equal to five.
Common Uses Comparing two populations. For example: men vs. women, before vs. after, east vs. west. The variable is categorical with more than two possible response values.
Do male and female college students have the same distribution of living arrangements? Use a level of significance of 0.05. Suppose that 250 randomly selected male college students and 300 randomly selected female college students were asked about their living arrangements: dormitory, apartment, with parents, other. The results are shown in Table 11.19 . Do male and female college students have the same distribution of living arrangements?
72 | 84 | 49 | 45 | |
91 | 86 | 88 | 35 |
H 0 : The distribution of living arrangements for male college students is the same as the distribution of living arrangements for female college students. H a : The distribution of living arrangements for male college students is not the same as the distribution of living arrangements for female college students. Degrees of freedom ( df ): df = number of columns – 1 = 4 – 1 = 3 Distribution for the test: χ 3 2 χ 3 2 Calculate the test statistic: χ 2 = 10.1287 (calculator or computer) Probability statement: p -value = P ( χ 2 >10.1287) = 0.0175
Compare α and the p -value: Since no α is given, assume α = 0.05. p -value = 0.0175. α > p -value. Make a decision: Since α > p -value, reject H 0 . This means that the distributions are not the same. Conclusion: At a 5 percent level of significance, from the data, there is sufficient evidence to conclude that the distributions of living arrangements for male and female college students are not the same. Notice that the conclusion is only that the distributions are not the same. We cannot use the test for homogeneity to draw any conclusions about how they differ.
Do families and singles have the same distribution of cars? Suppose that 100 randomly selected families and 200 randomly selected singles were asked what type of car they drove: sport, sedan, hatchback, truck, van/SUV. The results are shown in Table 11.20 . Do families and singles have the same distribution of cars? Test at a level of significance of 0.05.
Sport | Sedan | Hatchback | Truck | Van/SUV | |
---|---|---|---|---|---|
5 | 15 | 35 | 17 | 28 | |
45 | 65 | 37 | 46 | 7 |
Both before and after a recent earthquake, surveys were conducted asking voters which of the three candidates they planned on voting for in the upcoming city council election. Has there been a change since the earthquake? Use a level of significance of 0.05. Table 11.21 shows the results of the survey. Has there been a change in the distribution of voter preferences since the earthquake?
167 | 128 | 135 | |
214 | 197 | 225 |
H 0 : The distribution of voter preferences was the same before and after the earthquake. H a : The distribution of voter preferences was not the same before and after the earthquake. Degrees of freedom ( df ): df = number of columns – 1 = 3 – 1 = 2 Distribution for the test: χ 2 2 χ 2 2 Calculate the test statistic: χ 2 = 3.2603 (calculator or computer) Probability statement: p -value= P ( χ 2 > 3.2603) = 0.1959
Press the MATRX key and arrow over to EDIT . Press 1:[A] . Press 2 ENTER 3 ENTER . Enter the table values by row. Press ENTER after each. Press 2nd QUIT . Press STAT and arrow over to TESTS . Arrow down to C:χ2-TEST . Press ENTER . You should see Observed:[A] and Expected:[B] . Arrow down to Calculate . Press ENTER . The test statistic is 3.2603 and the p -value = 0.1959. Do the procedure a second time but arrow down to Draw instead of Calculate .
Compare α and the p -value: α = 0.05 and the p -value = 0.1959. α < p -value.
Make a decision: Since α < p -value, do not reject H o .
Conclusion: At a 5 percent level of significance, from the data, there is insufficient evidence to conclude that the distribution of voter preferences was not the same before and after the earthquake.
Ivy League schools receive many applications, but only some can be accepted. At the schools listed in Table 11.22 , two types of applications are accepted: regular and early decision.
Application Type Accepted | Brown | Columbia | Cornell | Dartmouth | Penn | Yale |
---|---|---|---|---|---|---|
2,115 | 1,792 | 5,306 | 1,734 | 2,685 | 1,245 | |
577 | 627 | 1,228 | 444 | 1,195 | 761 |
We want to know if the number of regular applications accepted follows the same distribution as the number of early applications accepted. State the null and alternative hypotheses, the degrees of freedom and the test statistic, sketch the graph of the p -value, and draw a conclusion about the test of homogeneity.
This book may not be used in the training of large language models or otherwise be ingested into large language models or generative AI offerings without OpenStax's permission.
Want to cite, share, or modify this book? This book uses the Creative Commons Attribution License and you must attribute Texas Education Agency (TEA). The original material is available at: https://www.texasgateway.org/book/tea-statistics . Changes were made to the original material, including updates to art, structure, and other content updates.
Access for free at https://openstax.org/books/statistics/pages/1-introduction
© Apr 16, 2024 Texas Education Agency (TEA). The OpenStax name, OpenStax logo, OpenStax book covers, OpenStax CNX name, and OpenStax CNX logo are not subject to the Creative Commons license and may not be reproduced without the prior and express written consent of Rice University.
A discussion of qualitative & quantitative research design, focus groups: heterogeneity vs. homogeneity.
The following is a modified excerpt from Applied Qualitative Research Design: A Total Quality Framework Approach (Roller & Lavrakas, 2015, pp. 107-109).
Fundamental to the design of a focus group study is group composition. Specifically, the researcher must determine the degree of homogeneity or heterogeneity that should be represented by the group participants. As shown below, there are many questions the researcher needs to contemplate, such as the extent of similarity or dissimilarity in participants’ demographic characteristics, as well as in their experiences and involvement with the subject matter.
A few of the questions the focus group researcher might consider when determining the desired heterogeneity or homogeneity among group participants include: |
Whether or not—or the degree to which—group participants should be homogeneous in some or all characteristics has been at the center of debate for some years. On the one hand, Grønkjær, Curtis, Crespigny, and Delmar (2011) claim that at least some “homogeneity in focus group construction is considered essential for group interaction and dynamics” (p. 23)—for example, participants belonging to the same age group may have similar frames of reference and feel comfortable sharing their thoughts with people who have lived through the same experience. In the same vein, Sim (1998) states that, “the more homogeneous the membership of the group, in terms of social background, level of education, knowledge, and experience, the more confident individual group members are likely to be in voicing their [own] views” (p. 348). Even among strangers, there is a certain amount of comfort and safety in the group environment when the participants share key demographic characteristics, cultural backgrounds, and/or relevant experience.
A problem arises, however, if this comfortable, safe environment breeds a single-mindedness (or “groupthink”) that, without the tactics of a skillful moderator, can stifle divergent thinking and result in erroneous, one-sided data. Heterogeneity of group participants (e.g., including users and nonusers of a particular child care service within the same focus group) potentially heads off these problems by stimulating different points of view and a depth of understanding that comes from listening to participants “defend” their way of thinking (e.g., product or service preferences). As Grønkjær et al. (2011) state, “a group may be too homogeneous; thus influencing the range and variety of the data that emerges” (p. 26). The tension that heterogeneity may create in a group discussion can serve to uncover deeper insights into what is being studied, providing the moderator is able to channel this tension in constructive directions. In addition to a heightened level of diversity, heterogeneous groups may also be a very pragmatic choice for the researcher who is working with limited time and financial resources, or whose target population for the research is confined to a very narrow group (e.g., nurses working at a community hospital).
Ultimately, the answer to the question of whether group participants should be homogeneous or heterogeneous is “it depends.” As a general rule, group participants should have similar experiences with, or knowledge of, the research topic (e.g., using the Web to diagnose a health problem, weekly consumption of fat-free milk), but the need for “sameness” among participants on other parameters can fluctuate depending on the circumstance. Halcomb, Gholizadeh, DiGiacomo, Phillips, and Davidson (2007), for example, report that homogeneity of age is particularly important in non-Western countries where younger people may believe it is disrespectful to offer comments that differ from those stated by their elders. Homogeneous groups are also important when investigating sensitive topics, such as drug use among teenagers, where a more mixed group of participants with people who are perceived as “different” (in terms of demographics and knowledge/experience with drugs) may choke the discussion and lead to a struggle for control among participants (e.g., one or more participants trying to dominate the discussion).
Homogeneity of gender, on the other hand, may or may not be important to the success (usefulness) of a focus group study. For example, an organization conducting employee focus group research to explore employees’ attitudes toward recent shifts in management would need to conduct separate groups with men and women in order to discover how the underlying emotional response to new management differs between male and female employees. In contrast, a focus group study among city residents concerning public transportation might benefit from including both men and women in the same discussion, among whom the varied use and perceptions of the transportation services would serve to stimulate thinking and enrich the research findings. The heightened level of dynamics in groups that are heterogeneous in gender and other aspects may also provoke conversations on taboo subjects (e.g., racism) that might not be forthcoming in other methods such as in-depth interviews.
Grønkjær, M., Curtis, T., de Crespigny, C., & Delmar, C. (2011). Analysing group interaction in focus group research: Impact on content and the role of the moderator. Qualitative Studies , 2 (1), 16–30.
Halcomb, E. J., Gholizadeh, L., DiGiacomo, M., Phillips, J., & Davidson, P. M. (2007). Literature review: Considerations in undertaking focus group research with culturally and linguistically diverse groups. Journal of Clinical Nursing , 16 (6), 1000–1011. https://doi.org/10.1111/j.1365-2702.2006.01760.x
Sim, J. (1998). Collecting and analysing qualitative data: Issues raised by the focus group. Journal of Advanced Nursing , 28 (2), 345–352. Retrieved from http://www.ncbi.nlm.nih.gov/pubmed/9725732
Images captured/created from: https://www.thoughtco.com/heterogeneous-definition-and-example-606355
Design decisions about focus groups – should be deliberate (unless it’s with a naturally occurring group). However, as with any research endeavor – what drives design decisions – is/are the research question(s) being asked.
Thank you, Joe, for the comment. Indeed, it is all about the research question(s) — and the type of participants needed to address the objectives — that guide all sorts of design decisions. Reflecting on the type of participants and the objectives is important to the heterogeneity-homogeneity consideration.
This site uses Akismet to reduce spam. Learn how your comment data is processed .
--> |
> > to return to the Sampling page |
| © RWJF 2008 | | |
This website may not work correctly because your browser is out of date. Please update your browser .
Homogenous sampling involves selecting similar cases to further investigate a particular phenomenon or subgroup of interest.
The logic of homogenous sampling is in contrast to the logic of maximum variation sampling.
In a recent evaluation of village level revitalization in Aceh, post-tsunami, leadership was identified as a contributing factor to village success. Those villages with effective leaders were able to rebuild more productively than those without effective leadership. A homogenous sample of village leaders could be a useful augment to the study, identifying common leadership characteristics and circumstances.
This page is a Stub (a minimal version of a page). You can help expand it. Contact Us to recommend resources or volunteer to expand the description.
Framework/guide.
Back to top
© 2022 BetterEvaluation. All right reserved.
An official website of the United States government
The .gov means it’s official. Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.
The site is secure. The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.
Preview improvements coming to the PMC website in October 2024. Learn More or Try it out now .
Sampling is a key feature of every study in developmental science. Although sampling has far-reaching implications, too little attention is paid to sampling. Here, we describe, discuss, and evaluate four prominent sampling strategies in developmental science: population-based probability sampling, convenience sampling, quota sampling, and homogeneous sampling. We then judge these sampling strategies by five criteria: whether they yield representative and generalizable estimates of a study’s target population, whether they yield representative and generalizable estimates of subsamples within a study’s target population, the recruitment efforts and costs they entail, whether they yield sufficient power to detect subsample differences, and whether they introduce “noise” related to variation in subsamples and whether that “noise” can be accounted for statistically. We use sample composition of gender, ethnicity, and socioeconomic status to illustrate and assess the four sampling strategies. Finally, we tally the use of the four sampling strategies in five prominent developmental science journals and make recommendations about best practices for sample selection and reporting.
When we undertake to study some phenomenon, we wish to know something about that phenomenon in a population, but in practice we study the phenomenon in a group of individuals who purportedly represent the target or reference population to whom we wish our results to generalize. That is, we sample the population. We sample because we normally do not command the resources (time, money, or personnel) to assess the entire population of interest. Sampling is therefore a key feature of every study in developmental science, and sampling has far-reaching implications in all studies. This article is concerned with sampling in developmental science. As we point out, different sampling strategies exist, and each has its implications. Employing sub-optimal sampling strategies is far too common in developmental research, compromises the validity and utility of the research, renders replication and cross-study comparisons difficult, and most generally impedes progress in the field of developmental science.
In this article, we briefly describe and illustrate four prominent strategies that answer the sampling challenge, and we evaluate each in terms of some fundamental, meaningful, and practical criteria. The four strategies include (a) population-based probability sampling as well as nonprobability sampling strategies such as (b) convenience sampling, (c) quota sampling, and (d) homogeneous sampling. The five criteria by which we appraise these sampling strategies include (a) whether they yield representative and generalizable estimates of a study’s target population (e.g., estimates of intelligence among the population when all sociodemographic groups are collapsed), (b) whether they yield representative and generalizable estimates of sociodemographic group differences within a study’s target population (e.g., how estimates of intelligence vary across a population’s ethnic groups), (c) the recruitment efforts and costs they entail, (d) whether they provide sufficient power to detect sociodemographic group differences, and (e) whether they introduce noise related to variation in sociodemographic factors and whether that noise can be accounted for statistically. After overviewing the four sampling strategies, we examine how the sociodemographic composition of a sample in terms of gender, ethnicity, and SES can compromise a study’s findings – regardless of the study goals. We then recount the use of each prominent sampling strategy in five high-profile journals in contemporary developmental science. On these bases, we arrive at conclusions and recommendations about best practices and practical considerations, including ethical issues, and discuss the importance of weighing the research question when considering the merits of various sampling strategies.
This article is not comprehensive, and we have not assumed some related burdens. By now demographers, sociologists, and others in many disciplines have weighed the pros and cons of different sampling strategies ( Davis-Kean & Jager, 2011 ; Henry, 1990 ; Onwuegbuzie & Collins, 2007 ; Sue, 1999 ; Watters & Biernacki, 1989 ). This article does not provide a tutorial on sampling (see http://stattrek.com/statistics/data-collection-methods.aspx?Tutorial=Stat ). We also eschew technical details in favor of highlighting “big picture” issues of design and practicality in an accessible way. Although our examples and arguments are applicable to any single sociodemographic factor or set of sociodemographic factors, here we limit our focus to gender, ethnicity, and SES. Also, although we fully recognize that gender, ethnicity, and SES are non-independent (ethnicity and SES in particular) but interact in myriad complex ways, when discussing the implications of these three sociodemographic factors we typically limit our examples to a single factor for the sake of conceptual clarity. Finally, for the purposes of this exposition about sampling, we combine “race” and “ethnicity” as used by the U.S. Government and its agencies and define six ethnic categories (see Table 1 ). We acknowledge that these ethnic groups are also heterogeneous in that each group contains people who originated from many different countries with different cultural practices.
Ethnicity distribution in the United States in 2010
Ethnicity | Percentage |
---|---|
White | 63.75% |
Hispanic | 16.35% |
Black | 12.21% |
Asian | 4.69% |
American Indian/Alaskan Native | 0.73% |
Hawaiian/Other Pacific Islander | 0.16% |
Note . Adapted from Table 1 in Humes, Jones, and Ramirez (2011) .
Here we describe four of the most used sampling strategies, and we assess their advantages, disadvantages, and limitations. How each of the four sampling strategies fares on the five criteria is summarized in Table 2 .
Evaluation of sampling strategies based on five criteria
Population-based | Convenience | Quota | Homogeneous | |
---|---|---|---|---|
Generalizable estimates of target population | Yes | No | No | No |
Generalizable estimates of sociodemographic differences | Yes | No | No | No |
Ease of recruitment | Low | High | Intermediate | High |
Power to detect sociodemographic group differences | Yes | No | Yes | No |
Absence of sociodemographic “noise” in results | No | No | No | Yes |
A principal set of sampling strategies falls under the category of “population-based probability sampling.” These strategies include simple random sampling as well as more complex sampling designs such as stratified sampling and cluster sampling (and its variants such as probability proportional to size sampling). Because a detailed exposition of these strategies is beyond this scope of this paper, here we only provide basic descriptions (for more thorough reviews see Cochran, 1977 , or Levy & Lameshown, 2011 ). In simple random sampling, a random subset ( n ) of the target population ( N ) is selected, with each member of N having an equal probability of selection. In stratified sampling, the population is divided into separate groups called “strata” (such as ethnic groups), and then a probability sample (often a simple random sample) is drawn from each stratum. With cluster sampling, the target population is divided into separate geographic groups called “clusters” (such as schools, neighborhoods, businesses), a simple random sample of clusters is selected from the population, and data collection is limited to those who fall within these randomly selected clusters. Within each selected cluster, data collection can be probability based (i.e., based on a simple random sample or a stratified design) or complete (i.e., every individual within a given cluster is eligible to participate in the study). Although these population-based probability sampling strategies differ from one another in important ways, they all, when carried out properly, yield an unbiased sample that is representative of the target population (i.e., the sociodemographic characteristics of the sample faithfully reflect the sociodemographic characteristics of the target population). For example, the National Longitudinal Study of Adolescent Health (Add Health; Bearman, Jones, & Udry, 1997 ), which used a clustered sampling design that applied stratified sampling within cluster (school), is a population-based probability sample of 7 th -12 th graders in the United States during the 1994–1995 school year. The sample has been re-interviewed three times, the most recent being in 2008, when the sample was ages 24 to 32. Because Add Health is a population-based probability sample with a clear target population, researchers and practitioners can be confident that findings from studies utilizing Add Health data generalize to the U.S. population of adolescents and young adults as a whole.
Based on our five criteria, population-based probability sampling appears to have more advantages than disadvantages. Focusing first on its advantages, in terms of representativeness and generalizability, when carried out properly, these sorts of samples yield generalizable estimates of the target population (i.e., all sociodemographic groups combined) and of sociodemographic group differences (e.g., gender, ethnic, or SES differences within a target population). Additionally, assuming the subsamples for a given sociodemographic factor are sufficiently large (say ≥ 45; based on .80 power to detect a medium effect of f = .25, α = .05, in an ANOVA design with four groups 1 ; Faul, Erdfelder, Lang, & Buchner, 2007 ), this sampling strategy yields sufficient power to detect differences among sociodemographic subgroups within the target population. Finally, for researchers not interested in subgroup differences, probability samples also allow accounting for noise introduced by variation in sociodemographic factors. Taking advantage of the substantial sociodemographic variation in these studies, researchers can take steps to control for sociodemographic group differences, or, better, researchers can simply examine their research question separately for each sociodemographic subgroup (i.e., examine questions separately by gender using multiple-group analyses) and compare the findings.
Regarding its disadvantages, when done properly the recruitment costs and efforts for population-based probability sampling are high. Regardless of the population-based sampling strategy used, researchers need to carefully define the target population and clarify its sociodemographic composition. Depending on the target population, doing so can be straightforward (e.g., the ethnic composition of the U.S. population is tracked by the U.S. Census). However, in many cases the sociodemographic compositions of other, smaller target populations are not fully known and may require great efforts to accurately determine. In addition, when researchers sample from geographic areas smaller than their country, geographical decisions can be arbitrary. Suppose one were to sample from the geographic home of the National Institutes of Health (NIH). What would define that sampling area? The main campus of the NIH? The town of Bethesda? Montgomery Country? The state of Maryland? The Middle Atlantic States? Second, population-based probability-sample sizes need to be quite large, often coming at great costs in terms of money, time, and effort. For example, consider a research group that wishes to collect ethnicity data that is representative of a given target population via simple random sampling. Unless the target population is a single ethnic group (see Homogeneous Sampling below), any target population will consist of a set of ethnic groups that are not distributed equally (i.e., some comprise a far lesser proportion of the target population than others). Therefore, to yield a subsample of each ethnic group that is statistically useful 2 , including those that comprise a small proportion of the target population, requires collecting a substantial amount of data. For example, using the U.S. population as the target population, to yield a Hispanic n of 45 (a rather modest n for making generalizations to the entire U.S. population of Hispanics) calls for a total representative sample N of 275. To yield n s of 45 for non-Hispanic Blacks, Asians, American Indians, and Hawaiian/Other Pacific Islanders would call for total representative sample N s of 369, 959, 6,164, and 28,125, respectively. Of course the required sample size balloons even higher when multiple sociodemographic factors are considered (e.g., to yield an n of 45 for Hawaiian/Other Pacific Islander females would call for a total sample N of 56,250).
One of the benefits of clustered and stratified sampling designs is that, relative to simple random sampling, the recruitment costs and efforts are often lower ( Groves, 1989 ; Heeringa et al., 2004 ), although still considerable. Within stratified designs, strata that are underrepresented within the population (e.g., the American Indian stratum among ethnicity strata or the highly affluent stratum among SES strata) can be oversampled (termed “disproportionate allocation”) reducing the overall N required to yield a statistically useful subsample of each subgroup. Sample weights, which “down weight” the oversampled strata, can then be applied to the data to yield estimates that are generalizable to the total population. Additionally, although cluster designs do not necessarily reduce the number of individuals that need to be recruited, they can reduce recruitment costs. In terms of effort and time, it is far easier to sample, for example, 100 individuals within a single sampling cluster (i.e., neighborhood or school) than it is to sample 100 individuals scattered across a number of different sampling clusters. Stratified and clustered sampling designs also have their disadvantages relative to simple random sampling. Complex sampling strategies are less straightforward to implement and require the use of specialized analytical techniques to obtain accurate variance estimates ( Davis-Kean & Jager, 2011 ).
In summary, population-based probability sampling strategies allow for clear generalizability to both the target population and its sociodemographic subpopulations, and they enable researchers to account for the noise introduced by variation in sociodemographic factors. Despite these important advantages, population-based sampling strategies are often prohibitively costly and labor-intensive.
Due to the substantial costs of population-based sampling strategies, the use of nonprobability samples, which are typically less expensive in all ways, is much more common in developmental research. Next, we review and critique three specific types of nonprobability samples: convenience sampling, quota sampling, and homogenous sampling.
Unlike population-based sampling strategies, convenience sampling is a nonprobability sampling strategy where participants are selected based on their accessibility and/or proximity to the research. One of the most common examples of convenience sampling within developmental research is the use of student volunteers as study participants. This strategy entails recruiting a sample that has some ad hoc sociodemographic composition, which is tailored neither to the United States nor to any other identifiable target population but rather settles for whatever convenience sample the researcher recruits (presumably) on a first come-first recruited basis.
This strategy’s clear advantage is that, of all the sampling strategies, convenience sampling is the easiest, least time-intensive, and least expensive to implement, perhaps accounting for its popularity in developmental research. Regarding its disadvantages, results that derive from convenience sampling have known generalizability only to the sample studied. Thus, any research question addressed by this strategy is limited to the sample itself. The same limitation holds true for estimates of differences between sociodemographic subgroups. As another disadvantage, convenience samples typically include small numbers of underrepresented sociodemographic subgroups (e.g., ethnic minorities) resulting in insufficient power to detect subgroup differences within a sociodemographic factor or factors. Moreover, although small in number, these underrepresented sociodemographic subgroups introduce modest amounts of variation into the sample, enough variation to produce statistical noise in the analyses but not enough variation to harness or control statistically. Indeed, the widespread use of convenience sampling may be partly responsible for the host of small and inconsistent effects that pervade developmental science, why sizes of effects often vary depending on the variables considered, and why research shows links between particular setting conditions and outcomes for some, but not other, groups ( Bornstein, 2013 ).
In summary, convenience sampling is a common strategy, but its scientific disadvantages appear to outweigh its practical advantages. Relative to population-based probability sampling, convenience sampling is far easier and less expensive to implement. However, unlike population-based probability sampling, convenience sampling produces estimates that lack generalizability to any identifiable target population or subpopulations (except for the sample studied), provides insufficient power to detect differences among sociodemographic subgroups, and includes noise due to sociodemographic variation that cannot be controlled or accounted for.
Because of the well-intentioned movement to improve the representation of underrepresented groups in developmental research, there has developed a sampling strategy of recruiting fixed numbers of participants from different sociodemographic groups (e.g., sample n s of 45 for each ethnic group within the target population). Like convenience sampling, quota samples are typically nonprobability samples. Although quota sampling (also referred to as “equal sampling”) bears some resemblance to stratified sampling using disproportionate allocation (i.e., oversampling underrepresented groups or strata), it is distinct from stratified sampling in two important ways. First, stratified sampling using disproportionate allocation draws a probability sample from each stratum under investigation, but quota sampling typically draws a nonprobability sample from each group under investigation (i.e., a non-random sample of each ethnic group within the target population). Second, stratified sampling using disproportionate allocation is typically accompanied by the use of sample weights, which render estimates that are generalizable to the target population, but quota sampling rarely involves the calculation or application of sample weights.
Based on our five criteria, quota sampling has more disadvantages than advantages. One advantage of quota sampling is that, because it often entails oversampling of underrepresented groups, quota sampling typically provides sufficient statistical power to detect group differences. Another advantage of quota sampling is that, for researchers not interested in differences across subgroups of a sociodemographic factor, quota sampling permits accounting statistically for noise introduced by sociodemographic variation. However, because quota sampling typically draws nonprobability samples from each group under investigation, it does not yield generalizable estimates of the target population or of subgroup differences within the target population. Indeed, that sample weights are rarely applied renders estimates of the target population all the more biased. A second disadvantage of quota sampling is that, when done properly, the recruitment costs and efforts are intermediate. Although not as large as population-based samples, the size of the overall sample still needs to be fairly large. That is, the sample still needs to include a statistically useful sub-sample of each subgroup of a sociodemographic factor, which at the low end amounts to an n = 45 (for a 4-group ANOVA). For example, for studies focused on ethnicity, based on the six-category definition of ethnicity, a minimum total sample N of 270 would be required (i.e., a subgroup n of 45, by 6 subgroups, equals a total N of 270). Locating high numbers of those in underrepresented groups to participate in a study can be challenging.
In summary, relative to convenience samples, quota samples are equally disadvantaged when it comes to producing estimates of the target population and of subgroup differences that are generalizable and more disadvantaged with respect to recruitment costs. However, unlike convenience samples, quota sampling typically provides the statistical power necessary to detect subgroup differences among sociodemographic factor(s) of interest (i.e., the sociodemographic factor(s) whose subgroups are quota sampled) and permits accounting statistically for noise introduced by variation in the sociodemographic factor(s) of interest. Therefore, relative to convenience samples, quota samples are better suited for examining subgroup differences within a sociodemographic factor or factors.
In homogenous sampling, researchers undertake to study a sociodemographically homogeneous population (e.g., the overall sample is comprised of just females). Homogenous samples can vary in their degree of sociodemographic homogeneity. For example, samples that restrict variation on multiple sociodemographic factors (e.g., a sample limited to female, European American adults) are more homogeneous than samples that restrict variation on a single sociodemographic factor (e.g., a sample limited to females that includes all ethnic groups). Like convenience and quota samples, homogenous samples are typically nonprobability samples.
Based on our five criteria, homogeneous sampling has important advantages and disadvantages. Focusing first on its advantages, the recruitment costs and efforts are generally low for homogeneous samples, although relative to homogenous samples of overrepresented sociodemographic groups (e.g. European Americans in the United States), homogeneous samples of underrepresented sociodemographic groups (e.g., Native Americans) can be more costly and time consuming. An additional advantage is, because the homogeneous sampling design eliminates all variation associated with one or more sociodemographic factors, it adds no noise associated with those sociodemographic factors to the overall results. That is, by including only one ethnic group, for example, the research intentionally avoids the noise associated with ethnic confounds that can cloud the findings if different ethnic groups are combined, thereby improving the accuracy and quality of the resultant data. Provided homogeneous samples are nonprobability samples, which in practice is typically the case, a key shortcoming is that they yield estimates that are not generalizable to the target population (e.g., a non-probability, homogenous sample of Native Americans would not yield estimates that are generalizable to the population of Native Americans). However, this limitation does not hold for probability homogeneous samples. Additionally, because the target population is defined as a particular sociodemographic group, homogeneous samples are often ill-suited to examine sociodemographic differences. For example, homogeneous samples limited to females cannot be used to examine gender differences, and homogeneous samples limited to African American females cannot be used to examine ethnic differences, gender differences, or ethnic differences within gender.
In summary, because homogeneous samples are typically nonprobability samples, they are equally disadvantaged relative to convenience samples and quota samples when it comes to producing estimates of the target population that are generalizable. Like convenience samples, recruiting costs for homogeneous samples are typically low, and lower than the costs associated with population-based probability and quota samples. Unlike convenience samples and quota samples, homogeneous samples eliminate noise due to variation in one or more sociodemographic factors. Therefore, homogeneous samples provide key advantages over both convenience samples (i.e., noise due to variation in one or more sociodemographic factors is eliminated in homogenous samples) and quota samples (i.e., recruitment costs are typically lower for homogeneous samples).
Given the advantages and disadvantages of the four sampling strategies, it is important to note how sociodemographic characteristics can affect study outcomes and the interpretation of study results. Gender, ethnicity, and SES variation in many characteristics—physical and mental health ( Adkins, Wang, Dupre, van den Oord, & Elder, 2009 ; American Public Health Association, 2004 ; Breslau et al., 2005 ; Crimmins & Saito, 2001 ; National Center for Health Statistics, 2011 ), beliefs and cognitions ( Burke et al., 1992 ; Courtenay, McCreary, & Merighi, 2002 ; Diala et al., 2000 ), and behaviors and practices ( Blum et al., 2000 ; Burke et al., 1992 ; Courtenay et al., 2002 ; National Center for Health Statistics 2011 ; Snowden & Yamada, 2005 )—is ubiquitous. Beyond mean-level differences in these constructs, cross-sectional associations as well as developmental linkages among them and other salient constructs also vary by gender ( Card, Stucky, Sawalani, & Little, 2008 ; Griffin, Botvin, Scheier, Diaz, & Miller, 2000 ), ethnicity ( Amey, Albrecht, & Miller, 1996 ; Burke et al., 1992 ; Jager, 2011 ), and SES ( Geoffroy et al., 2007 ; Sachs-Ericsson et al., 2007 ). These broad and entrenched sociodemographic differences in both the levels and the correlates of significant developmental characteristics—constructs, structures, functions, or processes—place demands on developmental researchers regarding the sociodemographic composition of their analytic samples ( Bornstein, 2010 ; Davis-Kean & Jager, 2011 ; Henry, 1990 ; Onwuegbuzie & Collins, 2007 ; Sue, 1999 ; Watters & Biernacki, 1989 ). Moreover, the practical implications of a sample’s sociodemographic composition vary depending on whether a study is (a) focused on one or more sociodemographic factors as a source of heterogeneity or (b) focused on broad developmental patterns (not focused on sources of heterogeneity of any kind).
The roots of sociodemographic differences – including gender, ethnicity, and SES differences - in developmental characteristics are complex and likely the product of layered interactions among biological, behavioral, and sociocultural factors ( Betencourt & Lopez, 1993 ; Courtenay et al., 2002 ; Crimmins & Saito, 2001 ; Phinney, 1996 ). Nonetheless, they are important to unpackage because without a scientific base of knowledge regarding human health and behavior that takes into account the sociodemographic diversity of the population, health care delivery, planning, and policy making would be compromised by inadequate information and potentially misleading generalizations ( Betencourt & Lopez, 1993 ; Hahn & Stroup, 1994 ; Mays, Ponce, Washington, & Cochran, 2003 ). Indeed, a sizable amount of developmental research has been devoted to examining sociodemographic differences in health and development. However, if research does not include a broad, representative sample of the sociodemographic groups under examination, then a distorted view of sociodemographic differences likely results. Additionally, if research does not include adequately large samples of each sociodemographic group under examination, then sociodemographic differences that exist in the target population may go undetected (i.e., Type II errors) due to a lack of power. Thus, a proper examination of socioeconomic differences in a given developmental characteristic demands an analytical sample whose sociodemographic composition is (a) varied enough to represent the diversity of the sociodemographic groups under examination and (b) large enough to yield sufficient power to detect differences among the sociodemographic groups under examination.
Studies whose sociodemographic compositions do not meet these criteria are problematic because (a) when considered individually it is unclear whether a study’s findings generalize to the intended target population and (b) when considered collectively, the findings from these studies are difficult to synthesize. For purposes of illustration, consider the following three hypothetical studies exploring ethnic differences in depression within the United States, all of which poorly sample ethnicity (either in terms of representation, number, or both). Study A utilized a sample of 1,000 middle-aged, married adults from an affluent East Coast suburban community and found that European Americans reported higher rates of major depressive disorders than all other ethnic groups; here, the sample is large but not representative of the ethnic distribution of the United States. Study B utilized a diverse sample of 200 adults and found that all ethnic groups reported equivalent rates of major depressive disorders; here the sample is representative but not large. Study C, whose sample is neither large nor representative, utilized an impoverished, rural sample of 150 adults and found one group difference: African Americans reported higher rates of major depressive disorders than did European Americans. When considering any one of these three studies individually, can one confidently conclude that its findings regarding ethnic differences in depression are a true reflection of the United States population? Although Study A’s sample is adequately large to yield sufficient power to detect group differences, its estimates of group difference may be biased due to the characteristics of the sample (i.e., affluent, married, suburban adults). Although Study B’s sample is more diverse, it is also much smaller and likely insufficiently powered to detect ethnic group differences. As a result, its finding that the ethnic groups did not differ from one another could simply reflect lack of power (i.e., a Type II error). Finally, the one group difference found by Study C could be an artifact of its biased sample, and its failure to detect any other group differences could be ascribable to insufficient power on account of the small sample size. Additionally, when considered collectively, how does one go about integrating the findings from these three studies? Their findings regarding ethnic differences in depression are inconsistent; therefore, they cannot all be correct. But is one more correct (or less incorrect)? Unfortunately, determining the answers to these questions is difficult because each study’s sample is deficient in representation and/or sample size. Although this example involved a set of hypothetical studies, substantial variation in the size and sociodemographic composition of samples is all too common across studies examining sociodemographic differences in a given developmental characteristic in an equivalent target population. These variations make it difficult to determine whether inconsistencies across studies represent true population differences or instead are artifacts of differences in sample composition.
Even for developmental research not focused on sociodemographic factors as a source of heterogeneity, issues pertaining to the sociodemographic composition of the sample still warrant consideration ( LaVeist, 2005 ; Phinney, 1996 ; Williams, 1999 ). For example, much developmental research focuses on developmental patterns in a particular target population (e.g., mental health trajectories among children, romantic relationship formation among young adults, and so forth). Even though this sort of research is concerned with population norms and not sources of heterogeneity (such as gender, ethnicity, and SES), the nature of the sociodemographic composition of the sample still influences the patterns found. In studies where heterogeneity due to sociodemographic factors is not the focus, variation introduced by sociodemographic factors can be thought of as “noise” or nuisance variance that, although tangential to the topic of study, should ideally be accounted for. When left unaccounted for, this noise at the very least introduces additional variance resulting in unnecessarily inflated standard errors and increases the likelihood of Type II errors. In addition to inflating the standard errors of parameter estimates, this noise can also bias or alter parameter estimates themselves. For example, consider two studies designed to examine developmental linkages between alcohol abuse and major depressive disorders in a target population. The two studies’ samples are the same except for their gender composition; Study A’s sample is 60% male, and Study B’s sample is 60% female. Because the association between alcohol abuse and major depressive disorders is markedly higher among females ( Kessler et al., 1997 ; Poulin, Hand, Boudreau, & Santor, 2005 ; Zilberman, Tavares, Blume, & el-Guebaly, 2003 ), the association between alcohol abuse and major depressive disorders will vary across the two studies. Specifically, the association is likely higher for Study B because its sample included a higher proportion of females. Thus, two studies focusing on population norms and examining the same question (in this case the association between alcohol abuse and major depressive disorders) could reach different conclusions due to differences in sample composition (in this case gender).
A popular option for dealing with the noise introduced by sociodemographic factors is to control for it statistically, but in many cases attempts to do so prove unsatisfactory. To be able to adequately control for a sociodemographic factor, there must be sufficient variance in it. Thus, if a study has only a small number of individuals in one or more ethnic groups, for example, then the ability of the study design to effectively control for ethnicity is reduced. Because gender, ethnicity, and (in some instances) SES are categorical variables, a common method for controlling for the noise they introduce is to use a series of dummy variables. However, this method of controlling for sociodemographic “noise” is problematic for two reasons. First, this technique only controls for sociodemographic differences in levels of variables; it does not account for sociodemographic differences in associations among variables (i.e., the technique assumes that the slope of the regression line in each sociodemographic group is the same). Thus, it only partially accounts for the noise attributed to sociodemographic differences. Second, this technique successfully controls for level differences across sociodemographic groups; however, if dummy-coding (as opposed to effects-coding) is used, it yields findings that generalize only to the reference group. To account properly for the noise introduced by sociodemographic variation, we suggest using one of the two following options. One option is to conduct preliminary analyses testing for sociodemographic differences (in both means and covariances) and if differences are found to conduct primary analyses separately for each sociodemographic group (i.e., multiple-group analyses). Option two is to recruit a sample that contains no variation in a sociodemographic factor (i.e., a data set limited to a single SES group; see Homogeneous Sampling above).
To determine the frequency with which the four different sampling methods are used in contemporary developmental science, we surveyed five years (2007–2011) of five high-profile developmental journals. The five journals included two that generally focus on abnormal development, ( Journal of the American Academy of Child and Adolescent Psychiatry (JAACAP) and Development and Psychopathology ), two that generally focus on normative development ( Developmental Psychology and Child Development ), and one that focuses on experimental studies related to development ( Developmental Science ). As we have done previously, for purposes of illustration and the sake of simplicity we limit our focus to just one of the three sociodemographic factors – ethnicity. For all published articles that included sample descriptions, we recorded the country of data collection, the sample size, and the percentages of participants who were European American or White, Hispanic American or Latino, African American or Black, Asian American or Asian, and another ethnicity. We also noted if the study did not report ethnicity and if the study mentioned that the sample was nationally representative. If ethnicity was reported on multiple groups or in multiple studies within an article, each sample was documented separately. Meta-analyses, reviews, commentaries, and editorials were ignored unless they presented new data. Because we were only interested in documenting the sampling distributions of U.S. samples for this illustration, we excluded any non-U.S. samples from our analysis. Any article that did not report the sample distribution of ethnicity, or that did not provide enough information to evaluate the ethnic distribution (e.g., “most participants were European American”), was coded as “not reported.”
Population-based probability sampling was coded if the authors reported that the study sample was nationally representative or if the distribution of ethnicities in the sample was ordered as it is in the population of the United States (with European Americans represented in higher proportions than Hispanic Americans, Hispanic Americans higher than African Americans, and African Americans higher than Asian Americans; Table 1 ). This decision constitutes a generous way to code population-based probability sampling because some studies that approximate the distribution of ethnicities in the United States were not truly representative. Quota sampling was coded when the authors mentioned that a particular ethnic group was oversampled (but not part of a probability sample), or when the sample included only 2 ethnic groups, each representing 45–55% of the sample, or when the sample included only 3 ethnic groups, each representing 28–38% of the sample. Homogeneous sampling was coded when the entire sample was comprised of a single ethnic group. Finally, convenience sampling was coded for the remainder of the samples.
Table 3 presents the categorization of samples in each of the five journals. The number of samples drawn from the United States ranged from 226 to 647 across the five journals. The range of sample sizes for each journal was wide, and the median sample sizes varied across journals, Kruskal-Wallis χ 2 (4, N =2104) = 293.96, p < .001. Post-hoc tests indicated that the median sample size for Developmental Science was smaller than all other journals, and the median sample size for Development and Psychopathology was larger than all other journals. Notably, depending on the journal, ethnicity was not adequately reported for one-quarter to more than two-thirds of the samples. As expected, the most prevalent type of sample in every journal was the convenience sample (78–88% of codable samples). Homogeneous sampling was the next most prevalent (5–19% of codable samples), followed by population-based probability samples (3–7% of codable samples) and quota samples (0–2% of codable samples). It could be that most samples for which ethnicity was not reported were convenience samples. If we assumed that the “not reported” samples were all convenience samples, then 88–98% (overall 91%) of samples would be convenience.
Characteristics of U.S. samples that fell into each sampling category from 2007 to 2011 in five high-profile developmental journals
Journal | 5-year impact factor | samples | Median sample size | Range of sample sizes | Percentage of samples drawn using each sampling strategy | ||||
---|---|---|---|---|---|---|---|---|---|
Population-based | Convenience | Quota | Homogeneous | Not reported | |||||
JAACAP | 5.976 | 443 | 126 | 10–16,128,828 | 3.8/7.2 | 45.6/85.6 | 1.1/2.1 | 2.7/5.1 | 46.7 |
D&P | 6.688 | 226 | 220 | 20–390,350 | 3.1/4.2 | 65.5/88.1 | 0.9/1.2 | 4.9/6.5 | 25.7 |
DP | 5.123 | 647 | 121 | 8–67,124 | 3.9/5.5 | 58.1/82.8 | 1.4/2.0 | 6.8/9.7 | 29.8 |
CD | 5.700 | 517 | 101 | 6–21,255 | 3.5/5.1 | 57.8/84.7 | 1.7/2.5 | 5.2/7.6 | 31.1 |
DS | 4.597 | 289 | 36 | 10–10,200 | 0.3/3.1 | 8.7/78.1 | 0.0/0.0 | 2.1/18.8 | 88.9 |
Total | -- | 2122 | 105 | 6–16,128,828 | 3.2/5.5 | 49.5/84.5 | 1.2/2.0 | 4.7/8.0 | 41.4 |
Note . The 5-year Impact Factor was abstracted from the 2011 Thomson Reuters Journal Citation Reports Database. JAACAP = Journal of the American Academy of Child and Adolescent Psychiatry . D&P = Development and Psychopathology . DP = Developmental Psychology . CD = Child Development . DS = Developmental Science .
Overall, the five journals were remarkably consistent in their representation of different types of samples. Developmental Science (the experimental journal) had a much larger percentage of samples where ethnicity was not reported adequately, χ 2 (8, N =2122) = 330.66, p < .001, but otherwise the patterns were very similar across journals. We explored whether the type of journal (abnormal, normative, or experimental) was associated with the prevalence of different sampling strategies. When samples where ethnicity was not reported were excluded, the distribution of sampling strategies was similar across journal types, χ 2 (6, N =1243) = 9.85, p = .13 (compare percentages after the slash in Table 3 ).
The median sample sizes for the four sampling strategies differed, Kruskal-Wallis χ 2 (4, N =2104)=326.45, p < .001. Post-hoc tests indicated that the median sample size for population-based probability sampling (Median = 2,741) was larger than the median of all other sampling strategies except quota sampling (Median = 486). The median sample size for quota sampling was larger than the median sample size of homogeneous (Median = 152), convenience (Median = 153), and “not reported” samples (Median = 52). Finally, the median sample sizes for homogeneous and convenience samples were both larger than the median sample size for “not reported” samples, but the median sample sizes for homogeneous and convenience samples did not significantly differ. These results support the proposition that population-based probability sampling requires a relatively large sample size and that there is little difference in sample size for convenience and homogeneous samples.
The analysis of published articles reveals several inconsistencies and inadequacies in basic reporting of demographic characteristics of study samples in developmental science. Overall, 41.4% of articles published in five high-profile developmental science journals failed to report ethnicity at all or only reported that the sample was “predominantly White” or “about half minority”, which is not detailed enough to draw conclusions about the representativeness of the sample to a particular population. Many studies also grouped different prominent ethnicities into an “other” category. We recommend that published studies report the percentages of their sample that fall into (at least) the four major ethnic groups (White/European American, Hispanic/Latino American, Black/African American, and Asian American), but more detailed accounting of groups and subgroups is always preferred. Of course, if the study includes an oversampling of other ethnicities or ethnic subgroups, they should also be reported. Similarly, any other sociodemographic characteristics that would help to evaluate the study sample’s representativeness or generalizability (e.g., proportions of females and males, SES, education levels, and ages of participants) should be reported in detail.
A second related recommendation is that each published study explicitly report the population to which the study generalizes. Not all studies intend to represent the entire population of the United States, but in most cases authors do not explain what population their study sample is intended to represent. For example, knowing whether a study generalizes to the entire population of 8- to 10-year-old children in the United States or only to 8- to 10-year-olds in five public schools in Maine frames the interpretation and application of the study’s findings considerably. Furthermore, a study that was designed to represent public school children in Maine should have very different sociodemographics from a study designed to represent 8- to 10-year-olds in the entire United States.
Here we have recounted four common sampling strategies (population-based probability sampling, convenience sampling, quota sampling, and homogenous sampling) and evaluated each by five meaningful criteria. We find that by far the most common sampling strategy (convenience sampling) is the least desirable in terms of representativeness, generalizability, and noise. Convenience samples require less practical investment in recruitment costs and efforts, but this advantage does not offset its aforementioned consequential scientific disadvantages. We find population-based probability sampling the most desirable strategy, but potentially cost and resource prohibitive for many researchers. The sample size required to adequately represent all sociodemographic groups would be large and the recruitment costs and efforts for this method are consequently considerable. Furthermore, large samples often involve a trade-off with detailed measurement. For example, a small sample may allow a researcher to investigate a topic in greater depth, such as with open-ended questions or more extensive measurement. Larger samples may only allow for more blunt instruments or require even greater allocation of resources for detailed measurement. Whether homogeneous sampling or quota sampling is a better nonprobability option depends on the research question. For research questions focused on sociodemographic factors as a source of heterogeneity (e.g., the sociodemographic factor is a focal variable of interest in the study), quota sampling is the better choice. In this case, quota sampling allows the researcher to recruit sociodemographic groups (e.g., ethnic or SES groups) that are sizeable enough to address the research questions. For research questions not focused on sociodemographic factors as a source of heterogeneity (e.g., the sociodemographic factor is not a focal variable and would likely be used as a covariate), homogeneous samples are preferable because they eliminate sociodemographic noise in the sample that could cloud the study results. Collecting a homogeneous sample might marginally increase recruitment costs and efforts compared to convenience samples (especially if the targeted group is infrequently occurring in the population), but pays off in terms of reducing noise.
One concern many researchers have about homogeneous samples is the success in obtaining protocol approval by internal review boards or to secure funding. Many funding agencies require (or strongly recommend) inclusion of all major sociodemographic groups. For example, the NIH released guidelines about including women and minorities in clinical research in 1994 (revised in 2001; Hohmann & Parron, 1996 ; NIH Office of Extramural Research, 2001 ) which indicate that all grant applications are evaluated for the inclusion of sociodemographic groups, and if groups are omitted, a strong justification is required. As we have outlined, there are circumstances when a homogenous sample is the best available option (e.g., probability samples are too expensive and sociodemographic differences are of little interest). Including sociodemographic variation that cannot be properly addressed statistically (as in a small convenience sample) should not be considered preferable to a homogeneous sample. Researchers are encouraged to make principled theoretical and statistical arguments to support their choice of a better sampling strategy, even if that strategy excludes one or more sociodemographic groups.
Sampling is a necessary evil of developmental science. It is practically and feasibly impossible to conduct a true population study, and so scientific researchers are forced to resort to sampling. Insofar as scientists must sample, they are confronted with the question of defining the profile of their sample. There are different types of samples, and with respect to different scientific questions different sampling strategies have different advantages and disadvantages. Although they are clearly preferable in terms of their generalizability and power, probability samples are infrequently used because of the cost and effort required to plan and execute them. Instead, nonprobability samples are much more common, with faulty convenience samples making up the largest share. However, with only a small amount of extra effort and planning, researchers could recruit a quota or homogeneous sample, thereby improving the characteristics of their sample and study.
We thank A. Bradley, O. M. Haynes, P. Horn, A. Mahler, C. Padilla, and C. Yuen. Supported by the Intramural Research Program of the NIH, NICHD.
1 This is only one example for illustrative purposes. The sample size needed for different analyses should always be estimated using an a priori power analysis prior to selecting a sample.
2 A sample large enough to yield sufficient power and to reasonably account for other confounds via random sampling.
Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.
Home » Purposive Sampling – Methods, Types and Examples
Table of Contents
Definition:
Purposive sampling is a non-probability sampling technique used in research to select individuals or groups of individuals that meet specific criteria relevant to the research question or objective.
This sampling technique is also known as judgmental sampling or selective sampling, and it is often used when the population being studied is too small, too difficult to access, or too heterogeneous to use probability sampling methods.
Purposive Sampling Methods are as follows:
Here are the general steps involved in conducting purposive sampling:
Here are some examples of how purposive sampling might be used in research:
Purposive sampling has a wide range of applications across different fields of research. Here are some examples of how purposive sampling can be used:
The purpose of purposive sampling is to select participants based on specific criteria relevant to the research question or objectives. Unlike probability sampling techniques, which rely on random selection to ensure representativeness, purposive sampling allows researchers to select participants who are most relevant to their research question or objectives.
Purposive sampling is often used when the population of interest is rare, hard to reach, or has specific characteristics that are important to the research question. By selecting participants who meet specific criteria, researchers can gather valuable insights that can help inform their research.
The ultimate goal of purposive sampling is to increase the validity and reliability of research findings by selecting participants who are most relevant to the research question or objectives. This can help researchers to make more accurate inferences about the population of interest and to develop more effective interventions or solutions based on their findings.
Purposive sampling is appropriate when researchers need to select participants who meet specific criteria relevant to their research question or objectives. Here are some situations where purposive sampling might be appropriate:
Purposive sampling has several characteristics that distinguish it from other sampling methods:
Purposive sampling has several advantages over other sampling methods:
Some Disadvantages of Purposive Sampling are as follows:
Researcher, Academic Writer, Web developer
Purposive sampling, also known as judgmental , selective or subjective sampling, is a type of non-probability sampling technique . Non-probability sampling focuses on sampling techniques where the units that are investigated are based on the judgement of the researcher [see our articles: Non-probability sampling to learn more about non-probability sampling, and Sampling: The basics , for an introduction to terms such as units , cases and sampling ]. There are a number of different types of purposive sampling, each with different goals. This article explains (a) what purposive sampling is, (b) the eight of the different types of purposive sampling, (c) how to create a purposive sample, and (d) the broad advantages and disadvantages of purposive sampling.
Types of purposive sampling, advantages and disadvantages of purposive sampling.
Purposive sampling represents a group of different non-probability sampling techniques . Also known as judgmental , selective or subjective sampling, purposive sampling relies on the judgement of the researcher when it comes to selecting the units (e.g., people, cases/organisations, events, pieces of data) that are to be studied. Usually, the sample being investigated is quite small, especially when compared with probability sampling techniques .
Unlike the various sampling techniques that can be used under probability sampling (e.g., simple random sampling, stratified random sampling, etc.), the goal of purposive sampling is not to randomly select units from a population to create a sample with the intention of making generalisations (i.e., statistical inferences ) from that sample to the population of interest [see the article: Probability sampling ]. This is the general intent of research that is guided by a quantitative research design .
The main goal of purposive sampling is to focus on particular characteristics of a population that are of interest, which will best enable you to answer your research questions. The sample being studied is not representative of the population, but for researchers pursuing qualitative or mixed methods research designs , this is not considered to be a weakness. Rather, it is a choice, the purpose of which varies depending on the type of purposing sampling technique that is used. For example, in homogeneous sampling , units are selected based on their having similar characteristics because such characteristics are of particular interested to the researcher. By contrast, critical case sampling is frequently used in exploratory , qualitative research in order to assess whether the phenomenon of interest even exists (amongst other reasons).
During the course of a qualitative or mixed methods research design , more than one type of purposive sampling technique may be used. For example, critical case sampling may be used to investigate whether a phenomenon is worth investigating further, before adopting a maximum variation sampling technique is used to develop a wider picture of the phenomenon. We explain the different goals of these types of purposive sampling technique in the next section.
There are a wide range of purposive sampling techniques that you can use (see Patton, 1990, 2002; Kuzel, 1999, for a complete list). Each of these types of purposive sampling technique is discussed in turn:
Homogeneous sampling, typical case sampling, extreme (or deviant) case sampling, critical case sampling, total population sampling, expert sampling.
Maximum variation sampling, also known as heterogeneous sampling , is a purposive sampling technique used to capture a wide range of perspectives relating to the thing that you are interested in studying; that is, maximum variation sampling is a search for variation in perspectives, ranging from those conditions that are view to be typical through to those that are more extreme in nature. By conditions , we mean the units (i.e., people, cases/organisations, events, pieces of data) that are of interest to the researcher. These units may exhibit a wide range of attributes, behaviours, experiences, incidents, qualities, situations, and so forth. The basic principle behind maximum variation sampling is to gain greater insights into a phenomenon by looking at it from all angles. This can often help the researcher to identify common themes that are evident across the sample.
Homogeneous sampling is a purposive sampling technique that aims to achieve a homogeneous sample; that is, a sample whose units (e.g., people, cases, etc.) share the same (or very similar) characteristics or traits (e.g., a group of people that are similar in terms of age, gender, background, occupation, etc.). In this respect, homogeneous sampling is the opposite of maximum variation sampling . A homogeneous sample is often chosen when the research question that is being address is specific to the characteristics of the particular group of interest, which is subsequently examined in detail.
Typical case sampling is a purposive sampling technique used when you are interested in the normality/typicality of the units (e.g., people, cases, events, settings/contexts, places/sites) you are interested, because they are normal/typical . The word typical does not mean that the sample is representative in the sense of probability sampling (i.e., that the sample shares the same/similar characteristics of the population being studied). Rather, the word typical means that the researcher has the ability to compare the findings from a study using typical case sampling with other similar samples (i.e., comparing samples, not generalising a sample to a population). Therefore, with typical case sampling, you cannot use the sample to make generalisations to a population, but the sample could be illustrative of other similar samples. Whilst typical case sampling can be used exclusively, it may also follow another type of purposive sampling technique, such as maximum variation sampling, which can help to act as an exploratory sampling strategy to identify the typical cases that are subsequently selected.
Extreme (or deviant) case sampling is a type of purposive sampling that is used to focus on cases that are special or unusual , typically in the sense that the cases highlight notable outcomes , failures or successes . These extreme (or deviant) cases are useful because they often provide significant insight into a particular phenomenon, which can act as lessons (or cases of best practice) that guide future research and practice. In some cases, extreme (or deviant) case sampling is thought to reflect the purest form of insight into the phenomenon being studied.
Critical case sampling is a type of purposive sampling technique that is particularly useful in exploratory qualitative research, research with limited resources , as well as research where a single case (or small number of cases) can be decisive in explaining the phenomenon of interest. It is this decisive aspect of critical case sampling that is arguably the most important. To know if a case is decisive, think about the following statements: ?If it happens there, it will happen anywhere?; or ?if it doesn?t happen there, it won?t happen anywhere?; and ?If that group is having problems, then we can be sure all the groups are having problems? (Patton, 202, p.237). Whilst such critical cases should not be used to make statistical generalisations , it can be argued that they can help in making logical generalisations . However, such logical generalisations should be made carefully.
Total population sampling is a type of purposive sampling technique where you choose to examine the entire population (i.e., the total population ) that have a particular set of characteristics (e.g., specific experience, knowledge, skills, exposure to an event, etc.). In such cases, the entire population is often chosen because the size of the population that has the particular set of characteristics that you are interest in is very small. Therefore, if a small number of units (i.e., people, cases/organisations, etc.) were not included in the sample that is investigated, it may be felt that a significant piece of the puzzle was missing [see the article, Total population sampling , to learn more].
Expert sampling is a type of purposive sampling technique that is used when your research needs to glean knowledge from individuals that have particular expertise . This expertise may be required during the exploratory phase of qualitative research, highlighting potential new areas of interest or opening doors to other participants. Alternately, the particular expertise that is being investigated may form the basis of your research, requiring a focus only on individuals with such specific expertise. Expert sampling is particularly useful where there is a lack of empirical evidence in an area and high levels of uncertainty, as well as situations where it may take a long period of time before the findings from research can be uncovered. Therefore, expert sampling is a cornerstone of a research design known as expert elicitation .
Whilst each of the different types of purposive sampling has its own advantages and disadvantages, there are some broad advantages and disadvantages to using purposive sampling, which are discussed below.
There are a wide range of qualitative research designs that researchers can draw on. Achieving the goals of such qualitative research designs requires different types of sampling strategy and sampling technique . One of the major benefits of purposive sampling is the wide range of sampling techniques that can be used across such qualitative research designs; purposive sampling techniques that range from homogeneous sampling through to critical case sampling , expert sampling , and more.
Whilst the various purposive sampling techniques each have different goals, they can provide researchers with the justification to make generalisations from the sample that is being studied, whether such generalisations are theoretical , analytic and/or logical in nature. However, since each of these types of purposive sampling differs in terms of the nature and ability to make generalisations, you should read the articles on each of these purposive sampling techniques to understand their relative advantages.
Qualitative research designs can involve multiple phases, with each phase building on the previous one. In such instances, different types of sampling technique may be required at each phase. Purposive sampling is useful in these instances because it provides a wide range of non-probability sampling techniques for the researcher to draw on. For example, critical case sampling may be used to investigate whether a phenomenon is worth investigating further, before adopting an expert sampling approach to examine specific issues further.
Purposive samples, irrespective of the type of purposive sampling used, can be highly prone to researcher bias . The idea that a purposive sample has been created based on the judgement of the researcher is not a good defence when it comes to alleviating possible researcher biases, especially when compared with probability sampling techniques that are designed to reduce such biases. However, this judgemental, subjective component of purpose sampling is only a major disadvantage when such judgements are ill-conceived or poorly considered ; that is, where judgements have not been based on clear criteria, whether a theoretical framework, expert elicitation, or some other accepted criteria.
The subjectivity and non-probability based nature of unit selection (i.e., selecting people, cases/organisations, etc.) in purposive sampling means that it can be difficult to defend the representativeness of the sample. In other words, it can be difficult to convince the reader that the judgement you used to select units to study was appropriate. For this reason, it can also be difficult to convince the reader that research using purposive sampling achieved theoretical/analytic/logical generalisation . After all, if different units had been selected, would the results and any generalisations have been the same?
Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.
Nature Medicine ( 2024 ) Cite this article
60 Altmetric
Metrics details
The role of antithrombotic therapy in the prevention of ischemic stroke after non-cardiac surgery is unclear. In this study, we tested the hypothesis that the association of new-onset postoperative atrial fibrillation (POAF) on ischemic stroke can be mitigated by postoperative oral anticoagulation therapy. Of 251,837 adult patients (155,111 female (61.6%) and 96,726 male (38.4%)) who underwent non-cardiac surgical procedures at two sites, POAF was detected in 4,538 (1.8%) patients. The occurrence of POAF was associated with increased 1-year ischemic stroke risk (3.6% versus 2.3%; adjusted risk ratio (RRadj) = 1.60 (95% confidence interval (CI): 1.37–1.87), P < 0.001). In patients with POAF, the risk of developing stroke attributable to POAF was 1.81 (95% CI: 1.44–2.28; P < 0.001) without oral anticoagulation, whereas, in patients treated with anticoagulation, no significant association was observed between POAF and stroke (RRadj = 1.04 (95% CI: 0.71–1.51), P = 0.847, P for interaction = 0.013). Furthermore, we derived and validated a computational model for the prediction of POAF after non-cardiac surgery based on demographics, comorbidities and procedural risk. These findings suggest that POAF is predictable and associated with an increased risk of postoperative ischemic stroke in patients who do not receive postoperative anticoagulation.
This is a preview of subscription content, access via your institution
Access Nature and 54 other Nature Portfolio journals
Get Nature+, our best-value online-access subscription
24,99 € / 30 days
cancel any time
Subscribe to this journal
Receive 12 print issues and online access
195,33 € per year
only 16,28 € per issue
Buy this article
Prices may be subject to local taxes which are calculated during checkout
Data availability.
Access to anonymized patient data is subject to a data-sharing agreement and protocol approval from the institutional review board committee. The study-specific analyzable dataset is, therefore, not publicly available. Due to the sensitive nature of the patient data collected for this study, requests to access the dataset may be directed to the corresponding author, M.E., at [email protected] (14-d response time).
Codes for inclusion criteria, baseline characteristics and outcome events were developed based on the International Classification of Diseases code classification, Current Procedural Terminology or data from electronic health records, unless otherwise specified. These are described in the supplementary materials of this manuscript. Statistical analyses were conducted using Stata (version 17), GraphPad Prism (version 9) and G*Power (version 3.1.9.4). Detailed information about the codes and packages used in the analyses can be found in the Supplementary Information . For additional clarification of these codes, contact the corresponding author at [email protected] (14-d response time).
Gialdini, G. et al. Perioperative atrial fibrillation and the long-term risk of ischemic stroke. JAMA 312 , 616–622 (2014).
Article CAS PubMed PubMed Central Google Scholar
Butt, J. H. et al. Risk of thromboembolism associated with atrial fibrillation following noncardiac surgery. J. Am. Coll. Cardiol. 72 , 2027–2036 (2018).
Article PubMed Google Scholar
Bhave, P. D., Goldman, L. E., Vittinghoff, E., Maselli, J. & Auerbach, A. Incidence, predictors, and outcomes associated with postoperative atrial fibrillation after major noncardiac surgery. Am. Heart J. 164 , 918–924 (2012).
Article PubMed PubMed Central Google Scholar
Day, R. W. et al. Incidence and impact of postoperative atrial fibrillation after minimally invasive esophagectomy. Dis. Esophagus 29 , 583–588 (2016).
Article CAS PubMed Google Scholar
Christians, K. K., Wu, B., Quebbeman, E. J. & Brasel, K. J. Postoperative atrial fibrillation in noncardiothoracic surgical patients. Am. J. Surg. 182 , 713–715 (2001).
Joshi, K. K., Tiru, M., Chin, T., Fox, M. T. & Stefan, M. S. Postoperative atrial fibrillation in patients undergoing non-cardiac non-thoracic surgery: a practical approach for the hospitalist. Hosp. Pract. (1995) 43 , 235–244 (2015).
McIntyre, W. F. et al. Incidence and recurrence of new-onset atrial fibrillation detected during hospitalization for non-cardiac surgery: a systematic review and meta-analysis. Can. J. Anaesth. 68 , 1045–1056 (2021).
Jokinen, J. D. V. et al. Wireless single-lead ECG monitoring to detect new-onset postoperative atrial fibrillation in patients after major noncardiac surgery: a prospective observational study. Anesth. Analg. 135 , 100–109 (2022).
CAS PubMed Google Scholar
McIntyre, W. F. et al. Atrial fibrillation recurrence in patients with transient new-onset atrial fibrillation detected during hospitalization for noncardiac surgery or medical illness: a matched cohort study. Ann. Intern. Med. 176 , 1299–1307 (2023).
Fragão-Marques, M. et al. Impact of oral anticoagulation therapy on postoperative atrial fibrillation outcomes: a systematic review and meta-analysis. Thromb. J. 19 , 89 (2021).
Yao, R. J. R. et al. Anticoagulation management of postoperative atrial fibrillation after cardiac surgery: a systematic review. J. Card. Surg. 36 , 2081–2094 (2021).
Neves, I. A. et al. Anticoagulation therapy in patients with post-operative atrial fibrillation: systematic review with meta-analysis. Vasc. Pharm. 142 , 106929 (2022).
Article CAS Google Scholar
Elharram, M. et al. Anticoagulant use and the risk of thromboembolism and bleeding in postoperative atrial fibrillation after noncardiac surgery. Can. J. Cardiol. 37 , 391–399 (2021).
Hindricks, G. et al. 2020 ESC guidelines for the diagnosis and management of atrial fibrillation developed in collaboration with the European Association for Cardio-Thoracic Surgery (EACTS). Eur. Heart J. 42 , 373–498 (2021).
Joglar, J. A. et al. 2023 ACC/AHA/ACCP/HRS guideline for the diagnosis and management of atrial fibrillation: a report of the American College of Cardiology/American Heart Association Joint Committee on Clinical Practice Guidelines. Circulation 149 , e1–e156 (2024).
Feigin, V. L. et al. Global, regional, and national burden of neurological disorders, 1990–2016: a systematic analysis for the Global Burden of Disease Study 2016. Lancet Neurol. 18 , 459–480 (2019).
Article Google Scholar
Meersch, M. et al. Effect of intraoperative handovers of anesthesia care on mortality, readmission, or postoperative complications among adults. JAMA 327 , 2403–2412 (2022).
Lang, R. M. et al. Recommendations for cardiac chamber quantification by echocardiography in adults: an update from the American Society of Echocardiography and the European Association of Cardiovascular Imaging. J. Am. Soc. Echocardiogr. 28 , 1–39 (2015).
Stronati, G. et al. Derivation and validation of a clinical score for predicting postoperative atrial fibrillation in noncardiac elective surgery (the HART score). Am. J. Cardiol. 170 , 56–62 (2022).
Patel, A. Y., Eagle, K. A. & Vaishnava, P. Cardiac risk of noncardiac surgery. J. Am. Coll. Cardiol. 66 , 2140–2148 (2015).
Bruins, P. et al. Activation of the complement system during and after cardiopulmonary bypass surgery. Circulation 96 , 3542–3548 (1997).
Amar, D., Zhang, H., Miodownik, S. & Kadish, A. H. Competing autonomic mechanisms precedethe onset of postoperative atrial fibrillation. J. Am. Coll. Cardiol. 42 , 1262–1268 (2003).
Prince-Wright, L. H. et al. Postoperative atrial fibrillation following non-cardiac surgery: predictors and risk of mortality. Am. J. Surg. 224 , 1062–1067 (2022).
Lee, S.-H. et al. New-onset atrial fibrillation predicts long-term newly developed atrial fibrillation after coronary artery bypass graft. Am. Heart J. 167 , 593–600 (2014).
Lin, M. H. et al. Perioperative/postoperative atrial fibrillation and risk of subsequent stroke and/or mortality: a meta-analysis. Stroke 50 , 1364–1371 (2019).
Huynh, J. T. et al. Association between perioperative atrial fibrillation and long-term risks of stroke and death in noncardiac surgery: systematic review and meta-analysis. CJC Open 3 , 666–674 (2021).
Siontis, K. C. et al. Associations of atrial fibrillation after noncardiac surgery with stroke, subsequent arrhythmia, and death. Ann. Intern. Med. 175 , 1065–1072 (2022).
Hindricks, G. et al. Corrigendum to: 2020 ESC Guidelines for the diagnosis and management of atrial fibrillation developed in collaboration with the European Association for Cardio-Thoracic Surgery (EACTS): the task force for the diagnosis and management of atrial fibrillation of the European Society of Cardiology (ESC) developed with the special contribution of the European Heart Rhythm Association (EHRA) of the ESC. Eur. Heart J. 42 , 4194 (2021).
Devereaux, P. J. et al. Effects of extended-release metoprolol succinate in patients undergoing non-cardiac surgery (POISE trial): a randomised controlled trial. Lancet 371 , 1839–1847 (2008).
Conen, D. et al. Effect of colchicine on perioperative atrial fibrillation and myocardial injury after non-cardiac surgery in patients undergoing major thoracic surgery (COP-AF): an international randomised trial. Lancet 402 , 1627–1635 (2023).
Ha, A. C. T. et al. Effect of continuous electrocardiogram monitoring on detection of undiagnosed atrial fibrillation after hospitalization for cardiac surgery. JAMA Netw. Open 4 , e2121867 (2021).
El‐Chami, M. F. et al. Management of new‐onset postoperative atrial fibrillation utilizing insertable cardiac monitor technology to observe recurrence of AF (MONITOR‐AF). Pacing Clin. Electrophysiol. 39 , 1083–1089 (2016).
Yao, X. et al. Artificial intelligence–enabled electrocardiograms for identification of patients with low ejection fraction: a pragmatic, randomized clinical trial. Nat. Med. 27 , 815–819 (2021).
Attia, Z. I. et al. Screening for cardiac contractile dysfunction using an artificial intelligence–enabled electrocardiogram. Nat. Med. 25 , 70–74 (2019).
Friedrich, S. et al. Patent foramen ovale and long-term risk of ischaemic stroke after surgery. Eur. Heart J. 40 , 914–924 (2019).
Ng, P. Y. et al. Association of preoperatively diagnosed patent foramenovale with perioperative ischemic stroke. JAMA 319 , 452–462 (2018).
Marcucci, M., Chan, M. T. V., Smith, E. E., Absalom, A. R. & Devereaux, P. J. Prevention of perioperative stroke in patients undergoing non-cardiac surgery. Lancet Neurol. 22 , 946–958 (2023).
Ke Wang, M. et al. Anticoagulation use in perioperative atrial fibrillation after noncardiac surgery: a systematic review and meta-analysis. Swiss Med. Wkly 153 , 40056 (2023).
Kotalczyk, A., Mazurek, M., Kalarus, Z., Potpara, T. S. & Lip, G. Y. H. Stroke prevention strategies in high-risk patients with atrial fibrillation. Nat. Rev. Cardiol. 18 , 276–290 (2021).
Spyropoulos, A. C. & Douketis, J. D. How I treat anticoagulated patients undergoing an elective procedure or surgery. Blood 120 , 2954–2962 (2012).
Vandenbroucke, J. P. et al. The Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) statement: explanation and elaboration. PLoS Med. 4 , e296 (2007).
Benchimol, E. I. et al. The REporting of studies Conducted using Observational Routinely-collected health Data (RECORD) statement. PLoS Med. 12 , e1001885 (2015).
Schneeweiss, S. et al. Graphical depiction of longitudinal study designs in health care databases. Ann. Intern. Med. 170 , 398–406 (2019).
Siontis, K. C. et al. Association of new-onset atrial fibrillation after noncardiac surgery with subsequent stroke and transient ischemic attack. JAMA 324 , 871–878 (2020).
Brott, T. et al. Measurements of acute cerebral infarction: a clinical examination scale. Stroke 20 , 864–870 (1989).
Shay, D. et al. Preoperative heart failure treatment prevents postoperative cardiac complications in patients with lower risk: a retrospective cohort study. Ann. Surg. 277 , e33–e39 (2023).
Ahrens, E. et al. Prevalence and association of non-medical cannabis use with post-procedural healthcare utilisation in patients undergoing surgery or interventional procedures: a retrospective cohort study. EClinicalMedicine 57 , 101831 (2023).
Scheffenbichler, F. T. et al. Effects of the level and duration of mobilization therapy in the surgical ICU on the loss of the ability to live independently: an international prospective cohort study. Crit. Care Med. 49 , e247–e257 (2021).
Schaefer, M. S. et al. What factors predict adverse discharge disposition in patients older than 60 years undergoing lower-extremity surgery? The Adverse Discharge in Older Patients after Lower-extremity Surgery (ADELES) risk score. Clin. Orthop. Relat. Res. 479 , 546–547 (2021).
Aalen, O. O. & Johansen, S. An empirical transition matrix for non-homogeneous Markov chains based on censored observations. Scand. J. Stat. 5 , 141–150 (1978).
Google Scholar
Kurth, T. et al. Results of multivariable logistic regression, propensity matching, propensity adjustment, and propensity-based weighting under conditions of nonuniform effect. Am. J. Epidemiol. 163 , 262–270 (2006).
Webster‐Clark, M. et al. Using propensity scores to estimate effects of treatment initiation decisions: state of the science. Stat. Med. 40 , 1718–1735 (2021).
Lip, G. Y. H., Nieuwlaat, R., Pisters, R., Lane, D. A. & Crijns, H. J. G. M. Refining clinical risk stratification for predicting stroke and thromboembolism in atrial fibrillation using a novel risk factor-based approach. Chest 137 , 263–272 (2010).
Yaggi, H. & Mohsenin, V. Obstructive sleep apnoea and stroke. Lancet Neurol. 3 , 333–342 (2004).
Barnes, M. E. et al. Left atrial volume in the prediction of first ischemic stroke in an elderly cohort without atrial fibrillation. Mayo Clin. Proc. 79 , 1008–1014 (2004).
Bouzas-Mosquera, A. et al. Left atrial size and risk for all-cause mortality and ischemic stroke. Can. Med. Assoc. J. 183 , E657–E664 (2011).
Mannina, C. et al. Association of left atrial strain with ischemic stroke risk in older adults. JAMA Cardiol. 8 , 317–325 (2023).
Zou, G. A modified Poisson regression approach to prospective studies with binary data. Am. J. Epidemiol. 159 , 702–706 (2004).
Yelland, L. N., Salter, A. B. & Ryan, P. Performance of the modified Poisson regression approach for estimating relative risks from clustered prospective data. Am. J. Epidemiol. 174 , 984–992 (2011).
Pedroza, C. & Truong, V. T. T. Estimating relative risks in multicenter studies with a small number of centers—which methods to use? A simulation study. Trials 18 , 512 (2017).
Mandrekar, J. N. Receiver operating characteristic curve in diagnostic test assessment. J. Thorac. Oncol. 5 , 1315–1316 (2010).
Faul, F., Erdfelder, E., Lang, A.-G. & Buchner, A. G*Power 3: a flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behav. Res. Methods 39 , 175–191 (2007).
Download references
The authors declare no specific grants for this research from any funding agency in the public, commercial or not-for-profit sector.
These authors contributed equally: Omid Azimaraghi, Maíra I. Rudolph.
Department of Anesthesiology, Montefiore Medical Center, Albert Einstein College of Medicine, Bronx, NY, USA
Omid Azimaraghi, Maíra I. Rudolph, Karuna Wongtangman, Felix Borngaesser, Maya Doehne, Annika Eyth, Christopher Tam, William J. Sauer, Michael E. Kiyatkin, Ibraheem M. Karaye, Ling Zhang, Carina P. Himes, Aline M. Grimm, Christian Mpody, Aiman Suleiman, Annika Bald, Jay J. Im, Jonathan Leff, Can M. Luedeke, Timothy Pulverenti, Tina Ramishvili, Flora T. Scheffenbichler, Sujatha Ramachandran & Matthias Eikermann
Department of Anesthesiology and Intensive Care Medicine, University of Cologne, Faculty of Medicine and University Hospital Cologne, Cologne, Germany
Maíra I. Rudolph
Department of Anesthesiology, Faculty of Medicine, Siriraj Hospital, Mahidol University, Bangkok, Thailand
Karuna Wongtangman
Carl von Ossietzky Universität Oldenburg, University Clinic for Anesthesiology, Intensive Care, Emergency Medicine, and Pain Therapy, Klinikum Oldenburg AöR, Oldenburg, Germany
Felix Borngaesser & Simon T. Schaefer
Critical Care Medicine Unit, School of Clinical Medicine, The University of Hong Kong, Pokfulam, Hong Kong SAR, China
Pauline Y. Ng
Department of Adult Intensive Care, Queen Mary Hospital, Hong Kong SAR, China
Center for Anesthesia Research Excellence, Beth Israel Deaconess Medical Center, Harvard Medical School, Boston, MA, USA
Dario von Wedel, Maximilian S. Schaefer & Dario von Wedel
Department of Anesthesia, Critical Care, and Pain Medicine, Beth Israel Deaconess Medical Center, Harvard Medical School, Boston, MA, USA
Institute of Medical Informatics, Charité University Medicine Berlin, Berlin, Germany
Dario von Wedel & Dario von Wedel
Department of Cardiology, Montefiore Medical Center, Albert Einstein College of Medicine, Bronx, NY, USA
Fengwei Zou, Luigi Di Biase, Mario J. Garcia, Luigi Di Biase & M. Azeem Latib
Department of Anesthesia, Critical Care and Pain Medicine, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
Timothy T. Houle
Department of Population Health, Hofstra University, Hempstead, NY, USA
Ibraheem M. Karaye
Department of Anesthesiology, Duesseldorf University Hospital, Duesseldorf, Germany
Maximilian S. Schaefer
Department of Anesthesiology & Pain Medicine, Nationwide Children’s Hospital and The Ohio State University, Columbus, OH, USA
Olubukola O. Nafiu
Department of Cardiovascular and Thoracic Surgery, Montefiore Medical Center, Albert Einstein College of Medicine, Bronx, New York, NY, USA
Brendon M. Stiles, Joseph J. DeRose & Stephen J. Forest
Mount Sinai Fuster Heart Hospital, Icahn School of Medicine at Mount Sinai, New York, NY, USA
Deepak L. Bhatt
Klinik für Anästhesiologie und Intensivmedizin, Universität Duisburg-Essen, Essen, Germany
Matthias Eikermann
Department of Medicine, Montefiore Medical Center and Albert Einstein College of Medicine, Bronx, NY, USA
Fran Ganz-Lord
Department of Anesthesiology and Intensive Care Medicine, Ulm University, Ulm, Germany
Flora T. Scheffenbichler
You can also search for this author in PubMed Google Scholar
O.A., L.Z., M.S.S. and M.E. had access to the data. M.E. is the guarantor of the manuscript and takes full responsibility for the integrity of the data and the accuracy of the data analysis. Concept and design: O.A., M.K., W.J.S., M.S.S., M.G., D.L.B., L.D. and M.E. Acquisition, analysis or interpretation of the data: O.A., T.T.H., L.Z., M.D., M.I.R., F.B. and D.W. Drafting of the manuscript: O.A., M.I.R., F.Z., A.E., D.L.B. and M.E. Critical revision of the manuscript: all authors. Statistical analysis: O.A., I.M.K., M.D., M.I.R., F.B. and K.W. D.L.B. and M.E. supervised critical revisions.
Correspondence to Matthias Eikermann .
Competing interests.
M.E. receives funding from the National Institutes of Health (NIH) (R01AG065554 and R01HL132887) that does not pertain to this manuscript. He holds two patents for acyclic curcubiturils for reversal of drugs of abuse and neuromuscular blocking agents (patent numbers 9956229 and 9469648). He is a member of the associated editorial board for the British Journal of Anaesthesia . D.L.B. discloses the following relationships. Advisory Board: Angiowave, Bayer, Boehringer Ingelheim, CellProthera, Cereno Scientific, Elsevier Practice Update Cardiology, High Enroll, Janssen, Level Ex, McKinsey, Medscape Cardiology, Merck, MyoKardia, NirvaMed, Novo Nordisk, PhaseBio, PLx Pharma and Stasys; Board of Directors: American Heart Association New York City, Angiowave (stock options), Bristol Myers Squibb (stock), DRS.LINQ (stock options) and High Enroll (stock); Consultant: Broadview Ventures, Hims, SFJ and Youngene; Data Monitoring Committee: Acesion Pharma, Assistance Publique-Hôpitaux de Paris, Baim Institute for Clinical Research (formerly Harvard Clinical Research Institute, for the PORTICO trial, funded by St. Jude Medical, now Abbott), Boston Scientific (chair, PEITHO trial), Cleveland Clinic, Contego Medical (chair, PERFORMANCE 2), Duke Clinical Research Institute, Mayo Clinic, Mount Sinai School of Medicine (for the ENVISAGE trial, funded by Daiichi Sankyo; for the ABILITY-DM trial, funded by Concept Medical; for ALLAY-HF, funded by Alleviant Medical), Novartis, Population Health Research Institute and Rutgers University (for the NIH-funded MINT Trial); Honoraria: American College of Cardiology (ACC) (senior associate editor, Clinical Trials and News ; chair, ACC Accreditation Oversight Committee), Arnold and Porter law firm (work related to Sanofi/Bristol Myers Squibb clopidogrel litigation), Baim Institute for Clinical Research (formerly Harvard Clinical Research Institute; RE-DUAL PCI clinical trial steering committee, funded by Boehringer Ingelheim; AEGIS-II executive committee, funded by CSL Behring), Belvoir Publications (editor in chief, Harvard Heart Letter ), Canadian Medical and Surgical Knowledge Translation Research Group (clinical trial steering committees), CSL Behring (American Heart Association lecture), Cowen and Company, Duke Clinical Research Institute (clinical trial steering committees, including for the PRONOUNCE trial, funded by Ferring Pharmaceuticals), HMP Global (editor in chief, Journal of Invasive Cardiology ), Journal of the American College of Cardiology (guest editor and associate editor), K2P (co-chair, interdisciplinary curriculum), Level Ex, Medtelligence/ReachMD (CME steering committees), MJH Life Sciences, Oakstone CME (course director, Comprehensive Review of Interventional Cardiology), Piper Sandler, Population Health Research Institute (for the COMPASS operations committee, publications committee and steering committee and USA national co-leader, funded by Bayer), WebMD (CME steering committees) and Wiley (steering committee); Other: Clinical Cardiology (deputy editor); Patent: sotagliflozin (named on a patent for sotagliflozin assigned to Brigham and Women's Hospital, assigned to Lexicon; neither D.L.B. nor Brigham and Women's Hospital receive any income from this patent); Research Funding: Abbott, Acesion Pharma, Afimmune, Aker Biomarine, Alnylam, Amarin, Amgen, AstraZeneca, Bayer, Beren, Boehringer Ingelheim, Boston Scientific, Bristol Myers Squibb, Cardax, CellProthera, Cereno Scientific, Chiesi, CinCor, Cleerly, CSL Behring, Eisai, Ethicon, Faraday Pharmaceuticals, Ferring Pharmaceuticals, Forest Laboratories, Fractyl, Garmin, HLS Therapeutics, Idorsia, Ironwood, Ischemix, Janssen, Javelin, Lexicon, Eli Lilly, Medtronic, Merck, Moderna, MyoKardia, NirvaMed, Novartis, Novo Nordisk, Otsuka, Owkin, Pfizer, PhaseBio, PLx Pharma, Recardio, Regeneron, Reid Hoffman Foundation, Roche, Sanofi, Stasys, Synaptic, The Medicines Company, Youngene and 89Bio; Royalties: Elsevier (editor, Braunwald’s Heart Disease ); Site Co-Investigator: Abbott, Biotronik, Boston Scientific, CSI, Endotronix, St. Jude Medical (now Abbott), Philips, SpectraWAVE, Svelte and Vascular Solutions; Trustee: ACC; Unfunded Research: FlowCo. M.S.S. received funding for investigator-initiated studies from Merck & Co., which does not pertain to this manuscript. He is an associate editor for BMC Anesthesiology . He received honoraria for lectures from Fisher & Paykel Healthcare and Mindray Medical International Limited. He received an unrestricted philanthropic grant from Jeff and Judy Buzen. All other authors have no support from any organization for the submitted work, no financial relationships with any organizations that might have an interest in the submitted work and no other relationships or activities that could appear to have influenced the submitted work.
Peer review information.
Nature Medicine thanks Behnood Bikdeli, Jeffrey Weitz, William McIntyre and the other, anonymous, reviewer(s) for their contribution to the peer review of this work. Primary Handling Editor: Michael Basson, in collaboration with the Nature Medicine team.
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Extended data fig. 1 kaplan-meier survival estimates..
The figure illustrates the adjusted Kaplan-Meier survival estimates analysis conducted for the occurrence of stroke in patients experiencing new onset postoperative atrial fibrillation(red line) and patients who did not(blue line). The x-axis represents time after surgery in days, while the y-axis indicates the probability of being stroke-free (Hazard Ratio: 2.07 (95% CI 1.92-2.23), p < 0.001).
( A ) New-onset atrial fibrillation (POAF) after non-cardiac surgery is associated with postoperative complications, stroke, heart failure, and myocardial infarction. ( B ) For patients who were prescribed oral anticoagulants, the absolute risk difference was 0.34% in all patients and 0.68% in high-risk patients (upper quintile in computational prediction model [>11 points]). There was no significant absolute risk difference in patients who were prescribed oral anticoagulants. The images used were created by the authors using Procreate for iOS17 (Savage Interactive Pty Ltd., Hobart, Australia).
Graphical description of the study design, visualizing temporal anchors of exposure (new-onset postoperative atrial fibrillation), outcome/follow-up (ischemic stroke/transient ischemic attack), covariates (baseline demographics, procedure related factors, comorbidities and preexisting medication) and primary effect modifier (oral anticoagulation). The cohort entry date (CED), serving as the primary anchor date for patients entering the study analysis, was the date of the patients' index surgery (Day 0). We excluded patients with preexisting atrial fibrillation, age below 18 years, ASA physical status greater than 4 or missing data for exposure, outcome or covariates. Covariates were assessed during a window extending one year prior to the CED if not otherwise specified [-365,-1]. The time window for the primary exposure - postoperative atrial fibrillation (POAF) - was between day 0 and the end of day post-CED. Oral anticoagulation prescriptions were included between postoperative day 1 and the end of day 365. The outcome ischemic stroke was assessed between day 31 and 365 after surgery. Abbreviations: y, years; ASA, American Society of Anesthesiologists; POAF, new-onset postoperative atrial fibrillation.
Supplementary information.
Supplementary Figs. 1–7, Supplementary Tables 1–7 and Supplementary Notes 1 (STROBE checklist) and 2 (Statistical Analysis Plan).
Rights and permissions.
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
Reprints and permissions
Cite this article.
Azimaraghi, O., Rudolph, M.I., Wongtangman, K. et al. Role of anticoagulation therapy in modifying stroke risk associated with new-onset atrial fibrillation after non-cardiac surgery. Nat Med (2024). https://doi.org/10.1038/s41591-024-03206-0
Download citation
Received : 10 November 2023
Accepted : 19 July 2024
Published : 23 August 2024
DOI : https://doi.org/10.1038/s41591-024-03206-0
Anyone you share the following link with will be able to read this content:
Sorry, a shareable link is not currently available for this article.
Provided by the Springer Nature SharedIt content-sharing initiative
Sign up for the Nature Briefing newsletter — what matters in science, free to your inbox daily.
IMAGES
COMMENTS
In homogeneous convenience sampling researchers undertake to study (and therefore sample) a population that is homogeneous with respect to one or more sociodemographic factors (e.g., the overall population is composed of just Blacks or Whites). Thus, the target population (not just the sample studied) is a specific sociodemographic subgroup.
Homogenous populations are alike and heterogeneous populations are unalike. Homogenous means alike. Heterogenous means unalike or distinct from one another. Thus, a homogenous population has little variation. You could refer to a specific trait, such as hair color or you could refer to genetic diversity. For example, a population of humans that ...
Homogeneity of Variance. Homogeneity of variance (also called homoscedasticity) is used to describe a set of data that has the same variance. Visually, the data will have the same scatter on a scatter plot. If data does not have the same variance, it will show a heteroscedastic ("not the same") scatter pattern.
Items on an assembly line.Y might be the weight of the item. If there were no auxiliary variables, then the homogenous model would apply. If the assembly line is dedicated to the production of a single item, then Y might be expected to follow a fairly well-behaved distribution, as production and legal standards are usually such that the items produced have to be as alike as possible.
In the broad context of human populations - one of the central objects of gene-environment interaction (GEI) research and of this article - a homogeneous population is one that is believed to have, or that can be made to have, a uniform character, where all the constituents are of the same or similar nature, and are therefore recombinable ...
A part of population that repre sents it completely is known as sample. It means, the units, selected from the population as a sample, must represent all kind of characteristics of different ...
Simple populations surveys may start from the idea that responses will be homogeneous across the whole of a population. Assessing the homogeneity of the population would involve looking to see whether the responses of certain identifiable subpopulations differ from those of others. For example, car-owners may differ from non-car-owners, or ...
Introduction. Research methodology relies heavily on the precise definition and differentiation between the. population under study and the target population, as these concepts serve as the ...
Stratified Sampling | Definition, Guide & Examples. Published on September 18, 2020 by Lauren Thomas.Revised on June 22, 2023. In a stratified sample, researchers divide a population into homogeneous subpopulations called strata (the plural of stratum) based on specific characteristics (e.g., race, gender identity, location, etc.).
A population is a complete set of people with a specialized set of characteristics, and a sample is a subset of the population. The usual criteria we use in defining population are geographic, for example, "the population of Uttar Pradesh". In medical research, the criteria for population may be clinical, demographic and time related.
Data Homogeneity. It is often important to determine if a set of data is homogeneous before any statistical technique is applied to it. Homogeneous data are drawn from a single population. In other words, all outside processes that could potentially affect the data must remain constant for the complete time period of the sample.
size than a study that requires one only to describe population parameters. Research frequently has multiple target populations, each critically important to the objectives of the study. A health survey may target all ... The more homogeneous the population in terms of the variables of interest, the more consideration should be given to choosing
A population-based approach was proposed and the sample frame was from the National Cancer Database, which includes more than 40 million historical records from over 1500 treatment sites. This was used to create the study population (women with T1-3N1 breast cancer before chemotherapy) by refining the initial dataset to match the research ...
Should there be research investigating efficacy of a therapy in lung cancer patients in the USA, it would not be possible for all 541,000 patients to be enrolled in the research. Instead, only a representative number of the total number of patients would be enrolled. This representative subset of the lung cancer population is called a 'sample'.
The goodness-of-fit test can be used to decide whether a population fits a given distribution, but it will not suffice to decide whether two populations follow the same unknown distribution. A different test, called the test for homogeneity, can be used to draw a conclusion about whether two populations have the same distribution. To calculate ...
The following is a modified excerpt from Applied Qualitative Research Design: A Total Quality Framework Approach (Roller & Lavrakas, 2015, pp. 107-109). Fundamental to the design of a focus group study is group composition. Specifically, the researcher must determine the degree of homogeneity or heterogeneity that should be represented by the group participants.
The process of selecting a small homogeneous group of subjects or units for examination and analysis. Why use this method? Homogeneous sampling is used when the goal of the research is to understand and describe a particular group in depth. Citation: Cohen D, Crabtree B. "Qualitative Research Guidelines Project."
Clinical research is essential for the advancement of medicine; however, trials often enrol homogeneous populations that do not accurately represent the patient populations served. Representative ...
A sample is a representative portion of the larger population. In research, sampling is the process of acquiring this subset from a population. ... The scientific merits of homogeneous convenien ...
Homogenous sampling. Homogenous sampling involves selecting similar cases to further investigate a particular phenomenon or subgroup of interest. The logic of homogenous sampling is in contrast to the logic of maximum variation sampling.
Whether homogeneous sampling or quota sampling is a better nonprobability option depends on the research question. For research questions focused on sociodemographic factors as a source of heterogeneity (e.g., the sociodemographic factor is a focal variable of interest in the study), quota sampling is the better choice.
Homogeneous sampling: In homogeneous sampling, the researcher selects participants who have similar characteristics or experiences. This can be useful when studying a specific subpopulation that shares common traits or experiences. ... Ethical considerations: Purposive sampling can be used to ensure that vulnerable or marginalized populations ...
In this respect, homogeneous sampling is the opposite of maximum variation sampling. A homogeneous sample is often chosen when the research question that is being address is specific to the characteristics of the particular group of interest, which is subsequently examined in detail. Typical case sampling
An empirical transition matrix for non-homogeneous Markov chains based on censored observations. ... funded by Alleviant Medical), Novartis, Population Health Research Institute and Rutgers ...