Qualitative Sampling Methods The following module describes common methods for collecting qualitative data. Describe common types of qualitative sampling methodology. Explain the methods typically used in qualitative data collection. Describe how sample size is determined. Purposeful Sampling is the most common sampling strategy.
In this type of sampling, participants are selected or sought after based on pre-selected criteria based on the research question. For example, the study may be attempting to collect data from lymphoma patients in a particular city or county. The sample size may be predetermined or based on theoretical saturation, which is the point at which the newly collected no longer provides additional insights.
Click on the following link for a desciption of types of purposeful sampling: Types of Purposeful Sampling. Quota Sampling is a sampling technique whereby participant quotas are preset prior to sampling. Typically, the researcher is attempting to gather data from a certain number of participants that meet certain characteristics that may include things such as age, sex, class, marital status, HIV status, etc.
Click here for more information on this type of sampling: Snowball Sampling is also known as chain referral sampling. In this method, the participants refer the researcher to others who may be able to potentially contribute or participate in the study. This method often helps researchers find and recruit participants that may otherwise be hard to reach. For more information, click here: The following sampling methods that are listed in your text are types of non-probability sampling that should be avoided:.
Since such non-probability sampling methods are based on human choice rather than random selection, statistical theory cannot explain how they might behave and potential sources of bias are rampant.
In your textbook, the two types of non-probability samples listed above are called "sampling disasters. The article provides great insight into how major polls are conducted. When you are finished reading this article you may want to go to the Gallup Poll Web site, https: It is important to be mindful of margin or error as discussed in this article. We all need to remember that public opinion on a given topic cannot be appropriately measured with one question that is only asked on one poll.
Such results only provide a snapshot at that moment under certain conditions. The concept of repeating procedures over different conditions and times leads to more valuable and durable results. Within this section of the Gallup article, there is also an error: In 5 of those surveys, the confidence interval would not contain the population percent.
Eberly College of Science. Printer-friendly version Sampling Methods can be classified into one of two categories: Sample has a known probability of being selected Non-probability Sampling: Sample does not have known probability of being selected as in convenience or voluntary response surveys Probability Sampling In probability sampling it is possible to both determine which sampling units belong to which sample and the probability that each sample will be selected.
Simple Random Sampling SRS Stratified Sampling Cluster Sampling Systematic Sampling Multistage Sampling in which some of the methods above are combined in stages Of the five methods listed above, students have the most trouble distinguishing between stratified sampling and cluster sampling. With stratified sampling one should: With cluster sampling one should divide the population into groups clusters.
Stratified sampling would be preferred over cluster sampling, particularly if the questions of interest are affected by time zone. For example the percentage of people watching a live sporting event on television might be highly affected by the time zone they are in. Cluster sampling really works best when there are a reasonable number of clusters relative to the entire population.
In this case, selecting 2 clusters from 4 possible clusters really does not provide much advantage over simple random sampling.
In some cases, investigators are interested in "research questions specific" to subgroups of the population. For example, researchers might be interested in examining whether cognitive ability as a predictor of job performance is equally applicable across racial groups.
SRS cannot accommodate the needs of researchers in this situation because it does not provide subsamples of the population. Systematic sampling also known as interval sampling relies on arranging the study population according to some ordering scheme and then selecting elements at regular intervals through that ordered list. Systematic sampling involves a random start and then proceeds with the selection of every k th element from then onwards.
It is important that the starting point is not automatically the first in the list, but is instead randomly chosen from within the first to the k th element in the list.
A simple example would be to select every 10th name from the telephone directory an 'every 10th' sample, also referred to as 'sampling with a skip of 10'. As long as the starting point is randomized , systematic sampling is a type of probability sampling. It is easy to implement and the stratification induced can make it efficient, if the variable by which the list is ordered is correlated with the variable of interest. For example, suppose we wish to sample people from a long street that starts in a poor area house No.
A simple random selection of addresses from this street could easily end up with too many from the high end and too few from the low end or vice versa , leading to an unrepresentative sample. Note that if we always start at house 1 and end at , the sample is slightly biased towards the low end; by randomly selecting the start between 1 and 10, this bias is eliminated. However, systematic sampling is especially vulnerable to periodicities in the list.
If periodicity is present and the period is a multiple or factor of the interval used, the sample is especially likely to be un representative of the overall population, making the scheme less accurate than simple random sampling. For example, consider a street where the odd-numbered houses are all on the north expensive side of the road, and the even-numbered houses are all on the south cheap side.
Under the sampling scheme given above, it is impossible to get a representative sample; either the houses sampled will all be from the odd-numbered, expensive side, or they will all be from the even-numbered, cheap side, unless the researcher has previous knowledge of this bias and avoids it by a using a skip which ensures jumping between the two sides any odd-numbered skip.
Another drawback of systematic sampling is that even in scenarios where it is more accurate than SRS, its theoretical properties make it difficult to quantify that accuracy. In the two examples of systematic sampling that are given above, much of the potential sampling error is due to variation between neighbouring houses — but because this method never selects two neighbouring houses, the sample will not give us any information on that variation.
As described above, systematic sampling is an EPS method, because all elements have the same probability of selection in the example given, one in ten. It is not 'simple random sampling' because different subsets of the same size have different selection probabilities — e. When the population embraces a number of distinct categories, the frame can be organized by these categories into separate "strata.
There are several potential benefits to stratified sampling. First, dividing the population into distinct, independent strata can enable researchers to draw inferences about specific subgroups that may be lost in a more generalized random sample. Second, utilizing a stratified sampling method can lead to more efficient statistical estimates provided that strata are selected based upon relevance to the criterion in question, instead of availability of the samples. Even if a stratified sampling approach does not lead to increased statistical efficiency, such a tactic will not result in less efficiency than would simple random sampling, provided that each stratum is proportional to the group's size in the population.
Third, it is sometimes the case that data are more readily available for individual, pre-existing strata within a population than for the overall population; in such cases, using a stratified sampling approach may be more convenient than aggregating data across groups though this may potentially be at odds with the previously noted importance of utilizing criterion-relevant strata.
Finally, since each stratum is treated as an independent population, different sampling approaches can be applied to different strata, potentially enabling researchers to use the approach best suited or most cost-effective for each identified subgroup within the population. There are, however, some potential drawbacks to using stratified sampling.
First, identifying strata and implementing such an approach can increase the cost and complexity of sample selection, as well as leading to increased complexity of population estimates. Second, when examining multiple criteria, stratifying variables may be related to some, but not to others, further complicating the design, and potentially reducing the utility of the strata.
Finally, in some cases such as designs with a large number of strata, or those with a specified minimum sample size per group , stratified sampling can potentially require a larger sample than would other methods although in most cases, the required sample size would be no larger than would be required for simple random sampling.
Stratification is sometimes introduced after the sampling phase in a process called "poststratification". Although the method is susceptible to the pitfalls of post hoc approaches, it can provide several benefits in the right situation.
Implementation usually follows a simple random sample. In addition to allowing for stratification on an ancillary variable, poststratification can be used to implement weighting, which can improve the precision of a sample's estimates. Choice-based sampling is one of the stratified sampling strategies. In choice-based sampling,  the data are stratified on the target and a sample is taken from each stratum so that the rare target class will be more represented in the sample.
The model is then built on this biased sample. The effects of the input variables on the target are often estimated with more precision with the choice-based sample even when a smaller overall sample size is taken, compared to a random sample. The results usually must be adjusted to correct for the oversampling.
In some cases the sample designer has access to an "auxiliary variable" or "size measure", believed to be correlated to the variable of interest, for each element in the population. These data can be used to improve accuracy in sample design. One option is to use the auxiliary variable as a basis for stratification, as discussed above. Another option is probability proportional to size 'PPS' sampling, in which the selection probability for each element is set to be proportional to its size measure, up to a maximum of 1.
In a simple PPS design, these selection probabilities can then be used as the basis for Poisson sampling. However, this has the drawback of variable sample size, and different portions of the population may still be over- or under-represented due to chance variation in selections. Systematic sampling theory can be used to create a probability proportionate to size sample. This is done by treating each count within the size variable as a single sampling unit.
Samples are then identified by selecting at even intervals among these counts within the size variable. This method is sometimes called PPS-sequential or monetary unit sampling in the case of audits or forensic sampling. The PPS approach can improve accuracy for a given sample size by concentrating sample on large elements that have the greatest impact on population estimates. PPS sampling is commonly used for surveys of businesses, where element size varies greatly and auxiliary information is often available—for instance, a survey attempting to measure the number of guest-nights spent in hotels might use each hotel's number of rooms as an auxiliary variable.
In some cases, an older measurement of the variable of interest can be used as an auxiliary variable when attempting to produce more current estimates. Sometimes it is more cost-effective to select respondents in groups 'clusters'. Sampling is often clustered by geography, or by time periods. Nearly all samples are in some sense 'clustered' in time — although this is rarely taken into account in the analysis. For instance, if surveying households within a city, we might choose to select city blocks and then interview every household within the selected blocks.
Clustering can reduce travel and administrative costs. In the example above, an interviewer can make a single trip to visit several households in one block, rather than having to drive to a different block for each household. It also means that one does not need a sampling frame listing all elements in the target population. Instead, clusters can be chosen from a cluster-level frame, with an element-level frame created only for the selected clusters. In the example above, the sample only requires a block-level city map for initial selections, and then a household-level map of the selected blocks, rather than a household-level map of the whole city.
Cluster sampling also known as clustered sampling generally increases the variability of sample estimates above that of simple random sampling, depending on how the clusters differ between one another as compared to the within-cluster variation. For this reason, cluster sampling requires a larger sample than SRS to achieve the same level of accuracy — but cost savings from clustering might still make this a cheaper option.
Cluster sampling is commonly implemented as multistage sampling. This is a complex form of cluster sampling in which two or more levels of units are embedded one in the other. The first stage consists of constructing the clusters that will be used to sample from. In the second stage, a sample of primary units is randomly selected from each cluster rather than using all units contained in all selected clusters. In following stages, in each of those selected clusters, additional samples of units are selected, and so on.
All ultimate units individuals, for instance selected at the last step of this procedure are then surveyed. This technique, thus, is essentially the process of taking random subsamples of preceding random samples.
Multistage sampling can substantially reduce sampling costs, where the complete population list would need to be constructed before other sampling methods could be applied.
By eliminating the work involved in describing clusters that are not selected, multistage sampling can reduce the large costs associated with traditional cluster sampling. In quota sampling , the population is first segmented into mutually exclusive sub-groups, just as in stratified sampling. Then judgement is used to select the subjects or units from each segment based on a specified proportion. For example, an interviewer may be told to sample females and males between the age of 45 and It is this second step which makes the technique one of non-probability sampling.
In quota sampling the selection of the sample is non- random. For example, interviewers might be tempted to interview those who look most helpful. The problem is that these samples may be biased because not everyone gets a chance of selection. This random element is its greatest weakness and quota versus probability has been a matter of controversy for several years. In imbalanced datasets, where the sampling ratio does not follow the population statistics, one can resample the dataset in a conservative manner called minimax sampling.
The minimax sampling has its origin in Anderson minimax ratio whose value is proved to be 0. This ratio can be proved to be minimax ratio only under the assumption of LDA classifier with Gaussian distributions. The notion of minimax sampling is recently developed for a general class of classification rules, called class-wise smart classifiers. In this case, the sampling ratio of classes is selected so that the worst case classifier error over all the possible population statistics for class prior probabilities, would be the.
Accidental sampling sometimes known as grab , convenience or opportunity sampling is a type of nonprobability sampling which involves the sample being drawn from that part of the population which is close to hand.
Simple random sampling. Many dissertation supervisors advice the choice of random sampling methods due to the representativeness of sample group and less room for researcher bias compared to non-random sampling techniques.
Sampling Methods can be classified into one of two categories: Probability Sampling: Sample has a known probability of being selected. Non-probability Sampling: Sample does not have known probability of being selected as in convenience or voluntary response surveys.
In this technique, each member of the population has an equal chance of being selected as subject. The entire process of sampling is done in a single step with each subject selected independently of the other members of the population. There are many methods to proceed with simple random sampling. Random sampling refers to a variety of selection techniques in which sample members are selected by chance, but with a known probability of selection.
Sampling Methods in Qualitative and Quantitative Research , views. Share; Like; Sampling Methods in Qualitative and Quantitative Research This is the hard part! It must not systematically exclude anyone. b. Remember the famous sampling mistake?2. Generate random numbers3. Select one person per random number. Sampling and types of sampling methods commonly used in quantitative research are discussed in the following module. Learning Objectives: Define sampling and randomization.