Methodology  Open  Published:
The effect of clustering on lot quality assurance sampling: a probabilistic model to calculate sample sizes for quality assessments
Emerging Themes in Epidemiologyvolume 10, Article number: 11 (2013)
Abstract
Background
Traditional Lot Quality Assurance Sampling (LQAS) designs assume observations are collected using simple random sampling. Alternatively, randomly sampling clusters of observations and then individuals within clusters reduces costs but decreases the precision of the classifications. In this paper, we develop a general framework for designing the cluster(C)LQAS system and illustrate the method with the design of data quality assessments for the community health worker program in Rwanda.
Results
To determine sample size and decision rules for CLQAS, we use the betabinomial distribution to account for inflated risk of errors introduced by sampling clusters at the first stage. We present general theory and code for sample size calculations.
The CLQAS sample sizes provided in this paper constrain misclassification risks below userspecified limits. Multiple CLQAS systems meet the specified risk requirements, but numerous considerations, including percluster versus perindividual sampling costs, help identify optimal systems for distinct applications.
Conclusions
We show the utility of CLQAS for data quality assessments, but the method generalizes to numerous applications. This paper provides the necessary technical detail and supplemental code to support the design of CLQAS for specific programs.
Background
Lot quality assurance sampling (LQAS) was originally developed as an industrial control tool [1]. From a lot of goods, a random sample of n items are selected and inspected for defects. If a certain number, d, or more goods have a defect, then the lot is rejected from distribution and the industrial process reviewed to address the problem. If fewer then d defects are detected, then the lot is determined to be of acceptable quality and shipped. In the 1980s, LQAS emerged as a useful tool in public health to identify areas with suboptimal program coverage or unacceptable levels of disease [2–6].
The binomial model, assuming simple random sampling, is ordinarily used to determine the LQAS design when dealing with large populations. This approach allows for the calculation of both the required sample size, n, and the decision rule, d. These two design parameters are governed by four parameters: 1) the upper threshold, p_{ u } — the proportion of defective items at or above which a lot from such a population should be classified as suboptimal; 2) the maximum allowable risk of misclassification error at the upper threshold, α_{ max }; 3) the lower threshold, p_{ l } – the proportion of defective items, at or below which a lot should be classified to have sufficient quality; and 4) the maximum allowable risk of misclassification error at the lower threshold, β_{ max }. The requisite relationships are:
and
where X is a binomial random variable representing the number of 'defects’ observed in the sample of size n, with probability of observing a defect set to p_{l} or p_{u}.
Sometimes a simple random sample is not available—for example, geographic spread may make a simple random sample infeasible [7–9]— and one must resort to other sampling designs. One such instance is the data quality assessment of the community health worker (CHW) program in southern Kayonza, Rwanda. For this program supported by Partners In Health, CHWs visit their assigned households monthly to document key demographic and health indicators. These data are used to manage CHW activities, forecast health needs in the area, and evaluate program implementation. Thus, ensuring high quality CHW data is important.
Each CHW is responsible for 40–60 households, with 2–4 CHWs per village, and about seven villages per cell [10]. Since CHWs are supervised at the cell level, we are interested in classifying the data quality for each cell. Cells identified as having low data quality would receive additional support, such as extra training for the CHWs and the cell supervisor. Cells identified as having sufficiently high quality data would receive some positive reinforcement, such as certificates of achievement.
Applying the traditional LQAS system, we would randomly sample n households from the approximately 1000 households in a cell. Each household would be visited and remeasured. The new data would then be compared to the data documented in the CHW household register. If d or more households captured in the CHW registers have discrepant data, then the cell is classified as having insufficient quality. If fewer then d households have discrepant data, then the cell is classified as high data quality. The challenge, however, is the potential for large geographic spread of the sampled households requiring an intensive amount of resources and time. In the most extreme case, the n households could come from n different and widespread villages. To lighten the burden on the program by providing a more concentrated sample, we explore a two stage clustersampling procedure: first randomly sample a number of CHWs, and second sample households within these CHWs’ domain.
This method of sampling lowers the resources required for the data quality assessments, as sampling households in the same village requires significantly fewer resources than sampling households in different villages. However, previous work has shown that the implementation of an LQAS system in such a clustered manner when n and d are incorrectly calculated under the assumption of simple random sampling raises the misclassification risks above the specified bounds because of the correlation within clusters [7, 8, 11, 12]. In 2010, Pezzoli et al. describe an approach for developing a clusterLQAS (CLQAS) system where the total sample size, n, is based on the binomial distribution [12]. They divide the total sample size into m clusters, each of size k and then estimate the true error rates of the CLQAS systems (adjusting for correlation) via simulations. Their approach for designing the CLQAS system has several limitations. The method calculates errors for a specific system and determines whether these errors are sufficiently low instead of designing a system to restrict misclassification risks below programspecified bounds. Restricting systems to those with a total sample size of n provides a limited number of options, and it is possible that for certain conditions, none of the designs will have misclassification risks below desired levels.
In this paper, we provide a general probability framework for determining the appropriate sample size for a CLQAS system that meets the desired constraints without restricting it to systems with a fixed total sample size of n. The theory for CLQAS is presented in the methods section. In the results section, we illustrate the method by calculating the sample sizes and decision rules for the CHW data quality assessment in Rwanda and present sensitivity analyses using simulation results to assess the performance of the selected CLQAS sampling designs when the assumptions used to develop the system are not fully met. Finally, the discussion presents considerations for designing CLQAS systems for other contexts.
Methods
Calculating sample sizes and decision rules for clusterLQAS
Basic twoway LQAS aims to classify an area into one of two categories – defective or acceptable quality if referring to a lot of goods, high or low performance if referring to program coverage. The sample size, n, and decision rule, d, are determined by the upper and lower thresholds (p_{ u } and p_{ l }) and the maximum allowable misclassification risks at each threshold (α_{ max } and β_{ max }, respectively). These parameters are set a priori by program managers and are determined by a variety of factors. The thresholds can be determined by program targets, key epidemiological parameters, or knowledge of underlying distribution. The maximum misclassification risks reflect the amount of uncertainty tolerable in an evaluation with consideration of the resulting consequences of failing to allot resources to areas in need or to spread limited resources too thin by allocating additional resources to high performing areas.
For basic twoway LQAS actual probability of a certain classification for a given p is based on the binomial model. Areas at or above the upper threshold have a high probability (≥1 α_{ max }) of being correctly classified in the upper category (in this case, being classified as defective) and areas at or below the lower threshold have a high probability (≥1 β_{ max }) of being classified in the lower category (in this case, being classified as of acceptable quality). The area between the two thresholds, often referred to as the grey area, has less discernible classification properties. Typically in the grey area, the closer to a threshold, the more likely to be classified into the corresponding category though in the most extreme, areas in the between the upper and lower thresholds can have a 50–50 chance of being classified into a given category.
The basic twoway LQAS system is defined by two numbers: n, the number of items randomly sampled from the lot and the decision rule, d. If d or more of the n items are defective, then the lot is rejected. Otherwise, the lot is classified to be of acceptable quality. The design of a CLQAS system is defined by three numbers: the number of clusters sampled, m; the number of items per cluster sampled, k; and the decision rule for classification, d[7, 8]. The number of defective 'units’ across the m*k elements are tallied and compared to the decision rule in the same way as the basic twoway LQAS. If d or more are defective, then the lot is rejected; if fewer than d are defective, then the lot is classified to be of acceptable quality.
If there is no correlation in the presence or absence of defects within clusters, then calculating the sample size, n = m*k, and the decision rule based on the binomial distribution assuming simple random sampling is sufficient to maintain the risk of errors below specified levels. However, if the sample size is calculated using the binomial model that assumes simple random sampling but there is clustering of the outcome, then the true misclassification risks are likely greater than the specified limits, α_{ max } and β_{ max }[7, 8, 11, 12].
To overcome this problem, we propose using the betabinomial model to incorporate the clustering effect into the sample size and decision rule calculations. In addition to specifying the four parameters above (p_{l}, p_{u}, α_{ max } and β_{ max }), one must also specify ρ, the intraclass correlation (ICC). Reliable estimates of ICC are critical for obtaining appropriate designs and more detail on this topic is presented in the discussion.
Based on these five parameters, the program manager finds a recommended number of clusters and number of individuals per cluster to sample, along with a corresponding decision rule, that reduces the risk of misclassification below the specified levels (Additional file 1). The methods and results presented assume that the population sizes within clusters are equal. The risk limits are not always achievable. Even for infinitely large samples in a cluster, we must sample at least ρ* n_{ min } clusters, where n_{ min } is the minimum sample size needed for an LQAS system with no clustering, to meet the specified constraints for the CLQAS system. For a given ρ, and any m > ρ* n_{ min }, there are multiple k to satisfy the desired risk profile. We choose the smallest such k.
We provide code developed for R version 2.15.1 that calls the VGAM package [13] for the probability distribution function of the betabinomial to calculate the number of samples per cluster and the corresponding decision rules for a range of number of clusters, m, up to m = n_{ min } (Additional file 2).
Application to the CHW data quality assessment
To illustrate the use of this method, we apply the above theory to calculate the sample size for the CHW data quality assessment in Rwanda. The aim of the data quality assessment is to monitor, at the cell level, the accuracy of household data collected by CHWs. For each cell, using the CLQAS system, m CHWs and k households per CHW are randomly chosen. At each sampled household, information collected during a supervision visit of the household are compared to the information in the CHW register. As an example, we consider just one data element collected by CHWs – whether the number of children (under five years old) in the household is correctly documented in the CHW register.
One possible approach to selecting a particular CLQAS design for a set of parameters is to introduce a cost function into the decision making. Consider the per cluster and per individual costs, C_{ m } and C_{ k } respectively, and define the total cost for the implementation, C_{ total } = C_{ m }*m + C_{ k }*m*k. The total costs for CLQAS designs under different assumed per cluster and per individual costs are provided.
Simulation of performance of clusterLQAS
Using simulations we evaluated the performance of various CLQAS systems. The results are based on 10,000 simulation draws from a betabinomial distribution for a given set of values of m, k, ρ, and p. The first set of simulations looked at the effect of misspecification of ICC in the design on the misclassification errors. The second set of simulations also evaluated the misclassification errors for misspecification of ICC as well as the mean and standard error of the ICC using the Analysis of Variance (ANOVA) estimator [14]. For simulations that resulted in no events in any of the clusters, the ANOVA ICC is inestimable and we thus excluded them from the results. We report the percent of simulations with valid ANOVA estimates.
Results
Application to the CHW data quality assessment
For the design of the CLQAS system for the data quality assessment in Rwanda, the following parameters are specified:

1.
p_{u} = 0.25 – cells with 25% or higher error rates in documentation of the number of children are considered to have insufficient data quality;

2.
α_{ max } = 0.10 – restrict the probability of classifying cells with 25% or higher error rates in documentation as sufficient data quality to less than 10%;

3.
p_{l} = 0.05 – cells with 5% or lower error rates in documentation of the number of children are considered to have sufficient data quality;

4.
β_{ max } = 0.10 – restrict the probability of classifying cells with 5% or lower error rates in documentation as insufficient data quality to less than 10%; and

5.
ρ = 0.1 – the ICC describing the amount that data quality clusters by CHW.
Using these parameter values, and ignoring the clustered nature of the sampling, the recommended sample size is 20 with d = 3. The expected probability of misclassification of such a system, based on the (nonclustered) binomial distribution are α = 0.091 and β = 0.075. Table 1 presents the recommended sample sizes and expected misclassification errors for a range of ICCs (ρ = 0.01, 0.025, 0.05, 0.1, 0.15, and 0.2) and number of clusters (m = 220) to meet the constraints specified above. When ρ = 0.1, total sample sizes range from 20 (m = 20, k = 1) to 136 (m = 2, k = 68). For small number clusters sampled and large ICC, some of the designs may not be suitable for the CHW data quality assessment as the recommended number of samples per cluster exceed the average number of households available in a village. For a fixed m, the per cluster sample size increases as the ICC increases. As a general rule, for a fixed ICC, the total sample size increases as the number of clusters sampled decreases. This rule may not hold when the sample sizes are small and the requirement that all clusters be of the same size is imposed because of the discrete nature of the betabinomial model.
In Table 1, we also present the estimated costs for each system under two possibilities, one with large per cluster costs (C_{ m } = $500) and small per individual costs (C_{ k } = $10) and the other with moderate per cluster costs (C_{ m } = $300) and larger per individual costs (C_{ k } = $50). These two disparate costing ratios (C_{ m }/C_{ k }), 50 to 1 and 6 to 1, cover a spectrum of possibilities. The least expensive CLQAS for a given ICC depends on the costing ratio. For example, for ρ = 0.1, the design with {m = 3, k = 15, d = 6} is the least expensive ($1950) under the first cost structure and the designs with {m = 4, k = 9, d = 5} and {m = 5, k = 6, d = 4} are the cheapest ($3000) under the second cost structure.
Simulation of performance of clusterLQAS
To examine the impact of misspecification of ρ, we first evaluate two different CLQAS systems for different values of ρ. The two designs are based on the parameters specified above for the CHW data quality assessment assuming ρ = 0.1 – namely, {m = 4,k = 9,d = 5} and {m = 9,k = 3,d = 4}. The effect of misspecifying ICC in the design on misclassification risks is evident in the Operating Characteristic Curves, the curve that fits the probability of classifying an area as poor data quality via a specific CLQAS system for a range of proportions of true errors in data. As seen in Figure 1a and b, when the true ICC is equal to that specified in the design (in this case ρ = 0.1) the probability of classifying an area as having poor data quality remains below β_{ max } = 0.1 when the proportion of registers with errors is less than 5% (line falls in the lower shaded box for p < 5%) and is above 0.9 (1α_{ max }) when the proportion of registers with errors is greater than 25% (the line falls in the upper shaded box for p < 25%). However, when the ICC is above the value estimated for the design of the system, then the misclassification errors often exceed α_{ max } and β_{ max }. Despite the bigger overall sample sizes, the systems with fewer sampled clusters (m = 4) are more likely to exceed α_{ max } and β_{ max } than the system with more sampled clusters (m = 9) when the true ICC is larger than expected, illustrating that, when cost is not an issue, sampling more clusters is a more conservative design. Any other ICC values smaller than the specified ρ = 0.1 result in systems with misclassification risks smaller than α_{ max } and β_{ max. }
Next, we present three CLQAS systems that satisfy the misclassification constraints for ρ = 0.01 (Table 2), ρ = 0.1 (Table 3), and ρ = 0.2 (Table 4). As above, when the ICC used for a set of simulations was equal to or less than the value assumed for designing the system, the observed α and β values were less than the constraints. With a true ICC larger than the value assumed in the design, the risks often exceeded the misclassification constraints. The more clusters in the sample, the smaller the impact of the ICC, and thus a lower risk of exceeding misclassification constraints even when the ICC is underestimated in the design.
Discussion
Lot quality assurance sampling has already proven a valuable tool for public health program monitoring and evaluation [6]. Combining sampling in clusters and LQAS increases the versatility of the tool. However, as others have noted, failing to account for the clustering in the design of the LQAS system typically leads to inflated misclassification errors [7, 8]. This paper provides the general theory for sample size calculations for CLQAS that accounts for the clustering in the design using a betabinomial model. We applied this theory to the CHW data quality assessments, using one set of parameters – p_{ u } = 0.25, p_{ l } = 0.1, α_{ max } = 0.1 and β_{ max } = 0.1, and ρ = 0.1. Conclusions about the design of CLQAS parallel standard cluster sampling theory. First, for a given set of parameters, there is an interplay between the number clusters and number of individuals per cluster sampled, and sampling fewer clusters will lead to a larger total sample size. Second, if the ICC is correctly specified, or overestimated, then the resulting CLQAS systems will classify areas while restricting misclassification risks below the bounds set by the program manager. However, underestimating the ICC can lead to misclassification risks above the desired bounds.
Determining the final design of the clusterLQAS system
One challenge for CLQAS, as with any cluster design, is determining a plausible value of ICC for the first implementation. Relevant ICC values may not be published, can change over time, and are population and design dependent. One can look to previous surveys, such as the Demographic Health Survey (DHS) for the country, to give some indication of the magnitudes of the clustering expected, especially if local information is provided [15]. In our application, the DHS does not collect data on indicators that could be directly linked to the performance of CHWs, so that option is not available. The recent CONSORT extension for cluster randomized trials requires that ICC values be reported [16]. These estimates, particularly from control groups [17], could provide insight into a range of values of ICC for the design of CLQAS. Two studies offer insight into the ICC of CHW data quality  one study estimated the ICC of the quality of communication in nurses to be 0.28 [18] and another the median estimated ICC across a variety of qualityofcare indicators in health facilities to be 0.2 [19]. These two studies focused on quality of care, and not quality of documentation, and evaluated performance in a different group of professionals and so the relevance is limited.
As Table 1 typifies, a conservative value of the ICC is ideal to protect against high misclassification errors, although it may be a costly solution. Once the data are collected, the ICC can be estimated with confidence intervals and used for subsequent CLQAS designs. Ridout et al. demonstrate that averaged across different settings, the ANOVA estimator has low bias and standard deviation compared to other methods to estimate ICC for binary data [14]. However, the context presented here for the assessment of CHW data quality where few clusters and few individuals per cluster are sampled, we observe large biases and standard deviations (Tables 2, 3 and 4). Most worrisome, when the fewest clusters (m = 4) are used, the ANOVA estimator is small relative to the true ρ, thus indicating a bias that would lead to underestimating the risk of misclassification. Similar results were observed by Minetti et al. for designs that had more clusters (m = 10) and few individuals per cluster [9]. Therefore, practitioners should exercise caution when choosing the best ICC value for future designs, especially if estimated from previous implementations where very few clusters were sampled.
Cost is one approach to determining the preferred design, and other factors may come into play. For example, when ρ = 0.1, the cost of the {m = 5, k = 6, d = 4} design is $3,000 compared to $3,500 for the {m = 7, k = 4, d = 4} design. For an extra $500, the latter design would still meet the misclassification constraints for larger ICCs than anticipated (up to ρ = 0.15). Further, more clusters would provide a better estimate of the ICC for future designs. Other factors such as time requirements, available human resources, geographic spread, and topography may contribute to the final decision as well.
Limitations of the CLQAS design
In the design of a CLQAS system, users should be aware of several possible limitations. The sample size and decision rules, and corresponding misclassification risks, hinge on the assumed value of ICC. As discussed earlier, if the ICC is underestimated in the design, the risk of misclassification may be higher than desired. Additionally, CLQAS is subject to the same limitations as LQAS; specifically, when many areas have true prevalences in the grey area (between p_{ l } and p_{ u }), the survey will have very poor classification properties (commonly referred to as specificity) [9, 20, 21]. Prior knowledge about the underlying distribution of prevalence among the areas in the study region can be used to inform the choice of the upper and lower thresholds and the choice of design. For example, Olives and Pagano propose an alternative Bayesian approach to incorporate this prior knowledge into the design [22, 23]. Additionally, the betabinomial model assumes that there are an infinite amount of clusters in the population. When the population consists of few clusters, the classification risks presented here will overestimate the true misclassification risks. Therefore, investigators may be able to sample fewer clusters or fewer individuals per cluster than presented here while restricting misclassification risks below the specified bounds. A final limitation of the CLQAS design presented here is the assumption that the clusters are the same size and that m clusters are sampled using simple random sampling without replacement. For the CHW data quality assessment, this is a reasonable assumption, as there is little variability in the number of households per village.
Conclusion
CLQAS has already proven useful for monitoring of neonatal tetanus eradication, vaccination coverage, and acute malnutrition [8, 11, 24–28]. This paper provides a probabilistic model to design CLQAS systems in the future. We illustrate the method with the CHW data accuracy assessment in southern Kayonza, Rwanda, but the methodology presented and supplemental code are generalizable to any health program activity. Extending to other programs requires that program managers specify parameters useful for planning and evaluation – namely thresholds that meaningfully define areas that are high or low performing and allowable misclassification risks at these thresholds. Selection of these parameters is often more challenging than the determining the design once these parameters are specified as it requires consideration of program resource allocation and program objectives. Further, as highlighted in this paper, correct design of CLQAS hinges on the assumed value of the ICC.
Along with additional evaluation of CLQAS in the field, more methodological research is needed as well. Future work should investigate adjusting the misclassification risks to account for finite cluster sizes, accounting for varying cluster sizes, or incorporating knowledge about underlying prevalence of the trait of interest (BayesCLQAS). Advances in the CLQAS methodology will add to the set of evaluation tools that support local public health program evaluation and management, particularly in resource limited settings.
References
 1.
Dodge H, Romig H: A method of sampling inspection. The Bell System Technical Journal. 1929, 8: 613631. 10.1002/j.15387305.1929.tb01240.x.
 2.
Lanata CF, Stroh G, Black RE: Lot quality assurance sampling in health monitoring. Lancet. 1988, 1: 1223.
 3.
Valadez JJ: Assessing child survival programs in developing countries: testing lot quality assurance sampling. Cambridge, MA: Harvard University Press; 1991.
 4.
Lwanga S, Lemeshow S: Sample size determination in health studies: a practical manual. Geneva: World Health Organization; 1991.
 5.
Lemeshow S, Taber S: Lot quality assurance sampling: single and doublesampling plans. Bull World Health Organ. 1991, 44: 11532.
 6.
Robertson SE, Valadez JJ: Global review of health care surveys using lot quality assurance sampling (LQAS), 1984–2004. Soc Sci Med. 2006, 63: 164860. 10.1016/j.socscimed.2006.04.011
 7.
Olives C, Pagano M, Deitchler M, et al: Cluster designs to assess the prevalence of acute malnutrition by lot quality assurance sampling: a validation study by computer simulation. J R Stat Soc A Stat Soc. 2009, 172: 495510. 10.1111/j.1467985X.2008.00572.x.
 8.
Pezzoli L, Tchio R, Dzossa AD, et al: Clustered lot quality assurance sampling: a tool to monitor immunization coverage rapidly during a national yellow fever and polio vaccination campaign in Cameroon, May 2009. Epidemiol Infect. 2012, 140: 1426. 10.1017/S0950268811000331
 9.
Minetti A, RieraMontes M, Nackers F, et al: Performance of small cluster surveys and the clustered LQAS design to estimate locallevel vaccination coverage in Mali. Emerg Themes Epidemiol. 2012, 9: 6. 10.1186/1742762296
 10.
Government of Rwanda: Economic development & poverty reduction strategy 2008–2012. Kigali: Government of Rwanda; 2007.
 11.
Pezzoli L, Pineda S, Halkyer P, et al: Clustersample surveys and lot quality assurance sampling to evaluate yellow fever immunisation coverage following a national campaign, Bolivia, 2007. Trop Med Int Health. 2009, 14: 35561. 10.1111/j.13653156.2009.02231.x
 12.
Pezzoli L, Andrews N, Ronveaux O: Clustered lot quality assurance sampling to assess immunisation coverage: increasing rapidity and maintaining precision. Trop Med Int Health. 2010, 15: 5406.
 13.
Yee TW, Wild CJ: Vector generalized additive models. J R Stat Soc Ser B Methodol. 1996, 58: 481493.
 14.
Ridout MS, Demétrio CG, Firth D: Estimating intraclass correlation for binary data. Biometrics. 1999, 55: 13748. 10.1111/j.0006341X.1999.00137.x
 15.
Fenn B, Morris SS, Frost C: Do childhood growth indicators in developing countries cluster? implications for intervention strategies. Public Health Nutr. 2004, 7: 82934.
 16.
Campbell MK, Elbourne DR, Altman DG: CONSORT statement: extension to cluster randomized trials. BMJ. 2004, 328: 7028. 10.1136/bmj.328.7441.702
 17.
Pagel C, Prost A, Lewycka S, et al: Intracluster correlation coefficients and coefficients of variation for perinatal outcomes from five clusterrandomized trials in low and middleincome countries: results and methodological implications. BMC Trials. 2011, 12: 151. doi:10.1186/1745621512151. 10.1186/1745621512151
 18.
Gilroy K, Winch PJ, Diawara A, et al: Impact of IMCI training and language used by provider on quality of counseling provided to parents of sick children in Bougouni District, Mali. Patient Educ Couns. 2004, 54: 3544. 10.1016/S07383991(03)001897
 19.
Rowe AK, Lama M, Onikpo F, et al: Design effects and intraclass correlation coefficients from a health facility cluster survey in Benin. Int J Qual Health Care. 2002, 14: 521523. 10.1093/intqhc/14.6.521
 20.
Sandiford P: Lot quality assurance sampling for monitoring immunization programmes: costefficient or quick and dirty?. Health Policy Plan. 1993, 8 (3): 217223. 10.1093/heapol/8.3.217
 21.
Bilukha OO, Blanton C: Interpreting results of cluster surveys in emergency settings: is the LQAS test the best option?. Emerg Themes Epidemiol. 2008, 5: 25. 10.1186/17427622525
 22.
Olives C, Pagano M: BayesLQAS: classifying the prevalence of global acute malnutrition. Emerg Themes Epidemiol. 2010, 7: 3. 10.1186/1742762273
 23.
Olives C, Pagano M: Choosing a design to fit the situation: improving specificity and positive predictive values using BayesianLQAS. Int J Epidemiol. 2013, 42 (1): 346355. 10.1093/ije/dys230
 24.
Cotter B, Bremer V, Stroh G, et al: Assessment of neonatal tetanus elimination in an African setting by lot quality assurance cluster sampling (LQACS). Epidemiol Infect. 2003, 130: 221226. 10.1017/S0950268802008233
 25.
Deitchler M, Deconinck H, Bergeron G: Precision, time, and cost: a comparison of three sampling designs in an emergency setting. Emerging Themes in Epidemiology. 2008, 5: 6. 10.1186/1742762256
 26.
Deitchler M, Valadez JJ, Egge K, et al: A field test of three LQAS designs to assess the prevalence of acute malnutrition. Int J Epidemiol. 2007, 36: 85864. 10.1093/ije/dym092
 27.
Greenland K, Rondy M, Chevez A, et al: Clustered lot quality assurance sampling: a pragmatic tool for timely assessment of vaccination coverage. Trop Med Int Health. 2011, 16: 8638. 10.1111/j.13653156.2011.02770.x
 28.
Kim SH, Pezzoli L, Yacouba H, et al: Whom and where are we not vaccinating? coverage after the introduction of a New conjugate vaccine against group a meningococcus in Niger in 2010. PLoS ONE. 2012, 7: e29116. 10.1371/journal.pone.0029116
Acknowledgements
The team acknowledges the support and efforts by the IMB/PIH team in Rwanda, specifically the Community Health Department, which will implement the first accuracy assessment using this clusterLQAS design.
Funding statement
This work was supported by National Institutes of Health grant numbers 5R01AI09701502 to BHG and MP, T32 AI007358 to LH. CO was supported in part by the UW NIEHSsponsored Biostatistics, Epidemiologic and Bioinformatic Training in Environmental Health (BEBTEH) Training Grant, Grant #: NIEHS T32ES015459. BHG received support from the Department of Global Health and Social Medicine Research Core at Harvard Medical School. TM received support through the Doris Duke Charitable Foundation’s African Health Initiative.
Author information
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors’ contributions
BHG led the design, analysis and interpretation of data and the drafting of the article. TM, LH, CO, and MP all made substantial contributions to the design, analysis and interpretation of data and the drafting of the article. All authors approved of the final version of the article.
Authors’ original submitted files for images
Rights and permissions
About this article
Received
Accepted
Published
DOI
Keywords
 ClusterLQAS
 Lot quality assurance sampling
 Program evaluation
 Survey
 Community health workers