- We are sorry, but NCBI web applications do not support your browser and may not function properly. More information

# Non-differential measurement error does not always bias diagnostic likelihood ratios towards the null

^{1}Department of Veterinary Integrative Biosciences, College of Veterinary Medicine and Biomedical Sciences, Texas A&M University, College Station, TX, 77843-4458, USA

^{}Corresponding author.

This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

## Abstract

Diagnostic test evaluations are susceptible to random and systematic error. Simulated non-differential random error for six different error distributions was evaluated for its effect on measures of diagnostic accuracy for a brucellosis competitive ELISA. Test results were divided into four categories: <0.25, 0.25 – 0.349, 0.35 – 0.499, and ≥ 0.50 proportions inhibition for calculation of likelihood ratios and diagnostic odds ratios. Larger variance components of the error structure resulted in larger accuracy attenuations as measured by the area under the receiver-operating characteristic curve and systematic components appeared to cause little bias. Added error caused point estimates of likelihood ratios to be biased towards the null value (1.0) for all categories except 0.25 – 0.349. Results for the 0.35 – 0.499 category also extended beyond the null value for some error structures. Diagnostic odds ratios were consistently biased towards the null when the <0.25 category was considered the reference level. Non-differential measurement error can lead to biased results in the quantitative evaluation of ELISA and the direction is not always towards the null value.

## Background

The goal of epidemiologic investigations is the collection of valid data leading to a precise estimate of a population parameter (e.g. measure of association). For the purpose of this discussion, an estimate of a parameter will be considered biased if the expected value (over indefinite replications) is not the true value [1,2]. A study or process is considered biased if a systematic error is present in study design, data collection, or data analysis [2,3]. Systematic error, using a slight modification of a standard dictionary definition [4], can be defined as a persistent error having a nonzero mean that cannot be attributed entirely to chance but to inaccuracy inherent in the system of measurement. A random error develops from imprecision in a measuring instrument or protocol used to collect data for study. A random error in absence of systematic error will not result in bias if on average the measured value is still the true population value. The effect of random errors will be reduced by increasing sample size or number of measurements taken from each sampling unit. Systematic error will not be reduced by increasing sample size because it does not result from imprecise measurements.

Epidemiologic investigations must consider the potential effects of both systematic and random errors on study results. The odds ratio (OR) is frequently the measure of association estimated in studies concerning etiology and the likelihood ratio (LR) is commonly estimated for evaluation of diagnostic tests. Odds ratios for diagnostic purposes can also be estimated that quantify the change in the odds of infection (or disease) resulting from a positive test result [5].

Estimates of LRs and diagnostic ORs can be affected by random and systematic errors similar to other epidemiologic measures of association. The error in detection of the analyte (biologic substance measured by a diagnostic assay) must exert its effect through misclassification of the test result. The ability of the analyte to predict infection (or lack of infection) in an individual determines its usefulness or accuracy for diagnosis. Accuracy of a testing system is measured by its sensitivity (probability of correctly classifying infected individuals) and specificity (probability of correctly classifying uninfected individuals). Accuracy can be measured at a single or over multiple positive cutoff values.

The evaluation of tests over multiple cutoffs can be performed through estimation of LRs or performing receiver-operating characteristic (ROC) analyses. The LR is a measure of association that quantifies how many more times likely a test result is from an infected individual compared to one that is uninfected. It is calculated as the ratio of the probability that an infected individual will have that test result to the probability that an uninfected individual would have that same result [6]. Calculation of LRs for tests with quantitative outcomes (e.g. titers, optical densities) necessitates dividing up the possible range of test results into categories. Likelihood ratios are also mathematically related to ROC curves as the slope between adjacent test result categories [7].

Receiver-operating characteristic (ROC) curves are formed by plotting 1 – specificity (x-axis) by sensitivity (y-axis) over multiple positive cutoff values [8]. The area under a ROC curve (AUC) is defined as the probability that a randomly selected infected individual will have a greater test result than a randomly selected uninfected individual, and is considered a measure of overall discriminating ability of the test [9]. The precision of a diagnostic testing system will affect the overall accuracy and is often measured as the coefficient of variation (CV), which is calculated as the standard deviation of measurements on the same sample divided by the mean of the measurements. The CV quantifies the random measurement error inherent in the diagnostic system.

Measurement error associated with the analyte could theoretically be differential or non-differential. Differential measurement error is defined as an error whose magnitude or direction is different for individuals who have the outcome (eg. infection) compared to those without the outcome. Non-differential measurement error is an error that is independent of outcome status; the direction and magnitude is equal for those with and without the outcome. Measurement error can lead to misclassification that is differential or non-differential. The effect of measurement error and misclassification on LRs could not be found in the currently available peer-reviewed literature. The direction of bias in estimates of ORs and risk ratios with differential misclassification cannot be predicted [10-12], however, non-differential misclassification of an exposure has been shown to result in measures of association to be consistently biased towards the null when evaluated in a 2 × 2 table [1,10-14] except in unrealistically extreme situations [1,10,15]. When the exposure is classified into more than two categories (higher-order tables) the direction of deviation is no longer consistently towards the null value with non-differential misclassification [10,15-18]. However, even in situations with more than two levels of exposure, the measures of association will be biased towards the null when calculated for the categories involved in the misclassified exposure [19,20].

A quantitative exposure that is categorized into three levels will often assign the lowest category (level 1) as the reference level. The usual ORs evaluating the effect of exposure are calculated comparing level 2 to level 1 and level 3 to level 1. If non-differential misclassification occurs only between exposure levels 2 and 3, for instance, then the usual ORs could be biased towards or away from the null value, however, the OR calculated between level 2 and level 3 (not usually reported) would consistently be biased towards the null. The exception to this rule is when misclassification is so extreme that the probability of incorrect classification is more likely than correct classification [19,20].

The effects of misclassification on measures of association are often studied by creating hypothetical data distributions, but simulation studies using actual data can also be employed [21,22]. Simulation studies have the advantage of defining probability distributions rather than creating extreme and potentially unrealistic situations. The objective of this study was to investigate the effects of non-differential measurement error on AUC, LRs, and diagnostic ORs calculated for a test categorized into four levels using real data and simulated error structures.

## Analysis

### Data source

Brucellosis is a major disease problem worldwide [23] associated with chronic debilitating infections in people and reproductive failure in domestic animals. *Brucella *species that cause disease in people include *B. abortus *(primary reservoir is cattle and water buffalo), *B. melitensis *(sheep and goats), and *B. suis *(swine) [24]. Cattle and domestic water buffalo in Trinidad have been found to be infected with *B. abortus *[25] and the data used for this simulation study are results from a brucellosis competitive ELISA (c-ELISA) in 391 cattle and 381 domestic water buffalo (*Bubalus bubalis*) of Trinidad. Evaluation of this assay has been reported elsewhere [26] and results from both species were pooled in a single analysis for purpose of these evaluations. The brucellosis status (infected or uninfected with *B. abortus*) was determined using results from multiple diagnostic tests in a no gold standard analysis. The most likely infection status based on this analysis was assumed the true status. This classification resulted in 126 cattle and water buffalo infected with *B. abortus *and 656 uninfected cattle and water buffalo.

The c-ELISA is a quantitative test where sample results are reported as the proportion of inhibition compared to a conjugate-only control (no serum added). Each test and control sample had optical density (OD) values measured in duplicate and the formula to calculate the proportion inhibition (PI) is included below.

The theoretical limits are therefore zero to one with values closer to one being more positive (higher level of competing antibodies). Negative values occur infrequently when the OD of the sample is greater than the conjugate control.

### Data simulation

The data measured when performing an ELISA is the degree of color change, or OD, that quantifies the amount of antibodies in the serum. The observed mean OD values for test sera and conjugate-only controls from each ELISA plate were assumed to represent the true biologic value for purpose of these simulations. Commercially available software [27] was used to incorporate error distributions to both sample and control mean values independently. After addition of error to original mean OD values, the PI was re-calculated for each sample.

Conjugate-only control samples contain no competing antibodies and therefore the color change (i.e. OD) should be equal to a baseline level. Variation in the measured values for these controls represents the random error associated with the assay. Therefore, mean ODs measured for the duplicate conjugate controls on the original ELISA plates were used to estimate the inherent error of the testing system and determine simulation error distributions. Normal distributions with means of 0, 0.1, -0.1 and standard deviation of 0.12 and mean of 0 and standard deviation of 0.24 were evaluated as part of the study. A value of 0.1 was chosen for a mean because it was the interquartile range for the average of duplicate conjugate control values on each ELISA plate. A standard deviation of 0.12 was chosen because this was the standard deviation of all original mean conjugate control values. A random sample from these distributions was added to observed mean OD values. Lognormal distributions were used to add an error structure that varied depending upon observed ODs. The scale (μ) parameter of these distributions was calculated as the observed mean OD of the particular sample divided by the mean OD of all sample values. The shape (σ) parameters investigated were 0.12 and 0.24. A random selection from these distributions was multiplied by the observed mean OD to calculate simulated values. Simulated mean ODs were not truncated in range and calculated PI values could be less than zero and greater than one.

Test results were divided into four categories: <0.25, 0.25 – 0.349, 0.35 – 0.499, and ≥ 0.50 PI. These categories were based on an evaluation of this assay [28] with the original six categories collapsed to four to reduce complexity of simulations and increase the number of infected and uninfected individuals in the lower most and upper most categories, respectively. Category-specific LRs [7] were calculated for each of the four categories as the proportion of infected individuals in each category divided by the proportion of uninfected individuals within that same category. Diagnostic ORs were calculated comparing the three higher test result categories to the lowest category as the baseline, or reference level. Sensitivity and specificity were calculated for the c-ELISA at all possible cutoff values from 0.01 to 0.99 PI at 0.01 intervals. Area under the ROC curve was calculated as an overall measure of diagnostic accuracy using the trapezoid method [29]. The average sensitivity between adjacent cutoffs was the mean height of the trapezoid and base width was the difference in adjacent specificities.

Six simulation studies were performed independently assessing the impact of added error distributions to the original observed data. Monte Carlo sampling was performed of these error distributions independently for each test sample and conjugate control over 10,000 iterations. Error was added to all mean OD values at each iteration, new PIs were calculated, and diagnostic accuracy measures (AUC, LR, OR) were determined. The mean, median, standard deviation, minimum, and maximum values of PIs for infected and uninfected individuals were calculated at each iteration. Median values and percentiles over these 10,000 iterations were used as point estimates and confidence intervals, respectively for descriptive statistics and all investigated AUCs, LRs, and ORs.

### Simulation results

The six added error structures caused mean PI values to have greater range and larger standard deviations for both infected and uninfected groups of individuals compared to original values and decreased overall test accuracy as measured by AUC (Table (Table1).1). Distribution of c-ELISA PI values for Normal (0, 0.12) and lognormal (0.24) error structures were noticeably different from the original distribution for uninfected individuals (Fig (Fig1)1) and relatively similar for infected individuals (Fig (Fig2).2). Added error with different means but the same standard deviations resulted in visually similar distributions (data not shown). Distribution of PIs in uninfected individuals peaked at zero because all lower extreme values were included in the 0–5% test result category. In general, distribution of PIs with added error had a wider (less precise) distribution, which resulted in more overlap with distribution of infected individuals and lowered overall test accuracy. Added error caused point estimates of LRs to be biased towards the null value (1.0) for all categories except 0.25 – 0.349 (Table (Table2).2). Results for the 0.35 – 0.499 category also extended beyond the null value for some error structures. Diagnostic ORs calculated with the lowest category as baseline were consistently biased towards the null for all evaluated error structures. Error structures with larger variance resulted in more bias for both LRs and ORs.

**...**

*B. abortus*uninfected cattle and water buffalo from Trinidad with added error from a single iteration of a simulation study and summed over 5% proportion inhibition intervals.

*B. abortus*infected cattle and water buffalo with added error from a single iteration of a simulation study and summed over 5% proportion inhibition intervals.

## Conclusion

The effect of non-differential random error in exposure measurement has been discussed in previous publications [21,30-32], and leads to measures of association being biased towards the null value except in unrealistically extreme situations. Overall accuracy of a quantitative diagnostic test, measured via the AUC, has been shown here to also be decreased (biased towards null value of 0.5) through addition of non-differential measurement error. The variance component of the measurement error structure appears to have an important effect on decreasing AUC and the systematic (mean) component of the error structure has little or no bearing on results when applied equally to all samples (ie. non-differential). This attenuation in accuracy is due to added variability spreading out the distribution of test results and creating more overlap between results from infected and uninfected individuals as shown in the figures. Values were not truncated during simulations despite the fact that biologically unusual values were observed as evidenced by the PI ranges. These observations did not unduly influence the analysis because they were considered equivalent to the boundary categories.

Likelihood ratios are derived from the odds version of Bayes' theorem [28], correspond to the added information provided by a test, and are used to update the prior odds of infection. The null value of a LR is one, which would correspond to a particular test result being equally likely in infected and uninfected individuals (would not affect prior probabilities). A previous study [31] demonstrated that non-differential, random measurement error in exposure determination without a systematic component, consistently led to attenuation in effect measures such as the OR. Results for diagnostic ORs agreed with this previous finding; however, LRs estimated in the present study were not consistently biased towards the null value. For example, the baseline (without error) LR for the 0.25 – 0.349 category was 0.675 and all evaluated error structures resulted in this LR (based on median simulated value) to be biased further away from one. The baseline LR for the 0.35 – 0.499 category was 3.32 and four of the evaluated error structures resulted in this measure to be biased to such an extent that the point estimates extended below the null value of one. Estimates of LRs and ORs were also mildly affected by the systematic component (mean) of the error structure. Unlike the AUC, these measures are dependent upon the underlying distribution of values because they are calculated for a small number of fixed categories.

The direction of bias is not easily described as being away from or towards the null value for investigated LRs. However, all LRs estimated from evaluated error structures could be described as being biased in a manner resulting in the test having less discriminating ability for its usual function at that category. For instance, the lowest category is often useful for "ruling out" infection given that (i.e. "negative") test result. The observed bias towards one causes a test result falling within this category to be less useful for that purpose. Larger test result categories are more positive (0.30 is the usual positive cutoff for the brucellosis c-ELISA [33]) and therefore a higher post-test probability of infection (compared to pre-test probability) would be the desired effect. All error structures resulted in LRs for these categories to be biased towards zero, which means that the test is less useful for this purpose (calculated post-test probability of infection lower than the true value). The observed direction of the biases in this study might have resulted from the underlying distributions of test results for infected and uninfected individuals and a different dataset might not demonstrate the same relationships.

A strength of the employed simulation procedure is that error distributions were added to the mean ODs measured from duplicate test and control samples. Optical densities, and not PIs, would be the values actually affected by measurement error. A similar analysis adding error to the PIs would not directly simulate this type of error. Lognormal error distributions were evaluated to simulate measurement error that depended on the magnitude of the measured value. In the example of a c-ELISA, higher OD values correspond to more negative (fewer competing antibodies) samples. Therefore, this error structure has a greater impact on the distribution of values in uninfected individuals as was seen in the presented figures. Investigated error structures might overestimate true measurement error and only a limited number of distributions were evaluated leading to difficulty in generalizing results to all possible error situations. However, added error distributions were based on true observations from the mean conjugate-only controls that have no competing antibodies. Therefore, variability inherent in these measurements should be a valid representation of the true variability of the testing system. It is expected that some sources of error would be dependent upon plate-level and day-level factors such as reagents, laboratory temperature, and incubation times that would be equal for both test and control samples. Therefore, the addition of non-differential error independently to test and control values represents the upper limit of possible effects on test accuracy measures.

Non-differential random error added via a probability distribution might result in differential misclassification of test result categories as evidenced in data presented in Table Table2.2. The proportion of individuals misclassified in the four test result categories does not appear to be equal between infected and uninfected individuals. A similar finding has been reported for non-differential measurement error of exposure [21,31,32]. In this study, however, it is impossible to know which particular individuals were misclassified because only total counts could be calculated and an accurate assessment of the magnitude of misclassification could not be determined. It is only possible to know the net result of the misclassification and not the number of individuals incorrectly entering or leaving each category. The misclassification across the test result categories also depends upon the underlying distribution of values.

The true infection status of individuals in the evaluated dataset was not known and classification of individuals was performed based on results of a no gold standard test evaluation study. Therefore, the original data is expected to contain some results that were misclassified based on infection status. These errors are not expected to unduly affect results of the simulation study because they would apply equally to the baseline and error-augmented situations. The underlying distributions of test results in infected and uninfected individuals, however, might not adequately reflect the true distributions because of this potential misclassification.

This study shows that non-differential measurement error can lead to biased results in the evaluation of diagnostic tests with quantitative outcomes. It is especially important to recognize that LRs are not consistently biased towards the null even when measurement error is exclusively non-differential. These biases will not be reduced by simply increasing the sample size; it would be necessary to increase the number of observations on each sampling unit to reduce the impact of this error. It is therefore possible for an unbiased study (presence of random error without a systematic component) to yield biased population values through non-differential measurement error. This situation is possible when the population parameter to be estimated by the study (e.g. LR of a test) is not a simple one to one transformation of the data affected by measurement error (e.g. OD). The observed attenuation in AUC would be expected to occur in all situations involving non-differential measurement error, but the direction of bias in measured LRs would be expected to vary depending upon the amount of error and underlying distribution of test results.

## Abbreviations

OR – odds ratio

LR – likelihood ratio

ROC – receiver-operating characteristic

AUC – area under receiver-operating characteristic curve

CV – coefficient of variation

c-ELISA – competitive enzyme-linked immunosorbent assay

OD – optical density

PI – proportion inhibition

## Competing interests

The author declares that he has no competing interests.

## Authors' contributions

GTF performed all analyses and wrote the manuscript without substantive contributions from other investigators.

## Acknowledgements

I would like to thank Dr. Saraya Tavornpanich for helpful suggestions that led to improvements in the paper.

## References

- Wacholder S, Hartge P, Lubin JH, Dosemeci M. Non-differential misclassification and bias towards the null: a clarification.[comment] Occupational & Environmental Medicine. 1995;52:557–558. [PMC free article] [PubMed]
- Szklo M, Nieto FJ. Epidemiology : beyond the basics. Gaithersburg, Md: Aspen; 2000. pp. 125–126.
- Rothman KJ. Epidemiology : an introduction. New York, N.Y.: Oxford University Press; 2002. pp. 94–95.
- Webster's new millennium dictionary ofEnglish Preview edn. Long Beach, CA: Lexicon Publishing Group; 2005. http://dictionary.reference.com/search?q=systematic%20error 3-16-06.
- Pepe MS, Janes H, Longton G, Leisenring W, Newcomb P. Limitations of the odds ratio in gauging the performance of a diagnostic, prognostic, or screening marker. American Journal of Epidemiology. 2004;159:882–890. doi: 10.1093/aje/kwh101. [PubMed] [Cross Ref]
- Simel DL, Samsa GP, Matchar DB. Likelihood ratios for continuous test results – making the clinicians' job easier or harder?[erratum appears in J Clin Epidemiol 1993 Nov;46(11);1295] Journal of Clinical Epidemiology. 1993;46:85–93. doi: 10.1016/0895-4356(93)90012-P. [PubMed] [Cross Ref]
- Choi BC. Slopes of a receiver operating characteristic curve and likelihood ratios for a diagnostic test. American Journal of Epidemiology. 1998;148:1127–1132. [PubMed]
- Zweig MH, Campbell G. Receiver-operating characteristic (ROC) plots: A fundamental evaluation tool in clinical medicine. Clinical Chemistry. 1993;39:561–577. [PubMed]
- Greiner M, Pfeiffer D, Smith RD. Principles and practical application of the receiver-operating characteristic analysis for diagnostic tests. Preventive Veterinary Medicine. 2000;45:23–41. doi: 10.1016/S0167-5877(00)00115-X. [PubMed] [Cross Ref]
- Armstrong BG. Effect of measurement error on epidemiological studies of environmental and occupational exposures. Occupational & Environmental Medicine. 1998;55:651–656. [PMC free article] [PubMed]
- Mertens TE. Estimating the effects of misclassification.[see comment] Lancet. 1993;342:418–421. doi: 10.1016/0140-6736(93)92820-J. [PubMed] [Cross Ref]
- Copeland KT, Checkoway H, McMichael AJ, Holbrook RH. Bias due to misclassification in the estimation of relative risk. American Journal of Epidemiology. 1977;105:488–495. [PubMed]
- Gladen B, Rogan WJ. Misclassification and the design of environmental studies. American Journal of Epidemiology. 1979;109:607–616. [PubMed]
- Barron BA. The effects of misclassification on the estimation of relative risk. Biometrics. 1977;33:414–418. doi: 10.2307/2529795. [PubMed] [Cross Ref]
- Birkett NJ. Effect of nondifferential misclassification on estimates of odds ratios with multiple levels of exposure. American Journal of Epidemiology. 1992;136:356–362. [PubMed]
- Correa-Villasenor A, Stewart WF, Franco-Marina F, Seacat H. Bias from nondifferential misclassification in case-control studies with three exposure levels. Epidemiology. 1995;6:276–281. [PubMed]
- Wacholder S. When measurement errors correlate with truth: surprising effects of nondifferential misclassification. Epidemiology. 1995;6:157–161. [PubMed]
- Dosemeci M, Wacholder S, Lubin JH. Does nondifferential misclassification of exposure always bias a true effect toward the null value?[see comment] American Journal of Epidemiology. 1990;132:746–748. [PubMed]
- Peeters PHM. Re: "Does Nondifferential Misclassification of Exposure Always Bias a True Effect toward the Null Value". American Journal of Epidemiology. 1991;134:439–440.
- Brenner H. Re: "Does Nondifferential Misclassification of Exposure Always Bias a True Effect toward the Null Value". American Journal of Epidemiology. 1991;134:438–439. [PubMed]
- Flegal KM, Keyl PM, Nieto FJ. Differential misclassification arising from nondifferential errors in exposure measurement. American Journal of Epidemiology. 1991;134:1233–1244. [PubMed]
- Sorahan T, Gilthorpe MS. Non-differential misclassification of exposure always leads to an underestimate of risk: an incorrect conclusion.[see comment] Occupational & Environmental Medicine. 1994;51:839–840. [PMC free article] [PubMed]
- Corbel MJ. Brucellosis: An overview. Emerging Infectious Diseases. 1997;3:213–221. [PMC free article] [PubMed]
- Young EJ. An overview of human brucellosis. Clinical Infectious Diseases. 1995;21:283–290. [PubMed]
- Fosgate GT, Adesiyun AA, Hird DW, Hietala SK, Ryan J. Isolation of Brucella abortus biovar 1 from cattle and water buffalo of Trinidad. The Veterinary Record. 2002;151:272–273. [PubMed]
- Fosgate GT, Adesiyun AA, Hird DW, Johnson WO, Hietala SK, Schurig GG, Ryan J. Estimation of receiver-operating characteristic curves to determine accuracy of a competitive enzyme-linked immunosorbent assay for the serodiagnosis of Brucella infection in domestic water buffalo (Bubalus bubalis) and cattle. American Journal of Veterinary Research. 2003;64:57–64. doi: 10.2460/ajvr.2003.64.57. [PubMed] [Cross Ref]
- @Risk Version 452. Palisade Corporation, Ithaca, NY, USA;
- Fosgate GT, Adesiyun AA, Hird DW, Hietala SK. Likelihood ratio estimation without a gold standard: a case study evaluating a brucellosis c-ELISA in cattle and water buffalo of Trinidad. Preventive Veterinary Medicine. 2006;75:189–205. [PubMed]
- Munem MA, Foulis DJ. Calculus with analytic geometry. 2. New York, N.Y.: Worth Publishers; 1984. pp. 338–342.
- Gustafson P, Le Nhu D. Comparing the effects of continuous and discrete covariate mismeasurement, with emphasis on the dichotomization of mismeasured predictors. Biometrics. 2002;58:878–887. doi: 10.1111/j.0006-341X.2002.00878.x. [PubMed] [Cross Ref]
- Brenner H, Loomis D. Varied forms of bias due to nondifferential error in measuring exposure. Epidemiology. 1994;5:510–517. [PubMed]
- Brenner H, Blettner M. Misclassification bias arising from random error in exposure measurement: implications for dual measurement strategies. American Journal of Epidemiology. 1993;138:453–461. [PubMed]
- Nielsen KH, Kelly L, Gall D, Bossé J, Pulkkinen W. A competitive enzyme immunoassay for the detection of serum antibody to Brucella abortus: APHD Standard Protocol Version 20. Nepean, Ontario, Canada: Animal Diseases Research Institute; 1994.

**BioMed Central**

## Formats:

- Article |
- PubReader |
- ePub (beta) |
- PDF (259K)

- Varied forms of bias due to nondifferential error in measuring exposure.[Epidemiology. 1994]
*Brenner H, Loomis D.**Epidemiology. 1994 Sep; 5(5):510-7.* - Receiver-operating characteristic curves and likelihood ratios: improvements over traditional methods for the evaluation and application of veterinary clinical pathology tests.[Vet Clin Pathol. 2006]
*Gardner IA, Greiner M.**Vet Clin Pathol. 2006 Mar; 35(1):8-17.* - The number needed to treat (NNT) can be adjusted for bias when the outcome is measured with error.[J Clin Epidemiol. 2004]
*Marschner IC, Emberson J, Irwig L, Walter SD.**J Clin Epidemiol. 2004 Dec; 57(12):1244-52.* - A non-inferiority test for diagnostic accuracy based on the paired partial areas under ROC curves.[Stat Med. 2008]
*Li CR, Liao CT, Liu JP.**Stat Med. 2008 May 10; 27(10):1762-76.* - Most clinical tests cannot accurately diagnose rotator cuff pathology: a systematic review.[Aust J Physiother. 2008]
*Hughes PC, Taylor NF, Green RA.**Aust J Physiother. 2008; 54(3):159-70.*

- Bayesian adjustment for measurement error in continuous exposures in an individually matched case-control study[BMC Medical Research Methodology. ]
*Espino-Hernandez G, Gustafson P, Burstyn I.**BMC Medical Research Methodology. 1167* - Characteristics of Mammography Facility Locations and Stage of Breast Cancer at Diagnosis in Chicago[Journal of Urban Health : Bulletin of the N...]
*Tarlov E, Zenk SN, Campbell RT, Warnecke RB, Block R.**Journal of Urban Health : Bulletin of the New York Academy of Medicine. 2009 Mar; 86(2)196-213*

- MedGenMedGenRelated information in MedGen
- PubMedPubMedPubMed citations for these articles
- TaxonomyTaxonomyRelated taxonomy entry
- Taxonomy TreeTaxonomy Tree

- Non-differential measurement error does not always bias diagnostic likelihood ra...Non-differential measurement error does not always bias diagnostic likelihood ratios towards the nullEmerging Themes in Epidemiology. 2006; 3()7PMC

Your browsing activity is empty.

Activity recording is turned off.

See more...