Chapter 54 Screening: Relationship to Diagnosis and Therapy M. Cristina Leske, Barbara S. Hawkins, Leslie Hyman and Päivi H. Miskala Table Of Contents |
OVERVIEW APPLICATIONS OF SCREENING REFERENCES |
OVERVIEW | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Although primary prevention is directed to averting the initial occurrence
of disease, for example, through immunizations, secondary prevention
aims to improve outcome of disease by early detection and treatment.1 Screening, typically considered as secondary prevention, seeks to detect
disease at a preclinical stage with the expectation that early treatment
will deter disease progression. DEFINITIONS Screening has been defined as the presumptive identification of unrecognized disease or defect by the application of tests, examinations, or other procedures that can be applied rapidly. Screening tests sort out apparently well persons who probably have a disease from those who probably do not.2 By this definition, screening is conducted on asymptomatic, presumably healthy individuals who are tested to determine whether they have high or low probabilities of disease. The results of screening, therefore, are tentative, and neither positive nor negative findings can be considered conclusive. Positive screening results do not necessarily imply disease; they only indicate the need for additional evaluation, typically by more extensive and more expensive methods. Similarly, negative screening results cannot be interpreted as conclusive evidence of the absence of disease. Screening activities can be classified as “mass” or “selective.”3 Mass screening, or population-based screening, is provided on a large scale to unselected populations, such as visitors to health fairs, shopping centers, or other ad hoc settings. Selective screening, or high-risk screening, is limited to persons who have already been identified as being at an increased risk for a particular disease. Tools suitable for mass screening are typically less invasive and less expensive than screening tools to be applied in select population settings.4 Screening may be performed on volunteers in different community settings or on patients seeking health care for unrelated symptoms. Identifying previously unrecognized disease by screening among patients consulting for an unrelated disease (e.g., glaucoma screening of a patient consulting for myopia) is sometimes defined as case finding. When implementing a screening program, it is appropriate to provide follow-up evaluation for persons with positive screening tests. Screening is more likely to identify cases with a long preclinical phase of the disease than cases with a short preclinical phase.5 CRITERIA Screening raises ethical, clinical, as well as scientific issues, and the decision to screen for a particular disease must be evaluated carefully. Several general principles have been proposed to assist in this evaluation.3,6–9 To be suitable for population-based screening, a disease should meet the following criteria:
The rationale for each criterion is discussed below. Morbidity and Mortality A condition merits screening only if it has an important public health impact and is serious enough to affect the quality or quantity of life. Prevalence Ideally, screening should be aimed at detecting relatively common conditions within the population targeted. As the prevalence of a condition increases in the target population, screening yields more cases and the cost per case detected decreases.9 Natural History Knowledge of the course of the disease process is essential for screening.10 In some conditions, there is a marked overlap in measurements between diseased and nondiseased persons, and the diagnosis is not clear cut. To be suitable for screening, a disease must have a clearly recognized biologic onset and a diagnosis that can be confirmed by accepted criteria. The disease should also have a latent, asymptomatic stage before it becomes clinically apparent (see Fig. 1). The presence of this asymptomatic stage will permit detection and intervention before the usual time of diagnosis. The time lapse between early detection by screening and the usual time of diagnosis is known as the lead time.8,11
Effectiveness and Availability of Treatment Screening is of value when the disease can be effectively treated or controlled. With incurable hereditary conditions, control of disease may be possible with counseling.12 In addition to being effective, the treatment must be accessible. Screening is justified only when facilities for diagnosis and treatment are available to persons with positive screening results. A major issue to emphasize, because it is often overlooked, is that resources for effective follow-up must be an integral part of a screening program. Better Outcome with Early Treatment The rationale for screening is that early detection and treatment will improve disease outcome. This goal can be attained only if intervention in the asymptomatic stage (Fig. 1) produces better results than when treatment is begun after the usual time of diagnosis. Screening Tests An important prerequisite for a screening test is its acceptability to the person being screened and to those performing and interpreting the test. Other requirements are simplicity, so the test can be easily administered and reliability or reproducibility, so that consistent results may be expected on repeated measurements by the same or a different observer. An important requirement is test validity (i.e., the ability to correctly identify diseased and nondiseased persons).1,4 Validity is measured by sensitivity and specificity (see Table 1). Sensitivity is the ability of a test to identify cases (i.e., diseased persons) correctly. Thus, a screening test with 90% sensitivity will be positive in 90 of 100 cases screened. Specificity is the ability of a test to identify noncases (i.e., nondiseased persons) correctly. Thus, a test with 95% specificity will be negative in 95 of every 100 noncases tested. In the example given in Table 1, Test A has a 70% sensitivity and 80% specificity.
TABLE 1. Example of Sensitivity, Specificity, and Predictive Values of a Screening
Test, Assuming an Eye Disease with a Prevalence of 10% in the
Target Population
Sensitivity and Specificity The denominator for these calculations is the number of cases (a + c) or noncases (b + d) of Eye Disease X.
Predictive Values The denominator for these calculations is the number of persons with positive tests (a+b) or negative tests (c + d).
Sensitivity and specificity should always be evaluated together. A test with high sensitivity will detect most of the cases and will thus have few false-negative results. However, if the same test has low specificity, many false-positive results will occur and lead to overreferrals. In contrast, a test with low sensitivity and high specificity will have many false-negative results but few false-positive results. Ideally, a test should have high sensitivity and high specificity, but this goal is difficult to achieve in practice. Because the test results of diseased and nondiseased persons usually overlap, some cases have low (negative) test values and some noncases have high (positive) values (see Fig. 2). The specific test value chosen as a cutoff to define a “positive” screening result will affect both the sensitivity and the specificity. If a low cutoff value is chosen, the sensitivity will be high, because most cases will have “positive” screening results and will be referred for further diagnostic examinations. The specificity, however, will be low because many noncases will also be referred. Increasing the test value chosen as “positive” to trigger referral will increase the specificity, at the expense of the sensi-tivity. The decision to select a specific test value to determine referral depends on the disease being detected. If the failure to detect cases has serious consequences, for example, nondetection of malignant disease, a low test value should be chosen for referral, thus increasing sensitivity even though the specificity will decrease. If an excessive number of overreferrals is unacceptable, for example, when follow-up of screening requires invasive procedures, a high screening test value should be chosen for referral to increase specificity, although the sensitivity will decrease.
The evaluation of tests also includes “predictive values.” Sensitivity and specificity evaluate the ability of a test to separate cases correctly from noncases and are measures of validity, but predictive values are not. The predictive value of a positive test refers to the percentage of cases found among all those with positive tests (Table 1). In the example in Table 1, where Disease X had a prevalence of 10%, 250 persons had positive tests and 70 truly had the disease, resulting in a positive predictive value of 28% for Test A in this population. The predictive value of a negative test is defined as the percentage of noncases found among persons with negative tests (Table 1). Therefore, if 720 of 750 persons with negative tests were noncases, the negative predictive value was 96%. Predictive values depend not only on sensitivity and specificity of the test, but also on the prevalence of disease in the population screened. As the prevalence of the disease increases, the positive predictive value increases (see Table 2). Therefore, false-positive results are reduced when screening is performed in populations where the disease is common; conversely, screening leads to a large number of overreferrals when carried out in populations where the disease is rare. For this reason, the cost per case detected increases when the population screened has a low prevalence of disease.
TABLE 2. Positive Predictive Values by Disease Prevalence at Selected Levels of
Sensitivity and Specificity
TWO-STAGE SCREENING. Sometimes a screening program is multiphasic, that is, involves a series of sequential tests.4 Usually, an initial screening test that is inexpensive and noninvasive is performed first; then those with positive results are retested using a more accurate test, which is typically more expensive and/or more invasive.4 In this situation, sensitivity and specificity of the two screening tests can be combined and used in sequence, being referred to as net sensitivity and net specificity.4 These concepts are illustrated through the example that assumes multiphasic screening for Disease X, first using Test A (see Table 1) and then applying Test B (see Table 3) to the subset that screened positive with Test A.
TABLE 3. Example of Two-Stage Screening, Net Sensitivity and Net Specificity
for a Sample Eye Disease X, Based on Screening Test A Followed by
Screening Test B
(Test B: Sensitivity = 90%; specificity = 90%) Net Sensitivity True positives from Test B: 63 = 63% All cases in population for Test A: 100 Net specificity: True negatives from Test A + True negatives from Test B: 720 +162 = 98% All noncases in population for Test A: 900
Net sensitivity and net specificity are derived in two stages. Stage 1 evaluates the initial screening test for the total population, as presented for Test A in Table 1. For Stage 2, individuals who tested positive by Test A are rescreened using Test B. In the example, the 250 individuals who tested positive, as in Table 1, would be retested with Test B, which has 90% sensitivity and 90% specificity as indicated in Table 3. To calculate net sensitivity, the numerator is the number of individuals who were identified as true positives (i.e., tested positive and have the disease) by Test B, which is 63 in the example, and the denominator includes the total number of cases (i.e., the sum of true positives and false negatives) in the target population initially screened with Test A, which is 100 in the example. Therefore, net sensitivity is 63% and is lower than the sensitivity of either test, being equivalent to the sensitivity of Test A times the sensitivity of Test B. Net specificity is calculated by defining the numerator as the sum of the true negatives (i.e., tested negative and do not have the disease) identified by Tests A (n = 720) and B (n = 162) and the denominator as the total number of noncases (i.e., a sum of true negatives and false positives) in the population (n = 900), evaluated by Test A. Therefore, the net specificity is 98%, which is higher than the specificity of Test A and Test B, resulting in an overall gain by using the two screening tests. As demonstrated by this example, retesting individuals who initially test positive will increase specificity, thus decreasing the likelihood of over-referrals due to false-positive tests. Positive predictive value also increases by retesting persons in this group, because they have a higher prevalence of the disease. Cost-Effectiveness Aside from the purely humanitarian and social benefits of preventing morbidity, the cost-benefit of screening must be evaluated. Although screening may ultimately reduce the public health impact of a disease and bring about economic savings, the screening process itself generates costs. A common problem is the lack of inexpensive tests that can effectively separate cases from noncases. In addition to the resources expended in the testing process itself, these costs include the follow-up and diagnosis of persons with positive tests, as well as the costs of treating all the newly detected cases. The errors in classifying persons as “positives” and “negatives” also have a cost. Besides being subjected to unnecessary diagnostic tests for a condition they do not have, persons with false-positive tests may suffer other undesirable consequences, such as anxiety and worry at being considered a disease suspect. Additionally, costs are incurred by persons with false-negative screening tests, who may derive false reassurance from screening. Thus, screening for eye conditions may be justified if the cost of a screening program and associated services is less than the costs incurred when the disease is detected at the usual time of diagnosis, for example, the costs of providing long-term services for the visually disabled. It is also justified if earlier diagnosis will diminish psychologic and/or other negative consequences associated with the ophthalmologic condition. |
APPLICATIONS OF SCREENING | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Few conditions meet all the criteria for screening, and the issues related
to screening may be different depending on the condition. To illustrate
issues related to screening in ophthalmology, the remaining section
will focus on glaucoma and amblyopia screening. GLAUCOMA SCREENING Primary open-angle glaucoma is a major cause of blindness and visual impairment throughout the world, particularly affecting persons of African descent.13 Because the visual disability caused by glaucoma is knownto be preventable by early treatment, efforts have been made to identify the disease in its asymptomatic stages. To assess the value of these efforts, it is necessary to determine how well the disease meets the criteria required for screening. First, the disease must be defined. Definitions The glaucomas are a group of conditions characterized by progressive optic nerve damage.14 Primary open-angle glaucoma occurs in the presence of an open iridocorneal angle and without any apparent cause; the diagnosis is made after other possible types of glaucoma have been excluded. The primary open-angle type of glaucoma is the most frequent among adults of European and African descent and is an important cause of visual disability. The discussion that follows is limited to screening for primary open-angle glaucoma, referred to hereafter as glaucoma, for brevity. The diagnosis of glaucoma is made in the presence of nerve fiber bundle defects of the visual field, accompanied by structural changes of the optic disc (cupping or excavation), in the absence of other causes. An intraocular pressure over the population mean often, but not always, accompanies the disease. It is well recognized that field defects and disc cupping may occur at any intraocular pressure level. Therefore, glaucoma definitions from large epidemiologic studies are based on the presence of disc and/or field changes and, with few exceptions, do not use specific numerical values of intraocular pressure as a diagnostic criterion.13–29 Glaucoma definitions, however, are not standardized. Some reports distinguish normal tension glaucoma (disc and field changes with intraocular pressure less than or equal to 21 mm Hg) as a subcategory of open-angle glaucoma.22 For the purposes of this discussion, the diagnosis of primary open-angle glaucoma requires characteristic visual field and disc changes in the absence of other possible causes, regardless of intraocular pressure. Persons with pressures over 21-mm Hg, but without disc or field changes, are categorized as having ocular hypertension not glaucoma.30,31 Criteria Traditionally, the most commonly used method for early identification of glaucoma has been mass tonometry screening, implemented in many community settings. This approach, however, has limitations, and its effectiveness has long been in question,32–39 thus leading to the search for better screening methods. The rationale for developing new methods and approaches to population screening rests on the assumption that detecting undiagnosed glaucoma is a desirable public health measure. To evaluate the validity of this assumption, a discussion of the criteria to justify screening, as applied to glaucoma, follows. PUBLIC HEALTH IMPACT. Glaucoma is a leading cause of blindness worldwide and is among the three leading causes of blindness in the United States, being the major reason for blindness registration among the African-American population.40 Although precise data are difficult to obtain, over 2 million Americans were estimated to have glaucoma in 2003 and about 90,000 were thought to be blind from the disease.41 Available sources suggest that glaucoma is responsible for 11% to 13% of existing blindness,42,43 a frequency that increases to over one-fourth among persons of African origin.44,45 In one U.S. study, primary open-angle glaucoma was 6.6 times higher in the black than in the white participants, and glaucoma blindness was seen at an earlier age, markedly increasing with age.43 These and other health care utilization data13 indicate that glaucoma is an important cause of visual impairment, especially among the African-derived population. When assessing the impact of glaucoma, it is necessary to consider its effects on quality of life, which are largely dependent on the stage of the disease. The quality of life of glaucoma patients has been studied using various methods and approaches. Results consistently show a clear relationship with degree of visual function, particularly visual acuity.46–50 Most glaucoma patients, however, do not have important visual losses until advanced stages of the disease, thus suggesting that early glaucoma may have few or no effects on vision function-related quality of life. Alternatively, it may be that current instruments are not sufficiently sensitive to assess these effects. Using available methods, the impact of glaucoma is mainly evident among patients who have later stages of the disease. PREVALENCE. The prevalence of glaucoma is best determined by multiple diagnostic methods, including tonometry, ophthalmoscopy, and visual field testing of every survey participant. The first prevalence studies of glaucoma were conducted in Europe, starting in the 1960s, but a large number of studies from various parts of the world have followed (see Table 4).15–29 The earlier prevalence studies initially screened for glaucoma with tonometry and sometimes with ophthalmoscopy, with referral for visual field tests being limited to a subset of participants, such as the glaucoma suspects. This two-stage testing protocol leads to an undercounting of cases having ocular tensions below the cutoff for positive tonometry screening. In contrast, recent studies have emphasized methods other than tonometry for preliminary screening and have not included elevated intraocular pressure as a diagnostic criterion. As seen in Table 4, the variations in definitions of elevated IOP or ocular hypertension have led, at least in part, to differences in the prevalence of these conditions.
TABLE 4. Estimated Prevalence of High Intraocular Pressure and Glaucoma Defects
at Screening in Population-Based Studies of Persons Older Than 40 Years of Age When glaucoma is defined on the basis of optic disc changes with field defects, prevalence is around 1% to 2% for white populations aged 40 years and older (Table 4). Prevalences of open-angle glaucoma are markedly higher among black populations,19–21,29 especially those reported from studies in the Caribbean Islands of St. Lucia20 and Barbados.21 The prevalence of intraocular pressure over 21 mm Hg is also higher in these areas (Table 4).20–21 All studies show a sharp increase in glaucoma prevalence with age. At 70 years of age, prevalences reach 1% to 3% in white populations13,18,19 and 8% to 14% in black populations.19–21 For the purposes of screening, the prevalence of undetected cases is of most interest, since these persons are the target for screening efforts. A consistent result among population studies is that at least half of the cases were newly discovered and were unaware of their glaucoma diagnosis.19,21,24,26,27 Although the degree of severity of these undetected cases is seldom reported or compared to the known cases, they are likely to have an earlier stage of disease than persons with a known glaucoma diagnosis. Supportive evidence is provided by a Swedish study, where glaucoma cases newly identified in a population screening had significantly better visual field status and lower intraocular pressure than self-selected cases.51 If the goal of a screening program is to find all undetected cases, these results indicate that the program must include tests that are sufficiently sensitive to identify early glaucoma. NATURAL HISTORY. The reasons why open-angle glaucoma develops are not well understood, but age, high intraocular pressure, family history, African ancestry, and myopia have been confirmed as risk factors in many studies. (Hypertension and diabetes are positively associated with elevated intraocular pressure, but a similar relationship to glaucoma has not been shown consistently.13,27,52–54) Since high intraocular pressure often accompanies the disease, the distinction between “ocular hypertension” and glaucoma is not always clear cut, leading to a potential overlap between diseased and nondiseased persons.30,31 However, the natural history does not always consist of an orderly progression from an asymptomatic stage with high intraocular pressure to a clinical stage with optic nerve damage and visual field loss. Glaucoma damage may occur at any tension level and is not necessarily preceded by an increase in intraocular pressure.55,56 Although many persons sustain high intraocular pressures without damage, others develop field loss; it is not yet possible to distinguish between persons who will develop damage from glaucoma and those who will not. Ocular hypertensives are at higher risk of optic nerve damage, but the magnitude of this risk is small, estimated at less than 1% per year.13 Incidence was somewhat higher in the black population of the Barbados Eye Study cohort, where 5% of ocular hypertensives developed OAG after 4 years of follow-up.56 Consistent results were found in the Ocular Hypertension Treatment Study, suggesting a higher glaucoma risk in African-American participants.57 A requisite for screening is that the disease have a long preclinical stage to allow early detection. The natural course of glaucoma has been difficult to determine, because under usual standards of practice, all diagnosed patients are given intraocular pressure-lowering treatment. Recently, some information on natural history has been obtained from clinical trials with untreated and treated arms. These data indicate there is large variability in clinical course among individuals, with some patients progressing rapidly and others remaining very stable for years, with and without treatment.58,59 This variability could be explained by the presence of factors related to progression,60 which should be considered when planning patient management. EFFECTIVENESS AND AVAILABILITY OF TREATMENT. At present, medical and surgical therapies for open-angle glaucoma are based on lowering the intraocular pressure. This approach to treatment has been assumed effective in preventing, but not reversing, visual loss. Until recently, the available evidence on the effectiveness of intraocular pressure–lowering treatment to decrease the progression of glaucoma was mainly derived from nonrandomized studies and from clinical trials including ocular hypertensives.61 To demonstrate effectiveness, it was necessary to have evidence from randomized, controlled clinical trials that compared the frequency of visual field progression in treated and untreated patients with glaucoma. The Collaborative Normal Tension Glaucoma Study compared progression in treated versus untreated eyes of patients with a median IOP of 20-mm Hg or less.58 Although results showed a slower progression in treated eyes after controlling for the effects of cataract, the intent-to-treat analysis revealed no significant difference in visual field progression between groups.58 Definitive evidence on the effectiveness of treatment to slow progression in various types of glaucoma was provided by the Early Manifest Glaucoma Trial.59 The results are highly relevant to population-based screening, because participants in the trial were all previously undetected and largely identified from a specific population. After a median follow-up of 6 years, the progression of patients randomized to treatment was significantly slower than in the untreated control patients, with the overall risk being reduced in half.60 Results were consistent across various patient categories, such as older and younger ages, high and normal tension glaucoma, and eyes with more or less visual field loss at baseline.59 Similar results on the effectiveness of lowering the intraocular pressure to reduce the incidence of glaucoma were reported by the Ocular Hypertension Treatment Study, a large randomized clinical trial.57 In this study, ocular hypertensive patients treated with topical medications experienced conversion to glaucoma at less than half the rate of untreated patients, although the incidence of glaucoma in either group was low. This finding indicates that therapeutic lowering of intraocular pressure can slow disease onset and as such, it has potential for possible primary prevention of glaucoma. There are many caveats to consider, because the prevalence of ocular hypertension in the population is high (Table 4), yet the incidence of glaucoma is low, so that most ocular hypertensives will never develop optic nerve damage. Universal treatment for ocular hypertension is thus not recommended and should be limited to persons at higher risk. Additional information is needed to develop guidelines regarding which individuals with ocular hypertension would benefit most from treatment. The effectiveness of treatment in ocular hypertension relates to primary prevention of glaucoma and not directly to screening, which is a secondary prevention measure aimed to identify undetected glaucoma. BETTER OUTCOME WITH EARLY TREATMENT. A crucial issue to justify screening is that the outcome must be clearly and significantly improved by bringing the patient under clinical care in the asymptomatic preclinical stage. Evidence to indicate that early treatment improves visual field outcomes was provided by the Early Manifest Glaucoma Trial, based on a highly sensitive method to detect early changes.59 In addition to clinical outcomes, it is important to know whether early detection and treatment of glaucoma lead to improved quality of life for the patient. Because glaucoma treatment is lifelong, the early identification of cases through screening will lead to a prolongation of the usual length of therapy, as compared to no screening. This prolongation has consequences for the patient, because current treatments have a number of side effects and potential complications, which may affect quality of life. Further evaluation is needed of the long-term effects of early treatment on clinical and nonclinical outcomes. The additional costs of extended glaucoma treatment must be considered in the calculation of cost-effectiveness of screening. Evaluation of Glaucoma Screening Methods. Are acceptable and valid screening tests available? Can these be applied at a reasonable cost? Should they aim to detect very early disease stages (which requires highly sensitive tests), or to detect moderate or more advanced disease (which has more impact on the patient)? These issues are important in glaucoma, which has no symptoms until vision is impaired. The most frequently used screening methods aim to detect different manifestations of glaucoma: (1) tonometry to detect high intra-ocular pressure; (2) assessment of the disc to detect structural optic nerve changes, usually through ophthalmoscopy; and (3) perimetry to detect functional losses of the visual field. An overview of these screening methods is presented in Table 5.
TABLE 5. Comparison of Methods for Glaucoma Screening INTRAOCULAR PRESSURE Historically, the most commonly used test in glaucoma screening has been tonometry, although its shortcomings to detect the disease are now well recognized.32–37 Ocular tensions are extremely variable, and values are affected by age, time of day, sex, type of tonometer used, central corneal thickness, and many other factors.62,63 Some screening programs have used the Schiotz tonometer because it is less costly and simpler to use, but its measurements are subject to interinstrument and interobserver variations. Furthermore, Schiotz measurements may differ from those obtained with the Goldmann applanation tonometer. Interpretation of the Goldmann readings must consider the potential effects of central corneal thickness, because persons with thick corneas may have artifactually high readings and the reverse may occur in those with thin corneas.63,64 As such, screening with Goldmann tonometry could potentially lead to over- and under-referrals. Most tonometry screenings consist of a one-time measurement of intraocular pressure. The results of these screenings are difficult to evaluate. The large variability in measurements and the increasing skewness of the intraocular pressure distribution with age make it difficult to define the boundary between “normal” and “high” tensions. In addition, the significance of one high measurement is uncertain, since repeat measurements may not be elevated. Usually, levels of 21 or 22-mm Hg have been used as the cutoff point for referral. With this criterion, a large number of persons will be referred after screening, because such values are found in at least 5% of white adults, especially at older ages, and even more frequently in African-Caribbeans (Table 4). Available data suggest that intraocular pressure is lower in Japan,22 where levels decrease with age. Evaluations of tonometry screening must exclude the one-third to one-half of the prevalent cases who have a previous glaucoma diagnosis.19,21,24,26,27 These evaluations should also be based on studies that use several methods to detect structural and functional signs of glaucoma damage. A summary of such tonometry evaluations is presented in Table 6, which includes results for new and existing cases separately, if available.
TABLE 6. Evaluation of Tonometry Screening to Detect Glaucoma in Population Studies Despite methodologic differences, the proportion of open-angle glaucoma cases found to have intraocular pressure over 21-mm Hg at screening generally varies from about one-half to two-thirds of the total cases, being even lower in some studies. As a consequence, the sensitivity of tonometry is consistently low (see Table 6), with somewhat higher sensitivities in African-Caribbean populations. In Barbados, West Indies, the sensitivity of tonometry to detect new cases of open-angle glaucoma was 74%. Sensitivity decreased to 63% when including known cases under treatment, and the same results were found in St. Lucia, West Indies; positive predictive values were also higher in these studies, in which glaucoma prevalence and mean intraocular pressure are also higher than in other populations (Table 4). The use of an intraocular pressure criterion over 21-mm Hg at screening yields comparable specificities across studies, with positive predictive values depending on the population prevalence (Table 6). The above results are based on a one-time tonometry measurement. Retesting of glaucoma patients increases the frequency of elevated pressures. In Baltimore, the frequency of intraocular pressures over 21-mm Hg in all open-angle glaucoma cases rose from about 50% at screening to 75% when screening and/or one follow-up pressure was considered.66 For newly diagnosed open-angle glaucoma, comparable results from Barbados show increases from 74% to 81%.67 Most screenings, however, are limited to a one-time tonometric examination. A related problem is that one-fourth68 to one-half69 of the presumed ocular hypertensives will have normal ocular tensions at retesting. On the other hand, about 7% of those with normal pressures will have elevated measurements on repeat tonometry.68 To reduce the large number of overreferrals and decrease costs of tonometry screening, one option is to increase the referral level, for example, to 25 mm-Hg or higher. This change decreases the number of referrals and increases the positive predictive value but leads to a large decrease in sensitivity. In population studies, for example, only about one-third of cases have pressures of 25 mm-Hg or more at screening.66 Therefore, an unacceptably high percentage of the cases may be missed, if referral levels are raised. These findings have several implications for screening programs. Although measurements of the intraocular pressure on more than one occasion increase sensitivity, the use of one-time tonometry for screening will miss about one-half (or more) of cases. At least 5% to 10% of those screened will be referred for further testing, but most will not have glaucoma. One-half to three-fourths of those referred will be confirmed as ocular hypertensives and asked to return for periodic follow-up visits. Although they are at increased risk of field loss, the magnitude of this risk is low, estimated to be around 1% to 2% per year of follow-up. The magnitude of these estimates will vary depending on the age and racial composition of the population screened. In sum, although tonometry is useful to detect ocular hypertension, it is widely recognized that this method alone is not acceptable for glaucoma screening. STRUCTURAL ASSESSMENT Techniques that allow a direct appraisal of the structural changes indicative of glaucoma have the potential for high sensitivity and specificity. Methods for this examination include ophthalmoscopy, which is most commonly used in screening, as well as photographic gradings of various anatomic features of the optic disc, such as cup-disc ratios. Other methods include nerve fiber layer photography and a variety of new and sophisticated imaging techniques. The requirements and resources to implement these techniques vary widely, and their methodology and application need further development and assessment for screening. Screening by ophthalmoscopy requires skilled and reproducible observers. Rigorous training and standardization are needed for accurate observation of the optic disc and for appropriate referral. The specific disc findings that will result in a positive screening also require strict definition and evaluation of reproducibility within and among observers. Screening by ophthalmoscopy in the Framingham Eye Study showed that 76% of the persons with field defects had a vertical cup-disc ratio of greater than 0.4 and that 71% had high horizontal cup to disc ratios.65 In the Bedford Glaucoma Survey, however, screening based on disc appearance had a sensitivity under 40%.69 A recent evaluation of direct opthalmoscopic assessment of the optic disc reported limited sensitivities and specificities for individual disc features.70 Best results were achieved by combining data on narrowest rim width and parapapillary atrophy, which achieved a sensitivity of 81% and a specificity of 90%.70 In general, evaluations of ophthalmoscopy confirm the influence of observer variation on the results, although satisfactory agreement has been reported in some settings.70 The use of fundus photography for screening has limitations because of the additional resources and trained personnel required to take and interpret the photographs, as well as the pupil dilatation required by some techniques. In addition, media opacities, small pupils, and technical problems lead to incomplete population coverage, even while using nonmydriatic cameras.71 As with ophthalmoscopy, the sensitivity and specificity of photographic evaluations vary according to the criteria used to define abnormal disc structure (e.g., horizontal or vertical cup-disc ratio, narrowest remaining neuroretinal rim, “cupping” or “excavation”); they also vary with the disc criteria used to define the presence of glaucoma changes, even among experts.72 Population studies using photographic gradings have documented the limitations of existing disc criteria. For example, the Baltimore Eye Survey found 63% sensitivity and specificity for cup-disc ratios of 0.35, with 18% sensitivity and 98% specificity for cup-disc ratios of 0.7; similar results were found when using narrowest remaining neuroretinal rim as a criterion.66 Optic disc parameters from fundus photographs are thus not sufficiently sensitive or specific for glaucoma screening. Insufficient evidence exists for the use of nerve fiber layer assessment in screening. To date, it has been used among groups of selected patients, but not on a population-wide basis.71,73–76 Further development and assessment of these methods, which allow the detection of early glaucoma damage, would enhance their use in screening. Considerable interest exists in developing advanced imaging modalities to enhance glaucoma detection and follow-up; these include optical coherence tomography, scanning laser polarimetry, and scanning laser tomography and others.77–79 These modalities offer advantages in terms of quantification, reproducibility, and possibly high sensitivity, as compared to other qualitative techniques, such as photographic gradings. These advantages have to be balanced against the high cost of the instruments and related feasibility issues, which limit their widespread use in screening. Various evaluations of these techniques are ongoing. FUNCTIONAL ASSESSMENT Visual field testing provides a direct functional evaluation of optic nerve status and is important in determining the diagnosis and extent of glaucoma damage. Other factors, such as cataract and retinal abnormalities that might impact the visual field, also need to be taken into account. Automated, computerized perimetry is in widespread use and is potentially useful for screening, because it does not require highly trained examiners and provides more standardized testing conditions than manual perimetry. Its use in screening has been infrequent because of limitations such as lengthy testing protocols, costly and poorly portable instruments, and variability of results, which are greatly influenced by patient-related factors, such as understanding and cooperation. The high frequency of unreliable tests, and difficulties experienced by some individuals in completing perimetry are further deterrents in screening.80 Evaluation of several automated perimeters in screening has found sensitivities of over 90%, with specificities being somewhat dependent on the type of instrument and algorithm used to detect visual field defects.81 Several analytic strategies and algorithms have been developed to enhance the testing and interpretation of results provided by automated perimeters.82–86 Although evaluations of their sensitivity and specificity to distinguish glaucoma patients from selected visually normal individuals give encouraging results, their performance in population-based settings has the potential for low specificity, because many conditions or artifacts may trigger a positive visual field test. Standard threshold perimetry is the most commonly used method to test the visual field in practice. Newer techniques include short wavelength automated perimetry, frequency doubling technology perimetry, high pass resolution perimetry and motion-detection perimetry.81 Some of these are not suitable for screening because of cost or logistic issues. The development of shorter testing algorithms, such as SITA (Swedish Interactive Testing Algorithm) and SITA Fast,85,86 obviate the problems of prolonged testing times. As compared to standard threshold perimetry, these new algorithms may reduce testing time by more than 50%, while increasing sensitivity and reproducibility.81 The development of less costly and portable automated instruments such as the frequency doubling technology perimeter, which offers shorter algorithms, has increased advantages for screening.87 Results to date show high sensitivity and specificity in distinguishing glaucoma patients from controls, particularly when detecting moderate to advanced disease.88 Because the results are not affected by pupil size or blur, this instrument has potential for screening but needs additional testing on a population-wide basis. In sum, significant advances in automated perimetry have occurred. To date, most of these advantages are applicable for clinical purposes, but much remains to be done to increase the feasibility of perimetry screening.89 OTHER TECHNIQUES Because of the limitations of all of the previous techniques, other screening methods are being investigated. Efforts to increase the sensitivity and specificity of screening also include the use of multiple testing methods, combinations of referral criteria, and questionnaires to identify possible predictive factors, such as medical or family history.66,90,91 Further evaluation is needed to determine fully the role of all of these approaches in screening. Conclusions Past assessments of the scientific justification for glaucoma screening have concluded that supportive evidence was lacking, because some of the criteria to justify screening were not met or incompletely met.36–39,89 As such, national task forces on prevention have not recommended glaucoma screening at the time of this writing.38,39 One major weakness has been the lack of evidence on the effectiveness of intraocular pressure-lowering treatment in previously undetected glaucoma. Although this evidence is now available, other issues remain to be resolved. A major challenge is the unavailability of acceptable and cost-effective screening methods that can be applied on a community-wide basis. None of the current screening tests for glaucoma is ideal. Mass tonometry screening of unselected populations, which has been used in the past, cannot be justified because of the large number of false-positive referrals, the low prevalence and incidence of disease in confirmed ocular hypertensives, and the substantial number of persons with glaucoma who will be missed. Instead, screening methods should be based on structural and/or functional assessment to detect glaucoma damage. Aside from logistic considerations, these techniques need further evaluation before being applied on a population-wide basis. Most are likely to have high sensitivity and specificity to detect moderate or advanced glaucoma, with increased efforts needed to detect early disease. Whenever feasible, a comprehensive eye examination has been a recommended approach to detection of glaucoma,39,92 particularly because it can identify other eye conditions. Another strategy proposed is the training of family practitioners in ophthalmoscopy.93 The effectiveness of ophthalmoscopy and automated perimetry to detect glaucoma in primary care and other settings needs further evaluation, as well as the use of various types of personnel in screening.94–96 At present, conventional perimeters are not a practical alternative, owing to their cost and other limitations, but new approaches are now available and show promise for use in screening. Another remaining issue is to document the impact of detecting and treating very early disease on quality of life. Such documentation requires instruments that are sensitive to vision problems in glaucoma. Screening to find patients with early disease requires strong evidence of benefit, because it will require considerable efforts and highly sensitive tests. Additionally, this approach will prolong the usual time span of treatment and affect cost-effectiveness. It will be difficult to estimate cost-effectiveness until this issue is resolved and decisions made regarding the screening tests to use. Rather than mass screening, a more effective approach at present is identifying high-risk groups and targeting them for selective screening or case-finding. Because these groups have an increased prevalence of disease, positive predictive values will increase and costs per case will decrease. Until the uncertainties of population screening are resolved, it seems justified to concentrate on selective screening, for example, testing relatives of glaucoma patients97 or persons of African origin. In sum, although many advances have been made in glaucoma screening, a considerable number of issues need to be addressed. As such, it seems premature to provide a recommendation on the advisability of community-based glaucoma screening. AMBLYOPIA SCREENING Definitions The word amblyopia has often been used in a broad sense to denote nonspecific losses of visual acuity.98,99 For the purposes of the present discussion, amblyopia is defined as low vision that occurs in infants or young children who have experienced some disturbance of binocular and/or foveal vision during development of the visual system but who have no detectable ophthalmologic defect.100 Strabismic or squint amblyopia is often observed in children whose eyes are improperly aligned. Amblyopia ex anopsia designates the condition in children who have had one eye obstructed at an early age.100,101 In these types of amblyopia, visual acuity remains impaired after the original defect has been corrected, and there is no discernible reason for poor vision. Except in cases in which media opacities or severe bilateral refractive errors have degraded the visual image falling on the retinas of both eyes, amblyopia is usually unilateral. The remainder of this discussion focuses on the unilateral condition. Visual impairment ranges from slight abnormality to functional blindness; light perception is always maintained.102 Criteria PUBLIC HEALTH IMPACT. Although amblyopic eyes retain some sight, visual acuity may fall below the level of legal blindness, which is 20/200 (6/60) in the United States and 20/400 by World Health Organization criteria. As long as good vision is maintained in the other eye, unilateral low vision is generally considered to be a mild handicap. Because only one eye is usually affected, amblyopes typically are not included in statistics regarding blindness.103–105 Amblyopia may place the individual at greater risk of mishap, both to the good eye and in general, because as much as one-fourth of the normal bilateral visual field may be lost in unilateral blindness.106 Feman et al.107 studied the causes of blindness of 539 eyes of new applicants for aid to the blind in Tennessee between July 1980 and June 1981. Seventy-four (14%) of the eyes were blind because of amblyopia. These findings are similar to those from the Framingham Eye Study; in 18 (11.8%) of the 152 blind eyes, blindness could be attributed only to amblyopia.17 In contrast, based on registration of partially sighted individuals in the county of Avon, England, Grey et al.108 found that amblyopia accounted for only 10 (0.9%) of 1146 blind eyes and 7 (1.3%) of 546 eyes with partial sight. From nationwide survey samples of 14,577 settled persons and 2233 bedouins in Saudi Arabia, Tabbara and Ross-Degnan109 calculated the prevalence of amblyopia to be 0.82%. Amblyopia was the third most common reason, after refractive errors and cataract, for reduced visual acuity (<20/60 [6/18]) in this population. Tommila and Tarkkanen110 collected histories of amblyopes in Finland who lost vision in the good eye from trauma or other causes and concluded that amblyopes are at greatly increased risk of bilateral blindness. Although this conclusion was based on a small number of patients and imperfect methodology, it implies that the public health problem may be greater than current data suggest. A more recent population-based study in the United Kingdom supported the findings of the Finnish study.111 Jakobsson et al.112 investigated the prevalence of amblyopia among patients at Visual Rehabilitation Centres in southeast Sweden. They estimated that 195 patients (1.7%) of 11,365 patients (average age 69 years) with visual handicap had amblyopia in one or both eyes. Of these 195 individuals, 40% had macular degeneration in the nonamblyopic eye. Stewart-Brown et al.113 confirmed earlier reports that children with amblyopia perform poorly at reading in comparison to their peers without any visual defect or with myopia only. No quality of life data have been published for amblyopes to date; assessment of the social impact on children is difficult. PREVALENCE. Prevalence rates of amblyopia in the United States are not well documented. A prevalence of 2% is frequently quoted,100 but this figure may be an underestimate.114 Prevalence data from several studies in different populations are summarized in Tables 7, 8, and 9.115–151 These rates are probably underestimates, because of failures in almost every study either to screen a random sample of the populations studied, to achieve complete or nearly complete follow-up on all suspected cases to confirm diagnosis, to obtain cooperation from all individuals who presented for screening, to identify cases already successfully treated, or to use reliable screening tests. Furthermore, as noted by Flom and Bedell,152 the level of visual acuity selected to define amblyopia or to select screeners for referral has a major impact on estimates of prevalence.
TABLE 7. Prevalence Rates of Amblyopia Among Preschool Children
VA, visual acuity; —, not reported.
TABLE 8. Prevalence Rates of Amblyopia Among School Children
VA, visual acuity; —, not reported.
TABLE 9. Prevalence Rates of Residual Amblyopia Among Military Recruits and Personnel
*Examined in 1968 †Examined in 1976 VA, visual acuity; —, not reported.
The similarity of prevalence rates shown in Tables 7 and 8 suggests that amblyopia develops in the preschool-aged child. This conclusion is supported by a follow-up study of children who were screened at 4 years of age and screened again when they were 7 years old. Although 44 (1.8%) of 2391 4-year-olds had functional amblyopia, only one new case of amblyopia (0.07%) was found among the 1530 7-year-olds who completed both screenings.121,133 Furthermore, this conclusion agrees with findings from the Avon Longitudinal Study of Pregnancy and Childhood (ALSPAC) in which Williams et al.147 compared methods of following and testing children during the first 3 years of life who were selected randomly from a population birth cohort. Most of the data summarized in Tables 7, 8, and 9 were derived from predominantly or totally white populations (European descent), although more recent studies have been conducted in other ethnic groups. Even in the United States, a country in which there are sizable populations of both blacks (African descent) and whites, there are few data that permit valid comparisons between the two races. In the study by Preslan and Novak,148 75% of the school children participants were black, and this may account for the high prevalence rate. However, more data from black preschool and school children are required before such a conclusion can be reached. No differences in prevalence rates between males and females have been reported. DESCRIPTION AND NATURAL HISTORY. Amblyopia is not a single homogeneous entity. Several distinct clinical syndromes exist, and each may have different etiology and prognosis. Different methods of classification have been proposed on the basis of presence of strabismus, refractive errors, accommodation, fixation, or other factors.98,142,153,154 They have in common incongruity of visual information received by the two eyes, decrease in visual input, or a combination of both factors.155 Strabismic amblyopia associated with late-onset esotropia is perhaps the most common form of amblyopia. It is characterized by onset of an inward deviation of one eye between the ages of 1 and 5 years, best corrected visual acuity worse than 20/50 (6/15) in the deviating eye, at least a two-line difference in acuity between the two eyes, and the absence of any other detectable ophthalmologic defect. The esotropia may appear suddenly after illness, trauma, or stress. The deviation may be constant or intermittent and marked or so slight as to be almost unnoticeable. Various hypotheses have been proposed to explain why good vision is lost or fails to develop in the deviating eye. Most authorities agree with Von Noorden,156 saying that the strabismus develops first, perhaps as a result of some insult to the nervous system. The retinas of the fixating eye and the deviating eye transmit conflicting images to the visual center. To avoid confusion, the child suppresses the image received by the deviating eye. Eventually, this suppression becomes a permanent loss of visual function if not treated. The median age of presentation with strabismic amblyopia is before the fourth birthday.157 Congenital esotropic amblyopia is characterized by an inward deviation of one or both eyes associated with monolateral fixation and is observed during the first 6 months of life. These cases tend to be less amenable to treatment than those associated with late-onset esotropia. Microstrabismic amblyopia is also known as microtropia and monofixation syndrome.152,158,159 Misalignment of the eyes is not obvious, but careful measurement reveals a small-angle esotropia and inability of the affected eye to hold fixation under binocular conditions. This form of amblyopia usually remains undetected until the child undergoes vision testing in school or has an ophthalmologic examination for some other condition. Anisometropic amblyopia occurs even in the absence of strabismus. It is usually detected only during vision screening. There is no deviation, even upon careful ophthalmologic evaluation, but there is a difference in the refractive error of the two eyes (sphere or cylinder) of at least 1.5 diopters and a difference in corrected visual acuity. Age at presentation with anisometropic amblyopia is typically later than with strabismic amblyopia; Shaw et al.157 reported the median age at presentation to be after the sixth birthday. Stimulus deprivation amblyopia (amblyopia ex anopsia) is found in children with a congenital opacity of the media, such as cataract, or ptosis that obstructs vision from birth.100 Decreased vision may persist in the deprived eye even after repair or resolution of the original problem. Stimulus deprivation amblyopia is the best-understood form of amblyopia, but it is a relatively rare clinical problem, accounting for no more than 3% of cases. Prognosis for true amblyopia ex anopsia is usually poor, unless the opacity is removed in timely fashion and aggressive therapy is pursued. Other amblyopias are much less common. These forms include strabismic amblyopia other than the esotropic form, bilateral amblyopia, amblyopia associated with congenital neurologic defects, and amblyopia associated with late occlusion.100,160–162 Strabismic amblyopia and anisometropic amblyopia account for most of the cases.152 Evidence is accumulating that there are fundamental differences between strabismic and anisometropic amblyopia with respect to visual function and pathophysiology and that the better-seeing fellow eyes of amblyopes may also have abnormal visual function. Without treatment, vision in an amblyopic eye may deteriorate to functional blindness. In a few cases, loss of the good eye has been reported to have induced recovery of vision in the amblyopic eye. However, the frequency of such recovery without treatment is unknown and is likely to be very small. It is commonly accepted that the younger the child when therapy is initiated, the greater the likelihood that full vision will be restored. Treatment after the age of 6 to 9 years, depending on the developmental status of the child's visual system, is anticipated by most ophthalmologists to be, at best, only partially successful. Ongoing randomized clinical trials of treatment of older children, noted in the next section, should provide useful data to confirm or refute this position. EFFECTIVENESS AND AVAILABILITY OF CURRENT METHODS OF TREATMENT. Numerous procedures have been used for treating amblyopia.100,163 However, information from well-designed, randomized, multicenter clinical trials was not available until recently, largely due to the efforts of the Pediatric Eye Disease Investigator Group (PEDIG).164 The first completed PEDIG trial enrolled 419 children who were between 3 and 7 years of age and had moderate amblyopia (visual acuity in the amblyopic eye between 20/40 and 20/100). In this initial PEDIG trial, daily atropine drops were compared with 6-hour daily patching.164 After 6 months of follow-up, the investigators concluded that children with moderate amblyopia can be treated equally well with either treatment, evidenced by the fact that 79% of children in the patching group, and 74% in the atropine group had visual acuity 20/30 or better in the amblyopic eye or had gained three or more lines of vision in that eye by 6 months of follow-up. In another PEDIG randomized trial, 189 children between 3 and 7 years of age enrolled.165 The investigators demonstrated that 2-hour per day patching was equally effective in treating moderate amblyopia (visual acuity 20/40 to 20/80 inclusive) as compared with 6-hour per day patching when combined with near vision activities for at least 1 hour a day. Sixty-two percent of patients in each treatment arm in this study had visual acuity 20/32 or better in the amblyopic eye at 4 months of follow-up or had gained three or more lines of vision in that eye. Although long-term comparison of outcomes from these studies were not provided by the PEDIG trials, the short-term data suggest that parents and physicians can choose whether part-time patching or atropine penalization is a more appropriate treatment option for individual children. The benefit of patching is that visual acuity recovery may be faster than with atropine penalization.165 However, atropine penalization results in better treatment compliance than patching. A clinical trial investigating daily use of atropine as compared with Saturday and Sunday use only is currently underway at (http://www.nei.nih.gov/neitrials/static/study94.htm, accessed on January 8, 2004). The mainstay of treatment of severe amblyopia is patching of the healthy eye because the blurring effect of atropine may not be sufficient to switch fixation from the healthy eye to the amblyopic eye. However, there is evidence to support that 6 hours of daily patching is equally effective to full-time patching, in combination with one hour or more of near activities daily, to improve visual acuity in the severely affected amblyopic eyes of children.166 All the treatment regimens currently proposed for amblyopia require both parental and child cooperation. Parental understanding of the importance of early treatment is critical for a successful outcome.167 BETTER OUTCOME WITH EARLY TREATMENT. To achieve a successful long-term outcome, it is generally believed that treatment must be initiated while the visual system is in the plastic stage of development. Another rationale for early treatment is that correction of amblyopia before beginning school ensures that the child does not start formal education with a handicap. If treatment fails, proper provisions for minimizing the psychologic and educational impact of the visual handicap can be made by the parents and teachers. Most investigators report achievement of better visual acuity in a larger proportion of patients when amblyopia therapy is initiated in the first few years of life.168–173 However, the PEDIG has recently reported successful outcomes in children up to 7 years of age and currently are conducting an amblyopia treatment trial in children aged 7 to 17 years of age (http://www.nei.nih.gov/neitrials/static/study93.htm; accessed on January 8, 2004). A Swedish team is investigating atropine treatment in previously untreated children when initiated at 6 to 10 years of age.174 EVALUATION OF AMBLYOPIA SCREENING METHODS. Disappointing results of therapy in older children and adults strengthen the argument for early screening and detection. However, screening young children for amblyopia is a difficult and complex task. Visual function may not be easily tested in a child who does not yet read, is unable to verbalize well, and is uncoordinated. An additional problem is differentiation of immature development of the visual system from vision abnormalities. A major difficulty in the design of preschool vision screening projects is the absence of a readily accessible population, as one has with school children.175 Ehrlich et al.176 estimated that only 21% of preschool children in the United States receive any form of vision screening. Young children are usually unaware of their visual problem because they have no basis for comparison. Unless defects are cosmetically severe, parents may fail to observe any ocular or visual problems. Because of the implications of diagnosis and treatment at a young age, amblyopia screening should be conducted among preschool-age children. However, vision tests in current use in screening programs are not usually practical for evaluating children younger than 3 years old.177–182 Recent experiences with preferential looking tests and acuity cards for screening younger children and infants have been mixed.183–185 Friendly175 and Lippmann186 have written extensively about preschool vision screening. Screening programs designed for preschool-age children usually include visual acuity testing and observation by the examiner for overt, intermittent, or latent strabismus. Tests of stereoscopic or binocular vision may be included.187 Family history has been found to be useful when positive but unreliable when negative.188 Ingram,189 after many years devoted to preschool screening and evaluation of screening programs, adopted a pessimistic view of their rationale and effectiveness. However, Vinding et al.190 concluded that obligatory vision testing of 4-year-olds in Denmark and treatment of visual defects had reduced the prevalence of amblyopia from 3% to 1%. Kentucky is unique in the United States in requiring that all children entering preschool, Head Start, or public school for the first time have a diagnostic eye examination, not just vision screening, by an optometrist or ophthalmologist.191 Because reduced visual acuity is the most clinically obvious sign of amblyopia, visual acuity testing has been the mainstay of most screening programs for preschoolers and school children. Many methods of visual acuity testing in young children have been proposed and reported. More than 30 tests were reviewed by Fern and Manny.192 Tests for young children can be classified into the following types:
None of these tests is totally satisfactory. The letter or number recognition tests rely on the child's ability to recall the name of the symbol. Single letter (or number) presentation usually results in higher visual acuity scores than line or chart presentation and has been shown to be an insensitive method for detecting amblyopia because of the “crowding phenomenon” associated with this disorder.175,198 Directional tests, in which the child points a hand in the direction indicated by the position of the symbol (E or hand), overcome the recognition and naming problems of the letter and number charts. However, young children can discriminate between up and down better than between right and left so that errors in the horizontal position are more common than in the vertical direction.175 Symbol matching tests avoid some of the problems discussed earlier. The child has only to hold up or point to the replica of the symbol that matches the one shown. However, single optotype presentation, which is generally used in these tests, demonstrates the same insensitivity to amblyopia as recognition tests that use this method of presentation. The Electronic Visual Acuity Tester181 developed to facilitate and standardize testing of visual acuity in clinical trials in pediatric eye disease uses matching but incorporates crowding bars that frame the letter displayed and has been demonstrated to have good reliability in children. Picture tests were developed as a means of maintaining the child's interest.195–197,199 The difficulty with these tests is that few objects are universally recognizable as silhouettes or outlines. Furthermore, even when the picture is recognized, the child may not be able to recall its name or may be reluctant to say it if not confident of its pronunciation. Only modest attention has been given to determining the sensitivity and specificity of individual screening procedures, their predictive value in different populations, and the comparability among tests. Table 10 displays calculations of these quantities from some reports that have included sufficient data to compare methods. Aaberg201 compared the E test, monocular cover-uncover, Hirschberg corneal light reflex, and the Worth 4-dot test in a pilot screening project among 248 children. Children who failed screening and a 20% sample of those who passed were evaluated by an ophthalmologist the same day. The overall sensitivity of the screening was 78%, and the specificity was 95%. The Worth 4-dot test had the lowest sensitivity; however, it identified two cases in children who had passed all the other tests. The specificities of all tests were high.
TABLE 10. Estimated Sensitivity, Specificity, and Predictive Values Reported for
Tests Used to Screen for Amblyopia
ACT, Atlantic City Test; AVT, Allen Vision Test; TAC, Teller Acuity Cards; —, not reported.
Friendly175 compared the E test and the letter-matching STYCAR test in 633 children. The letter-matching test was found to be superior from the standpoint of testability, instruction time, and performance time. However, the E test had better reliability and validity. Frydman and Allen202 compared the Atlantic City Test168 with the Allen Vision Test203 and calculated individual and joint sensitivities and specificities. Both tests measure visual acuity, test for excessive hyperopia, and check on abnormal muscle balance. Ellis et al.183 compared Teller Acuity Cards with the clinical judgment of a pediatric ophthalmologist in 108 strabismic children. When an interocular difference of one octave in visual acuity or larger was chosen as the screening cutoff point, sensitivity of 68% and specificity of 82% were obtained. However, children who could not be tested monocularly were eliminated. Thus, agreement of 80% may be an overestimate. Borg and Sundmark204 compared three directional tests in 71 children and found the reliability of all three to be low. Hedin et al.205 compared a letter-matching test to the E chart and concluded that performance was better on the letter-matching test. Jonkers200 compared test results using letter and number recognition charts, a picture chart, Landolt's rings (a directional test), and E chart and calculated conversion formulas for comparing results from the different charts. However, amblyopic children were eliminated from his investigation and the results may not apply to them. Keith et al.206 compared seven tests of visual acuity in a group of 74 children that included those with amblyopia and strabismus. One of the picture tests (Beale Collins) gave good results, but the other was poorly performed. The results of the E test, STYCAR test, and Ffooks test were similar, but the STYCAR matching test was the easiest for the youngest children to grasp and perform. Ohlsson et al.200 evaluated the Lang II stereo card in 1046 children 12 and 13 years of age. They concluded that the sensitivity and positive predictive value were too low for screening purposes. In addition, a number of children with normal eyes and vision misidentified one or more of the three images. Williams et al.147 compared screening methods for strabismus and/or amblyopia. The findings given in Table 10 are based on test results in 876 children tested at 37 months of age. They found that specificity of photorefraction, although low in infancy, was better than 95% by 25 months of age and recommended combining it with cover testing or visual acuity testing to achieve acceptable sensitivity. No clear consensus regarding the best visual acuity test for amblyopia screening can be derived from these comparative studies. Similarly, no recommendation can be made regarding supplemental tests to be added to the visual acuity testing to improve identification of amblyopic children. Kohler and Stigmar121 found that 97% of the eye defects found in children referred from screening had been detected by visual acuity testing; however, Hall et al.207 found that children with suspected strabismus formed the bulk at high risk of amblyopia among referrals for evaluation. Photorefraction to screen infants and young children has become increasingly more widely used. The findings of Williams et al.147 support adding it to vision testing for screening children 2 years of age and older. In spite of the measurement and logistic problems, visual acuity testing has few rivals in screening programs for eye defects among children 3 years of age or older because of the desirability of using a single method to measure visual acuity at different ages and stages of development, and throughout screening, diagnosis, treatment, and follow-up. The Electronic Visual Acuity Tester developed for the PEDIG may become more widely used.180,181 Doran208 has encouraged adoption of Bailey-Lovie charts to standardize test interpretation across a wide age range. Conclusions The need for standard operational definitions required to describe the various clinical presentations of amblyopia, to define amblyopia screening criteria, and to evaluate treatment of amblyopia may be met by the PEDIG or other collaborative groups.163,164 Until standard definitions are used by all amblyopia researchers, comparison of results from screening projects and evaluation of testing methodology will remain difficult. The incidence of amblyopia is high among children with strabismus, and the child with obvious strabismus is likely to come to medical attention earlier than the child with microstrabismus or amblyopia in the absence of strabismus.157,175,209 Better epidemiologic data concerning amblyopia are required to identify other high-risk groups who should receive special attention in screening programs. Objective comparison and evaluation of screening methods for use in preschool-aged children must continue to establish which are best for the detection of different types of amblyopia and for children of different ages. Problems of cooperation and confounding developmental factors are not likely to be eliminated for any single “universal” screening procedure. It has been stated that early vision screening is as important to the child as immunization.114 Combining vision screening with immunization visits to pediatricians and well-child clinics may be one way to ensure that the greatest possible numbers of young children are screened in the least disruptive fashion. The increasing number of preschool children in day-care centers suggests another venue for screening.210,211 |