Top-Rated Free Essay
Preview

Experimental Estimates of Education Production Functions

Powerful Essays
14418 Words
Grammar
Grammar
Plagiarism
Plagiarism
Writing
Writing
Score
Score
Experimental Estimates of Education Production Functions
EXPERIMENTAL ESTIMATES OF EDUCATION PRODUCTION FUNCTIONS* ALAN B. KRUEGER
This paper analyzes data on 11,600 students and their teachers who were randomly assigned to different size classes from kindergarten through third grade. Statistical methods are used to adjust for nonrandom attrition and transitions between classes. The main conclusions are (1) on average, performance on standardized tests increases by four percentile points the first year students attend small classes; (2) the test score advantage of students in small classes expands by about one percentile point per year in subsequent years; (3) teacher aides and measured teacher characteristics have little effect; (4) class size has a larger effect for minority students and those on free lunch; (5) Hawthorne effects were unlikely.

I. INTRODUCTION The large literature on the effect of school resources on student achievement generally finds ambiguous, conflicting, and weak results. Even quantitative summaries of the literature tend to reach conflicting conclusions. For example, based on the fact that most estimates of the effect of school inputs on student achievement are statistically insignificant, Hanushek [1986] concludes, ‘‘There appears to be no strong or systematic relationship between school expenditures and student performance.’’ By contrast, Hedges et al. [1994] conduct a meta-analysis of (a subset of ) the studies enumerated by Hanushek and conclude, ‘‘the data are more consistent with a pattern that includes at least some positive relation between dollars spent on education and output, than with a pattern of no effects or negative effects.’’ Much of the uncertainty in the literature derives from the fact
* I thank Helen Bain, a founder and principal director of Project STAR, for providing me with the data used in this study, Jayne Zaharias, DeWayne Fulton, and Van Cain for answering several questions regarding the data, and Jessica Baraka, Aaron Saiger, and Diane Whitmore for providing outstanding research assistance. The STAR data have been collected and maintained by the Center of Excellence for Research in Basic Skills at Tennessee State University. The STAR data are available from www.telalink.net/ heros. Helpful comments on my research were provided by Charles Achilles, Jessica Baraka, Ronald Ehrenberg, William Evans, Jeremy Finn, John Folger, Victor Fuchs, Joseph Hotz, Lawrence Katz, Cecilia Rouse, James P. Smith, two referees, and seminar participants at the Milken Institute, Massachusetts Institute of Technology, National Bureau of Economic Research, Princeton University, Vanderbilt University, University of California at Los Angeles, the Kennedy School (Harvard University), the London School of Economics, Stockholm University, the Econometric Society, World Bank, and Society of Labor Economists. Financial support was provided by the National Institute of Childhood Health and Development.
1999 by the President and Fellows of Harvard College and the Massachusetts Institute of Technology. The Quarterly Journal of Economics, May 1999

497

498

QUARTERLY JOURNAL OF ECONOMICS

that the appropriate specification—including the functional form, level of aggregation, relevant control variables, and identification—of the ‘‘education production function’’ is uncertain.1 Some specifications do consistently yield significant effects, however. Notably, estimates that use cross-state variation in school resources typically find positive effects of school resources, whereas studies that use within-state data are more likely to find insignificant or wrong-signed estimates (see Hanushek [1996]).2 Many of these specification issues arise because of the possibility of omitted variables, either at the student, class, school, or state level. Moreover, functional form issues are driven in part by concern for omitted variables, as researchers often specify education production functions in terms of test-score changes to difference out omitted characteristics that might be correlated with school resources (although such differencing could introduce greater problems if the omitted characteristics affect the trajectory of student performance). A classical experiment, in which students are randomly assigned to classes with different resources, would help overcome many of these specification issues and provide guidance for observational studies. This paper provides an econometric analysis of the only large-scale randomized experiment on class size ever conducted in the United States, the Tennessee Student/Teacher Achievement Ratio experiment, known as Project STAR. Project STAR was a longitudinal study in which kindergarten students and their teachers were randomly assigned to one of three groups beginning in the 1985–1986 school year: small classes (13–17 students per teacher), regular-size classes (22–25 students), and regular/aide classes (22–25 students) which also included a full-time teacher’s aide. After their initial assignment, the design called for students to remain in the same class type for four years. Some 6000–7000 students were involved in the project each year. Over all four years, the sample included 11,600 students from 80 schools. Each school was required to have at least one of each class-size type, and random assignment took place within schools. The students
1. There is also debate over what should be the appropriate measure of school outputs (see Card and Krueger [1996]). Whereas education researchers tend to analyze standardized test scores, economists tend to focus on students’ educational attainment and subsequent earnings. 2. Hanushek attributes this difference to omitted state-level variables that bias the state-level studies, although it is possible that endogenous resource decisions within states (e.g., assignment of weaker students to smaller classes as required by compensatory education) bias the within-state micro-data estimates, and that the interstate estimates are unbiased.

EXPERIMENTAL ESTIMATES

499

were given a battery of standardized tests at the end of each school year. In a review article Mosteller [1995] described Project STAR as ‘‘a controlled experiment which is one of the most important educational investigations ever carried out and illustrates the kind and magnitude of research needed in the field of education to strengthen schools.’’ The STAR data have been examined extensively by an internal team of researchers. This analysis has found that students in small classes tended to perform better than students in larger classes, while students in classes with a teacher aide typically did not perform differently than students in regular-size classes without an aide (see Word et al. [1990], Finn and Achilles [1990], and Folger and Breda [1989]). Past research primarily consists of comparisons of means between the assignment groups, and analysis of variance at the class level. In this research, little attention has been paid to potential threats to the validity of the experiment or to the longitudinal structure of the data. As in any experiment, there were deviations from the ideal experimental design in the actual implementation of Project STAR. First, students in regular-size classes were randomly assigned again between classes with and without full-time aides at the beginning of first grade, while students in small classes continued on in small classes, often with the same set of classmates.3 Re-randomization was done to placate parents of children in regular classes who complained about their children’s initial assignment. Because analysis of data for kindergartners did not indicate a significant effect of a teacher aide on achievement in regular-size classes, it was felt that this procedure would create few problems. But if the constancy of one’s classmates influences achievement, then the experimental comparison after kindergarten is compromised by the re-randomization. A second limitation of the experiment is that approximately 10 percent of students switched between small and regular classes between grades, primarily because of behavioral problems or parental complaints. These nonrandom transitions could also compromise the experimental results. Furthermore, because some students and their families naturally relocate during the school year, actual class size varied more than intended in small classes (11 to 20) and in regular classes (15 to 30). Finally, as in most
3. If a school had more than one small class, students could be moved between small classes.

500

QUARTERLY JOURNAL OF ECONOMICS

longitudinal studies of schooling, sample attrition was common— half of students who were present in kindergarten were missing in at least one subsequent year. And some students may have nonrandomly switched to another public school or enrolled in private school upon learning their class-type assignments. These limitations of the experiment have not been adequately addressed in previous work. This paper has three related goals. First, to probe the sensitivity of the experimental estimates to flaws in the experimental design. Second, to use the experiment to identify an appropriate specification of the education production function to estimate with nonexperimental data. Third, to use the experimental results to interpret estimates from the literature based on observational data. The conclusion makes a rough attempt to compare the benefits and costs of reducing class size from 22 to 15 students.

II. BACKGROUND

ON

PROJECT STAR

AND

DATA

A. Design and Implementation
Project STAR was funded by the Tennessee legislature, at a total cost of approximately $12 million over four years.4 The Tennessee legislature required that the study include students in inner-city, suburban, urban, and rural schools. The research was designed and carried out by a team of researchers at Tennessee State University, Memphis State University, the University of Tennessee, and Vanderbilt University. To be eligible to participate in the experiment, a public school was required to sign up for four years and be large enough to accommodate at least three classes per grade, so within each school students could be assigned to a small class (13–17), regular class (22–25 students), or regular plus a full-time aide class.5 The statewide pupil-teacher ratio in kindergarten in 1985–1986 was 22.3, so students assigned to regular classes fared about as well as the average student in the state [Word et al. 1990]. Schools with more than 67 students per grade had more than three classes. One limitation of the comparison between regular and regular/aide classes is that in grades 1–3 each regular class had the services of a part-time aide 25–33
4. This section draws heavily from Word et al. [1990] and Folger [1989]. 5. Participating schools had an average per-pupil expenditure in 1986–1987 of $2724, compared with the statewide average of $2561.

EXPERIMENTAL ESTIMATES

501

percent of the time on average, so the variability in aide services was restricted.6 The cohort of students who entered kindergarten in the 1985–1986 school year participated in the experiment through third grade. Any student who entered a participating school in a relevant grade was added to the experiment, and participating students who repeated a grade, skipped a grade, or left the school exited the sample. Entering students were randomly assigned to one of the three types of classes (small, regular, or regular/aide) in the summer before they began kindergarten.7 Students were typically notified of their initial class assignment very close to the beginning of the school year. Students in regular classes and in regular/aide classes were randomly reassigned between these two types of classes at the end of kindergarten, while students initially in small classes continued on in small classes. Notice, however, that results from the kindergarten year are uncontaminated by this feature of the experiment. Because kindergarten attendance was not mandatory in Tennessee at the time of the study, many new students entered the program in first grade. Additionally, students were added to the sample over time because they repeated a grade or because their families moved to a school zone that included a participating school. In all, some 2200 new students entered the project in first grade and were randomly assigned to the three types of classes. Another 1600 and 1200 new students entered the experiment in the second and third grades, respectively. Newly entering students were randomly assigned to class types, although the uneven availability of slots in small and regular classes often led to an unbalanced allocation of new students across class types. A total of 11,600 children were involved in the experiment over all four years. After third grade, the experiment ended, and all students were assigned to regular-size classes. Although data have been collected on students through ninth grade, the present study only has access to data covering grades K–3. Data were
6. The reason that regular classes often had a teacher aide is that the ethic underlying the study was that students in the control group (i.e., regular classes) would not be prevented from receiving resources that they ordinarily would receive. 7. The procedure for randomly assigning students was as follows. Each school prepared an alphabetized enrollment list. Algorithms were centrally prepared which assigned every kth student to a class type; the algorithm was tailored to the number of enrolled students. A random starting point was used by each school to apply the algorithm. The schools were audited to ensure that they followed procedures for random assignment.

502

QUARTERLY JOURNAL OF ECONOMICS

collected on students each fall and spring during the experiment. Class type is based on the class attended in the fall. All students who attended a STAR class in either the fall or spring are included in the database. Unfortunately, the STAR data set does not contain students’ original class type assignments resulting from the randomization procedure; only the class types that students actually were enrolled in each year are available. It is possible that some students were switched from their randomly assigned class to another class before school started or early in the fall. To determine the frequency of such switches, we obtained and (double) entered data on the initial random assignments from the actual enrollment sheets that were compiled in the summer prior to the start of kindergarten for 1581 students from 18 participating STAR schools.8 It turns out that only 0.3 percent of students in the experiment were not enrolled in the class type to which they were randomly assigned in kindergarten. Moreover, only one student in this sample who was assigned a regular or regular/aide class enrolled in a small class. Consequently, in the analysis below, we will refer to the class type in which students are enrolled during the first year they enter the experiment as their initial random assignment. A limitation of the experiment is that baseline test score information on the students is not available, so one cannot examine whether the treatment and control groups ‘‘looked similar’’ on this measure before the experiment began. Nonetheless, if the students were successfully randomly assigned between class types, one would expect those assigned to small- and regular-size classes to look similar along other measurable dimensions at base line. Tables I and II provide some evidence on the differences among students assigned to the three class types. Table I disaggregates the data into waves, based upon the grade the students entered the program, because this was the first time the students were randomly assigned to a class type. The sample consists of all students who were enrolled in a STAR class when the fall or spring data were collected. Sample means by class type for several variables are presented. As one would expect, students assigned to small classes had fewer students in their class than those in regular classes, on average. There are small
8. I thank Jayne Zaharias for providing the enrollment sheets. The sample I analyze excludes twins; schools were allowed to assign twins to the same class if that was the school’s ordinary practice.

EXPERIMENTAL ESTIMATES

503

COMPARISON

OF

TABLE I MEAN CHARACTERISTICS OF TREATMENTS UNADJUSTED DATA

AND

CONTROLS:

A. Students who entered STAR in kindergartenb Variable 1. 2. 3. 4. 5. 6. Free lunchc White/Asian Age in 1985 Attrition rated Class size in kindergarten Percentile score in kindergarten Small .47 .68 5.44 .49 15.1 54.7 Regular .48 .67 5.43 .52 22.4 49.9 Regular/Aide .50 .66 5.42 .53 22.8 50.0 Joint P-Valuea .09 .26 .32 .02 .00 .00

B. Students who entered STAR in first grade 1. 2. 3. 4. 5. 6. Free lunch White/Asian Age in 1985 Attrition rate Class size in first grade Percentile score in first grade .59 .62 5.78 .53 15.9 49.2 .62 .56 5.86 .51 22.7 42.6 .61 .64 5.88 .47 23.5 47.7 .52 .00 .03 .07 .00 .00

C. Students who entered STAR in second grade 1. 2. 3. 4. 5. 6. Free lunch White/Asian Age in 1985 Attrition rate Class size in third grade Percentile score in second grade .66 .53 5.94 .37 15.5 46.4 .63 .54 6.00 .34 23.7 45.3 .66 .44 6.03 .35 23.6 41.7 .60 .00 .66 .58 .01 .01

D. Students who entered STAR in third grade 1. 2. 3. 4. 5. 6. Free lunch White/Asian Age in 1985 Attrition rate Class size in third grade Percentile score in third grade .60 .66 5.95 NA 16.0 47.6 .64 .57 5.92 NA 24.1 44.2 .69 .55 5.99 NA 24.4 41.3 .04 .00 .39 NA .01 .01

a. p-value is for F-test of equality of all three groups. b. Sample size in panel A ranges from 6299 to 6324, in panel B ranges from 2240 to 2314, in panel C ranges from 1585 to 1679, and in panel D ranges from 1202 to 1283. c. Free lunch pertains to the fraction receiving a free lunch in the first year they are observed in the sample (i.e., in kindergarten for panel A; in first grade in panel B; etc.) Percentile score pertains to the average percentile score on the three Stanford Achievement Tests the students took in the first year they are observed in the sample. d. Attrition rate is the fraction that ever exits the sample prior to completing third grade, even if they return to the sample in a subsequent year. Attrition rate is unavailable in third grade.

504
P-VALUES
FOR

QUARTERLY JOURNAL OF ECONOMICS
TABLE II WITHIN-SCHOOL DIFFERENCES AND REGULAR/AIDE CLASSES

TESTS

OF

AMONG

SMALL, REGULAR,

Variable K 1. 2. 3. 4. 5. 6. Free lunch White/Asian Age Attrition rate Actual class size Percentile score .46 .66 .38 .01 .00 .00

Grade entered STAR program 1 .29 .28 .12 .07 .00 .00 2 .58 .15 .48 .58 .00 .46 3 .18 .21 .40 NA .00 .00

Each p-value is for an F-test of the null hypothesis that assignment to a small, regular, or regular/aide class has no effect on the outcome variable in that grade, conditional on school of attendance. All rows except 4 pertain to the first grade in which the student entered the STAR program. The attrition rate in row 4 measures whether the student ever left the sample after initially being observed.

differences in the fraction of students on free lunch, the racial mix, and the average age of students in classes of different size, although some of these differences are statistically significant (see rows 1–4).9 Because random assignment was only valid within schools, these differences suggest the importance of controlling for school effects as is done in Table II. Table II presents p-values for joint F-tests of the differences among small, regular, and regular/aide classes for the variables presented in Table I. Unlike results reported in Table I, these p-values are conditional on school effects. None of the three background variables displays a statistically significant association with class-type assignment at the 10 percent level, which suggests that random assignment produced relatively similar groups in each class size, on average. As an overall test of random assignment, I regressed a dummy variable indicating assignment to a small class on the three background measures in rows 1–3 and school dummies. For each wave, the student characteristics had no more than a chance association with class-type assignment. Furthermore, if the same regression model is estimated for a sample that pools all four entering waves of students together, the three student characteristics are still insignificantly related to assignment to a small class ( p-value .58). Within schools, there
9. To be precise, the fraction on free lunch actually measures the fraction who receive free or reduced-price lunch.

EXPERIMENTAL ESTIMATES

505

is no apparent evidence that initial assignment to class types was correlated with student characteristics. To check whether teacher assignment was independent of observed teacher characteristics, I regressed each of three teacher characteristics (experience, race, or education) on dummies indicating the class type the teachers were assigned to and school dummies, and then performed an F-test of the hypothesis that the class-type dummies jointly had no effect. These regressions were calculated for each of the four grade levels, so there was a total of twelve regressions. In each case, the p-value for the class-type dummies exceeded .05.10 These results are as one would expect with random assignment of teachers to the different class types. There was a high rate of attrition from the project. Only half the students who entered the project in kindergarten were present for all grades K–3. For the kindergarten cohort, students in small classes were three–four percentage points more likely to stay in the sample than those in regular-size classes. This pattern was reversed among those who entered in first grade, however. Attrition could occur for several reasons, including students moving to another school, students repeating a grade, and students being advanced a grade. Although I lack data on retention rates for the early grades, Word et al. [1990] report that over the four years of the project, 19.8 percent of students in small classes were retained, while 27.4 percent of students in regular classes were retained. This is consistent with the lower attrition rate of students in small classes. Some of the analysis that follows makes a crude attempt to adjust for possible nonrandom attrition. It is virtually impossible to prescribe the exact number of students in a class: families move in and out of a school district during the course of a year; students become sick; and varying numbers of students are enrolled in schools. As a result, in some cases actual class size deviated from the intended ranges. Table III reports the frequency distribution of class size for first graders, by assignment to small, regular, or regular/aide classes. Although students assigned to small classes clearly were more likely to attend classes with fewer students, there was considerable variability in class size within each class-type assignment, and some overlap between the distributions.
10. In two cases the p-value was less than .10. Third grade teachers assigned to small classes were less likely to have a master’s degree or higher than were teachers assigned regular-size classes, and first grade teachers in small classes had two more years of experience than those in regular-size classes (although less experience than those in regular/aide classes).

506
DISTRIBUTION
OF

QUARTERLY JOURNAL OF ECONOMICS
TABLE III ACTUAL CLASS SIZES GROUP IN FIRST GRADE

CHILDREN

ACROSS

BY

RANDOM ASSIGNMENT

Actual class size in first grade 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 Average class size

Assignment group in first grade Small 24 182 252 465 256 561 108 57 20 0 0 0 0 0 0 0 0 0 0 15.7 Regular 0 0 0 0 16 17 36 76 200 378 594 437 384 175 130 54 28 29 30 22.7 Aide 0 0 0 0 0 0 0 57 120 378 329 460 264 225 234 108 56 58 30 23.4

Actual class was determined by counting the number of students in the data set with the same class identification.

It is also virtually impossible to prevent some students from switching between class types over time. Table IV shows a transition matrix between class types for students who continued from K–1, 1–2, and 2–3 grades. If students remained in their same class type over time, all the off-diagonal elements would be zero. The re-randomization of students in regular classes in first grade is apparent in panel A. But in second and third grades, when students were supposed to remain in their same type of class, 9–11 percent of students switched class-size types. Students were moved between class types because of behavioral problems or, in some cases, parental complaints. Obviously, if the movement between class types was associated with student characteristics (e.g., students with stronger academic backgrounds more likely to move into small classes), these transitions would bias a simple comparison of outcomes across class types.

EXPERIMENTAL ESTIMATES
TABLE IV TRANSITIONS BETWEEN CLASS-SIZE IN ADJACENT GRADES NUMBER OF STUDENTS IN EACH TYPE OF CLASS A. Kindergarten to first grade First grade Kindergarten Small Regular Aide All Small 1292 126 122 1540 Regular 60 737 761 1558 Reg/aide 48 663 706 1417

507

All 1400 1526 1589 4515

B. First grade to second grade Second grade First grade Small Regular Aide All Small 1435 152 40 1627 Regular 23 1498 115 1636 Reg/aide 24 202 1560 1786 All 1482 1852 1715 5049

C. Second grade to third grade Third grade Second grade Small Regular Aide All Small 1564 167 40 1771 Regular 37 1485 76 1598 Reg/aide 35 152 1857 2044 All 1636 1804 1973 5413

To address this potential problem, and the variability of class size for a given type of assignment, in some of the analysis that follows initial random assignment is used as an instrumental variable for actual class size.

B. Data and Standardized Tests
Students were tested at the end of March or beginning of April of each year. The tests consisted of the Stanford Achievement Test (SAT), which measured achievement in reading, word recognition, and math in grades K–3, and the Tennessee Basic Skills First (BSF) test, which measured achievement in reading and math in grades 1–3. The tests were tailored to each grade level. Because there are no natural units for the test results, I scaled the test scores into percentile ranks. Specifically, in each grade level the regular and regular/aide students were pooled

508

QUARTERLY JOURNAL OF ECONOMICS

together, and students were assigned percentile scores based on their raw test scores, ranging from 0 (lowest score) to 100 (highest score). A separate percentile distribution was generated for each subject test (e.g., Math-SAT, Reading-SAT, Word-SAT, etc.). For each test I then determined where in the distribution of the regular-class students every student in the small classes would fall, and the students in the small classes were assigned these percentile scores. Finally, to summarize overall achievement, the average of the three SAT percentile rankings was calculated.11 If the performance of students in the small classes was distributed in the same way as performance of students in the regular classes, the average percentile score for students in the small classes would be 50. An examination of the correlations among the tests indicates that the strongest correlations typically are between tests of the same subject matter; for example, in second grade the SAT and BSF reading tests have a correlation of .80. Tests of the same subjects tend to have a higher correlation from one grade to the next than tests of different subjects. The SAT and BSF tests are also highly correlated with each other: the correlation between the average SAT percentile and average BSF percentile is .79 in first grade and .85 in second grade. For most of the subsequent analysis, the SAT exam is the primary focus of study because this test has been used on a national level for a long period of time. The main findings are similar for the BSF test, however. The average of the three SAT exams by class type is presented in the last row of Table I. Figure I displays the kernel density of the average test score distributions for students in small and regular classes at each grade level.12 In all grades, the average student in small classes performed better on this summary test measure than did those in regular or regular/aide classes. There does not seem to be a very strong or consistent effect of the teacher aide, however. The rest of the paper probes the robustness of these findings.
11. Formally, denote the cumulative distribution of scores on test j (denoted j T j ) of students in the regular and regular/aide classes as F R (T j ) prob [T Ri j j T j] y j. For each student i in a small class, we then calculated F R (T Si ) ySi. Naturally, the distribution of y j for students in regular classes follows a uniform distribution. We then calculated the average of the three (or two for BSF) percentile rankings for each student. If one subtest score was missing, we took the average of the two percentiles that were available; and if two were missing, we used the percentile score corresponding to the only available test. 12. Note that because we have averaged over three percentile scores, the distributions are not uniform for students assigned to regular classes.

EXPERIMENTAL ESTIMATES

509

FIGURE I Distribution of Test Percentile Scores by Class Size and Grade

Observe also that the average test score of students in all class types tends to be lower for those who entered the experiment in higher grades. This pattern is likely to reflect the fact that kindergarten was optional and higher-achieving students were more likely to attend kindergarten, as well as the tendency of lower-achieving students to be retained and disproportionately added to the sample at higher grade levels. Because of this feature of the data, I control for the grade in which the student entered Project STAR in some of the analysis below. The Appendix presents means for several variables that are available in the data set. III. STATISTICAL MODELS To see the advantage of a randomized experiment in estimating the effect of school resources on student achievement, consider the following general model: (1)

Yij

aSij

bFij

ij,

510

QUARTERLY JOURNAL OF ECONOMICS

where Yij is the achievement level of student i in school j, Sij is a vector of school characteristics, Fij is a vector representing the family background of the student, and ij is a stochastic error component. In principle, Sij and Fij include information cumulated over the student’s life; for example, classroom size and teacher qualifications for each year the student attended school. The entire history of family background variables and school resources may contribute to students’ achievement in a given year. In addition, children’s unobserved inherent ability may contribute to their achievement. In any actual application we will generally lack data on some relevant school, family, or student characteristics. These omitted variables will then appear in the error term. If the omitted variables are correlated with the included variables, then the estimated parameters will be biased. If a school characteristic such as class size is determined by random assignment, however, it will be independent of the omitted variables. Thus, with random assignment, a simple comparison of mean achievement between children in small and large classes provides an unbiased estimate of the effect of class size on achievement. We begin analyzing the STAR data by estimating the following regression equation for students in each grade level: (2) Yics
0 1SMALLcs 2REG/Acs 3Xics

s

ics,

where Yics is the average percentile score on the SAT test of student i in class c at school s, SMALLcs is a dummy variable indicating whether the student was assigned to a small class that year, REG/Acs is a dummy variable indicating whether the student was assigned to a regular-size class with an aide that year, and Xics is a vector of observed student and teacher covariates (e.g., gender). The independence between class-size assignment and other variables is only valid within schools, because randomization was done within schools. Consequently, a separate dummy variable is included for each school to absorb the school effects, s. The equation is estimated by ordinary least squares (OLS). In calculating the standard errors, however, the error term ics is modeled in a components-of-variance framework. Specifically, ics is assumed to consist of two components: ics µcs ics, where µcs is a class-specific random component that is common to all members of the same class, and ics is an idiosyncratic error term. The class-specific component µcs may exist because of unobserved

EXPERIMENTAL ESTIMATES

511

teacher characteristics, or because some students may exert a common influence over others in the class. Because several students were reassigned to different classes after their initial random assignment, in part based on their performance, equation (1) was also estimated using dummies indicating students’ initial assignment the first year they entered the program, rather than their actual assignment each year. Models including initial assignment are labeled ‘‘reduced-form’’ models, because one can think of initial assignment as an excluded variable that is correlated with actual class size. The initial assignment and actual assignment variables are identical in kindergarten, so the OLS and reduced-form estimates are identical for kindergarten students. Regression results are presented in Table V.13 Columns 1–4 use actual assignment, and columns 5–8 use initial class assignment. Columns 1 and 5 omit the school dummies. As earlier analyses of the data have found, students in small classes tend to perform better than those in regular and regular/aide classes. Here, the gap in average performance is about 5 percentile points in kindergarten, 8.6 points in first grade, and 5–6 points in second and third grade. Columns 2 and 6 add unrestricted school dummies to the model. In three of four grades, including the school dummies leads to a slight increase in the effect of being assigned to a small class. If class size were truly randomly assigned, including additional exogenous variables would not significantly alter the coefficient on the class-size dummies. In fact, including covariates seems to have a very modest effect on the class-size coefficients conditional on school effects. The student characteristics in columns 3 and 5 add considerable explanatory power. White and Asian students tend to score eight percentile points higher than black students in kindergarten, and this gap is about six points in third grade.14 Students on free lunch score thirteen percentile points less than those not on free lunch, and girls score three–four points higher than boys in each grade level. The teacher characteristics have notably weak explanatory power. Teacher education—as proxied by a dummy indicating
13. The robust standard errors are about two-thirds larger than the OLS standard errors. The estimated standard deviation of the class effects (µcs ) is about 8 in the models in column 4. 14. Ninety-nine percent of the students are white or black. The small number of Asian students are included with white students in the analysis. The small number of hispanic students and others are included with the black students.

512

QUARTERLY JOURNAL OF ECONOMICS

whether the teacher has a master’s degree—does not have a systematic effect. Hardly any of the teachers are male, so the gender results are not very meaningful. Teacher experience has a small, positive effect. Experimentation with a quadratic in experience indicated that the experience profile tends to peak at about

OLS

AND

TABLE V REDUCED-FORM ESTIMATES OF EFFECT OF CLASS-SIZE ASSIGNMENT AVERAGE PERCENTILE OF STANFORD ACHIEVEMENT TEST OLS: actual class size

ON

Reduced form: initial class size (5) (6) (7) (8)

Explanatory variable

(1)

(2)

(3)

(4)

A. Kindergarten Small class 4.82 5.37 (2.19) (1.26) Regular/aide class .12 .29 (2.23) (1.13) White/Asian (1 — — yes Girl (1 yes) — — Free lunch (1 yes) White teacher Teacher experience Master’s degree School fixed effects R2 — — — — No .01 — — — — Yes .25 5.36 (1.21) .53 (1.09) 8.35 (1.35) 4.48 (.63) 13.15 (.77) — — — Yes .31 5.37 4.82 5.37 (1.19) (2.19) (1.25) .31 .12 .29 (1.07) (2.23) (1.13) 8.44 — — (1.36) 4.39 — — (.63) 13.07 — — (.77) .57 — — (2.10) .26 — — (.10) .51 — — (1.06) Yes No Yes .31 .01 .25 5.36 (1.21) .53 (1.09) 8.35 (1.35) 4.48 (.63) 13.15 (.77) — — — Yes .31 5.37 (1.19) .31 (1.07) 8.44 (1.36) 4.39 (.63) 13.07 (.77) .57 (2.10) .26 (.10) .51 (1.06) Yes .31

B. First grade Small class 8.57 (1.97) Regular/aide class 3.44 (2.05) White/Asian (1 — yes) Girl (1 yes) — Free lunch (1 yes) White teacher Male teacher Teacher experience Master’s degree School fixed effects R2 — — — — — No .02 8.43 (1.21) 2.22 (1.00) — — — — — — — Yes .24 7.91 (1.17) 2.23 (0.98) 6.97 (1.18) 3.80 (.56) 13.49 (.87) — — — — Yes .30 7.40 (1.18) 1.78 (0.98) 6.97 (1.19) 3.85 (.56) 13.61 (.87) 4.28 (1.96) 11.82 (3.33) .05 (0.06) .48 (1.07) Yes .30 7.54 (1.76) 1.92 (1.12) — — — — — — — No .01 7.17 (1.14) 1.69 (0.80) — — — — — — — Yes .23 6.79 (1.10) 1.64 (0.76) 6.86 (1.18) 3.76 (.56) 13.65 (.88) — — — — Yes .29 6.37 (1.11) 1.48 (0.76) 6.85 (1.18) 3.82 (.56) 13.77 (.87) 4.40 (1.97) 13.06 (3.38) .06 (.06) .63 (1.09) Yes .30

EXPERIMENTAL ESTIMATES

513

TABLE V (CONTINUED) OLS: actual class size Explanatory variable (1) (2) (3) (4) Reduced form: initial class size (5) (6) (7) (8)

C. Second grade Small class 5.93 (1.97) Regular/aide class 1.97 (2.05) White/Asian (1 — yes) Girl (1 yes) — Free lunch (1 yes) White teacher Male teacher Teacher experience Master’s degree School fixed effects R2 — — — — — No .01 6.33 (1.29) 1.88 (1.10) — — — — — — — Yes .22 5.83 (1.23) 1.64 (1.07) 6.35 (1.20) 3.48 (.60) 13.61 (.72) — — — — Yes .28 5.79 5.31 5.52 (1.23) (1.70) (1.16) 1.58 .47 1.44 (1.06) (1.23) (0.87) 6.36 — — (1.19) 3.45 — — (.60) 13.61 — — (.72) .39 — — (1.75) 1.32 — — (3.96) .10 — — (.06) 1.06 — — (1.06) Yes No Yes .28 .01 .21 5.27 (1.10) 1.16 (0.81) 6.27 (1.21) 3.48 (.60) 13.75 (.73) — — — — Yes .28 5.26 (1.10) 1.18 (0.81) 6.29 (1.20) 3.44 (.60) 13.77 (.73) .43 (1.76) .82 (4.23) .10 (.07) 1.16 (1.05) Yes .28

D. Third grade 5.32 5.58 (1.91) (1.22) Regular/aide class .22 .16 (1.95) (1.12) White/Asian (1 — — yes) Girl (1 yes) — — Free lunch (1 yes) White teacher Male teacher Teacher experience Master’s degree School fixed effects R2 — — — — — No .01 — — — — — Yes .17 Small class 5.01 (1.19) .33 (1.11) 6.12 (1.45) 4.16 (.66) 13.02 (.81) — — — — Yes .22 5.00 5.51 5.42 (1.19) (1.46) (1.08) .75 .30 .12 (1.07) (1.17) (0.85) 6.11 — — (1.44) 4.16 — — (.65) 12.96 — — (.81) .64 — — (1.75) 7.42 — — (2.80) .04 — — (.06) 1.10 — — (1.15) Yes No Yes .23 .01 .16 5.30 (1.03) .13 (0.81) 5.97 (1.44) 4.17 (.66) 13.21 (.82) — — — — Yes .22 5.24 (1.04) .10 (0.78) 5.96 (1.43) 4.18 (.66) 13.16 (.81) .19 (1.75) 6.83 (2.76) .03 (.06) .88 (1.15) Yes .22

All models include constants. Robust standard errors that allow for correlated residuals among students in the same class are in parentheses. Sample size is 5861 for kindergarten, 6452 for first grade, 5950 for second grade, and 6109 for third grade.

514

QUARTERLY JOURNAL OF ECONOMICS

twenty years of experience, and students in classes where the teacher has twenty years of experience tend to score about three percentile points higher than those in classes where the teacher has zero experience, all else being equal. As a whole, however, consistent with much of the previous literature, the STAR data suggest that measured teacher characteristics explain relatively little of student achievement on standardized tests. Estimates of the effect of being in a small class which use initial assignment (columns 5–8) are only slightly smaller than the estimates which use the actual class assignment (columns 1–4), and are always statistically significant. This finding suggests that possible nonrandom movement of students between small and regular classes was not a major limitation of the experiment. To summarize these results, based on column 4 it appears that students in small classes score about five–seven percentage points higher than those assigned to regular-size classes. Students assigned to a regular/aide class perform slightly better (one or two percentile points, on average) than students assigned to a regular class without a full-time aide, but the gap is only statistically significant in one grade level. Thus, it is possible that a teacher aide has only a trivial effect on student achievement, or that the availability of part-time aides in regular classes confounds the true effect of an aide. Is the impact of attending a small class big or small? Unfortunately, it is unclear how percentile scores on these tests map into tangible outcomes. Nevertheless, a couple of comparisons are informative. First, relative to the standard deviation (S.D.) of the average percentile score, the effect sizes are .20 in kindergarten, .28 in first grade, .22 in second grade, and .19 in third grade (based on the model in column 4). Second, one could compare the estimated class-size effects with the effects of other student characteristics. For example, in kindergarten the impact of being assigned to a small class is about 64 percent as large as the white-black test score gap, and in third grade it is 82 percent as large. By both metrics, the magnitudes are sizable.

A. Effects of Attrition
Table VI provides some simple evidence on the impact of sample attrition. As is common in longitudinal studies of education, attrition was very high from Project STAR classes. If the students originally assigned to regular classes who left the sample

EXPERIMENTAL ESTIMATES
TABLE VI EFFECT OF ATTRITION DEPENDENT VARIABLE: AVERAGE PERCENTILE SCORE ON SAT

515

EXPLORATION

OF

Actual test data Coefficient on small class dum. 5.32 (.76) 6.95 (.74) 5.59 (.76) 5.58 (.79) Sample size 5900 6632 6282 6339

Actual and imputed test data Coefficient on small class dum. 5.32 (.76) 6.30 (.68) 5.64 (.65) 5.49 (.63) Sample size 5900 8328 9773 10919

Grade K 1 2 3

Estimates of reduced-form models are presented. Each regression includes the following explanatory variables: a dummy variable indicating initial assignment to a small class; a dummy variable indicating initial assignment to a regular/aide class, unrestricted school effects; a dummy variable for student gender; and a dummy variable for student race. The reported coefficient on small class dummy is relative to regular classes. Standard errors are in parentheses.

had higher test scores, on average, than students assigned to small classes who also left the sample, then the small class effects will be biased upwards. One reason why this pattern of attrition might occur is that high-income parents of children in larger classes might have been more likely to subsequently enroll their children in private schools over time than similar parents of children in small classes. At heart, adjusting for possible nonrandom attrition is a matter of imputing test scores for students who exited the sample. With longitudinal data, this can be done crudely by assigning the student’s most recent test percentile to that student in years when the student was absent from the sample.15 The sample used in the first panel of Table VI includes the largest number of students with nonmissing data available each grade. These results correspond to the model estimated in column 7 of Table V, except the free lunch variable is omitted because it
15. In the case of a student who left the sample but later returned, the average test score in the years surrounding the student’s absence was used. Test scores were also imputed for students who had a missing test score but did not exit the sample (e.g., because they were absent when the test was conducted). This technique is closely related to the ‘‘last-observation-carry-forward’’ method that has been used in clinical studies.

516

QUARTERLY JOURNAL OF ECONOMICS

changes over time.16 For simplicity, only the coefficient on the dummy variable indicating initial assignment to a small class is reported in the table. The sample used in the second panel is larger than the sample in column 1 because it includes the column 1 sample plus any student who entered the program in an earlier grade and exited the sample by the current grade, assigning imputed test percentiles to students who exited the sample. To be included in the sample, it is necessary to have test data in at least some year. (Because kindergarten students could not have previously exited the sample, the sample size is the same in the first row.) The estimates using imputed test percentiles for missing observations are qualitatively quite similar to the estimates using the subsample of observations who were present in each particular grade.17 Thus, nonrandom attrition does not appear to bias the estimated class size effects in Table V. The sample used in column 2 of Table VI excludes students who were listed on the enrollment logs for kindergarten but withdrew from school prior to the start of school. For example, if a parent chose to withdraw a child from a STAR public school and enroll him or her in a private school immediately upon learning that the child was assigned to a regular-size class, the student is excluded from the sample. This type of behavior appears to have been rare (based on our inspection of notes on a sampling of enrollment sheets), but 12 percent of students who were listed on the enrollment logs and assigned to a class prior to the start of school were not actually enrolled in the school the following fall. These students moved to another school zone, enrolled in private school, or were withdrawn from kindergarten over the summer for some other reason. Using data for eighteen of the participating schools for which we were able to obtain initial kindergarten enrollment sheets, we calculate that 10.4 percent of students who were listed on the enrollment sheets and assigned to small kindergarten classes were missing from our sample by the start of kindergarten; the corresponding figures for regular/aide and regular classes are 12.2 percent and 14.3 percent, respectively. The differential withdrawal rate between the regular and small classes is statistically significant (t 1.86), while the difference
16. The estimated model uses initial class assignment to avoid imputing actual class size for missing observations. The sample in column 1 is a little larger than that in column 7 of Table V because Table V uses a balanced sample, and some observations were excluded due to missing data on free lunch status and teacher characteristics. 17. The coefficient on the regular/aide initial assignment dummy is also quite similar if the model is estimated with or without the imputed data.

EXPERIMENTAL ESTIMATES

517

between the regular/aide and small class is not (t 0.86). These findings suggest that 2 to 4 percent of students may have been withdrawn from the STAR schools because they were not assigned to a small class. An upper bound estimate of the impact on test scores of the higher kindergarten withdrawal rate for students in regular and regular/aide classes can be calculated. Suppose that the 2–4 percent extra students who withdrew from regular and regular/ aide classes all would have scored in the one-hundredth percentile of the SAT exams. With this intentionally extreme assumption, the average score of students in the regular and regular/aide classes would only have increased by one–two percentile points if the extra students had not withdrawn from kindergarten. At the opposite extreme, if the higher withdrawal rate is due to the lowest achieving students leaving regular-size classes, the regularsize class students would have scored one–two points lower, on average, if they had remained. The actual impact is probably even smaller, however, because the extra withdrawals probably would have scored closer to the average student if they remained in the STAR schools. But even the upper and lower bounds estimates suggest that the higher withdrawal rate from regular-size classes does not have much impact on the results.

B. Two-Stage Least Squares (2SLS) Models
As noted, students in the Project STAR experiment who were assigned to small classes had varying numbers of students in their classes because of student mobility and enrollment differences across schools. Similarly, students in the regular-size classes had variable class sizes. A more appropriate model of achievement would take actual class size into account. A natural model for this situation is a triangular model of student achievement in which the actual number of students in the class is included on the right-hand side, and initial assignment to a class type is used as an instrumental variable for actual class size. Specifically, we estimate the following model by 2SLS: (3) (4)

CSics

0 0

1Sios 1CSics

2Rios 2Xics

3Xics

s

ics

Yics

s

ics,

where CSics is the actual number of students in the class, Sios is a dummy variable indicating assignment to a small class the first year the student is observed in the experiment, Rios is a dummy variable indicating assignment to a regular class the first year the

518
OLS Grade K 1 2 3
AND

QUARTERLY JOURNAL OF ECONOMICS
TABLE VII 2SLS ESTIMATES OF EFFECT OF CLASS SIZE ON ACHIEVEMENT DEPENDENT VARIABLE: AVERAGE PERCENTILE SCORE ON SAT OLS (1) .62 (.14) .85 (.13) .59 (.12) .61 (.13) 2SLS (2) .71 (.14) .88 (.16) .67 (.14) .81 (.15) Sample size (3) 5,861 6,452 5,950 6,109

The coefficient on the actual number of students in each class is reported. All models also control for school effects; student’s race, gender, and free lunch status; teacher race, experience, and education. Robust standard errors that allow for correlated errors among students in the same class are reported in parentheses.

student is observed in the experiment, and all other variables are defined as before. Again, the error term ( ics ) is treated as consisting of a common class effect and an idiosyncratic individual effect, and the standard errors are adjusted for correlation in the residuals among students in the same class.18 In this setup, only variation in class size due to initial assignment to a regular or small class is used to provide variation in actual class size in the test score equation. Due to the random assignment of initial class type, one would expect that this excluded instrumental variable is uncorrelated with ics, as required for 2SLS to be consistent.19 If attending a small class has a beneficial effect on students’ test scores, 1 would be negative. Table VII presents OLS and 2SLS estimates of equation (4). The 2SLS estimates tend to be a little larger in absolute value, especially in third grade. According to the 2SLS estimates, a reduction of ten students is associated with a seven-to-nine point increase in the average percentile ranking of students, depending on the grade. There is no obvious trend over grade levels in the effect of class size in these data.
18. Because the teacher aide was found to have a small effect in Table V, we do not hold constant the availability of an aide in equation (4). One could, however, add a dummy indicating the presence of a full-time aide to equation (4). 19. To interpret this model as yielding the causal effect of current class size on achievement, it is necessary to assume that initial class assignment only affects current test scores by affecting current class size. If previous class sizes affect current performance, initial assignment will be correlated with the error term in equation (4). Of course, in the kindergarten year this assumption is not controversial, but it may not hold in later grades.

EXPERIMENTAL ESTIMATES

519

Table VIII presents additional 2SLS estimates of the effect of actual class size on achievement, disaggregating the sample by the grade the student entered Project STAR and current grade. The model and identification strategy are the same as in Table VII, column 2. The results indicate that for each cohort of students, those attending smaller classes tend to score higher on the standardized test by the end of the first year they entered the experiment. If assignment to small or regular classes was somehow nonrandom, then the initial assignment would have to have been skewed in the direction of producing higher test scores in the small classes for each wave of students who entered the program—an unlikely event. Interestingly, for the wave of students who entered in kindergarten, the beneficial effect of attending a small class does not appear to increase as students spend more time in their class assignment. For students entering the experiment in first or second grade, however, the test score gap between those in small- and regular-size classes grows as students progress to higher grades. The effect of time spent in a small class is explored further by pooling students in all grades together below.

C. Models with Pooled Data
To explore the cumulative effects of having been in a small or regular class, several models were estimated with the data pooled
TABLE VIII 2SLS ESTIMATES OF EFFECT OF CLASS SIZE ON ACHIEVEMENT, BY ENTRY GRADE AND CURRENT GRADE DEPENDENT VARIABLE: AVERAGE SCORE ON STANFORD ACHIEVEMENT TEST Current grade K 1 2 3 K .71 (.15) .89 (.17) .49 (.16) .66 (.17) 1 Entering grade 2 3

.49 (.23) .70 (.29) 1.21 (.34)

.24 (.21) .71 (.28)

.66 (.21)

The coefficient on the actual number of students in each class is reported. All models also control for school effects; student’s race, gender, and free lunch status; teacher race, experience, and education. Robust standard errors that allow for correlation of residuals among students in the same class are reported in parentheses. Sample size in column 1 begins at 5901 and ends at 3124; sample size in column 2 begins at 2190 and ends at 1110; sample size in column 3 begins at 1492 and ends at 1010; sample size in column 4 is 1110.

520

QUARTERLY JOURNAL OF ECONOMICS

over students and grades. The general model was of the form, (5) Yig
0 1Sio 2REG/Aio

S 3N ig

A 4N ig g 5Xig

f

s

ig,

where g indicates grade level (K, 1, 2, or 3) and i indicates students, Yig is the test score, Sio and REG/Aio are dummy variables indicating a student’s class type in the first year he or S A she participated in the program, N ig and N ig are the cumulative number of years (including the current grade) the student has spent in a small or regular/aide class, Xig is a vector of student, teacher, and class characteristics, g is a set of three current grade dummies, f is a set of three dummies indicating the first year the student entered the STAR sample, and s is a set of school fixed effects. Estimation is done by OLS, and robust standard errors that allow for a random individual component in the error term are reported. Results including various sets of explanatory variables are reported in Table IX. Estimates shown in column 1 exclude student, teacher, and classmate characteristics. In column 2, regressors for measured student and teacher characteristics are included. Both of these models indicate that achievement of students in small classes jumps up by about four percentile points the first year a student attends a small class ( 1 3 ), and increases by about one percentile point for each additional year the student spends in a small class thereafter. Both the initial effect of being in a small class and the cumulative effect are statistically significant in these models. Column 3 adds four variables reflecting the composition of a student’s classmates. Students in small classes were more likely to remain with their classmates in first grade because students in regular classes were randomly reassigned between regular classes with and without full-time aides. Two variables are included to control for the impact of the constancy of one’s classmates. First, the fraction of each student’s classmates who were in that student’s class the preceding year is included. If a student is new to the school in a particular grade, this variable will have a value of 0; and if a student attends a class that consists only of students who were in that student’s class the preceding year, the variable will have a value of 1. As a second measure of the environment in the class, we take the average of this variable over all the other students in the class. This variable might influence achievement because the extent to which other students in a class know each other could influence one’s adjustment to the class.

EXPERIMENTAL ESTIMATES

521

TABLE IX ESTIMATES OF POOLED MODELS DEPENDENT VARIABLE: AVERAGE PERCENTILE RANKING ON SAT TEST COEFFICIENT ESTIMATES WITH ROBUST STANDARD ERRORS IN PARENTHESES Variable Initial class small (1 yes) yes) (1) 2.87 (.83) .29 (.69) 1.19 (.39) .37 (.39) — — — — No Yes .18 25,249 (2) 3.16 (.80) .49 (.67) 1.05 (.38) .25 (.37) — — — — Yes Yes .23 24,350 (3) 2.99 (.80) .58 (.67) .65 (.39) .14 (.37) .60 (1.03) .46 (1.52) 2.73 (1.62) 6.85 (1.67) Yes Yes .23 24,349

Initial class regular/aide (1

Cumulative years in small class Cumulative years in reg/aide class Fraction of classmates in class previous year Average fraction of classmates together previous year Fraction of classmates on free lunch Fraction of classmates who attended kindergarten Student and teacher characteristics 3 current grade dummies; 3 dummies indicating first grade appeared in sample; school effects R2 Sample size

Student and teacher characteristics are as follows: student race, gender, and free lunch status; and teacher race, gender, experience, and master’s degree or higher status. OLS estimates are reported, with robust standard errors that adjust for a possible correlation of residuals for the same student over time in parentheses.

In addition to these two ‘‘class constancy’’ variables, the regression includes the fraction of students in a class who receive free lunch and the fraction of students in the class who were present in the experiment during kindergarten. Because students on free lunch score lower on standardized tests than other students, a higher proportion of classmates on free lunch in a class may lower overall performance. The fraction of a class that attended kindergarten could affect achievement because kindergarten attendance is likely to make the class more socialized for school, which should enable the teacher to convey more material. Due to the random assignment of students, these variables should be uncorrelated with any omitted variables within schools. Including these four variables hardly changes the initial jump in test scores associated with attending a small class (see column 3), although the cumulative effect of time spent in a small

522

QUARTERLY JOURNAL OF ECONOMICS

class is reduced by one-third when they are included, and is only on the margin of statistical significance (t 1.66). Also notice that attendance in classes with a higher proportion of classmates who attended kindergarten has a large, positive effect on a student’s own achievement. A two-standard-deviation change in the fraction of one’s classmates who attended kindergarten is associated with about a three-percentile-point change in test scores. Test scores are not significantly related to the variables measuring the constancy of one’s classmates. However, these variables are set to zero in kindergarten as all kindergarten students are new to the class. If the model in column 3 is estimated using the subsample from first grade on, students who are new to classes that include many students who were together the previous grade tend to score significantly lower on the SAT exam (t 3.2). Thus, if a student is new to a class, he or she does better if most of the other students are new to the class as well. A higher fraction of classmates on free lunch has a negative, marginally statistically significant effect on achievement in this sample. The pooled models in Table IX allow for a one-time, discrete improvement in test scores from attending a small class, and for a constant increase for each additional year the student spends in a small class. One could estimate a more general model. Most obviously, the initial effect of being in a small class could vary by grade level (i.e., interact grade dummies and SMALL ), and the linear effect of cumulative years in a small class could be generalized by including a set of unrestricted dummies indicating the number of past years spent in a small class. In results not presented here, such a less restrictive model was estimated. The estimates in Table IX are nested in this model, so they can be tested against it. An F-test rejects the parsimonious specification in Table IX at the .01 level. However, inspection of the coefficients suggests that the main reason for the rejection is that the initial effect of being in a small class is smaller in grades 2 and 3 than in kindergarten and first grade; the linear trend appears to be a plausible representation of the cumulative effect of time spent in a small class. Despite this rejection, the parsimonious model is a convenient way to summarize the dynamic effects of attending a small class in the early grades. The relationship between the pooled model and the ‘‘valueadded’’ specification, which Hanushek and Taylor [1990] suggest is superior to other specifications of the education production function, should be emphasized. The value-added model only identifies the cumulative effect of time spent in a small class; the

EXPERIMENTAL ESTIMATES

523

initial effect is differenced out. This can be seen by taking the first-difference of equation (5). If the estimates in Table IX had indicated that there was no effect of the initial year spent in a small class, the value-added specification would capture the only parameter of interest. But the pooled estimates and the results in Table VIII indicate that perhaps the most important benefit of attending a small class occurs the first year a student is placed in a small class. This benefit is missed in the value-added specification. This point is illustrated by estimating the following valueadded specification by OLS: (6)

Yics, g

Yics, g

1

0

1SMALLics, g

2Xics, g

g

s

ics, g,

where the dependent variable is the change in students’ percentile test scores between the end of grade g and g-1, and SMALLics, g is class size during grade g. The coefficient 1 essentially corresponds to the coefficient on cumulative time spent in a small class in equation (5). When this specification is estimated, the estimate of 1 is 1.2, with a t-ratio of 3.1.20 This value-added effect is of similar magnitude to the coefficient on the cumulative years in a small class variable in the models in Table IX. Thus, although the estimated value-added specification indicates that students gain from attending small classes, the benefit is less than the full effect that accounts for the discrete gain that occurs the first year students are in a small class. Prais [1996] and Hanushek [1998] interpret the STAR experiment as providing evidence that smaller classes did not improve performance because previously published cross-sectional results do not show the achievement test gap between students in small and regular classes expanding significantly over time. For example, Hanushek [1998] writes: ‘‘If smaller classes were valuable in each grade, the achievement gap would widen. It does not. In fact, the gap remains essentially unchanged through the sixth grade. . . . The inescapable conclusion is that the smaller classes at best matter in kindergarten.’’ This conclusion strikes me as questionable for two reasons. First, the mix of students compared at various grade levels in the results cited by Hanushek changes over time; half of the students exit or enter the sample after kindergarten. When the same students are tracked over time, the value-added and pooled specifications show students in small
20. The other covariates in this regression are the same as in column 3 of Table IX.

?

524

QUARTERLY JOURNAL OF ECONOMICS

classes gaining about one percentile rank per year relative to students in regular classes. Second, students appear to benefit particularly from attending a small class the first year they attend one, whether that is in kindergarten, first, second, or third grade (see Table VIII). The discrete jump in scores occurring the first year students attend a small class, combined with the entry of new waves of students over time, can distort the simple crosssectional comparison of gains for the changing mix of students.

D. Heterogeneous Treatment Effects
The effect of being in a small class may vary for students with different backgrounds. Table X presents OLS estimates of the pooled model (equation (5)) for several subsamples of students. The pooled model was selected to summarize the class-size effects over all grade levels, although a less restrictive model would fit the data better. Smaller classes tend to have a larger initial effect, but a smaller cumulative effect, for boys as compared with girls. Students on free lunch and black students tend to have both a larger initial effect and larger cumulative effect than those not on free lunch and white students. Finally, inner-city students tend to have a more beneficial effect of attending a small class in the first year they attend one than students from other areas, and a sharper gain over time from remaining in a small class.21 Word et al. [1990] similarly found that smaller classes had a more beneficial effect for black students, students on free lunch, and inner-city students, but did not examine whether these differences were due to the initial effect or cumulative effect of time spent in a small class. In general, the pattern of effects reported in Table X suggests that the lower achieving students benefit the most from attending smaller classes. Summers and Wolfe [1977] also find that attending a small class is more beneficial for low achieving students than for high achieving students. The effect of attending a small class can also be estimated for each of the 80 schools in Project STAR. To estimate school-level small-class effects, I pooled the data for students across grades, and for each school regressed the percentile score on dummies indicating attendance in small and regular/aide classes, current
21. Inner-city schools were defined as schools in metropolitan areas in which more than half of students received free lunch; suburban was defined as the balance of metropolitan area schools; towns were defined as areas with more than 2500 inhabitants; and rural was defined as areas with fewer than 2500 inhabitants.

EXPERIMENTAL ESTIMATES

525

TABLE X SEPARATE ESTIMATES FOR SELECT SAMPLES DEPENDENT VARIABLE: AVERAGE PERCENTILE RANKING ON SAT TEST COEFFICIENT ESTIMATES WITH ROBUST STANDARD ERRORS IN PARENTHESES Boys Small Cumulative years in small class Sample size 4.18 (1.11) .60 (.56) 12,576 Girls 1.28 (1.13) .92 (.54) 11,773 Not on free lunch 2.85 (1.12) .55 (.51) 12,285

Free lunch Small Cumulative years in small class Sample size 3.14 (1.10) .94 (.59) 12,064

Black Small Cumulative years in small class Sample size Inner city Small Cumulative years in small class Sample size 3.74 (1.68) 1.71 (.90) 5,154 3.84 (1.29) 1.04 (.68) 8,150

White 2.58 (1.02) .66 (.48) 16,069

Metropolitan 2.92 (1.55) .57 (.83) 5,906

Towns 3.09 (2.83) 1.35 (1.50) 1,872

Rural 2.58 (1.23) 1.03 (.56) 11,417

Model and covariates are the same as column 3 of Table IX.

grade dummies, and dummies indicating the grade the student entered project STAR. A parsimonious model was estimated for simplicity and to preserve degrees of freedom. A kernel density for the coefficients on the small-class dummy is shown in Figure II. Two-thirds of the school-specific small-class effects are positive, while one-third are negative. Furthermore, 2.5 percent of the 80 coefficients had t-ratios less than 2, while 30 percent had t-ratios

526

QUARTERLY JOURNAL OF ECONOMICS

FIGURE II Kernel Density of School-Level Small-Class Effects

exceeding 2. The mean coefficient estimate is 4.6. The standard deviation of the coefficients (after adjusting for sampling variability) 7.5 percentage points.22 Thus, some schools are more adept at translating smaller classes into student achievement than are other schools.

E. Hawthorne and John Henry Effects
It has been suggested by some that the effectiveness of small classes found in the STAR experiment may have resulted from ‘‘Hawthorne effects,’’ in which teachers in small classes responded to the fact that they were part of an experiment, rather than a true causal effect of small classes themselves.23 Others have suggested that the effect sizes might actually be larger than measured by the STAR experiment because teachers in regular classes provided greater than normal effort to demonstrate that
22. To adjust for sampling variability in the coefficient estimates, the average squared standard error was subtracted from the variance of the estimated coefficients. 23. For an interesting study that finds little evidence of Hawthorne effects in the original Hawthorne experiments, see Jones [1992]. One could argue in the current context that each individual teacher in small classes has an incentive to free ride rather than work extra hard.

EXPERIMENTAL ESTIMATES

527

they could overcome the bad luck of being assigned more students: a ‘‘John Henry’’ effect. Either set of responses could limit the external validity of the results of the STAR experiment. As a partial check on these potential ‘‘reactive’’ effects, I examined the relationship between class size and student achievement just among students assigned to regular-size classes. Recall that there is considerable variability in class size even in the regular-size classes (see Table III).24 Obviously, Hawthorne and John Henry effects do not apply to a sample in which all teachers were randomly assigned to the control group. On the other hand, variability in class size is likely to be due primarily to idiosyncratic factors in this sample, such as integer effects in assigning classes and student mobility during the school year. Moreover, there is limited variability in class sizes within schools because many schools had only one regular-size class per grade. To estimate the effect of class size on achievement for the control sample, I pooled the sample of students in regular-size classes across all grade levels, and regressed the average SAT test score on the number of students in the class, grade level dummies, and student and teacher characteristics.25 The coefficient on class size in this regression is .55, with a t-ratio of 4.3. If school dummies are added to this model, the coefficient on class size falls to .39, but remains statistically significant (t 3.1). Based on these estimates, an eight-student reduction in class size is associated with a three-to-four-percentile increase in test scores, which is insignificantly different from estimates derived from the experimental variations in class size. These regression results do not provide much evidence of either Hawthorne or John Henry effects. And given that much of the variability in class size in the control group may be due to measurement errors (e.g., students moving in and out of class during the school year), it is noteworthy that these regressions find any evidence of class-size effects.

F. Separate Subject Test Results for SAT and BSF
Table XI presents estimates of the pooled data model corresponding to column 3 of Table IX for each of the main subsections of the SAT test, as well as for the subsections of the BSF test and the average of the math and reading percentile scores on the BSF test. These results indicate relatively minor differences in the initial and cumulative effects of attending a small class on the
24. The standard deviation of class size in the sample of students assigned to regular classes is 2.3, as compared with 4.1 among all students in the experiment. 25. These regression results are reported in Table 12 of Krueger [1997].

528

QUARTERLY JOURNAL OF ECONOMICS
TABLE XI ESTIMATES OF POOLED DATA MODEL BY SUBJECT TEST DEPENDENT VARIABLE: PERCENTILE SCORE ON SAT OR BSF TEST COEFFICIENT ESTIMATES WITH ROBUST STANDARD ERRORS IN PARENTHESES Stanford Achievement Test Math Reading 3.52 (.88) .43 (.43) 23,461 Word Basic Skills First Math Reading Avg.

Small

2.83 (.88) Cumulative years in small .45 class (.42) Sample size 23,794

2.97 1.09 3.04 2.02 (.87) (1.05) (1.08) (.96) .80 1.23 .41 .83 (.42) (.44) (.46) (.41) 23,630 18,174 18,010 18,250

Model and covariates are the same as in column 3 of Table IX.

math, reading and word recognition tests. Furthermore, the BSF test shows the same basic pattern as the SAT test—a discrete increase in performance for attending a small class, with a small (statistically insignificant) increase thereafter. On the whole, little seems to have been lost by focusing on the average of the SAT tests as the mainstay of the analysis. IV. CONCLUSION One well-designed experiment should trump a phalanx of poorly controlled, imprecise observational studies based on uncertain statistical specifications. The implementation of the STAR experiment was not flawless, but my reanalysis suggests that the flaws in the experiment did not jeopardize its main results. Adjustments for school effects, attrition, re-randomization after kindergarten, nonrandom transitions, and variability in actual class size do not overturn the main findings of Word et al. [1990], Folger and Breda [1989], and Finn and Achilles [1990]: students in small classes scored higher on standardized tests than students in regular-size classes. The results also indicate that the provision of a full-time teacher aide has only a modest effect on student achievement, although this effect may be attenuated because of the frequent availability of part-time aides in regular classes. Interestingly, at least for the early grades, my analysis suggests that the main benefit of attending a small class seems to arise by the end of the initial year a student attends a small class. After the first year, additional time spent in a small class has a positive but smaller effect on test scores. One possible explanation for this pattern is that attending a small class in the lower grades

EXPERIMENTAL ESTIMATES

529

may confer a one-time, ‘‘school socialization effect’’ which permanently raises the level of student achievement without greatly affecting the trajectory. Because much of the previous literature estimates class-size effects using a ‘‘value-added’’ specification that uses student test score gains as the dependent variable and current class size as the main explanatory variable, much of the past research may miss the main benefit of smaller classes. More research is needed to develop an appropriate model of student learning. But for now, one should be concerned that the value-added specification may miss much of the value that is added from attending a smaller class. Moreover, studies that identify class-size effects by comparing differences in the level of test scores between students who were subject to different class sizes for exogenous reasons, such as Angrist and Lavy’s [1999] clever use of Maimonides’ law, may stand a better chance of uncovering the total effect of class size than estimates based on the value-added specification. No single study, even an experimental one, could be definitive. The STAR results suggest that the magnitude of the achievement gains from attending smaller classes varies across schools and student characteristics. It is possible (though probably unlikely) that Tennessee has a much higher concentration of students or schools that benefit from smaller classes than other states. It is also possible that reducing class size does not have a beneficial effect for students after the third grade. Obviously, more experimentation would help resolve these issues. It would also be helpful to compare the STAR findings with the rest of the literature. Before concluding that the weight of the literature suggests that attending a small class does not matter for the average student, it would be useful to know how many of the studies enumerated in Hanushek’s [1986, 1996] surveys have sufficient power to reject either the level effect (for level specifications) or cumulative effect (for value-added specifications) of attending a small class that is implied by the Project STAR data. Experiments of the scale and quality of Project STAR are disappointingly rare in the education field. When these experiments are conducted, they should be analyzed and followed up to the fullest extent possible. The students who participated in Project STAR were returned to regular classes after third grade, and have been followed up through the ninth grade. Nye et al. [1994] find that students who were placed in small classes have lasting achievement gains through at least the seventh grade, although it is difficult to compare the magnitude of the benefits

530

QUARTERLY JOURNAL OF ECONOMICS

with those at earlier grades because of changes in the tests that were administered. The students studied in Project STAR are currently in high school. To learn more about the long-term benefits of attending smaller classes, it would be useful to continue studying the academic—and just as importantly, nonacademic—outcomes of the STAR participants as they enter early adulthood. In the meantime, we can perform the following rough benefitcost analysis to gauge the likely order of magnitudes of the economic effects of reducing class size in the early grades. The STAR experiment reduced class size by seven or eight students, or about by one-third. Folger and Parker [1990] estimate that the cost of reducing class size in Tennessee (including capital costs) would be proportional to the total annual educational expenditures per student. In 1995–1996 total expenditures per enrolled public school student in the United States were $6459 [National Center for Educational Statistics 1996], so reducing class size by one-third would increase costs per student by about $2151 per year. We discount all benefits and costs to the present. Using a 3 percent real discount rate, the present value of the additional costs of reducing class size by one-third for the wave of entering kindergarten students for four years would be approximately $7400. The economic benefits of the STAR experiment are much more difficult to assess than the costs. The Table V results suggest that test scores for students in small classes rose by about 0.22 standard deviations. I am not aware of any study that links achievement on the Stanford Achievement Test to later economic outcomes. Furthermore, it is possible that the cognitive gains from attending smaller classes will dissipate or grow by the time the STAR students enter the labor market. As a rough assumption, suppose that the 0.22 S.D. gain persists. How does this translate into economic benefits? Estimates based on the High School and Beyond sample in Murnane, Willet, and Levy [1995] indicate that male high school seniors who score 0.22 S.D.’s higher on the basic math achievement test in 1980 earned 1.7 percent higher earnings six years later. The comparable figure for females was 2.4 percent. Average earnings for workers age 18 and older in the United States in 1996 were $34,705 for men and $20,570 for women [U. S. Census Bureau 1996]. If we assume that real earnings will be unchanged in the future and that Murnane, Willet, and Levy’s estimates can be applied to the STAR experiment, then the present value of the earnings gain from raising

EXPERIMENTAL ESTIMATES

531

test scores .22 S.D.’s is $9603 for men and $7851 for women, assuming that students enter the workforce at age 20 and retire at age 65, and using a real discount rate of 3 percent. Many assumptions underlying this cost-benefit calculation could turn out to be wrong, including the following: real earnings may grow or shrink; the effect of test scores on future earnings may be different than assumed; class size may influence other economic outcomes, such as crime and dependency; the cost of reducing class size may be different than assumed. There is no substitute for directly measuring the economic outcomes that may be affected by reducing class size. Nonetheless, these calculations suggest that the benefit of reducing class size in terms of future earnings is in the same ballpark as the costs.
APPENDIX: SUMMARY STATISTICS STANDARD DEVIATIONS IN PARENTHESES Grade Variable Class size Percentile score avg. SAT Percentile score avg. BSF Free lunch White Girl Agea Exited sampleb Retained Percent of teachersc with MA degree Percent of teachers who are White Percent of teachers who are male No. of schools No. of students No. of small classes No. of reg. classes No. of reg/aide classes K 20.3 (4.0) 51.4 (26.7) NA .48 .67 .49 5.43 (0.35) .29 NA .35 .83 .00 79 6,323 127 99 99 1 21.0 (4.0) 51.5 (26.9) 51.8 (26.1) .52 .67 .48 6.58 (0.49) .26 NA .35 .82 .00 76 6,828 124 115 100 2 21.1 (4.1) 51.2 (26.5) 51.6 (26.2) .51 .65 .48 7.67 (0.56) .21 NA .37 .79 .01 75 6,839 133 100 107 3 21.3 (4.4) 51.0 (27.0) 51.4 (26.1) .50 .66 .48 8.70 (0.59) NA .04 .44 .79 .03 75 6,801 140 89 107 All 20.9 (4.1) 51.3 (26.8) 51.6 (26.1) .51 .66 .47 7.12 (1.31) .43 NA .38 .81 .01 80 11,599 524 403 413

MEANS

WITH

a. Age as of September of the school year they are observed. b. The fraction that exited the sample in the next year, for K–2; for All it is the fraction that ever exited the sample. c. Teacher characteristics are weighted by the number of students in each teacher’s class.

532

QUARTERLY JOURNAL OF ECONOMICS

PRINCETON UNIVERSITY AND THE NATIONAL BUREAU OF ECONOMIC RESEARCH

REFERENCES
Angrist, Joshua, and Victor Lavy, ‘‘Using Maimonides’ Rule to Estimate the Effect of Class Size on Children’s Academic Achievement,’’ Quarterly Journal of Economics, CXIV (1999), 533–575. Card, David, and Alan B. Krueger, ‘‘Labor Market Effects of School Quality: Theory and Evidence,’’ in Gary Burtless, ed., Does Money Matter? The Effect of School Resources on Student Achievement and Adult Success (Washington, DC: Brookings Institution, 1996). Finn, Jeremy D., and Charles M. Achilles, ‘‘Answers and Questions about Class Size: A Statewide Experiment,’’ American Educational Research Journal, XXVIII (1990), 557–577. Folger, John, ‘‘Editor’s Introduction: Project STAR and Class Size Policy,’’ Peabody Journal of Education, LXVII (1989), 1–16. Folger, John, and Carolyn Breda, ‘‘Evidence from Project STAR about Class Size and Student Achievement,’’ Peabody Journal of Education, LXVII (1989), 17–33. Folger, John, and Jim Parker, ‘‘The Cost-Effectiveness of Adding Aides or Reducing Class Size,’’ Vanderbilt University, mimeo, 1990. Hanushek, Eric A., ‘‘The Economics of Schooling: Production and Efficiency in Public Schools,’’ Journal of Economic Literature, XXIV (1986), 1141–1177. , ‘‘A More Complete Picture of School Resource Policies,’’ Review of Educational Research, LXVI (1996), 397–409. , ‘‘The Evidence on Class Size,’’ Occasional Paper Number 98-1, W. Allen Wallis Institute of Political Economy, University of Rochester, Rochester, NY, February 1998. Hanushek, Eric A., and Lori Taylor, ‘‘Alternative Assessments of the Performance of Schools,’’ Journal of Human Resources, XXV (1990), 179–201. Hedges, Larry V., Richard Laine, and Rob Greenwald, ‘‘Does Money Matter? A Meta-Analysis of Studies of the Effects of Differential School Inputs on Student Outcomes,’’ Education Researcher, XXIII (1994), 5–14. Jones, Stephen, ‘‘Was There a Hawthorne Effect?’’ American Journal of Sociology, III (1992), 451–468. Krueger, Alan, ‘‘Experimental Estimates of Education Production Functions,’’ Princeton University, Industrial Relations Section Working Paper No. 379, Princeton, NJ, 1997. Mosteller, Frederick, ‘‘The Tennessee Study of Class Size in the Early School Grades,’’ The Future of Children: Critical Issues for Children and Youths, V (1995), 113–127. Murnane, Richard, John Willet, and Frank Levy, ‘‘The Growing Importance of Cognitive Skills in Wage Determination,’’ Review of Economics and Statistics, LXXVII (1995), 251–266. National Center for Education Statistics, Digest of Education Statistics (Washington, DC: 1996), Table 166. Nye, Barbara, Jayne Zaharias, B. D. Fulton, C. M. Achilles, and Richard Hooper, ‘‘The Lasting Benefits Study: A Continuing Analysis of the Effect of Small Class Size in Kindergarten through Third Grade on Student Achievement Test Scores in Subsequent Grade Levels,’’ Seventh grade technical report, Nashville: Center of Excellence for Research in Basic Skills, Tennessee State University, 1994. Prais, S. J., ‘‘Class-Size and Learning: The Tennessee Experiment—What Follows?’’ Oxford Review of Education, XXII (1996), 399–414. Summers, Anita A., and Barbara L. Wolfe, ‘‘Do Schools Make a Difference?’’ American Economic Review, LXVII (1977), 639–652. U. S. Census Bureau, Historical Income Tables (Washington, DC: 1996), Table P25. Word, Elizabeth, J. Johnston, Helen Bain, et al., ‘‘The State of Tennessee’s Student/Teacher Achievement Ratio (STAR) Project: Technical Report 1985– 1990’’ (Nashville: Tennessee State Department of Education, 1990).

You May Also Find These Documents Helpful

  • Powerful Essays

    CITE) Knowing this as a teacher can help reduce the amount of money teachers have to take out of their pockets to spend on the classroom. This can be used for updating school technology, buying more school supplies, and updating school textbooks. However, I find that the amount of funds spent in school classroom vary on the schools. Leggett school in Akron spends 73.9% of funds in their classroom, but are failing in almost every category. (CITE) One would assume that spending more funds in the classroom would raise scores. Seeing this as a teacher would make this information not as important as others. This information has many factors that play into it like the community that surrounds the school. Depending on the school, I can use community resources like the library to borrow books for my classroom. This will help my classroom if the school that I am in is not getting the resources that I need to provide a good education for my…

    • 1473 Words
    • 6 Pages
    Powerful Essays
  • Good Essays

    Schools lacking social utilities that are needed to promote the academic status of its students is an issue. Whether these utilities should be kept opened or closed is widely debated in most communities. The condition of such schools is an important issue because it determines the future of its students academically. Some issues facing schools include social, public and economical issues; this essay will consider arguments concerning the social, public and economical causes of this problem through the use of Jonathan Kozol's "TITLE OF ARTICLE", as well as the discussion of the reasons why some schools do not receive sufficient funds to care for public schools.…

    • 1028 Words
    • 5 Pages
    Good Essays
  • Powerful Essays

    This paper includes a reference list of literature relating to the impact socioeconomic status and race/ethnicity has on academic achievement and what can be done to combat the achievement gap. In general, the literature seems to indicate that socioeconomic status has a greater impact then race or ethnicity on achievement. However, these constructs are often intertwined. The greatest source for combatting the achievement gap are teachers high in self-efficacy, strong and well directed principal leadership, having a positive and accepting racial climate, smaller class sizes, less harsh discipline with more support to reduce bad behavior, and better communication between parents and the school. Most authors agree, making education more meaningful and purposeful to the students would increase motivation to learn, which would increase success in academics. Almost all authors agree, professional development concentrated in these areas will reduce the achievement gap due to the increased academic success. There seems to be a combination quantitative and qualitative research.…

    • 3212 Words
    • 10 Pages
    Powerful Essays
  • Powerful Essays

    acknowledgements We would like to thank Barbara Hamill and Trisha Carr for their valuable assistance in creating and fielding the survey, and we are extremely grateful to Grant Olscamp for creating the data tables and Cathryn Dembek for helping prepare the report. We also thank Sarah DeLucia, Roberta Friedman, Carol Hazen, and Jennifer Pomeranz for their very helpful comments on earlier drafts. Support for this project was provided by grants from the Robert Wood Johnson Foundation and the Rudd Foundation.…

    • 23027 Words
    • 93 Pages
    Powerful Essays
  • Good Essays

    Socioeconomic Status

    • 974 Words
    • 4 Pages

    Throughout the United States, there are many regular disparities among our education system. First, there is an alarming disparity in education especially in the United States. Students from lower socioeconomic statuses do not always receive the same education as those from higher socioeconomic statuses for many reasons. In areas with lack of resources there tends to be poorer school institutions in comparison to wealthier neighborhoods. In addition, public schools are funded by taxes and therefore, the quality of teachers and amount of resources depends on the quantity of taxes individuals pay. Within these areas, families…

    • 974 Words
    • 4 Pages
    Good Essays
  • Good Essays

    Inner City School Systems

    • 687 Words
    • 3 Pages

    The school system in America has long been an issue of discussion and debate amongst people everywhere. The discussions and debates often stem from evaluating the current educational system in order to determine if significant social issues, including increasing regional poverty and declining literacy rates in specific urban regions are related to economic differentiations in the educational system. Many policy analysts have considered the issue of educational funding allotments in order to determine a system that provides greater equity between socio-economically disadvantaged inner-city schools and wealthier suburban, middle class schools (Kozol 83). The foundation for the necessary funding changes have stemmed from the recognition that school funding differences relate directly to sociological issues, including the creation of a cycle of poverty and illiteracy in under funded urban settings.…

    • 687 Words
    • 3 Pages
    Good Essays
  • Good Essays

    Elementary students had begun taking standardized tests, the tests ranked teachers in grades 3-5 accordingly. Any person is able to see the “value-added performance” (Kuehn, Larry). Anxiety led to devastating occurrences; “One teacher, distressed by being singled out, committed suicide days after the individual teacher results were released” (Kuehn, Larry). The government did not take this incident seriously, even though they attempted to think about the issues, the final answer was “test better” (Kuehn, Larry). The tests are not accurate, voluminous students do not take the test completely and honestly; those students tend to lower the teacher’s ranking. A teacher can never actually make students try their hardest on the standardized tests; the students must put forth the effort in order for all scores and teacher’s rankings to be a reflection of their…

    • 861 Words
    • 4 Pages
    Good Essays
  • Best Essays

    Weiss, C. C., Carolan, B. V., & Baker-Smith, E. (2010). Big School, Small School: (Re)Testing Assumptions about High School Size, School Engagement and Mathematics Achievement. Journal Of Youth & Adolescence, 39(2), 163-176. doi:10.1007/s10964-009-9402-3…

    • 3187 Words
    • 13 Pages
    Best Essays
  • Good Essays

    school systems than are any other group of students. A recent report of the Harvard…

    • 689 Words
    • 3 Pages
    Good Essays
  • Powerful Essays

    No Child Left Behind Act

    • 1177 Words
    • 5 Pages

    Since 1965, when the federal government embarked on its first major elementary and secondary education initiative federal policy has strongly influenced America's schools (www.nclb.gov). Over the years, Congress has created hundreds of programs intended to address problems in education without asking whether or not the programs produce results or knowing they're impact on local needs. This "program for every problem" solution has begun to add up – so much so that there is hundreds of education programs spread across thirty nine federal agencies at the cost of $120 billion dollars a year. Yet, after years of spending billions of dollars on education, the United States have fallen short in meeting the goals for educational excellence. The academic achievement gap between rich and poor and Anglo and minority are not wide, but in some cases is still growing wider.…

    • 1177 Words
    • 5 Pages
    Powerful Essays
  • Powerful Essays

    Pulic School Funding

    • 2077 Words
    • 9 Pages

    Should a child's wealth determine whether or not he or she can get a quality education? Funding for public schools is an enormous issue that affects many people. A major issue is unequal funding throughout the country. There are various reasons for these differences, from the governmental level to the lack of public support. There is a need to re-work the current formula used by states to determine how school funds are disbursed and also to increase public support for education funding. After all, whether a child is wealthy or poor, it is in everyone's interests to ensure that America's future generations are both skilled and educated. Limited funds for public schools have primarily affected the poor and have put them at a disadvantage in getting a quality education.…

    • 2077 Words
    • 9 Pages
    Powerful Essays
  • Better Essays

    "Schools as Communities, Poverty Levels ofStudent Population and Attitudes, otives and Performance." American Eduactional Research Journal (1995): 627-658.…

    • 1245 Words
    • 5 Pages
    Better Essays
  • Powerful Essays

    Standardized Tests

    • 2501 Words
    • 11 Pages

    For years, children of all ages have been required to take standardized tests in order to assess their capabilities and theoretically prepare them for further education. These tests are often the main factor in acceptance into colleges and universities, used in grade advancement for younger ages, and an overall indicator of a person’s ability. However, because of the lack of variety in the subjects they evaluate, standardized tests are often under scrutiny for whether or not they are actually as necessary as the school systems all across the county make them out to be. With ample evidence supporting both sides of the argument, there are an outstanding amount of students, principles, and community members that stand against the excessive and dependent use of standardized tests. Standardized tests are not accurate depictions of a student’s ability to perform in an academic setting whether it is college or the next grade that they face.…

    • 2501 Words
    • 11 Pages
    Powerful Essays
  • Good Essays

    Texas Education

    • 917 Words
    • 4 Pages

    Cited: Indriago, Kristen, and Nancy Druart. "Money Isn 't the Only Answer to School Woes." Statesman. N.p., 24 Feb. 2013. Web.…

    • 917 Words
    • 4 Pages
    Good Essays
  • Powerful Essays

    Raudenbush, Stephen W. Schooling, Statistics, and Poverty: Can We Measure School Improvement? Princeton, NJ: Educational Testing Service, Policy Evaluation and Research Center, Policy Information Center, 2004. Print.…

    • 1223 Words
    • 5 Pages
    Powerful Essays