Description: Tags: 20064001r
Description: Tags: 20064001r
Description: Tags: 20064001r
Interim Report
Volume I: Implementation
Institute of Education Sciences
National Center for Education National
Evaluation
and Regional Assistance Assessment of
U.S. Department of Education
Title I:
Interim Report
Volume I:
Implementation of
Title I
A Report Prepared for IES
by the Office of Planning,
Evaluation, and Policy
Development, Policy and
Program Studies Service
February 2006
NCEE 2006-4001R
National Assessment of Title I
Interim Report
Stephanie Stullich
Elizabeth Eisner
Joseph McCrary
Collette Roney
Institute of Education Sciences
National Center for Education Evaluation and Regional Assistance NCEE 2006-
4001R
U.S. Department of Education
February 2006
U. S. Department of Education
Margaret Spellings
Secretary
Call in your request toll free to 1-877-4ED-Pubs. If 877 service is not yet
available in your area, call 800-872-5327 (800-USA-LEARN). Those who
use a telecommunications device for the deaf (TDD) or a teletypewriter
(TTY) should call 800-437-0833.
I. Introduction.................................................................................. ..........................1
A. National Assessment of Title I......................................................................1
References....................................................................................................... .........85
Acknowledgments....................................................................................................87
Appendices.............................................................................................. .................89
Appendix A. Descriptions of Major Data Sources Included in This Report......89
Appendix B. Supplemental Exhibits...............................................................99
Appendix C. Standard Error Tables..............................................................109
Endnotes................................................................................................................ .131
............................................................................................................ .........................
............................................................................................................ .........................
............................................................................................................ .........................
............................................................................................................ .........................
............................................................................................................ .........................
............................................................................................................ .........................
............................................................................................................ .........................
............................................................................................................ .........................
............................................................................................................ .........................
............................................................................................................ .........................
............................................................................................................ .........................
............................................................................................................ .........................
Executive Summary
The Title I program began in 1965 as part of the Elementary and Secondary
Education Act (ESEA) and is intended to help ensure that all children have the
opportunity to obtain a high-quality education and reach proficiency on
challenging state standards and assessments. The No Child Left Behind Act of
2001 (NCLB) built upon and expanded the assessment and accountability
provisions that had been enacted as part of the ESEA’s previous reauthorizing
legislation, the Improving America’s Schools Act (IASA), while also creating new
provisions related to parental choice and teacher quality. These and other
changes were intended to increase the quality and effectiveness not only of the
Title I program, but also of the entire elementary and secondary education
system in raising the achievement of all students, particularly those with the
lowest achievement levels.
As part of the No Child Left Behind Act, the Congress mandated a National
Assessment of Title I to evaluate the implementation and impact of the program.
The mandate specifically requires a longitudinal study of Title I schools, as well
as an Independent Review Panel composed of expert researchers and
practitioners to advise the U.S. Department of Education on the conduct of the
National Assessment. An interim report is due in 2005 and a final report is due
in 2007.
The final report will provide more complete data on Title I implementation and
outcomes, including information about the targeting and uses of Title I funds,
services for private school students, findings from a parent survey about
parents’ experiences with choice options, and analyses of a) student outcomes
associated with participation in the Title I choice and supplemental services
options and b) the impact on student achievement of identifying schools for
improvement.
3. Graduation Rates
Under NCLB, high schools are held accountable for graduation rates, but
methods for calculating graduation rates vary considerably across states. The
averaged freshman graduation rate (calculated by NCES based on data from the
Common Core of Data) is useful for providing a common standard against which
state-reported graduation rates may be compared. The median state
graduation rate in 2002 was 84 percent based on state reports and 75 percent
based on the averaged freshman graduation rate.5
The recent trend in the averaged freshman graduation rate has been
fairly level, and the mean graduation rate in 2002 (73 percent) was the
same as in 1996.
While some states have standards and assessments in place in all of the
required grade levels, most states need to implement additional assessments to
meet the NCLB requirements by 2005-06 for reading and mathematics and by
2007-08 for science. As of March 2005, 27 states had completed their first full
administration of all required reading assessments; 26 states had done so for all
required mathematics assessments; and 22 states had done so for all required
iii
science assessments. Most of the remaining states had at least field-tested all of
the required assessments.6
The lowest participation rates were for students with disabilities. While
states missing the test participation requirement for other subgroups often
missed by just one or two percentage points, states that failed to assess 95
percent of students with disabilities typically had lower participation rates for
those students (as low as 77 percent in one state).
How fully are states meeting NCLB requirements for reporting state
assessment data?
The number of states that report student achievement data has more
than doubled since NCLB was enacted. Fifty states present data
disaggregated by race/ethnicity and gender and for limited English proficient
students, students with disabilities, and low-income students on state report
cards.9
iv
E. Accountability and Support for School Improvement
1. School Identification for Improvement
v
Exhibit E-5
Number and Percentage of Identified Title I Schools,
1996-97 to 2004-05
10,000
9,028
8,000 8,408 8,348 8,375 18%
7,353 20% 19% 18%
7,021
6,000 17% 6,441
16% 6,094 5,963
13%
12% 12%
4,000
2,000
0
1996- 1997- 1998- 1999- 2000- 2001- 2002- 2003- 2004-
97 98 99 00 01 02 03 04 05
Exhibit reads: In 2004-05, 9,028 Title I schools had been identified for
improvement based on test scores for 2003-04 and earlier years; these identified
schools represented 18 percent of all Title I schools in that year.
Note: The first year that schools were identified for improvement based in part on NCLB AYP
definitions was 2003-04, based on assessments administered in 2002-03. However, schools are
identified when they miss AYP for two consecutive years, and 2004-05 was the first year that includes
schools identified because they missed NCLB AYP targets for two consecutive years.
What are the reasons schools did not make adequate yearly progress
(AYP)?
vi
Schools most
commonly missed Exhibit E-6
AYP for the Reasons Schools Missed AYP, 2003-04
achievement of all
students and/or
multiple subgroups;
only in a minority of
cases did schools
miss only one AYP
target. Based on
data from 33 states,
among schools that
missed AYP in 2003-
04, 33 percent did not
meet achievement
targets for the “all
students” group in
reading or
mathematics, and
another 18 percent
missed AYP for the
achievement of two or
more subgroups (see Exhibit reads: In 2003-04 testing, 33 percent of
Exhibit E-6). Only 23 schools missed AYP for the achievement of the all
percent missed AYP students group in reading and/or mathematics.
solely due to the
achievement of a Source: Study of State Implementation of Accountability and Teacher
Quality Under NCLB (based on data from 33 states and 15,731
single subgroup. schools that missed AYP in these states).
Twenty percent missed
AYP due to the “other
academic indicator,” but only 7 percent missed for this indicator alone. More
than one-fourth (29 percent) missed AYP due to insufficient test participation
rates, but only 6 percent missed solely due to test participation. The remaining
13 percent of schools that missed AYP missed for other combinations of AYP
targets.13
However, schools that were held accountable for more subgroups were
less likely to make AYP. Among schools for which AYP was calculated for six
or more subgroups, 39 percent did not make AYP, compared with 10 percent of
schools for which AYP was calculated based on only one subgroup. More than
one-fifth of those schools that were held accountable for the achievement of
African-American students, LEP students, or students with disabilities did not
make AYP for those subgroups in 2003-04 testing. Schools with subgroups of
students from low-income families, Hispanic students, or Native American
students were somewhat less likely to miss AYP for those subgroups (12 to 15
percent). Schools were much less likely to miss AYP due to the achievement of
white or Asian students (1 percent and 4 percent of schools with these
subgroups, respectively).14
vii
3. School Improvement Activities
All states notified schools about their identification status for 2004-05
based on 2003-04 testing, and a majority provided preliminary results
before September 2004, but 20 states did not, and only 15 states
provided final results by that time.15 NCLB regulations require states to
notify schools and districts of their school improvement status prior to the
beginning of the school year; this is important to enable districts with identified
schools to notify parents of eligible students about their Title I choice options in
a timely manner.
viii
Title I schools in corrective action status nearly universally experienced
the interventions NCLB defines for schools in this stage of
improvement. Corrective actions were implemented in 95 percent of Title I
schools in corrective action status in 2004-05. The most common corrective
actions experienced by Title I schools in this status in 2003-04 and 2004-05
resembled forms of technical assistance rather than sanctions. For instance, 90
percent of Title I schools in corrective action were required to implement new
research-based curricula or instructional programs and 58 percent had an
outside expert appointed to advise the school.21
ix
F. School Choice and Supplemental Educational Services
1. Eligibility and Participation
x
The number of state- Exhibit E-8
approved supplemental Supplemental Service Providers:
service providers has tripled Share of Providers and Participants, by Provider Type, 2003-04
over the past two years, 100%
Percent of Approved Providers Percent of Participating Students
rising from 997 in May 2003
80%
to 2,734 in May 2005. 70%
59%
Private firms accounted for 76 60%
percent of approved providers 40%
40%
in May 2005 and served 59 25%
percent of participating 20%
6%
students in the previous school 0% 2% 0%
0%
year (2003-04). A growing All Private Faith-Based Districts and Colleges and
number and percentage of Providers Public Schools Universities
2. Parental Notification
The timing of parental notification was often too late to enable parents
to choose a new school before the start of the 2004-05 school year.
Almost half (49 percent) of districts notified parents after the school year had
already started, and in these districts this notification occurred, on average, five
weeks after the start of the school year.24
States report that they are working to develop and implement systems
for monitoring and evaluating the performance of supplemental service
providers, but, as of early 2005, 15 states had not established any
xi
monitoring process, 25 states had not yet established any standards
for evaluating provider effectiveness, and none had finalized their
evaluation standards. Seventeen states say they will evaluate student
achievement on state assessments, although only one of these plans to use a
matched control group. The most common approaches that states have
implemented to monitor providers are surveying the districts about provider
effectiveness (25 states) and using providers’ reports on student-level progress
(18 states).25
xii
G. Teacher Quality and Professional Development
1. State Definitions of Highly Qualified Teachers
xiii
elementary teachers in non-identified schools said they were considered not
highly qualified, compared with 5 percent in schools that were in the first or
second year of being identified for improvement, 8 percent in schools in
corrective action, and 6 percent of schools in restructuring.30
3. Professional Development
xiv
I. Introduction
The Title I program began in 1965 as part of the Elementary and Secondary
Education Act (ESEA) and is intended to help ensure that all children have the
opportunity to obtain a high-quality education and reach proficiency on
challenging state standards and assessments. As the largest federal program
supporting elementary and secondary education (funded at $12.7 billion in
FY 2006), Title I, Part A, targets these resources primarily to high-poverty
districts and schools, where the needs are greatest. Title I provides flexible
funding that may be used to provide additional instructional staff, professional
development, extended-time programs, and other strategies for raising student
achievement. The program focuses on promoting schoolwide reform in high-
poverty schools and on ensuring students’ access to scientifically based
instructional strategies and challenging academic content. Title I holds states,
school districts, and schools accountable for improving the academic
achievement of all students and turning around low-performing schools, while
providing alternatives to students in such schools to enable those students to
receive a high-quality education.
The No Child Left Behind Act of 2001 (NCLB), which went into effect beginning
with the 2002-03 school year, reauthorized the Title I program and made a
number of significant changes in key areas. NCLB strengthened the assessment
and accountability provisions of the law, requiring that states annually test all
students in grades 3-8 and once in grades 10-12 on assessments that are
aligned with challenging state standards. States must also set targets for school
and district performance that lead to all students achieving proficiency on state
reading and mathematics assessments by the 2013-14 school year. Schools and
districts that do not make adequate yearly progress (AYP) towards this goal are
identified as needing improvement and are subject to increasing levels of
interventions designed to improve their performance, as well as provide
additional options to their students. NCLB also required that states establish
definitions for “highly qualified” teachers and that all teachers of core academic
subjects become highly qualified. These and other changes were intended to
increase the quality and effectiveness not only of the Title I program, but also of
the entire elementary and secondary education system in raising the
achievement of all students, particularly those with the lowest achievement
levels.
As part of the No Child Left Behind Act, the Congress mandated a National
Assessment of Title I to evaluate the implementation and impact of the
program.34 This mandate specifically requires a longitudinal study of Title I
schools to examine the implementation and impact of the Title I program. In
addition, the law also requires the establishment of an Independent Review
Panel to advise the Secretary on methodological and other issues that arise in
carrying out the National Assessment and the studies that contribute to this
1
assessment. An interim report to Congress is due in 2005 and the final report is
due in 2007.
1
The Elementary and Secondary Education Act defines the term “state” to include the District of
Columbia and Puerto Rico (Section 9101(40)). Accordingly, this report presents data on all 52 “states”,
except in cases where data for one or more states was not reported.
2
Study of Title I Accountability Systems and School Improvement
Efforts (TASSIE). This study examines implementation of Title I
accountability provisions during the transition years from 2001-02 (prior
to implementation of NCLB) through 2003-04 (the second year of NCLB
implementation). The study surveyed a nationally representative sample
of 1,200 districts and 740 schools that had been identified for
improvement under the previous authorization of ESEA.37
References in the text to differences between groups or over time that are
based on nationally representative samples highlight only those differences that
are statistically significant using the t statistic and a significance level of 0.05.
The significance level, or alpha level, reflects the probability that a difference
between groups as large as the one observed could arise simply due to
sampling variation, if there were no true difference between groups in the
population. The tests were conducted by calculating a t value for the difference
between a pair of means and comparing that value to a published table of
critical values for t. Analyses of data on student achievement on state
assessments, percentages of schools and districts identified for improvement,
and reasons for schools not making adequate yearly progress were based on the
full population of schools as reported by each state.
The final report will provide more complete data on Title I implementation and
outcomes, including information about the targeting and uses of Title I funds,
services for private school students, findings from the NLS-NCLB parent survey
and supplemental service provider survey, and exploratory analyses of student
outcomes associated with participation in the Title I choice and supplemental
services options and of the impact on student achievement of identifying
schools for improvement.
3
A. Key Provisions of Title I Under the No Child Left Behind
Act
The No Child Left Behind Act built upon and expanded the assessment and
accountability provisions that had been enacted as part of the ESEA’s previous
reauthorizing legislation, the Improving America’s Schools Act (IASA), while also
creating new provisions related to parental choice and teacher quality. These
changes were intended to strengthen the Title I program’s ability to leverage
systemic improvements throughout states, districts, and schools, in order to
help ensure that all children have the opportunity to obtain a high-quality
education and to reach proficiency on challenging state academic standards and
assessments.
IASA initiated the Title I requirements for states to develop and implement state
standards and aligned assessments in reading and mathematics that were to be
used for all students, not just Title I students; states were required to implement
assessments aligned with state standards at least once in each of three grade
spans: grades 3-5, 6-9, and 10-12. NCLB extended the state assessment
requirements to cover testing in additional grades, requiring that states
establish reading and mathematics assessments in each grade from 3-8 and
once in grades 10-12, as well as requiring adoption of state standards and
assessments in science. In addition, NCLB established the expectation that all
students should be included in the state assessment, including students with
disabilities or limited English proficiency (LEP). NCLB also instituted a
requirement to assess the English language proficiency of LEP students.
NCLB created new educational options for students in schools that have been
identified for improvement. Allowing students to transfer to a non-identified
4
school, a rarely used “corrective action” under the previous law, is now an
option that districts must offer for all students in identified schools. In addition,
in identified schools that miss AYP for a third time, districts must offer students
from low-income families the opportunity to receive supplemental educational
services such as tutoring from a state-approved provider.
NCLB also established minimum qualification requirements for teachers and for
Title I paraprofessionals, provisions that were not previously part of the law.
Notably, the requirement that teachers must be “highly qualified” applies to all
teachers of core academic subjects and not just to teachers in Title I schools.
Exhibit 1 compares the No Child Left Behind Act with the Improving America’s
Schools Act on key provisions in the areas of assessments, accountability, and
teacher quality.
Exhibit 1
Comparison of Key Provisions of the No Child Left Behind Act of 2001 (NCLB)
and the Improving America’s Schools Act of 1994 (IASA)
NCLB IASA
State States must implement annual state States must implement annual
assessments assessments in reading and state assessments in reading and
mathematics in grades 3-8 and at least mathematics at least once in
once in grades 10-12, and in science at each of three grade spans: 3-5,
least once in each of three grade spans: 6-9, and 10-12. Assessments
3-5, 6-9, and 10-12. Assessments must must be aligned with challenging
be aligned with challenging state state content and performance
content and academic achievement standards. States must provide
standards. States must provide for for participation of all students,
participation of all students, including including students with
students with disabilities and limited disabilities and LEP students.
English proficient (LEP) students.
States must provide for the annual
assessment of English language
proficiency of all LEP students.
Adequate States must set annual targets that States must set annual targets
yearly lead to the goal of all students’ for continuous and substantial
progress achieving proficiency in reading and improvement sufficient to
(AYP) mathematics by 2013-14. For each achieve the goal of all Title I
measure of school performance, states students achieving proficiency in
must include absolute targets that must reading and mathematics, but no
be met by key subgroups of students specific timeline is mandated.
(major racial/ethnic groups, low-income Targets for school performance
students, students with disabilities, and may be absolute or relative and
LEP students). Schools and districts apply to the school as a whole,
must meet annual targets for each not to individual subgroups within
student subgroup in the school, and a school. No minimum test
must test 95% of students in each participation requirement.
subgroup, in order to make “adequate
yearly progress.” States also must
define an “other academic indicator”
that schools must meet in addition to
5
proficiency targets on state
assessments.
Schools Schools and districts that do not make Schools and districts that do not
identified AYP for two consecutive years are make AYP for two consecutive
for identified for improvement and are to years are identified for
improvemen receive technical assistance to help improvement. When a school
t them improve. Those that miss AYP for continues to miss AYP for three
additional years are identified for additional years, districts must
successive stages of intervention, take corrective action. To leave
including corrective action and “identified for improvement”
restructuring. To leave “identified for status, a school or district must
improvement” status, a school or make AYP for two consecutive
district must make AYP for two years.
consecutive years.
Exhibit 1 (continued)
Comparison of Key Provisions of NCLB and IASA
NCLB IASA
Public Districts must offer all students in Districts must offer all students in
school identified schools the option to transfer identified schools the option to
choice to a non-identified school, with transfer to a non-identified school
transportation provided by the district. unless a) the district is in a state
receiving a minimum grant (small
states), or b) the school choice
option is prohibited by state or
local law.
Supplement In schools that miss AYP for a third Not applicable.
al year, districts also must offer low-
educational income students the option of
services supplemental educational services from
a state-approved provider.
Corrective In schools that miss AYP for a fourth In schools that miss AYP for a fifth
actions year, districts also must implement at year, districts must implement
least one of the following corrective corrective actions which may
actions: replace school staff members include: withhold funds; provide
who are relevant to the failure to make health, counseling, and social
AYP; implement a new curriculum; services; revoke authority for
decrease management authority at the schoolwide program; decrease
school level; appoint an outside expert decision making authority at the
to advise the school; extend the school school level; create a charter
day or year; or restructure the internal school; reconstitute the school
organization of the school. staff; authorize students to
transfer to another school; or
implement opportunity-to-learn
standards.
Restructurin In schools that miss AYP for a fifth year, Not applicable.
g districts also must begin planning to
implement at least one of the following
restructuring interventions: reopen the
school as a charter school; replace all
or most of the school staff members;
contract with a private entity to
manage the school; turn over operation
of the school to the state; or adopt
6
some other major restructuring of the
school’s governance. Districts must
spend a year planning for restructuring
and implement the school restructuring
plan the following year.
Highly All teachers of core academic subjects Not applicable.
qualified must be “highly qualified” as defined
teachers by NCLB and the state. To be highly
qualified, teachers must have a
bachelor’s degree, full state
certification, and demonstrated
competence in each core academic
subject that they teach. Subject-matter
competency may be demonstrated by
passing a rigorous state test,
completing a college major or
coursework equivalent, or (for veteran
teachers) meeting standards
established by the state under a “high,
objective uniform state standard of
evaluation” (HOUSSE).
7
Schools that have been identified for improvement under NCLB are divided
among four stages of improvement status: 1) “Year 1” of identification, when
they must make school choice available; 2) “Year 2” of identification, when they
must also offer supplemental services; 3) corrective action status; and 4)
restructuring status. Schools move to the next stage of improvement status
when they miss AYP again, and not just because they have remained in
improvement status for another year. For example, a school that missed AYP in
2002-03 and 2003-04 would be in the first stage of improvement in 2004-05. If
the school then made AYP in 2004-05 testing, it would remain in the first stage
of improvement status and would not have to offer supplemental services. In
2005-06 testing, if the school made AYP again (for a second consecutive year), it
would move out of improvement status, and if it missed AYP, it would then move
to the next stage of improvement and would have to offer supplemental
services. Note that once a school is identified, it does not need to miss AYP in
consecutive years to move to the next stage of improvement, but it does need
to make AYP in consecutive years to move out of improvement status.
The NCLB provisions went into effect beginning with the 2002-03 school year,
but a number of important provisions do not take effect until later years (see
Exhibit 2). States developed AYP definitions using the new NCLB criteria during
2002-03 and were to use these criteria for AYP determinations, beginning with
the 2002-03 state assessment data. These determinations first affected schools
that were identified for improvement for 2003-04. However, NCLB-specified
interventions for identified schools (such as school choice, supplemental
educational services, and technical assistance) were first implemented in 2002-
03, for schools that had been identified under the AYP procedures already in
place based on the IASA provisions.
Exhibit 2
Timeline for Implementation of Key NCLB Provisions
2002-03 • States use results from assessments administered in this year to make AYP
determinations under the new NCLB provisions
• Districts implement Title I school choice and supplemental services
• Newly hired teachers and paraprofessionals must meet NCLB qualification requirements
2003-04 • First year that schools are identified for improvement based on NCLB AYP definitions
2005-06 • States implement reading and mathematics assessments in additional grades
• States develop or adopt science standards
• Existing teachers and paraprofessionals must meet NCLB qualification requirements
2007-08 • States implement science assessments in three grade spans
The new state assessment requirements are not due to be implemented until
2005-06 for reading and mathematics and 2007-08 for science. Thus, AYP
determinations prior to 2005-06 are based on state assessments adopted under
the previous law. The highly qualified teacher requirements went into effect in
8
2002-03 for newly hired staff members in Title I schools, but existing staff
members teaching core academic subjects in all schools have until the end of
the 2005-06 school year to meet the requirements. Due to the extended
timeline for implementing a number of NCLB requirements, this report often
examines progress towards deadlines that have not yet arrived, as well as the
extent to which states, districts, and schools are implementing NCLB
requirements already in effect.
9
B. Profile of Title I Participants and Resources
Title I Part A funds go to nearly all (93 percent) of the nation’s school
districts and to 55 percent of all public schools. Schools may use Title I
funds for one of two approaches: schoolwide programs, or targeted assistance
programs. High-poverty schools2 (those with 40 percent or more students from
low-income families) are eligible to adopt schoolwide programs to raise the
achievement of low-achieving students by improving instruction throughout the
entire school. Schools that are not eligible for (or do not choose to operate)
schoolwide programs must use Title I funds to provide targeted services to
specifically identified low-achieving students. Schoolwide programs accounted
for 54 percent of all Title I schools in 2002-03, and the use of the schoolwide
option has been growing steadily over the past decade (see Exhibit 3).39
Exhibit 3
Number of Schoolwide Programs and Targeted
Assistance Schools, 1994-95 to 2002-03
Schoolwide Programs
60,000
Targeted Assistance Schools
50,000 5,050
30,000
46,638
20,000
31,763
27,084 23,977
10,000 24,842
0
1994-95 1996-97 1998-99 2000-01 2002-03
Exhibit reads: The number of school wide programs increased from 5,050 in
1994-95 (10 percent) to 28,162 in 2002-03 (54 percent).
Source: Consolidated State Performance Reports (for 50-52 states).
2
School poverty levels are usually based on the percentage of students eligible for the free and
reduced-price lunch program, although districts have the flexibility to use certain other measures
described in the law. In this report, survey data for “high-poverty schools” included schools where at
least 75 percent of the students were eligible for free or reduced-price lunches, and “low-poverty
schools” included schools where fewer than 35 percent were eligible for such lunches. For NAEP data,
“high-poverty schools” included schools where 76-100 percent of the students were eligible for free or
reduced-price lunches, and “low-poverty schools” were defined as those with 0-25 percent eligible for
subsidized lunches).
10
Fueled by the growth in schoolwide programs, the number of students
counted as Title I participants has more than doubled in recent years,
rising from 6.7 million in 1994-95 to 16.5 million in 2002-03 (a 146
percent increase). The dramatic increase in participation is due in part to the
way that students are counted: when a school converts from targeted assistance
to a schoolwide program, all students in the school are counted as Title I
participants, instead of just the lowest-achieving students who are receiving
specific targeted services. In 2002-03, 84 percent of Title I participants were in
schoolwide programs and only 15 percent were in targeted assistance schools;
the remaining participants were in private schools (1 percent) or local
institutions for neglected or delinquent children (1 percent).40
11
Funding for Title I Part A has increased by 61 percent over the past five
years, from $7.9 billion in FY 2000 to $12.7 billion in FY 2005. After
adjusting for inflation, Title I funding increased by 46 percent from 2000 to 2005
and by 138 percent since the program’s inception in 1965 (see Exhibit 5).
A variety of formulas used to allocate Title I funds are intended to target these
resources to high-poverty districts and schools, where the needs are greatest.
About half of all Title I Part A funds are allocated to districts in the highest-
poverty quartile, which have 25 percent of all children and 49 percent of the
nation’s poor children. The targeting of Title I funds to high-poverty districts
increased slightly from FY 1997 to FY 2004, with the share of funds for the
highest-poverty districts rising from 50 percent to 52 percent (see Exhibit 6).
The final report will provide more detailed information on the targeting and uses
Exhibit 5
Appropriations for Title I Grants to LEAs,
FY 1966 to FY 2005
(in 2005 Constant Dollars)
$ in Billions
14
12
10
0
1966 1970
66 70 75 1980
80 85 1990
90 95 2000
00 05
Year
12
Exhibit 6
Distribution of Title I Funds
by District Poverty Quartile, FY 1997 and FY 2004
100%
Highest Poverty
80% Quartile
50% 52%
Second Highest
60% Poverty Quartile
Second Lowest
40%
27% 27% Poverty Quartile
Note: District poverty quartiles are based on Census Bureau estimates of the number of school-
age children and poor children living in each district. The poverty quartiles were created by
ranking all districts by the percentage of poor school-age children and then dividing these
districts into quartiles that each contain 25 percent of the school-age children.
Sources: Study of Education Resources and Federal Funding (FY 1997); National Longitudinal
Study of NCLB (FY 2004) (based on data for 51 states).
13
Key Findings on Trends in Student Achievement
In the 23 states that had consistent three-year trend data from 2000-01
to 2002-03, most student subgroups showed gains in the percentage of
students performing at or above the state’s proficient level in 4th- and
8th-grade reading and mathematics. The increases in student
proficiency were often small.
Recent trends on the Main NAEP assessment (from 2000 to 2005) show
gains in 4th-grade reading and mathematics for black and Hispanic
students and for students in high-poverty schools. Gains were larger
for mathematics than for reading, and achievement trends were less
positive for older students. Gains were especially large on the Long-
Term Trend NAEP, and the most recent gains for black and Hispanic 9-
year-olds from 1999 to 2004 substantially extended the gains these
groups had made since the 1970s.
14
Under NCLB, high schools are held accountable for graduation rates,
but methods for calculating graduation rates vary considerably across
states. State-reported graduation rates for 2002 ranged from a high of
97 percent in North Carolina to a low of 62 percent in Georgia.
15
III. Trends in Student Achievement
This chapter examines trends in student achievement using both state
assessment data and the National Assessment of Educational Progress (NAEP).
Student achievement on state assessments represents the primary criterion that
the Title I legislation applies to measure school success, but these data cannot
be aggregated across states to examine national trends or used to make
comparisons among states. Because each state has developed its own
standards and assessments, the content and rigor of these assessments are not
comparable across states. In addition, many states have revised their
assessment systems in recent years, so they often do not have the trend data
needed to assess student progress. The National Assessment of Educational
Progress provides a high-quality assessment that is consistent across states,
making the data useful for examining national trends in student achievement.
However, the NAEP is not aligned with individual state content and achievement
standards, so it does not necessarily measure what students are expected to
learn in their states. This report draws on both types of assessment to examine
the best available information about the recent progress of our schools in raising
student achievement.
This interim report examines trends on the Main NAEP from the early 1990s
through 2005, with a focus on the most recent period from 2000 to 2005, in
order to show trends in NAEP results during the early years of NCLB
implementation. We also examine long-term trends on the Trend NAEP from the
1970s through 2004. For state assessments, we examine recent three-year
trends (2000-01 through 2002-03) in 23 states that had consistent assessments
in place over this period. The report focuses on presenting achievement trends
for 4th-grade reading3 and mathematics assessments (because Title I funds are
predominantly used at the elementary level), although assessments for other
grades are shown as well.
For both state assessment and NAEP results, recent achievement trends are
positive overall and for key subgroups. At this early stage of NCLB
implementation—states, districts, and schools only began to implement the
NCLB provisions in 2002-03—it is too early to say whether these trends are
attributable to NCLB, to other improvement initiatives that preceded it, or to a
combination of both. The data presented in this chapter provide a baseline
indicator of achievement levels and trends that existed at the time that NCLB
implementation began, rather than an indicator of outcomes associated with
NCLB. They may very well reflect pre-existing state standards-based reform
efforts and accountability systems that NCLB was intended to strengthen.
Moreover, even when additional years of assessment data become available,
3
For simplicity, the term “reading” is used throughout this report to refer to the set of subjects that
may be variously known as reading, English, or language arts.
16
such data will be limited in their ability to precisely address the impact of NCLB,
because it is difficult to separate the impact of NCLB from the effects of other
state and local improvement efforts.
17
Key Evaluation Questions for Student Achievement
The report focuses on presenting achievement trends for 4th-grade reading and
mathematics; however, many states did not administer assessments in the 4th
grade in all three years and in such cases we used state assessment data for an
adjacent grade. Assessments for other grade levels were also examined and are
presented in an appendix.
18
used to make comparisons across states. In addition, caution should be used
when examining changes over time in the proportion of students performing at
or above each state’s proficiency level. The data come from the Consolidated
State Performance Reports submitted by each state to the U.S. Department of
Education, and cannot speak to the reasons for observed losses or gains over
time within each state. Observed losses or gains could reflect a number of
things, including changes in the assessment system, population changes, or
changes in the proficiency level of a stable population.
19
Exhibit 7
Percentage of 4th-Grade Students Achieving At or Above the "Proficient" Level
on NAEP and State Assessments in Reading, 2003
Connecticut 69
43
Massachusetts 56
40
New Hampshire 77
40
New Jersey 78
39
Colorado 87
37
Minnesota 76
37
Vermont 37
Maine 49
36
Iow a 76
35
Montana 77
35
Virginia 72
35
Missouri 34
34
New York 34
Ohio 66
34
Wyoming 41
34
Delaw are 79
33
Indiana 74
33
Kansas 69
33
North Carolina 81
33
Pennsylvania 58
33
South Dakota 85
33
Washington 67
33
Wisconsin 80
33
Florida 61
32
Maryland 58
32
Michigan 66
32
Nebraska 83
32
North Dakota 74
32
Utah 78
32
Illinois 60
31
Kentucky 62
31
Oregon 83
31
Idaho 75
30
Rhode Island 62
29
West Virginia 29
Alaska 71
28
Arkansas 61
28
Georgia 80
27
Texas 86
27
Oklahoma 65
26
South Carolina 32
26
Tennessee 81
26
Arizona 64
23
Alabama 63
22
California 39
21
Haw aii 43
21 State Assessment, 2003
Louisiana 61
20 NAEP, 2003
Nevada 45
20
New Mexico 70
19
Mississippi 87
18
District of Columbia 46
10
0 20 40 60 80 100
Source: Consolidated State Performance Reports and National Center for Education Statistics, Main
NAEP. Note: The preferred grade for this table was 4th grade; however, in states
that did not consistently assess students in 4th-grade reading, we used either 3rd- or 5th-grade
20
assessment results.43
21
Center for Education Statistics has not determined that they are “reasonable,
valid, and informative to the public.” NAEP and current state assessments were
established at different times to meet different purposes, and there is no one
“right” level that should be defined as “proficient.” Under NCLB, each state has
been given the responsibility to establish standards and assessments and to
define a “proficient” level that all students are expected to reach by 2013-14. In
contrast, when the NAEP proficiency levels were created about 15 years ago,
there was no expectation that all students must reach the NAEP Proficient level
by a particular date. Assessment systems vary tremendously, both between
NAEP and state systems, as well as across states that are using different
approaches with the NCLB framework, and similar-sounding terms often may not
be comparable.
4
The total number of states examined here varies because states often did not have disaggregated
assessment data available for all student subgroups for the full time period examined here (dating
back to the 2000-01) school year).
22
Exhibit
Exhibit109
Proportion
Proportion
of of
Students
Students
Performing
PerformingAtAt
or or
AboveExhibit
Above
Their 8State’s
Their State’s
Proficient
Proficient
Level
Level
forfor
Mathematics
Reading in in
2002-03,
2002-03,
and ChangeProportion of Students
from 2000-01, Performing
in 4th Grade At or Above
or Another Their Grade,
Elementary State’sfor
Proficient
VariousLevel
Student Subgroups
in Reading and Mathematics, in 4th Grade or Another Elementary Grade, 2000-01 to 2002-03
Low-Income Black Hispanic
Reading White LEP
Mathematics Migrant Disabilities
Percent Chang Percent
2000-01Change Percent2002-03
2001-02 Change Change
Percent Change
2000-01 Percent
2001-02 Chang
2002-03PercentChangeChang Percent Change
Proficien e from Proficien from Proficien
Alabama 64 63 from Proficien
-1 from
69 Proficien e from
64 Proficient -5 e from Proficien from
t in 2000- t in 2000-01 t in 2000-01 t in 2000-01 t in 2000- in 2002- 2000- t in 2000-01
Arizona
2002-03 01 2002-0375 2002-03 64 -11
2002-03 57 2002-03 01 57 03 0 01 2002-03
Connecticut 71 69 69 -2 81 80 81 2
Arizona 37 18
6 33
37 18
5 28
32 -2
6
Delaware 75 80 79 4 73 72 74 1
Connecticut Illinois 62 63 60 -2 74 1874
45 1 73
5 -1 38
47 -3
4
Delaware 68
62
Kansas 7
3 63 63 69 6 67 6767
51 28
11 74 7 44
41 14
13
Illinois Kentucky
41
56 1
3 58 60 62 4 34 4136
49 5 38
-9 34
48 4 -1 0 33
54 2
3
Kansas Louisiana
55
61 10
12 44
48 59 14
11 57 52
56 61 19
13 2 79
74 54
11
9 5050 29 60
12 51
52 6 21 11 49
59 16
13
Maine 51 49 49 -2 23 23 28 5
Kentucky 51
26 6
5 43
19 -8
6 54
31 -2
5 65
40 4
9 39
28 0
4 47
19 -2
5 43
19 11
8
Massachusetts 51 54 56 5 51 39 40 -11
Louisiana Mississippi 81 84 87 6 63 5672
61 0 74
6 11 30
35 14
11
Maine Missouri 30
7 32 -3 36 40
22 34 10
15 2 29
50 374
8 2938
18 7 37
1 20 0 9 10
20 12
-3
Massachusetts Montana 30
15 79 6
5 73 26
15 77 7
4 -2 48
65 737
6 1669
14 2 75
4 25
17 2 5 9 26
18 9
6
Mississippi New Jersey 80
61 79 148 79 91
80 78 3
2 -1 88
95 663
8 66 68 2
North Carolina 74 77 81 7 87 89 92 5
Missouri 22
24 3
2 14
21 3
2 23
21 14
2 20
18 5
3
Ohio 56 66 66 10 59 62 59 0
Montana 65
64
Oklahoma 0
2 66 63 65 -1 64 2662
32 -8 65
-9 1 36
40 1
4
North Carolina 70
87
Oregon 10
9 71
87 84 14
13 85 64
82 83 -1
1 -1 95
89 756
2 4877
72 -3
1 78 60
80 3 4 9 48
71 4
3
Ohio Pennsylvania 56 57 58 2 54 4253 5 56
-5 27
25 2 -18
-14 36
34 -1
-9
Oklahoma South64 Carolina11
63 37 34 32 -5 26 3836
48 1 33
-5 59
69 7 17 9 19
23 -1
2
Utah 82 80 78 -4 73 74 74 1
Oregon 53
51 -6
4 50
48 -6
5 49
51 -10
-1
Virginia 64 71 72 8 77 80 83 6
Pennsylvania 36
35
Washington 1
2 67 66 67 0 43 1952
28 7 55 25
29 12 3 4 22
24 6
5
South Carolina # of20
18
states with-3
6 17 12
-1 22
26 12
-6 43
47 201 14
7 -7
2 14
12 -2
-5 35
37 24
29
11 out of 23 states 17 out of 23 states
Utah achievement
65 gains
-4 38
36 -16
-11 65
60 14
12 41 -10
-5
Virginia 56
75 30
11 35
56 16
13 54
64 19
29
Washington Exhibit reads: The proportion of students performing at20 24or above 0
8 Alabama’s
30
24 11
4 31
25 1
8
# of states with 8 “proficient”
out
10 of
out11ofstates
10 level in7 states
5 out of 4th-grade 6 outreading
5 of 7 states (or another nearby 15
7 out of 7 states elementary
12 out of 20 11grade)
out
12 out
of 15
of states
16 16 out of 20 states
achievement states states states 14 out of 20 states
gains
declined from 64 percent in 2000-01 to 63 percent in 2002-03. Overall, states
that had consistent assessments during this period showed increases in the
Note:
Note:The percent
Thepreferred
preferred grade proficient
gradefor
forthis tableon
thistable was
wasthese
4th elementary
4thgrade;
grade; however, reading
however,ininstates
statesthatassessments
thatdid
didnot in 11 out
notconsistently
consistently of students
assess
assess 23
studentsin
in4th-grade
4th-grade
reading,
mathematics,
a nearbystates.
a nearby
grade grade
was used
was(3rd
usedgrade
(gradefor3Arizona,
for Arizona,
Delaware,
Delaware,
Illinois,
Illinois,
Missouri,
Missouri,
Oregon,
Oregon,
and and
Virginia
Virginia
and and
5th grade for
5 for
Kentucky,
Kentucky,Oklahoma,
Oklahoma,and andPennsylvania).
Pennsylvania). Gray cells indicate that the state did not report disaggregated assessment data for that
subgroup for all three
Note:years included grade
The preferred in the analysis; however,
for this table all grade;
was 4th of these states have
however, sincethat
in states developed
did not the capacity to report
disaggregated
Source: Consolidated
data. State Performance Reports (for 21 states).
consistently assess students in 4th-grade reading and mathematics, a nearby grade was used
(3rd grade for Arizona, Delaware, Illinois, Missouri, Oregon, and Virginia and 5th grade for
Source: Consolidated State
Kansas Performance
(reading), Reports
Kentucky (forOklahoma,
(math), 21 states).and Pennsylvania).
23
cases, the gap reduction was from one to three percentage points; however, a few
states showed larger reductions in the gap.
Exhibit 11
Change in the Achievement Gap: Difference Between the Proportion of Low-Income Students and All
Students Performing At or Above Their State’s Proficient Level, in 4th Grade or Another Elementary Grade,
2000-01 to 2003-03
24
An important question is whether these recent growth rates will be sufficient to
bring states to the goal of 100 percent of their students’ performing at or above
their state’s proficient level by the 2013-14 school year. To examine this
question, we calculated the average annual change in each state’s percent
proficient over the period from 2000-01 to 2002-03, and determined the percent
proficient that would be attained by 2013-14 if the state continued to progress at
that rate.5 Exhibit 12 shows these calculations for the low-income subgroup, and
Exhibit 13 summarizes the number of states that would be predicted to meet the
100 percent goal for six different student subgroups.
Based on trend data for 21 states, most would not meet the goal of 100
percent proficiency by 2013-14 unless the percentage of students
achieving at the proficient level increases at a faster rate. For example,
among the 11 states that had consistent elementary reading assessment data for
low-income students, four states would meet the 100 percent goal by 2013-14 for
this subgroup if they sustained the same rate of growth that they achieved from
2000-01 to 2002-03. Not surprisingly, states that began the period with a
relatively low percentage of students performing at the proficient level defined by
the state were often less likely to be predicted to meet the 100 percent goal.
Exhibit 12
Predicted Percentage of Low-Income Students That Would Reach Their State’s Proficient Level in 2013-14,
in Elementary Reading, If Achievement Trajectories from 2000-01 to 2002-03 Continued Through 2013-14
5
More specifically, we multiplied the annualized percentage-point change from 2000-01 to 2002-03 by
the number of years remaining to 2013-14 (11 years), and added that figure to the percent proficient in
2002-03. If the product was greater than 100 percent, the predicted percent proficient in 2013-14 is 100
percent (since there cannot be more than 100 percent of students reaching the proficient level). It
should be noted that this method assumes no variation in the rate of change.
25
Exhibit reads: The percent of low-income students reaching Delaware’s
proficient level in 3rd-grade reading rose from 61 percent in 2000-01 to 68
percent in 2002-03, an average gain of 3.5 percentage points per year. If this
rate of increase were sustained over the next 11 years from 2002-03 to 2013-14,
Delaware would succeed in having 100 percent of these students reach the
proficient level.
26
Looking across six different student subgroups (low-income, black,
Hispanic, LEP, migrant, and students with disabilities), an average of 33
percent of the subgroups within these states would be predicted to
reach 100 percent proficiency based on current growth rates. This
percentage declines to 26 percent of subgroups in elementary mathematics and
to 13 to 18 percent in 8th-grade reading and mathematics (see Exhibit 13).
Exhibit 13
Predicted Number of States That Would Reach the Goal of 100% Proficient by 2013-14,
for Various Subgroups, If Achievement Trajectories from 2000-01 to 2002-03 Continued Through 2013-14
Grade 3, 4, or 5 Grade 6, 7, or 8
Student Subgroup
Reading Mathematics Reading Mathematics
Low-income 4 out of 11 states 3 out of 10 states 3 out of 10 states 1 out of 9 states
Black 3 out of 6 states 3 out of 5 states 4 out of 10 states 3 out of 10 states
Hispanic 2 out of 6 states 1 out of 5 states 2 out of 10 states 2 out of 10 states
Limited English proficient 3 out of 19 states 3 out of 19 states 1 out of 16 states 2 out of 18 states
Migrant 6 out of 15 states 5 out of 15 states 3 out of 13 states 2 out of 15 states
Students with disabilities 7 out of 19 states 4 out of 19 states 1 out of 17 states 0 out of 17 states
Average proportion of state
33% 26% 18% 13%
subgroups predicted to reach 100%
Exhibit reads: For the low-income student subgroup, four out of 11 states would
reach the state’s proficient level on an elementary reading assessment, if the rate
of change from 2000-01 to 2002-02 were to continue through 2013-14.
Note: The average shown at the bottom of each column is based on summing the numerators and
denominators reflected in the cells of that column, and dividing the total of the numerators by the
total of the denominators.
Source: Consolidated state performance reports (for 21 states).
Most state AYP targets do not project an even growth rate over the full period
from 2002-03 to 2013-14; indeed, states use a variety of growth trajectories for
their AYP targets, and many are planning for achievement growth rates to
accelerate as 2013-14 approaches. Based on recent achievement trajectories,
such acceleration will often be necessary if states are to meet the goal of 100
percent proficient by 2013-14.
In general, the Main NAEP places greater emphasis on open-ended and extended
response items and less emphasis on multiple-choice questions. In reading, the
27
Trend NAEP features shorter passages and focuses on locating specific
information, making inferences, and identifying the main idea of a passage,
whereas the Main NAEP requires students to read longer passages and also asks
students to compare multiple texts on a variety of dimensions. In mathematics,
the Trend NAEP focuses on basic computational skills in four content areas—
numbers and operations, measurement, geometry, and algebra—while the Main
NAEP also includes data analysis and probability.45
Results from the Main NAEP and Trend NAEP are not comparable because they
cover different content and also different samples. Students are sampled by
grade for the Main NAEP (grades 4, 8, and 12) and by age for the Trend NAEP
(ages 9, 13, and 17). In addition, the Main NAEP reports on the percentages of
students performing at various achievement levels (Basic, Proficient, and
Advanced) as well as average scale scores, while the Trend NAEP reports only
scale scores. The National Center for Education Statistics (NCES) has stated that,
although results from these two NAEP assessments cannot be compared directly,
comparisons of the patterns they show over time, especially for student
demographic groups, may be informative.
The most recent NAEP results are from 2004 on the Trend NAEP and 2005 on the
Main NAEP. The discussion below examines both recent trends (since 1999 on the
Trend NAEP and since 2000 on the Main NAEP), in order to show trends in NAEP
results during the early years of NCLB implementation, as well as longer-term
trends on both NAEP assessments.
1. Main NAEP
28
Exhibit 14 Exhibit 15
Reading Achievement on the Main NAEP, 1992 to 2005: Mathematics Achievement on the Main NAEP, 1990 to 2005:
Average Scale Scores by School Grade Level Average Scale Scores by School Grade Level
320 320
12th Grade
290*
12th Grade 300 303*
300
286 289 * 300
285 294* 297
* 8th Grade
280 280
272* 276 * 278
8th Grade 260 267* 269*
260 262*
261 263* 261* 260 237
258* 257* 234*
240
240
222* 224*
219* 4th Grade
4th Grade 220 212*
220
* Indicates that the score is significantly different from the most recent score (2005 for 4th and 8th
grade, 2002 for 12th grade reading, and 2000 for 12th grade mathematics) (p<.05).
29
Looking at high-poverty schools, defined as those with 75 percent or more of their
students eligible for free or reduced-price lunches, average scale scores rose from
2000 to 2005 by 14 points in 4th-grade reading and 16 points in 4th-grade
mathematics (see Exhibits 16 and 17). Over the complete period during which
the Main NAEP assessment was administered, there was no significant change in
reading, but a markedly higher gain in mathematics (a 27-point gain). In short,
NAEP mathematics scores in high-poverty schools rose over the period from 1990
to 2005, including the most recent five-year period, while NAEP reading trends for
these schools fluctuated up and down but were at about the same level in 2005
as in 1992.
Exhibit 16 Exhibit 17
Reading Achievement on the Main NAEP, 1992 to 2005: Mathematics Achievement on the Main NAEP, 1990 to 2005:
Average Scale Scores in 4th Grade by School Poverty Level Average Scale Scores in 4th Grade by School Poverty Level
Low-Poverty Schools
260 251
260 247*
Low-Poverty Schools 239*
240 235* All Schools
240 233 232* 233 230*
230 230 237
225 * 225 * 234 *
218 *
All Schools 220
220 222 * 224 *
219 * 221
217 216* 217 216*
215 212 * 213 * 211 *
212 *
209 *
200 200 205* High-Poverty Schools
160 160
1992 1994 1998 2000 2002 2003 2005
1990 1992 1996 2000 2003 2005
* Indicates that the score is significantly different from the one in 2005 (p<.05).
Note: “High-poverty” was defined as schools with 76 to 100 percent of their students eligible for free or
reduced-price lunches, and “low-poverty” indicates that 0 to 25 percent were eligible for subsidized
lunches.
Source: National Center for Education Statistics, Main NAEP, unpublished tabulations.
30
Recent NAEP trends by race/ethnicity also show gains in both 4th-grade
reading and mathematics for black and Hispanic students (see Exhibits 18
and 19). From 2000 to 2005, black students gained 10 points in 4th-grade
reading and Hispanic students gained 13 points, both greater than the 5-point
gain for white students over the same time period. In 4th-grade math, black
students gained 17 points from 2000 to 2005 and Hispanic students gained 18
points, again greater than the 13-point gain for white students.
Exhibit 18 Exhibit 19
Reading Achievement on the Main NAEP, 1992 to 2005: Mathematics Achievement on the Main NAEP, 1990 to 2005:
Average Scale Scores in 4th Grade by Race/Ethnicity Average Scale Scores in 4th Grade by Race/Ethnicity
260 260 White
246
243*
240 White 240 233 *
231 * Hispanic
227 227 228 227 *
223 * 222* 223* 223 * 225
219 * 221*
220 220
Hispanic 207 * 207* 220
216*
199 199 201 199 * 201 * Black
200 194 * 192 * 200
186* 188 * 203 *
198 197* 199 198 *
191 * 192 * Black 192 *
189 *
180
184* 180 187 *
160 160
1992 1994 1998 2000 2002 2003 2005 1990 1992 1996 2000 2003 2005
* Indicates that the score is significantly different from the one in 2005 (p<.05).
Over the longer term, 4th-grade mathematics scores show even larger gains from
1990 to 2005 for black and Hispanic students (33 points and 26 points,
respectively), while white students gained 27 points. In 4th-grade reading, the
13-year trend from 1992 to 2005 shows somewhat smaller gains for black and
Hispanic students (eight points and seven points, respectively).
31
Looking at the trends on 4th-grade NAEP assessments in terms of the
percentage of students achieving at or above the proficient level,
patterns are mixed. For the recent period from 2000 to 2005, black students
show a modest increase on the reading assessment from 9 percent proficient in
2000 to 12 percent proficient in 2005, while Hispanic or white students show no
significant change (see Exhibits 20 and 21). On the mathematics assessment,
however, all three racial/ethnic groups show significant gains in the percentage
achieving at or above the proficient level, with black students rising from 4
percent proficient to 13 percent proficient, and Hispanic students rising from 7
percent to 19 percent. Over the longer period since the early 1990s, 4th-graders
in all three racial/ethnic groups show significant gains in both subjects,
particularly in mathematics.
Exhibit 20 Exhibit 21
Reading Achievement on the Main NAEP, 1992 to 2005: Mathematics Achievement on the Main NAEP, 1990 to 2005:
Percent Proficient in 4th Grade by Race/Ethnicity Percent Proficient in 4th Grade by Race/Ethnicity
50% 50% 47%
White 42%*
39% 39% 39% White
40%
35%* 36% * 36% 40%
33% *
30%*
30% 30% 26% *
Hispanic
22% *
Hispanic 19%
20%
20%
14% 14% 15% 15% * 15%*
11% 12% 12%
10% *
10%
12% 12% 12% 10% 7%* 7% * 13%
10% * 9% * 4% * 5% *
8% * 8% * Black 10%* Black
2%*
0% 1%* 3% * 4%*
0%
1992 1994 1998 2000 2002 2003 2005
1990 1992 1996 2000 2003 2005
* Indicates that the score is significantly different from the one in 2005 (p<.05).
32
2. Trend NAEP
Exhibit 22 Exhibit 23
Reading Achievement on the Trend NAEP, 1971 to 2004: Mathematics Achievement on the Trend NAEP, 1973 to 2004:
Average Scale Scores by Student Age Group Average Scale Scores by Student Age Group
* Indicates that the score is significantly different from the one in 2004 (p<.05).
208 in 1971 to 219 in 2004, an 11-point gain, compared with a 4-point gain for
13-year-olds and no change for 17-year-olds.
33
Black and Hispanic students show substantial gains on the Trend NAEP,
both in the most recent period as well as over the full three decades
covered by the assessment (see Exhibits 24 and 25). From 1999 to 2004,
black 9-year-olds gained 14 points in reading and 13 points in mathematics; long-
term gains were 30 points in reading (since 1971) and 34 points in mathematics
(since 1973). Similarly, Hispanic 9-year-olds gained 12 points in reading and 17
points in mathematics from 1999 to 2004, with long-term gains of 22 points in
reading and 28 points in mathematics. In reading, black and Hispanic students
made strong gains in the 1970s, but the trends leveled out during the 1980s and
Exhibit 24 Exhibit 25
Reading Achievement on the Trend NAEP, 1971 to 2004: Mathematics Achievement on the Trend NAEP, 1978 to 2004:
Average Scale Scores for 9-Year-Olds by Race/Ethnicity Average Scale Scores for 9-Year-Olds by Race/Ethnicity
160
160 1973 1978 1982 1986 1990 1992 1994 1996 1999 2004
1971 1975
1975 1980
1980 1984 1988 1990
19901992 1994 1996 1999 2004
* Indicates that the score is significantly different from the one in 2004 (p<.05).
1990s until the most recent jump in scores from 1999 to 2004. In mathematics,
black and Hispanic scores rose through the late 1970s and 1980s, were fairly flat
during the 1990s, and then increased dramatically from 1999 to 2004.
34
C. Graduation Rates
Under NCLB, in addition to reading, math, and eventually science achievement,
high schools are held accountable for graduation rates. In AYP determinations for
2003-04, one-third of high schools did not meet their state’s graduation rate
target.
States have differing methods for calculating and reporting graduation rates, so
they are not consistent across the country and cannot provide a national picture
of progress on this indicator. To provide more consistent data, the Task Force on
Graduation, Completion, and Dropout Indicators recommended the use of a
graduation rate measure called the Exclusion-Adjusted Cohort Graduation
Indicator (EACGI), which is also referred to as a true cohort graduation rate.46 In
order to calculate this indicator, longitudinal individual student-level data systems
are needed, and most states do not yet have such data systems in place.
The recent trend in the averaged freshman graduation rate, from 1996
to 2002, has been fairly level, and the graduation rate in 2002 (73
percent) was the same as in 1996. Graduation rates calculated for the
preceding five years were slightly higher (e.g., 76 percent in 1991), but these
6
For more information about various graduation indicators, see Marilyn Seastrom, Chris Chapman,
Robert Stillwell, Daniel McGrath, Pia Peltola, Rachel Dinkes, and Zeyu Xu (forthcoming), A Review and
Analysis of Alternative High School Graduation Rates, Volume I. User’s Guide to Computing High School
Graduation Rates, Volume 2, An Analysis of Alternative High School Graduation Rates (NCES 2006-602).
Washington, DC: U.S. Department of Education, National Center for Education Statistics.
35
data may not be strictly comparable because of improvements in reporting over
time.
36
Exhibit 26
Comparison of Averaged Freshman Graduation Rates
and State-Reported Graduation Rates, 2002
New Jersey 86 90
North Dakota 85
85
Wisconsin 91
85
Iow a 89
84
Minnesota 88
84
Nebraska 85
84
Vermont 90
82
Connecticut 89
80
Maryland 84
80
Montana 84
80
Pennsylvania 86
80
Utah 86
80
Idaho 81
79
South Dakota 96
79
Massachusetts 94
78
New Hampshire 85
78
Illinois 85
77
Kansas 83
77
Missouri 82
77
Ohio 83
77
Virginia 85
77
Maine 86
76
Oklahoma 76
Rhode Island 81
76
Arkansas 84
75
Arizona 76
75
Colorado 82
75
Texas 81
74
West Virginia 83
74
Wyoming 77
74
California 87
73
Indiana 91
73
Michigan 86
73
Haw aii 79
72
Nevada 75
72
Washington 79
72
Oregon 81
71
Delaw are 83
70
Kentucky 81
70
District of Columbia 68
North Carolina 97
68
New Mexico 67
Alaska 67
66
Louisiana 64
Florida 63 68
Alabama 62
Georgia 6162
Mississippi 81
61
New York 77
61
Tennessee 76
60
South Carolina 78
58
40 60 80 100
Sources: U.S. Department of Education, National Center for Education Statistics, averaged freshman
graduation rates calculated from data in the Common Core of Data.48 U.S. Department of Education,
Policy and Program Studies Service, analysis of state-reported graduation rates from Consolidated
State Performance Reports and State Education Agency Web sites; state-reported rates for 2003 or
2004 were used for 16 states where 2002 rates were not available.49
37
38
Conclusions
NCLB established the ambitious goal of, by 2013-14, having all children achieve
proficiency in reading and math according to state standards. Recent data from
both state assessments and the National Assessment of Educational Progress
show promising trends. Although the NAEP and state assessment data were not
designed to address questions about the causal impact of NCLB, they can still be
informative for examining changes over time in student achievement. Student
achievement as measured by state assessments rose from 2000-01 to 2002-03
for most student subgroups—such as low-income students, blacks, Hispanics,
migrants, and those with limited English proficiency or with disabilities—in a
majority of the states where consistent assessment practices make it possible to
track trends from 2001 to 2003. Similarly, recent trends on the Main NAEP
assessment show gains in 4th-grade reading and mathematics for black and
Hispanic students and for students in high-poverty schools; however, recent
trends are mixed for 8th-grade students and not available for 12th grade students.
The long-term achievement trends measured by the Trend NAEP show significant
gains for 9-year-olds and 13-year-olds in both reading and mathematics, although
both recent and long-term trends on the Trend NAEP are flat for 17-year-olds. It
remains to be seen whether the current trajectories will remain steady or
accelerate in the years to come; the latter will be required if all states are to reach
the 100 percent proficient target within the next decade.
39
Key Findings on Implementation of State Assessment Systems
Many states have put in place all of the assessments they intend to use
to meet NCLB requirements. As of March 2005, 27 states had completed
their first full administration of all required reading assessments; 26
states had done so in all required mathematics assessments; and 22
states had done so in all required science assessments. Most of the
remaining states had at least field-tested all of the required assessments.
The lowest test participation rates were for students with disabilities.
While states missing the test participating requirement for other
subgroups often missed by just one or two percentage points,
participation rates for students with disabilities were often well below the
95 percent threshold—as low as 77 percent in Texas and 84 percent in
the District of Columbia.
40
students with disabilities, and students from low-income families. In
contrast, in 2002-03, only 20 states disaggregated data by race/ethnicity
on report cards. Most states are also providing assessment data to
districts and schools, including individual student-level data as well as
longitudinal data.
41
IV. Implementation of State Assessment Systems
A central feature of the No Child Left Behind Act is its emphasis on high
expectations for all students, schools, and districts. To support this goal, Title I
requires states (including the District of Columbia and Puerto Rico) to develop or
adopt challenging state content and achievement standards as well as
assessments that are aligned with them. By 2005-06, all states are to assess all
students in grades 3-8 and once in grades 10-12 in reading and mathematics. By
2007-08, states also must administer annual science assessments at least once in
grades 3-5, 6-9, and 10-12.
NCLB establishes a high standard for inclusion of all students in the state
assessment system, requiring that each state, district, and school ensure that at
least 95 percent of students participate in the state assessment system, both
overall and within specific subgroups, including the major racial/ethnic categories
as well as economically disadvantaged students, students with disabilities, and
limited English proficient students. State assessment systems must produce
results for individual students and be reported at the school, district, and state
levels. When reporting assessment data, states, districts, and schools must
disaggregate for each of the above subgroups as well as by gender and migrant
status.
NCLB also requires states to provide for the assessment of English language
proficiency of all limited English proficient students, beginning in 2002-03. The
assessments, aligned to state English language proficiency standards, must
include the four domains of reading, writing, speaking, and listening.
4. How fully are states meeting NCLB requirements for reporting state
assessment data?
42
standards since the passage of NCLB; since 2002, 29 states have adopted or
revised reading content standards and 28 states have adopted or revised
mathematics standards.51
Because NCLB
expanded the Title I Exhibit 27
State Approaches to Developing Assessments Required by 2005-06
state assessment
requirements to Percentage of Grade 3-8
additional grades, Assessments Across All States
states needed to Reading Math
implement additional (N=312) (N=312)
assessments in a Kept existing assessment 31% 30%
number of grades. For Modified existing assessment 5% 5%
about one-third of all of Adopted new assessment:
the required reading • Augmented existing off-the-shelf test 12% 12%
and mathematics • Developed new assessment 45% 46%
assessments in grades • Other approach 4% 4%
3 though 8, states are Data not available 3% 3%
using existing
assessments that were
Exhibit reads: In order to meet the NCLB requirements
already in place by
for state assessments in grades 3 though 8 in reading,
2004-05 (see Exhibit
states used existing assessments for 31 percent of the
27). For the remaining required assessments and modified existing assessments
required assessments in an additional 5 percent of the cases for 2004-05.
in these grades, states
are adopting new Source: Study of State Implementation of Accountability and Teacher
assessments, typically Quality Under No Child Left Behind.
by developing a new
assessment (45 to 46 percent of the necessary reading and mathematics
assessments), although in some cases states are augmenting an existing “off-the-
shelf” assessment published by a test developer (12 percent of the needed
assessments). Individual states may have used different approaches for different
grade levels and/or subjects.
43
and mathematics. In 45 of these states, at least one of these assessments was
based on alternate achievement standards. Twenty-five states also had alternate
assessments based on grade-level achievement standards.53
Under Title III requirements, states must develop ELP standards (which must be
linked to state academic content standards); states are required to provide for the
assessment of their limited English proficient (LEP) students on ELP assessments
aligned with their ELP standards.55 Half of the states (25) indicated that they had
linked their ELP assessment to ELP standards and 22 states either have not made
that linkage or have linked their ELP standards with the ELP assessment that will
be used in 2005-06.
States were asked if they used any of four approaches in developing their ELP
assessments that were administered in 2004-05, and several reported taking
more than one approach. Six states modified an out-of-state source, such as an
existing published test by a test developer, for their ELP assessment. Twenty-nine
states adopted their entire ELP assessment from an out-of-state source by, for
example, purchasing a test from a testing company. Twelve states developed
their ELP assessment as part of a multi-state consortium. Eight states developed
their own ELP assessment in-house or had it developed specifically for their state.
44
Most of the states that did not assess at least 95 percent of their black, Hispanic,
or Native American students missed the mark by a small margin, typically less
than two percent. Five states—Alabama, Connecticut, the District of Columbia,
Kentucky, and Georgia—missed by a wider margin for at least one of the
racial/ethnic groups reported.
Exhibit 28
Participation of Selected Student Subgroups in State Assessment Systems, 2003-04
Reading Math Reading Math Reading Math Reading Math Reading Math Reading Math
Number of states assessing
46 44 43 45 44 44 47 46 41 45 42 42
at least 95% of students
45
Percent of students
assessed, in states
assessing less than 95%:
Alabama 93.8% 94.5% 91.6% * * * 86.1% 87.8% ** ** 90.5% 91.6%
Arkansas * 93.4% * * ** ** 90.0% 91.0% * * ** **
California * 93.0% * * * 92.0% * 91.0% * * * *
Connecticut 92.2% 93.6% 91.7% 92.6% 93.0% * * * 88.2% 91.9% 93.2% 93.7%
Delaware * * * * 93.8% 92.7% * * 94.6% * * *
District of Columbia 92.7% 92.4% 94.8% 94.8% 90.0% 90.0% 85.1% 84.2% * * 94.4% 94.0%
Georgia 92.9% 92.8% 90.0% 90.0% 88.6% 88.5% * * 90.4% 90.5% 94.6% 94.6%
Kansas * * * * * * * * 94.9% * * *
Kentucky * * 92.0% 94.0% * * * * 83.0% 89.0% * *
Maine * * * * * * * * 91.0% * * *
Missouri * * * * * * * * 92.5% * * *
New York * * * * * * * 91.0% * 94.0% * *
Tennessee * * 93.3% * * * * * 85.2% * * *
Texas 94.1% 94.2% 93.0% 93.2% 94.6% 94.8% 77.4% 77.4% 83.3% 83.8% 93.0% 93.2%
Exhibit reads: Forty-six states assessed at least 95 percent of their black students
in reading in 2003-04.
Alabama, Connecticut, the District of Columbia, Georgia, and Texas assessed less
than 95 percent of their
black students in reading.
Notes: Six states did not report data for the participation of at least one subgroup.58 Data from the
2003-04 Consolidated State
Performance Reports have not been verified by states, and should be considered preliminary.
Most states also met the 95 percent assessment criterion for limited English
proficient (LEP) students in 2003-04. In reading, 41 states assessed at least 95
percent of their LEP students, while 45 did so for mathematics (see Exhibit 28). In
their assessment of LEP students, most states provided some sort of
accommodation, such as modifying the presentation (47 states), timing or
scheduling (46), or setting (46). The most frequent presentation accommodations
included the use of dictionaries, reading aloud the questions in English, and
reading aloud or explaining the directions. Most states gave timing or scheduling
accommodations in the form of extra assessment time. Forty-four states made
the setting accommodations of small-group or individual or separate room
administration available to LEP students. Nearly half of the states (23) made
response accommodations—such as allowing responses in the student’s native
language and writing answers directly in the test booklet—available to their LEP
students.59
46
C. Reporting Assessment Data for Use in School
Improvement Efforts
Most states are also providing data to districts and schools. Forty-three
states (of the 45 for which there were data) were providing individual student
data to school districts as of 2003-04 (see Exhibit 29). Most states were also
showing school-level and subgroup-level results over time. However, states
provided individual student data showing change over time less frequently (18
states).
Exhibit 29
Reporting of State Assessment Results to Districts or Schools
in Various Formats and by Various Groups, 2003-04
Number of States
(Out of 45 Responding)
Percentage scoring at or above proficient 45
School or district
results showing… Percentage scoring at each achievement level 43
Scale score or other similar score 41
School as a whole 45
Subgroups 45
Results for… Each grade level 42
Each classroom 29
Individual students 43
School results 37
Trends in… Subgroups within the school 34
Individual student results 18
47
Conclusions
NCLB requires states to assess annually all students in reading and mathematics
in grades 3-8, and at least once in grades 10-12, beginning with the 2005-06
school year. By 2007-08, annual science assessments must also be in place.
Most states have already administered or field-tested all the assessments needed
to meet the law’s requirements, although many of these assessments are still
subject to review by the U. S. Department of Education. NCLB also requires states
to provide for the assessment of English language proficiency of all limited English
proficient students, beginning in 2002-03; all states had some kind of ELP
assessment in place in 2004-05, but these assessments did not necessarily meet
NCLB requirements, and most states indicated that they expect to revise their ELP
assessments.
Most states are also meeting the law’s requirement to assess at least 95 percent
of their students and a similar percent of key subgroups. A few states have not
met this inclusion standard, and those states often fall short by a substantial
margin for students with disabilities or students with limited English proficiency.
NCLB has led to increased reporting on student achievement to the public as well
as to states and districts. As required, state report cards now disaggregate
achievement data by race/ethnicity, for limited English proficient students, and by
disability status and poverty. Almost all states now also publish report cards that
include assessment results by school and district. Furthermore, states are
reporting results back to districts and schools, most commonly annual assessment
results for the school, subgroups, and individual students. Trend data showing
improvement from year to year is less common, particularly trend data for
individual students.
48
Key Findings on Accountability and Support for School
Improvement
Schools in large and urban districts, and with high concentrations of poor
and minority students, were much more likely to be identified than other
schools. More diverse schools that were held accountable for more
student subgroups were more likely to be identified; for example, 37
percent of Title I schools with six or more subgroups were identified,
compared with 8 percent of those with only one subgroup.
Schools most commonly missed AYP for the achievement of all students
in reading and/or mathematics (38 percent of schools that missed AYP
based on 2003-04 testing). Smaller percentages of schools missed AYP
for only one student subgroup or test participation rates.
49
How fully have state and districts implemented other key
accountability provisions (such as unitary accountability
systems, reporting, accountability under Title III, etc.)?
50
V. Accountability and Support for School Improvement
The intent of NCLB is to promote improved achievement for all students by
requiring states to establish accountability systems that hold all schools, including
Title I schools and non-Title I schools, to the same academic standards. Under
Title I, states must assess all students and use the results to determine whether
schools and districts make adequate yearly progress. States have developed
definitions for the adequate yearly progress (AYP) expected of schools and
districts, with annual targets leading to the ultimate goal: namely, that students
from all groups—including students from low-income families and each major
racial and ethnic group, students with disabilities, and limited English proficient
(LEP) students—reach the proficient level on state assessments by 2013-14.
Required state and district report cards must present the assessment results and
other information related to school performance to parents and the public.
Schools and districts become identified for improvement when they miss AYP for
two consecutive years. Title I prescribes specific consequences for identified Title
I schools, and states may choose to apply the same consequences to non-Title I
schools. Districts must provide identified Title I schools with technical assistance
in developing or revising school improvement plans, analyzing assessment data,
identifying and implementing proven professional development and instructional
strategies, and developing budgets. Identified Title I schools also must reserve 10
percent of their Title I allocations for professional development. States, in turn,
must provide assistance to identified Title I schools through statewide systems of
support, including school support teams and distinguished educators.
When a Title I school becomes identified for improvement, the district also must
provide parents of each student at the school the option to transfer their child to a
non-identified school in the district. If the school misses AYP again after being
identified, the district must give students from low-income families the option to
receive supplemental educational services (e.g., tutoring) from state-approved
providers. If such schools miss AYP for another year after identification, districts
must take at least one of a series of “corrective actions” at the school, such as
requiring a new curriculum or replacing school staff members. If a school does
not make AYP after one year of corrective action, NCLB calls for major
restructuring of the school, beginning with a year of planning for restructuring
followed by actual restructuring the next year. Identified schools and districts exit
improvement status when they make AYP for two consecutive years.
Title III, Language Instruction for Limited English Proficient and Immigrant
Students, along with Title I, outlines additional accountability requirements
related to LEP students. Under Title III, states implement English language
proficiency (ELP) standards and assessments aligned with those standards. For
Title III subgrantees, states also define annual measurable achievement
objectives (AMAOs), which include targets for student progress in gaining and
attaining English language proficiency and AYP under Title I, as well as
consequences for subgrantees that repeatedly do not meet these targets.
51
Key Evaluation Questions for Accountability
1. What types of schools and districts are identified for improvement and thus
subject to NCLB accountability requirements?
2. What are the reasons schools do not make adequate yearly progress (AYP)?
4. How fully have states and districts implemented other key accountability
provisions (such as unitary accountability systems, reporting, accountability
under Title III, etc.)?
52
The numbers and percentages of schools identified for improvement
varied considerably across states (see Exhibit 31). States differ in the
content and rigor of their assessments and academic achievement standards as
well as other features of their accountability systems. As a result, variation across
states in the numbers and percentages of identified schools likely reflects
differences in state accountability systems as well as differences in student
achievement; states with more identified schools are not necessarily lower
performing than states with fewer identified schools. Seven states had identified
5 percent or fewer of their Title I schools, while six states had identified more than
one-third their Title I schools in 2004-05. Similarly, the numbers of Title I schools
in corrective action or restructuring status varied by state, from none in several
states to more than 100 in a few states.65
Exhibit 30
Number and Percentage of Identified Title I Schools,
1996-97 to 2004-05
10,000
9,028
8,000 8,408 8,348 8,375 18%
7,353 20% 19% 18%
7,021
6,000 17% 6,441
16% 6,094 5,963
13%
12% 12%
4,000
2,000
0
1996- 1997- 1998- 1999- 2000- 2001- 2002- 2003- 2004-
97 98 99 00 01 02 03 04 05
Exhibit reads: In 2004-05, 9,028 Title I schools had been identified for
improvement based on test scores for 2003-04 and earlier years; these identified
schools represented 18 percent of all Title I schools in that year.
Note: The first year that schools were identified for improvement based in part on NCLB AYP
definitions was 2003-04, based on assessments administered in 2002-03. However, schools are
identified when they miss AYP for two consecutive years, and 2004-05 was the first year that includes
schools identified because they missed NCLB AYP targets for two consecutive years.
53
2003-04 testing, and these states identified 2,503 non-Title I schools for 2004-05.
Few states required NCLB consequences of public school choice and supplemental
services for identified non-Title I schools (three states each). Similarly, only eight
states had assigned non-Title I schools to corrective action status, and only eight
states had put non-Title I schools in restructuring status in 2004-05. Overall,
states had placed fewer than 300 non-Title I schools in corrective action or
restructuring.66
Most districts with identified schools have very few, though a smaller number of
districts had large numbers of identified schools. Of the 2,912 districts that had
one or more identified schools in 2004-05, nearly three-fourths (73 percent) of
these districts had only one or two identified schools. However, 4 percent of
districts with identified schools (129 districts) contained 13 or more identified
schools each.67
54
Exhibit 31
Number and Percentage of Identified Schools, by State, 2004-05*
Note: This table shows data reported by 51 states from October 2004 to April 2005. Some states decided
appeals prior to this data collection, and others made appeal decisions later; for example, Texas later
approved more than 100 appeals, resulting in a final count of 91 identified schools. This chapter uses the
numbers that states reported for this data collection.
Source: Study of State Implementation of Accountability and Teacher Quality Under NCLB.
55
Schools with high
concentrations of poor Exhibit 32
Percentage of Identified Schools with
and minority students Selected Characteristics, 2004-05
were much more likely
100%
to be identified than
other schools. Just over 80%
56
they had identified districts, 1,511 districts had been identified for improvement
in 2004-05. Nineteen states had identified 10 percent or fewer of their districts,
and 12 states had identified a third or more of their districts. Among the
identified districts, 49 districts in 11 states were identified for corrective action.
Twenty-six percent of all students, or about 12.6 million students, attend schools
in identified districts across 48 states with available data. 72
Large and urban districts with high concentrations of poor, minority, and
LEP students were more likely to be identified than other districts.
District size mattered most, with one-third of large districts identified in 2004-05,
compared with 17 percent of medium districts and 5 percent of small districts.
Nineteen percent of districts with high concentrations of minority students were
identified, compared with 5 percent of districts with low concentrations of
minority students; distributions were similar for other district characteristics, with
greater likelihood of identification for districts that were more urban and served
high concentrations of students from low-income families and LEP students.73
57
Exhibit 33
Number and Percentage of Identified Districts, by State, 2004-05*
Note: This table shows data reported by 51 states from October 2004 to April 2005. Some states
decided appeals prior to this data collection, and others made appeal decisions later; for example,
California later increased its number of identified districts to 58. This chapter uses the numbers that
states reported for this data collection.
Source: Study of State Implementation of Accountability and Teacher Quality Under NCLB.
58
B. Adequate Yearly Progress Ratings for Schools and
Districts
In determining whether a school or district makes adequate yearly progress (AYP),
NCLB requires states to consider state assessment results, student participation
rates in assessments, and an “other academic indicator.” For state assessment
results, states must set absolute annual targets that lead to the goal of all
students achieving proficiency in reading and mathematics by 2013-14. For test
participation, schools and districts must assess at least 95 percent of their
students in order to make AYP. For the other academic indicator, states must use
graduation rates for high schools, but they have flexibility in selecting a measure
for elementary and middle schools.
Calculating AYP separately for key subgroups of students is a key feature of the
NCLB accountability system. AYP must be calculated for up to nine student
groups in a school or district: all students, five major racial and ethnic groups,
economically disadvantaged students, students with disabilities, and LEP
students. To enhance validity and reliability, states also define a minimum
subgroup size that must be met before AYP is calculated for that subgroup for a
school or district. The number of AYP targets a school must meet will vary by
state definitions of AYP, enrollment size, and the demographic composition of the
school. Schools that serve diverse populations often must meet more AYP targets
than those whose enrollments are homogeneous.
NCLB includes a “safe harbor” provision: Schools may make AYP if the percentage
of students in a subgroup that did not meet the AYP target decreases by 10
percent from the preceding school year, and if the school makes AYP for the
relevant subgroup for the other academic indicator and participation rate. In
addition, regulations allow states and districts to count as proficient scores of
students with disabilities who take alternate assessments based on alternate
achievement standards, as long as the number of those proficient scores does not
exceed one percent of all students in the grades assessed.7
States first applied NCLB AYP definitions to state assessment results from 2002-
03; these determinations first affected schools that were identified for
improvement for the following year, 2003-04. Findings below include all schools
(both Title I and non-Title I schools).
7
The U.S. Department of Education has provided additional flexibility concerning AYP determinations for
LEP students and students with disabilities. Flexibility for LEP students, which became effective in
February 2004, includes permitting LEP students in their first year of enrollment in U.S. schools to take
an English language proficiency assessment instead of the state reading assessment; permitting states
to exclude those students’ reading and mathematics scores from AYP calculations; and permitting states
to retain formerly LEP students in the LEP subgroup for AYP calculations for up to two years after they
attain English proficiency. New flexibility for students with disabilities, announced in May 2005, will allow
eligible states to make adjustments to their AYP decisions based on 2004-05 assessment results to
reflect the need for modified achievement standards for students with disabilities who may not be able
to reach grade-level standards in the same timeframe as other students, while the Department proposes
new regulations to address this matter.
59
1. Schools and Districts Making Adequate Yearly Progress
Schools most commonly missed AYP for the achievement of all students
and/or multiple subgroups in 2003-04; only in a minority of cases did
schools miss just
one AYP target. Exhibit 34
Failure to meet AYP for Reasons Schools Missed AYP, 2003-04
the all students group
or for multiple
subgroups suggests
systemic low
performance of a
school, especially
when it misses AYP in
multiple subjects.
Overall, by subject area, 64 percent of schools that missed AYP missed for a
reading achievement target and 58 percent missed for a target in mathematics,
either for all students or a subgroup, while 42 percent missed AYP in both
subjects.78
61
A majority of students from most racial and ethnic groups and from low-
income families attended schools held accountable for the performance
of their subgroups. Across 34 states with available data, nearly 80 percent or
more of all white, African-American, and Hispanic students, as well as students
from low-income families, attended schools where AYP was calculated for these
subgroups based on 2003-04 test results. Percentages were smaller for Native
American and Asian students (25 percent and 45 percent, respectively).79
The numbers of
subgroups for AYP Exhibit 35
were calculated and Percentage of Schools Held Accountable for Subgroup That
the numbers of Missed AYP for Achievement for That Subgroup, 2003-04
grades tested both
appear related to
the likelihood of a Low-Income Students 14%
Schools that were held accountable for more subgroups were less likely
to make AYP. Among schools for which AYP was calculated for six or more
subgroups, 39 percent did not make AYP, compared with 10 percent of schools for
which AYP was calculated based on only one subgroup. In high-poverty schools
that had six or more subgroups, 66 percent missed AYP, compared with 35
62
percent of low-poverty schools that had six or more subgroups. Similarly, the
numbers of grades tested also appeared related to the likelihood of a school
making AYP. In states where the scores of students in grade 3 through grade 8
and at least one high school grade were used to determine AYP for 2003-04,
schools were less likely to make AYP (71 percent) than were schools in states that
used test results from fewer grades (82 percent).81
63
3. Other Academic Indicators in Adequate Yearly Progress
Twenty percent of all schools that did not make AYP missed for the other
academic indicator in 2003-04, and one-third of all high schools missed
AYP for their other academic indicator, graduation rate. States varied
considerably in the percentage of high schools that missed AYP for graduation
rate targets, from zero to 82 percent of high schools across states. Elementary
and middle schools less frequently missed AYP for this reason (10 percent). In
many states (22), fewer than 10 percent of elementary or middle schools missed
AYP for the other academic indicator, though percentages in other states ranged
from 11 percent to 64 percent.83
For limited numbers of students with disabilities, regular assessments, even with
accommodations, are not appropriate. Title I regulations allow for students with
the most significant cognitive disabilities to be assessed on alternate assessments
based on alternate achievement standards; states and districts may count as
proficient for AYP the scores of such students as long so those proficient scores
did not exceed 1.0 percent of all students tested. The regulations also allow
states and districts to receive exceptions to exceed the 1.0 percent cap.
64
approvals ranging from zero to 100 percent. At least four states approved more
than 100 appeals for schools. Thirteen of 32 states also said that one or more
districts had appealed their AYP designations based on 2003-04 testing, and these
states approved 50 percent of the 236 district appeals. Most appeals involved
either errors in data or the misclassification of students by subgroup.85
All states notified schools about their identification status for 2004-05
based on 2003-04 testing, and a majority of states provided preliminary
results before September 2004, but 20 states did not, and only 15 states
provided final results by that time. In order to be able to plan for the
upcoming school year, including school improvement planning and notifying
parents about choice options for children in identified schools, school districts and
schools need to receive notification of their school improvement status prior to
the beginning of the school year. This is also a requirement under Title I
regulations. To enable such planning at the local level, many states first notified
schools whether they had been identified for improvement based on preliminary
data, then followed up with final data later. Thirty-one states reported releasing
preliminary data on whether schools were identified for improvement based on
2003-04 testing before September 2004, though six did not release preliminary
data until November 2004 or later. Fifteen states released final data on school
identification status prior to September 2004, whereas 21 states did not notify
schools of their final designation for 2004-05 until November 2004 or later.86
Correspondingly, among principals who indicated their school was identified, 57
percent reported learning of their improvement status before September 2004,
with 35 percent of identified schools being notified in August. Only 9 percent of
identified schools reported they did not learn of their identification status until
November 2004 or later.87
65
2. Report Cards
State report cards released in 2004-05 (presenting data from 2003-04)
generally included the percentages of all students and student
subgroups scoring at proficient levels in reading and mathematics, as
well as the percentage tested, but did not universally include other
required elements. Only for the migrant student subgroup did states less
commonly report the percentage of students scoring at proficient levels (37
states). About two-thirds of state report cards included data on other academic
indicators used for AYP, and under half of state report cards contained
comparisons of student achievement to AYP targets, graduation rates
disaggregated for student subgroups, and required data on the professional
qualifications of teachers.89
About a third of Title I districts did not prepare district report cards in
2003-04. District report cards commonly included assessment data along with
attendance and graduation rates, but half or fewer of district report cards
included data on teacher quality, a new requirement under NCLB, and on the
number and percentage of district Title I schools identified for improvement.90
Though not required under NCLB, most districts also reported preparing
school report cards. Eighty-seven percent of Title I districts reported preparing
school report cards in 2003-04, up from 81 percent in 2002-03, but down from 93
percent in 2001-02. School report cards for 2003-04 more commonly included
data on whether a school was identified for improvement than they had in 2002-
03 (69 percent vs. 35 percent of Title I districts with school report cards).91
66
statewide systems of support. While these individuals may serve the roles NCLB
outlines for distinguished principals and distinguished teachers, often more
general terms defined their work (e.g., school improvement specialists, principal
mentor, coaches). School support teams and specialized individuals in many of
these states serve non-Title I identified schools as well as Title I identified schools
(19 states and 17 states, respectively). Twenty states incorporated a triage
approach in which the level of support they provided to identified schools was
attuned to the severity of individual schools’ needs.92
Though states often used multiple strategies to improve identified schools, most
states focused their efforts on one of five primary support strategies: school
support teams (19 states); specialized individuals (13 states); regional centers,
area educational agencies, or county offices (9 states); providing resources or
hosting statewide meetings (6 states); or depending on districts to provide
support (3 states).93
67
Identified schools, however, reported receiving more days of assistance from their
districts than non-identified schools. For 2004-05 and the previous school year,
75 percent of identified schools reported six or more days of assistance from their
districts, compared with 56 percent of non-identified schools. Forty-eight percent
of identified schools received at least 11 days of assistance, and 25 percent
received more than 25 days of assistance.98
68
non-identified schools. Similarly, 74 percent of identified schools reported
needing assistance to get parents more engaged in their child’s education,
compared with 46 percent of non-identified schools.103 Across all types of
assistance shown in Exhibit 36, about half or less of non-identified schools
reported needing assistance.
69
Exhibit 36
Percentage of Non-Identified and Identified Schools That Reported Needing Various Types
of Technical Assistance and Whether Identified Schools Received Assistance, 2003-04 to 2004-05
70
2. School Improvement Strategies
Exhibit 37
Percentage of Schools Reporting Major Focus
on Various School Improvement Strategies, 2004-05
Identified Non-Identified
Schools Schools
(n=430) (n=881)
Using student achievement data to inform instruction and school 82%* 67%
improvement
Providing additional instruction to low-achieving students 78%* 60%
Aligning curriculum and instruction with standards and/or assessments 72% 70%
Implementing new instructional approaches or curricula in 61%* 49%
reading/language arts/English
Increasing the intensity, focus, and effectiveness of professional 60%* 42%
development
Implementing new instructional approaches or curricula in mathematics 59%* 41%
Restructuring the school day to teach core content areas in greater depth 52%* 31%
Providing extended-time instructional programs 51%* 31%
Implementing strategies for increasing parents’ involvement in their 32%* 13%
children’s education
Increasing instructional time for all students 26%* 13%
71
Teacher reports also indicate widespread use of state assessment
results to support instruction and school improvement. For example,
about 70 percent of reading and mathematics teachers reported using 2003-04
state assessment data moderately or extensively for identifying and correcting
gaps in the curriculum and for identifying areas where they needed to strengthen
their knowledge or skills. In 2004-05, teachers in identified schools were more
likely than teachers in non-identified schools to report using the previous year’s
state assessment data in several ways. For example, among teachers who teach
mathematics at either the elementary or secondary level, 74 percent of those in
identified schools reported using the state assessment data to identify students
who need remedial assistance, compared with 60 percent in non-identified
schools. Similarly, mathematics teachers in identified schools were more likely to
report using the information to tailor instruction to individual student needs (75
percent vs. 61 percent) and to recommend tutoring or other educational services
for students (60 percent vs. 45 percent). Similar patterns were reported for these
uses of state reading assessments.105
Exhibit 38
Change in Instructional Time Per Day at Elementary Schools,
by Subject Area, 2003-04 to 2004-05
72
Exhibit reads: Thirty percent of identified schools reported that instructional
time spent per day on reading in their schools increased more than 30 minutes
from 2003-04 to 2004-05, compared with 13 percent of non-identified schools.
73
At the secondary school level, about two-fifths of all schools reported increasing
instructional time for low-achieving students in reading (40 percent) and
mathematics (42 percent), with identified secondary schools reporting increasing
time for reading at significantly higher rates than non-identified schools
(55 percent vs. 36 percent). Smaller percentages of secondary schools reported
increasing instructional time in science (13 percent) and social studies
(11 percent) for low-achieving students. Few elementary or secondary schools
reported decreasing instructional time in any subject.107
Some identified schools have not followed NCLB requirements for school
improvement planning and professional development for identified
schools. Since they were first identified, only 82 percent of identified Title I
schools in 2004-05 had developed a joint school improvement plan with their
district or state, despite the requirement that all identified Title I schools do so.108
Similarly, only 89 percent of districts required identified Title I schools to spend at
least 10 percent of their Title I allocation on professional development in 2003-04,
though this represents an increase over the 79 percent of districts that
implemented this requirement in 2002-03.109
States placed 1,047 Title I schools in corrective action and 1,065 Title I schools in
restructuring for 2004-05, making these schools subject to the particular menus
of interventions outlined in NCLB.
Many of the interventions that NCLB defines as corrective actions were also
implemented in schools in earlier stages of identification for improvement. For
example, 66 percent of schools in their second year of improvement were
required to implement new research-based curricula or instructional programs.111
74
restructuring status reported state take-over of the school (7 percent), re-opening
of the school as a public charter school (2 percent), contracting with a private
entity to manage the school (2 percent), or replacement of all of the school staff
(2 percent).112 About one-fourth (24 percent) of schools in restructuring status
reported that a new principal had been appointed, although this may partly reflect
normal principal turnover, as similar percentages of schools in other stages of
improvement status also reported this.8
8
The NLS-NCLB survey question did not exactly parallel the law on one intervention: the law gives the
option of “replacing all or most of the school staff (which may include the principal) who are relevant to
the failure to make adequate yearly progress,” while the survey asked if the state or district had
“replaced all of the school staff” or “appointed a new principal.”
75
Exhibit 39
Percentage of Identified Title I Schools Experiencing Various Types of Interventions
Since Identification for Improvement, 2004-05
Action Required for Identified Schools That Miss AYP After Identification
Eligible students were offered supplemental 46% 90% 94% 100%
educational services from a state-approved
provider
Corrective Actions
Implemented a new research-based curriculum or 48% 66% 90% 70%
instructional program
Significantly decreased management authority at 4% 5% 27% 26%
the school level
Appointed outside expert to advise the school 30% 34% 58% 57%
Extended length of school day 24% 29% 42% 31%
Extended length of school year 9% 15% 32% 16%
Restructured internal organization of the school 12% 22% 21% 33%
Replaced school staff members relevant to 2% 17% 5% 18%
school’s low performance
Restructuring Interventions
Reopened the school as a public charter school 0% 0% 0% 2%
Entered into a contract with a private entity to 0% 1% 0% 2%
manage the school
Operation of school turned over to state 2% 0% 1% 7%
Replaced all of the school staff 0% 1% 0% 2%
Appointed new principal 21% 20% 18% 24%
76
Schools in restructuring status frequently reported experiencing actions
that NCLB specifies for the “corrective action” stage of school
improvement, such as implementing a new research-based curriculum or
instructional program (70 percent), appointment of an outside expert to advise
the school (57 percent), restructuring the internal organization of the school (33
percent), extending the length of the school day (31 percent), significantly
decreasing management authority at the school level (26 percent), and replacing
school staff members relevant to school’s low performance (18 percent).
77
Exhibit 40
Percentage of Districts Taking Various Actions in Response to
Being Identified for Improvement, 2004-05
Exhibit reads: In 2004-05, 80 percent of identified districts reported that they had
offered or required specific professional development for teachers in response to
being identified for improvement.
78
locally determined tests, or tests in subjects other than reading and
mathematics), and two states used different rules for how to include students.
Additionally, some of these other state initiatives used different designations of
school performance (such as using letter grades or identifying “high-improving”
schools) or reported the results of the state initiatives separately from reporting
for NCLB.117
79
targets based on 2003-04 testing (1,898 subgrantees). States varied
considerably in the proportion of their subgrantees that met AMAO targets; nine
states had 25 percent or fewer of their subgrantees make AMAO targets, while 13
states had more than 75 percent of subgrantees meet targets. At the state level,
33 states (out of 42 responding) reported meeting their AMAO targets for students
making progress in learning English, while 41 out of 45 met some or all of their
AMAO targets for students’ attainment of English language proficiency.9 120
About two-thirds of the states (35 states) calculated AMAOs for 2003-04
testing for Title III districts only, while another 13 states reported
calculating AMAO performance data for all districts with LEP students.
Most of these states also reported the results to their districts (34 and 11 states,
respectively).121 Though NCLB requires that states assess the English language
proficiency of all LEP students, Title III only requires states to calculate whether
Title III districts meet the AMAOs and report the results to these districts, but
states also may choose to calculate and report to non-Title III districts whether
they met AMAOs.
Fewer than half of the states had articulated a specific strategy for
providing technical assistance to subgrantees that missed AMAO targets
in 2003-04 testing. Such assistance was not required in 2004-05; instead, it
must be provided beginning in 2005-06 for subgrantees that miss AMAO targets
for a second consecutive year. Twelve states reported they planned to conduct
needs assessments among subgrantees that miss AMAO targets twice, and 10
states reported they would provide technical assistance through existing technical
assistance frameworks in the state.122
Conclusions
Under NCLB, state accountability systems are intended to drive substantial
improvement in student achievement. In the latest school year for which data is
available, 2004-05, 11,530 schools were identified for improvement (13 percent of
all schools); nearly 80 percent were Title I schools. The proportion of schools
identified varied widely across states, but to a large extent this reflects
differences in state accountability systems and not necessarily differences in
school performance. Schools serving large numbers of poor, minority and LEP
students were most likely to be identified. African-American students and
Hispanic students were three times more likely to attend identified schools than
were white students. Although schools can miss making adequate yearly
progress because of the performance of a single subgroup, schools most
commonly missed AYP for the achievement of all students and/or multiple
subgroups.
9
For more information on student achievement on English language proficiency assessments and states’
AMAOs, see the Biennial Evaluation Report to Congress on the Implementation of the State Formula
Grant Program, 2002-2004, English Language Acquisition, Language Enhancement and Academic
Achievement Act (ESEA, Title III, Part A).
80
States, districts, and schools generally reported pursuing improvement efforts for
identified schools and districts consistent with NCLB requirements, and identified
schools reported receiving more hours of assistance than did non-identified
schools. However, states and districts indicated limited capacity to assist all
identified schools.
81
Key Findings on School Choice and Supplemental Educational
Services
The timing of parental notification was often too late to enable parents to
choose a new school before the start of the 2004-05 school year. Almost
half of districts notified parents after the school year had already started,
and in these districts this notification occurred, on average, five weeks
after the start of the school year.
States report that they are working to develop and implement systems
for monitoring and evaluating the performance of supplemental service
providers, but as of early 2005, 15 states had not established any
monitoring processes and 21 states had not finalized their monitoring
processes; 25 states had not yet established any standards for evaluating
provider effectiveness and none had finalized their evaluation standards.
82
The most common approaches that states have implemented to monitor
providers are surveying the districts about provider effectiveness (25
states) and using providers’ reports on student progress (18 states). The
most common standard that states have adopted to evaluate the
effectiveness of providers is student achievement on state assessments,
although only one state plans to use a matched control group.
83
VI. School Choice and Supplemental Educational Services
In Title I schools that have been identified as in need of improvement, NCLB
provides parents with new options for their children, including the option to
transfer to another public school or to receive supplemental educational services
(most commonly, after-school tutoring). The choice and supplemental services
provisions are designed not only to improve educational opportunities for
individual students, but also to provide an incentive for low-performing schools to
improve.
Districts are required to offer students the option to transfer to another school in
the first year that a school is identified for improvement; all students in the school
are eligible for this option, and the district must provide transportation for
participating students. Supplemental educational services are not required until
an identified school misses AYP again (for a third time), and only low-income
students in these schools are eligible to receive the services; the district is not
required to provide transportation.
States must develop criteria for approving supplemental service providers and
must provide school districts with a list of available approved providers in their
area. States also have the responsibility for monitoring the performance of
participating providers.
Districts must notify parents of their school choice and supplemental service
options and disseminate information about school performance and provider
qualifications and effectiveness that parents need to make informed decisions.
Each district that must offer these options must allocate an amount equal to
20 percent of its Title I Part A allocation to provide supplemental services and
transportation for students using the school choice option, unless a lesser amount
is needed to satisfy all requests. In addition, each such district must make
available, for each child receiving supplemental services, an amount equal to the
district’s Title I Part A allocation per low-income student, unless the actual cost of
such services is less than that amount.
1. How many students are eligible to participate, and how many actually do so?
2. How and when do districts and schools inform parents of eligible children
about the Title I school choice and supplemental services options?
84
A. Eligibility and Participation
Although more students were eligible to participate in the Title I school
choice option, a larger number actually participated in the supplemental
educational services option. Nearly three times as many students were
eligible to transfer to another school under the Title I choice option in 2003-04
(3.9 million) as were eligible to receive supplemental services (1.4 million). More
students are eligible for the choice option because it applies to all identified
schools and all students in those schools are eligible, whereas the supplemental
services option only applies to identified schools that have missed AYP for a third
year and only low-income students in those schools are eligible. Nevertheless, six
times as many students actually participated in the supplemental services option
(233,000) as participated in the school choice option (38,000) in that year (see
Exhibit 41).123
Exhibit 41
Student Eligibility and Participation for
Title I School Choice and Supplemental Services, 2003-04
Exhibit reads: More than 3.9 million students were eligible for Title I school choice
in 2003-04, while 1.4 million were eligible for supplemental services.
Sources: Study of State Implementation of Accountability and Teacher Quality Under NCLB; National
Longitudinal Study of NCLB, District and Principal Surveys. Estimates of eligible students are based on
data reported by 51 states. Estimates of participating students areExhibit
based42on an n of 109 responding
Number of
schools for school choice and 92 districts for supplemental services.Students Participating in
Title I School Choice and Supplemental Services
300,000
Student participation
in the supplemental
services option Exhibit reads: The number of students
participating in Title I school choice rose from
increased more than
18,000 in 2002-03 to 45,000 in 2004-05.
five-fold over the two-
year period from Sources: Study of Title I Accountability Systems and School
Improvement Efforts, District Survey (2002-03); National
Longitudinal Study of NCLB, District and Principal Surveys (2003-
04 and 2004-05). School choice estimates are based on an n of
852002-03, 109 schools in 2003-04, and 121 schools
247 districts in
in 2004-05. Supplemental services estimates are based on an n
of 90 districts in 2002-03 and 92 districts in 2003-04.
2002-03 to 2003-04, rising from 42,000 to 233,000 participants. Data on
supplemental services participants is available only through 2003-04 because the
NLS-NCLB survey was administered in fall 2004 and the total number of
supplemental services participants is usually not known until later in the school
year (because students may begin supplemental services as late as the spring,
whereas school choice transfers typically occur before or near the start of the
school year).
The number of
Exhibit 43
schools where
Number of Schools Where Title I School Choice
supplemental and Supplemental Services Were Offered
services were
offered tripled from 8,000
2002-03 2003-04 2004-05
2002-03 to 2003-04, 6,200
while the number 6,000
where Title I school 5,100
4,600
choice was offered
increased from 4,000
5,100 in 2002-03 to 2,500
6,200 in 2004-05 2,000
(see Exhibit 43). Title I 800
school choice was
offered in about 6,200 0
School Choice Supplemental Services
schools and 1,800
districts in 2004-05,
and supplemental
services were offered Exhibit reads: The number of schools where
in 2,500 schools and supplemental services were offered rose to 2,500 in
500 districts in 2003- 2003-04, while the number of schools where choice
04. Most districts was offered grew to 6,200 in 2004-05.
required to offer
supplemental services Sources: Study of Title I Accountability Systems and School
Improvement Efforts, District Survey (2002-03 and 2003-04);
reported that they did National Longitudinal Study of NCLB, Principal Survey (2004-05).
offer such services (89 School choice estimates are based on an n of 314 districts in
percent in 2003-04). 2002-03, 327 districts in 2003-04, and 308 schools in 2004-05.
Supplemental services estimates are based on an n of 71
districts in 2002-03 and 206 districts in 2003-04.
More than one-third
(39 percent) of
districts required to offer the school choice option in 2004-05 did not do
so, but often such districts had no non-identified schools in the district
to which students could transfer. Among districts that were required to offer
school choice in 2004-05, 20 percent reported that having no non-identified
86
schools within the district, either because there was only one school per grade
level or because all schools in the district were identified for improvement, was a
major challenge to implementing Title I school choice. Some districts pointed to a
lack of space in non-identified schools (25 percent) or an inability to negotiate
agreements with other districts to receive students who wished to transfer
(16 percent) as major challenges.124
approved providers
(25 percent) might
suggest.
Exhibit reads: Private providers accounted for 70
percent of state-approved providers in May 2004
Nevertheless, private
and 59 percent of participating students during the
organizations served a 2003-04 school year.
majority of
participating students Sources: Policy and Program Studies Service review of State
Education Agency Web sites, May 2004; National Longitudinal
Study of NCLB, Principal Survey (2003-04). Percentages of
providers are based on data reported by 51 states. Percentages
of participating students
87 are based on an n of 71 districts.
(59 percent) in 2003-04; about one-third of participants were served by national
for-profit companies (34 percent); while 12 percent were served by other for-profit
companies and 13 percent by community-based organizations. Colleges and
universities accounted for a small proportion of approved providers (2 percent)
and an even smaller share of participants (less than one percent). Charter
schools also served less than one percent of participants.
B. Parental Notification
Districts that were required to offer Title I school choice and
supplemental services in identified schools most frequently reported
notifying parents about their choice options through written notification
materials (60 percent and 88 percent, respectively), but districts also
reported using other strategies to communicate with parents (see Exhibit
45). The most common approach was to hold special meetings to inform parents
about their school choice (37 percent) and supplemental services options (72
percent). Notices in district or school newsletters, enrollment fairs or open
houses, and notices in public newspapers were other approaches used. The
percentage of students in districts using each notification strategy was always
higher, sometimes considerably higher, than the percentage of districts would
suggest, indicating that large districts were more likely to use each type of
notification strategy. For example, districts providing written notification enrolled
80 percent of the students in districts required to provide the school choice
option, although they accounted for only 60 percent of such districts.
Exhibit 45
District Strategies for Communicating with Parents
About Title I School Choice and Supplemental Services Options, 2004-05 130
88
Notices in public newspapers 18% 39% 20% 42%
Public service announcements 7% 29% 17% 37%
Outreach through a local community partner
(e.g., Parent Information & Resource Center) 7% 18% 17% 38%
Other 19% 21% 24% 35%
Exhibit reads: Districts that were required to offer Title I school choice most
frequently reported notifying parents about their choice options through written
materials (60 percent). Districts providing written notification about the school
choice option enrolled 80 percent of all students in districts required to offer this
choice option.
However, the timing of parental notification was often too late to enable
parents to choose a new school before the start of the school year. Only
29 percent of affected districts notified parents about the school choice option
before the beginning of the 2004-05 school year. Another 21 percent notified
parents at the beginning of the school year, which would have given parents very
little time to make important decisions about which school their child should
attend. The remaining 50 percent of districts notified parents after the school
year had already started; in these districts, notification occurred, on average,
five weeks after the start of the school year.131 Districts that notified parents
before the start of the school year accounted for 52 percent of the students in
districts offering Title I school choice.
One reason for the delay in notifying parents about their choice options may be
that some states did not provide final determinations about schools’ AYP and
identification status until late in the summer or, in some cases, after the school
year had begun (see Chapter V).
89
In interviews with a small sample of parents10 in schools where supplemental
services were being offered in 2003-04, about half of the parents interviewed said
they had received enough information to choose good providers for their children,
while nearly as many reported that they knew little or were confused about the
services available to them. The parents often indicated that teacher and principal
recommendations were important factors in their decision whether to enroll their
child in supplemental services and in choosing a provider. Parents interviewed
also said that location and the availability of transportation were critical issues in
selecting a provider.132
As of early 2005, half of the states had not yet established any standards for
evaluating provider effectiveness and none had finalized their evaluation
standards.
90
children’s grades had improved, while others pointed to improved mathematics or
reading skills. However, some parents reported that they were disappointed with
the services and saw no improvement in their children’s reading or mathematics
skills.135
91
Conclusions
NCLB requires Title I schools that have been identified for improvement to offer
options for parents to transfer their children to another public school or to obtain
supplemental educational services, most typically after-school tutoring. Although
many more students are eligible to use the school choice option, the early
experience with these provisions indicates that after-school tutoring is by far the
more popular option. In the 2003-04 school year, six times as many students
participated in the supplemental services option (233,000) as participated in the
school choice option (38,000). Stated differently, only one percent of eligible
students changed schools under the NCLB provision, and 17 percent of eligible
students enrolled to receive supplemental services.
Future reports will examine additional issues related to choice and supplemental
services—what factors parents consider when making decisions about whether to
choose and what to choose, the characteristics of participating students, and
student achievement outcomes associated with participation in the Title I school
choice or supplemental services options.
92
Key Findings on Teacher Quality and Professional Development
In all but two states, teachers may take exams to demonstrate their
subject-matter competency for the purposes of meeting the NCLB highly
qualified teacher requirement, frequently one of the Praxis II subject
assessments (41 states). States vary considerably in the scores that they
require teachers to obtain in order to be certified to teach and/or to be
deemed highly qualified under NCLB.
The large majority of teachers across the country have been designated
by their states as “highly qualified” under NCLB. According to state-
reported data for 42 states, 86 percent of classes were taught by highly
qualified teachers in 2003-04. Principal and teacher reports for 2004-05
provide somewhat lower estimates of the percentage of teachers who are
highly qualified. For example, 74 percent of all teachers reported that
they were considered highly qualified under NCLB, although 23 percent
responded that they did not know if they were considered highly
qualified.
93
the remaining aides (26 percent), principals reported that they did not
know their qualifications status. However, 87 percent of Title I
instructional aides indicated that they had at least two years of college
(and/or an associate’s degree) or had passed a paraprofessional
assessment.
94
VII. Teacher Quality and Professional Development
Although there has not been an extensive examination of the relationship
between instructors’ content knowledge and students’ achievement, some
research suggests that teachers who have strong preparation in the subjects they
teach are more effective than teachers without strong subject-area preparation.136
Because many policymakers have been concerned that some teachers graduate
from their teacher preparation programs without adequate subject-matter
preparation and that other teachers are assigned to teach subjects for which they
have not been certified to teach, NCLB requires all teachers of core academic
subjects to be “highly qualified” by the end of the 2005-06 school year. NCLB
specifies the core academic subjects to be English, reading or language arts,
mathematics, science, foreign languages, civics and government, economics,
arts, history, and geography.137
To be “highly qualified” each teacher must have: (1) a bachelor’s degree; (2) full
state certification; and
(3) demonstrated competency, as defined by the state, in each core academic
subject he or she teaches. The law requires new elementary teachers to
demonstrate subject-matter competency by passing a rigorous state test; new
secondary teachers must either pass a subject-matter test or have a college
major (or coursework equivalent), advanced degree, or advanced certification in
the subject(s) they plan to teach. For veteran teachers, the law allows each state
to create its own “high, objective, uniform state standard of evaluation” (HOUSSE)
to measure subject-matter competency.
NCLB also increased the minimum qualification requirements for Title I-funded
paraprofessionals who provide instructional services. Specifically, NCLB requires
that aides providing instructional services must have at least two years of college
or an associate’s degree, or they must meet a rigorous standard of quality
through a formal state or local assessment. All new Title I instructional aides
must be qualified upon hire, and all existing Title I instructional aides must
become qualified by the end of the 2005-06 school year.
95
2. How many elementary teachers meet the NCLB requirement to be “highly
qualified” as defined by their state? What percentage of secondary classes
in core academic subjects are taught by teachers who are highly qualified?
How does this vary across states, grade levels, and school poverty levels?
Most states meet the requirement to test the content knowledge of new
teachers through the Praxis II subject assessments developed by the
Educational Testing Service (ETS).139 Based on an analysis of state Web sites
and the ETS Web site in early 2005, of the 41 states that use one or more of the
various Praxis II examinations, 25 use the Praxis II exams alone and 16 list the
Praxis II exams as well as other exams. Ten states do not list the Praxis II exams
but list other exams, such as tests developed for use in specific states (e.g., the
Massachusetts Test for Educator Licensure).140
The two states (Iowa and Montana) that do not identify any exams that
elementary or secondary teachers can take in order to demonstrate subject-
matter competency also do not require teachers to pass any subject-matter
exams in order to become initially certified to teach.141
States vary considerably in the qualifying scores that they use on Praxis
II subject assessments for the purposes of initial teacher certification
and for determining whether teachers are “highly qualified” under NCLB
(see Exhibit 46). States set different qualifying scores (often called “cut scores”
or “passing scores”) for reasons involving each state’s individual context and
challenges; each state assembles a panel of experts that reviews the test and
recommends a cut score to the state licensing board or state department of
96
education.142 Twenty-nine of the 35 states that use the Praxis II Mathematics
Content Knowledge exam set their cut scores below the national median and nine
states set theirs below the 25th percentile (ranging from the 14th to the 22nd
percentile).143 In contrast, four states set the cut score above the national
median, and one of those four states set its cut score at the 74th percentile.
As far as could be determined from extant sources (state Web sites and the ETS
Web site), states are, with the exception of Alabama, using the same cut scores
for both highly qualified determinations and initial teacher certification
requirements. (The scores listed for Alabama are used for the highly qualified
designation only; the state is in the process of determining the cut scores for
initial certification.) Note that this analysis did not distinguish between the use of
exams for teachers at different grade levels; in particular, states may vary in
whether middle school teachers take a general elementary examination or a
specific subject-matter examination.144
Nearly all states (47) officially offered a HOUSSE option to their veteran
teachers as of spring 2005 (see Exhibit 47).145 The five states that did not
offer a HOUSSE option at that time were Colorado, the District of Columbia,
Mississippi, Missouri, and Puerto Rico.
97
98
Exhibit 46
State Definitions of “Highly Qualified Teacher”: Use of Praxis II Exams and Cut Scores, 2004-05
State Uses At Least Praxis II: Elementary Praxis II: English Praxis II: Mathematics
One Exam from Education Content Language, Literature Content Knowledge
Praxis II Series for Knowledge and Composition:
Some/All Teachers Content Knowledge
Total Number of
States Using 41 20 36 36
Praxis II: Subject
Assessments
Alabama X 137 151 118
Alaska X 143 158 146
Arkansas X 159 116
California X
Colorado X 147 162 156
Connecticut X 172 137
Delaware X 159 121
District of Columbia X 142 141
Georgia X 168 136
Hawaii X 164 136
Idaho X 143 158 119
Indiana X 153 136
Kansas X 165 137
Kentucky X 148 160 125
Louisiana X 150 160 125
Maine X 145 169 126
Maryland X 142 164 141
Minnesota X 140 148 124
Mississippi X 153 157 123
Missouri X 158 137
Nevada X 150 144
New Hampshire X 164 127
New Jersey X 141 162 137
New Mexico X
North Carolina X Composite with other tests
North Dakota X 151 139
Ohio X 143 167 139
Oklahoma X
Oregon X 159 138
Pennsylvania X 160 136
Rhode Island X 145
South Carolina X 162 131
South Dakota X 137 154 124
Tennessee X 140 157 136
Utah X 150 168 138
Vermont X 148 172 141
Virginia X 143 172 147
Washington X 141 158 134
West Virginia X 155 133
Wisconsin X 147 160 135
Wyoming X
National Median Score 163 178 143
Range from 25th to 75th Percentile 150-175 166-188 127-156
Range from 10th to 90th Percentile** 139-185 156-196 111-171
99
The most common type of HOUSSE option offered in spring 2005
involved a system wherein teachers were allowed to accumulate a state-
determined number of points in order to earn a highly qualified status
(29 states). Most states allowed points to be earned retroactively for such
100
things as successful completion of certain college courses (28 states) or
publishing articles and/or receiving teaching awards or honors (23 states). If
teachers could not document successful completion of college courses or
professional development in their specific content(s), they were required to earn
the points by successfully completing college courses or professional
development activities. Four states (Florida, Georgia, Minnesota, and Oklahoma)
allowed teachers to earn some points for evidence of improved student
achievement.146
Exhibit 47
Number of States Offering Various Types of HOUSSE Options
for Determining Whether Veteran Teachers Are "Highly Qualified" Under NCLB
Number of States
State offers a HOUSSE option 47
Uses a point system for HOUSSE 29
Uses teacher performance evaluation as a HOUSSE 7
Uses teacher certification systems (or the on-going evaluation
8
components of those systems) as an official HOUSSE
Uses a HOUSSE that provides teachers a menu of options for
demonstrating “highly qualified” status. 5
State does not offer a HOUSSE option 5
Note: Two states (Pennsylvania and Tennessee) are counted twice because they use more than one
of these approaches.
Source: Study of State Implementation of Accountability and Teacher Quality Under NCLB (n=52
states).
101
Tennessee stood out among the five states that provided teachers a menu of
options for demonstrating their highly qualified status. Tennessee allowed a
teacher to be deemed highly qualified if the teacher had demonstrated improved
student achievement on state tests of reading and mathematics over three
consecutive years.
102
Exhibit 48
Percentage of Classes Taught by Teachers Who Are Highly Qualified Under NCLB,
as Reported by States, 2003-04
Percent Percent
Total 86%
Alabama 77% Montana 99%
Alaska -- Nebraska 91%
Arizona 96% Nevada 64%
Arkansas -- New Hampshire 73%
California 52% New Jersey 94%
Colorado 91% New Mexico 67%
Connecticut 99% New York 92%
Delaware 73% North Carolina 85%
District of Columbia -- North Dakota 77%
Florida 89% Ohio 93%
Georgia 97% Oklahoma 98%
Hawaii 73% Oregon 87%
Idaho 97% Pennsylvania 97%
Illinois 98% Puerto Rico --
Indiana 96% Rhode Island 76%
Iowa 95% South Carolina 75%
Kansas 95% South Dakota 93%
Kentucky 95% Tennessee 58%
Louisiana 90% Texas 92%
Maine 90% Utah Exhibit 49 69%
Maryland Percentage
67% of Teachers
VermontReporting that They Are Considered
--
Massachusetts 94% Highly Qualified
Virginia under NCLB, 2004-05 95%
Michigan 92% Washington 99%
Minnesota 99% West Virginia 96%
Mississippi 93% Wisconsin 98%
Missouri 100% 96% Wyoming 99%
23% 20% 24% Don't Know
29%
80% 2% 6%
Note: Forty-seven states provided data for this table, but the national
8% estimate is based on 42 states
that reported both a numerator and 15% of classes taught
60%a denominator for calculating the percentage
by highly qualified teachers. Not Highly
40% 75% 74% Qualified
68%
Source: Consolidated State Performance Reports. 52%
20%
Highly
0% Qualified
Elementary Secondary Secondary Special
Compared with the Teachers English Math Education
state-reported data, (n=4,059) Teachers Teachers Teachers
(n=1,787) (n=1,627) (n=1,158)
principal and
teacher reports
provide somewhat
lower estimates of
the percentage of Exhibit reads: Seventy-five percent of
teachers who are elementary teachers reported that they were
highly qualified; considered highly qualified under NCLB, while 2
however, this percent said they were considered not highly
appears to be qualified, and 23 percent said they did not know
because they often their highly qualified status.
did not know
teachers’ highly Note: The percentages for “special education teachers” do
not total 100 because special educators were offered a fourth
response category – “do not need to meet highly qualified
requirement.” Four percent of special educators gave this
response. 103
Middle school teachers and special education teachers were more likely
to report that they were considered not highly qualified under NCLB
than were elementary teachers or high school teachers. For example,
although 6 percent of secondary English teachers reported in 2004-05 that they
were not highly qualified (see Exhibit 49), middle school English teachers were
twice as likely as high school English teachers to say they were not highly
qualified (8 percent vs. 4 percent). Similarly, 12 percent of middle school
mathematics teachers said they were not highly qualified, compared with 5
percent of high school mathematics teachers.152 These findings are not
surprising, since middle school teachers are less likely to have majors in English
or mathematics than their high school counterparts. For example, in a 1999-2000
survey, only 28 percent of middle school mathematics teachers reported that they
had a major in mathematics, compared with 79 percent of high school
mathematics teachers.153 Few elementary teachers (2 percent) reported that they
were not highly qualified. However, 15 percent of special education teachers said
they were not highly qualified.154
Principals reported similar but slightly higher rates of highly qualified teachers for
all categories of teachers. For example, according to principals, 82 percent of
elementary teachers were highly qualified and 77 percent of secondary
mathematics classes were taught by highly qualified teachers.
104
Students in schools that were identified for improvement for 2004-05
were more likely to be taught by teachers who were not highly qualified
under NCLB than were students in non-identified schools (see Exhibit 50).
For example, only one percent of elementary teachers in non-identified schools
said they were considered not highly qualified, compared with 5 percent in
schools that were in the first or second year of being identified for improvement,
8 percent in schools in corrective action, and 6 percent of schools in restructuring.
At the secondary level, 15 percent of teachers in schools identified for
restructuring said they were considered not highly qualified, as did 12 percent of
teachers in schools in the first or second year of improvement status.157
Exhibit 50
Percentage of Teachers Reporting That They Are
Considered Not Highly Qualified Under NCLB,
20% by School Improvement Status, 2004-05
15% *
15%
12% *
10% 8% * 8%
6% *
5%*
5% 4%
1%
0%
Elementary Teachers Secondary Teachers
Exhibit reads: In schools that were not identified for improvement, one percent of
elementary teachers reported that they were considered to not be highly qualified
under NCLB.
* Indicates that percentage was significantly different from percentage for non-identified schools
(p<.05).
Source: National Longitudinal Study of NCLB, Teacher Survey (n= 4,051 elementary teachers and
3,218 secondary teachers).
105
Reasons for teachers being considered not highly qualified under NCLB
differed by school grade level. Elementary teachers most commonly
reported that the reason was lack of full certification, while secondary
teachers were more likely to report that they had not demonstrated
subject-matter competency (see Exhibit 51). About one-third (35 percent) of
elementary teachers who said that they were not highly qualified reported that
this was because they lacked full certification, compared with only 16 percent of
secondary English teachers and 19 percent of secondary mathematics teachers.
Over half (59 percent) of secondary mathematics teachers who were not highly
qualified indicated that lack of subject-matter competency in mathematics was
the reason, while only 18 percent of secondary English teachers who were not
highly qualified indicated that lack of subject-matter competency in English was
the reason.
Exhibit 51
Reasons Why Teachers Were Considered
Not Highly Qualified Under NCLB, 2004-05
40% 35%
31%
19%
20% 14% 16% 18% 13%
0%
Elementary Teachers Secondary English Secondary Math
(n=135) Teachers (n=152) Teachers (n=243)
Note: Elementary teachers who reported that they were not highly qualified due to “lack of full
certification” represented fewer than one percent of all elementary teachers nationally.
106
Teachers who reported
Exhibit 52
that they had not met Percentage of Secondary Teachers Who Are Novice
the NCLB highly Teachers or Lack a College Major in the Subject That
qualified requirement They Teach, by Highly Qualified Status, 2004-05
also appeared less 100% English teachers with fewer than 3 years of teaching experience
qualified on other Math teachers with fewer than 3 years of teaching experience
85%
measures; for 80%
English teachers who do not have a major in English
Math teachers who do not have a major in math 75%
example, they were
more likely to lack a 60%
59%
college major in the 46%
subjects they taught or 40%
three or more years of
teaching experience 20%
18%
11%
(see Exhibit 52). 7% 9%
Among secondary 0%
English teachers, 75 Highly Qualified Not Highly Qualified
percent of those who
reported that they
were not highly
qualified under NCLB Exhibit reads: Secondary English teachers who
did not have a major in said they were not highly qualified under NCLB were
English, compared with more likely to be novice teachers with fewer than
46 percent of those three years of teaching experience (18 percent) than
who said they were those who were considered highly qualified (7
highly qualified. percent).
C. Professional Development
Research indicates that professional development that places a strong
emphasis on academic content, and on how students learn specific
content, is associated with gains in student achievement.159 Research also
indicates that teachers reported that professional development enhanced their
knowledge and skills when it was sustained and intensive; connected to state
standards and to teachers’ goals or other learning experiences; involve teams of
teachers from the same grade levels, departments, or schools; and allow teachers
to observe and practice the skills and techniques being introduced or to actively
engage in conversations about teaching and learning.160
108
development is more likely to have an impact if it involves many contact hours
over a long time period.163
Exhibit 53
Percentage of Teachers Participating in Professional Development
Focused on Instructional Strategies for Reading and Mathematics,
2003-04
100% 9%
20% 16%
22%
More than
80%
24 Hours 26%
39% 6 to 24 30%
60% 36% Hours
1 to 5 37%
40% hours 31%
31% 30% None
20%
29% 23%
10% 12%
0%
Elementary Secondary Elementary Secondary
Teachers English Teachers Math
(n=4,007) Teachers (n=3,994) Teachers
(n=1,740) (n=1,580)
PD in Reading PD in Mathematics
110
special educators reported that they participated in at least some professional
development that provided them with the opportunity to practice what they had
learned and receive feedback; more than half of secondary English and
mathematics teachers participated in training that involved this kind of activity.
In addition, more than 50 percent of these same groups of teachers reported that
they reviewed student work or scored assessments as part of some of their 2003-
04 professional development activities. Elementary teachers were more likely
than secondary mathematics and English teachers to report that they engaged in
each of these activities.168
Teachers who reported that they were not considered highly qualified
were no more likely to report that they participated in content-focused
professional development than were highly qualified teachers. However,
elementary teachers who said they were not highly qualified under NCLB were
more likely to report participating in a sustained mentoring or new-teacher
induction program (47 percent, compared with 26 percent of highly qualified
elementary teachers) during the 2003-04 school year. However, no significant
differences were found for secondary teachers or for other types of support, such
as peer coaching or release time for course preparation or college courses.169
More than three-quarters of Title I instructional aides reported that they spent at
least some of their work day tutoring students one-on-one (79 percent) or working
111
with students in groups (87 percent); on average these aides reported spending
about 57 percent of their time on these two activities. Nearly one-quarter (23
percent) reported that, of the time that they spent tutoring or working
with students in a classroom, a teacher was present for half or less of
this time.173
Among Title I instructional aides who said they were not qualified under
NCLB, 30 percent reported “not enough money or funding to become
qualified” as a major challenge and 21 percent reported “not enough
time to get qualified.” Other major challenges reported by aides were
insufficient encouragement from school and district (17 percent), level of difficulty
of the test (13 percent), and insufficient information about what they needed to
do (8 percent).174
The majority of states, districts, and schools reported that they had
adopted at least one strategy to help Title I aides comply with the NCLB
“qualified” requirements as of fall 2004. At the state level, the most
common strategies were working with local colleges and universities to design
needed courses or offering evening and weekend courses to Title I aides (21
states) and offering test preparation courses for aides wishing to take the state
competency exam (13 states). Other common strategies included offering
funding for course tuition (10 states) and paying the state test fee for interested
aides (six states).175
Conclusions
Due to concern that too many teachers, particularly those in low-performing
schools, had not met state certification requirements or lacked expertise in the
subjects they were teaching, NCLB requires that all teachers be “highly qualified”
by 2005-06. Although most states are well on the way to meeting the law’s
requirements, we do not have evidence about whether the qualifications of the
teaching workforce have actually changed.
The large majority of teachers (86 percent) have been designated as “highly
qualified” according to state-reported data for 2003-04. However, teachers in
schools with high concentrations of low-income students or minority students
were more likely to be considered not highly qualified under NCLB. In addition,
almost one-fourth of teachers surveyed said they did not know their highly
qualified status.
112
The way in which states are implementing the HOUSSE option for veteran
teachers has been the subject of considerable debate. Nearly all states (47) offer
a HOUSSE option, with the most common type involving a point system. Most
states allow points to be earned for completion of college courses, published
articles, and teaching awards or honors. Four states recognize improved student
achievement. Twenty-six states allow a substantial percentage of required points
to be earned for prior teaching experience.
Professional development has been and remains a key strategy for improving
teacher effectiveness. Most teachers reported receiving some professional
development in reading and math, but a relatively small proportion participated in
such training for an extended period of time. For example, only 20 percent of
elementary school teachers reported receiving more than 24 hours of training in
reading instruction in 2003-04. Teachers were less likely to receive training in
instructional strategies for teaching mathematics or in-depth study of topics in
reading or mathematics. Special education teachers were less likely than general
education teachers to receive training focused on reading and mathematics.
Classroom teachers in high-poverty schools received more training in both
reading and mathematics.
113
References
Kennedy, Mary (1998). Research Monograph No. 13: Form and Substance in
Inservice Teacher Education. Madison: National Institute for Science Education,
University of Wisconsin.
Center on Education Policy (2005). From the Capital to the Classroom: Year 3 of
the No Child Left Behind Act. Washington, D.C.: Center on Education Policy.
Cohen, David K., Milbrey W. McLaughlin, and Heather C. Hill (1998). Instructional
Policy and Classroom Performance: The Mathematics Reform in California (RR-39).
Philadelphia: Consortium for Policy Research in Education.
Garet, Michael S., Beatrice F. Birman, Andrew C. Porter, Laura Desimone, Rebecca
Herman, and Kwang Suk Yoon (1999). Designing Effective Professional
Development: Lessons From the Eisenhower Professional Development Program.
Washington, D.C.: U.S. Department of Education, Office of the Under Secretary,
Planning and Evaluation Service.
Goldhaber, Dan D., and Dominic J. Brewer (2000). “Does Teacher Certification
Matter? High School Certification Status and Student Achievement” in
Educational Evaluation and Policy Analysis, 22: 129-145.
Padilla, Christine, Katrina Woodworth, Andrea Lash, Patrick M. Shields, and Katrina
G. Laguarda (2005). Evaluation of Title I Accountability Systems and School
Improvement Efforts: Findings From 2002-03. Washington, DC: U.S. Department
114
of Education, Office of Planning, Evaluation, and Policy Development, Policy and
Program Studies Service.
Perie, Marianne, Wendy S. Grigg, and Patricia L. Donahue (2005). The Nation's
Report Card: Reading 2005 (NCES 2006-451). Washington, DC: U.S. Department
of Education, National Center for Education Statistics.
Perie, Marianne, Wendy S. Grigg, and Gloria S. Dion (2005). The Nation's Report
Card: Mathematics 2005 (NCES 2006-453). Washington, DC: U.S. Department of
Education, National Center for Education Statistics.
Schiller, Ellen, Ellen Bobronnikov, Fran O’Reilly, Cristofer Price, and Robert
St.Pierre (2005). The Study of State and Local Implementation and Impact of the
Individuals with Disabilities Education Act: Final 2nd Interim Report (2002-2003
School Year). Bethesda, MD: Abt Associates.
Seastrom, Marilyn, Lee Hoffman, Chris Chapman, and Robert Stillwell (2005). The
Averaged Freshman Graduation Rate for Public High Schools From the Common
Core of Data: School Years 2001-02 and 2002-03 (NCES 2006-601). Washington,
DC: U.S. Department of Education, National Center for Education Statistics.
Seastrom, Marilyn, Chris Chapman, Robert Stillwell, Daniel McGrath, Pia Peltola,
Rachel Dinkes, and Zeyu Xu (forthcoming). A Review and Analysis of Alternative
High School Graduation Rates, Volume I. User’s Guide to Computing High School
Graduation Rates, Volume 2, An Analysis of Alternative High School Graduation
Rates (NCES 2006-602). Washington, DC: U.S. Department of Education, National
Center for Education Statistics.
Seastrom, Marilyn, Kerry Gruber, Robin Henke, Daniel McGrath, and Benjamin
Cohen (2004). Qualifications of the Public School Teacher Workforce: Prevalence
of Out-of-Field Teaching: 1987-88 to 1999-2000 (NCES 2002-603). Washington,
DC: U.S. Department of Education, National Center for Education Statistics.
Shields, Patrick M., Camille Esch, Andrea Lash, Christine Padilla, Katrina
Woodworth, Katrina G. Laguarda, and Nicholas Winter (2004). Evaluation of Title I
Accountability Systems and School Improvement Efforts (TASSIE): First-Year
Findings. Washington, DC: U.S. Department of Education, Office of the Under
Secretary, Policy and Program Studies Service.
Beth Sinclair (forthcoming). State ESEA Title I Participation Information for 2002-
03: Final Summary Report. Washington, DC: U.S. Department of Education,
Office of Planning, Evaluation, and Policy Development, Policy and Program
Studies Service.
115
Enhancement and Academic Achievement Act (ESEA, Title III, Part A). Washington,
D.C: U.S. Department of Education.
Wagner, Mary, Lynn Newman, Renee Cameto, and Phyllis Levine (2005). Changes
Over Time in the Early Postschool Outcomes of Youth with Disabilities: A Report of
Findings from the National Longitudinal Transition Study (NLTS) and the National
Longitudinal Transition Study-2 (NLTS2). Menlo Park, CA: SRI International.
Williams, Andra, Rolf K. Blank, and Carla Toye (forthcoming). State Education
Indicators With a Focus on Title I: 2002-03. Washington, DC: U.S. Department of
Education, Office of Planning, Evaluation, and Policy Development, Policy and
Program Studies Service.
116
Acknowledgments
This report benefited from the contributions of many individuals and organizations
that provided valuable information and advice. Although they are too numerous
to mention each by name, we appreciate their support, and would like to
specifically acknowledge the contributions of the following individuals.
The report was prepared under the leadership and direction provided by Alan
Ginsburg, director of the Policy and Program Studies Service (PPSS), David
Goodwin, director of program and analytic studies in PPSS, Daphne Kaplan, PPSS
team leader, Ricky Takai, associate commissioner for the evaluation division of the
National Center for Education Evaluation and Regional Assistance (NCEE), and
Audrey Pendleton, senior education researcher at NCEE.
Other researchers who provided useful assistance for this report include Brian
Gong of the Center for Assessment, Allison Henderson and Beth Sinclair of Westat,
and Andra Williams of the Council of Chief State School Officers.
117
Teachers, principals, school district staff, and state education agency
representatives across the country took time out of their busy schedules to
respond to our surveys, interviews, and requests for information. Without their
efforts, this report would not have been possible, and we greatly appreciate their
support for this national assessment as well as their core work of educating
America’s children.
Many Department staff reviewed drafts of this report and provided useful
comments and suggestions as well as, in some cases, providing data for the
report. We would like to acknowledge the assistance of: Andrew Abrams, Millicent
Bentley-Memon, Kerri Briggs, Chris Chapman, William Cordes, Thomas Corwin,
Laurette Crum, Sarah Dillard, Kathryn Doherty, Brian Fu, Arnold Goldstein, Patricia
Gonzalez, Lee Hoffman, Rene Islas, Jacquelyn Jackson, Stacy Kreppel, Milagros
Lanauze, Kathleen Leos, Jeannette Lim, David Malouf, Darla Marburger, Carlos
Martinez, Richard Mellman, Meredith Miller, Michael Petrilli, Anne Ricciuti, Kay
Rigling, Krista Ritacco, Patricia O’Connell Ross, Ross Santy, Zollie Stevenson, Bob
Stonehill, and Elizabeth Witt.
While we appreciate the assistance and support of all of the above individuals,
any errors in judgment or fact are of course the responsibility of the authors.
118
Appendix A: Description of Major Data Sources
Included in This Report
119
National Longitudinal Study of NCLB (NLS-NCLB)
Purpose
Sample Design
The nationally representative sample includes 300 districts and 1,483 schools
within those districts, including both Title I and non-Title I schools. In order to
ensure sufficient sample sizes of schools identified for improvement under Title I,
the study oversampled high-poverty districts and schools, as well as oversampling
Title I schools. The distribution of sample schools by grade level is similar to the
distribution of all schools. The original sample included 1,502 schools, but 19
were determined to be out-of-scope and the net sample was 1,483 schools.
Exhibit A.1
Characteristics of NLS-NCLB District and School Sample,
Compared with the Universe of Districts and Schools
Sample Universe
Number Percent Number Percent
Districts, by Poverty Quartile (Census poverty) 300 14,972
Highest poverty quartile 163 54% 3,743 25%
Second highest poverty quartile 41 14% 3,743 25%
Second lowest poverty quartile 50 17% 3,743 25%
Lowest poverty quartile 46 15% 3,743 25%
Schools, By Poverty Level 1,502 83,298
75-100% eligible for free or reduced price lunch 596 40% 11,282 13%
50-74% eligible for free or reduced price lunch 363 24% 15,461 19%
35-49% eligible for free or reduced price lunch 106 7% 12,844 15%
<35% eligible for free or reduced price lunch 291 19% 33,884 41%
Missing 146 10% 9,827 12%
Schools, by Title I Status 1,502 83,298
Title I 1,163 77% 46,048 55%
Non Title I 259 17% 31,312 38%
Missing 80 5% 5,938 7%
Schools, by Grade Level 1,502 83,298
Elementary 906 60% 50,597 61%
120
Middle 298 20% 15,700 19%
High 298 20% 17,001 20%
District poverty quartiles were based on Census Bureau estimates of the number
of school-age children and poor children living in each district (2002 Small Area
Income and Poverty Estimates). The poverty quartiles were created by ranking all
districts by the percentage of poor school-age children and then dividing these
districts into quartiles that each contain 25 percent of the school-age children.
School poverty levels were based on the percentage of students eligible for free
or reduced-price lunches. The eligibility threshold for the subsidized lunch
program is looser than the official poverty definition (eligibility for reduced-price
lunches is set at 185 percent of the official poverty definition), so school poverty
rates are generally higher than district poverty rates.
The teacher sample includes approximately seven teachers per school (six
classroom teachers and one special education teacher). School staff rosters were
collected and divided into teacher strata by grade level; a stratum of Title I
paraprofessionals was also created. After school rosters were stratified,
independent random sampling took place within each stratum. At the elementary
level, one teacher was selected per grade. At the secondary level, about three
math teachers and three English teachers were selected per school. One Title I
paraprofessional was selected from each Title I school that uses such
paraprofessionals. The resulting sample included a total of 8,791 classroom
teachers (including 4,772 elementary teachers, 2,081 secondary English teachers,
and 1,938 secondary mathematics teachers), 1,408 special education teachers,
and 950 paraprofessionals.
The parent sample consists of a maximum of 400 parents in each of eight districts
for a total of 3,094 parents. In each district, the 400 parents were selected
randomly from four groups: 100 parents of students receiving supplemental
services in schools identified for improvement; 100 parents of students not
receiving supplemental services in schools identified for improvement; 100
parents of students who moved from an identified to a non-identified school; and
100 parents of students in non-identified schools. Some districts had fewer than
100 students who moved from an identified to a non-identified school. The eight
districts were selected based on availability of the necessary longitudinal
individual student achievement data and sufficient numbers of students
participating in the Title I school choice and supplemental services options to
support the above target sample sizes.
Data Collection
121
Data collection instruments for this study that were used in this report include
mail surveys of district federal program coordinators, school principals, classroom
teachers, and Title I paraprofessionals; survey administration for the 2004-05
school year began in October 2004 and was completed in March 2005. Topics
covered in the survey questionnaires included accountability systems, AYP and
identification for improvement, technical assistance, improvement strategies, use
of assessment results, Title I school choice and supplemental educational
services, teacher quality, and professional development.
Surveys of parents and supplemental service providers were also conducted but
were not completed in time for this report. Other components of the data are also
ongoing, including review of extant data such as state report cards and school
improvement plans, analyses of state assessment data, and targeting and
resource allocation data.
The study includes two exploratory achievement analyses that are examining
achievement outcomes for students participating in the Title I choice and
supplemental services options (in nine districts) and the impact of identifying
schools for improvement on student achievement (in two states). Both analyses
are using quasi-experimental designs.
122
For the targeting and resource allocation component, the study is collecting data
from each of the 50 states on state suballocations of federal program funds to
school districts for the six programs included in this component: Title I Part A, Title
II Part A, Title III, Reading First, Comprehensive School Reform (CSR), and Perkins
Vocational Education. Districts in the 300-district sample were asked to provide
budget, expenditure, and administrative records, including personnel and payroll
records, for these six programs. The sample districts were also asked to provide
their allocations to schools for Title I, Reading First, and CSR, as well as school-
level budgets and plans for the uses of these funds in the sample schools within
their district. For schools operating schoolwide programs under Title I, the study
is collecting schoolwide plans and budgets if applicable. The information on
targeting and resource allocation will be collected one time only, for the 2004-05
school year.
Survey response rates for 2004-05 were 96 percent for the school district survey,
89 percent for the principal survey, 84 percent for the teacher surveys, and 87
percent for the Title I paraprofessional survey.
Survey data were weighted in order to produce national estimates. At the school
level, for example, the base weight for each school is the reciprocal of the
school's two-stage selection probability, equal to the product of the probability of
selecting the district and the conditional probability of selecting the school, given
the district. In addition, the weights were adjusted, controlling for covariates, to
handle instances of total school non-response. School weights were raked to
population counts of schools in four dimensions: school size, region by poverty
stratus, metro status, and school type. Two sets of weights were finally produced
for schools: (1) a set for estimating the proportion of schools with a defined
attribute, and (2) a set for estimating the proportion of students attending schools
with a defined attribute. Similar weighting procedures were employed for the
district, teacher, and paraprofessional survey data.
Missing data were imputed for principal survey data on the total number of
elementary classroom teachers and secondary classes, which were used as
denominators for calculating the percentage of elementary teachers who were
considered highly qualified under NCLB and the percentage of secondary classes
that were taught by highly qualified teachers (reported on page 78). There were
18 out of 930 elementary school principals that did not answer the survey item
asking about the total number of classroom teachers at their schools, and 36 out
of 385 secondary school principals that did not answer the survey item about the
total number of class sections. Data for elementary classroom teachers were
imputed by taking the student teacher ratios for the principals who answered the
item and then fitting a regression model on this ratio by the total number of
students enrolled and the school poverty level as the predictors. Using the
regression coefficients, the predicted student teacher ratio was computed for
each of the 18 schools and then converted to the estimated number of classroom
teachers in the school. Data on the total number of secondary class sections
123
were imputed in a similar manner. There were two elementary school principals
and five secondary school principals whose values could not be imputed due to
missing values in the predictor variables.
Reporting
This study will issue a series of reports in collaboration with the Study of State
Implementation of Accountability and Teacher Quality Under NCLB. Interim
reports on accountability and teacher quality are due in Spring 2006, an interim
report on public school choice and supplemental services under Title I is due in
Summer 2006, and a report on targeting and resource allocation is due in Fall
2006. Reports from the second wave of the data collection are due in late 2007.
The reports will be available at:
www.ed.gov/about/offices/list/opepd/ppss/reports.html.
124
Study of State Implementation of Accountability and
Teacher Quality Under NCLB (SSI-NCLB)
Purpose
The SSI-NCLB study is being conducted by the American Institutes for Research in
collaboration with the Council of Chief State School Officers and REDA
International.
Study Design
Data Collection
The study has conducted telephone interviews with state-level personnel with
responsibilities in the key areas of this evaluation, such as state federal program
coordinators responsible for administering Title I, Title II, and Title III, as well as
state assessment directors. The interviews began in September 2004 and were
completed in March 2005. Topics covered in the interviews included state
assessment and accountability systems, state implementation of supplemental
educational services, state teacher quality and professional development
initiatives, and accountability and teacher quality under Title III.
The study also collected extant data from a range of sources including
consolidated state applications and consolidated state performance reports, state
report cards, and state educational agency websites. In particular, the study
compiled a detailed school-level database on the identification status of schools
and whether the schools met or missed AYP targets. The database contains the
identification status of 88,160 schools (Title I and non-Title I) in 50 states and the
District of Columbia. The database also contains the AYP status of 87,892 schools
located in approximately 15,000 districts across 50 states and the District of
Columbia. Some states did not report data on certain AYP targets; as a result, the
number of states and schools for which data is available on individual AYP targets
varies from 33 states (including 15,731 schools missing AYP and 61,868 schools
overall) to the full dataset.
Response Rates
125
Interviews for 2004-05 were completed for all 50 states plus the District of
Columbia and Puerto Rico.
Reporting
This study will issue a series of reports in collaboration with the National
Longitudinal Study of NCLB. Interim reports on accountability and teacher quality
are due in Spring 2006, and an interim report on public school choice and
supplemental services under Title I is due in Summer 2006. Reports from the
second wave of the data collection are due in late 2007. The reports will be
available at: www.ed.gov/about/offices/list/opepd/ppss/reports.html.
126
Study of Title I Accountability Systems and School
Improvement Efforts (TASSIE)
Purpose
Study Design
Data Collection
Data collection instruments for this study include mail surveys of district Title I
administrators and school principals, a telephone survey of state Title I
administrators, and site visit protocols for the case studies. The district and
school surveys, along with the case studies, were conducted in 2001-02, 2002-03
and 2003-04. The state survey was conducted twice, in 2002-03 and 2003-04.
Topics covered in the surveys included school and district identified for
improvement, school improvement activities in identified schools, corrective
actions and restructuring activities for identified schools, and the implementation
of public school choice and supplemental services under Title I.
TASSIE district and school samples are both stratified random samples in which
the probability of selection into the sample varies across strata. To estimate
population parameters, the sampled districts or schools are weighted so that the
total of the weights within a stratum equals the number of districts or schools in
that stratum in the sampling frame.
Survey response rates ranged from 88 to 90 percent for the district survey and
from 83 to 85 percent for the principal survey. To estimate population parameters
from the survey respondents, the weights assigned to respondents within any
127
stratum were modified to absorb the weights that would otherwise accrue to non-
responding schools in the stratum (thus, respondents’ weights were adjusted to
sum to the total number in the stratum). A new set of weights were derived for
each year of the survey since the set of respondents varied from one year to
another. The longitudinal estimates presented in this report use the analysis
weights assigned for the 2001-02, 2002-03, and 2003-04 respondent pools,
respectively.
There were limited cases of missing data, and there was no imputation for missing
data.
Reporting
Reports on the 2001-02 and 2002-03 data collections have been released and are
available at www.ed.gov/about/offices/list/opepd/ppss/reports.html#title. A final
report that includes data from the third and final year of the study (2003-04) is
forthcoming.
Shields, Patrick M., Camille Esch, Andrea Lash, Christine Padilla, Katrina
Woodworth, Katrina G. Laguarda, and Nicholas Winter (2004). Evaluation of
Title I Accountability Systems and School Improvement Efforts (TASSIE):
First-Year Findings. Washington, DC: U.S. Department of Education, Office of
the Under Secretary, Policy and Program Studies Service.
128
Case Studies of the Early Implementation of Supplemental
Educational Services
Purpose
This study examined how states and districts were implementing the
supplemental educational services provisions of NCLB during the first two years
they were in effect, the 2002-03 and 2003-04 school years.
The study was conducted by Policy Studies Associates under subcontract to SRI
International.
The case studies focused on nine school districts in six states implementing NCLB
supplemental services during the 2002-03 and 2003-04 school years. In each
district, case studies included visits to approximately three schools and three
supplemental services providers. Case studies also included telephone interviews
of state personnel; in-person interviews with district administrators, school
principals and providers; and focus groups with teachers and parents
Reporting
The interim and final reports from this study have been released and are available
at www.ed.gov/about/offices/list/opepd/ppss/reports.html#title.
129
Consolidated State Performance Reports
Purpose
Section 1111 of the No Child Left Behind Act (NCLB) requires states to provide an
annual report to the Secretary that includes data on student achievement on
state assessments, disaggregated for various student subgroups specified in the
law, as well as the number and names of schools identified for improvement
under Title I, the reasons why each school was so identified, the percentage of
classes taught by teachers who are highly qualified under NCLB, and other
information. Section 9303 gives States the option of reporting on multiple ESEA
programs through a single consolidated report, and all states do in fact use the
consolidating reporting option. The Consolidated State Performance Reports also
collect basic descriptive information about programs, such as numbers of
participating schools and students, and numbers of schools identified for
improvement.
The Consolidated State Performance Reports are divided into Part I, which
includes achievement data on state assessments, implementation of Title I
accountability requirements, and other information considered high-priority, and
Part II, which includes the remaining required information and has a later due
date. For 2002-03, Part I reports were due to the U.S. Department of Education in
December 2003 and Part II reports were due in June 2004. State reports for 2003-
04 were due in early 2005 (January 2005 for Part I and April 2005 for Part II);
however, these data were not available in time for inclusion in this report due to
delays in state submissions and the need for data cleaning which has not yet
been completed.
Reporting
Two annual reports summarize data from the Title I Part A portion of the
Consolidated State Performance Reports. Reports presenting the 2001-02 data
have been released and are available at
www.ed.gov/about/offices/list/opepd/ppss/reports.html#title. Reports presenting
the 2002-03 data used in this report are forthcoming:
Williams, Andra, Rolf K. Blank, and Carla Toye (forthcoming). State Education
Indicators With a Focus on Title I: 2002-03. Washington, DC: U.S. Department
of Education, Office of Planning, Evaluation, and Policy Development, Policy
and Program Studies Service.
130
Beth Sinclair (forthcoming). State ESEA Title I Participation Information for
2002-03: Final Summary Report. Washington, DC: U.S. Department of
Education, Office of Planning, Evaluation, and Policy Development, Policy and
Program Studies Service.
131
Appendix B: Supplemental Exhibits
132
Exhibit B-1
Percentage of 8th-Grade Students Achieving At or Above the "Proficient" Level
on NAEP and State Assessments in Mathematics, 2003
Minnesota 77
44
Massachusetts 37
38
North Dakota 44
36
Connecticut 77
35
Montana 70
35
New Hampshire 74
35
South Dakota 55
35
Vermont 35
Wisconsin 73
35
Colorado 69
34
Kansas 60
34
Iowa 72
33
New Jersey 57
33
Nebraska 75
32
New York 32
North Carolina 82
32
Oregon 59
32
Washington 37
32
Wyoming 35
32
Indiana 71
31
Utah 70
31
Virginia 75
31
Alaska 64
30
Maryland 40
30
Ohio 53
30
Pennsylvania 51
30
Illinois 52
18 29
Maine 29
Idaho 53
28
Michigan 40
28
Missouri 14
28
Delaware 47
19 26
South Carolina 26
Texas 73
25
Kentucky 31
24
Rhode Island 34
24
Florida 57
23
California 29
22
Georgia 67
18 22
Arizona 21
Tennessee 79
21
Nevada 38
20
Oklahoma 65
20
West Virginia 20
Arkansas 22
19 State Assessment, 2003
Hawaii 17
17 NAEP, 2003
Louisiana 52
17
Alabama 56
16
New Mexico 64
15
Mississippi 48
12
District of Columbia 40
6
0 20 40 60 80 100
Source: Consolidated State Performance Reports and National Center for Education Statistics, Main
NAEP. Note: The preferred grade for this table was 8th grade; however, in states
that did not consistently assess students in 8th-grade mathematics, we used either 7th- or 6th-grade
133
assessment results.
134
Exhibit B-2
Proportion of Students Performing At or Above Their State’s Proficient Level
in Reading and Mathematics, in 8th Grade or Another Middle School Grade, 2000-01 to 2002-03
Reading Mathematics
2000-01 2001-02 2002-03 Change 2000-01 2001-02 2002-03 Change
Alabama 64 58 -6 71 54 -17
Arizona 42 46 4 18 18 0
Colorado 64 65 89 25 37 39 69 32
Connecticut 77 78 78 1 76 77 77 1
Delaware 72 72 70 -2 43 48 47 4
Illinois 66 68 63 -3 50 52 52 2
Kansas 66 67 71 5 57 56 60 3
Kentucky 54 56 57 3 27 26 31 4
Louisiana 51 48 55 4 46 41 52 6
Maine 41 43 45 4 20 21 18 -2
Massachusetts 67 64 34 34 37 3
Mississippi 49 48 57 8 40 45 48 8
Missouri 34 32 32 -2 14 14 14 0
Montana 73 71 71 -2 69 68 70 1
New Jersey 73 73 74 1 62 58 57 -5
North Carolina 83 85 26 -57 80 83 82 2
Ohio 58 56 65 7 61 59 53 -8
Oklahoma 70 70 71 1 63 64 65 2
Oregon 62 64 60 -2 55 58 59 4
Pennsylvania 60 58 64 4 51 52 51 0
South Carolina 24 27 20 -4 18 19 19 1
Utah 36 78 67 31 66 40 70 4
Virginia 73 70 70 -3 68 70 75 7
Washington 40 44 48 8 27 30 37 10
# of states with
14 out of 23 states 17 out of 24 states
achievement gains
Note: The preferred grade for this table was 8th grade; however, in states that did not
consistently assess students in 8th-grade reading and mathematics, a nearby grade was used
(7th grade for Kansas (math), Kentucky (reading), Missouri (reading), and Washington, and 6th
grade for Alabama and Ohio).
135
Exhibit B-3
Proportion of Students Performing At or Above Their State’s Proficient Level for Reading in 2002-03,
and Change from 2000-01, in 8th Grade or Another Elementary Grade, for Various Student Subgroups
Note: The preferred grade for this table was 8th grade; however, in states that did not consistently assess students in 8th-grade reading, a nearby
grade was used (6th grade for Ohio and 8th grade for Kentucky, Massachusetts, Missouri, and Washington). Gray cells indicate that the state did not
report disaggregated assessment data for that subgroup for all three years included in the analysis; however, all of these states have since developed
the capacity to report disaggregated data.
136
Exhibit B-4
Proportion of Students Performing At or Above Their State’s Proficient Level for Mathematics in 2002-03,
and Change from 2000-01, in 8th Grade or Another Elementary Grade, for Various Student Subgroups
Note: The preferred grade for this table was 8th grade; however, in states that did not consistently assess students in 8th-grade mathematics, a
nearby grade was used (6th grade for Ohio and 8th grade for Kansas and Washington).
137
Exhibit B-5
Change in the Achievement Gap: Difference Between the Proportion of Low-Income Students and
All Students Performing At or Above Their State’s Proficient Level, in 8th Grade or
Another Middle School Grade, 2000-01 to 2003-03
* Indicates that the score is significantly different from the one in 2005 (p<.05).
Note: “High-poverty” was defined as schools with 76 to 100 percent of their students eligible for free or
reduced-price lunches, and “low-poverty” indicates that 0 to 25 percent were eligible for subsidized
lunches.
Source: National Center for Education Statistics, Main NAEP, unpublished tabulations.
138
Exhibit B-8 Exhibit B-9
Reading Achievement on the Main NAEP, 1992 to 2005: Mathematics Achievement on the Main NAEP, 1990 to 2005:
Average Scale Scores in 8th Grade by Race/Ethnicity Average Scale Scores in 8th Grade by Race/Ethnicity
300 300 White
287* 288
283*
279*
280 White 280 276*
271* 270* 269
268 269 *
265 * 265* Hispanic
* 261
258
260 260
Hispanic 252*
249*
245 244 245 245 * 247*
241 254 Black
239* 252*
238 *
240 244* 244* 242 Black 240
243*
242
239*
236 * 235* 236 * 236*
220 220
1992 1994 1998 2002 2003 2005 1990 1992 1996 2000 2003 2005
* Indicates that the score is significantly different from the one in 2005 (p<.05).
30%
29% *
30%
25%*
Hispanic
20%
20% 18% * Hispanic
13% 14% 14% 14%
11%* 12% 13%
11%*
10% 13% 12% 7% * 7%* 8% *
11% 11% 10% 6% *
8% * 9%* Black Black
5%* 7%* 8%
0% 2%* 4%* 5% *
0%
1992 1994 1998 2000 2002 2003 2005
1990 1992 1996 2000 2003 2005
* Indicates that the score is significantly different from the one in 2005 (p<.05).
139
Exhibit B-12 Exhibit B-13
Reading Achievement on the Trend NAEP, 1971 to 2004: Mathematics Achievement on the Trend NAEP, 1978 to 2004:
Average Scale Scores for 13-Year-Olds by Race/Ethnicity Average Scale Scores for 13-Year-Olds by Race/Ethnicity
300 300
288 White
* * 283*
279* 281 281
280 280 274 * 274 * 274* 276*
White 272 * Hispanic
266 265 266 267 266 265
263*
261* 262*
264
261* 262* 259* 259*
260 260 254* 255* 256* 256 *
252 * 262 Black
Hispanic
244 * 252* 252 *
240 238 242
239* 251*
240 239 238 238 * 249* 249* 250
240 237 235* 240
244 Black
232 * 243
241 238 238 240 *
236 * 234 *234 *
233 *
222 *
226 * 228 * 230 *
220 220
1971 1975
1975 1980
1980 1984 1988 1990
19901992 1994 1996 1999 2004 1973 1978 1982 1986 1990 1992 1994 1996 1999 2004
* Indicates that the score is significantly different from the one in 2004 (p<.05).
140
Appendix C: Standard Error Tables
141
Exhibit C-1
Reading and Mathematics Achievement on the Main NAEP, 1990 to 2005:
Average Scale Scores by School Grade Level
142
Exhibit C-2
Reading and Mathematics Achievement on the Main NAEP, 1990 to 2005:
Average Scale Scores in 4th Grade by School Poverty Level
143
Exhibit C-3
Reading and Mathematics Achievement on the Main NAEP, 1990 to 2005:
Average Scale Scores in 4th Grade by Race/Ethnicity
144
Exhibit C-4
Reading and Mathematics Achievement on the Main NAEP, 1990 to 2005:
Percent Proficient in 4th Grade by Race/Ethnicity
Black Hispanic White
Reading
1992 8 (1.4) 10 (1.7) 33 (1.8)
1994 8 (0.9) 11 (2.1) 35 (1.5)
1998 10 (1.0) 12 (1.7) 36 (1.2)
2000 9 (1.0) 12 (1.9) 36 (1.4)
2002 12 (0.5) 14 (0.8) 39 (0.5)
2003 12 (0.4) 14 (0.5) 39 (0.3)
2005 12 (0.3) 15 (0.5) 39 (0.3)
Mathematics
1990 1 (0.5) 4 (1.6) 15 (1.7)
1992 2 (0.6) 5 (1.2) 22 (1.5)
1996 3 (0.7) 7 (1.4) 26 (1.5)
2000 4 (0.8) 7 (1.0) 30 (1.4)
2003 10 (0.3) 15 (0.5) 42 (0.3)
2005 13 (0.3) 19 (0.3) 47 (0.3)
Source: National Center for Education Statistics, Main NAEP.
145
Exhibit C-5
Reading and Mathematics Achievement on the Trend NAEP, 1971 to 2004:
Average Scale Scores by Student Age Group
146
Exhibit C-6
Reading and Mathematics Achievement on the Trend NAEP, 1971 to 2004:
Average Scale Scores for 9-Year-Olds by Race/Ethnicity
147
Exhibit C-7
Reading and Mathematics Achievement on the Main NAEP, 1990 to 2005:
Average Scale Scores in 8th Grade by School Poverty Level
148
Exhibit C-8
Reading and Mathematics Achievement on the Main NAEP, 1990 to 2005:
Average Scale Scores in 8th Grade by Race/Ethnicity
149
Exhibit C-9
Reading and Mathematics Achievement on the Main NAEP, 1990 to 2005:
Percent Proficient in 8th Grade by Race/Ethnicity
Black Hispanic White
Reading
1992 8 (1.1) 11 (1.3) 33 (1.4)
1994 9 (1.2) 12 (1.3) 33 (1.2)
1998 11 (1.6) 13 (1.0) 37 (1.3)
2002 13 (0.7) 14 (0.8) 39 (0.7)
2003 12 (0.4) 14 (0.6) 39 (0.3)
2005 11 (0.4) 14 (0.4) 37 (0.3)
Mathematics
1990 5 (1.1) 7 (2.1) 18 (1.4)
1992 2 (0.7) 6 (1.0) 25 (1.2)
1996 4 (0.7) 7 (1.2) 29 (1.4)
2000 5 (0.7) 8 (1.0) 33 (1.1)
2003 7 (0.3) 11 (0.5) 36 (0.4)
2005 8 (0.3) 13 (0.4) 37 (0.3)
Source: National Center for Education Statistics, Main NAEP.
150
Exhibit C-10
Reading and Mathematics Achievement on the Trend NAEP, 1971 to 2004:
Average Scale Scores for 13-Year-Olds by Race/Ethnicity
151
Exhibit C-11
Percentage of Identified Schools That Reported Needing and Receiving Various Types
of Technical Assistance, 2003-04 to 2004-05
152
Exhibit C-12
Percentage of Schools Reporting Major Focus
on Various School Improvement Strategies, 2004-05
153
Exhibit C-13
Change in Instructional Time Per Day at Elementary Schools,
by Subject Area, 2003-04 to 2004-05
154
Exhibit C-14
Percentage of Identified Title I Schools Experiencing Various Types of Interventions
Since Identification for Improvement, 2004-05
155
Exhibit C-15
Percentage of Districts Taking Various Actions in Response to
Being Identified for Improvement, 2004-05
Exhibit C-16
Number of Schools Where Title I School Choice and Supplemental Services Were Offered,
and Number of Participating Students, 2002-03 to 2004-05
Note: The estimated number of participating schools in 2004-05 is based on two data sources: a
count from the SSI-NCLB study of the number of Title I schools identified for improvement in
2004-05, and an estimate from the NLS-NCLB of the proportion of Title I identified schools that
reported that they were required to offer Title I school choice. Because this estimate is based on
a combination of two data sources, a standard error cannot be calculated.
156
Exhibit C-17
Share of Students Receiving Supplemental Services, by Type of Provider, 2003-04
Exhibit C-18
District Strategies for Communicating with Parents
About Title I School Choice and Supplemental Services Options, 2004-05
157
Exhibit C-19
Percentage of Teachers Reporting that They Are Considered
Highly Qualified Under NCLB, 2004-05
Exhibit C-20
Percentage of Teachers Reporting that They Are Considered Highly Qualified
under NCLB, 2004-05, by School Improvement Status, 2004-05
158
Exhibit C-21
Reasons Why Teachers Were Designated as Not Highly Qualified Under NCLB, 2004-05
Exhibit C-22
Percentage of Secondary Teachers Who Are Novice Teachers or Lack a College Major in the Subject
That They Teach, by Self-Reported Highly Qualified Status, 2004-05
159
Exhibit C-23
Percentage of Teachers Participating in Professional Development Focused on
Instructional Strategies for Reading and Mathematics, 2003-04
Exhibit C-24
Percentage of Teachers Participating in Professional Development Focused on
In-Depth Study of Topics n Reading and Mathematics, 2003-04
160
Endnotes
161
1
U.S. Department of Education, Budget Service.
2
Jay Chambers, Joanne Lieberman, Tom Parrish, Daniel Kaleba, James Van Campen, and Stephanie Stullich
(2000). Study of Education Resources and Federal Funding: Final Report. Washington, DC: U.S. Department
of Education, Office of the Under Secretary, Planning and Evaluation Service. The estimate of the percentage
of public schools receiving Title I funds was updated based on the number of Title I schools reported on
Consolidated State Performance Reports for 2002-03 divided by the total number of public elementary and
secondary schools in 2001-02 from the NCES Common Core of Data.
3
Beth Sinclair (forthcoming). State ESEA Title I Participation Information for 2002-03: Final Summary Report.
Washington, DC: U.S. Department of Education, Office of Planning, Evaluation, and Policy Development, Policy
and Program Studies Service.
4
Beth Sinclair (forthcoming). State ESEA Title I Participation Information for 2002-03: Final Summary Report.
Washington, DC: U.S. Department of Education, Office of Planning, Evaluation, and Policy Development, Policy
and Program Studies Service.
5
U.S. Department of Education, National Center for Education Statistics, unpublished data on averaged
freshman graduation rates. U.S. Department of Education, Policy and Program Studies Service analysis of
state-reported graduation rates from Consolidated State Performance Reports and State Education Agency
websites. State-reported rates for 2003 or 2004 were used for 16 states where 2002 rates were not available.
6
Consolidated State Performance Reports, 2003-04.
7
Source: U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study
of State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
8
Consolidated State Performance Reports, 2003-04.
9
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
10
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
11
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
12
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
13
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
14
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
15
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality under No Child Left Behind.
16
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
17
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
18
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
19
Center on Education Policy (2005). From the Capital to the Classroom: Year 3 of the No Child Left Behind
Act. Washington, DC: Center on Education Policy.
20
Christine Padilla, Katrina Woodworth, Andrea Lash, Patrick M. Shields, and Katrina G. Laguarda (2005).
Evaluation of Title I Accountability Systems and School Improvement Efforts: Findings From 2002-03.
Washington, DC: U.S. Department of Education, Office of Planning, Evaluation, and Policy Development, Policy
and Program Studies Service.
21
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
22
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
23
Policy and Program Studies Service monthly reviews of State Education Agency Web sites, conducted by
Westat from May 2003 through May 2005.
24
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
25
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
26
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
27
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
28
Consolidated State Performance Reports, 2003-04.
29
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
30
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
31
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
32
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
33
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
34
Section 1501 of the Elementary and Secondary Education Act, as reauthorized by the No Child Left Behind
Act.
35
The National Longitudinal Study of No Child Left Behind is being conducted by the RAND Corporation in
collaboration with the American Institutes for Research and the National Opinion Research Center. Response
rates for the 2004-05 surveys that have been completed are 96 percent for the school district survey,
89 percent for the principal survey, 84 percent for the teacher surveys, and 87 percent for the Title I
paraprofessional survey. Other components of the study, including surveys of parents and supplemental
service providers, are still in progress.
36
The Study of State Implementation of Accountability and Teacher Quality Under No Child Left Behind is
being conducted by the American Institutes for Research in collaboration with the Council of Chief State
School Officers and REDA International. Interviews were completed for all 50 states plus the District of
Columbia and Puerto Rico.
37
The Study of Title I Accountability Systems and School Improvement Efforts is being conducted by SRI
International. Survey response rates ranged from 88 to 90 percent for the school district survey and from 83
to 85 percent for the principal survey.
38
The Case Studies of the Early Implementation of Supplemental Educational Services are being conducted
by Policy Studies Associates.
39
Beth Sinclair (forthcoming). State ESEA Title I Participation Information for 2002-03: Final Summary Report.
Washington, DC: U.S. Department of Education, Office of Planning, Evaluation, and Policy Development, Policy
and Program Studies Service.
40
A separate program, Title I Part D, serves students in state institutions for neglected and delinquent
children and youth.
41
Beth Sinclair (forthcoming). State ESEA Title I Participation Information for 2002-03: Final Summary Report.
Washington, DC: U.S. Department of Education, Office of Planning, Evaluation, and Policy Development, Policy
and Program Studies Service. U.S. Department of Education, National Center for Education Statistics.
Common Core of Data (CCD), “State Nonfiscal Survey of Public Elementary/Secondary Education, 2001-02.”
Ungraded students are not included in these calculations; they account for one percent of both Title I
participants and all public school students.
42
Beth Sinclair (forthcoming). State ESEA Title I Participation Information for 2002-03: Final Summary Report.
Washington, DC: U.S. Department of Education, Office of Planning, Evaluation, and Policy Development, Policy
and Program Studies Service.
43
In states that did not assess students in 4th-grade reading, a nearby grade was used (3rd grade for Arizona,
Delaware, the District of Columbia, Hawaii, Illinois, Indiana, Maryland, Minnesota, Missouri, New Hampshire,
Oregon, Tennessee, and Virginia, and 5th grade for Kansas, Oklahoma, and Pennsylvania).
44
State participation in NAEP was not required prior to NCLB, and as a result, NAEP results for years prior to
2003 are based on a subset of the states. For example, for the 4th-grade NAEP reading assessment, 39 states
participated in 1998 and 44 states participated in 2002, compared with 51 states in 2003.
45
NAEP changed its approach to testing accommodations for students with disabilities an LEP students during
the period examined in this report. Before 1996, no testing accommodations were provided to such students
participating in NAEP assessments. Beginning in 1996 for the mathematics assessment and 1998 for the
reading assessment, the Main NAEP was administered to two reporting samples—“accommodations
permitted” and “accommodations not permitted.” Beginning in 2002 for reading and 2003 for mathematics,
NAEP administered the Main NAEP test with “accommodations permitted” as its only administration
procedure. For the National Assessment of Title I interim report, we report Main NAEP results with no
accommodations up through 1994 and with accommodations permitted thereafter. For the Trend NAEP, 2004
was the first year that accommodations were permitted, but a sample was also assessed with “no
accommodations permitted”; we present Trend NAEP results with no accommodations for the full time period
examined in this report.
46
National Institute of Statistical Sciences, Education Statistics Services Institute (2005). Task Force on
Graduation, Completion, and Dropout Indicators (NCES 2005-105). Washington, DC: U.S. Department of
Education, National Center for Education Statistics.
47
U.S. Department of Education, Policy and Program Studies Service analysis of state-reported graduation
rates from Consolidated State Performance Reports (based on reports that have not yet been verified) and
State Education Agency Web sites.
48
Marilyn Seastrom, Lee Hoffman, Chris Chapman, and Robert Stillwell (2005). The Averaged Freshman
Graduation Rate for Public High Schools From the Common Core of Data: School Years 2001-02 and 2002-03
(NCES 2006-601). Washington, DC: U.S. Department of Education, National Center for Education Statistics.
49
For 16 states, state-reported graduation rates for 2002 were not available and 2003 or 2004 graduation
rates were used instead; these 16 states are: Alaska, California, Connecticut, Idaho, Massachusetts,
Minnesota, Mississippi, Nevada, New Hampshire, North Carolina, Oregon, Rhode Island, South Carolina, South
Dakota, Tennessee, and West Virginia.
50
OESE review of Consolidated State Performance Reports and State Education Agency Web sites.
51
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
52
U.S. Department of Education, Office of Elementary and Secondary Education, review of Consolidated State
Performance Reports and State Education Agency Web sites.
53
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
54
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
55
U.S. Department of Education, Office of English Language Acquisition, Language Enhancement, and
Academic Achievement for Limited English Proficient Students (2003). Part II: Final Non-Regulatory Guidance
on the Title III State Formula Grant Program -Standards, Assessments, and Accountability. Available online at
www.ed.gov/programs/nfdp/NRG1.2.25.03.doc.
56
Ellen Schiller, Ellen Bobronnikov, Fran O’Reilly, Cristofer Price, and Robert St.Pierre (2005), The Study of
State and Local Implementation and Impact of the Individuals with Disabilities Education Act: Final 2nd
Interim Report (2002-2003 School Year), Bethesda, MD: Abt Associates, available at
www.abt.sliidea.org/Reports/complete%20Interim%20Report%2003-05%20for%20web.pdf (see Exhibit 6.2 on
page 105). The Study of State and Local Implementation and Impact of the Individuals with Disabilities
Education Act (SLIIDEA) focuses on policies, practices and resources used to implement the goals set forth in
IDEA. The SLIIDEA study is collecting data over a 6-year period through mail surveys at the state, district, and
school levels as well as a set of case studies. The study sample includes all 50 states and the District of
Columbia and a nationally representative sample of 959 school districts and 4,448 schools within those
districts. The school response rate was 74 percent.
57
SRI International (2004). Facts from OSEP's National Longitudinal Studies: Standardized Testing Among
Secondary School Students with Disabilities. Menlo Park, CA: Author. Available at
www.nlts2.org/pdfs/fact_sheet4%20_05_04.pdf. These data are from the National Longitudinal Transition
Study 2 (NLTS2), a study of the experiences of a national sample of students receiving special education who
were 13 to 16 years of age in 2000 as they move from secondary school into adult roles. The NLTS2 sample
draws from a nationally representative sample of LEAs and a sample of state-supported special schools. The
initial sample was approximately 11,500 students. For the data reported here, almost 6,000 parents and
guardians completed phone interviews in 2005, for a 70% response rate. Almost 3,000 youth interviews also
took place. More information on the methodology for this study is presented in Mary Wagner, Camille Marder,
Jose Blackorby, Renee Cameto, Lynn Newman, Phyllis Levine, Elizabeth Davies-Mercier, et al. (2003), The
Achievements of Youth With Disabilities During Secondary School: A Report From the National Longitudinal
Transition Study-2, Menlo Park, CA: SRI International.
58
Data for all subgroups were not available for Louisiana and Utah. Arkansas and Kentucky did not report
participation rates for Native American/Alaska Native students. Alabama did not report on the participation of
limited English proficient students. Vermont did not report on the participation of Hispanic students. Arizona,
Arkansas, and Indiana did not report on the participation of low-income students. Minnesota and New Mexico
reported 100 percent participation for all students, but did not report participation rates for subgroups. The
subgroup participation rates in Minnesota and New Mexico were assumed to be 100 percent, since the
participation rate of all students was 100 percent in both states.
59
U.S. Department of Education, Office of English Language Acquisition (2005). Biennial Evaluation Report to
Congress on the Implementation of the State Formula Grant Program, 2002-2004, English Language
Acquisition, Language Enhancement and Academic Achievement Act (ESEA, Title III, Part A), Table 2.3.
Washington, D.C: U.S. Department of Education.
60
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
61
Education Week, Quality Counts, 2003 through 2005 annual reports, Standards and accountability tables.
62
Data were unavailable for seven states. Additionally, one state did not report on all three trend indicators,
one state did not report on presenting results for each classroom, and one state did not report on presenting
results for trends in subgroups within a school.
63
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
64
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
65
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
66
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
67
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
68
Another 503 identified schools were of other types. The sum across levels is somewhat less than the total
number of identified schools reported earlier due to some missing data on school grade level. U.S.
Department of Education, Policy and Program Studies Service, unpublished data from the Study of State
Implementation of Accountability and Teacher Quality Under No Child Left Behind.
69
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
70
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
71
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
72
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
73
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
74
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
75
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
76
Data in Exhibit 34 are based on data provided by 33 states. Data on the percentage of schools missing AYP
that did not meet achievement targets for all students in reading or math is also available for a larger set of
states (48), and this figure is consistent with the Exhibit 34 figure for this group (34 percent based on the 48
states, vs. 38 percent based on the 33 states).
77
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality under No Child Left Behind. Findings involving
subgroups are based on data that was available from 33 states. There is some variation in findings involving
differing numbers of states.
78
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind. Findings are based
on data from 48 states.
79
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind. Data for calculating
percentages for limited English proficient students and students with disabilities were not available.
80
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
81
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
82
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
83
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind. Findings on
elementary, middle, and high schools meeting AYP for the other academic indicator are based on data that
were available from 36 states.
84
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
85
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
86
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
87
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
88
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind. Patrick M. Shields, Camille Esch, Andrea Lash, Christine Padilla,
Katrina Woodworth, Katrina G. Laguarda, and Nicholas Winter (2004). Evaluation of Title I Accountability
Systems and School Improvement Efforts (TASSIE): First-Year Findings. Washington, DC: U.S. Department of
Education, Office of the Under Secretary, Policy and Program Studies Service.
89
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
90
Christine Padilla, Katrina Woodworth, Andrea Lash, Patrick M. Shields, and Katrina G. Laguarda (2005).
Evaluation of Title I Accountability Systems and School Improvement Efforts: Findings From 2002-03.
Washington, DC: U.S. Department of Education, Office of Planning, Evaluation, and Policy Development, Policy
and Program Studies Service.
91
Christine Padilla, Heidi Skolnik, Alejandra Lopez-Torkos, Katrina Woodworth, Andrea Lash, Patrick M. Shields,
Katrina G. Laguarda, and Jane L. David (forthcoming). Title I Accountability and School Improvement Efforts
From 2001 to 2004. Washington, DC: U.S. Department of Education, Office of Planning, Evaluation, and Policy
Development, Policy and Program Studies Service. Christine Padilla, Katrina Woodworth, Andrea Lash, Patrick
M. Shields, and Katrina G. Laguarda (2005). Evaluation of Title I Accountability Systems and School
Improvement Efforts: Findings From 2002-03. Washington, DC: U.S. Department of Education, Office of
Planning, Evaluation, and Policy Development, Policy and Program Studies Service.
92
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
93
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind. Data are not
included for two states.
94
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
95
Center on Education Policy. From the Capital to the Classroom: Year 3 of the No Child Left Behind Act.
Washington, DC: Center on Education Policy, 2005.
96
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
97
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
98
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
99
Christine Padilla, Heidi Skolnik, Alejandra Lopez-Torkos, Katrina Woodworth, Andrea Lash, Patrick M. Shields,
Katrina G. Laguarda, and Jane L. David (forthcoming). Title I Accountability and School Improvement Efforts
From 2001 to 2004. Washington, DC: U.S. Department of Education, Office of Planning, Evaluation, and Policy
Development, Policy and Program Studies Service.
100
Christine Padilla, Katrina Woodworth, Andrea Lash, Patrick M. Shields, and Katrina G. Laguarda (2005).
Evaluation of Title I Accountability Systems and School Improvement Efforts: Findings From 2002-03.
Washington, DC: U.S. Department of Education, Office of Planning, Evaluation, and Policy Development, Policy
and Program Studies Service.
101
Center on Education Policy (2005). From the Capital to the Classroom: Year 3 of the No Child Left Behind
Act. Washington, DC: Center on Education Policy.
102
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
103
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
104
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
105
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
106
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
107
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
108
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
109
Christine Padilla, Heidi Skolnik, Alejandra Lopez-Torkos, Katrina Woodworth, Andrea Lash, Patrick M. Shields,
Katrina G. Laguarda, and Jane L. David (forthcoming). Title I Accountability and School Improvement Efforts
From 2001 to 2004. Washington, DC: U.S. Department of Education, Office of Planning, Evaluation, and Policy
Development, Policy and Program Studies Service.
110
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
111
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
112
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
113
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
114
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
115
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
116
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
117
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
118
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
119
U.S. Department of Education, Office of English Language Acquisition (2005). Biennial Evaluation Report
to Congress on the Implementation of the State Formula Grant Program, 2002-2004, English Language
Acquisition, Language Enhancement and Academic Achievement Act (ESEA, Title III, Part A). Washington,
D.C.: Author.
120
U.S. Department of Education, Office of English Language Acquisition (2005). Biennial Evaluation Report
to Congress on the Implementation of the State Formula Grant Program, 2002-2004, English Language
Acquisition, Language Enhancement and Academic Achievement Act (ESEA, Title III, Part A). Washington,
D.C.: Author.
121
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
122
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
123
The eligibility counts are based on state-reported data, while the participation numbers are estimated from
district and principal survey responses.
124
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
125
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
126
U.S. Department of Education, National Center for Education Statistics, Common Core of Data, 2002-03.
Special tabulation conducted by RAND Corporation for the National Longitudinal Study of NCLB.
127
Policy and Program Studies Service monthly reviews of State Education Agency Web sites, conducted by
Westat from May 2003 through May 2005.
128
Policy and Program Studies Service, review of State Education Agency Web sites.
129
Leslie M. Anderson and Katrina G. Laguarda (2005). Case Studies of Supplemental Services Under the No
Child Left Behind Act: Findings From 2003-04. Washington, DC: U.S. Department of Education, Office of
Planning, Evaluation, and Policy Development, Policy and Program Studies Service.
130
For supplemental services, districts were asked if they used the strategy in either 2003-04 or 2004-05,
while for school choice, districts were asked separately about strategies used during each school year.
131
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
132
Leslie M. Anderson and Katrina G. Laguarda (2005). Case Studies of Supplemental Services Under the No
Child Left Behind Act: Findings From 2003-04. Washington, DC: U.S. Department of Education, Office of
Planning, Evaluation, and Policy Development, Policy and Program Studies Service.
133
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
134
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
135
Leslie M. Anderson and Katrina G. Laguarda (2005). Case Studies of Supplemental Services Under the No
Child Left Behind Act: Findings From 2003-04. Washington, DC: U.S. Department of Education, Office of
Planning, Evaluation, and Policy Development, Policy and Program Studies Service.
136
Bradford Chaney (1995). Student Outcomes and the Professional Preparation of 8th Grade Teachers.
NSF/NELS:88 Teacher Transcript Analysis. Rockville, MD: Westat. Dan D. Goldhaber and Dominic J. Brewer
(2000). “Does Teacher Certification Matter? High School Certification Status and Student Achievement” in
Educational Evaluation and Policy Analysis, 22: 129-145.
137
Section 9101(11) of the No Child Left Behind Act.
138
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
139
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
140
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
141
However, Montana is considering adoption of the Praxis II series; the state is currently piloting some of the
Praxis II tests.
142
Personal communication on November 29, 2005 with Rick Tannenbaum, Director of Assessment Design and
Scoring, Educational Testing Service.
143
Educational Testing Service, unpublished data provided on August 19, 2005. The national median scores
are based on scores of all individuals who took these tests from October 1, 2001, to July 31, 2004.
144
Educational Testing Service Web site, state requirements for Praxis II exams, accessed on September 8,
2005.
145
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
146
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
147
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
148
Pennsylvania’s HOUSSE for elementary teachers relies on the initial certification process, but the state’s
HOUSSE for secondary teachers is a point system.
149
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
150
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
151
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
152
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
153
Similarly, only 38 percent of middle school English teachers reported that they had a major in English,
compared with 81 percent of high school English teachers. Seastrom, Marilyn, Kerry Gruber, Robin Henke,
Daniel McGrath, and Benjamin Cohen (2004), Qualifications of the Public School Teacher Workforce:
Prevalence of Out-of-Field Teaching: 1987-88 to 1999-2000 (NCES 2002-603), Washington, DC: U.S.
Department of Education, National Center for Education Statistics. The sample for this survey included
56,354 teachers, with a response rate of 83 percent. More information on the methodology for this study can
be found at http://nces.ed.gov/surveys/SASS/methods9900.asp.
154
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
155
The term “classes” is used here because the data come from a survey of principals that asked secondary
principals about the number of secondary “class sections” taught by highly qualified teachers. The survey
asked elementary school principals about the number of teachers who were highly qualified; however, since
most elementary teachers teach only one class of students, we use the term “classes” here to describe both
the elementary and secondary data collected from principals.
156
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind. “High minority” schools are those in which 75 percent or more of
the students are minorities and “low minority” schools are those in which fewer than 35 percent of the
students are minorities.
157
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
158
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
159
Mary Kennedy (1998). Research Monograph No. 13: Form and Substance in Inservice Teacher Education.
Madison: National Institute for Science Education, University of Wisconsin. David K. Cohen, Milbrey W.
McLaughlin, & Heather C. Hill (1998). Instructional Policy and Classroom Performance: The Mathematics
Reform in California (RR-39). Philadelphia: Consortium for Policy Research in Education.
160
Michael S. Garet, Beatrice F. Birman, Andrew C. Porter, Laura Desimone, Rebecca Herman, and Kwang Suk
Yoon (1999). Designing Effective Professional Development: Lessons From the Eisenhower Professional
Development Program. Washington, DC: U.S. Department of Education, Office of the Under Secretary,
Planning and Evaluation Service.
161
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
162
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
163
Michael S. Garet, Beatrice F. Birman, Andrew C. Porter, Laura Desimone, Rebecca Herman, and Kwang Suk
Yoon (1999). Designing Effective Professional Development: Lessons From the Eisenhower Professional
Development Program. Washington, DC: U.S. Department of Education, Office of the Under Secretary,
Planning and Evaluation Service.
164
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
165
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
166
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
167
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
168
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
169
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
170
These findings are consistent with research by the Center on Education Policy (2005), From the Capital to
the Classroom: Year 3 of the No Child Left Behind Act, Washington, DC: Center on Education Policy.
171
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
172
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
173
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
174
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.
175
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the Study of
State Implementation of Accountability and Teacher Quality Under No Child Left Behind.
176
U.S. Department of Education, Policy and Program Studies Service, unpublished data from the National
Longitudinal Study of No Child Left Behind.