Emma C., et al v. Eastin, et al
Filing
2520
ORDER RE STATE'S COMPLIANCE AT PHASE 2. Signed by Judge Vince Chhabria on 7/5/2019. (vclc1S, COURT STAFF) (Filed on 7/5/2019)
UNITED STATES DISTRICT COURT
NORTHERN DISTRICT OF CALIFORNIA
EMMA C., et al.,
Case No. 96-cv-04179-VC
Plaintiffs,
ORDER RE STATE’S COMPLIANCE
AT PHASE 2
v.
TONY THURMOND, et al.,
Defendants.
After inheriting this consent decree from the previously assigned judge, the Court set up
a process for examining whether the State of California does a legally adequate job of
monitoring school districts for compliance with the Individuals with Disabilities Education Act
(IDEA). The process has four phases, the first of which is complete. That phase addressed the
state’s data collection activities, and the Court held that the state will – after correcting a couple
isolated but significant defects – be deemed in compliance with its federal statutory obligations
in this area.
Thus began the second phase, which examines how the state analyzes the data it collects
to determine which districts are in need of intervention, and to determine what type of
intervention is called for. The matters addressed in this phase are more central to the state’s
responsibility to monitor school districts to ensure that disabled children are receiving an
appropriate education as required by the IDEA.
Unfortunately, California’s system for flagging school districts for intervention is riddled
with serious defects. To give just three examples:
•
The state attaches the classification of “meets requirements” to hundreds of small school
districts despite conducting virtually no analysis of those districts.
•
The state takes minimal steps to flag problems with the delivery of services to preschoolaged children with disabilities, even while acknowledging the importance of early
intervention to IDEA compliance.
•
The state’s system for identifying the most troubled districts appears irrational, resulting
in the selection of districts for intensive intervention that are less in need of it than the
ones passed over.
It appears, at least so far, that current leaders in the Special Education Division of the
California Department of Education don’t deserve much of the blame for this. They inherited
many of the problems, candidly acknowledge most of them, and are committed to improving the
state’s monitoring activities. They have been forthcoming and cooperative, even while their
lawyers from the Attorney General’s Office have sometimes been nonresponsive or
obstructionist. Accordingly, for now, there is no reason to seriously consider holding the state in
contempt for failure to comply with the consent decree. Nonetheless, the defects in the current
system are so serious, and so numerous, that they significantly interfere with California’s ability
to monitor how school districts are serving disabled children. The state cannot move on to Phase
3 until it addresses – or shows that it’s well on its way to addressing – these problems.
I. BACKGROUND
A.
The IDEA requires states that accept federal assistance to provide a free and appropriate
education to all children with disabilities. 20 U.S.C. § 1412(a)(1)(A). As a practical matter, this
responsibility falls largely on individual school districts. In recognition of this, the statute
requires states, in turn, to conduct effective oversight of the school districts. Id. § 1416(a)(1)(C).
The consent decree in this case requires the state to demonstrate that it has an adequate oversight
system in place.1
1
As discussed in the May 18, 2018 order, the precise question is whether the state has a
monitoring and enforcement system that is adequate with respect to Ravenswood City School
District. The state has explained, however, that it intends to establish compliance with the
consent decree with reference to its normal statewide system for monitoring school districts,
2
This Court’s May 18 order established a four-phase monitoring process for assessing the
legal adequacy of the state’s oversight system. Dkt. No. 2387 at 1–2. Phase 1 involved scrutiny
of the data the state collects from school districts. Specifically, the inquiry was whether the state
collects enough data, and the right types of data, to enable it to effectively monitor districts. The
Court concluded that although isolated legal deficiencies in the state’s data collection system
must be addressed, the state was largely compliant in this area. The Court thus concluded that the
state could move on to Phase 2 and could cure the isolated data collection deficiencies in the
subsequent phases.
Phase 2 – the current phase – examines how the state analyzes the data it collects. To
satisfy its monitoring obligations under the IDEA, the state must do an adequate job of flagging
the districts that are failing, or struggling, to provide an appropriate education to students with
disabilities. And it must do an adequate job of deciding what type of intervention is necessary in
a given district. Phase 3 will examine how the state, after it has decided which districts to select
for further monitoring, actually executes that monitoring. Phase 4 will examine the state’s
written policies and directions governing school district compliance with the IDEA.
Phase 2 essentially poses two questions. The first is whether the state translates the data
it collects into metrics capable of identifying school districts that may be falling short on their
obligation to provide an appropriate education to students with disabilities. The second question
is whether the state’s methods for sorting districts for inclusion in (and exclusion from) its
various monitoring activities is adequate.
In addition, the IDEA’s implementing regulations require states to issue annual
compliance determinations for school districts. 34 C.F.R. § 300.600(a)(2). The state’s process for
selecting districts for monitoring also determines whether the districts are classified by the state
as complying with federal law. School districts that are selected for monitoring are classified as
rather than developing a narrower remedy that would apply to Ravenswood only. See Dkt. No.
2387 at 1. Because Ravenswood primarily serves preschool students and students in grades
kindergarten through eighth grade, this order does not consider elements of the state’s
monitoring program that would apply, for example, exclusively to high school students.
3
“needs assistance,” “needs intervention,” or “needs substantial intervention” if problems persist.
States must prohibit these school districts from reducing their “maintenance of effort,” meaning
their allocation of non-federal funds for special education services. 34 C.F.R. § 300.608(a).
School districts that the state decides are not in need of any monitoring are typically classified as
“meets requirements.” Phase 2 therefore inherently includes an inquiry into whether the state’s
sorting methods adequately ensure that districts labeled as “meets requirements” do not suffer
serious deficiencies in serving disabled children.
Phase 2 took roughly the same format as Phase 1. The Court received written
submissions from the parties, a report from the court monitor outlining his conclusions, and an
amicus brief from the Morgan Hill Concerned Parents Association. The written submissions
were followed by two days of evidentiary hearings. Kristen Wright, the Director of the Special
Education Division for the California Department of Education, Shiyloh Duncan-Bercerril, the
Division’s Education Administrator, and Alison Greenwood, the Division’s Quality Assurance
Administrator, testified during Phase 1 and returned to testify for Phase 2. After the hearings, the
Court ordered the monitor to conduct supplemental data analyses and ordered the state to provide
all data necessary for those analyses. The monitor and his data consultant, Dr. Susan Wagner,
presented their conclusions at a third evidentiary hearing, and state policymakers offered further
testimony in response.2 The monitor then conducted a final set of data analyses.
B.
As explained in previous orders, the purpose of this oversight is to ensure that the state
complies with federal law. The IDEA and its implementing regulations offer only general
guidance for what states must do to satisfy their monitoring and enforcement obligations. The
state’s monitoring activities must focus on “improving educational results and functional
outcomes for all children with disabilities,” and ensuring that states meet the IDEA’s
requirements, with a special focus on “priority areas” enumerated in the statute. 20 U.S.C.
2
Alison Greenwood did not testify at the third evidentiary hearing. Stacy Wedin, a policy
consultant in the state’s Special Education Division, testified in her place.
4
§ 1416(a)(2), (a)(3). The state must use “quantifiable indicators and such qualitative indicators as
are needed to adequately measure performance in the priority areas.” 34 C.F.R. § 300.600(c).
When a district appears from the data to fall short on its obligations, the state must respond with
an appropriate enforcement action to correct the noncompliance “as soon as possible.” 34 C.F.R.
§ 300.600(e). The IDEA, however, does not require states to adopt any particular approach for
monitoring. By specifying the ends, but leaving the means to the states, the IDEA strikes a
balance between federal authority and the states’ historic discretion in the design and control of
their education systems.
For purposes of this consent decree, however, the standard for compliance is relatively
straightforward. The state will not be found to be out of compliance simply because the plaintiffs
or the court monitor have identified isolated deficiencies or ways in which the monitoring system
could become more effective. But if the state’s chosen procedures are so deficient that they
significantly hinder its ability to monitor school districts, the state will not be found compliant
merely because the statute does not expressly forbid those choices.
This standard has important implications for this phase. Any evaluation of the state’s data
analysis activities must pay close attention to how that data is actually used. Because certain
problems may occur in tandem – for example, districts that frequently suspend students with
disabilities may also have poor performance on statewide assessments – different metrics may
yield similar information. Therefore, an imprecise metric may not compromise the overall
monitoring system. Further, the importance of precision may turn on the importance of the
metric. The state need not use the most granular data possible if the data is less important in the
context of the overall system, or if the state has set targets for that data that sufficiently ensure
that any poor performance will lead to monitoring.
These realities underscore the importance of reviewing the state’s data analysis activities
as a whole. No monitoring system is perfect. Identifying a theoretical concern with an individual
metric is not sufficient to find the state out of compliance with federal law. It must be shown that
the concern, either alone or in combination with other issues, significantly interferes with the
5
state’s ability to evaluate school districts and to identify the ones in need of intervention.
C.
As discussed in detail in the Phase 1 order, during the school year, school districts submit
large swaths of data about all students, with and without disabilities, to databases maintained by
the state. See Dkt. No. 2428 at 3. This process constitutes the “first tier” of the state’s monitoring
system.
The state then translates those data into metrics that align with key directives of the
IDEA. For example, one directive is that schools educate students with disabilities in the “least
restrictive environment.” 20 U.S.C. § 1412(a)(5). In general, this means that students with
disabilities must be taught in general education classrooms, alongside their nondisabled peers, as
often as reasonably possible. Schools may only remove students with disabilities from general
education classrooms when the nature or severity of the child’s disability requires it. To evaluate
district performance in this area, the state uses three different metrics. For school-age children,
the state calculates the percentages of disabled students who are taught (i) in general education
classrooms for greater than 80% of the day; (ii) in general education classrooms for less than
40% of the day; and (iii) in separate placements. For preschool-aged children, the state looks to
the percentages of disabled students in (i) general early-childhood programs; and (ii) separate
placements. Dkt. No. 2455-1 at 11–12, 14. The state compares school districts’ performance on
each metric to a target that policymakers set in consultation with various groups, including local
administrators, parent groups, and advocacy groups. Dkt. No. 2455-1 at 4.
In other areas, the state uses a system called the “Dashboard.” The Dashboard is a visual
depiction of district performance on a grid. The vertical axis of the grid represents the school
district’s “status,” or the district’s current performance. The horizontal axis on the grid represents
the degree to which the school district improved or regressed from the previous year. Combining
information from both axes results in a color: red, orange, yellow, green, or blue. An example
from the state’s written submissions is included here:
6
Dkt. No. 2455-1 at 24.
This is the Dashboard that the state uses to evaluate disabled students’ performance on
statewide assessments. The state uses a metric called “distance from standard.” See Dkt. No.
2455-1 at 22–23. For each student, the state calculates the distance between the student’s score
and the score needed to establish that the student met the academic standards relevant to that
assessment. The state then calculates the average “distance from standard” for each school
district and reduces the score if the district failed to ensure that enough disabled students actually
took the assessment. The circled squares show that districts with a distance from standard of 10–
44.5 points that improved their performance between 3–14 points from the prior year will receive
a “green.” Districts that receive a “red” or “orange” are treated as missing the target for this
metric. In addition to performance on statewide assessments, the state currently uses the
Dashboard to evaluate districts’ suspension practices.
D.
After the data are translated into a set of metrics and compared to targets, the next step is
to determine whether a district should be chosen for further monitoring, and if so, what type of
monitoring activity to conduct. The degree to which the state effectively executes its monitoring
7
activities will be explored in depth during Phase 3. But evaluating how the state decides which
monitoring activities to conduct, and for which districts, requires at least some understanding of
the substance of those monitoring activities and how they differ from one another.
In general, a district that performs poorly on any individual metric will be selected for a
type of targeted monitoring. A district that performs poorly across many different metrics, or in
certain instances, the same metrics for several consecutive years may be selected for more
intensive monitoring. In all, the state currently performs five types of monitoring relevant to this
case – three targeted monitoring activities that relate to a specific area of poor performance by a
district, and two intensive monitoring activities that involve greater intervention.3
The state has dubbed the three targeted monitoring activities “performance indicator
review,” “data identified noncompliance review,” and “disproportionality review.” A district is
selected for performance indicator review if it fails to meet a target for a particular metric that
the state believes is closely tied to student outcomes. Dkt. No. 2506 at 6. For example, a district
that receives a red or orange on the Dashboard for poor performance on statewide assessments
by disabled students, or for excessive suspension of disabled students, will be selected for
performance indicator review in the pertinent area. A district selected for review must perform a
“root cause” analysis into the reasons for its inadequate performance in that area and must
submit an improvement plan to the state. Dkt. No. 2501 at 7–9. The state does not formally
3
The monitoring activities discussed in this order exist alongside other accountability measures
developed by the state. For example, state officials testified during the hearings that the
California Department of Education may separately offer school districts “differentiated
assistance” based on some of the same performance criteria that it uses in determining whether
that district is providing an adequate education to disabled students. Dkt. No. 2506 at 120–21. In
addition, a significant portion of school districts receiving differentiated assistance do so
specifically because of their outcomes for students with disabilities. Dkt. 2478-1 at 4. Because
differentiated assistance is not administered by the Special Education Division of the Department
of Education, which is the division responsible for IDEA compliance and the division whose
activities are being scrutinized here, the state provided little information about it in its written
submissions. Further, the information presented at the evidentiary hearings does not explain in
detail the forms of assistance that these districts may receive, and leaves it unclear whether those
forms of assistance may assist the state and school districts in complying with the IDEA. If the
state believes that a fuller picture of differentiated assistance would assist the Court in evaluating
the state’s monitoring systems, then it should provide more information in future submissions for
Phase 2 and/or Phase 3, when the inquiry turns to the content of the state’s monitoring activities.
8
supervise the district’s implementation of the plan but looks to see whether the targets are met
for the following year. If the district remains in performance indicator review for multiple years,
it must perform a “record review,” meaning it must select ten individual student records and
review them for any patterns related to the noncompliance. Dkt. No. 2506 at 161–63.
The second targeted monitoring activity, which has the unfortunate name of data
identified noncompliance review, addresses poor performance on metrics that are related to
timeliness. For example, before a child may receive special education services, the district must
conduct an initial evaluation to determine whether the child has a disability. The IDEA requires
districts to conduct these evaluations within 60 days after receiving parental consent. 20 U.S.C.
§ 1414(a)(1)(C)(i)(I). The state flags districts for monitoring whenever they fail to meet this and
any other statutory deadline for any student. The district must correct all identified
noncompliances (for example, by holding any past-due initial evaluations). The district must also
submit a root cause analysis that explains why the district has missed the deadline (or deadlines)
and must submit data to the state demonstrating that it has corrected the problem. If the district
continues to miss deadlines after that, the state may order it to take further corrective actions, and
may even withhold funding to the district. Dkt. No. 2506 at 24.
The third form of targeted monitoring, disproportionality review, relates to the IDEA’s
requirement that states prevent discrimination against children on the basis of race or ethnicity.
20 U.S.C. § 1418(d). To use one concrete example, school districts must take care to avoid
suspending Hispanic disabled children at a disproportionally higher rate than all disabled
children who are not Hispanic. See 34 C.F.R. § 300.646(a)(3). The state is given the discretion to
determine how marked the discrepancy must be before districts become “significantly
disproportionate.” 34 C.F.R. § 300.647(a)(7). Once selected, the state evaluates the districts’
policies and procedures and a sample of student records for compliance with federal law (for
example, to ensure that all IEP meetings were conducted with a general education teacher
present). Dkt. No. 2455-1 at 54; Dkt. No. 2506 at 25–26. The district must correct all
noncompliance identified by the state.
9
The state’s two intensive monitoring activities are comprehensive review and significant
disproportionality review. Comprehensive review is reserved for districts experiencing the most
serious performance issues. The decision to select districts for comprehensive review is based on
a total score derived from performance on many different metrics. Poor performance on any
individual metric will, in addition to potentially leading to selection for targeted review, cause
the district to lose points from its total score. Dkt. No. 2469 at 57. After all districts are scored,
the state sets a cut score and selects all districts falling below that score for comprehensive
review. In contrast to performance indicator review and data identified noncompliance review –
which largely involve self-analysis by the district – the state puts its own boots on the ground
during comprehensive review. The state, rather than the district, develops the monitoring plan
after analyzing the district’s compliance history. It will interview parents and administrators and
conduct staff trainings as needed. A district selected for comprehensive review may remain in
monitoring for several years. Dkt. No. 2506 at 30–34.
The state’s other intensive monitoring activity, significant disproportionality review,
applies to districts that California has determined to be “significantly disproportionate” in any
area for three consecutive years. During this monitoring activity, the state takes an even harder
look at the district’s policies and practices. In addition to the activities performed during targeted
disproportionality review, districts in intensive monitoring must develop an improvement plan
that the state dubs an “early intervening services plan,” which must be approved by the
California Department of Education. The state also makes various forms of technical assistance
available to districts, both through the state directly and through contractors, to assist the district
in developing their plan and getting back into compliance. Districts must set aside 15% of their
IDEA-based funds to finance the implementation of the plan and must report on their progress to
the state. Dkt. No. 2506 at 28–30.
II. OBVIOUS DEFECTS IN THE STATE’S DATA ANALYSIS SYSTEM
The state’s process of determining which school districts need further monitoring, as well
as its process for deciding how intensive that monitoring should be, is riddled with defects. Some
10
of these defects are small and easily fixable, but several are so fundamental, or so obviously
contrary to the IDEA (or both), that the state cannot get out from under this consent decree
without fixing them. Furthermore, in contrast to Phase 1– where the state largely established
compliance, such that it made sense to move on to Phase 2 of the monitoring process concurrent
with the state’s efforts to fix a couple of outstanding legal defects in its data collection program –
the flaws in the state’s system for identifying districts in need of intervention are so severe that
they preclude a transition to the next phase. Before moving on to Phase 3, the state will be
required to demonstrate that it has meaningfully addressed these problems (or at least that it’s
well into the process of addressing them).
A.
As a preliminary matter, there are several areas where the state essentially conducts no
assessment of school district performance at all, even though the IDEA requires it. The
policymakers testified at the hearings that they are currently developing data analysis procedures
for almost all of these areas. But there is nothing in place now.
Small school districts. The state requires school districts to have a minimum number of
disabled students (or “n-size”) before assessing those districts with the Dashboard and with most
other metrics. See Dkt. No. 2478-3 at 2. Almost all of the state’s metrics (including all metrics
for academic performance, least restrictive environment, discipline, and disproportionality) have
an associated minimum n-size. The only real exception is for the metrics relating to timeliness.
Thus, the state is conducting no actual substantive evaluation of small districts. For example, the
state did not evaluate performance on English and Language Arts and Mathematics assessments
for school districts serving approximately 60,000 disabled students. It did not examine
suspension practices for districts serving approximately 40,000 disabled students. Dkt. No. 2494,
Exhibit 1 at 6.
Because these small districts are not being assessed on many metrics, they are
functionally exempt from many targeted and intensive monitoring activities. During the hearings,
state officials testified that over 1,000 school districts – comprising roughly 600 charter schools
11
(which the state treats as separate school districts for its data analysis activities) and roughly 400
small school districts – had fewer than thirty disabled students. Dkt. No. 2507 at 36; Dkt. No.
2494, Exhibit 1 at 6. Consequently, these districts could not be selected for targeted review for
academic assessments or for their suspension rates and could not be evaluated for comprehensive
review based on their performance on those metrics. See Dkt. No. 2478-3 at 2.
Moreover, it bears recalling that the state classifies any district not selected for
monitoring activities as “meets requirements” for the purposes of the state’s federal reporting
obligations. Therefore, these small districts are particularly likely to be classified by the state as
compliant with the IDEA irrespective of their actual performance. Of the 499 school districts not
selected for any monitoring process and determined by the state to meet requirements, only 132
had a census count of greater than 30 students with disabilities. Dkt. No. 2516 at 8. In other
words, the state is telling the federal government that hundreds of school districts “meet
requirements” under the IDEA even though the state has effectively exempted them from
monitoring.
To be clear, the state has identified sound reasons for excluding small districts from the
data analyses it ordinarily performs. A small number of disabled students in a given district may
cause that district’s metrics to vacillate wildly based on the performance of just a small handful
of students (or even just one student). Dkt. No. 2506 at 202. But this cannot account for the
state’s failure to develop any alternative protocol for assessing these districts, given that the
IDEA requires states to monitor all districts. The policymakers acknowledged this at the
hearings, and they testified that they have already begun researching methods of closing this gap,
including by aggregating small districts – perhaps with similar features or within the same
county – into larger units for data analysis purposes. There may be several different ways to
adequately include these districts, and how best to do so is a question for the policymakers. For
now, it’s enough to say that the state’s failure to include them is a failure to comply with its
monitoring obligations under the IDEA.
Preschool review. There are roughly 90,000 preschool-aged children with disabilities in
12
California. Dkt. No. 2478-1 at 23. A district that does poorly at placing preschool-aged disabled
children in regular early-childhood programs may be selected for targeted performance indicator
review. If that district does not put individualized education programs (“IEPs”) in place in a
timely manner for very young children transitioning into early-childhood programs from early
intervention programs (the “Part C to Part B” transition), or fails to conduct initial evaluations on
time, that district will be selected for data identified noncompliance review. Finally, a district
that performs very poorly with respect to preschoolers and school-aged children may be selected
for comprehensive review, although, as discussed in the next section, the state’s selection
formula makes this somewhat unlikely. This represents the full extent of the state’s monitoring
for preschool-aged children.
State policymakers candidly testified that a separate preschool-related monitoring activity
is necessary to close this gap, because preschools face unique issues that cannot be adequately
addressed through other monitoring activities. Dkt. No. 2506 at 93–94. The state is currently in
the process of designing such a system, on similar lines as comprehensive review. In addition to
looking at districts’ preschool placements and timeliness in carrying out the Part C to Part B
transition, the state would consider districts’ suspension practices, participation on preschool
assessments, and performance on those assessments in selecting districts for review.
In its written submissions, the state described a selection methodology for identifying
preschools for intensive monitoring, and described a set of monitoring activities, which at times
made it seem as if this system were already in place. See, e.g., Dkt. No. 2455-1 at 10 (“[T]he
State addresses [least restrictive environment] in a number of monitoring activities, including
Performance Indicator Review, Preschool Monitoring … and Comprehensive Review.”), 52
(“The Preschool Review examines issues related to placement, suspension, child find, and the
provision of FAPE for students 3 through 5 years of age.”); Dkt. No. 2478-1 at 8 n.16 (“[The
state] identified . . . 39 [school districts] for Preschool Review.”). It did not become apparent
until the hearings that this monitoring activity does not yet exist, in any form. As with the small
school districts, the state’s failure to adequately monitor preschool children puts it out of
13
compliance with federal law.4
Mediation. The IDEA requires school districts to establish mediation procedures. 20
U.S.C. § 1415(e). In mediation, an impartial party helps to resolve disputes between parents and
the school district regarding the district’s obligation to provide special education services. The
IDEA requires states to pay for the costs of mediation and to prioritize mediation in their
monitoring activities. 34 C.F.R. §§ 300.506(b)(4), 300.600(d)(2). But the state does not assess
districts’ mediation practices. This is true even though the state collects data regarding districts’
mediation practices and reports the percentage of mediations resulting in successful agreements
to the federal government. Dkt. No. 2390-1 at 40–41.
The state defends this by noting that mediation is a voluntary process that requires the
consent of both the parents and the school district, which means that the state cannot force
recalcitrant districts to mediate. Even assuming that’s true on a case-by-case basis, it’s far from
obvious that the state couldn’t take action against districts that refuse to use mediation as a
blanket matter (or, for that matter, seriously neglect this tool). Nor does it follow from the
voluntary nature of mediation that the state need not analyze whether districts are doing a good
enough job of using it. As the drafters of the IDEA have signaled, mediation is a useful
alternative for parents who lack the resources to hire a lawyer to assist with a due process
complaint, or who prefer more informal and nonadversarial methods for resolving disputes.
Analyzing mediation data could therefore reveal trends not separately captured by an analysis of
formal due process complaints or complaints made to the state. The state could also combine
mediation data with other data – for example, districts’ suspension rates or assessment
outcomes – to form a more complete picture of district performance.
To be fair, the state’s failure to assess districts for their use of mediation is less
significant than some of the other problems discussed in this section. There is even an argument
It would be premature to make any formal conclusion regarding the state’s anticipated selection
methodology before any monitoring system is actually in place, but it’s worth noting that the
state’s proposed formula closely mirrors the formula used for comprehensive review, which, as
discussed in subsection II.B, contains serious flaws.
4
14
that the failure to analyze mediation data does not, on its own, significantly interfere with the
state’s ability to fulfill its monitoring obligations under the IDEA. But in contrast to some of the
other areas discussed in this ruling, the monitoring obligation in the statute is specific rather than
general. Federal regulations require states to prioritize mediation in their monitoring activities.
See 34 C.F.R. § 300.600(d)(2). Because the state does not do this, it is out of compliance with
federal law.
B.
The next two issues relate to areas where the state is conducting some analysis, but the
design of its processes prevents it from effectively monitoring school districts.
Targets. Another flaw in the state’s monitoring system relates to the performance targets
it has set for school districts. An adequate monitoring system requires adequate targets. If targets
are too modest, states may fail to identify districts that are falling short on their obligation to
provide an adequate education to disabled children.
While the IDEA grants states discretion in setting targets, that discretion is not limitless.
For example, the IDEA requires states to design a state performance plan, in which states report
to the federal government their progress in implementing the IDEA in school districts, using
metrics and calculation methodologies that the Department of Education selects. States are given
the discretion to set targets for those metrics, but the regulations specify that those targets must
be “measurable and rigorous” across all priority areas. See 34 C.F.R. § 300.601(a)(3). The state
uses many of the same metrics and targets both for monitoring districts and for the state
performance plan. Dkt. No. 2478-1 at 6–7. The targets are set on a cyclical basis, and the state
will revisit its targets in the coming months.
The problems with the state’s chosen targets are clearest in the area of least restrictive
environment, which is one of the IDEA’s “priority areas” for monitoring. 20 U.S.C.
§ 1416(a)(3)(A); 34 C.F.R. § 300.600(d)(1). As discussed previously, the state uses three metrics
to assess whether districts are placing school-aged disabled children in the least restrictive
environment (placement in general education classes for greater than 80% of the day, for less
15
than 40% of the day, and in separate schools). And it has two metrics for preschool-aged children
(placement in regular early childhood programs and in separate programs). In 2016, all five
targets for these metrics were below statewide performance. Dkt. No. 2469 at 23, 26–27. And in
many instances, statewide performance already fell significantly behind national performance.
Thus, even though the data suggest that California performs poorly relative to the rest of the
country at placing disabled students in general education placements, the state has set targets
even below its already-poor statewide performance levels. This is particularly concerning in an
area like least restrictive environment, because the state has elsewhere emphasized its strong
belief that improving placements for disabled students will also improve outcomes for those
students. See, e.g., Dkt. No. 2506 at 62–63. Setting low targets increases the likelihood that
districts will escape selection for targeted or intensive monitoring – and that districts will be
classified by the state as compliant with the IDEA – even if they fall significantly short in
placing disabled children in general education classes “to the maximum extent appropriate.” 20
U.S.C. § 1412(a)(5)(A).
The state’s targets for “child find” may be another example of flawed targets. The IDEA
requires states to ensure that all children with disabilities, including homeless children and
children attending private schools, are “identified, located, and evaluated.” 20 U.S.C.
§ 1412(a)(3)(A). Similar to least restrictive environment, the IDEA instructs states to prioritize
child find in their monitoring activities. 34 C.F.R. § 300.600(d)(2). The state considers districts’
identification rates – the number of children identified as needing special education services over
the total number of children served by the district. Districts that are more than two standard
deviations from the statewide identification rate are selected for monitoring. But California’s
identification rate is only 11.7% as compared to 13.2% nationally. Dkt. No. 2469 at 31. And in
the 2017–2018 monitoring cycle, setting the target two standard deviations below California’s
low mean resulted in a system that selected only those 38 districts (out of 1,296 districts) with an
16
identification rate below 3.6%.5 Dkt. No. 2455-1 at 35.
To be clear, in some instances it may be necessary to set modest targets to encourage
incremental but stable progress. The state has explained that it follows a model of “continuous
improvement,” which involves setting a baseline and encouraging development from that
baseline. Dkt. No. 2478-1 at 3, 5. And for purposes of these court proceedings, the question is
not whether the state has developed the best method for encouraging improvement (a question
best left for the policymakers), but only whether the state’s system is capable of identifying
districts that are out of compliance with federal law. But the monitor and plaintiffs have raised a
genuine concern that the state’s targets are so low that they significantly impede its ability to
identify districts experiencing serious performance issues.
The state first argued in its written submissions that all of its targets were adequate
because they were the result of an inclusive process that considered input from various
stakeholders. Dkt. No. 2478-1 at 5–7. That argument is unsatisfactory – the purpose of these
court proceedings (and of the IDEA) is to ensure that the state has an adequate monitoring and
enforcement system in place. Dkt. No. 2387 at 1. If the state’s performance targets are legally
deficient – that is, if they prevent the state from adequately monitoring school districts – they
cannot be defended by asserting that they were the result of a process that included many
stakeholders. During the hearings, however, the policymakers agreed that, at minimum, the least
restrictive environment targets are not adequate. They stated that these are not the targets they
would have chosen had they been in charge at the time they were set. Dkt. No. 2507 at 101–03.
The policymakers further confirmed that performance targets will be revisited in the coming
5
The plaintiffs and monitor also argue that the state is out of compliance because it only assesses
districts with respect to child find for students ages 6–21. The state is interested in developing a
measure for child find applicable to preschool-aged children, but finding an appropriate method
has proven elusive. Because there’s no universal preschool in California, the state cannot assess
how well districts are doing at identifying children with disabilities, because it lacks adequate
information about the total number of children from which disabled students are being identified.
The state has explained that there are projections available to estimate numbers of preschoolaged children based on many different variables, but those projections provide estimates by
county (rather than by school district) and are not sufficiently reliable for use in the state’s data
analysis activities. Dkt. No. 2507 at 132–38. This explanation is adequate.
17
months, and that they plan to adopt a different approach. Dkt. No. 2507 at 171.
The plaintiffs and monitor raised concerns with other targets set by the state. In response,
at the hearings the policymakers often responded with arguments and statistics that were not
raised in the state’s written submissions, for example, concerning the number of districts that
were selected using the relevant targets. Given that the state plans to revisit its performance
targets so soon, there’s no need to perform a target-by-target analysis. At this stage, what matters
is that the record suggests that the state’s targets do not enable it to adequately identify school
districts for further monitoring. The state’s new targets must be crafted with its monitoring
obligations in mind, and it must do a far better job of explaining and justifying its targets during
the renewed Phase 2.6
Comprehensive review. The next set of issues relate to the state’s method for selecting
districts for comprehensive review, the state’s intensive monitoring activity. One might object
that comprehensive review should receive less scrutiny for purposes of these proceedings,
because the state’s monitoring system is tiered. As a result, all of the school districts that are
under serious consideration for comprehensive review have already been selected for one or
more targeted monitoring activities. It is clear, however, from the state’s written submissions and
policymakers’ testimony during the evidentiary hearings that there is a significant difference
between the quality and depth of the state’s monitoring activities, as well as the resources
available to school districts, during targeted review versus intensive review. A closer look is
therefore needed to determine whether the state’s data analysis activities adequately sort districts
between the targeted monitoring activities and comprehensive review.
This look reveals two major flaws in the state’s current approach. The first is that the
state appears to select the wrong districts. Under the state’s formula, it appears that betterperforming districts will often be far more likely to get selected for comprehensive review than
The state need not, however, present further information about its targets for the IDEA’s
timeliness requirements or for its risk-ratio for purposes of disproportionality monitoring,
assuming those targets remain the same. As discussed later in this order, these targets do not
appear to present any cause for concern and the state’s explanations are adequate.
6
18
worse-performing districts, and the state has not provided an adequate explanation for the
discrepancy. Second, the state’s formula appears not to select enough districts for comprehensive
review. The state’s chosen metrics suggest that more districts need close scrutiny than currently
receive it.
The state uses a point system to select districts for comprehensive review. For each
metric, a district receives 1–4 points. Districts receive points for meeting the target, and more
points for improving their performance from the prior year. For example, recall that the state
evaluates districts’ efforts to place disabled children in the least restrictive environment in part
by calculating the percentage of disabled students in general education classes for greater than
80% of the day. The target for this metric is 51.2%. A district receives a “1” if it fails to meet the
target and declines in performance from the previous year (for example, if the district places 45%
of disabled children in general education classrooms for more than 80% of the day, but it placed
48% in the prior year). A district receives a “2” if it fails to meet the target but either improved
or maintained its performance from the prior year. A district receives a “3” if it meets or exceeds
the 51.2% target but declined in performance from the prior year. Finally, a district receives a
“4” if it meets or exceeds the 51.2% target and either improves or maintains its performance
from the prior year. Dkt. No. 2478-1 at 27.
Recall further that for the Dashboard metrics, school districts receive a color based on a
grid that combines information about the district’s current performance and any improvements or
regressions from the prior year. In scoring districts for comprehensive review, districts receive a
“1” for a red, a “2” for an orange, a “3” for a yellow, and a “4” for a “blue” or “green.”
After the districts are scored, the state calculates a cut score and then selects all districts
that fall below it for comprehensive monitoring. The cut score is expressed as a percentage of the
total number of points that were available to that district, for all metrics for which the district was
actually scored. According to the testimony of the policymakers, in the last monitoring cycle the
state originally set the cut score at 65%, meaning that districts scoring less than 65% of all points
available to that district would be selected for comprehensive review. But then, after the state ran
19
the numbers, it concluded that it did not have, or did not wish to spend, the resources necessary
to put all those districts through comprehensive monitoring. So it reduced the cut score to 62%.
Dkt. No. 2506 at 108–09, 153.
There are a number of problems with this system for selecting districts for comprehensive
review. First, two features of the scoring methodology combine to produce odd results: (i) in
awarding points, the state’s formula considers only whether a target is met or missed, without
considering the extent by which a district missed or exceeded the target, meaning that near
misses and wide misses receive the same number of points; and (ii) the state’s formula heavily
prioritizes year-to-year improvements, even if the improvements are quite minor, meaning that
minor fluctuations in performance can produce major differences in scoring. The court monitor
identified several examples of how these two features influence districts’ scores. As discussed
above, the state has set a target of 51.2% with respect to the placement of disabled students in
general education classrooms for greater than 80% of the day.
In the last round of selection for comprehensive review, one particular district placed
62.8% of disabled students in general education classrooms for greater than 80% of the school
day. This rate represented a 0.1% improvement from the prior year’s rate of 62.7%. The state
awarded four points to that district. Another district had a rate of 92%, which represented a
regression of 0.1% from the prior year’s rate of 92.1%. The state gave that district only three
points, although it appears to have performed significantly better at placing students in general
education classrooms, and the year-to-year differences for both are small. Dkt. No. 2469 at 60. In
a different area, the state considers the number of times that it determined that a district violated
federal law after investigating a parent’s complaint. One district went from 101 to 90
noncompliances and received a “2.” Dkt. No. 2506 at 106. By contrast, 17 districts increased
from one noncompliance in the previous year to two in the current year and received a “1.”7
The state’s reasons for relying on annual changes rather than improvements over a more
sustained period remain somewhat unclear. The Morgan Hill Concerned Parents Association
makes a colorable argument in its amicus brief that comparing one year to the prior year does not
provide the state with sufficient information to draw a reliable conclusion about the district’s
7
20
Another problem is that the state’s formula is unweighted. It does not prioritize the
metrics most closely related to the IDEA’s guarantee of an appropriate public education for
disabled students. For example, under the state’s formula, school districts can receive up to four
points merely for having high numbers of disabled students take statewide tests (regardless of
how they perform). But districts can receive no more than four points for meeting actual
performance targets for those assessments. Dkt. No. 2455-1 at 57. To use another example,
districts can receive no more than two points for meeting five out of six preschool assessment
targets, and for improving or maintaining performance from the prior year. But districts can also
receive two points merely for submitting between 50% and 69% of all required reports to the
state in a timely manner. Dkt. No. 2455-1 at 58, 62. The selection formula also entirely omits
districts’ performance on child find, although the IDEA singles out child find as a priority area
for monitoring. 20 U.S.C. §§ 1412(a)(3)(A), 1416(a)(3)(B); 34 C.F.R. § 300.600(d)(2).
These problems appear to result in the selection of the wrong districts for comprehensive
review. To assist the Court in evaluating the state’s selection methodology, the court monitor
developed a competing methodology and applied it to the same set of school districts evaluated
by the state. The monitor’s approach considered districts’ relative distance from the target (rather
than an all-or-nothing score), did not emphasize annual improvements, and focused only on four
metrics that are particularly tied to disabled students’ outcomes—performance on English and
Language Arts and Math assessments, suspension rates, and placement of school-aged children
in general education classes. Dkt. No. 2510 at 2. The monitor also expanded the analysis to
include two more least restrictive environment metrics for school-aged children and preschool
placements in regular early-childhood programs. Dkt. No. 2510 at 4. State policymakers agreed
that the monitor chose metrics that are most closely tied to the provision of a free and appropriate
education. Dkt. No. 2518 at 27. Finally, the monitor chose a cut score that would select roughly
trends in performance, as opposed to natural fluctuations in the data. Dkt. No. 2485 at 14; Dkt.
No. 2485-1 at 3. The state did not offer a clear response in its written submissions or during the
evidentiary hearings.
21
the same percentage of districts for monitoring that the state itself selected, to provide a fair point
for comparison.
The monitor’s method, when limited to four metrics, would select none of the districts
selected by the state. What’s more, all of the districts selected using this method scored 70% or
more under the state’s system—safely above the state’s 62% cut-off for monitoring—or were not
scored at all. Dkt. No. 2516 at 4. When the analysis is limited to districts with at least 31 students
with disabilities, six districts selected by the state would be selected under the monitor’s
approach, but the majority of districts selected under the monitor’s approach scored 70% or more
under the state’s approach. Expanding the analysis to seven metrics produces similar results. Dkt.
No. 2516 at 5–6. The monitor’s analysis suggests that the state’s design choices significantly
affect which districts are selected for comprehensive review, and that the districts selected are
not likely the ones most in need of intensive intervention.
The data also suggest that not enough districts are selected. In all, the state scored 1,330
school districts for monitoring.8 Dkt. No. 2516 at 2. The state selected only 34 districts for
comprehensive review. At the same time, excluding the first-year charters and very small school
districts that the state simply ignored, the state classified only 48 districts as in compliance with
federal law. Dkt. No. 2516 at 9. Perhaps these numbers would not provide much cause for
concern if the data suggested that a significant proportion of districts were found out of
compliance only in areas that are not as closely tied to the provision of an appropriate education.
But the state’s metrics further suggest that this is not the case. For example, 68 districts received
a “red” on all three Dashboard metrics for suspension and performance on English and Language
Arts and Math assessments, but only seven were selected for comprehensive review.9 (37 of
those districts received a score of 70% or above under the state’s selection formula).
8
This number excludes first-year charter schools. The state did not have prior-year performance
data for these schools and therefore did not consider them for selection for comprehensive
review.
9
This number also excludes first-year charter schools.
22
Policymakers testified that 83% of school districts were not meeting targets for the state’s
English and Language Arts assessment, and 85% were not meeting targets for the state’s Math
assessment. Dkt. No. 2506 at 192.
Furthermore, the state’s cut score does not appear to be tethered to a clear theory of
compliance. In fact, it appears somewhat arbitrary. The state did not explain in its written
submissions why a score of 62% (as opposed to a score of 52% or 72%) sufficiently identifies
those districts urgently in need of intensive monitoring. And, as mentioned earlier, at the
evidentiary hearings the policymakers suggested that the cut score was set, and then later
adjusted, based on resource constraints. Dkt. No. 2506 at 153. Although it is inevitable that
resource constraints and other practical considerations will play some role in the number of
districts chosen for monitoring in a given cycle, it cannot, given the IDEA’s legal requirements,
be the primary driver of the decision about where to set the cut score.
All these flaws in the state’s system for selecting districts for comprehensive review
combine to put it in violation of the IDEA. The system does not result in meaningful choices
about which districts, and how many districts, should receive intensive intervention.10 This is not
to suggest that the state may never consider improvement in selecting school districts to monitor,
or that the state may never use all-or-nothing scoring for an individual metric, if the targets for
that metric are sufficiently demanding. Nor would the state be required to adopt the competing
methodology designed by the court monitor. But the current system is so defective that it is not
As mentioned earlier, it’s worth noting that the state’s proposed selection methodology for its
intensive preschool monitoring activity bears close similarities to the methodology used for
comprehensive review. For preschool review, the state proposes to use an unweighted formula
with all-or-nothing scoring (although the formula would not consider year-to-year performance).
Districts would receive points across ten different metrics, and districts scoring “6” or more
would be selected for review. (Confusingly, for comprehensive review, better-performing
districts receive more points, while for preschool review, better-performing districts would
receive fewer points). The cut score of “6” was set with resource constraints in mind and does
not appear to be tied to a clear theory of statutory compliance. Given the similarities between the
two methodologies, it seems reasonably likely that the same problems identified for
comprehensive review would extend to preschool review as well.
10
23
merely inadvisable from a policy standpoint; it is legally inadequate.11
III. FURTHER ISSUES
The issues described in Section II put the state out of compliance with federal law and
prevent it from proceeding to Phase 3. The plaintiffs and monitor contend there are still other
flaws in the state’s procedures for identifying school districts for intervention. But in contrast to
the issues discussed in Section II, the alleged deficiencies discussed in this section can all be
placed in one of two categories: (i) the record does not justify a conclusion that the flaw exists in
the first instance; or (ii) the flaw may exist, and perhaps it compounds the problems discussed in
the previous section, but in the overall scheme, it does not appear significant enough to warrant a
finding that the state is out of compliance with its statutory monitoring obligations. (That is,
assuming the problems discussed in Section II will be adequately addressed.) Therefore, absent
some significant revelation, or some significant change in the way the state conducts itself in
these areas, the Court will not entertain further discussion on them during the renewed Phase 2.
A.
Disaggregation. The monitor and plaintiffs argue that the state is not in compliance across
many metrics because, except as necessary for its disproportionality analyses, the state does not
disaggregate its data. The state’s monitoring system is designed to address outcomes for disabled
students in general, and the state does not separately examine outcomes for disabled students
who are also members of vulnerable subgroups, such as students in foster care, homeless
students, English-language learners and students eligible for free and reduced-price lunch. The
state also generally does not disaggregate its data by disability type.
The monitor’s concern is that, if these vulnerable groups are small enough, the aggregate
11
In addition to the issues discussed above, two issues identified during Phase 1 remain
outstanding. During Phase 1, the state was found out of compliance for failing to collect data
related to districts’ implementation of their students’ IEPs and the use of restraints and seclusion.
Dkt. No. 2428 at 13–18, 26–28. The state is currently in the process of designing data collection
protocols to address both issues. Because the state will need to accommodate these areas within
all levels of its monitoring system, these issues remain relevant here, although they will be
addressed in greater depth at a later date.
24
data will mask important information specific to these groups, so that a district that performs
well with respect to disabled students as a whole will not be selected for monitoring even if it
may fail to provide adequate services to, for example, disabled students that are in foster care.
This, however, is a policy issue rather than a legal one. Neither the monitor nor the
plaintiffs have identified a legal basis for requiring the state to analyze these data, and the
monitor stated at the hearing that he was unaware of any other state that disaggregates data in
this manner. Dkt. No. 2506 at 169. By comparison, a separate set of statutory provisions,
regulations, and guidance documents issued by the Department of Education govern a state’s
obligation to assess disproportionality for race and ethnicity, and even this form of analysis is
limited to three discrete areas: discipline, identification, and placement.
Nor does it appear the failure to disaggregate as recommended by the monitor is
significantly interfering with the state’s ability to monitor school districts. To assess this issue,
the monitor chose three metrics (two for least restrictive environment, and one for suspension)
and applied them to a disaggregated sample of data. The monitor concluded that the
disaggregation would result in the selection of districts for targeted reviews that were not
previously selected by the state. In response, state officials observed that, assuming the validity
of this approach, it would result only in incremental changes in the number of districts selected
for monitoring. The state’s “error rate,” – meaning the number of school districts that were not
selected for monitoring but should have selected under the monitor’s approach, divided by the
total number of school districts not selected – would range from somewhere between less than
1.0% to 3.6% for each metric. Dkt. No. 2518 at 81–82. Moreover, all of those districts are in
some form of monitoring (and so none have been determined to “meet requirements”). Several
are receiving assistance through the state’s other accountability programs for poor outcomes
related to foster children, regardless of disability.
Discipline. The state uses the Dashboard to assess district’s discipline practices. The state
considers only suspensions for longer than ten days. The plaintiffs and monitor argue that the
state is out of compliance, because it does not consider multiple suspensions, and so a student
25
who is suspended once for one event is counted once, and a student suspended four times for
four events in a year is also counted once. The state also does not consider the cumulative length
of time a student is suspended. Dkt. No. 2469 at 12.
But neither the plaintiffs nor monitor have adequately explained how these issues
significantly hinder the state’s assessment of school discipline. It is not known, for example,
whether one would expect to see many districts with many short suspensions but few lengthy
suspensions. It is similarly unknown whether focusing on the number of suspensions per student
would enhance the state’s ability to red-flag districts. Perhaps one would ordinarily expect
suspensions to be concentrated among smaller groups of students, and so focusing on students
rather than suspension events would not add much. At this point, it is only apparent that the state
does not use the most granular metric possible. It is not apparent that this practice materially
affects the state’s ability to monitor discipline practices.
Placements close to home. As part of the least restrictive environment principle, the
IDEA requires districts to ensure that educational placements are as close as possible to the
child’s home, and in the school the child would attend if he or she was not disabled, “unless the
IEP . . . requires some other arrangement.” 34 C.F.R. § 300.116(b)(3), (c). At Phase 1, the state
was found in compliance because it collected sufficient information to permit it to monitor
district performance in this area, including information about students’ “districts of residence”
and “districts of service.” See Dkt. No. 2435 at 175; Dkt. No. 2428 at 36–37.
The plaintiffs and monitor believe that the state should be ruled out of compliance
because it does not do any data analysis for this requirement, or more generally into patterns of
placement. State policymakers explained that this principle does not translate well to the state of
California, because California is a “school choice” state, and so officials cannot presume that a
child would attend the school closest to home if he or she were not disabled.
As applied to a state like California, the regulations are best understood as enshrining a
general principle that districts must attempt to avoid concentrating disabled children in remote
placements if more desirable placements are available. But it is not clear that data analysis
26
regarding placement is the best way to guard against this, for the reasons explained by the state.
Perhaps analysis of IEP implementation – an issue on which the Court has ruled the state must
collect data – would assist in this area. Or perhaps it’s an issue that should be considered only
when the state is actually monitoring school districts (as opposed to when the state is assessing
districts for different types of monitoring). In any event, the Court cannot conclude on this record
that the state’s decision not to analyze “close to home” data when selecting districts for
monitoring significantly interferes with its ability to meet its federal statutory obligations.
Complaints. The IDEA gives parents two mechanisms for enforcing their rights. First, it
allows parents and advocates to file complaints with the state if they believe that a school district
is not complying with federal or state special education law. 34 C.F.R. §§ 300.151–300.153. The
state must carry out an investigation, if necessary, and resolve the complaint within 60 days. 34
C.F.R. § 300.152(a). Second, the IDEA permits parents or school districts to file due process
complaints relating to the identification, evaluation, or educational placement of a disabled child,
or the provision of a free and appropriate public education to the child. 34 C.F.R.
§ 300.507(a)(1). The parent or school district is entitled to a hearing before an administrative
officer to resolve the issues raised in the complaint. 34 C.F.R. § 300.511(a), (c). A party
dissatisfied with the officer’s decision can file a lawsuit in state court or federal district court. 34
C.F.R. § 300.516(a).
To evaluate district performance in this area, the state counts the number of instances in
which a school district is determined to be out of compliance with federal law, through either
process. These numbers do not lead to selection for any targeted review but are considered in
selecting districts for comprehensive review. Districts receive one point for having more than
one documented noncompliance, and another point if the number has decreased from the prior
year. Districts with zero noncompliances receive four points. Dkt. No. 2455-1 at 60.
The plaintiffs and monitor object to the state’s use of a simple count rather than a per
capita rate (the number of noncompliances divided by the total number of students within a
district). Under the state’s method, two districts with three noncompliances each will be treated
27
the same, although three might be a troubling number for a very small district but less
concerning for a very large district. The state responds that the target for this metric treats any
noncompliance as giving significant cause for concern, and so using a more specific metric
would not add materially to its analysis. In other words, the target for this metric is zero: districts
with any noncompliances receive the lowest scores.
To be sure, there are problems with the state’s explanation. The state does not merely rely
on whether a district met or missed the target; it also considers whether the district has gotten
better or gotten worse from the prior year. It’s hard to see how the state could meaningfully
determine whether the district is improving or regressing from raw totals alone and without any
sense of the size of the school district. A district’s population may fluctuate from year to year,
and so a district’s number of noncompliances could conceivably increase as the result of
population changes, rather than as a result of poor district performance.
That said, the record does not demonstrate that this issue, which represents one point in a
complicated selection formula, materially affects the state’s ability to monitor school districts.
And because substantial changes must already be made to the state’s selection methodology for
comprehensive review, this issue is likely to become even less important by the time of the
renewed Phase 2 proceedings.
Preschool achievement. The state administers assessments for both school-aged children
and preschool-aged children. For preschoolers, the state uses the “Desired Results
Developmental Profile,” which measures social development, knowledge acquisition, and the use
of appropriate behaviors to meet needs. Dkt. No. 2455-1 at 25. The state has established six
targets for disabled students taking these assessments. However, there is no targeted review for
preschool achievement. Poor performance in preschool assessments is one factor in the state’s
proposed preschool review, but a minor one. A district that misses all six assessment targets
would receive one point, and a district needs a total of six points for selection. Performance on
preschool assessments also comprises only one of many different selection metrics for
comprehensive review. As a result, a school district that performs poorly on its preschool
28
assessments will not receive any monitoring, unless it also performs poorly across other
dimensions. The plaintiffs and monitor believe that poor performance on preschool assessments
should receive a separate targeted review.
The state explained that it has made a policy determination to focus its efforts on placing
more preschoolers in regular early childhood programs rather than separate programs. Dkt. No.
2506 at 204. State officials testified that placements in general education settings (where
expectations are often higher) are associated with better performance on assessments. In
addition, the state’s data establishes that all districts that missed one or more preschool
assessment targets have been selected for at least one form of monitoring. Dkt. No. 2503 at 1. Of
those districts, half are monitored for inadequate placements. Further, because all of these
districts are in some form of monitoring, none has received a positive annual compliance
determination from the state. Given this record, the Court will not second-guess the state’s policy
determination that for now, the best way to evaluate and monitor districts with respect to
preschool-aged students is to focus on preschool placements.
Referrals. As discussed above, the state assesses school discipline by looking at districts’
rates of suspensions greater than ten days. The state does not collect data regarding districts’
referrals of disabled students to law enforcement. During Phase 1, the Court noted the
importance of this data, because disabled students are often referred to law enforcement at
disproportionate rates. However, the state was not ruled out of compliance for failing to
independently collect referral data, because the federal government routinely collects the data
and makes it publicly available. Accordingly, the Court found at Phase 1 that the state had
sufficient access to referral data and reserved for Phase 2 the question whether the state
adequately analyzes this data in determining which districts to select for monitoring.
During Phase 2, the state took the position that it doesn’t need to separately analyze
referral data, because school districts’ suspension rates are correlated with their rates of referrals
to law enforcement, such that suspension data alone is sufficient to red-flag districts who may be
overusing referrals to law enforcement. The record supports this. Although the referral data that
29
is publicly available is of poor quality, the monitor’s supplemental analysis suggests a
meaningful correlation between school districts’ suspension rates and their referral rates. Dkt.
No. 2510 at 10–12. The analysis also suggests that there are few districts with high referral rates
but low suspension rates. Dkt. No. 2510 at 12. While it remains conceivable that the state may
need to show that during its actual monitoring activities it sufficiently disentangles the reasons
that referrals to law enforcement occur, there is not an issue at this phase with respect to law
enforcement referrals. As with preschool achievement, the Court will not second-guess the
state’s decision to use suspension practices as a proxy for referral practices on this record.
Parent input. The state also analyzes parent input as part of its monitoring processes. The
state asks parents to respond to the question: “Did the school district facilitate parent
involvement as a means of improving services and results for your child?” Dkt. No. 2455-1 at
30. To analyze this data, the state records the percentage of “yes” responses for each school
district. Districts that fall below the target (92% of parents responding “yes”) are selected for
performance indicator review and move closer to selection for comprehensive review. The
plaintiffs and monitor believe that this target is unambitious because in 2016, state performance
was at 99.4%. Dkt. No. 2469 at 40. But for many of the reasons discussed in the Court’s Phase 1
order and at the Phase 1 hearings, data about parents’ responses to this question do not appear to
be particularly useful (nor does the Court have confidence that any sort of statewide parent
survey data would be actionable), so any concerns about the target set by the state in connection
with it will not give rise to a noncompliance finding.
B.
In addition to the above metric-level issues, the Court finds that two of the state’s
selection methods for monitoring activities present no cause for legal concern on this record:
Data identified noncompliance. The state’s method of selecting districts for data
identified noncompliance monitoring (the monitoring activity relevant to the IDEA’s timeliness
30
requirements) appears to be adequate.12 Because the targets require perfect performance, a single
untimely IEP meeting or initial evaluation is sufficient to flag a district for monitoring. The
plaintiffs note that the state currently does not monitor districts for compliance with the IDEA’s
deadlines for resolution sessions, which are meetings that districts must convene between parents
and members of the disabled student’s IEP team after the parent files a due process complaint
against the district. The resolution session must be held within 15 days after the district receives
notice of the complaint. 20 U.S.C. § 1415(f)(1)(B)(i)(I); 34 C.F.R. § 300.510(a)(1). But going
forward, the state plans to include timeliness in convening resolution sessions as a metric for
targeted review. Dkt. No. 2506 at 15; Dkt. No. 2455-1 at 8 n.17.
Disproportionality. The state’s selection formulas for its three disproportionality reviews
are, once the state completes a couple of simple fixes, adequate. These are the state’s monitoring
activities for districts where students appear to be disproportionality represented among those
receiving special education services based on their racial or ethnic group. Federal regulations
require the state to use a specific formula to assess disproportionality, although states retain
discretion to set an appropriate target for determining when a school’s numbers suggest that it
has become “significantly disproportionate.” 34 C.F.R. §§ 300.646, 300.647.
One concern is that while the state appears to have set an appropriate target, its current
formula for assessing disproportionality in discipline and placement is incorrect. The state’s
formula compares the risk of disciplinary action or restrictive placements for disabled students of
a particular race or ethnicity to the risks faced by students of other races and ethnicities who are
in general education settings. Dkt. No. 2469 at 17. The United States Department of Education,
however, has explained that the relevant comparator is the risk faced by disabled students of
other races and ethnicities. The state has conceded the error and plans to put the correct formula
The state currently assesses compliance with the IDEA’s deadlines for (1) evaluating students
to determine eligibility for special education services, 34 C.F.R. § 300.301(c)(2)(i); (2) putting
IEPs in place for toddlers transitioning to preschool, 34 C.F.R. § 300.124(b); (3) performing
required reviews of IEPs, 34 C.F.R. § 300.324(b)(1)(i); (4) conducting required reevaluations of
students with IEPs, 34 C.F.R. § 300.303(b)(2).
12
31
in place for the 2019–2020 school year. Dkt. No. 2507 at 68. Given this commitment, and the
absence of any other outstanding issues with respect to the targeted disproportionality review,
this issue does not provide a basis for finding the state out of compliance with its statutory
obligations.
The state is also in compliance with respect to significant disproportionality review, the
intensive form of monitoring. The state selects all districts that are disproportionate (as defined
by the state’s chosen threshold) for three years for intensive monitoring. Federal law permits a
state to exempt districts from monitoring if it determines, in its discretion, that a district is
making “reasonable progress” towards resolving its issues. 34 C.F.R. § 300.647(d)(2). The state
does not exercise this discretion but selects for monitoring all districts that exceed the threshold.
Because the state’s formula selects more districts for monitoring than federal law requires, the
state is currently in compliance. If the state adopts a “reasonable progress” definition in the
future, and if the plaintiffs believe that the definition is so deficient that it would significantly
interfere with the state’s ability to identify and select school districts for disproportionality
monitoring, they may bring that to the Court’s attention. But absent any such development, there
is no need for further inquiry during this phase.
Private school monitoring. During the Phase 1 hearings, policymakers testified that the
state monitors private schools that educate students with disabilities who are placed there by
virtue of their IEPs. These schools must be certified by the Special Education Division, and
although the state collects data from these schools, they are all monitored on a cyclical basis,
irrespective of whether the data suggests they need it. Dkt. No. 2518 at 94–98. Because the state
monitors all private schools, this area presents no issues for purposes of Phase 2, although
Phase 3 will involve further inquiry into the content and quality of the state’s monitoring
activities.
Effectiveness of monitoring. The IDEA states that the “primary focus” of any state’s
monitoring system must be on “[i]mproving educational results and functional outcomes for all
children with disabilities,” and ensuing that school districts comply with the law, including
32
“those requirements that are most closely related to improving educational results for children
with disabilities.” 34 C.F.R. § 300.600(b). This focus logically requires the state to conduct some
assessment of the effectiveness of its monitoring activities – in effect, to monitor its monitoring
system. The monitor and the plaintiffs argue that the state is out of compliance because it did not
present any information in this phase about how it does this.
It’s true that, to get out from under the consent decree, the state must show that it has
institutionalized a process for self-examination, to ensure that its monitoring actually improves
student outcomes and that its data analysis activities do not become detached from realities on
the ground. The deficiencies identified in this order underscore the importance of such a system.
However, it makes little sense to require the state to present evidence about its self-evaluation
process at this stage, given that it is already engaging in self-reflection as part of these
proceedings, and already considering substantial changes to its monitoring system. Accordingly,
the state need not revisit this requirement until Phase 4.
IV. NEXT STEPS
Given the concerns discussed in this ruling, the state must repeat Phase 2. In the next set
of written submissions and at the next set of evidentiary hearings, the state must explain the
changes that it has made to address the concerns identified in this order. The sequence of written
submissions will remain the same, although the parties are urged to include data analyses where
appropriate.
A case management conference is scheduled for Wednesday, August 28, 2019 at
9:30 a.m. to discuss the next phase and the state’s proposals for addressing IEP implementation.
A case management statement is due seven days prior. The parties should include a proposed
schedule for the written submissions and evidentiary hearings.
The parties may also, if they choose, file any materials with the case management
statement that would assist the Court in facilitating a discussion between the parties regarding the
state’s proposals for integrating IEP implementation into its monitoring protocols. If the parties
no longer need to hold such a discussion with the Court (or if they wish to defer the matter to the
33
next set of evidentiary hearings), then the policymakers do not need to attend the case
management conference, although they are welcome to do so, either in person or by phone.
IT IS SO ORDERED.
Dated: July 5, 2019
______________________________________
VINCE CHHABRIA
United States District Judge
34
Disclaimer: Justia Dockets & Filings provides public litigation records from the federal appellate and district courts. These filings and docket sheets should not be considered findings of fact or liability, nor do they necessarily reflect the view of Justia.
Why Is My Information Online?