برچسب: takes

  • State takes another step toward mandatory testing for reading difficulties in 2025

    State takes another step toward mandatory testing for reading difficulties in 2025


    Students at Theodore Roosevelt Elementary School in the Burbank Unified School District practice their reading skills.

    Credit: Jordan Strauss/AP Images

    A panel of reading experts has designated the tests that school districts can use to identify reading difficulties that kindergartners through second graders may have, starting next fall.

    Gov. Gavin Newsom’s announcement Tuesday of the selection of the reading risk screeners marks a milestone in the nearly decadelong campaign to mandate that all young students be measured for potential reading challenges, including dyslexia. California will become one of the last states to require universal literacy screening when it takes effect in 2025-26.

    To learn more

    For Frequently Asked Questions about the screening instruments for risk of reading difficulties, go here.

    For more about the screeners selected for district use, go here.

    For the letter on screening sent to district, county office and charter school superintendents, go here.

    For more on the Reading Difficulties Risk Screener Selection Panel, go here.

    Between now and then, districts will select which of four approved reading screeners they will use, and all staff members designated as the testers will undergo state-led training. The Legislature funded $25 million for that effort.

    “I know from my own challenges with dyslexia that when we help children read, we help them succeed,” Gov. Gavin Newsom said in a statement.

    Students will be tested annually in kindergarten through second grade. In authorizing the screeners, the Legislature and Newsom emphasized that screening will not serve as a diagnosis for reading disabilities, including dyslexia, which is estimated to affect 5% to 15% of readers. Instead, the results could lead to further evaluation and will be used for classroom supports and interventions for individual students. Parents will also receive the findings of the screenings.

    “This is a significant step toward early identification and intervention for students showing early signs of difficulty learning to read. We believe that with strong implementation, educators will be better equipped to support all learners, fostering a more inclusive environment where every child has the opportunity to thrive,” said Megan Potente, co-director of Decoding Dyslexia CA, which led the effort for universal screening. 

    A reading-difficulty screener could consist of a series of questions and simple word-reading exercises to measure students’ strengths and needs in phonemic awareness skills, decoding abilities, vocabulary and reading comprehension.  A student may be asked, for example, “What does the ‘sh’ sound like in ‘ship’”?

    Among the four designated screeners chosen is Multitudes, a $28 million, state-funded effort that Newsom championed and the University of California San Francisco Dyslexia Center developed. The 10 to 13-minute initial assessment will serve K–2 grades and be offered in English and Spanish.

    The other three are:

    Young-Suk Kim, an associate dean at UC Irvine’s School of Education, and Yesenia Guerrero, a special education teacher at Lennox School District, led the nine-member Reading Difficulties Risk Screener Selection Panel that held hearings and approved the screeners. The State Board of Education appointed the members.

    The move to establish universal screening dragged out for a decade. The California Teachers Association and advocates for English learners were initially opposed, expressing fear that students who don’t speak English would be over-identified as having a disability and qualifying for special education.

    In 2015, then-Gov. Jerry Brown signed legislation requiring schools to assess students for dyslexia, but students weren’t required to take the evaluation.   

    In 2021, advocates for universal screening were optimistic legislation would pass, but the chair of the Assembly Education Committee, Patrick O’Donnell, refused to give it a hearing.

    “Learning to read is a little like learning to ride a bike. With practice, typical readers gradually learn to read words automatically,” CTA wrote in a letter to O’Donnell.

    Sen. Anthony Portantino, D-Glendale, reintroduced his bill the following year, but instead Newsom included funding and requirements for universal screening in his 2023-24 state budget.

    The Newsom administration and advocates for universal screening reached out to advocates for English learners to incorporate their concerns in the requirements for approving screeners and to include English learner authorities on the selection panel.

    Martha Hernandez, executive director of Californians Together, an organization that advocates for English learners statewide, said Wednesday it was clear that the panel considered the needs of English learners and she is pleased that the majority of the screeners are available in Spanish and English. 

    “Their commitment to addressing the unique needs of English learners was evident throughout the process,” Hernandez said.

    However, she said it is important for the state to provide clear guidance to districts about what level of English proficiency is required in order for students to get accurate results from a screener in English.

    “The vast majority of English learners will be screened only in English, and without evidence that these screeners are valid and reliable across different English proficiency levels, there is a risk of misidentification,” Hernandez said.

    Hernandez said Californians Together emphasized to the panel that it is important for students who are not yet fluent in English to be assessed for reading in both their native language and English, “to capture the full scope of their skills.” In addition, Hernandez said it is crucial for the state Department of Education to offer guidance to districts on selecting or developing a screener in languages other than English or Spanish.

    The article was corrected on Dec. 18 to note that the initial Multitudes assessment takes 10 to 13 minutes, not 20 minutes, depending on the grade; a followup assessment can take an additional 10 minutes.





    Source link

  • California takes a big step in how it measures school performance, but there’s still more to do

    California takes a big step in how it measures school performance, but there’s still more to do


    Credit: Alison Yin / EdSource

    Accountability has been a central plank in California’s — and our nation’s — school reform efforts for over two decades. Over nearly that entire period, California has been criticized (including by me) for being one of the few states that does not include a measure of student achievement growth in our accountability system. The current approach, exemplified in the California School Dashboard, rates schools on their average performance levels on the state’s standardized tests, and on the difference between the school’s average performance this year and last year.

    But the state doesn’t have, and has never had, a student-level growth model for test scores. Student-level growth models are important because they do a much better job than the state’s existing measures of capturing school effectiveness at improving student achievement. This is because growth models directly compare students to themselves over time, asking how much individual children are learning each year and how this compares across schools and to established benchmarks for annual learning. The crude difference models the state currently displays in the dashboard could give the wrong idea about school performance, for instance, if there are enrollment changes over time in schools (as there have been since the pandemic).

    Growth models can help more fairly identify schools that are often overlooked because they are getting outsize results with underserved student groups. In other words, they send better, more accurate signals to report card users and to the state Department of Education about which schools need support and for which students. Along with Kansas, California has been the last holdout state in adopting a report card that highlights a growth model.

    Though the state’s task force on accountability and continuous improvement, on which I served, wrapped up its work and recommended a growth model almost nine years ago, the process of adopting and implementing a growth model has been — to say the least — laborious and drawn-out. Still, I was delighted to see that the California Department of Education (CDE) has finally started providing growth model results in the California School Dashboard! This is a great step forward for the state.

    Beyond simply including the results in the dashboard, there are some good things about how the state is reporting these growth model results. The growth model figures present results in a way I think many users will understand (points above typical growth), and results for different student groups can be easily viewed and compared.

    There is a clear link to resources to help understand the growth model, too. The state should be commended for its efforts to make the results clear and usable in this way.

    It doesn’t take a detailed look at the dashboard to see, however, that there are some important fixes that the State Board of Education should require — and CDE should adopt — as soon as possible. Broadly, I think these fixes fall into two categories: technical fixes about presentation and data availability, and more meaningful fixes about how the growth model results are used.

    First, the data are currently buried too deeply for the average user to even find them. As far as I can tell, the growth model results do not appear on the landing page for an individual school. You have to click through using the “view more details” button on some other indicator, and only then can you see the growth model results. The growth model results should, at minimum, be promoted to the front page, even if they are put alongside the other “informational purposes indicator” for science achievement. A downloadable statewide version of the growth model results should also be made available, so that researchers and other interested analysts can examine trends. Especially in light of the long shadow of Covid on California’s students, we need to know which schools could benefit from more support to recover.

    Second, the state should prioritize the growth model results in actually creating schools’ dashboard ratings. Right now, the color-coded dashboard rating is based on schools’ status (their average scale score) and change (the difference between this year’s average score and last year’s). It would be much more appropriate to replace the change score with these growth model results.

    There are many reasons why a growth model is superior, but the easiest to understand is that the “change” metrics the state currently uses can be affected by compositional changes in the student body (such as which kinds of students are moving into and out of the school). Researchers are unanimous that student-level growth models are superior to these change scores at accurately representing school effectiveness. Even for California’s highly mobile student population, growth models can accommodate student mobility and give “credit” to the schools most responsible for each child’s learning during that academic year.

    To be sure, I think there are other ways the dashboard can likely be improved to make it more useful to parents and other interested users. These suggestions have been detailed extensively over the years, including in a recent report that dinged the state for making it difficult to see how children are recovering post-Covid.

    The adoption of a growth model is a great sign that the state wishes to improve data transparency and utility for California families. I hope it is just the first in a series of improvements in California’s school accountability systems.

    •••

    Morgan Polikoff is a professor at the University of Southern California’s Rossier School of Education.

    The opinions expressed in this commentary represent those of the author. EdSource welcomes commentaries representing diverse points of view. If you would like to submit a commentary, please review our guidelines and contact us.





    Source link