will that work for us? interpreting research from the memphis striving readers project (msrp)...
Post on 26-Dec-2015
215 Views
Preview:
TRANSCRIPT
Will That Work for Us? Interpreting Research from
The Memphis Striving Readers Project (MSRP)
Presented by
Ric Potts, MCS; J. Helen Perkins, U of M; Elizabeth Heeren, MCS; Rorie Harris, MCS; and Jill Feldman, RBS
2008 International Reading Association Research Conference
Atlanta, GA
Session Overview
• Introduction to the Striving Reader’s grant• Overview of Memphis SR research design• Year One Impact Analyses• Collection of implementation fidelity data
– implications for practitioners and researchers• Planned (Ongoing) Analyses• Q & A /Group Discussion
Introduction: Memphis Striving Readers Project (MSRP)
Ric Potts, PI – MSRP
Memphis City Public Schools
Memphis-The City
The City of Memphis has a population of 642,251.
63.1% African American
31.3% Caucasian
4.1% Hispanic
And one Elvis
Approximately 70 percent of adolescentsstruggle to read. The young people enrolled inmiddle and high school who lack the broadliteracy skills to comprehend and learn advancedacademic subjects will suffer serious social,emotional, and economic consequences.
» Reading at Risk: The State Response to the Crisis in Adolescent Literacy, Oct. 2005
Urban Child InstituteThe State of Children in
Memphis and Shelby County2006
“Under-educated children have no future.”
Urban Child InstituteThe State of Children in
Memphis and Shelby County2006
• by U.S. standards roughly 75 percent of students in Tennessee fail to meet national grade appropriate standards, and Memphis is at the bottom in Tennessee. . . . Memphis is one of theleast-educated cities in America.
Motivation behindMemphis Striving Readers Project
• Memphis is among the poorest and least-educated cities in the US– 30.1% of all children live in poverty
– 24.3% of adults have less than a HS education
– 36.7% have HS diploma or equivalent
– 30.5% have Assoc. or some college
– 8.5% have at least a BA
• MCS is 21st largest K12 district in US >116,000 students– Over 95% of MCS’ 196 schools are Title I schools
– 71% of MCS students qualify for free/reduced price lunch
– MCS students are 87% AA; 9% White; 4% “other”
– In 85% of MCS schools, 33% of students change schools during year
– In 2003-04, the system-wide graduation rate was 61 percent
– 71% of students in grades 6-8 scored below the 50th percentile on TCAP (Reading/Language Arts)
Striving Readers – A Federal Response
• In 2005, the Department of Education called for proposals for the Striving Readers grant.
• In March, 2006, Memphis was one of eight sites awarded the grant.
Memphis Striving Reader Program Targeted Schools
School Grade Span
Total Enrollment Total # Of Non-Special Education Students Scoring In Bottom Quartile
In Reading
School 2 6-8 1,021 414
School 1 6-8 1,033 384
School 6 6-8 700 251
School 5 6-8 765 245
School 8 6-8 547 178
School 4 6-8 486 196
School 3 6-8 976 357
School 7 6-8 877 274
The Whole School Intervention: Memphis Content Literacy Academy
(MCLA)
Overview presented by
J. Helen Perkins, SR Co-PI
University of Memphis
A Change Model
No Knowledge
First Exposure
Deeper Learning withLimited Capacity
Practice with Coaching
Refined andExpandedCapacity
Expertise& Abilityto Coach Others
A Capacity-Building Model for Teacher Development
(Cooter & Cooter, 2003)
Emphasis: “Deep Training” (180 hours over two years) …
Memphis Content Literacy Academy Infusing Simultaneously Across Core Subject Areas Scientifically-based Reading Research (SBRR) Strategies in…
VocabularyReading ComprehensionReading Fluency
Benefits to Teacher – “Laureates”…
• Advanced Training (180 hours) on scientifically-based reading instruction (SBRR) for urban children
• A Master Teacher “Coach” to Assist (30 hours) with Implementing New Strategies (in their own classrooms!)
• Twelve (12) Graduate Semester Hours of Credit from University of Memphis (FREE) (applicable to an advanced degree)
• Can Seek “Highly Qualified” Endorsement in Reading• Books and Materials (FREE)• Success in Helping Children Achieve “AYP”• Principal Support
MCLA Year 1: Selected StrategiesFluency• Choral Reading• Paired reading• Guided, repeated, oral reading (pairs)
Comprehension• Question Generation• Three- Level Retelling
•Oral•Graphic Organizor•Written
• Comprehension monitoring• Expository Text Patterns• Multiple Strategies
Vocabulary Development• Pre-instruction of vocabulary• Repeated, multiple exposures• Semantic Maps
Classroom Organizational Tools & Strategies: Year 1
• CREDE Standards
• Whole class v. collaborative small group
• Reading Next Elements
• Use of leveled materials (e.g., National Geographic)
CREDE Formatting of Professional Development
Training
http://crede.berkeley.edu/standards/standards.html
Classroom Action Plans (CAPs)Spring 2008
Science, Social Studies, & ELA
Your task is to develop a series of class lessons where you teach academic vocabulary in a unit of your choice.
You must have at least one vocabulary learning strategy/activity that occurs:
1. BEFORE students read the assigned text,2. DURING the reading assignment, and
3. AFTER the reading assignment
MCLA Classroom Model
• Gradual release of responsibility (teacher modeling, guided practice, independent practice, independent use)
• Integration of 12 literacy strategies (vocabulary, fluency & comprehension)
• Development of Classroom Action Plans (CAPs) (content area lesson plans for strategy implementation including procedures for student assessment)
• On-site support provided by coaches
• Use of Curriculum Resource Center (CRC) materials
The Principals’ Fellowship
Literacy Leadership Practices Real World Problem Solving Create “Literacy Materials Centers” Early Identification w/ Intense/Focused Remediation Research-Informed Decision Making Involve Families Needs-Based Scheduling Matching the Most Successful Teachers
with “Critical Condition” Kids
READ 180, Our Targeted Intervention
Overview provided by
Elizabeth Heeren, SR Grant Coordinator
Memphis City Schools
Program Components
Student workbooks for Independent Practice in small and whole group rotations
Support materials for differentiated instruction in small group rotation
Tools for student placement and assessment
Key Elements of READ 180
• Fidelity of Implementation
• 90 minute classes
• Certified teachers (LA or Reading)
• District Instructional Support
• District Technological Support
• Scholastic training (site-based and on-line)
R180 Correlations to Reading Next Recommendations for Adolescent Literacy
• Direct, explicit comprehension instruction• Motivation and self-directed learning• Strategic tutoring• Differentiated texts (levels and topics)• Technology component• Ongoing formative assessment• Extended time for literacy• Professional development (long-term and on-
going)
Memphis Implementation
• We have 8 schools in the Striving Readers Grant, with up to 120 randomly selected R180 students at each school.
• Students receive R180 instruction for 2 years.• Each student placed in R180 falls in the lowest
quartile of TCAP (Reading score).• Each student in R180 is paired with a similar
student from the lowest quartile who does not receive the treatment (for impact comparison).
MSRP Research Design
Overview presented by
Jill Feldman, SR Research Director
Research for Better Schools
Overall MSRP Goals
To determine:
1. The effects of MCLA on core subject teachers’ knowledge and use of SBRR
2. The separate and combined effects of MCLA and Read 180 on students’ reading achievement levels, especially students who are identified as struggling readers
3. The separate and combined effects of MCLA and Read 180 on students’ achievement in core subjects, especially students who are identified as struggling readers
Funding, staff, curriculum resource center, facilities, incentives, research materials
Activities
Principals45 hours of Principal Fellowship participation
100% of principals incorporate plan for using MCLA strategies in SIP
100% attendance of key MCLA events
80% of principals report actively supporting the program
100% of MCLA schools have allocated space for the CRC
Teachers 90 of hours of MCLA training/yr for 2 years (180 hours)
Engage in weekly coaching sessions or as needed to meet teachers’ differentiated needs
8 CAP “cycles” completed each year for two years
100% of teachers complete performance measures identifying supplemental resources available/those necessary to support content area instruction
Students50% of students attend 4 classes taught daily by teachers participating in MCLA
Students learn to use 7 of 8 MCLA CAP strategies
Outputs
PrincipalsAwareness of and interest in staff implementation of MCLA concepts and strategies
Increased advocacy for school-wide use of MCLA strategies
TeachersIncreased knowledge about MCLA strategies
Improved preparedness to use research-based literacy strategies to teach core academic content
Increased use of direct, explicit instruction to teach research-based comprehension, fluency, and vocabulary strategies in content area classes
Integrated use of multiple MCLA strategies to support ongoing development of content-related instructional units
StudentsIncreased familiarity with and use of MCLA strategies when engaging with text
Increased internalization ofliteracy strategies
Increased confidence engaging with content related texts
Increased interest in school/learning
Short–term OutcomesInputs
PrincipalsImproved school climate
School-wide plans include focus on content literacy
Improved instructional leadership
TeachersIncreased effectiveness supporting students’ content literacy development
Continued collaboration among community of teachers to develop and implement CAPs
StudentsImproved reading achievement and content literacy:
10% increase in students scoring proficient in Reading/LA and other subject areas of TCAP
mean increase of five NCEs on ITBS
Increased performance on gateway and EOC exams
Long-term Outcomes
Higher Quality Teaching
PrincipalsAttend 45-hour sessions/yr (2 yrs)
Participate in motivational, recruitment and celebratory events
Discuss MCLA at faculty meetings
Conduct walkthrough observations
Provide opptys for teacher collab
Allocate space for CRC materials
TeachersAttend 30 weekly 3-hour MCLA training sessions/yr (2 years)
Develop and implement 8 CAPs per year in collab content-area groups
Meet with coaches for feedback to improve impl of MCLA strategies
Learn to use of leveled texts to support SR content literacy needs
StudentsLearn to use MCLA strategies to read/react to content related text (
MCLA Program Logic Model
Higher Student Achievement
Study Design and Analytic Approach: MCLA
Study Design MCLA:
• Evaluate teacher and student outcomes
– experimental design – randomly assigning schools (to treatment and control conditions)
• Teacher outcomes include– preparedness – frequency of literacy strategy use
Analytic Approach MCLA:
• Two-level HLM– spring ITBS and TCAP scores as
a function of teacher and school variables
Analytic Decisions• Missing Data
– students missing pretest score(s) deleted from impact analysis on relevant measure(s)
– teachers missing pretest score deleted from impact analysis on
measure • Covariates
– include all student- and school-level covariates in the model– run the model– eliminate the school covariate with the lowest significance level
(highest p-value) not less than 0.2– repeat steps 2 and 3 until the remaining covariates had p-values
less than 0.2– repeat steps 2-4 for the student covariates
MCLA: Random Assignment of Schools
Demographic Characteristics of Year 1 MCLA Student Sample
Student Characteristic Control a Treatment a All Schools a
Enrolled in Grade 6 817 (31.6%) 690 (28.4%) 1507 (30.1%) Enrolled in Grade 7 945 (36.6%) 883 (36.3%) 1828 (36.5%) Enrolled in Grade 8 821 (31.8%) 857 (35.3%) 1678 (33.5%) Female 1295 (50.1%) 1291 (53.1%) 2586 (51.6%) Male 1288 (49.9%) 1139 (46.9%) 2427 (48.4%) African-American 2375 (91.9%) 2374 (97.7%) 4749 (94.7%) Hispanic 193 (7.5%) 49 (2.0%) 242 (4.8%) Free or Reduced Lunch 2235 (86.5%) 2175 (89.5%) 4410 (88.0%) English Language Learner 143 (5.5%) 27 (1.1%) 170 (3.4%) Total 2583 (100%) 2430 (100%) 5013 (100%)
a Percentages are based on the total numbers of students in control, treatment, or all schools.
Comparison of Students in MCLA Tre atment and Control Schools on Baseline 2006 Scores on Each Achievement Test
Unadjusted Means Adjusted Means Test Score Control Treatment Control Treatment
Est. Impact
Effect Size
Signif. Level
Total Reading Standard Score
205.7 (2235)a
204.3 (2119)
208.6 200.8 -7.8 0.34 0.003
Comprehension Standard Score
203.8 (2240)
203.3 (2133)
207.7 198.6 -9.1 0.34 0.004 ITBS
Vocabulary Standard Score
207.5 (2244)
205.3 (2129)
207.9 204.5 -3.4 0.14 0.032
Reading/LA Scale Score
502.2 (2350)
502.9 (2294)
507.8 496.2 -11.6 0.36 0.107
Mathematics Scale Score
505.4 (2347)
502.9 (2293)
507.4 500.6 -6.8 0.19 0.126
Science Scale Score
187.7 (2308)
190.2 (2285)
189.3 188.4 -0.9 0.05 0.515 TCAP
Social Studies Scale Score
193.0 (2312)
192.0 (2278)
196.3 188.5 -7.8 0.47 0.071
a Numbers in parentheses are the number of students in each group having valid test scores from the baseline 2006 administrations and the Spring 2007 administrations.
Baseline Comparisons of Students in MCLA Treatment and Control Schools
Selected Characteristics of the Year 1 Teacher Sample for MCLA Impact Analyses
a These percentages are based on different numbers of teachers due to variations in response rates to different items on the teacher survey.
Teacher Characteristic Control a Treatment a Total a
Teaches Language Arts 32.1% 37.5% 34.8% Teaches Mathematics 20.1% 19.1% 19.6% Teaches Science 17.9% 18.4% 18.1% Teaches Social Studies 19.4% 20.6% 20.0% Female 74.2% 74.2% 74.2% Male 25.8% 25.8% 25.8% African-American 86.7% 88.0% 87.4% Masters Degree or Higher 53.9% 59.8% 56.9% Licensed in Grade/Subject Taught 85.4% 79.3% 82.3% Prior MCLA Participation 13.3% 5.4% 9.3% Prof. Dev. in Integrating Literacy in Class (more than 8 hours in past 12 months)
44.2% 39.5% 41.9%
More than 5 Years Full-Time Teacher 67.8% 57.6% 62.6% More than 5 Years Full-Time at Current School 14.4% 13.3% 13.9% More than 5 Years Full-Time in Memphis 52.2% 44.4% 48.4%
All Variables Included in MCLA Impact Analytical Models for Year 1
Variable Level Coding / Range
Dependent Year-End Preparedness Index Teacher 1-5; Not at All ; A L ittle; Prepared; Well Prepared; Could Teach Others Year-End Frequency Index Teacher 1-5; Never; Rarely; Sometimes; Often; Almost Always Independent School Receiving MCLA Intervention School Yes = 1; No = 0 Covariates Baseline Preparedness Index Teacher 1-5; 5 represents highest preparedness Baseline Frequency Index Teacher 1-5; 5 represents highest frequency English Language Arts Teacher Teacher Yes = 1; No = 0 Age Teacher 1-6: 20’s; 30’s; 40’s; 50’s; 60’s; 70’s Gender Teacher Female = 1; Male = 0 African-American Teacher Yes = 1; No = 0 Masters Degree or Higher Teacher Yes = 1; No = 0 Licensed in Grade/Subject Taught Teacher Yes = 1; No = 0 Prior MCLA Participation Teacher Yes = 1; No = 0 Prof Dev in Integrating Literacy in Class Teacher 1-4: None; 1-8 hrs; 9-32 hrs; 32+ hrs Years Full Time Teacher Teacher 1-7: Never; 0-2; 3-5; 6-10; 11-20; 21-30; 30+ Years Full Time at Current School Teacher 1-7: Never; 0-2; 3-5; 6-10; 11-20; 21-30; 30+ Percentage Female (Fall 2006) School 0-100 Percentage African-American (Fall 2006) School 0-100 Percentage Special Ed (Fall 2006) School 0-100 Percentage FRL (Fall 2006) School 0-100 Percentage ELL (Fall 2006) School 0-100 School Enrollment (Fall 2006) School 400-1200
READ 180 Logic Model
R180 Study Design Analytic Approach
Study Design:
• Evaluate student outcomes using RCT based on random assignment of students to conditions across schools
• Student outcome measures: – reading achievement (ITBS) – core content areas (TCAP)
Analytic Approach:
• Cross-sectional ITT analyses of reading and core content area achievement
• Two-level models using spring ITBS and TCAP scores as a function of student and school variables
READ 180: Enrolled Students
Demographic Characteristics of the Year 1 Read 180 ITT Sample Student Characteristic Control a Treatment a Total a
Enrolled in Grade 6 392 (37.6%) 239 (34.2%) 631 (36.3%) Enrolled in Grade 7 370 (35.5%) 233 (33.4%) 603 (34.7%) Enrolled in Grade 8 280 (26.9%) 226 (32.4%) 506 (29.1%) Female 465 (44.6%) 286 (41.0%) 751 (43.2%) Male 577 (55.4%) 412 (59.0%) 989 (56.8%) African-American 955 (91.7%) 657 (94.1%) 1612 (92.6%) Hispanic 86 (8.3%) 40 (5.7%) 126 (7.2%) Free or Reduced Lunch 931 (89.3%) 619 (88.7%) 1550 (89.1%) English Language Learner 83 (8.0%) 34 (4.9%) 117 (6.7%) Total 1042 (100%) 698 (100%) 1740 (100%)
Variable Level Coding / Range
Dependent
Spring 2007 ITBS Total Reading Student Standard Score 100-350
Spring 2007 ITBS Comprehension Student Standard Score 100-350
Spring 2007 ITBS Vocabulary Student Standard Score 100-350
Spring 2007 TCAP Reading/LA Student Scale Score 300-750
Spring 2007 TCAP Mathematics Student Scale Score 300-750
Spring 2007 TCAP Science Student Scale Score 100-300
Spring 2007 TCAP Social Studies Student Scale Score 100-300
Independent
Read 180 Participation Student Yes = 1; No = 0
Variables Included in READ 180 Impact Analytic Models (Year One):Dependent and Independent
Variable Level Coding / Range
Covariates
Fall 2006 ITBS Total Reading Student Standard Score 100-350
Fall 2006 ITBS Comprehension Student Standard Score 100-350
Fall 2006 ITBS Vocabulary Student Standard Score 100-350
Spring 2006 TCAP Reading/LA Student Scale Score 300-750
Spring 2006 TCAP Mathematics Student Scale Score 300-750
Spring 2006 TCAP Science Student Scale Score 100-300
Spring 2006 TCAP Social Studies Student Scale Score 100-300
Gender Student Female = 1; Male = 0
African-American Student Yes = 1; No = 0
Hispanic Student Yes = 1; No = 0
Free or Reduced Lunch (FRL) Student Yes = 1; No = 0
English Language Learner (ELL) Student Yes = 1; No = 0
Enrolled in Grade 7 Student Yes = 1; No = 0
Enrolled in Grade 8 Student Yes = 1; No = 0
Percentage Female (Fall 2006) School 0-100
Percentage Af rican-American (Fall 2006) School 0-100
Percentage Special Ed (Fall 2006) School 0-100
Percentage FRL (Fall 2006) School 0-100
Percentage ELL (Fall 2006) School 0-100
School Enrollment (Fall 2006) School 400-1200
Variables Included in READ 180 Impact Analytic Models (Year One): Covariates
Year One Impact
Comparison of Teachers in MCLA Treatment and Control Schools on Year-End Indices for
Preparedness and Frequency of Use
Unadjusted Means Adjusted Means Test Score
Control Treatment Control Treatment Est.
Impact Effect Size
Signif. Level
Preparedness Index 3.57 (49) a
3.92 (49)
3.52 3.93 0.41 0.75 0.012
Frequency Index 3.69 (49)
3.93 (43)
3.64 4.00 0.36 0.61 0.022
a Numbers in parentheses are the number of teachers in each group having valid index scores from the baseline 2006 administration and the Spring 2007 administration.
Comparison of Students in MCLA Tre atment and Control Schools on Spring 2007
Scores on Each Achievement Test Unadjusted Means Adjusted Means
Test Score Control Treatment Control Treatment Est.
Impact Effect Size
Signif. Level
Total Reading Standard Score
208.8 (1925) a
208.8 (1831)
207.8 207.6 -0.2 0.01 0.900
Comprehension Standard Score
205.7 (1932)
205.8 (1835)
202.9 207.1 4.2 0.13 0.067 ITBS
Vocabulary Standard Score
211.8 (1938)
210.2 (1854)
211.8 208.9 -2.9 0.12 0.125
Reading/LA Scale Score
517.0 (2301)
515.1 (2240)
519.3 513.6 -5.7 0.18 0.000
Mathematics Scale Score
522.4 (2297)
515.1 (2240)
521.2 515.1 -6.1 0.17 0.061
Science Scale Score
192.2 (2212)
193.1 (2222)
193.1 192.0 -1.1 0.07 0.355 TCAP
Social Studies Scale Score
193.5 (2205)
191.4 (2212)
193.2 191.3 -1.9 0.13 0.345 a Numbers in parentheses are the number of students in each group having valid test scores from the baseline
2006 administrations and the Spring 2007 administrations.
MCLA Impacts on Students (Year One)
Comparison of Read 180 Treatment and Control Groups on
Spring 2007 Scores on Each Achievement Test Unadjusted
Means Adjusted Means
Test Score Control Treat Control Treat
Est. Imp
Effect Size
Sig.
Total Reading Standard Score
191.8 (712) a
192.9 (511)
192.6 192.1 -0.5 0.03 0.532
Comprehension Standard Score
186.7 (718)
187.6 (519)
187.0 187.0 0.0 0.00 0.976 ITBS
Vocabulary Standard Score
197.0 (726)
198.3 (519)
197.5 197.6 0.1 0.00 0.937
Reading/LA Scale Score
495.8 (972)
498.0 (664)
496.9 497.1 0.2 0.01 0.882
Mathematics Scale Score
500.0 (971)
501.8 (661)
500.0 500.2 0.2 0.00 0.904
Science Scale Score
185.1 (906)
185.6 (643)
185.6 185.1 -0.5 0.03 0.573 TCAP
Social Studies Scale Score
185.1 (906)
186.1 (644)
185.0 185.8 0.8 0.05 0.323 a Numbers in parentheses are the number of students in each group having valid test scores from
the baseline 2006 administrations and the Spring 2007 administrations.
READ 180 Impacts on Students (Year 1)
Collection of Data about Implementation Fidelity
Implications for Researchers and Practitioners
What are our purposes for collecting implementation data?
1. To provide other districts with information about outcomes they might expect when implementing similar interventions with their struggling readers*
1. To set the context for understanding student outcomes
*Requires MCS to place the needs of the field above local concerns
Reasons to Collect “Double Data”
R180 evaluation is intended to test effects of a
replicable intervention in the real-world:
1. Without the support of external evaluators
2. In ways that emulate what districts will need to do to: • monitor implementation• obtain process feedback
Reasons to Collect “Double Data”
Collecting data about MCLA and R180 fidelity
• helps researchers explain patterns of impact findings
• can be useful in identifying predictors of outcomes
What Is the Role of the Researcher?
• RBS collects data about:– Impact (MCLA & R180)– Implementation fidelity
• To better understand impact or lack thereof
(MCLA & R180)• To support development of MCLA (only)
– Counterfactual• To compare effects to what would have happened
in SR schools in the absence of MSRP
What is the Role of MCS?
• Implement R180 & MCLA
• Monitor the implementation process– Ensure implementation is “on model”– Refine service delivery based on formative
data
Defining Implementation Fidelity: MCLA
Innovation Configuration Mapping
MCLA Implementation Framework
• Developing an Innovation Configuration (IC) Map
(Hall & Hord, 2006)– Operationally defines levels of implementation fidelity among
clusters of “key active ingredients”– Iterative process involving key stakeholders
• Development team (University of Memphis)• Grantee (Memphis City Public Schools)• Researchers (Research for Better Schools)
MCLA: Roles & Responsibilities
MCS Administrators:• Participate in Principal’s Fellowship• Support recruitment and retention efforts• Link MCLA w/School Improvement Plan• Observe MCLA teachers
(once/marking period)• Allocate space for CRC materials• Protect/respect role of coach
Developer: • Design MCLA curricula
(for teachers & principals)
• Facilitate writing team activities• Meet weekly with instructors (& coaches)• Disseminate research about adolescent SR
MCLA Training
Provided by the Developer:• 3-hour weekly principal meetings
(fall;Year 1)• 3-hour weekly teacher training sessions per content area
(180 hours over 2 years)*• PD for coaches in
Mentorship; Urban education; Adolescent lit
Provided by MCS (coaches):• On-site observation of CAPs• Model/co-teach strategies• Feedback• Supplemental resources
*has included coaches since spring 2007
MCLA Innovation Configuration Map Framework
Instrument Development
With the IC map guiding development, the following
measures were designed to collect data about MCLA
implementation:• Surveys
– Teacher knowledge about & preparedness to use MCLA strategies
– Teacher demographic characteristics– Teachers’ MCLA Feedback
• Interviews– Principals, coaches, development team, and MCS administrators
• Teacher Focus Group Discussions
Operationally defining components:“Job Definition”
Aligning the IC Map and Instrument Development: “Job Definition” – Teacher Survey
“Job Definition” - Principal Interviews
MCLA Innovation Configuration Map Framework
Where the rubber hits the “runway”…
MCLA Classroom Implementation
Operationally defining components: Implementation of Lesson Plans
Implementation of lesson plans:Collecting classroom observation data
MSR-COP Data Matrix
Interval 1 Interval 2 Interval 3 Interval 4 Record Interval Start & End Times : – : : – : : – : : – :
Instructional Mode(s)
Literacy Strategy(ies)
Cognitive Demand
Level of Engagement
Instructional Mode Codes
AD Administrative Tasks J Jigsaw SGD Small-group discussion A Assessment LC Learning center/station SP Student presentation CD Class discussion L Lecture TIS Teacher/instructor interacting w/ student
DI Direct, explicit instruction related to a literacy strategy
LWD Lecture with discussion/whole-class instruction
TA Think-alouds
DP Drill and practice (on paper, vocally, computer)
OOC Out-of-class experience TPS Think-Pair-Share
GO Graphic organizer TM Teacher modeling V Visualization (picturing in one’s mind)
HOA Hands-on activity/materials RSW Reading seat work (if in groups, add SGD) WW Writing work (if in groups, add SGD)
I Interruption RT Reciprocal teaching
Cognitive Demand Codes 1 = Remember Retrieve relevant knowledge from long-term memory (recognize, identify, recall) 2 = Understand Construct meaning from instructional messages, including oral, written, and graphic
communication (interpret, exemplify, classify, summarize, infer, compare, explain) 3 = Apply Carry out or use a procedure in a given situation (execute, implement, use) 4 = Analyze Break material into its constituent parts and determine how the parts relate to one another and to
an overall structure or purpose (differentiate, organize, attribute, outline) 5 = Evaluate Make judgments based on criteria and standards (check, coordinate, monitor, test, critique, judge) 6 = Create Put elements together to form a coherent or functional whole; reorganize elements into a new
pattern or structure (generate, hypothesize, plan, design, produce, construct)
Level of Engagement Codes LE = low engagement, ? 80% of students off-task ME = mixed engagement HE = high engagement, ? 80% engaged
Implementation of lesson plans:Collecting classroom observation data
LITERACY ACTIVITY CODES
VOCABULARY STRATEGIES
B Bubble or double-bubble map M Mnemonic strategies
CC Context clue PT Preteaching vocabulary
E Etymology SFA Semantic feature analysis, maps, word grid
G Glossary or dictionary use WS Word sorts
IW Interactive word wall use
FLUENCY STRATEGIES
CR Choral reading/whole group reading RR Repeated oral reading
LM Leveled content materials TRA Teacher models/reads aloud passage
PB Paired or buddy reading
COMPREHENSION STRATEGIES
APR Activate prior knowledge PV Previewing text (T.H.I.E.V.E.S., L.E.A.R.N., and S.E.A.R.C.H.)
CT Connecting text to students’ lives
RT Retelling/summarizing with guidance Q Questioning for focus/purpose
GR Retelling with graphics MU Monitoring understanding
OR Oral retelling QAR Question-answer relationships/ ReQUEST
REF Reflection/metacognition
SGQ Students generating questions
WRITING STRATEGIES
JU Journal or blog use WR Written retelling
SW Shared writing
4.2
MCLA: Implementation Barriers
Barriers:
• Limited development/planning time• Need for coaches with disciplinary content knowledge• Challenges in establishing a critical mass of enrolled teachers at
each school• CRC materials not received until spring 2007• Pressure to focus on TCAP test preparation (spring) • Difficulty maintaining principal attendance at weekly meetings
MCLA: Planned Implementation Changes
Changes:
• Adoption of CREDE (UC-Berkeley) JPA instructional model• Reduction in the number of CAPs required of teachers• Shortened class schedule/more intensive work with coaches• Inclusion of special education teachers among those eligible to
enroll• Restructured Principal Fellowship
(includes other school leaders; meets monthly)
Defining Implementation Fidelity: R180
Rorie Harris
Memphis City Public Schools
Findings Related to Implementation
• Scheduling– Scheduling 90 minute blocks in schools using the
Middle School concept is difficult. Teams of core content teachers traditionally have 55 minute classes.
– Interruptions to the 90 minute block occur.
• Special Education Students– READ 180 will only suffice as a SPED student’s
intervention if the teacher is SPED-certified.
Findings Related to Implementation
• Use of Technology– Technology issues can negatively affect instructional
time.
• Parents & Students – Some parents do not want their children in Reading
Intervention classes. They feel like this is a “label.”– Classroom management issues impact instruction. – Student mobility affects the scope and sequence of
reading instruction.
Findings Related to Implementation
• School Administration – Without administrator “buy-in” to the importance of
smaller classes and protection of the 90 minute block, fidelity is not supported.
• Read 180 Teachers– It is challenging to encourage ALL teachers to engage
in on-line professional development and/or to attend network meetings.
– Teacher turn-over brings out the need for repeated initial training and reduces the development of teacher leaders.
Indicators of Read 180 Implementation
• Scholastic identifies several key program aspects– Teacher Training/Professional Development– Computer Hardware/Software Use– Use of Read 180 Materials– Group Rotation– Class Size– Classroom Environment– Student Engagement
Sources of Implementation Data
• Classroom observations during the school year (Fall & Spring)
• Read 180 program databases (SAM) • Enrollment and course-related data from district
databases• Surveys administered to students (Fall & Spring)
and teachers (Spring)• Information collected during professional
development programs
MCS Data Linked to Implementation Indicators
MCS Data Source Key Program Area
Completion of Scholastic RED Course
•Teacher Training
Attendance at district-wide Read 180 Network Meetings
•Teacher Training
Fall & Spring Classroom Observations
•Computer Hardware & Software Use•Group Rotations•Class Size•Classroom Environment•Use of Read 180 Materials
Enrollment Data •Class Size
MCS Data Linked to Implementation Indicators
MCS Data Source Key Program Area
Student Usage Data from SAM •Computer Hardware & Software Usage
Student Surveys •Classroom Environment•Student Engagement•Use of Read 180 Materials
Teacher Survey •Computer Hardware/Software Use•Classroom Environment•Group Rotations•Use of Read 180 Materials
Overview of Year One Conclusions
Jill Feldman, RBS
(Brief) Conclusions & Discussion
READ 180: No significant Year One student impact
• Late startup• (Most) students will receive two years of intervention
Planned Future Analyses:• Three-level analyses planned to examine whether teacher
characteristics exert a moderating effect on student outcomes
• Exploratory analyses of relationships between amount of
READ 180 instruction and effects on student outcomes
(Brief) Conclusions & Discussion
MCLA: • Significant (moderate) impact on teachers’ frequency and
preparedness to use MCLA strategies• No significant impact on students’ achievement in reading or
core content areas
Discuss:– Subjectivity of measure (“Hawthorne Effect”)– Teacher findings support program logic model– Explore relationship between impact and participation in PD
Next Steps…
Planned Exploratory Analyses
• Re-run HLM impact analyses to test effects of teacher variables on outcomes– Preparedness and use of MCLA strategies– Age– Experience as teacher (& years at MCS)– PD in year prior to MCLA
Planned/ongoing analyses
• Individual student’s growth over time
• Rerun HLM with student-level variables– # MCLA teachers– Student’s school attendance
• ITS analyses – Using TCAP Spring 2003 & 2004 scores
• Correlating R180 data with TCAP & ITBS– for possible use as covariates in HLM
Now It’s Your Turn
• Ask the panel
• Share your experiences– Triumphs– Tribulations
Thank you for joining us!
For additional information contact:
feldman@rbs.org
top related