hosted by: colorado department of education

55
Unified Improvement Planning: Implementation and Progress Monitoring Hosted by: Colorado Department of Education Provided by : Center for Transforming Learning and Teaching

Upload: alohilani-matthew

Post on 03-Jan-2016

19 views

Category:

Documents


0 download

DESCRIPTION

Unified Improvement Planning: Implementation and Progress Monitoring. Hosted by: Colorado Department of Education Provided by : Center for Transforming Learning and Teaching. Introductions. Session Purpose. - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Hosted by: Colorado Department of Education

Unified Improvement Planning:

Implementation and Progress Monitoring

Hosted by: Colorado Department of Education

Provided by : Center for Transforming Learning and Teaching

Page 2: Hosted by: Colorado Department of Education

Introductions

Center for Transforming Learning and Teaching

Julie Oxenford-O’Brian

Colorado Department of Education

Judy HuddlestonChristina LarsonLisa Medler

Page 3: Hosted by: Colorado Department of Education

Session Purpose

Ensure planning teams are prepared to

monitor the progress of the implementation

of their unified improvement plan.

Page 4: Hosted by: Colorado Department of Education

IntroductionsShare:

– Name, Job Title, School/District

– Your role in supporting unified improvement planning within the district

– Your most burning question about monitoring the progress of UIP implementation.

Write your question on a sticky note.

Page 5: Hosted by: Colorado Department of Education

Materials

Page 6: Hosted by: Colorado Department of Education

Capturing Notes Today

• Plan for completing Progress Monitoring using the Planning Progress Monitoring note catcher (Toolkit, p. 1).

• Capture notes for the UIP Target Setting Form and UIP Action Planning Form– Use your partially completed version– Blank versions available in the Progress Monitoring

Toolkit (p. 5 and p. 9)

Page 7: Hosted by: Colorado Department of Education

The materials used during this session were developed in partnership with the Center for Transforming Learning and Teaching in the School of Education and Human Development at the University of Colorado Denver.

Page 8: Hosted by: Colorado Department of Education

Norms

The standards of behavior by which we

agree to operate while we are

engaged in learning together.

Page 9: Hosted by: Colorado Department of Education

Session Outcomes• Understand the statutory and regulatory

requirements for monitoring the progress of UIPs.

• Identify and fully describe at least one interim measures that will be used to interpret progress made towards annual performance targets.

• Develop implementation benchmarks that will be used to determine the level of implementation of each action step associated with at least one major improvement strategy.

• Plan for how remaining interim measures and implementation benchmarks will be identified.

• Develop a calendar for when progress monitoring will occur, what data will be reviewed, and who will participate.

• Plan for engaging in collaborative data-driven inquiry process as part of regular progress monitoring.

Engage in hands-on learning activities and dialogue with colleagues.

Access additional resources.

Complete follow-up activities.

Page 10: Hosted by: Colorado Department of Education

Agenda

The Role of Progress

Monitoring

Interim Measures

PlanningProgress

Monitoring

Data-Driven Collaborative

Inquiry

Implementation Benchmarks

Page 11: Hosted by: Colorado Department of Education

Section III: Data Narrative

Section IV:Target Setting

Ongoing:Progress Monitoring

Unified Improvement Planning Processes

Describe Notable Trends

Prioritize Performance Challenges

Identify Root

Causes

Set Performance

Targets

Identify Interim

Measures

Identify Major Improvement

Strategies

Identify Implementation

Benchmarks

Gather and Organize

Data

Review Performance

Summary

Preparing to Plan Section IV:

Action Planning

Page 12: Hosted by: Colorado Department of Education

Colorado Unified Planning Template

Major Sections:

I. Summary Information about the school or District

II. Improvement Plan Information

III.Narrative on Data Analysis and Root Cause Identification

IV.Action Plan(s)

Page 13: Hosted by: Colorado Department of Education

Section I Section II Section III Section IV

Summary Information about the School/District

Additional Information about the School/ District

Improvement Plan Information

Contact Information

Progress Monitoring of Prior Year’s Targets

Data WorksheetNotable TrendsPriority Performance ChallengesRoot Causes

School Target Setting FormPriority Performance ChallengesAnnual Performance Targets (2 years)Interim MeasuresMajor Improvement Strategies

Data NarrativeDescription of School/District and Process for Data AnalysisReview Current PerformanceTrend AnalysisPriority Performance ChallengesRoot Causes

Action Planning FormMajor Improvement StrategiesAssociated Root CausesAccountability ProvisionAction StepsTimelineKey PeopleResourcesImplementation BenchmarksStatus of Action Steps

Page 14: Hosted by: Colorado Department of Education

Progress Monitoring Terminology• Consider the UIP Handbook Excerpt: Planning

Terminology (Toolkit, p.13-14).

• Mark each term using the following legend:– ? I have questions about this term.

– √ I’ve got it.

– + I could explain this term to someone else.

• Answer these questions:– What is the difference between an interim measure and an

implementation benchmark?– What is the difference between a measure and a metric?

Page 15: Hosted by: Colorado Department of Education

Where and how is Progress Monitoring represented in the UIP?

• Consider: UIP Handbook Excerpts (Toolkit, p. 11) and UIP Quality Criteria Excerpts (Toolkit, p. 15)– Interim Measures

– Implementation Benchmarks

• Answer the following questions:

– Where in the UIP template will interim measures be captured? What information should be provided? How many should be included?

– Where in the UIP template will implementation benchmarks be captured? What information should be provided?

– How should progress monitoring be included in action steps?

Page 16: Hosted by: Colorado Department of Education

Statutory and Regulatory Requirements

Consider the District Accountability Handbook Excerpts (Toolkit p. 19) to answer the following questions:

– What is the School Accountability Committee (SAC) role in monitoring the progress of the school’s implementation of their UIP?

– How frequently must the SAC be involved in progress monitoring?

– What is the District Accountability Committee role in monitoring the progress of the district’s implementation of their UIP?

Page 17: Hosted by: Colorado Department of Education

Data Analysis: UIP Development vs. Progress Monitoring

• Turn to Annual UIP Development vs. Progress Monitoring, (Toolkit, p. 53)

• Work with a partner. Use a flip chart and markers.

• Develop a double-bubble map to describe how Annual UIP Develop and Progress Monitoring are the same and how they are different.

• What would you add/change in this table?

Page 18: Hosted by: Colorado Department of Education

UIP Development vs. Progress Monitoring

UIP Development

UIP Progress

Monitoring

Page 19: Hosted by: Colorado Department of Education

Agenda

The Role of Progress

Monitoring

Interim Measures

PlanningProgress

Monitoring

Data-Driven Collaborative

Inquiry

Implementation Benchmarks

Page 20: Hosted by: Colorado Department of Education
Page 21: Hosted by: Colorado Department of Education

Agenda

The Role of Progress

Monitoring

Interim Measures

PlanningProgress

Monitoring

Data-Driven Collaborative

Inquiry

Implementation Benchmarks

Page 22: Hosted by: Colorado Department of Education

Performance Targets and Interim Measures

• What is the relationship between performance targets and interim measures?

• With a partner, share a one-sentence description of how they relate to one another.

• Interim measures should provide data several times during the school year about the degree to which progress is being made towards each performance target.

Page 23: Hosted by: Colorado Department of Education

Effective Feedback

Clear, descriptive, criterion-based, and indicate:√ how their response differed from that reflected

in the UIP quality criteria, and

√ how they can move forward (what they might do next).

Page 24: Hosted by: Colorado Department of Education

Provide Feedback about Performance Targets

• Choose a partner group/table. Exchange Target Setting Forms (with performance targets identified).

• As you review their targets, consider the following questions:

– To which students do the targets pertain? On what content is the target focused? For what metric is the target set?

– To what degree do the targets meet the relevant quality criteria?

– How could they improve their annual performance targets?

• Provide feedback to your partner group/table.

Page 25: Hosted by: Colorado Department of Education

Responding to Feedback• Consider the feedback you received

– How will you respond to the feedback you received? What will you do next to incorporate this feedback into your annual performance targets?

– How did it go providing feedback about another district’s plan? What did you learn?

• Large-group share out

Page 26: Hosted by: Colorado Department of Education

Currently Available Performance Data

• Take out your Inventory of Performance Data Sources.

• Determine which of the identified resources could be used as “interim measures”– Administered more than once a year.– Provide data related to your performance

targets.

Page 27: Hosted by: Colorado Department of Education

Describing Interim Measures• Data Source (assessment name)

• Purpose

• Related Performance Target

• Content Focus

• Which Students

• When Available

• Metrics

• Comparison Point(s)

• Questions

Consider the Describing Interim Measures: Legend (Toolkit, p. 29).

Ensure you can explain what will information should be captured in each row of this table.

Page 28: Hosted by: Colorado Department of Education

Purposes of Interim Assessment• Consider excerpts from Benchmark Assessment

for Improved Learning (Toolkit, p. 23)

• Work with a partner to answer these questions:

– What are different purposes of benchmark (or interim) assessments?

– What best describes our purpose(s) for the benchmark (interim) assessments we currently administer?

– What purpose(s) are we serving when we use interim assessments as part of our school/district improvement efforts?

Page 29: Hosted by: Colorado Department of Education

How do Performance Targets Shape Interim Measures

Interim measures should:

• Provide data several times during the school year about the degree to which progress is being made towards each performance target.

• Provide data about the same students as the performance target

• Provide data about the same content focus.

Page 30: Hosted by: Colorado Department of Education

Describing Interim MeasuresFor at least one performance target, use the Describing Interim Measures worksheet (toolkit p. 29) and identify:

•The name of the interim measure.

•The purpose(s) of administering the interim measure.

•The performance target (including on which students and on what content the target is focused) about which this interim measure will provide data.

•The content focus (or foci) for interim measure analysis.

•Student group(s) results that will be the focus of the analysis.

•When data will be available (be specific).

Page 31: Hosted by: Colorado Department of Education

Metrics (Levels)• Two levels of metrics (interim assessment results):

– Individual– Group

• Examples of groups:– All students in the school– All students in a grade level– Students in a disaggregated group– Students in a classroom

• Group metrics are aggregates of individual metrics.

Page 32: Hosted by: Colorado Department of Education

Levels and Performance MetricsLevels

• Classroom (formal)/Individual

• Aggregate school/district over-all or grade-level

• Standard/strand (grade-level)

• Disaggregated group

Performance Metric (examples)

• individual performance rating, student growth percentile, scale score.

• %/number scoring at each performance level, MGP, & AGP, average scale score.

• %/number meeting scoring proficient or better by standard

• %/number (within group) scoring at each performance level, MGP, AGP (overall and by grade-level)

Group

Individual

Page 33: Hosted by: Colorado Department of Education

Individual Metrics/Scores• Raw Scores

– Number correct– Percent correct– Number of points earned (may incorporate item difficulty)

• Standardized scores (T-Scores)• Scaled Scores (often incorporates item difficulty IRT)• Norm Referenced

– Percentile Ranks– Grade Equivalent

• Criterion Referenced– Performance Category Ratings– Lexile Rating, Estimated Oral Reading Fluency, etc.

Page 34: Hosted by: Colorado Department of Education

Comparison Points• What is a comparison point? (Consider definition

in UIP Handbook Excerpt, Toolkit, p. 13)

• Comparison points can be norm or criterion referenced.

• Norm referenced scores often embed the comparison point in the score itself.

• Each interim assessment metric that is used to evaluate progress towards performance targets, needs an associated comparison point.

• What are examples of comparison points?

Page 35: Hosted by: Colorado Department of Education

Metrics and Comparison Points for Assessments most used by Colorado

Districts

• Consider Typical Interim Assessments: Metrics and Comparison Points, (Green Sheets on your table)

• Job aide for identifying metrics and comparison points for interim and early literacy assessments most frequently used by Colorado districts.

• Are any of these “typical” interim assessments being used in your school/district?

Page 36: Hosted by: Colorado Department of Education

Identifying Metrics and Comparison Points for Interim Assessments

• Consider Report Examples

• Review several reports.

• Use Identifying Metrics and Comparison Points (Toolkit, p. 31)

• For each report reviewed, identify:– What metrics are available on the report?– To what could performance on each metric be

compared (comparison points)? How good is good enough?

Page 37: Hosted by: Colorado Department of Education

Identify your metrics and Comparison Points

• Use the Describing Interim Measures worksheet (Toolkit, p. 29).

• For at least one interim measure, identify what you will use to track and evaluate progress toward the associated performance target :– Metrics – Comparison points

• If you know, identify what reports will be used to analyze the identified metric(s) and comparison points for your performance target group and content focus?

Page 38: Hosted by: Colorado Department of Education

Questions

Different metrics make it possible to answer different questions. What metrics would help answer the following questions:

•How many students in the school are likely to score proficient or better in mathematics within the next three years?

•Are most of our students meeting minimum expectations in reading?

•How does student achievement writing compare across the grades in the school?

•Are there differences in the rates of growth in math between ELL and non-ELL students in the school?

Page 39: Hosted by: Colorado Department of Education

Questions to Guide Analysis

• Work with a partner to identify what questions can be answered with the metrics and comparison points you have identified on the Describing Interim Measures worksheet.

• Capture those questions in the final row of that work sheet for at least one interim measure.

Page 40: Hosted by: Colorado Department of Education

Planning for Progress Monitoring

• How will remaining interim measures be identified and fully described?

• Take out: Planning for Progress Monitoring (Toolkit, p. 1)

• Make notes about:– Current status– How you will complete this task– Who will complete it? When?

Page 41: Hosted by: Colorado Department of Education

Agenda

The Role of Progress

Monitoring

Interim Measures

PlanningProgress

Monitoring

Data-Driven Collaborative

Inquiry

Implementation Benchmarks

Page 42: Hosted by: Colorado Department of Education

Role of Implementation Benchmarks

Consider UIP Handbook Excerpts (Toolkit, p. 11) to answer the following questions:

– What are implementation benchmarks?

– What would not be an example of an implementation benchmark?

– Where are implementation benchmarks captured in the UIP Template?

Page 43: Hosted by: Colorado Department of Education

Different Types of Implementation Benchmarks

• Two basic “types”;

– Outputs (e.g. professional development sessions held)

– Adult Outcomes (e.g. new instructional strategy implemented)

• Can be different “types” of data too:

– Qualitative

– Quantitative

Page 44: Hosted by: Colorado Department of Education

Logic ModelA Logic Model can be used to describe your theory of:

•How major improvement strategies will result in improved student performance.

•The outputs and adult outcomes that will result from each action step.

•The outputs and adult outcomes that will result in desired changes in student outcomes (performance).

•About what implementation benchmarks should provide information.

Page 45: Hosted by: Colorado Department of Education

Example Logic ModelConsider the Example Logic Model (Toolkit, p. 33) Work with a partner to answer the following questions:

•What outputs are expected for different action steps?

•What adult outcomes are expected?

•What student outcomes are expected?

•What evidence would you collect to determine if the expected outputs and adult outcomes are being met (implementation benchmarks)?

Be prepared to share some suggested implementation benchmarks for this major improvement strategy.

Page 46: Hosted by: Colorado Department of Education

Create a Logic Model

• Select at least one major improvement strategy on which to focus.

• Use the Logic Model Template (Toolkit, p. 35) to describe the logic model behind your Major Improvement Strategy.

• Include: – Action Steps– Associated Outputs and Adult Outcomes– If time, student outcomes

Page 47: Hosted by: Colorado Department of Education

Identifying Implementation Benchmarks

• Consider your Logic Model and your Inventory of Data Sources other than Performance Data.

• Use the Identifying Implementation Benchmarks worksheet (Toolkit, p. 37).

• Capture:

– Action Steps

– Associated Outputs and Adult Outcomes

• Identify Implementation Benchmarks associated with each Output and Adult Outcome. What will count as evidence?

Page 48: Hosted by: Colorado Department of Education

Planning for Progress Monitoring

• How will remaining implementation benchmarks be identified and fully described?

• Take out: Planning for Progress Monitoring (Toolkit, p. 1)

• Make notes about:

– Current Status

– How this will be completed

– Who will complete it and when

Page 49: Hosted by: Colorado Department of Education

Agenda

The Role of Progress

Monitoring

Interim Measures

PlanningProgress

Monitoring

Data-Driven Collaborative

Inquiry

Implementation Benchmarks

Page 50: Hosted by: Colorado Department of Education

Institutionalizing Progress Monitoring

• How can progress monitoring become key to how we do business?

• Consider the following resources:

– Case study of West Denver Prep: Using Data to Drive Instruction and Improve Achievement. (Toolkit, p. 39)

– Sample Planning Calendar for Developing and Revising UIPs (Toolkit, p.47).

• Institutionalizing Progress monitoring includes building it into the regular schedule of the school/district.

Page 51: Hosted by: Colorado Department of Education

Develop a Progress Monitoring Calendar

• Take out the Progress Monitoring Calendar template (Toolkit, p.49)

• With your team, identify progress monitoring activity that will occur between December 2012 and February 2013.

• Consider:

– Could this calendar template meet our needs?

– What might we change about this basic template?

Page 52: Hosted by: Colorado Department of Education

Preparing Data for Progress Monitoring• How will your team track:

– the changes in student performance at different points during the school year, and

– progress made towards implementing your action steps?

• Consider Example Tracking Tools (packet)

– What components do you like in these tools?

– What components would you change about these tools?

– Could you adapt one of these tools for your use?

Page 53: Hosted by: Colorado Department of Education

Planning for Progress Monitoring

• Turn to Planning for Progress Monitoring (Toolkit, p. 2)

• Make notes about how you will:– Develop a calendar for progress monitoring

sessions.– Determine how data will be tracked through

out the year.– Build capacity to engage in data-driven

collaborative inquiry.

Page 54: Hosted by: Colorado Department of Education

Preparing for Progress Monitoring Sessions

• Checklist for Preparing for Progress Monitoring Sessions (Toolkit, p. 3)

• Tasks include:– Determine what data will be reviewed:

– Generate appropriate interim assessment reports.

– Organize and prepare implementation benchmark data.

– Identify guiding questions

– Schedule time

– Determine participants.

• How could you use this checklist?

Page 55: Hosted by: Colorado Department of Education

Give us Feedback!!• Written: Use sticky notes

– + the aspects of this session that you liked or worked for you.

The things you will change in your practice or that you would change about this session.

– ? Question that you still have or things we didn’t get to today

– Ideas, ah-has, innovations

• Oral: Share out one ah ha!