Welcome to the week on technology supports for data-driven (DDDM) and evidence-based (EBDM) decision making in schools. Please read the following to prepare for lab.
Making Decisions about how best to teach is never an easy process. For this week’s lab, you should begin by considering what it might be like to be in an annual teacher evaluation meeting with your principal. You can get a sense of this by viewing This Teacher Evaluation video that includes a typical review of student progress that is, in part, as a metric of successful teaching. [WARNING: the audio on this video is not of high quality, as is often the case with real classroom video. So listen closely to what the principal (seated on the left behind the desk, and the teacher seated on the right) are discussing).] As you view this teacher evaluation meeting, consider the following:
- How did the 5th grade teacher generate these graphs? Do you see any other such charts around the principal’s desk?
- (around the 0:53 mark) In the classrooms you visited, how could you obtain data from the same students for 4 years (in the video it is 2nd grade to 5th grade)? Don’t students come in and out of the classroom? How did she get these data?
- (around the 1:00 mark) What does the grading language of “moving from ‘basic,’ to ‘proficient,’ and ‘advanced'” mean or indicate to you?
- (around the 1:30 mark) the 5th grade teacher makes the point that students should set goals to “maintain” rather than be ‘advanced’ – how does that relate to rubric grades in EPSY 5221?
- How do these educators use proximal level assessments to make decisions about individual students?
- How do these same results, that guide instruction, also speak to teacher skills? Are they valid for both purposes?
- (around the 2:13 mark) A specific student’s name is mentioned, and the teacher and principal exchange comments- drawing from your Ethics module content, what concerns does this event (in a YouTube public video) raise for you?
First, you should understand that there are many types of data, quantitative and qualitative, and that there may be additional forms of evidence that can be used to understand the impact of instruction and teacher success. To make this point, consider this diagram that shows that teachers can do many small “tests” and observations at the “Close” level, even within a single class period, but such data must be quick and easy to obtain, analyze, and use.
But they also can obtain data from unit end tests at the “Proximal” level, that test knowledge, attitudes, and skills related to the standards-based curriculum. And of course, there are always those high-stakes tests like the Smarter-balanced tests, the CMT and the SATs, at the “Distal” level, that we all love, that are norm referenced on national samples and have purposes well beyond a single classroom or teacher’s work.
All of this testing is often used to compute alphanumeric report card grades (e.g., 95% A). Certainly we have all experienced this system of grading and its limitations. In comparison, Standards-based or mastery-based grading, provide evidence of student success while removing confusing factors such as effort/attitude, extra credit, and student behavior (such as submitting things precisely on time). These other factors are graded separately from content-related mastery based on State and National standards.
PARCC and Smarter Balanced.
Testing in K-12 schools has changed, mainly related to technology access. Traditional standardized tests (like the SAT MCAT GRE LSAT) have been upgraded with artificial intelligence in the form of Computer Adaptive Test. The Smarter Balanced test is computer adaptive and so it changes with each answer students give, and then selects the next question presented based on the test takers history. Underlying the approach of computer adaptive testing are Item Response Theory, which uses research to optimize the diagnostic value of each wrong answer (distractors). Here are some slides with the core ideas of CAT and a 16 minute YouTube tutorial about key concepts. More issues with using CATs are address in “This test is for you.” Note ETS (for the GRE) and the College Board (for the SAT) are also considering test revisions that might incorporate computer adaptive programming and perhaps alignment with content objectives.
Purposes of Assessments.
Making the Grade.. On your mark… get set…
OK.. so reporting percent correct on a single test is different from calculating percentages across topics or units. If, as a future physician, I get an an A in my knowledge of the heart but a C on my knowledge of the brain, does that mean I am a B doctor, or would your still prefer me as your heart surgeon rather than your brain surgeon? Are grades supposed to reflect how much students have learned (criterion-based) or are they meant to show the relative performance of students to rank the top students (normative)? Grades cannot do both these things. Can you meaningfully give yourself or anyone, 1 single grade that meaningfully represents their knowledge of world history, or is it inevitable that I may know a bit more about my home country than I know about all other countries? Any what about effort and extra credit– should a student’s final grade reflect their level of effort as well as the level of their final achievement of understanding? Tom Gruskey asks these questions in his slides and handouts to get future teachers to think more deeply about how and why they grade and what those score mean when used for various purposes.
The role of Technology in DDDM
Take a peek at Infinite Campus, a school-based data system being widely adopted, including in my favorite schools in Ellington CT. Teacher, parents, and even students will have mobile access to a variety of school performance data, to help with decision making. Consider what it will be like to be a teacher in schools implementing these parent, teacher and student dashboard to data. Moving beyond data-driven and evidence-based toward “purpose driven” decision making with a variety of types of evidence, the technology is driving the availability of data and teachers will need to be fluent in data reduction and data summarization to access and help others makes sense of all types of data.
Resources
- Let Star Trek’s Capt. Kirk define DDDM for you! (humor value here)
- Youtube providing an overview of Standards-based Grading and its comparison to more common assessments
- Sample Gradebook (xls) Download this for Lab
- Image of standards-based gradebook
- MS Toolpack (with t-test and other stats) for Excel
- Free online t-test tool (for you non-Excel and google cloud users)
- R the free statistic package (if you want to go full bore into data analysis)
Technologies
- Classtime
- Poll Everywhere
- Socrative
- Kahoot!
- Google Forms for surveys
- Mentimeter for interactive presentations
Supplemental: References and Read More about…
t-tests vs ANCOVA analysis of gain scores
Analyzing gain scores vs using the pretest as covariate– the saga continues
Stiggins, R. (2018). Better Assessments Require Better Assessment Literacy. Educational Leadership, 75(5), 18-19.
Enduring Understanding- Student Learning Outcome from this module
This module intends to prepare teachers for ISTE Standard for Educators 7c that states: Use assessment data to guide progress and communicate with students, parents and education stakeholders to build student self-direction.
- Understand the Learning Sciences approach to co-research in schools
- Discuss the relative value of mastery vs alphanumeric grades
- Interrogate how to prepare students for Computer-Adaptive Tests