In the past, I was conflicted on how to perceive assessments results. During my first two years of teaching, I was extremely disappointed with the results that I would receive. Over half of the United States History students received failing grades, some as low as 45 or 50. I discovered through TVAAS, the issue was with me and my role in the classroom rather than the students. The students had no information concerning individual expectations and goals. The lessons and assessment were only sections of lectures formulated into a review day then churned into a test. These test were made from a test generator the department had received from a book company that claimed to be in line with the state standards. I was in survival mode: how I taught, how I assessed and especially how the results were analyzed.
Through implementing different types of assessments, I have challenged myself and the students to think differently about how standards are taught, assessed and analyzed. Now, at the beginning of the year, we as a department embrace data and enjoy breaking down test questions by words and phrases. At the middle of the first nine week period, I see that students are ‘hitting their stride’ or finding their pace to perform on assessments. My goal is to identify one or a group of students who are falling short and reward those who are going beyond their expected level of performance. Rather than using assessments as a 50/50 shot (you fail/ you pass), common assessments, norm referenced, formative and summative assessments shed a light on the students’ true and continuing academic performance.
Common Assessments are Vocabulary Quizzes
In the history department, we implemented common assessment three years after the push of PLCs. The assessment consisted of five questions, multiple choice, that were aligned with and taken from the item sampler online directly correlating with state SPI’s. We picked the questions based upon how often the standard appears on the practiced test. The validity of the assessment shifted to vocabulary issues rather than Historical knowledge. The most popularly missed questions consisted of political cartoons in which caricatures used a word that was beyond the students’ level of comprehension. As a department, we adjusted instructional strategies and addressing content validity by placing key words into lectures and conversation with students. We found that students became self-managing and would genuinely ask for the definition if the word was used in a causal manner. After analyzing common assessments, we mandated that the key words be integrated into formative assessments to measure the students’ performance of application skills. For manageability, we placed key questions of vocabulary focus on the first page of the formative assessment.
Integrating Norm-Referenced and Formative Assessments as a Learning Tool
While analyzing and addressing vocabulary skills, we tracked formative assessment data reliability based on how many questions the student would miss on a forty question assessment. During the period of the students, “hitting their stride”, I monitored the students achievement and compared performance based upon TVAAS projections, then converted the percentage into a baseline for how many questions should be missed on the assessment. Over a period of five weeks, the average student score would deviate from two to eight questions. The goal was for 70% of students to meet or go above their TVAAS projections. The assessments were 70% multiple choice, 30% short answer. For the class as a whole, I identified common questions the students missed on the first few assessments, linked the standard and composed spiraling questions addressing the standards in various ways. After week six, I requested that the students write down how many questions they should miss on an average assessment consisting of 40 questions. Over 80% of the students’ prediction matched or went above the TVAAS projectiles. If the student wrote down a score that was below TVAAS projections, we would have a conference to enable the students to self-monitor their progress. For example: students states they should miss ten and TVASS projections show they should miss six. The student and I set a goal for missing only eight in the next formative assessment, in one month only missing seven. The student became self-directed. In Paul Black’s words, the student has tasted their soup. The system was called a “magic number.” The students were provided clear targets and informed that if they fell below their own projections it would require remediation.
The frequencies of formative assessments were an issue that needed consideration in order to maintain reliability of the assessment findings. The timespan in between assessments was no more than two weeks at the most. If an assessment was delayed by an unforeseen issue, a short formative assessment was implemented with no prior review or forewarning. In order to receive honest data and reflection of student performance, we informed the students that the grading rubric would be more forgiving and stakes would be lowered if they did not meet their “magic number.” This version of an assessment could be an ego stroke or an opportunity to learning from mistakes. It was not a lengthy assessment, and was given in a casual manner. If the students fell short and felt guilt we identified that reaction as empowerment or taking ownership. For the purposes of self-modifying, I requested that students highlight missed standards in their individual notes.
Summative Assessments: The Showdown
Most of the pressure I observed the students placed upon themselves was during the days before a rigorous summative assessment. The stakes were raised and test rigor was based upon how the stage is set for the summative assessment. The main goal was gaining objective and reliable performance documentation. The soup was being delivered to the customer. In rigorous, high stakes norm referenced assessments, the results of positive growth had a substantial impact on increasing interest in subject matter and to improving behavioral patterns in class. The assessments were 100% teacher manufactured with 25% using multiple choice and 75% focusing on short answers. The short answers questions implemented spiral learning techniques with an objective goal of answering the questions using two to three words (if the students mastered a standard via multiple choice, the assessment question would shift to answering a question and applying knowledge in their words). Magic numbers were heavily discussed and student/class goals were provided, for instance: 70% class wide achievement reached equals extra day of enrichment activities.
When the class was motivated with reaching not only individual goals, but class-wide targets, I transitioned from recall and response pass/fail to an empowering learning coach.
Using these techniques, I was able to identify students who were falling short based upon specific SPI’s before any major high stakes assessment, as well as reward students who would not normally be recognized in other classes. Student felt empowered rather than embarrassed of their performance. Students needed, and secretly wanted, structure and feedback that could be provided through these types of communication through assessments. In the past, I handed back graded test and observed reactions of quiet apathy. Now when I given back assessments results, I observed a postive climate of students discussing questions with each other to looking for the item they missed then going into their note book and highlighting the SPI to making statements such as “not next time” or “OK, this is what I need to work on.” I felt that I made an impact by empowering students take ownership of their own targets of academic performance.
Always Adapting,
Joe Collier