Reflections on Public School Administration

If it’s a conceit of middle age to presume to have some accumulated wisdom or something to share, I apologize and ask your indulgence here. As I transition into retirement, I just can’t resist combing through my long career to try to find something of value.

For this exercise, I would like to imagine myself a guest speaker at a college course for aspiring school administrators. This is the kind of thing I would like to say…

Now, I’ve never been a public school administrator and nor have I ever aspired to such. I appreciate the enormous obstacles to success in this field even if not from personal experience. I have this idea that a person studying to become a school administrator could possibly find something useful in the views of a subordinate.

“Potestas (power) is the ability to do something, while auctoritas (authority) is the capacity to lead, and it comes from the respect one commands by reason of his dignity, his character, his knowledge, and his achievements. True leadership is the combination of both potestas and auctoritas.”

Cicero, De Legibus (On the Laws)

I taught middle and high school social studies. I got a chance in that time to teach about leadership and government in the past; what worked and what did not, maybe with some hypotheses as to why. An explanation from Roman civilization of note differentiated between two leadership powers: potestas and auctoritas. Potestas is the brute force to compel cooperation that societies confer upon their leaders. This is the power to force, physically or though threat of sanction, the cooperation of subordinates. Auctoritas, on the other hand, is the respect subordinates have for the leader that is born of leadership competence and compassion. Inevitably, a leader must get the cooperation of subordinates who may not choose that course of action. Leaders who rely on potestas use threats and sanctions. Leaders who employ auctoritas will get the participation of the unwilling out of respect. Roman philosophers regarded the leader who can lead with auctoritas, who seldom needs to resort to potestas, to be the superior leader. A society under such a leader will prosper and thrive in a more long-term and sustainable way.

Metaphors that work, metaphors that do not.

A metaphor that is apt for the skillful manager, the executive who operates with auctoritas: The Gardener. The gardener does not necessarily know how the tree produces its fruit, and they do not need to, but they do know how much sunlight the tree needs, what kind of solid promotes its growth, what pruning patterns best promote tree health and strong fruit yield. The good admin is a gardener who promotes the growth of their garden by creating the conditions for their success.

“Authority (auctoritas), not coercion (coercitio), creates obedience.”

Livy, Ab Urbe Condita (From the Founding of the City)

Here is a metaphor that is apt for the ineffective administrator, the executive who fails under the modus operandi of auctoritas: The School Bus Driver. The school bus driver is in charge. They are the only adult on the bus, so they are the only real expert on how to safely get to school. Everybody better sit down and behave on this ride! This admin sees themself as the central visionary of the journey; the only legitimate commander. The value of passengers is relative to their obedience, loyalty, and willingness to take the ride wherever the bus driver decides is the best road to take.

“To lead the people, walk behind them.”

Lao Tzu

The reason that the school bus driver is a bad admin is simple: teachers are not analogous to kids on a bus. Kids on a bus are young, innocent, inexperienced, limited by their immaturity. Teachers are highly educated professionals whose knowledge and experience in their subject will surpass that of the admin except for maybe subjects the admin once taught. You can tell when you’re in the school bus driver’s meeting when you realize the supervisor is hearing you but not listening to you.

A metaphor that is apt for the good admin, one who guides by auctoritas and finds success: The Tug-o-war Anchor. The Tug-o-war anchor is the person in the back of the rope. They signal the group to pull together when it is time. They arrange the team members along the rope to pull where they are most suited. The goal is created by the team, not by the leader’s particular vision or prejudices.

The first quality for a commander-in-chief is to be a man of virtue. His authority must be based not only on his power but also on his character.”

Sallust, The Jugurthine War
How can a leader best approach morale problems on their staff?

The proof of the existence of a morale problem is that someone says it; nothing more is needed. If a lot of people say it, then it is a serious problem. If many drop out of committees, it’s not a sign they are lazy, but it’s a sign they no longer have buy-in. Why did that happen?

Potestas leaders eventually have morale problems. There is high turnover is workplaces with weak leaders. Denial or debating whether things are actually all that bad denigrate the listener and trivialize their feelings in ways that have no possible good outcome. That path only earns the contempt of subordinates. An effective leader works to identify the causes of the morale issue and then seeks the counsel of those affected to arrive at reasonable remedies. Leaders accept all morale claims as unquestioningly legitimate and seek remedies as best as possible.

How can a leader best view themself in relation to the educational institution?

Unlike in the private sector where subordinates are often less skilled or knowledgeable than managers, in education the chief executive’s main purpose is to harness the potential of subordinates whose knowledge and expertise is superior to their own in their respective domains. The educational institution that reflects only the vision of the executive officer is culturally impoverished. School leaders understand they do not always know best. They seek regular information from department heads. They accept the direction to which research and regulations point even if it does not jive with their “gut”.

How can a leader enjoy the loyalty and support of subordinates even when they must take decisions they oppose?

Good leaders will have already built a reservoir of goodwill among subordinates. From time to time subordinates will need something extra or unusual: a little time off, an indulgence for a mistake, etc. Such little things are opportunities to build a reservoir of goodwill such that generosity and indulgence from time to time creates an attitude they will fall back on in times when executives must choose an unpopular but principled path. Executives in such cases are also prepared to give sound and convincing reasons for their claims that any reasonable person holding the opposing view would accept as defensible.

A leader’s power is not enough to inspire loyalty and obedience from their subordinates. Instead, a leader must possess personal virtues, such as courage, integrity, and wisdom, that inspire trust and respect in others. This idea reflects the Roman concept of virtus, or the combination of courage, excellence, and morality, as the foundation of leadership

Leaders Must Be Good Communicators

Leaders who are unskilled often communicate poorly. They often fail to convince the listener of an opposing view. They know they have the authority to impose their view and it shows in their manner. Being unresponsive is another communication flaw. Leaving messages unanswered is not a legitimate way to deny a query.

When a leader who relies on potestas attends a meeting to observe the input of subordinates, they arrive with their mind already made up and with a view to performing the show of hearing what subordinates have to say. When a leader who effectively manages with auctoritas arrives at a meeting with subordinates, the plan will be the one the group arrived at by consensus. These effective leaders say things like “once we all agree, we will…”

One should maintain an abiding and unwavering respect for the truth. One should strive to establish a workplace that is stable, predictable, and intellectually prosperous. Subordinates should not come to work wondering what will befall them that day; what aggression they will have to defend against. The ancients knew this well and it is a body of wisdom well worth considering in the present. I am not sure that good leadership can even be taught or studied. It seems very likely to me that some people are simply born with an innate leadership intelligence, a social intelligence, that other people can appreciate and respect.

Starting CRQ Work in Middle School Social Studies

The purpose for introducing the constructed response question (CRQ) in middle school is to prepare students for this kind of assessment later in their education. Ideally, the task should lay the groundwork for the habits of mind that promote success and should accustom students in a practical way to the assessment itself, its common form and its vocabulary. Experience teaching this to eighth graders shows that one of the first major obstacles is to get students to move away from the reflex ingrained in elementary school: to respond to a text by stating what it says. The second major obstacle to teaching this is that students coming out of elementary school are wholly unfamiliar with the idea that some materials they may be given are quite possibly not reliable. In addition, they lack the vocabulary to manage the concepts of text reliability.

It is difficult for upper elementary students to address primary source material for a variety of reasons. First and foremost is the text complexity. Secondly, a limited ability to comprehend given their severely limited background knowledge (class lessons should remedy this). The middle school CRQ needs to be accessible to most students while still preserving the “primary source” characteristic of the task; the opportunity to see what people of the past had to say in the way they said it. It is often a further revelation to people at this age that the English language has not always existed or that it has existed in variant forms they would find incomprehensible. An appreciation of language change and variety plays an important role in addressing primary sources for this age not only for a deeper understanding but to appreciate reliability concerns of translation and excerpts and secondhand accounts. The documents for analysis in the middle school CRQ will be carefully devised in the following ways:

  • An image of the source’s original format and language will be provided for purely observational purposes. This may be merely an incomplete image or fragment.
  • A standard translation of the source will be provided despite that it is at a text complexity above the grade level band. This is also purely for observational purposes, though some students may make the attempt to analyze it.
  • A translation of the source into a Lexile range of 800-1000 will be provided if necessary. This is the document on which students are to work.
  • There will be a citation of the source in the simplified version of the citation format used by genealogists. Students should consider the source in their analysis.

One task will consist of two pairs of documents. Students will give the historical or geographical context of the first document in each pair. Students will assess the reliability of the second document in each pair. In addressing the reliability of the source, students will need more support, naturally, than their compatriots at the high school level. The second document in each 2 pair will ask the student to address reliability in multiple-choice format. This will habituate the student to the typical phrases used in addressing reliability. First, students will be prompted in multiple-choice format to identify the document’s bias, point of view, audience, or purpose. Secondly, students will be asked to identify the best use of the document for a historian or anthropologist. Thirdly, a multiple-choice format question will ask the student to conclude how the reliability factor affects this document’s use as a reliable source of evidence to prove something specific. This latter point is important to help students understand that sources may have different reliability depending on what the historian wishes to do with them.

The last part of the task calls upon the student to synthesize a relationship between documents 1 and 2. This will also be in multiple-choice format. Students will be prompted to use both documents in one of three possible ways: (a) state a similarity or a difference between the documents; (b) explain one change associated with a turning point in history that the documents reflect (the turning point will be identified for the student); (c) explain how some development or idea is the cause of some event, idea, or historical development reflected in both documents.

Gradually toward the end of grade eight, students will move toward short answer format CRQ’s as they will see in high school. Having seen the same wording each unit across grades six through the first part of eighth, the idea of context, reliability, and turning point should be well established. A multiple-choice version for students who are still developing the skill could be offered for a reduced maximum score.

Task Models

Borrowing from the “task model” concept used in developing the New York State Global history and Geography Regents examination part one (stimulus-based multiple-choice), the following are the task models for the multiple-choice version of the middle school CRQ. It will be important to use similar language in constructing the questions for consistency.

  • Which of the following [statements | titles] best represents the [historical | geographic] context of the [document | map]? 
  • Correct answer will be historical background information not present in the document 
  • One incorrect option will be of the type “This document is about…” 
  • Correct answer will be geographic background information that explains the origin of the map’s information 
  • One incorrect option will be of the type “This map is showing is about…” 
  • Which of the following statements best represents the geographic context of the map? 
  • Which of the following would be the best use of this document for a historian? 
  • Which statement best describes the [point of view, intended audience, purpose, bias] of the document? 
  • When point of view is asked, one incorrect option will be “first person’ or “third person”. This is to teach the student to distinguish between how that term is used in English and how it is used in social science. 
  • Which of the following factor(s) would [weaken | strengthen] the reliability of this source for the purpose of __. 
  • The reliability factors taught are: authorship, format, point of view (objective or biased), time and place, intended audience, purpose. 
  • These will often have more than one correct answer. 
  • The factors are listed, followed by a colon and a description. Example: 
  • Point of view: The author is very biased. 
  • These two sources are artifacts from a turning point in history. Which would be the best title for that turning point? 
  • Which statement best describes a [similarity | difference] between the two sources? 
  • These two sources are artifacts from historical events. Which statement describes a cause-effect relationship of the historical events the sources represent? 
Assessment Task Comparison Across Three Assessments 

The purpose of developing this task is to create a logical early training step for students in middle school working toward the assessment tasks they will see in high school. 

Training the Innovation AI to Help you Grade Video Summaries

Teachers use video lessons a lot these days, some they create themselves, and other videos that they find on Youtube. Not everyone has time to build a set of cued comprehension questions for each video lesson. Some teachers assign their students to summarize what they saw in the video. This is a fantastic way to keep kids engaged in the video, but it’s a lot of work scoring them.

The Innovation AI grading assistant is the perfect solution. In this post, I would like to show you how to generate a video lesson in Innovation and then quickly train the AI to help you score the summaries very rapidly.

Quick Links to Short Video Tutorials

A Word about the Innovation AI

There are two broad categories of AI: those that work using a complex algorithm and those that work using large language model learning. The Innovation AI is of the former type. The latter, the machine-learning AI, is exemplified by ChatGPT and the Bing AI under development right now. They “learn” by analyzing vast amounts of data across the internet. The Innovation AI is trained on five to seven models that the teacher provides for comparison.

I developed the Innovation AI to help me grade summaries and short answer tests. When I taught social studies, I often assigned summaries of texts instead of comprehension questions. It works by comparing the student text to a number of models and scoring the comparison on a dozen features. These features include measures of similarity like cosine and Jaccard, as well as readability, number of words, level of text complexity, and so forth. The scoring rubric was designed using 500 of my students’ work submissions that I had scored manually so that the AI essentially grades as I would.

The Innovation AI is highly effective for helping you score summaries and short answer responses where the range of possible answers is fairly limited. The AI does effectively recognize different ways to say the same thing using natural language processing algorithms.

Training the AI

When you train the AI, you give it model answers to use in the comparison algorithm. For short answer tasks, the limit is presently five models. For writing samples such as summaries and compositions, you can store up to seven model answers.

The process, in summary, is this: (1) Compose your own summary of the task or let the Innovation AI generate a summary from the source text for you; (2) Manually score the work submissions of your students who usually get full credit. When you find a submission to which you would award full credit, ask the AI to score it. If the AI cannot recognize it as a full credit answer, you “add it to the corpus” of model answers. The next time you ask the AI to score a student submission, it will compare it to each of the models in the corpus and award the highest score earned by the student in those comparisons.

Since many of us reuse our assignments from year to year, you really only have to do this once. I trained the AI on most of my Global 9 and Global 10 assignments in 2018 and just continued to score with those for several years.

Creating an Inbox for a Task with Embedded Video

  1. Select the Inbox button from the new course playlist element dashboard.

2. Enter the title and some optional attributes. Paste in the embed code from the video you want students to watch and summarize.

Embed code from youtube.
Paste in the embed code.

3. Once created, you can click and drag the element to its right position in your class playlist.

Students Save their Summaries

When students access the task from the course playlist, they will see the video you embedded and the space underneath to compose their summary.

AI can coach students to write better!

The AI grading assistant can be engaged to coach students along the way. As they compose their summaries, they can periodically click the “Coach” button to get an estimate of their grade so far. In my experience, this promoted student prolonged engagement for a better work product.

Teaching Stimulus-Based Multiple-Choice for Document Analysis

The stimulus-based multiple-choice test item was introduced into the New York State social studies Regents examinations starting in 2019 for Global History and Geography II and for United States history. The task poses challenges for students such that it merits some regular, focused training throughout the year.

In a stimulus-based task, the student is directed to respond to a document, map, or image using their ability to analyze and their knowledge of historical context. In the case of the New York State exams, there are eighteen “task models” used when designing questions. For example, a student may be asked to evaluate and classify (identify) best use of a source or to respond based on knowledge of historical context. Principles of reliability assessment are applied here, such as when students are asked to identify point of view, purpose, context, bias, format of source, location of source in time and/or place, and/or intended audience of sources using background knowledge.

Click here to shop stimulus-based tasks at my store for grades seven through eleven social studies.

The first important habit of thought to train students to engage is to think beyond the document. Habit since their first reading lessons has asked them to find the answer in the text somehow. It takes a lot of practice and reinforcement to get students to activate their schema on the topic; to think of the story of which the document is but a fragment. The question cannot be answered without background knowledge.

  • part I of the Global and US History Regents
  • 25-30 questions
  • primary or secondary source documents
  • M-C questions are always paired with stimulus
  • primary or secondary source
  • maps
  • charts
  • cartoons
  • may have more than one stimulus tied to it
  • no more stand-alone questions
  • estimate 30-45 minutes for this part of the exam

I used almost exclusively primary source documents for my stimulus-based tasks. This can be challenging for weaker readers, but with practice in skills for addressing difficult texts, this obstacle can be addressed.

My custom was to assign a stimulus-based multiple-choice at the end of every unit starting in October. At first, students find these very difficult. I use a z-score standardization procedure to adjust the scores so as not to bomb out their GPA while they are just learning. Click here to read up on standardized scoring. It is a great way to score students in tasks they are not yet proficient at.

  • Practice! Students are generally not good at these at first. 
  • Read the question first. be certain you know what it is asking.
  • Remind yourself that the answer is rarely found in the document itself.
  • Identify the historical time period the documents go with.
  • Consciously call to mind the historical context of the document before you read. Try to recite to yourself who, what, when, where, why of the time period.
  • Use process of elimination to narrow down the options.

Some of my students always would wonder why ask questions this way. If the test writers want to know whether a student knows something, why not just ask? I don’t have a good answer for this. I strongly support instruction in social studies that calls upon students to think critically and make meaningful connections with knowledge. I also think that students should actually possess knowledge. This assessment method was no doubt inspired by AP history exams. It remains a question in my mind as to whether this level of complexity is necessary for an instrument for secondary school evaluation. If we want to know whether the student knows what caused the French Revolution, for example, maybe we should just ask them that?

Teaching the US History Regents Short Essay

The updated New York State Regents examination in United States History and Government, part II, is a short essay task designed to measure students’ ability to work with historic documents. It is a mature version of the “CRQ” found on the tenth grade Global Regents. Students are called upon to understand text, engage it with historical context, and assess a text’s reliability.

In document set 1, students describe the historical context surrounding two documents and identify and explain the relationship between the events and/or ideas found in those documents (Cause/Effect or Similarity/Difference or Turning Point).

Turning point is always the most challenging for students, mainly because it demands a strong knowledge of historical context which only the higher performing students usually possess. In stating similarities and differences, it is important to stress to students that this should be a substantial feature of the two texts, not trivial. For example, some students may respond something like this: “Document one is a cartoon and document two is a newspaper report”. This is trivial and should be discouraged. For cause and effect analyses, remind students that some events may lie outside the documents at hand, so they may need to rely on their historical knowledge.

Click here to short for short essay prompts at my TeachSimple store.

Document set 2 asks students to describe the historical context surrounding two documents and (for one identified document) analyze and explain how audience, or purpose, or bias, or point of view affects the document’s use as a reliable source of evidence.

A good strategy for ensuring students possess the skills to address this task is to be certain to assign one every month or so throughout the year, followed by a debriefing where the class can study their classmates’ work (anonymously) and develop strategies for improvement.

Weaker readers are particularly disadvantaged in this task, although since the test items are field tested before administration it is likely the field testing will mitigate some issues with the difficulty of reading some primary source texts. Students can be taught compensatory reading strategies to help deal with difficult texts.

As always, the challenge is to ensure that students have learned a strong body of historical context. That is, the best marks are reserved for those who actually recall the history and who can analyze it (cause-effect, turning point, etc.) This is best achieved by regularly administering quizzes on historical knowledge. I like to give students time in class to study for these. The apps here at Innovation Assessments are especially suited to that end. A lot of social studies assignments can tend to be just look-it-up and transfer kinds of exercises without real demands on students to remember. This is an easy instruction error to remedy.

Teaching factors affecting the reliability of sources is another matter. This takes a great deal of time and practice and, I would argue, is of upmost importance for a person’s education in this day and age. Students, for the most part, do not intuitively analyze and explain how audience, or purpose, or bias, or point of view affects a document’s use as a reliable source of evidence. They tend to take what we give them on face value. It is important to teach students to think critically and approach all historic documents with a healthy skepticism.

Click here to shop my store at TeachSimple.com

I think teaching reliability should begin young, down in middle school. Engaging students with documents that have very vague reliability weights is a good practice. In a debriefing after the task, it is useful to anonymously display some student reliability evaluations for all to see and to discuss. It is important to do this regularly, starting off right at the start of the school year. There are really no stock phrases that students can learn by rote for this, given the variety of context and source material. I had the benefit of working in a small school where I had the same students grades seven through ten or eleven, so I could implement reliability assessments early in my program. In larger school districts, it would be good to consider a commitment to reliability factor training from an early age. I assigned one longer primary source to analyze each topic (so about once a month). This short essay included an extended analysis of reliability in a conclusion paragraph. The training paid off and when my eleventh graders were preparing for the Regents in US History and Government, they had little difficulty with reliability factors.

A good piece of advice on this is to assign students to do this every month in grade eleven. I suggest assigning it as a test each time. Coach students on the historical context they have to memorize in advance. Lots of teachers assign these for homework, but this entirely misses the point of such training. Student independent work practices are highly efficient in applying the minimal effort to a task, including copying their colleagues’ work or copy-pasting from a source. If they are not doing these without notes, they’re not really practicing.

If you are afraid to assign your students this as a test because they are not likely to do well at first and don’t want to bother their GPA, I recommend using standardized scoring. You can use the z-score calculator here at Innovation Assessments. Use 78 as your standard mean and 14.8 as your standardized standard deviation. Read more about standardized scoring here and where I got those figures. The beauty of this system is you can apply this to their grades every month and as the class improves, as the class average approaches the standardized mean (78 in this case), then the algorithm affects their scores less and less.

Once you have scored their papers, select out some problem responses for class discussion in a debriefing. Keeping the responses anonymous, review how to improve the answers next time.

The short essay is 14% of the Regents score, so for passing the text it’s a good practice. But I would suggest that this kind of work as a regular lesson is extremely valuable as an educational tool. Studying primary sources rightly should take center stage in our social studies lessons. It is not just teaching to the test to do this. It is developing a critical thinking skill set and insisting on recall of historical context that are the values here. Remember, that the highest valued performance is that which is based on a substantial recall of historical context!

So here was my experience doing remote teaching in the pandemic…

I have been procrastinating writing this post. Probably because, like many of the readers who taught online during the pandemic, this is not exactly a fond memory. However, I have two reasons to move this task to the top of my to-do list. Firstly, I am aware that it has already been three years and I don’t want to forget anything important. Secondly, I will be teaching a course online next year and I want that to be a highly successful enterprise. I think there must be lessons to learn on reflection from teaching secondary level students online.

Pre-pandemic Experiments in Remote Teaching

My first foray into teaching online was a pilot online class I taught in 2016-2017. It was called Virtual United Nations. My school was a small, rural Adirondack district with limited offerings outside the core subjects and remediation. I was active on the technology committee and we found distance learning, as it was envisioned in the 1990s, was not going to work for us.

I had four students in the pilot course. It was an asynchronous course, the final project of which was participation in a regional high school United Nations simulation. For me personally, this course had the dual purpose of satisfying a graduate course requirement in digital teaching and learning through Empire State College. My paper is attached below.

Pandemic Remote Teaching for Me

At the time the pandemic hit, I was teaching social studies for grades 6-10 and a French class for grade 6. Schools in New York State closed 13 March 2020, supposedly for two weeks. I did not set foot in the school building again until September.

Like most schools in the region, our students all had Chromebooks. We scrambled to send middle and high school students home with these and chargers. Our plan evolved somewhat from its start, but eventually we settled on mostly synchronous remote classes that followed the same daily schedule as our regular day. At the end of each week, we were required to submit a summary of the week in our remote classes, including what we covered, student and teacher morale, and participation rates. I was able to search my email sent box to save copies of my reports from 13 March to 12 June 2020.

Lessons Learned

Here are some highlights of what I learned about online teaching from my weekly summaries in 2020:

  • 24 Apr I prefer delivering prerecorded video lessons to the live video. I think the live video sessions may turn out to be more PR than substance for a lot of reasons. I like them, but we’ll see. I embed my prerecorded video sessions with questions. 
  • 01 May It makes me feel happy to interact on video with [students]. I am still just learning how to conduct video classes. This week, I make them show their faces and I call on them to respond.
  • 8 May Sadly, my little Zoom “working together” idea did not work – no one showed. The streaming lessons are the highlight of my week because I get to see the kids, but I only get 6-8 now whereas I was getting the majority the first week. 
  • 08 May Education for people under ~17 needs to be a social, in-person experience and the governor’s “re-imagined” education might fail to see that. We are still primates, after all, and education requires the physical presence of teachers until upper high school. 
  • 15 May Participation in live streaming classes is very poor even with a reward system. I only get 3-6 kids out of classes of 20. These are among the highlights of my week, though. I put together nice presentations and I like seeing them.  
  • 15 May I got good participation in a survey asking successful online learners to share their secrets.  
  • 15 May I probably did not need to shorten 8th and 9th grade curriculum as much as I did. 
  • 15 May I have found breaking the tasks down to smaller steps so there is some small thing to do each day a highly effective practice for remote learning. 
  • 12 June  Breaking tasks down into daily 25-30 minute bits works great and I’ll do that in-building now too 
  • 12 June   “Remote learning” is NOT “homeschooling” and I worry that we burden parents too much. 
  • 12 June I think about 40% of my students need to be in school with a person teaching them. 

The paper “Successful Students Share Their Secret for Online Learning” is linked below. This was an extended survey of 21 (out of 87 total students on my roster that year) of my most successful students in remote learning during the pandemic. You can see for yourself the good advice they shared and some interesting survey results that may not be very surprising.

Participation

Overall, participation actually remained fairly stable, especially in grades nine and ten. I did a study in May of task completion rates of my freshmen and sophomores since 2017 and found only little changed in the pandemic. This was very surprising, since subjectively, it felt like it was a real struggle to get work in. I gave little assignments and took a grade a day. That’s how I teach in-person too, actually. The paper linked below, “Task Completion Study”, also includes a fairly comprehensive description of how I taught online. Many of those practices I retained when we returned to in-person and we had Chromebooks in class.

Remote Learning Conditions

The curriculum and assignment length were modified under remote learning. Assignments were decreased in size and duration. I avoided assigning multi-day tasks because many people just wait until the last minute to do it all, to poor effect. Students were also getting graded on “lectures” now, which consisted of 12-15 minute video tutorials embedded with questions. Some of the few multi-session tasks I normally assign in class, I now broke up into smaller assignments. The unit reading task, normally four pages to read and process, was reduced to three . In consequence of the setup of remote learning assignments, I did not complete the curriculum in grade nine to the point in history that I usually do and I moved some topics to the second part of the course to come in grade ten. I actually had more teaching time for grade ten under remote learning because, with Regents exams cancelled, I now had twelve extra lessons since I did not have to stop teaching and do review. There were on average 45 assignments across grades nine and ten in these courses over the date range. The 15% jump in the number of assignments under remote learning is because the tasks were broken down into more numerous, smaller steps. This cannot be interpreted as an increase in workload.

Resources in my online courses were curated through my own website, InnovationAssessments.com. Assignments were organized and posted at Planbook.com. Students were expected to complete one social studies assignment each school day. Courses were mostly asynchronous, save for a 30-minute live streaming lesson once a week which was optional and non-graded (there was a reward system built in for attendees). Video tutorials with multiple-choice questions ran 12-20 minutes, averaging around 14 minutes. Students were invited to attempt these twice for a higher grade. Textbook reading tasks consisted of either three- or four- page reading selection to process or multiple-choice comprehension questions. Each unit has an essay on a primary source where one session was reading and outlining the important points and one session was to compose the essay. There were a number of multiple-choice and matching quizzes which usually take about five minutes exclusive of pre-study time (students can rehearse the questions and answers in advance of some of these, so it’s very easy). Other assignment types follow similar patterns. In my opinion, a good estimate of the time commitment for the average student would be about 25 minutes a day on social studies in grades nine and ten under remote learning for the date range. I charge a late fee of ten points per day late just as in regular school and if students contact me with good reason or if students work to catch up a lot of work in a short time span, I always waive these late points.

Morale: Theirs and Mine

The short story on student morale as near as I could detect it was that it was always good. None of my weekly summaries reported adverse student morale. However, as the year progressed, some students dropped out of participation so I can only report on the students who were still involved.

My own morale story is more complex. In my memory of the events, it is all negative. But reading the contemporaneous notes of 2020, I see that it started out just fine. On 20 March, I actually wrote “My week went great. I absolutely love working from home. If I could, I would switch to just doing this.” Ugh. By my last entry on 12 June, I wrote “I endured the most serious aversion to teaching I ever experienced. This is not because it was online. I can teach online just fine. The reasons were complex… no room here… but I am getting better and look to the summer and next school year with a sense of optimism and hope.

The problems were not really connected to online teaching…

When on 12 June 2020 I wrote “The reasons were complex… no room here… ” I was alluding to a collection of things that were aggravating in the extreme. This only got worse in the coming school year. The things that made working in my district an aggravating and unpleasant experience are largely beyond the scope of this article, but these things, they were born of the pandemic. My school district imposed actions, directives, and policies which ranged from the insufferable to the outrageous. This trend gradually broke my working relationship with my district and inspired me to hasten my retirement and to request part-time work in my last year. Though I was not one of the 16% of the faculty who resigned or retired early that year in my district, I was aiming to get out as soon as feasible.

Innovations

The pandemic fostered a large number of innovations for my teaching in a digital environment. One issue of note, which I would like to address in a separate post, is the issue of online testing. Is it possible to offer valid, reliable assessments in a remote teaching situation? I am convinced that this will improve in the coming decade, but there is a lot that can be done now.

A few of the innovations that were inspired by the pandemic remote learning situation involved devising ways to give good online tests. My district subscribed to a service called “GoGuardian”, which allows me to view and control student screens. I cannot emphasize enough how important this was. But at my own site, I developed a number of useful apps. One was a parent proctoring system. Parents needed to enter a code to let students take a test. Another was a sort of AI proctor I developed in JavaScript that tracked student activity online while they worked.

I wrote a short paper on this experience. It is linked below. With the help of my colleague in the math department, we established that an examination of student scores on online tests showed no real change from performance in the classroom, thus supporting with this small study the idea that secure online testing is possible.

I am reminded now of a student who was taking one of my “credit by examination” tests. While he was writing his essay, I was able to watch his screen and my own app would prevent him from leaving the test web page in order to look something up (to cheat). I watched him type sentences I knew were not likely his — I discovered he was using his phone. Easy catch. I typed in his sentences into Google search and quickly located the web page he was copying from on his phone. It was easy to award no credit for the task and provide the very websites in rationale where he cheated.

Teaching in the pandemic for me was an adverse experience, though not for the reasons one might think. I believe it would be a good effort toward justifying the adversity if some lessons can be gleaned from my patient endurance.

The Virtue of Assigning Summaries

Dear fellow social studies teachers,

I would like to play the role of a crusader in this post, if you don’t mind. There’s something I would like you to consider about your teaching practice, or maybe reconsider as the case may be. I’d like you to abandon assigning your students comprehension questions on most kinds of texts you assign in your social studies class. Instead, I’d like you to consider assigning students to summarize instead.

When students do the search and copy instead of reading the text fully, they deny the author the chance to build with the reader an information schema of their own that lends itself to long-term recall.

Okay, so who do I think I am anyway to question a practice that’s at least a century old? Well, I cannot claim to be any super authority. I taught social studies for eighteen of my thirty-two years teaching. I worked in small, rural mountain schools and my roster averaged about one hundred students a year, which is likely smaller than that of my readers. I think these are good starters to convince you to at least hear me, but I have always felt that the argument from authority to be one of the most pernicious fallacies. So in honor of my thirty-two years, grant me the indulgence of reading further. As for accepting my suggestion, let the arguments sway you themselves.

When you assign your students to read a text selection and answer questions on it, you must know what the majority of them do. They skim the text to locate the answer, often they copy it wholesale, and then they’re done. Teachers who accept this kind of work are missing an important opportunity to foster greater reading comprehension skills and greater long-term recall. Writers, even writers of textbooks, build meaning in a schema of content to deliver to the reader, who constructs meaning related to their experience. Ideas are hierarchically arranged and organized to support a set of main ideas. When students do the search and copy instead of reading the text fully, they deny the author the chance to build with the reader an information schema of their own that lends itself to long-term recall. At the risk of putting too fine a point on it, questions on text are largely a useless exercise even when the questions are of high quality.

Once students leave sixth grade in my region, the formal lessons in reading stop. But a great number of sixth graders do not read at a sixth-grade level. They cannot advance their reading without texts at their level. And they have to actually read these. It is possible and desirable in middle and junior high school to promote the development of reading in youngsters by structuring what they do to process text. Teaching students to write competent summaries is one of the best ways to let them develop their reading skills. (This should be paired with offering texts at their independent reading level whenever possible).

What does the research say?

Summarizing and note-taking are two of the most useful academic skills students can have. (Marzano, Pickering, & Pollock, 2005)

Summarizing is in the top two most powerful writing tasks that support the development of student reading. (Graham & Hebert, 2010)

Click here to shop my TpT store for passcodes to give your students access to training videos with embedded auto-corrected questions on composing summaries.

Writing to Read

I was greatly influenced by a paper entitled Writing to Read: Evidence for how Writing can Improve Reading in 2010 published by the Carnegie Corporation of New York. A meta-analysis clearly demonstrates that composing summaries is superior to answering questions on text. The reader may argue that this research was conducted to address methods for using writing to enhance reading. But I would argue that (a) this is rightly a major goal of assigning most secondary students a textbook reading task and (b) the same cognitive processes that go into reading comprehension go into long-term recall of information. Students answering questions on text are less likely to be encoding from working memory into long-term memory than those composing summaries in their own words.

The Five and Three Summary Style

Around 2006 when I was making major investments in curriculum and methodology development, I sat down with a noted professor of literacy who happened to be working in our district, Trudy Walp. I explained my qualms about assigning students questions on text and I asked her, what is the very best thing I can have students do with text from a reading specialist’s perspective. She did not hesitate to say: have them retell what they read.

When I assigned my students summaries on their textbook articles early on, my top students wrote terribly long and detailed compositions that were not brief enough to be a summary. I needed to shorten my students’ summaries and to develop a rubric for evaluating the quality of a summary. The resulting task I used for some fifteen years: the Five and Three.

The assignment for the five and three is to summarize the assigned text (usually four pages in a typical high school textbook) in five sentences exactly, no more and no less. This is the “five”. The “three” is the personal reaction to the text. Students were assigned to connect this text to their own life somehow. What does it remind them of? What do they think about what they read? Why? These had to be exactly three sentences long. This was the second element that the reading specialist recommended to me. Students need to make the text meaningful to them in some personal way that makes new information integrate into their preexisting schema.

The five-and-three became the foundation of my students’ textbook reading work. They also had the option of composing Cornell notes for an assigned reading. I would invite the reader to return to the blog for a post dedicated to extolling the virtues of Cornell note-taking. The five-and-three faded in importance in certain classes where video lessons became an effective information delivery tool when reading instruction was no longer a priority.

The Mechanical Summary

The mechanical summary evolved in 2020 when I had a group of learners who struggled and who just would not produce summaries even given time to do so in class. I offered them the opportunity to write a summary by copying the first sentence of each paragraph word-for-word, then to connect and arrange these into complex sentences such that there were only five sentences. They had to compose the three-sentence personal reaction to the text just like normally done. For this, I offered a maximum score of 76 because, I argued, it had less value not being processed in their own words. I reason that not processing in their own words likely limited the ability of encoding to happen from working memory to long-term memory.

With the Assistance of AI

At this point you are likely wondering how you could possibly grade all these summaries. I will grant you, the workload takes some management, but I would argue that it is only slightly more time consuming than grading answers to a set of questions and the benefit to your students makes it all very worth it.

I’m an amateur programmer and I developed an AI grading assistant that is very good at scoring summaries. This app is available to subscribers to Innovation Assessments. Say, why not sign up for a free 60-day trial? The price is surely right if you decide to subscribe?

The AI grading assistant was accurate enough to save me tons of time scoring summaries. The summarizer app actually let me create models on which to train the AI that were highly accurate when compared to human-generated summaries… eerily so!

This blog post describes the AI grading assistant.

The ability to skillfully summarize is a lost art worth recovering for our students. It enhances the development of their reading ability and it promotes greater recall of the content we teach. It’s my hope my experience developing this system can make this suggestion seem do-able and worthwhile.

SOURCES

Graham, S., & Hebert, M. (2010). Writing to Read: Evidence for How Writing Can Improve Reading. New York: Carnegie Corporation.

Marzano, R. J., Pickering, D. J., & Pollock, J. E. (2005). Classroom Instruction that Works: Research-Based Strategies for Increasing Student Achievement. Upper Saddle River, NJ: Pearson Education, Inc.

Teaching the NYS Global Regents CRQ

When the new Frameworks for social studies in New York State came out about six years ago, the first item that caught my eye was the constructed-response question sets. If you are not familiar with this task, in brief, it calls upon the student to examine historical sources (I use exclusively primary sources) by providing historical or geographic context, identifying the point of view, intended audience, or purpose of a document and then using the two documents in either compare-contrast, cause-effect analysis, or turning point identification.

Originally, the task was supposed to call upon the student to evaluate the reliability of the source. This was scrapped, probably after field testing questions. I regret their decision to scrap this (though it does appear as part of the analogous task, the short essay, in US History grade 11). Students can be taught to evaluate the reliability of sources at an early age (I incorporated it into my sixth- through eighth-grade social studies work). Students can be trained to evaluate based on features like the point of view (bias), intended audience, purpose, time and place, and authorship. Teaching this does require patience and practice. What I think happened is this: few teachers address the reliability of sources and so when they field tested the new exam format, everyone did so poorly that they scrapped the question. I wish I could find out how my kids did. I had been teaching the reliability of sources since they were in middle school.

I advocate a strategy of assigning one CRQ in each unit of study in grades nine and ten without access to notes. I used this as one of the tests at the end of a unit of study. I would tell students about what kinds of documents would appear and what historical context they should be able to recall in advance. It is a challenging task for them.

The first challenge for novices is to understand what it means to provide context. Faced with the question “What is the historical context of this document?”, beginners will retell what the document says. The reason for this mistake is that, since they were little kids, teachers have asked them to relate what a text means to prove they understood it. This task asks students to bring to bear what they know about the history behind the document, which requires a level of recall I don’t think we’re asking students to do enough.

Another challenge for students writing the CRQ is the third question where they must use two documents in analysis. Cause-effect relationships are the easiest for students, it seems, so I teach those first. When students are asked to compare and contrast the documents, some students have to be cautioned to select significant elements to compare. The fact that one document is a map and one is a diary entry, for example, is not a significant fact. The turning point question is the hardest, mostly because it calls on students to recall history and to understand historical trends both before and after the event. I find it useful to break this up into small tasks: first say what the turning point is, then say what is so special about it.

After a difficult task, especially for writing, I like to do a “debriefing”. In the debriefing after the CRQ, I like to share student answers anonymously in a slide show and discuss them. It is a very effective strategy for stamping out common errors early and permanently. Sample debriefing PowerPoints from an eighth-grade and ninth-grade CRQ are posted below.

Certain constant reminders were recurring for all my classes:

— Things are not reliable because of what they say or show.

— Things are not reliable because they are in quotation marks.

— Things are not reliable because they happen to match what you think is already true.

— Unreliable statements can be true.

— When you’re asked for historical context, do not describe what the document says.

And what about middle school? Can the constructed-response task be modified for that level?

Students with an IEP often are very challenged by these tasks because of the reading level of the texts. Since I use primary sources exclusively in my CRQ’s, this is doubly difficult for students with reading limitations. The answer to this is to consider obtaining CRQ’s designed with a lower reading level.

If I may be so bold as to give advice, I would suggest that students do these throughout Global nine and ten, perhaps once a month. I suggest that they be as tests instead of open-book or take-home assignments because students need to prepare for recall situations. The debriefing afterward is vital to class improvement and it makes for a long-lasting correction. I would advocate resisting the temptation to leave this off for the Global ten teacher. It really does take a long time to learn to do this well.

Why didn’t anyone tell me about Standardized Scoring in Teacher School?

There was not a lot of math in my teacher preparation in the 1980s. Actually, I managed to take my BA without a math course (I took extra science instead). I sure wish I had statistics from the start!

Read this blog post on how it is done.

Sometimes no matter our experience or preliminary testing, we are surprised at how poorly a class does on a test. Most teachers resort, quite rightly, to some kind of “curve” to alter the scores. Even a valid, reliable test can be too difficult for students.

In effect, the standardized scoring was a useful set of training wheels that naturally disappeared once the class met the normal performance level!

A second issue of interest is in scaffolding difficult tasks for our students. Some things require time and practice to learn to do. Students may face low scores at first on such tasks. We don’t want their grades or their confidence to suffer. “Curving” the grades on a task while students are still in training for it is a good practice.

Enter the z-score standardization procedure. I don’t want this to be a post about mathematics (mostly because I am not confident to do so), but I would like to promote this as one of the best ways to alter a set of test scores in situations where (1) the group’s scores are lower than expected (say, more than 5 points below the class average overall in the course) or (2) the class is still practicing a difficult skill.

My interest in z-score began around 2010 when I was working to establish that the different capstone unit tasks I let kids choose were, in fact, of equal difficulty. I value differentiated instruction, but I also strongly value fairness. Z-score standardization let me establish how the rubrics for tasks in my class compared to a state test.

Standardizing the scores requires data that basically establishes a norm. How “should” the class have performed based on how a large set of previous students have performed? One of the things that makes grade standardizing hard is that one does not always have access to this data. How should my kids have done compared to how all my previous kids have done? Well, I saved my data.

I taught French for the first thirteen years of my career (plus two years later on) and then the other eighteen years I taught social studies. Now, mind you, I’m not a person to save a lot of stuff. My classroom was always pretty bare and I threw out stuff I wasn’t using. But data, that’s something I like to save. I have hard copies of my final grade sheets for all my students from 1994 to 2013. I also have all my Regents results (for those of you not in new York State, “Regents” are standardized state tests in different subjects). Permit me to share my data with the reader. This data will give you the mean and standard deviation on population sizes of around 100 (between 92 and 100) for French grades 8-10 and for social studies grades 8-11. See below.

I learned to do this with the help of my colleague in the math department, to whom I am grateful for answering a lot of my questions over the years and helping me learn basic statistics. I am a computer programmer and I wrote an app to do the calculations. It’s available for free and I invite you to use it. Just enter your class’ test scores, then the mean and standard deviation of the standard test you’re standardizing the scores to. The app generates a table of standardized scores and some statistical information.

A good example of using this is when I was teaching Global Studies 10 and US History 11. The new New York State Regents exams in these subjects have stimulus-based multiple-choice questions. These are hard for students at first. I had them do one each unit as a test. Standardized scoring let me modify their scores so that their grades were not harmed and their confidence preserved. So here’s the beauty of the standardized scoring: the method sets the mean of the current task to that of the standard, then adjusts everyone’s score using standard deviations. As the class improved on this task month by month, the class average approached the standard mean, so the grades were affected less and less. In effect, the standardized scoring was a useful set of training wheels that naturally disappeared once the class met the normal performance level!

Click here to shop for stimulus-based multiple-choice questions arranged by topics for Global Studies 10 and US History 11

A number of my education courses back in the ’80s were kind of useless. Hopefully, teacher training is better today. (I only had a one credit course in behavior management theory! Sheesh!) Statistics would have been a good course for me because I used it so extensively in my career. Readers who are interested might subscribe to InnovationAssessments.com to see the other statistical apps that you might find useful.

Basic Proficiency: A Classroom Differentiated Instruction Model

In the early 2000’s, there was a lot of talk about differentiated instruction picking up steam. Whether or not it was just another bandwagon remains to be seen, but it strikes me that differentiating middle and high school social studies classes is not only de rigeur right now but it is the right thing to do. Like anything else, how this is done is worth considering.

As a beginning teacher in the early ’90s, my unconscious goal was to teach the course so every kid would aim to get an A. The unspoken, unchallenged notion seemed to have been that every kid who works hard enough can, and therefore should, get an A. Disabusing myself of this notion improved my teaching practice and, therefore, my students’ progress.

I offer the analogy of a paper grocery bag, when overfilled, breaks at the bottom and all the groceries are lost. It is probably a question of cognitive load, or overload, as the case may be, that inhibits encoding of information into long-term memory.

The concept begins with the idea that the average student across the world earns probably a C or C+ in most courses. That is, after all, the definition of average. This is really the goal of instruction for the majority of students. The mandate is to pass. I was greatly influenced by a paper from 1993 (Dempster, F. N. (1993). Exposing Our Students to Less Should Help Them Learn More. Phi Delta Kappan) in which the author argued that limiting the amount of content delivered will increase how much students retain. Much instruction seems to be guided by the notion that if we inundate students with knowledge they will scoop up more of it, as much as they can. The opposite is the case. I offer the analogy of a paper grocery bag, when overfilled, breaks at the bottom and all the groceries are lost. It is probably a question of cognitive load, or overload, as the case may be, that inhibits encoding of information into long-term memory. The “Basic Proficiency” curriculum is a parallel set of modified tasks for the regular classroom that may be accessed by anybody any time whether special education or not.

The beneficiary of this program is the student who struggles to get that C+ in the course. Their performance and satisfaction is enhanced by offering less; more manageable chunks of data to process, integrate, and retain.

Click here to shop my TpT store for differentiated curriculum materials for social studies grades seven through ten.

So that’s the theory. In practice, it calls for a lot of preparation. I selected key assignments for modification: the reading task, the multiple-choice quiz, a writing task, and capstone options. The system worked thus (scroll to the bottom for sample parallel curricula): students electing basic proficiency at the start of a unit needed to declare their intention at the start. Then as the unit progressed, they accessed the modified versions of each assignment. When I moved my course materials completely online to go paperless in 2019, students changed from pulling modified assignments from a different folder to accessing a different set of links at Innovation Assessments.

For this kind of differentiated instruction to realize its full benefit, it must be paired with some remediation time outside class.

Fairness

I began developing this strategy about 2007. There are a number of key things I learned along the way. The first was to address the issue of fairness. A reasonable critique of differentiated instruction, and an element of this practice that was often overlooked by its proponents in the beginning, was the importance of ensuring that assignments that were different were actually of equivalent value. If student A is doing less than student B, how can student A reasonably expect to earn the same marks as student B? Furthermore, there were cases where strong students chose basic proficiency because they wanted an easy grade. Both of these problems were resolved by setting limits on the maximum score a student could earn on a modified task.

One of my friends in the special education department once made the case that a student with disabilities who was doing their best within their ability should be entitled to an A. I took a more conservative position on this, maintaining that the value of a work product was little influenced by the effort of the producer. There were valid and reliable ways of measuring the quality of student essays and the criteria were unaffected by ability.

Differentiated Reading

I would suggest that one of the key features of this plan, one which I embraced with some reservations at first, was to offer different levels of textbook reading for students on the basic plan. I was able to find history textbooks at a fifth grade reading level for all my classes (I was teaching US history in grades seven and eight and global history in grades nine and ten). In advance of the school year, I selected page numbers of articles that mirrored what was to be assigned from the standard textbook. Students process text in my classes using one of two methods, the “five and Three summary” (blog post coming soon on that) or Cornell Note Taking.

It should be noted that these texts contained about half as many words in a larger font and less than half the information of the standard text. Students choosing the easier reading were also choosing to learn less content. While they may learn some of the missing content from other activities such as my lecture, they were still having less served up to them.

Many assignments were the same for everybody. Each unit progressed through the same type of activities, key elements of which were offered on a modified basis. Management considerations dictated that students had to choose one package or another, Standard Inquiry or Basic Proficiency, on a unit-by-unit basis. the could not choose on a task-by-task basis. It was not possible to manage it.

Don’t Worry: There was plenty of enrichment for advanced students.

In the spirit of differentiation, I maintained a collection of college level books my ambitious learners could choose for their reading assignments. I had developed rubrics for more sophisticated versions of our essay work so students could attempt the next grade level of work. Differentiating for these students was done on a task-by-task basis and less formally, but being self-directed scholars they managed this mostly themselves. I only needed to provide the materials and encouragement to challenge them.

So How Did It Go?

I carried out the plan for a little over ten years in all my classes, grades seven through ten social studies. I found some interesting things. Firstly, I discovered that weaker students who chose basic proficiency in grades seven and eight mostly moved to standard inquiry by grades nine and ten. They tired of the maximum score limitations and they developed the skills to approach academics more effectively by having materials at their ability level to work on. Secondly, I discovered that some students who would be candidates for basic proficiency would sometimes choose standard inquiry if they liked the topic. The American Civil War was often one that had most people doing standard work. The effect of this was to give the weaker students confidence.

In 2012, I did a study of student progress on the plan to see whether I wished to continue it. Results were strong enough to continue the practice.

But then I could not do it anymore…

In my last year teaching social studies before retiring, I had to discontinue the program. The reasons were practical. My course assignments and rosters increased during this time period. A big disappointment for me was the loss of this extra remediation time in my schedule. I lost my remediation periods in the schedule in favor of teaching more courses (we taught six different subjects / grade levels where I worked). This showed me that modifying the work was sometimes not enough. Some students needed more time with their teacher outside of class and being denied this was a serious blow to my program. For this kind of differentiated instruction to realize its full benefit, it must be paired with some remediation time outside class even if only thirty minutes a week.

When I did a study of the work submission rates of my students during the remote learning of the pandemic, I discovered that there was a huge drop in work submission and homework completion overall starting after 2017 when my remedial class periods were cancelled. I invite the reader to return in the future and read my blog post on my experiences teaching in the pandemic. Pertinent to this discussion, the detrimental effect of eliminating extra help for my students was demonstrable over time.

I present this basic proficiency idea to the reader as a possibility they might consider for their classrooms. I found it to be a recipe for success, especially when paired with appropriate remediation. One needs to bear in mind that even when the assignments are all prepared in advance, there is a significant investment in time for management of the plan and for scoring a wider variety of assignments. Technology can help a lot with this (Like the Innovation Assessments learning platform!) but schools are advised to provide teachers who do this enough planning time and student contact time to make it happen. It is rewarding for the learner and well worth the investment.

Sample Parallel Curricula

Click here to shop my TpT store for differentiated curriculum materials for social studies grades seven through ten.