Teaching the NYS Global Regents Enduring Issue Essay

The first time I sat in on the regional scoring* for the new New York State Global History and Geography Regents, the other scorers and I had conversations about the difficulties teaching the enduring issue concept to students. Some waited until the end of the school year to practice these, since the essay prompt draws on documents across cultures and eras and calls on the student to observe some patterns. They argued that this could not be done during the year because it required many time periods. I think there’s a better method.

Learning to write the enduring issue essay calls upon the student to read sources, identify some common issue (as opposed to a theme), and then select three documents to combine with their own recollections of historical context to relate their conclusions. Frankly, I think it’s an outstanding task. I love that it calls upon students to draw conclusions from evidence — to synthesize for themselves. The class discussions we had while practicing these were interesting as students came to observe patterns I would not have thought of and to defend them admirably. Waiting until the end of the year during review sessions is a bad time to teach students how to do this. Most all students whose papers I scored from other districts scored 2 out of 5 on these essays.

Bringing most of my students’ scores to 3-4 out of 5 came from committing ourselves to write one of these essays every ten weeks in Global 9 and 10. For those of you not familiar, in New York State students take Global History and Geography I in grade nine and the second half, part II, in grade ten. The state Regents exam only now covers the tenth grade course. I mention this because there is a temptation for Global 9 teachers to skip the enduring issue essay and wait to grade ten just before the Regents. Permit me to suggest that this is a mistake and a missed opportunity.

Click here to visit my TeachersPayTeachers.com store where you can shop for enduring issue essay prompts for grades nine and ten.

The enduring issue essay was 45% of a student’s score on my ten week interim exams. The first few in grade nine were heavily supported, as I coached students to bring up historical context to connect with the documents they chose.

The first challenge for students was to learn to distinguish a “theme” from an “issue”. “Movement of People” is a theme. The violent conflicts caused by movement of people is an issue. Students who have a more narrowly defined issue that societies have to address now have a clear direction for their writing. When students merely notice a theme, such as “power”, they fall into the trap of just proving that this was something that was a “thing” because it’s in the three documents they chose. It may seem like splitting hairs here, but it is a very important distinction and it matters in the quality of their essay (and therefore their score). The essays we wrote at weeks ten and twenty both included a lot of coaching on my part on formulating an enduring issue that was focused enough to lay the groundwork for an excellent essay.

The problem mentioned by my colleagues from other districts, namely that they felt they could not teach this until they had covered a lot of history and therefore not until the end of the year, is resolved by composing essay prompts only on the topics we have already studied. So on the week ten essay in Global 9 an 10, the documents I selected were only from the civilizations / time periods we studied at the time. The reader is invited to browse my online store for essay prompts designed for different points in the course.

The next challenge to overcome for novice writers was to avoid the temptation to merely summarize what the document says. This pretty much goes against all their reading experience to date, where teachers demanded they say what they read or answer questions on it in order to prove they understood. Writing this essay well demands that students draw upon their recalled background knowledge. The tendency for students to just summarize the documents was a very difficult habit to break. I came to advise them to spend no more than a few sentences in a paragraph to summarize the document and to spend the rest of the words in that paragraph to bring in background history to the document and to state explicitly how it supports their issue.

My goal throughout the training is to get students to write a level 3 paper. I realize I am tempted to try to teach everyone to shoot for a 5, but that is not reasonable. A score of 5 represents above grade level. A score of 4 is reserved pretty much for those who recall a lot of history. Despite all our best efforts, our students do not really on average recall a lot of history. So in my training I shoot for polishing a level of writing that is still above what I was seeing in the compositions of neighboring schools and that was do-able given the typical memory of your average student. Those students capable of the 4’s and 5’s suffered no disadvantage from this because, once they perfected the method of identifying issues (as opposed to mere themes), selecting and interpreting documents, and then bringing in historical context then all they needed to do to impress the raters was to dump a ton of historical knowledge in there.

This essay assignment is a strong feature of the new New York State Global Regents examination. It gives evidence of critical thinking and it promotes a rich classroom experience. Training in this should not wait for the end of the year and is best done throughout both grades nine and ten. The effort pays off and students often come to like this task, the latter being a surprising result of this training program.

Click here to visit my TeachersPayTeachers.com store where you can shop for enduring issue essay prompts for grades nine and ten.

* For the reader not familiar with scoring high school Regents exams in New York State, about a decade ago the state instituted regional scoring. They felt there was some kind of funny business going on when teachers were scoring their own Regents exams, so they mandated that we could no longer grade our own students’ work. So since I worked in a tiny district, I had to schlepp off to another school where I would grade their exams and they would grade mine. The security was absurd: I could not even handle my students’ papers, lest I be accused of foul play. For the essays, each student’s essay was scored by two teachers trained on a rubric and sample papers from prior field testing. If the raters disagreed in score by more than 1 point, a third rater was called in. The system has some advantages but in hindsight it seems a little unnecessary.

AI Won’t Make Teachers Obsolete

When I started teaching a long time ago, I never imagined that I would see a little laptop on each of my students’ desks and something like ChatGPT. For a little less than half my career, I taught French. Google translate never occurred to me as anything more than science fiction. Yet here we are.

Could I write an app that would grade my students’ summaries? The answer turned out to be a pretty decent “yes!”

Like all technological advances, we’re going to do it even though it wouldn’t be so bad if we didn’t. Human beings just can’t help themselves. Luddites smashed the textile machines that took their jobs, but innovation in manufacturing went on anyway. Self-restraint is not a human virtue we can ever maintain for long. Our inner drives, born in Paleolithic desperation, eventually will have their way. We cannot ban AI development because somebody, somewhere is going to do it and then those that didn’t will be at a disadvantage. All things considered, I think educators can look forward to the AI developments soon to be upon us with a sense of optimism instead of dread.

AI cannot replace teachers. One big reason is that AI cannot develop the kind of personal rapport with students that has always been the foundation youngsters need in order to learn.

My interest in artificial intelligence grew from my interest in computer programming. When I started out learning BASIC in the early 1990s, I built a program that would block vulgar words in my students’ data entry fields. We were using pre-Windows DOS machines, 286’s donated from a closing Air Force base nearby. The project occurred to me to try to make what I now know to be called a “chatbot”. I tried devising software that would converse with me in simple sentences and such that, if it didn’t know how to respond, it would ask me and then store that as an option for future response. The reader will not be surprised to find that this project did not work. In hindsight, I now know I was way, way out of my league in attempting something like that. Besides that, the computing power necessary for machine learning, let alone the troves of digital data needed to train an AI on, did not exist in 1994. But I am contented to know that I had the basic gist of the idea of machine learning that real engineers would eventually put to use.

When I submitted the algorithm-generated summary into the AI grading assistant, which evaluates it based on comparison to human-composed ones, it scored 100%. Every. Time.

About five years ago, I started exploring the idea of automating some of my grading. I was teaching social studies and I would assign summarizing as the way students were to process textbook articles. I am convinced this is a far better method that having students answer questions on text they are reading. The problem was that I had about a hundred students across six different grade levels. The beginning of a unit would generate several hundred summaries a week to grade. Could I write an app that would grade my students’ summaries? The answer turned out to be a pretty decent “yes!”

AI-Scored Summaries

The AI grading assistant here at Innovation Assessments was trained on 500 human-scored summaries. The algorithm looks at eleven features of the text and compares it to the same text features of up to seven other models of summaries scoring 100%. These features include things like a Flesch-Kincaid readability measure, word count, common proper nouns and verb phrases, and statistical comparisons like cosine and Jaccard similarity. Before analysis, the app removes stop words, reduces words to their root form (lemma) and reduces many words to a common synonym (so the app can understand ideas written in slightly different wording). The method I used to establish the scoring algorithm was to chart these comparisons in a spreadsheet and adjust them until the AI scored the work about as I would have most of the time.

I was very pleased with the results on this. The scoring of student work became a lot faster. The app brings up the AI score estimate and I can check it to confirm. This is why I call it an “AI grading Assistant”: it still needs a human supervisor. As time went on, though, I came to trust the app more and more. When I set up the assignment, I would enter my own summary of the target text from the start. Once students completed the task, I went first to score the work of students who usually get 100%. I could add up to six of these to the “corpus”, which is the body of model text the software uses to judge. The next step was to run the AI grading assistant on the work submissions of the rest of the class.

The scoring of summaries in this way required one or more human-composed models. Next, I wondered whether I could write an algorithm that would summarize a text. I am not able to write code that can write “in its own words”. Instead, my little bot mainly extracts the first sentence of each paragraph and then some selected other sentences verbatim if they meet certain criteria (such as the presence of key words identified by frequency in the text). I had my doubts about how effective this would be. Surely, it would lose some important meaning sometimes since it was a formula and not really “reading” like a human would. Well, get this …

… When I submitted the algorithm-generated summary into the AI grading assistant, which evaluates it based on comparison to human-composed ones, it scored 100%. Every. Time.

AI-Scored Short Answer Tests

Another challenge of teaching social studies with a lot of reading and writing was the large volume of grading student work in the form of short answer tests, particularly document-based analyses. Could some similar software assist in scoring short answer tests?

The app development method was about the same: I had hundreds of student work samples to analyze. Using some similar methods as for grading summaries, the new app allowed the teacher to add up to five versions of full-credit answers into the corpus for comparison. One feature that was not examined in the summary AI grading assistant was the degree to which a student’s writing was analytical (as opposed to merely descriptive). This project went fairly well – well enough for an amateur programmer and accurate enough such that the short answer scoring was a huge help to me. Click here to read more about development of an algorithm to measure the degree of analysis in a student writing sample.

The AI-assisted scoring of short answer tests was most successful at evaluating responses that had a limited range of credit-worthy answers. The AI performed well for questions like “What caused the fall of the Roman Empire?” The AI did not perform well on questions such as evaluating the reliability of a primary source, since the range of possible correct answers would have required a lot more models to train on than the five the software allows. Nonetheless, the short answer AI grading assistant saved me tons of time. It allowed me to maintain a teaching method that was very time consuming by lightening the workload so I could spend my time in curriculum development.

Opportunities for AI to Coach Students

I came to have so much confidence in the AI grading assistant that I built in access for my students. Students composing their summaries at InnovationAssessments can access the coach, which gives them a pretty accurate score estimate while they write. This take a little mystery out of “how am I doing?” and helps develop strong summarizing skills. That’s reading comprehension and basic composition.

The AI grading assistant is also an effective coach in short answer exercises. Enabling the coach for a practice run at short answer tasks permits students to have instant estimates of the quality of their work submissions and the AI offers little hints and suggestions drawn from the corpus of model answers on which it was trained.

We’re Not Being Replaced Yet

AI used in the way described here did not replace me. It still required supervision. I would assert that it enhanced my work, allowing me to use a better teaching methodology that was not very practical otherwise. The way I wanted to teach was really a recipe for burnout in the context of my particular teaching job. Assigning three summary tasks to a hundred students over a two week period, well, the reader can do the math. AI assisted scoring let me do the best job I could without burning myself out. That is a great reason to continue AI development and research, even for amateur programmers like myself.

There is a very solid reason why AI will not replace us. AI cannot replace teachers. One big reason is that AI cannot develop the kind of personal rapport with students that has always been the foundation youngsters need in order to learn. AI cannot form emotional bonds with people. If the day ever comes that it can do this, then we have something more than intelligence that is artificial, we will have a consciousness.

Can Video Replace Textbook?

In the 2021-22 academic year, I had the opportunity to finally teach a section of Regents United States History and Government. It’s something I had looked forward to after teaching US History in grades seven and eight for nearly two decades. The students were awesome – I had taught them all since they were in eighth grade since I was working in such a small school. We worked together to devise a course that would be most suitable for them and in the process we discovered that classroom textbooks can be replaced by video lessons of a certain design. News to me!

I would like to term these “enhanced video lessons” to distinguish them from a lesson plan that merely asks students to watch a video. The enhanced video lesson includes embedded questions in order of appearance in the video …

My teaching practice for social studies always entailed a strong commitment to promoting reading. I am sold on the idea that middle school students, recently relieved of their elementary reading instruction, needed continued instruction in reading and that the content areas were a fine place to do that. I invite the reader to my other blog posts on reading in the content areas. The very idea that I would abandon this commitment was a little hard to swallow.

Check out this post “Teaching with Video: Three Paths to Engagement and Accountability” on research-based video teaching.

The school year began with lessons devised in my customary fashion. I selected textbook articles that gave good summaries of the US history my students needed to learn. I offered articles at a lower reading level for slightly reduced credit for those who do not yet read on grade level. My students were given a choice of either summarizing the selected text in a specific format or doing Cornell notes on the pages. This was my reading practice, developed over almost twenty years.

I should tell you about the class, warning you first of my bias about them since they stand as one of my favorites. The class was homogeneously grouped. Nearly all were enrolled half day in a vocational education program. Their path to earning a living was to be in the trades: heavy equipment operator, welder, mechanic, machinist, forestry management, etc. Those who were still pursuing a traditional educational program all shared a general distaste for traditional academic work. The textbook – summary or Cornell notes thing had never gone over well with them so when we began they were resigned to it at best.

Click here to Visit my TeachersPayTeachers store if you want to purchase access to my social studies enhanced videos.

In the system of my units of study, the reading days alternate with presentation days. On these days, I opened the class with a lecture I would classify as “semi-formal”. That is, I brought to it a list of ideas I wanted to discuss and a general plan of presentation which I put on the board in a few words, but which was not a “formal lecture” in the traditional sense. These lasted around 15 minutes, after which students would engage with a video lesson that was a more formal lesson. These ran about 15 minutes in length. The semi-formal presentation and the video lecture amplified what was in the text and zeroed in on important or interesting things that were not.

Click Here to Try one of the Enhanced Video Lessons for yourself. Go to TestDrive. Use the code 6MNU-9FBX-A11136Z-934-JON.

Presentation days were well received. I am known to be a good presenter and we often had good discussions. The video lessons also were well-received. It is this that actually came to replace the textbook.

I would like to term these “enhanced video lessons” to distinguish them from a lesson plan that merely asks students to watch a video. The enhanced video lesson includes embedded questions in order of the video which focus on what is important to remember or which ask students to reason out something important from the video. The videos themselves I made. They were simple voice-overs of slideshows. The embedded questions could be in both multiple-choice and short-answer formats. The multiple-choice was auto-corrected so students could see how they did right away. In addition, once the class had finished, I could open access to the answers so students could see the correct ones. This could be used as a study guide before tests later on.

Class averages on assessments throughout the year

The enhanced video lessons run here at InnovationAssessments.com. They are called “Etudes”. Thanks to the input of these students, an effective and elaborate learning tool evolved!

The test results were mostly consistent throughout the year and I was very satisfied with them. The class pretest average was 37 and it moved to 67 by the final exam. The class average on unit tests (numbered 11.1, 11.2, etc.) was mostly very good. I used only questions drawn from old New York State US History and Government Regents exams so that I would be sure the questions measured the standards and had been field-tested.

Click Here to check out the final exam. Use code: L5VC-JMM5-A11812Z-8452-JON

I do not recall how the suggestion came up to abandon the textbook assignments. Maybe I felt like giving up since so few actually did the work on time even though I did not assign homework in the class. The reading class periods were often punctuated by jokes or horsing around that I have no doubt was inspired by a collective willingness to procrastinate a task they did not like doing.

A younger version of myself would have clamped down on this misbehavior with stern reprimands and consequences. This older self takes a few progressive discipline steps first, one of which is to examine the lesson itself and the learners. There were reasons I could challenge the concept of the textbook assignment lesson plan. Firstly, these were not middle school students who needed continued reading support. These eleventh graders were nearly done with their program and it is well known that reading instruction has little impact on most older learners. Secondly, the tests showed they were learning the material despite not turning in the reading.

From course pre-test, through unit tests, and then to final exam. Textbook stopped after unit 2.

It could not have been the semi-formal lectures. Though popular and often entertaining, these did not provide the breadth of information on the test. So it must have been the video lessons.

So sometime in October of that year, I offered a deal: the class would take a pretest at the start of each unit and then a post-test. The test would have items from old New York State US History and Government Regents exams. I said that, as long as the class showed adequate progress from the pretest to the post-test and as long as the standard deviation was fairly small (showing the class mostly performing clustered together), then I would not require any textbook reading.

It worked! Throughout the rest of the year, the assessments showed that the enhanced video lessons were delivering the content that students had to retain to meet the New York State standards for social studies.

This is not to say that reading was abolished. The New York State Standards for social studies are very heavily document-based. My courses had been moving in this direction for years. Our reading experiences in the course now entailed excerpts from numerous primary source documents. This required more scaffolding for these students to be able to grasp, but they seemed willing to persevere through the difficulties with my assistance.

I used only test questions drawn from old New York State US History and Government Regents exams so that I would be sure the questions measured the standards and had been field-tested.

I would not advocate for a wholesale elimination of textbooks in secondary social studies classes. I know a strong case can be made that in middle school the kind of reading tasks described above serve a very important function. But in situations like this eleventh-grade class, it proves to be a viable alternative. Granted, I had some advantages: I had been making video lessons since 2011 and I had over 400 of them across US and global history subjects. I am also a programmer and so I could code apps that did exactly what I wanted. But I think teachers without these things can still adopt this strategy. YouTube has many good history lessons made by teachers. InnovationAssessments.com offers the Etude app that I wrote for a very reasonable annual fee.

Click Here to Visit my TeachersPayTeachers store if you want to purchase access to my social studies enhanced videos.