ESL Essay Writing Rubric

  • Resources for Teachers
  • Pronunciation & Conversation
  • Writing Skills
  • Reading Comprehension
  • Business English
  • TESOL Diploma, Trinity College London
  • M.A., Music Performance, Cologne University of Music
  • B.A., Vocal Performance, Eastman School of Music

Scoring essays written by English learners can at times be difficult due to the challenging task of writing larger structures in English. ESL / EFL teachers should expect errors in each area and make appropriate concessions in their scoring. Rubrics should be based on a keen understanding of English learner communicative levels . This essay writing rubric provides a scoring system which is more appropriate to English learners than standard rubrics. This essay writing rubric also contains marks not only for organization and structure, but also for important sentence level mistakes such as the correct usage of linking language , spelling , and grammar.

Essay Writing Rubric

  • ESL Presentation Rubric
  • How to Teach Essay Writing
  • Beginning Level Curriculum for ESL Classes
  • Writing Cause and Effect Essays for English Learners
  • Learn English
  • Beginning Guide to Teaching ESL
  • Standard Lesson Plan Format for ESL Teachers
  • Synonyms and Antonyms for ESL
  • ESL Tips to Improve Your English Online
  • Beginning Writing Short Writing Assignments
  • Sentence Connectors and Sentences
  • English Teaching Abbreviations Explained
  • Top Lesson Plans for ESL and EFL
  • Sentence Type Basics for English Learners
  • How to Build an ESL Class Curriculum
  • CALL Use in the ESL/EFL Classroom

Essay Rubric

Essay Rubric

About this printout

This rubric delineates specific expectations about an essay assignment to students and provides a means of assessing completed student essays.

Teaching with this printout

More ideas to try.

Grading rubrics can be of great benefit to both you and your students. For you, a rubric saves time and decreases subjectivity. Specific criteria are explicitly stated, facilitating the grading process and increasing your objectivity. For students, the use of grading rubrics helps them to meet or exceed expectations, to view the grading process as being “fair,” and to set goals for future learning. In order to help your students meet or exceed expectations of the assignment, be sure to discuss the rubric with your students when you assign an essay. It is helpful to show them examples of written pieces that meet and do not meet the expectations. As an added benefit, because the criteria are explicitly stated, the use of the rubric decreases the likelihood that students will argue about the grade they receive. The explicitness of the expectations helps students know exactly why they lost points on the assignment and aids them in setting goals for future improvement.

  • Routinely have students score peers’ essays using the rubric as the assessment tool. This increases their level of awareness of the traits that distinguish successful essays from those that fail to meet the criteria. Have peer editors use the Reviewer’s Comments section to add any praise, constructive criticism, or questions.
  • Alter some expectations or add additional traits on the rubric as needed. Students’ needs may necessitate making more rigorous criteria for advanced learners or less stringent guidelines for younger or special needs students. Furthermore, the content area for which the essay is written may require some alterations to the rubric. In social studies, for example, an essay about geographical landforms and their effect on the culture of a region might necessitate additional criteria about the use of specific terminology.
  • After you and your students have used the rubric, have them work in groups to make suggested alterations to the rubric to more precisely match their needs or the parameters of a particular writing assignment.
  • Print this resource

Explore Resources by Grade

  • Kindergarten K

Two ESL Writing Rubric Templates

Providing timely and high-quality feedback to students is key to ensure they make consistent progress, but grading (English as a Second Language) ESL writing assignments is time-consuming. An ESL writing rubric can make the process significantly easier for you and transparent for your students.

In this blog post, we will show you two ESL writing rubric templates that you can use to grade your student’s writing. But first, let’s understand what is an ESL writing rubric.

What is an ESL Writing Rubric?

Grading written essays objectively and consistently is difficult without standardized rubrics. With an ESL writing feedback rubric teachers can provide detailed, personalized, fair and unbiased feedback to their students.

Rubric-based feedback is helpful for students because it clearly communicates their instructor's expectations and how they measure against it, so don’t forget to craft a rubric that very clearly captures expectations and your feedback for the assignment.

ESL Writing Rubric Template

Esl essay writing rubric template.

Thank you for submitting your assignment. You can read my feedback below: Purpose {formtoggle: name=Clearly defined; default=yes}Purpose is clearly defined; includes many details that connect talking points and develop the assignment; may include minor irrelevant information.{endformtoggle: trim=yes} {formtoggle: name=good, not well supported; default=no}Purpose is comprehensible and defined; includes some details, although not all points are adequately supported; some irrelevant information.{endformtoggle: trim=yes} {formtoggle: name=not comprehensible, loose connections; default=no}Purpose is defined, but may not be comprehensible for readers who don’t know the writer; few details included; loose connections; a lot of irrelevant information.{endformtoggle: trim=yes}{formtoggle: name=no supporting details; default=no}Attempts to define the purpose; few to no supporting details.{endformtoggle} Organization {formtoggle: name=coherent and logical; default=yes}Coherent and logical throughout the assignment; clear beginning, middle and end; smooth transitions used to connect ideas.{endformtoggle: trim=yes} {formtoggle: name=follows a logical sequence; default=no}Assignment follows a logical sequence; contains a beginning, middle and end.{endformtoggle: trim=yes} {formtoggle: name=beginning or end is unclear; default=no}Attempts to follow a logical sequence; beginning or ending is unclear or ends abruptly.{endformtoggle: trim=yes} {formtoggle: name=incoherent, disconnected ideas; default=no}No coherent order; written with sentence fragments or disconnected ideas.{endformtoggle} Vocabulary {formtoggle: name=wide vocab; default=yes}Wide variety of vocabulary used; words help expand the topic; minor inaccuracies that do not affect the reader's ability to comprehend the assignment.{endformtoggle: trim=yes} {formtoggle: name=good vocab, few errors; default=no}Variety of words related to the topic; few errors.{endformtoggle: trim=yes} {formtoggle: name=basic vocab, unrelated words; default=no}Uses basic vocabulary; some words are not related to the topic.{endformtoggle: trim=yes} {formtoggle: name=limited vocab, very inaccurate; default=no}Uses a limited vocabulary that is below the expected level; vocabulary is largely inaccurate.{endformtoggle} Grammar/Punctuation/Conventions {formtoggle: name=excellent; default=yes}The writer demonstrates an excellent understanding of English conventions; limited errors that do not affect the reader’s ability to comprehend the assignment.{endformtoggle: trim=yes} {formtoggle: name=average; default=no}Demonstrates an average understanding of English conventions; errors are unlikely to affect reader comprehension; low amount of errors.{endformtoggle: trim=yes} {formtoggle: name=frequent errors; default=no}Shows some understanding of English conventions; frequent errors; errors hinder reader’s ability to comprehend the assignment.{endformtoggle: trim=yes} {formtoggle: name=incomprehensible; default=no}Demonstrates lower than expected knowledge of English conventions; the piece is nearly incomprehensible.{endformtoggle}

Fluency {formtoggle: name=noticeable effort to write; default=yes}The writer does not use translation; puts noticeable effort to write like a native speaker.{endformtoggle: trim=yes} {formtoggle: name=very few errors; default=no}Very little translation is used in the assignment; effort to write like a native speaker; errors are few.{endformtoggle: trim=yes} {formtoggle: name=hard to understand; default=no}Heavy translation throughout the document; writing is hard to understand.{endformtoggle: trim=yes} {formtoggle: name=incomprehensible; default=no}Writing is incomprehensible.{endformtoggle}

Hi there! You made it all the way down to the bottom of this article. Take a few seconds to share it.

Want to turbo charge your work with templates and snippets? Text Blaze is the fastest way to do that.

120 ESL Report Card Comments in 2024

eslwriting.org

A panoply of teaching resources.

ESL Writing: A Simple Scoring Rubric

Anyone who teaches ESL writing knows there comes a time when student papers pile up on the desk and they have to be marked. The hard part, when faced with 30 or more regular submissions, is finding a way to give English writing students meaningful input without spending hours generating feedback.

A writing rubric is one possible solution.

Below is a suggested scoring guide. It is simple, yes. Though I hope it provides useful guidance.

This rubric is suitable for one paragraph writing submissions. In addition, it is useful for classes where the emphasis is on three main elements:

  • a strong topic sentence (control sentence)
  • several sentences of supporting material (reasons, evidence or details)
  • a conclusion

ESL Writing Rubric

This ESL writing rubric uses a 1 to 4 scoring system.

  • A well-structured paragraph (clear topic sentence, good support and conclusion). Grammar and spelling errors are present though the meaning of the passage is clear.
  • A good topic sentence though the paragraph lacks unity (the main idea is not well supported) or there are too many grammar or spelling errors.
  • It looks like a rough draft. The topic sentence is not well formulated, the supporting data is unclear and there are too many errors.
  • Incomplete. It is not possible to understand the meaning of the paragraph because key sections are missing or they cannot be understood.

Leave a Comment

This site uses Akismet to reduce spam. Learn how your comment data is processed .

  • help_outline help

iRubric: ESL Writing Assessment (Intermediate) rubric

  • writing, esl, intermediate, composition
  • Foreign Languages

essay writing rubric esl

Rubric Best Practices, Examples, and Templates

A rubric is a scoring tool that identifies the different criteria relevant to an assignment, assessment, or learning outcome and states the possible levels of achievement in a specific, clear, and objective way. Use rubrics to assess project-based student work including essays, group projects, creative endeavors, and oral presentations.

Rubrics can help instructors communicate expectations to students and assess student work fairly, consistently and efficiently. Rubrics can provide students with informative feedback on their strengths and weaknesses so that they can reflect on their performance and work on areas that need improvement.

How to Get Started

Best practices, moodle how-to guides.

  • Workshop Recording (Fall 2022)
  • Workshop Registration

Step 1: Analyze the assignment

The first step in the rubric creation process is to analyze the assignment or assessment for which you are creating a rubric. To do this, consider the following questions:

  • What is the purpose of the assignment and your feedback? What do you want students to demonstrate through the completion of this assignment (i.e. what are the learning objectives measured by it)? Is it a summative assessment, or will students use the feedback to create an improved product?
  • Does the assignment break down into different or smaller tasks? Are these tasks equally important as the main assignment?
  • What would an “excellent” assignment look like? An “acceptable” assignment? One that still needs major work?
  • How detailed do you want the feedback you give students to be? Do you want/need to give them a grade?

Step 2: Decide what kind of rubric you will use

Types of rubrics: holistic, analytic/descriptive, single-point

Holistic Rubric. A holistic rubric includes all the criteria (such as clarity, organization, mechanics, etc.) to be considered together and included in a single evaluation. With a holistic rubric, the rater or grader assigns a single score based on an overall judgment of the student’s work, using descriptions of each performance level to assign the score.

Advantages of holistic rubrics:

  • Can p lace an emphasis on what learners can demonstrate rather than what they cannot
  • Save grader time by minimizing the number of evaluations to be made for each student
  • Can be used consistently across raters, provided they have all been trained

Disadvantages of holistic rubrics:

  • Provide less specific feedback than analytic/descriptive rubrics
  • Can be difficult to choose a score when a student’s work is at varying levels across the criteria
  • Any weighting of c riteria cannot be indicated in the rubric

Analytic/Descriptive Rubric . An analytic or descriptive rubric often takes the form of a table with the criteria listed in the left column and with levels of performance listed across the top row. Each cell contains a description of what the specified criterion looks like at a given level of performance. Each of the criteria is scored individually.

Advantages of analytic rubrics:

  • Provide detailed feedback on areas of strength or weakness
  • Each criterion can be weighted to reflect its relative importance

Disadvantages of analytic rubrics:

  • More time-consuming to create and use than a holistic rubric
  • May not be used consistently across raters unless the cells are well defined
  • May result in giving less personalized feedback

Single-Point Rubric . A single-point rubric is breaks down the components of an assignment into different criteria, but instead of describing different levels of performance, only the “proficient” level is described. Feedback space is provided for instructors to give individualized comments to help students improve and/or show where they excelled beyond the proficiency descriptors.

Advantages of single-point rubrics:

  • Easier to create than an analytic/descriptive rubric
  • Perhaps more likely that students will read the descriptors
  • Areas of concern and excellence are open-ended
  • May removes a focus on the grade/points
  • May increase student creativity in project-based assignments

Disadvantage of analytic rubrics: Requires more work for instructors writing feedback

Step 3 (Optional): Look for templates and examples.

You might Google, “Rubric for persuasive essay at the college level” and see if there are any publicly available examples to start from. Ask your colleagues if they have used a rubric for a similar assignment. Some examples are also available at the end of this article. These rubrics can be a great starting point for you, but consider steps 3, 4, and 5 below to ensure that the rubric matches your assignment description, learning objectives and expectations.

Step 4: Define the assignment criteria

Make a list of the knowledge and skills are you measuring with the assignment/assessment Refer to your stated learning objectives, the assignment instructions, past examples of student work, etc. for help.

  Helpful strategies for defining grading criteria:

  • Collaborate with co-instructors, teaching assistants, and other colleagues
  • Brainstorm and discuss with students
  • Can they be observed and measured?
  • Are they important and essential?
  • Are they distinct from other criteria?
  • Are they phrased in precise, unambiguous language?
  • Revise the criteria as needed
  • Consider whether some are more important than others, and how you will weight them.

Step 5: Design the rating scale

Most ratings scales include between 3 and 5 levels. Consider the following questions when designing your rating scale:

  • Given what students are able to demonstrate in this assignment/assessment, what are the possible levels of achievement?
  • How many levels would you like to include (more levels means more detailed descriptions)
  • Will you use numbers and/or descriptive labels for each level of performance? (for example 5, 4, 3, 2, 1 and/or Exceeds expectations, Accomplished, Proficient, Developing, Beginning, etc.)
  • Don’t use too many columns, and recognize that some criteria can have more columns that others . The rubric needs to be comprehensible and organized. Pick the right amount of columns so that the criteria flow logically and naturally across levels.

Step 6: Write descriptions for each level of the rating scale

Artificial Intelligence tools like Chat GPT have proven to be useful tools for creating a rubric. You will want to engineer your prompt that you provide the AI assistant to ensure you get what you want. For example, you might provide the assignment description, the criteria you feel are important, and the number of levels of performance you want in your prompt. Use the results as a starting point, and adjust the descriptions as needed.

Building a rubric from scratch

For a single-point rubric , describe what would be considered “proficient,” i.e. B-level work, and provide that description. You might also include suggestions for students outside of the actual rubric about how they might surpass proficient-level work.

For analytic and holistic rubrics , c reate statements of expected performance at each level of the rubric.

  • Consider what descriptor is appropriate for each criteria, e.g., presence vs absence, complete vs incomplete, many vs none, major vs minor, consistent vs inconsistent, always vs never. If you have an indicator described in one level, it will need to be described in each level.
  • You might start with the top/exemplary level. What does it look like when a student has achieved excellence for each/every criterion? Then, look at the “bottom” level. What does it look like when a student has not achieved the learning goals in any way? Then, complete the in-between levels.
  • For an analytic rubric , do this for each particular criterion of the rubric so that every cell in the table is filled. These descriptions help students understand your expectations and their performance in regard to those expectations.

Well-written descriptions:

  • Describe observable and measurable behavior
  • Use parallel language across the scale
  • Indicate the degree to which the standards are met

Step 7: Create your rubric

Create your rubric in a table or spreadsheet in Word, Google Docs, Sheets, etc., and then transfer it by typing it into Moodle. You can also use online tools to create the rubric, but you will still have to type the criteria, indicators, levels, etc., into Moodle. Rubric creators: Rubistar , iRubric

Step 8: Pilot-test your rubric

Prior to implementing your rubric on a live course, obtain feedback from:

  • Teacher assistants

Try out your new rubric on a sample of student work. After you pilot-test your rubric, analyze the results to consider its effectiveness and revise accordingly.

  • Limit the rubric to a single page for reading and grading ease
  • Use parallel language . Use similar language and syntax/wording from column to column. Make sure that the rubric can be easily read from left to right or vice versa.
  • Use student-friendly language . Make sure the language is learning-level appropriate. If you use academic language or concepts, you will need to teach those concepts.
  • Share and discuss the rubric with your students . Students should understand that the rubric is there to help them learn, reflect, and self-assess. If students use a rubric, they will understand the expectations and their relevance to learning.
  • Consider scalability and reusability of rubrics. Create rubric templates that you can alter as needed for multiple assignments.
  • Maximize the descriptiveness of your language. Avoid words like “good” and “excellent.” For example, instead of saying, “uses excellent sources,” you might describe what makes a resource excellent so that students will know. You might also consider reducing the reliance on quantity, such as a number of allowable misspelled words. Focus instead, for example, on how distracting any spelling errors are.

Example of an analytic rubric for a final paper

Example of a holistic rubric for a final paper, single-point rubric, more examples:.

  • Single Point Rubric Template ( variation )
  • Analytic Rubric Template make a copy to edit
  • A Rubric for Rubrics
  • Bank of Online Discussion Rubrics in different formats
  • Mathematical Presentations Descriptive Rubric
  • Math Proof Assessment Rubric
  • Kansas State Sample Rubrics
  • Design Single Point Rubric

Technology Tools: Rubrics in Moodle

  • Moodle Docs: Rubrics
  • Moodle Docs: Grading Guide (use for single-point rubrics)

Tools with rubrics (other than Moodle)

  • Google Assignments
  • Turnitin Assignments: Rubric or Grading Form

Other resources

  • DePaul University (n.d.). Rubrics .
  • Gonzalez, J. (2014). Know your terms: Holistic, Analytic, and Single-Point Rubrics . Cult of Pedagogy.
  • Goodrich, H. (1996). Understanding rubrics . Teaching for Authentic Student Performance, 54 (4), 14-17. Retrieved from   
  • Miller, A. (2012). Tame the beast: tips for designing and using rubrics.
  • Ragupathi, K., Lee, A. (2020). Beyond Fairness and Consistency in Grading: The Role of Rubrics in Higher Education. In: Sanger, C., Gleason, N. (eds) Diversity and Inclusion in Global Higher Education. Palgrave Macmillan, Singapore.

My Life! Teaching in a Korean University

in Life in the Classroom

ESL Writing Grading Rubric | Evaluating English Writing (For English Learners)

Do you teach writing to intermediate or advanced level English students, particularly in a university where you have to give grades? You’ve come to the right place! Keep on reading for all the details about my ESL rubric for writing that you can use to evaluate your students fairly. This ESL writing rubric is super easy to use and I think you’ll love it as much I do!

ESL-writing-grading-rubric

ESL writing rubric

Let’s get into all the details about this writing rubric for ESL students. Keep on reading!

ESL Writing Grading Rubric Introduction

For English teachers, grading writing and speaking are not easy when compared to something like grammar or vocabulary because there’s no right or wrong answer. Everything falls on a continuum from terrible to excellent.

The challenge for teachers is to grade in a way that’s fair and that also appears this way to the students. To do this, you’ll need an ESL writing rubric. I generally use the same one with all my writing classes, and for a wide variety of topics. Of course, feel free to adapt it to suit your needs for each individual situation.

esl-writing-rubric

Writing rubric for ESL students

ESL Writing Grading Rubric

My solution to ensure fair grading and also appear this way to my students is to use this ESL Writing Grading Rubric.  I use it, even though within reading the first couple sentences of a 5-paragraph, 500-word essay, I usually know what grade the student will get. This is especially true if it’s going to be an A, or D/F.

The middle ones require a bit of a closer look to separate the B students from the C ones. This ESL essay rubric started off more complicated, but over the years I’ve simplified it. The best ESL rubrics for writing are simple enough that students can clearly understand why they received the grade that they did.

The Categories I Evaluate for English Writing

Each of these five categories has an equal number of possible points (from 0-5), for a total of 25 points. To make your grading life easier, if the essay is worth 30%, you can adjust the rubric to make each category out of 6 points. Or, for 20%, make it out of four points.

If you make each section out of 4 points, you can use something like the following for an ESL writing rubric:

1- Inadequate

2- Needs improvement

3- Meets expectations

4- Exceeds expectations

Here are the various categories that I look at when evaluating ESL writing:

  • Sentences/paragraphs/format
  • Grammar/spelling/punctuation/vocabulary
  • Hook/thesis statement/topic sentences
  • Task completion/effect on the reader

Here are more details about each of those things.

#1: Sentence Structure, paragraphs, format

The best essays have sentences and paragraphs that are complete and easy to read. A nice variety of conjunctions and transitions are used to join them together.

#2: Grammar, spelling, punctuation, vocabulary

The best pieces of writing may only have 1-2 small errors in this area. There is a good use of higher level grammar and vocabulary.

This section includes pretty much all the formatting, language (vocabulary and grammar) that a student at their level would be expected to know, even if we explicitly haven’t discussed it in class.

#3: Hook, thesis statement, topic sentence

To get full marks, all of these will have to be very well done, and actually almost perfect. This is because I teach this extensively in class, and ensure that it students take away one thing from my writing course, it’s how to do this.

The ideas in the writing are clear, logical and well organized. Good supporting facts and information are used. If it’s a take-home assignment, I expect it to include real stats and facts. If the essay is for a test written in class, then good logic will have to be used.

#5: Task completion, the effect on the reader

The best essays are easy to understand on a first, quick read-through. The student followed the directions for the assignment (word count, etc.)

Learn More about Assessing Writing for ESL/EFL Students

Can I Evaluate ESL/EFL Writing Without a Rubric?

You may be tempted to evaluate student writing without a rubric because you think it will save time. I don’t recommend this unless you’re teaching informally and the grades/feedback you give students doesn’t actually count towards an official grade of some kind.

But, if you’re teaching at a university in a credit class, grades actually matter. Beyond that simple fact, here’s why you should use a rubric to evaluate written work:

  • It actually saves time because students will understand exactly how they got the grade that they did. You won’t have a steady stream of students in your office disputing their grade!
  • The feedback is more meaningful and students have a better idea about how to improve their writing.
  • You have some proof of why a student got the grade they did, if someone were to ask. It happened to me a number of times when teaching in Korea.

ESL Writing Activities, Games & Teaching Tips: Practical Ideas for the Classroom (ESL Activities for...

  • Bolen, Jackie (Author)
  • English (Publication Language)
  • 72 Pages - 12/09/2019 (Publication Date) - Independently published (Publisher)

Can I use This ESL Writing Grading Rubric for a Paragraph? 

I used to teach academic writing to advanced students where they had to write five paragraph essays. However, the above rubric would work equally well for a paragraph. Just about the only thing that would change is that you wouldn’t have a topic sentence because you’d only need a thesis statement for the entire piece.

Quick Tip to Cut Down on Grading Time for ESL Writing

The other tip besides using ESL rubrics for writing is for when grading ESL writing. This tip will reduce the number of complaints by a lot!

Tell the students that of course they are free to ask you to re-grade their essay and you will do your best to look at their essay with fresh eyes but you’ll spend 3-4 times the amount of time you spent the first time around and get out your red pen and circle every single mistake they’ve made and not overlook anything. I always mention that I’m very kind and generally overlook almost all the small-medium things.

I mention that it’s possible to get a higher grade but it’s also very possible to get a lower one. Finally, give the example of them getting 20/25, but they think they maybe should have gotten 21 or 22. In this case, it’s a terrible idea to ask me to re-grade it. However maybe I made a serious mistake and gave them 15/25, but their essay was actually quite good and they should have gotten 24/25. Then it would make sense to ask me to check again.

39 No-Prep/Low-Prep ESL Speaking Activities: For English Teachers of Teenagers and Adults Who Want...

Mean-Yes. Effective-also Yes

Is this mean? Perhaps. But, it’s also quite effective and this past semester in my 3 writing classes, a grand total of zero students asked me to take a second look. I strive for total professionalism at all times, but I’m also not willing to let my job consume every last second of free time that I have.

Teaching ESL writing, the easy way. It’s possible! Use this ESL rubric for writing and you’ll be well on your way!

How Much Feedback Do You Give when using this Rubric?

It’s an excellent question! When I teach English writing, I’m ALL about teaching self-editing instead of having students rely on me to correct all their errors. During the semester, students are free to come to my office during my allotted hours for me to have a quick read-through of their writing. I’ll usually give feedback along the lines of:

“Your thesis statement is kind of weak. Have a look at that and see if you can make it more concise.”

“I noticed that you have very few transitions in your essay. It makes it kind of hard to read.”

“You have many grammar errors. For example, subject-verb agreement.”

“Can you try to use some more complicated grammar or vocabulary? It’s fine, but all the sentences are so similar.”

What about Feedback on Assignments and Tests?

Along with this rubric, I’ll write some comments on my students’ work, usually 3-4 sentences next to their grade. Throughout their essay, I’ll pick out around 5 things to circle as problems or errors.

I’ll put a checkmark as a sign of a good thing like the thesis statement or topic sentences.

Is it necessary to correct every single error? Not really. It’s often more helpful to just point out mistakes that students have made more than once.

Which ESL Writing Textbook Do you Recommend?

Looking for an excellent textbook for teaching academic writing to ESL or EFL students? Stop looking right now and go buy this: Great Writing 4: From Great Paragraphs to Great Essays .

It’s an ideal introduction to writing an essay for high intermediate to advanced level English students. In my case, I used it when teaching 3rd or 4th year English majors at a university in South Korea.

writing-grading-rubric-esl

Writing grading rubric for ESL

How Do I Prevent Cheating in a Writing Class?

If you teach ESL writing in a for-credit class where you have to assign grades, you will almost certainly have students who try to cheat. There are a few things that I do to combat this and make things fair for the students.

Homework Assignments: Not Worth that Many Marks

During my course, I did have homework assignments. They just weren’t worth that many points, usually a maximum of 20% of the final grade.

The bulk of the grade was things we did in class: journalling and then the midterm and final exam that had to be physically written in class. This gave a better indication of who could actually write, without having the crutch of the Internet to assist them.

The Ultimate Thing to Do on the First Day of Class

On the first day of any writing class, I get students to complete a “Get to know you assignment.” I give them about 20 minutes to write 3 short paragraphs.

  • The past (high school days, growing up, etc.)
  • The present (university life and their thoughts about it)
  • The future (dreams, hopes, etc.)

This shows you how in a very clear way how proficient students are at the past/present/future verb tenses and it’s often quite obvious who will do well in your class and who will struggle.

Then, keep these papers and in case of a questionable homework assignment, you have something to compare to. For example, I had one student submit something that I myself probably couldn’t have produced. It was actually that good and had advanced level vocabulary that I had to look up.

I pulled out her assignment from the first day and found it riddled with simple mistakes like not using the correct past tense verb form and other comparable mistakes. She clearly could not have done that homework assignment herself and my guess is that she paid someone to do it because I was unable to find it through a Google search.

Midterm/Final Exam: Assign Random Topics

Some teachers assign a single topic for the exams and then allow the students to prepare their essay beforehand. During the exam, they just have to write it out basically. I try to avoid this.

Instead, I give students a list of around 20 possible topics. Then, I give students a slip of paper with two possible choices that they must choose from. Each student gets a random combination.

This allows students to prepare in terms of ideas and main points, but they can’t memorize an essay word for word and I find it a better test of actual writing ability.

Consider Using the Grammarly Plagiarism Checker

Don't get caught plagiarizing

If you’re serious about catching students who plagiarize, then you’ll want to invest in Grammarly. It’s just so much better than searching around on Google and can save you a ton of time when grading assignments. I get students to submit them electronically and then run them through this tool as the first step.

Head over there now to learn more about it:

Best Plagiarism Checker & Proofreader  

ESL Writing Rubric FAQs

There are a number of common questions that people have about ESL rubrics for writing. Here are the answers to some of the most popular ones.

What is ESL writing?

ESL writing is writing for people who don’t speak English as their first language (ESL = English as a Second language). These learners are also known as non-native speakers.

How do I teach ESL writing?

Teaching ESL writing starts with forming good sentences. Then, move from sentences to writing good paragraphs with a thesis statement or informal diaries. Finally, students can begin to write academic essays , newspaper articles, blog posts, and more.

How do beginners learn to write ESL?

ESL beginners can learn to write by following a few simple steps to construct good sentences.

  • Each sentence starts with a capital (big) letter.
  • Finish each sentence with a period, question mark or exclamation point.
  • Use a capital letter for proper names and “I.”
  • At a minimum, each sentence must contain a subject and verb. Usually, they also contain a complement like a prepositional phrase or direct object.

How do I write a paragraph for ESL?

Writing a paragraph for ESL involves a few simple steps:

  • Start with a good thesis statement that explains what your paragraph is about.
  • Give some facts, reasons or examples to support your opinion or idea.
  • Finish with a concluding sentence.

Have your say about this ESL Writing Grading Rubric

What do you use in your ESL writing classes to evaluate your students in a fair way? Do you like this ESL writing activity rubric, or do you prefer another one ? Leave a comment below and share your wisdom with us! We’d love to hear from you.

Also be sure to give this article a share on Facebook, Twitter, or Pinterest. It’ll help other busy teachers, like yourself find this useful resource.

ell-writing rubric

ELL writing rubric

Last update on 2022-06-17 / Affiliate links / Images from Amazon Product Advertising API

essay writing rubric esl

About Jackie

Jackie Bolen has been teaching English for more than 15 years to students in South Korea and Canada. She's taught all ages, levels and kinds of TEFL classes. She holds an MA degree, along with the Celta and Delta English teaching certifications.

Jackie is the author of more than 30 books for English teachers and English learners, including Advanced English Conversation Dialogues and 39 No-Prep/Low-Prep ESL Speaking Activities for Teenagers and Adults . She loves to share her ESL games, activities, teaching tips, and more with other teachers throughout the world.

You can find her on social media at: Facebook Pinterest

About and Contact

Jackie Bolen has been talking ESL South Korea since 2014. The goal is to bring you the best tips, ideas, and news for everything teaching English in Korea, including university jobs.

Contact Jackie Bolen and My Life! Teaching in a Korean University here.

Email: [email protected]

Address: 2436 Kelly Ave, Port Coquitlam, Canada

Find the Privacy Policy here .

Best Seller

101 ESL Activities: Games, Activities, Practical ideas, & Teaching Tips For English Teachers of...

As an Amazon Associate, I earn from qualifying purchases.

Advertise on My Life! Teaching in Korean University

Recent Articles

esl-body-parts

Parts of the Body ESL Activities, Games, Worksheets & Lesson Plans

essay writing rubric esl

Val Hamer: Life After EFL and her Return to the UK

wh-questions-esl

WH Questions ESL Activities and Games | ESL WH Question Ideas

english conversation partner

How to Find a Language Partner | English Conversation Partner

essay writing rubric esl

USC shield

Center for Excellence in Teaching

Home > Resources > Academic essay rubric

Academic essay rubric

This is a grading rubric an instructor uses to assess students’ work on this type of assignment. It is a sample rubric that needs to be edited to reflect the specifics of a particular assignment. 

Download this file

Download this file [63.33 KB]

Back to Resources Page

Essay Papers Writing Online

Effective essay writing rubrics to enhance students’ skills and performance.

Essay writing rubrics

Essay writing is a fundamental skill that students must master to succeed academically. To help students improve their writing, educators often use rubrics as a tool for assessment and feedback. An essay writing rubric is a set of criteria that outlines what is expected in a well-crafted essay, providing students with clear guidelines for success.

Effective essay writing rubrics not only evaluate the quality of the content but also assess the organization, coherence, language use, and overall structure of an essay. By using a rubric, students can better understand what aspects of their writing need improvement and how to achieve higher grades.

This article will explore the importance of essay writing rubrics in academic success and provide tips on how students can use them to enhance their writing skills. By understanding and mastering the criteria outlined in a rubric, students can not only improve their writing but also excel in their academic endeavors.

Learn the Key Elements

When it comes to effective essay writing, understanding the key elements is essential for academic success. Here are some important elements to keep in mind:

  • Thesis Statement: A clear and concise thesis statement that presents the main idea of your essay.
  • Structure: A well-organized structure with introduction, body paragraphs, and a conclusion.
  • Evidence: Use relevant evidence and examples to support your arguments.
  • Analysis: Analyze the evidence provided and draw conclusions based on your analysis.
  • Clarity and Coherence: Ensure that your essay is clear, coherent, and easy to follow.

By mastering these key elements, you can improve your essay writing skills and achieve academic success.

Elements of Effective Essay Writing

When it comes to writing a successful essay, there are several key elements that you should keep in mind to ensure your work is of high quality. Here are some essential aspects to consider:

Understand the Importance

Understanding the importance of essay writing rubrics is crucial for academic success. Rubrics provide clear criteria for evaluating and assessing essays, which helps both students and instructors. They help students understand what is expected of them, leading to more focused and organized writing. Rubrics also provide consistency in grading, ensuring that all students are evaluated fairly and objectively. By using rubrics, students can see where they excel and where they need to improve, making the feedback process more constructive and impactful.

of Clear and Concise Rubrics

of Clear and Concise Rubrics

One key aspect of effective essay writing rubrics is the clarity and conciseness of the criteria they outline. Clear and concise rubrics provide students with a roadmap for success and make grading more efficient for instructors. When creating rubrics, it is essential to clearly define the expectations for each aspect of the essay, including content, organization, style, and grammar.

By using language that is straightforward and easy to understand, students can easily grasp what is expected of them and how they will be evaluated. Additionally, concise rubrics avoid confusion and ambiguity, which can lead to frustration and misunderstanding among students.

Furthermore, clear and concise rubrics help instructors provide more targeted feedback to students, allowing them to focus on specific areas for improvement. This ultimately leads to better learning outcomes and helps students develop their writing skills more effectively.

Implement Structured Guidelines

Structured guidelines are essential for effective essay writing. These guidelines provide a framework for organizing your thoughts, ideas, and arguments in a logical and coherent manner. By implementing structured guidelines, you can ensure that your essay is well-organized, cohesive, and easy to follow.

One important aspect of structured guidelines is creating an outline before you start writing. An outline helps you plan the structure of your essay, including the introduction, body paragraphs, and conclusion. It also helps you organize your ideas and ensure that your arguments flow logically from one point to the next.

Another important aspect of structured guidelines is using transitions to connect your ideas and arguments. Transitions help readers follow the flow of your essay and understand how one point relates to the next. They also help maintain the coherence and clarity of your writing.

Finally, structured guidelines include formatting requirements such as word count, font size, spacing, and citation style. Adhering to these formatting requirements ensures that your essay meets academic standards and is visually appealing to readers.

for Academic Excellence

When aiming for academic excellence in essay writing, it is essential to adhere to rigorous standards and criteria set by the educational institution. The use of effective essay writing rubrics can greatly enhance the quality of academic work by providing clear guidelines and criteria for assessment.

Academic excellence in essay writing involves demonstrating a deep understanding of the subject matter, presenting a well-structured argument, and showcasing critical thinking skills. Rubrics can help students focus on key elements such as thesis development, evidence analysis, organization, and clarity of expression.

By following a well-designed rubric, students can ensure that their essays meet the desired academic standards and expectations. Feedback provided based on the rubric can also help students identify areas for improvement and guide them towards achieving academic success.

Master the Art

Mastering the art of essay writing requires dedication, practice, and attention to detail. To become a proficient essay writer, one must develop strong writing skills, a clear and logical structure, as well as an effective argumentative style. Start by understanding the purpose of your essay and conducting thorough research to gather relevant information and evidence to support your claims.

Furthermore, pay close attention to grammar, punctuation, and spelling to ensure your work is polished and professional. Use essay writing rubrics as a guide to help you stay on track and meet the necessary criteria for academic success. Practice makes perfect, so don’t be afraid to revise and edit your work to improve your writing skills even further.

By mastering the art of essay writing through consistent practice and attention to detail, you can set yourself up for academic success and develop a valuable skill that will benefit you in various aspects of your life.

of Constructing Detailed Evaluations

Constructing detailed evaluations in essay writing rubrics is essential for providing specific feedback to students. When designing a rubric, it is important to consider the criteria that will be assessed and clearly define each level of performance. This involves breaking down the evaluation criteria into smaller, measurable components, which allows for more precise and accurate assessments.

Each criterion in the rubric should have a clear description of what is expected at each level of performance, from basic to proficient to advanced. This helps students understand the expectations and enables them to self-assess their work against the rubric. Additionally, including specific examples or descriptors for each level can further clarify the expectations and provide guidance to students.

Constructing detailed evaluations also involves ensuring that the rubric aligns with the learning objectives of the assignment. By mapping the evaluation criteria to the desired learning outcomes, instructors can ensure that the rubric accurately reflects what students are expected to demonstrate in their essays. This alignment helps students see the relevance of the assessment criteria and encourages them to strive for mastery of the material.

Related Post

How to master the art of writing expository essays and captivate your audience, convenient and reliable source to purchase college essays online, step-by-step guide to crafting a powerful literary analysis essay, tips and techniques for crafting compelling narrative essays.

  • Grades 6-12
  • School Leaders

Enter Today's Teacher Appreciation Giveaway!

15 Helpful Scoring Rubric Examples for All Grades and Subjects

In the end, they actually make grading easier.

Collage of scoring rubric examples including written response rubric and interactive notebook rubric

When it comes to student assessment and evaluation, there are a lot of methods to consider. In some cases, testing is the best way to assess a student’s knowledge, and the answers are either right or wrong. But often, assessing a student’s performance is much less clear-cut. In these situations, a scoring rubric is often the way to go, especially if you’re using standards-based grading . Here’s what you need to know about this useful tool, along with lots of rubric examples to get you started.

What is a scoring rubric?

In the United States, a rubric is a guide that lays out the performance expectations for an assignment. It helps students understand what’s required of them, and guides teachers through the evaluation process. (Note that in other countries, the term “rubric” may instead refer to the set of instructions at the beginning of an exam. To avoid confusion, some people use the term “scoring rubric” instead.)

A rubric generally has three parts:

  • Performance criteria: These are the various aspects on which the assignment will be evaluated. They should align with the desired learning outcomes for the assignment.
  • Rating scale: This could be a number system (often 1 to 4) or words like “exceeds expectations, meets expectations, below expectations,” etc.
  • Indicators: These describe the qualities needed to earn a specific rating for each of the performance criteria. The level of detail may vary depending on the assignment and the purpose of the rubric itself.

Rubrics take more time to develop up front, but they help ensure more consistent assessment, especially when the skills being assessed are more subjective. A well-developed rubric can actually save teachers a lot of time when it comes to grading. What’s more, sharing your scoring rubric with students in advance often helps improve performance . This way, students have a clear picture of what’s expected of them and what they need to do to achieve a specific grade or performance rating.

Learn more about why and how to use a rubric here.

Types of Rubric

There are three basic rubric categories, each with its own purpose.

Holistic Rubric

A holistic scoring rubric laying out the criteria for a rating of 1 to 4 when creating an infographic

Source: Cambrian College

This type of rubric combines all the scoring criteria in a single scale. They’re quick to create and use, but they have drawbacks. If a student’s work spans different levels, it can be difficult to decide which score to assign. They also make it harder to provide feedback on specific aspects.

Traditional letter grades are a type of holistic rubric. So are the popular “hamburger rubric” and “ cupcake rubric ” examples. Learn more about holistic rubrics here.

Analytic Rubric

Layout of an analytic scoring rubric, describing the different sections like criteria, rating, and indicators

Source: University of Nebraska

Analytic rubrics are much more complex and generally take a great deal more time up front to design. They include specific details of the expected learning outcomes, and descriptions of what criteria are required to meet various performance ratings in each. Each rating is assigned a point value, and the total number of points earned determines the overall grade for the assignment.

Though they’re more time-intensive to create, analytic rubrics actually save time while grading. Teachers can simply circle or highlight any relevant phrases in each rating, and add a comment or two if needed. They also help ensure consistency in grading, and make it much easier for students to understand what’s expected of them.

Learn more about analytic rubrics here.

Developmental Rubric

A developmental rubric for kindergarten skills, with illustrations to describe the indicators of criteria

Source: Deb’s Data Digest

A developmental rubric is a type of analytic rubric, but it’s used to assess progress along the way rather than determining a final score on an assignment. The details in these rubrics help students understand their achievements, as well as highlight the specific skills they still need to improve.

Developmental rubrics are essentially a subset of analytic rubrics. They leave off the point values, though, and focus instead on giving feedback using the criteria and indicators of performance.

Learn how to use developmental rubrics here.

Ready to create your own rubrics? Find general tips on designing rubrics here. Then, check out these examples across all grades and subjects to inspire you.

Elementary School Rubric Examples

These elementary school rubric examples come from real teachers who use them with their students. Adapt them to fit your needs and grade level.

Reading Fluency Rubric

A developmental rubric example for reading fluency

You can use this one as an analytic rubric by counting up points to earn a final score, or just to provide developmental feedback. There’s a second rubric page available specifically to assess prosody (reading with expression).

Learn more: Teacher Thrive

Reading Comprehension Rubric

Reading comprehension rubric, with criteria and indicators for different comprehension skills

The nice thing about this rubric is that you can use it at any grade level, for any text. If you like this style, you can get a reading fluency rubric here too.

Learn more: Pawprints Resource Center

Written Response Rubric

Two anchor charts, one showing

Rubrics aren’t just for huge projects. They can also help kids work on very specific skills, like this one for improving written responses on assessments.

Learn more: Dianna Radcliffe: Teaching Upper Elementary and More

Interactive Notebook Rubric

Interactive Notebook rubric example, with criteria and indicators for assessment

If you use interactive notebooks as a learning tool , this rubric can help kids stay on track and meet your expectations.

Learn more: Classroom Nook

Project Rubric

Rubric that can be used for assessing any elementary school project

Use this simple rubric as it is, or tweak it to include more specific indicators for the project you have in mind.

Learn more: Tales of a Title One Teacher

Behavior Rubric

Rubric for assessing student behavior in school and classroom

Developmental rubrics are perfect for assessing behavior and helping students identify opportunities for improvement. Send these home regularly to keep parents in the loop.

Learn more: Teachers.net Gazette

Middle School Rubric Examples

In middle school, use rubrics to offer detailed feedback on projects, presentations, and more. Be sure to share them with students in advance, and encourage them to use them as they work so they’ll know if they’re meeting expectations.

Argumentative Writing Rubric

An argumentative rubric example to use with middle school students

Argumentative writing is a part of language arts, social studies, science, and more. That makes this rubric especially useful.

Learn more: Dr. Caitlyn Tucker

Role-Play Rubric

A rubric example for assessing student role play in the classroom

Role-plays can be really useful when teaching social and critical thinking skills, but it’s hard to assess them. Try a rubric like this one to evaluate and provide useful feedback.

Learn more: A Question of Influence

Art Project Rubric

A rubric used to grade middle school art projects

Art is one of those subjects where grading can feel very subjective. Bring some objectivity to the process with a rubric like this.

Source: Art Ed Guru

Diorama Project Rubric

A rubric for grading middle school diorama projects

You can use diorama projects in almost any subject, and they’re a great chance to encourage creativity. Simplify the grading process and help kids know how to make their projects shine with this scoring rubric.

Learn more: Historyourstory.com

Oral Presentation Rubric

Rubric example for grading oral presentations given by middle school students

Rubrics are terrific for grading presentations, since you can include a variety of skills and other criteria. Consider letting students use a rubric like this to offer peer feedback too.

Learn more: Bright Hub Education

High School Rubric Examples

In high school, it’s important to include your grading rubrics when you give assignments like presentations, research projects, or essays. Kids who go on to college will definitely encounter rubrics, so helping them become familiar with them now will help in the future.

Presentation Rubric

Example of a rubric used to grade a high school project presentation

Analyze a student’s presentation both for content and communication skills with a rubric like this one. If needed, create a separate one for content knowledge with even more criteria and indicators.

Learn more: Michael A. Pena Jr.

Debate Rubric

A rubric for assessing a student's performance in a high school debate

Debate is a valuable learning tool that encourages critical thinking and oral communication skills. This rubric can help you assess those skills objectively.

Learn more: Education World

Project-Based Learning Rubric

A rubric for assessing high school project based learning assignments

Implementing project-based learning can be time-intensive, but the payoffs are worth it. Try this rubric to make student expectations clear and end-of-project assessment easier.

Learn more: Free Technology for Teachers

100-Point Essay Rubric

Rubric for scoring an essay with a final score out of 100 points

Need an easy way to convert a scoring rubric to a letter grade? This example for essay writing earns students a final score out of 100 points.

Learn more: Learn for Your Life

Drama Performance Rubric

A rubric teachers can use to evaluate a student's participation and performance in a theater production

If you’re unsure how to grade a student’s participation and performance in drama class, consider this example. It offers lots of objective criteria and indicators to evaluate.

Learn more: Chase March

How do you use rubrics in your classroom? Come share your thoughts and exchange ideas in the WeAreTeachers HELPLINE group on Facebook .

Plus, 25 of the best alternative assessment ideas ..

Scoring rubrics help establish expectations and ensure assessment consistency. Use these rubric examples to help you design your own.

You Might Also Like

Collage of Interactive Notebook Ideas

How To Get Started With Interactive Notebooks (Plus 25 Terrific Examples)

It's so much more than a place to take notes during class. Continue Reading

Copyright © 2024. All rights reserved. 5335 Gate Parkway, Jacksonville, FL 32256

PrepScholar

Choose Your Test

Sat / act prep online guides and tips, act writing rubric: full analysis and essay strategies.

ACT Writing

feature_ACTessayrubric

What time is it? It's essay time! In this article, I'm going to get into the details of the newly transformed ACT Writing by discussing the ACT essay rubric and how the essay is graded based on that. You'll learn what each item on the rubric means for your essay writing and what you need to do to meet those requirements.

feature image credit: A study in human nature, being an interpretation with character analysis chart of Hoffman's master painting "Christ in the temple"; (1920) by CircaSassy , used under CC BY 2.0 /Resized from original.

ACT Essay Grading: The Basics

If you've chosen to take the ACT Plus Writing , you'll have 40 minutes to write an essay (after completing the English, Math, Reading, and Science sections of the ACT, of course). Your essay will be evaluated by two graders , who score your essay from 1-6 on each of 4 domains, leading to scores out of 12 for each domain. Your Writing score is calculated by averaging your four domain scores, leading to a total ACT Writing score from 2-12.

NOTE : From September 2015 to June 2016, ACT Writing scores were calculated by adding together your domain scores and scaling to a score of 1-36; the change to an averaged 2-12 ACT Writing score was announced June 28, 2016 and put into action September 2016.

The Complete ACT Grading Rubric

Based on ACT, Inc's stated grading criteria, I've gathered all the relevant essay-grading criteria into a chart. The information itself is available on the ACT's website , and there's more general information about each of the domains here . The columns in this rubric are titled as per the ACT's own domain areas, with the addition of another category that I named ("Mastery Level").

ACT Writing Rubric: Item-by-Item Breakdown

Whew. That rubric might be a little overwhelming—there's so much information to process! Below, I've broken down the essay rubric by domain, with examples of what a 3- and a 6-scoring essay might look like.

Ideas and Analysis

The Ideas and Analysis domain is the rubric area most intimately linked with the basic ACT essay task itself. Here's what the ACT website has to say about this domain:

Scores in this domain reflect the ability to generate productive ideas and engage critically with multiple perspectives on the given issue. Competent writers understand the issue they are invited to address, the purpose for writing, and the audience. They generate ideas that are relevant to the situation.

Based on this description, I've extracted the three key things you need to do in your essay to score well in the Ideas and Analysis domain.

#1: Choose a perspective on this issue and state it clearly. #2: Compare at least one other perspective to the perspective you have chosen. #3: Demonstrate understanding of the ways the perspectives relate to one another. #4: Analyze the implications of each perspective you choose to discuss.

There's no cool acronym, sorry. I guess a case could be made for "ACCE," but I wanted to list the points in the order of importance, so "CEAC" it is.

Fortunately, the ACT Writing Test provides you with the three perspectives to analyze and choose from, which will save you some of the time of "generating productive ideas." In addition, "analyzing each perspective" does not mean that you need to argue from each of the points of view. Instead, you need to choose one perspective to argue as your own and explain how your point of view relates to at least one other perspective by evaluating how correct the perspectives you discuss are and analyzing the implications of each perspective.

Note: While it is technically allowable for you to come up with a fourth perspective as your own and to then discuss that point of view in relation to another perspective, we do not recommend it. 40 minutes is already a pretty short time to discuss and compare multiple points of view in a thorough and coherent manner—coming up with new, clearly-articulated perspectives takes time that could be better spend devising a thorough analysis of the relationship between multiple perspectives.

To get deeper into what things fall in the Ideas and Analysis domain, I'll use a sample ACT Writing prompt and the three perspectives provided:

Many of the goods and services we depend on daily are now supplied by intelligent, automated machines rather than human beings. Robots build cars and other goods on assembly lines, where once there were human workers. Many of our phone conversations are now conducted not with people but with sophisticated technologies. We can now buy goods at a variety of stores without the help of a human cashier. Automation is generally seen as a sign of progress, but what is lost when we replace humans with machines? Given the accelerating variety and prevalence of intelligent machines, it is worth examining the implications and meaning of their presence in our lives.

Perspective One : What we lose with the replacement of people by machines is some part of our own humanity. Even our mundane daily encounters no longer require from us basic courtesy, respect, and tolerance for other people.

Perspective Two : Machines are good at low-skill, repetitive jobs, and at high-speed, extremely precise jobs. In both cases they work better than humans. This efficiency leads to a more prosperous and progressive world for everyone.

Perspective Three : Intelligent machines challenge our long-standing ideas about what humans are or can be. This is good because it pushes both humans and machines toward new, unimagined possibilities.

First, in order to "clearly state your own perspective on the issue," you need to figure out what your point of view, or perspective, on this issue is going to be. For the sake of argument, let's say that you agree the most with the second perspective. A essay that scores a 3 in this domain might simply restate this perspective:

I agree that machines are good at low-skill, repetitive jobs, and at high-speed, extremely precise jobs. In both cases they work better than humans. This efficiency leads to a more prosperous and progressive world for everyone.

In contrast, an essay scoring a 6 in this domain would likely have a more complex point of view (with what the rubric calls "nuance and precision in thought and purpose"):

Machines will never be able to replace humans entirely, as creativity is not something that can be mechanized. Because machines can perform delicate and repetitive tasks with precision, however, they are able to take over for humans with regards to low-skill, repetitive jobs and high-skill, extremely precise jobs. This then frees up humans to do what we do best—think, create, and move the world forward.

Next, you must compare at least one other perspective to your perspective throughout your essay, including in your initial argument. Here's what a 3-scoring essay's argument would look like:

I agree that machines are good at low-skill, repetitive jobs, and at high-speed, extremely precise jobs. In both cases they work better than humans. This efficiency leads to a more prosperous and progressive world for everyone. Machines do not cause us to lose our humanity or challenge our long-standing ideas about what humans are or can be.

And here, in contrast, is what a 6-scoring essay's argument (that includes multiple perspectives) would look like:

Machines will never be able to replace humans entirely, as creativity is not something that can be mechanized, which means that our humanity is safe. Because machines can perform delicate and repetitive tasks with precision, however, they are able to take over for humans with regards to low-skill, repetitive jobs and high-skill, extremely precise jobs. Rather than forcing us to challenge our ideas about what humans are or could be, machines simply allow us to BE, without distractions. This then frees up humans to do what we do best—think, create, and move the world forward.

You also need to demonstrate a nuanced understanding of the way in which the two perspectives relate to each other. A 3-scoring essay in this domain would likely be absolute, stating that Perspective Two is completely correct, while the other two perspectives are absolutely incorrect. By contrast, a 6-scoring essay in this domain would provide a more insightful context within which to consider the issue:

In the future, machines might lead us to lose our humanity; alternatively, machines might lead us to unimaginable pinnacles of achievement. I would argue, however, projecting possible futures does not make them true, and that the evidence we have at present supports the perspective that machines are, above all else, efficient and effective completing repetitive and precise tasks.

Finally, to analyze the perspectives, you need to consider each aspect of each perspective. In the case of Perspective Two, this means you must discuss that machines are good at two types of jobs, that they're better than humans at both types of jobs, and that their efficiency creates a better world. The analysis in a 3-scoring essay is usually "simplistic or somewhat unclear." By contrast, the analysis of a 6-scoring essay "examines implications, complexities and tensions, and/or underlying values and assumptions."

  • Choose a perspective that you can support.
  • Compare at least one other perspective to the perspective you have chosen.
  • Demonstrate understanding of the ways the perspectives relate to one another.
  • Analyze the implications of each perspective you choose to discuss.

To score well on the ACT essay overall, however, it's not enough to just state your opinions about each part of the perspective; you need to actually back up your claims with evidence to develop your own point of view. This leads straight into the next domain: Development and Support.

Development and Support

Another important component of your essay is that you explain your thinking. While it's obviously important to clearly state what your ideas are in the first place, the ACT essay requires you to demonstrate evidence-based reasoning. As per the description on ACT.org [bolding mine]:

Scores in this domain reflect the ability to discuss ideas, offer rationale, and bolster an argument. Competent writers explain and explore their ideas, discuss implications, and illustrate through examples . They help the reader understand their thinking about the issue.

"Machines are good at low-skill, repetitive jobs, and at high-speed, extremely precise jobs. In both cases they work better than humans. This efficiency leads to a more prosperous and progressive world for everyone."

In your essay, you might start out by copying the perspective directly into your essay as your point of view, which is fine for the Ideas and Analysis domain. To score well in the Development and Support domain and develop your point of view with logical reasoning and detailed examples, however, you're going to have to come up with reasons for why you agree with this perspective and examples that support your thinking.

Here's an example from an essay that would score a 3 in this domain:

Machines are good at low-skill, repetitive jobs and at high-speed, extremely precise jobs. In both cases, they work better than humans. For example, machines are better at printing things quickly and clearly than people are. Prior to the invention of the printing press by Gutenberg people had to write everything by hand. The printing press made it faster and easier to get things printed because things didn't have to be written by hand all the time. In the world today we have even better machines like laser printers that print things quickly.

Essays scoring a 3 in this domain tend to have relatively simple development and tend to be overly general, with imprecise or repetitive reasoning or illustration. Contrast this with an example from an essay that would score a 6:

Machines are good at low-skill, repetitive jobs and at high-speed, extremely precise jobs. In both cases, they work better than humans. Take, for instance, the example of printing. As a composer, I need to be able to create many copies of my sheet music to give to my musicians. If I were to copy out each part by hand, it would take days, and would most likely contain inaccuracies. On the other hand, my printer (a machine) is able to print out multiple copies of parts with extreme precision. If it turns out I made an error when I was entering in the sheet music onto the computer (another machine), I can easily correct this error and print out more copies quickly.

The above example of the importance of machines to composers uses "an integrated line of skillful reasoning and illustration" to support my claim ("Machines are good at low-skill, repetitive jobs and at high-speed, extremely precise jobs. In both cases, they work better than humans"). To develop this example further (and incorporate the "This efficiency leads to a more prosperous and progressive world for everyone" facet of the perspective), I would need to expand my example to explain why it's so important that multiple copies of precisely replicated documents be available, and how this affects the world.

body_theworld-1

World Map - Abstract Acrylic by Nicolas Raymond , used under CC BY 2.0 /Resized from original.

Organization

Essay organization has always been integral to doing well on the ACT essay, so it makes sense that the ACT Writing rubric has an entire domain devoted to this. The organization of your essay refers not just to the order in which you present your ideas in the essay, but also to the order in which you present your ideas in each paragraph. Here's the formal description from the ACT website :

Scores in this domain reflect the ability to organize ideas with clarity and purpose. Organizational choices are integral to effective writing. Competent writers arrange their essay in a way that clearly shows the relationship between ideas, and they guide the reader through their discussion.

Making sure your essay is logically organized relates back to the "development" part of the previous domain. As the above description states, you can't just throw examples and information into your essay willy-nilly, without any regard for the order; part of constructing and developing a convincing argument is making sure it flows logically. A lot of this organization should happen while you are in the planning phase, before you even begin to write your essay.

Let's go back to the machine intelligence essay example again. I've decided to argue for Perspective Two, which is:

An essay that scores a 3 in this domain would show a "basic organizational structure," which is to say that each perspective analyzed would be discussed in its own paragraph, "with most ideas logically grouped." A possible organization for a 3-scoring essay:

An essay that scores a 6 in this domain, on the other hand, has a lot more to accomplish. The "controlling idea or purpose" behind the essay should be clearly expressed in every paragraph, and ideas should be ordered in a logical fashion so that there is a clear progression from the beginning to the end. Here's a possible organization for a 6-scoring essay:

In this example, the unifying idea is that machines are helpful (and it's mentioned in each paragraph) and the progression of ideas makes more sense. This is certainly not the only way to organize an essay on this particular topic, or even using this particular perspective. Your essay does, however, have to be organized, rather than consist of a bunch of ideas thrown together.

Here are my Top 5 ACT Writing Organization Rules to follow:

#1: Be sure to include an introduction (with your thesis stating your point of view), paragraphs in which you make your case, and a conclusion that sums up your argument

#2: When planning your essay, make sure to present your ideas in an order that makes sense (and follows a logical progression that will be easy for the grader to follow).

#3: Make sure that you unify your essay with one main idea . Do not switch arguments partway through your essay.

#4: Don't write everything in one huge paragraph. If you're worried you're going to run out of space to write and can't make your handwriting any smaller and still legible, you can try using a paragraph symbol, ¶, at the beginning of each paragraph as a last resort to show the organization of your essay.

#5: Use transitions between paragraphs (usually the last line of the previous paragraph and the first line of the paragraph) to "strengthen the relationships among ideas" ( source ). This means going above and beyond "First of all...Second...Lastly" at the beginning of each paragraph. Instead, use the transitions between paragraphs as an opportunity to describe how that paragraph relates to your main argument.

Language Use

The final domain on the ACT Writing rubric is Language Use and Conventions. This the item that includes grammar, punctuation, and general sentence structure issues. Here's what the ACT website has to say about Language Use:

Scores in this domain reflect the ability to use written language to convey arguments with clarity. Competent writers make use of the conventions of grammar, syntax, word usage, and mechanics. They are also aware of their audience and adjust the style and tone of their writing to communicate effectively.

I tend to think of this as the "be a good writer" category, since many of the standards covered in the above description are ones that good writers will automatically meet in their writing. On the other hand, this is probably the area non-native English speakers will struggle the most, as you must have a fairly solid grasp of English to score above a 2 on this domain. The good news is that by reading this article, you're already one step closer to improving your "Language Use" on ACT Writing.

There are three main parts of this domain:

#1: Grammar, Usage, and Mechanics #2: Sentence Structure #3: Vocabulary and Word Choice

I've listed them (and will cover them) from lowest to highest level. If you're struggling with multiple areas, I highly recommend starting out with the lowest-level issue, as the components tend to build on each other. For instance, if you're struggling with grammar and usage, you need to focus on fixing that before you start to think about precision of vocabulary/word choice.

Grammar, Usage, and Mechanics

At the most basic level, you need to be able to "effectively communicate your ideas in standard written English" ( ACT.org ). First and foremost, this means that your grammar and punctuation need to be correct. On ACT Writing, it's all right to make a few minor errors if the meaning is clear, even on essays that score a 6 in the Language Use domain; however, the more errors you make, the more your score will drop.

Here's an example from an essay that scored a 3 in Language Use:

Machines are good at doing there jobs quickly and precisely. Also because machines aren't human or self-aware they don't get bored so they can do the same thing over & over again without getting worse.

While the meaning of the sentences is clear, there are several errors: the first sentence uses "there" instead of "their," the second sentence is a run-on sentence, and the second sentence also uses the abbreviation "&" in place of "and." Now take a look at an example from a 6-scoring essay:

Machines excel at performing their jobs both quickly and precisely. In addition, since machines are not self-aware they are unable to get "bored." This means that they can perform the same task over and over without a decrease in quality.

This example solves the abbreviation and "there/their" issue. The second sentence is missing a comma (after "self-aware"), but the worse of the run-on sentence issue is absent.

Our Complete Guide to ACT Grammar might be helpful if you just need a general refresh on grammar rules. In addition, we have several articles that focus in on specific grammar rules, as they are tested on ACT English; while the specific ways in which ACT English tests you on these rules isn't something you'll need to know for the essay, the explanations of the grammar rules themselves are quite helpful.

Sentence Structure

Once you've gotten down basic grammar, usage, and mechanics, you can turn your attention to sentence structure. Here's an example of what a 3-scoring essay in Language Use (based on sentence structure alone) might look like:

Machines are more efficient than humans at many tasks. Machines are not causing us to lose our humanity. Instead, machines help us to be human by making things more efficient so that we can, for example, feed the needy with technological advances.

The sentence structures in the above example are not particularly varied (two sentences in a row start with "Machines are"), and the last sentence has a very complicated/convoluted structure, which makes it hard to understand. For comparison, here's a 6-scoring essay:

Machines are more efficient than humans at many tasks, but that does not mean that machines are causing us to lose our humanity. In fact, machines may even assist us in maintaining our humanity by providing more effective and efficient ways to feed the needy.

For whatever reason, I find that when I'm under time pressure, my sentences maintain variety in their structures but end up getting really awkward and strange. A real life example: once I described a method of counteracting dementia as "supporting persons of the elderly persuasion" during a hastily written psychology paper. I've found the best ways to counteract this are as follows:

#1: Look over what you've written and change any weird wordings that you notice.

#2: If you're just writing a practice essay, get a friend/teacher/relative who is good at writing (in English) to look over what you've written and point out issues (this is how my own awkward wording was caught before I handed in the paper). This point obviously does not apply when you're actually taking the ACT, but it very helpful to ask for someone else to take a look over any practice essays you write to point out issues you may not notice yourself.

Vocabulary and Word Choice

The icing on the "Language Use" domain cake is skilled use of vocabulary and correct word choice. Part of this means using more complicated vocabulary in your essay. Once more, look at this this example from a 3-scoring essay (spelling corrected):

Machines are good at doing their jobs quickly and precisely.

Compare that to this sentence from a 6-scoring essay:

Machines excel at performing their jobs both quickly and precisely.

The 6-scoring essay uses "excel" and "performing" in place of "are good at" and "doing." This is an example of using language that is both more skillful ("excel" is more advanced than "are good at") and more precise ("performing" is a more precise word than "doing"). It's important to make sure that, when you do use more advanced words, you use them correctly. Consider the below sentence:

"Machines are often instrumental in ramifying safety features."

The sentence uses a couple of advanced vocabulary words, but since "ramifying" is used incorrectly, the language use in this sentence is neither skillful nor precise. Above all, your word choice and vocabulary should make your ideas clearer, not make them harder to understand.

body_adjective-1

untitled is also an adjective by Procsilas Moscas , used under CC BY 2.0 /Resized and cropped from original.

How Do I Use the ACT Writing Grading Rubric?

Okay, we've taken a look at the ACTual ACT Writing grading rubric and gone over each domain in detail. To finish up, I'll go over a couple of ways the scoring rubric can be useful to you in your ACT essay prep.

Use the ACT Writing Rubric To...Shape Your Essays

Now that you know what the ACT is looking for in an essay, you can use that to guide what you write about in your essays...and how develop and organize what you say!

Because I'm an Old™ (not actually trademarked), and because I'm from the East Coast, I didn't really know much about the ACT prior to starting my job at PrepScholar. People didn't really take it in my high school, so when I looked at the grading rubric for the first time, I was shocked to see how different the ACT essay was (as compared to the more familiar SAT essay ).

Basically, by reading this article, you're already doing better than high school me.

body_portraitofthemusician

An artist's impression of L. Staffaroni, age 16 (look, junior year was/is hard for everyone).

Use the ACT Writing Rubric To...Grade Your Practice Essays

The ACT can't really give you an answer key to the essay the way it can give you an answer key to the other sections (Reading, Math, etc). There are some examples of essays at each score point on the ACT website , but these examples assume that students will be at an equal level in each of domains, which will not necessarily be true for you. Even if a sample essay is provided as part of a practice test answer key, it will probably use different context, have a different logical progression, or maybe even argue a different viewpoint.

The ACT Writing rubric is the next best thing to an essay answer key. Use it as a filter through which to view your essay . Naturally, you don't have the time to become an expert at applying the rubric criteria to your essay to make sure you're in line with the ACT's grading principles and standards. That is not your job. Your job is to write the best essay that you can. If you're not confident in your ability to spot grammar, usage, and mechanics issues, I highly recommend asking a friend, teacher, or family member who is really good at (English) writing to take a look over your practice essays and point out the mistakes.

If you really want custom feedback on your practice essays from experienced essay graders, may I also suggest the PrepScholar test prep platform ? As I manage all essay grading, I happen to know a bit about the essay part of this platform, which provides you with both an essay grade and custom feedback. Learn more about PrepScholar ACT Prep and our essay grading here!

What's Next?

Desirous of some more sweet sweet ACT essay articles? Why not start with our comprehensive guide to the ACT Writing test and how to write an ACT essay, step-by-step ? (Trick question: obviously you should do this.)

Round out your dive into the details of the ACT Writing test with tips and strategies to raise your essay score , information about the best ACT Writing template , and advice on how to get a perfect score on the ACT essay .

Want actual feedback on your essay? Then consider signing up for our PrepScholar test prep platform . Included in the platform are practice tests and practice essays graded by experts here at PrepScholar.

Want to improve your ACT score by 4 points?

Check out our best-in-class online ACT prep program . We guarantee your money back if you don't improve your ACT score by 4 points or more.

Our program is entirely online, and it customizes what you study to your strengths and weaknesses . If you liked this Writing lesson, you'll love our program. Along with more detailed lessons, you'll get thousands of practice problems organized by individual skills so you learn most effectively. We're special in having expert instructors grade your essays and give you custom feedback on how to improve . We'll also give you a step-by-step program to follow so you'll never be confused about what to study next.

Check out our 5-day free trial:

Get 4 More Points on Your ACT, GUARANTEED

Laura graduated magna cum laude from Wellesley College with a BA in Music and Psychology, and earned a Master's degree in Composition from the Longy School of Music of Bard College. She scored 99 percentile scores on the SAT and GRE and loves advising students on how to excel in high school.

Ask a Question Below

Have any questions about this article or other topics? Ask below and we'll reply!

Improve With Our Famous Guides

  • For All Students

The 5 Strategies You Must Be Using to Improve 160+ SAT Points

How to Get a Perfect 1600, by a Perfect Scorer

Series: How to Get 800 on Each SAT Section:

Score 800 on SAT Math

Score 800 on SAT Reading

Score 800 on SAT Writing

Series: How to Get to 600 on Each SAT Section:

Score 600 on SAT Math

Score 600 on SAT Reading

Score 600 on SAT Writing

Free Complete Official SAT Practice Tests

What SAT Target Score Should You Be Aiming For?

15 Strategies to Improve Your SAT Essay

The 5 Strategies You Must Be Using to Improve 4+ ACT Points

How to Get a Perfect 36 ACT, by a Perfect Scorer

Series: How to Get 36 on Each ACT Section:

36 on ACT English

36 on ACT Math

36 on ACT Reading

36 on ACT Science

Series: How to Get to 24 on Each ACT Section:

24 on ACT English

24 on ACT Math

24 on ACT Reading

24 on ACT Science

What ACT target score should you be aiming for?

ACT Vocabulary You Must Know

ACT Writing: 15 Tips to Raise Your Essay Score

How to Get Into Harvard and the Ivy League

How to Get a Perfect 4.0 GPA

How to Write an Amazing College Essay

What Exactly Are Colleges Looking For?

Is the ACT easier than the SAT? A Comprehensive Guide

Should you retake your SAT or ACT?

When should you take the SAT or ACT?

Stay Informed

Follow us on Facebook (icon)

Get the latest articles and test prep tips!

Looking for Graduate School Test Prep?

Check out our top-rated graduate blogs here:

GRE Online Prep Blog

GMAT Online Prep Blog

TOEFL Online Prep Blog

Holly R. "I am absolutely overjoyed and cannot thank you enough for helping me!”

The Study Blog

Term Paper Writing Help

essay writing rubric esl

If you aren't sure whether you are good at expressing yourself through writing, then if you find it difficult to do so (e.g., when trying to write an english essay), we can help you overcome those obstacles by assisting you in improving your communication through writing. We help students compose essays or other types of papers for their courses. Now is the time to come visit us!

How to Overcome the Complexity of a Nursing Essay

There aren't many alternatives for professional translations. Before writing a good summary of something, you need to know your subject well enough to be able to write an accurate one. A research paper requires mastery of research language, a deep understanding of their subjects to be able to write about them clearly, and a careful consideration of possible problems before proposing solutions. Students often have trouble understanding medical terminology when they first encounter it, because they have never heard of these words before. When writing a cohesive psychology essay, students must be familiar with some psychological concepts. We have a wealth of experience under our belt, so we know where they need help. Although you may be able to find better deals elsewhere, there is no way to tell if these sites offer superior customer service and top-quality results. Read customer reviews before making any online purchases. If you don't think there's a market for them, it's perhaps best to skip them.

Professional Help from Copywriters

If you would like us to write anything from an essay in history to a term paper for you, we’d be happy to oblige. When writing something, there's a precise formula for choosing the best word. You can rest assured that you'll receive an expertly written paper from those who know exactly what they're doing. No need to write anything down today; there are no reasons why you shouldn't let others edit your document for you. Don't waste your time trying to convince them to do it for you, instead, invest it in something more productive! Order term papers online and go there! Founded in a simple belief that we are capable of delivering top-quality content to you, we offer a range of guarantees. Test it out yourself! The results must be presented after all the research has been completed.

Cheap Business Essay Writing Services

Before being accepted into our company, we underwent extensive background checks. Check their credentials to confirm that they have been writing professionally for some time. If they are members of professional associations, check, for instance.

essay writing rubric esl

Fun Tips to Spend Orthodox Easter Away from Home

In "Student Life"

Welcome to the New Bloggers

In "Degree Essentials"

Mastering Warwick as a Postgraduate

In "Looking After You"

Comments are closed.

Copyright, 2023

Improving Writing Feedback for Struggling Writers: Generative AI to the Rescue?

  • Original Paper
  • Open access
  • Published: 14 May 2024

Cite this article

You have full access to this open access article

essay writing rubric esl

  • Anya S. Evmenova 1 ,
  • Kelley Regan 1 ,
  • Reagan Mergen 1 &
  • Roba Hrisseh 1  

170 Accesses

1 Altmetric

Explore all metrics

Generative AI has the potential to support teachers with writing instruction and feedback. The purpose of this study was to explore and compare feedback and data-based instructional suggestions from teachers and those generated by different AI tools. Essays from students with and without disabilities who struggled with writing and needed a technology-based writing intervention were analyzed. The essays were imported into two versions of ChatGPT using four different prompts, whereby eight sets of responses were generated. Inductive thematic analysis was used to explore the data sets. Findings indicated: (a) differences in responses between ChatGPT versions and prompts, (b) AI feedback on student writing did not reflect provided student characteristics (e.g., grade level or needs; disability; ELL status), and (c) ChatGPT’s responses to the essays aligned with teachers’ identified areas of needs and instructional decisions to some degree. Suggestions for increasing educator engagement with AI to enhance teaching writing is discussed.

Similar content being viewed by others

essay writing rubric esl

Students’ voices on generative AI: perceptions, benefits, and challenges in higher education

essay writing rubric esl

Artificial Intelligence (AI) Student Assistants in the Classroom: Designing Chatbots to Support Student Success

Examining science education in chatgpt: an exploratory study of generative artificial intelligence.

Avoid common mistakes on your manuscript.

The advances in Generative Artificial Intelligence (generative AI) have transformed the field of education introducing new ways to teach and learn. Its integration is fast growing in all areas of education, including special education (Marino et al., 2023 ). Generative AI has the potential to increase the inclusion of students with disabilities in general education by providing additional assistive supports (Garg and Sharma, 2020 ; Zdravkova, 2022 ). Specifically, large language models like the one used by a popular AI tool, ChatGPT (Chat Generative Pre-trained Transformer) can generate human-like responses to prompts, similar to a conversation. It can facilitate learning for students with and without high-incidence disabilities (e.g., learning disabilities, ADHD) who struggle with writing (Barbetta, 2023 ). While experts continue to investigate the future of writing in the ChatGPT era, it is evident that it will significantly alter writing instruction (Wilson, 2023 ). ChatGPT can support students in choosing a topic, brainstorming, outlining, drafting, soliciting feedback, revising, and proofreading (Trust et al., 2023 ). This tool may also be a helpful resource for teachers in providing feedback on students’ writing. Timely and quality feedback by ChatGPT can encourage the use of higher-level thinking skills while improving the writing process including the planning, writing, and reviewing phases of that process (Golinkoff & Wilson, 2023 ).

Writing Instruction and Feedback for Struggling Writers

The writing process may be challenging for some students for many reasons. For example, planning is the first step of writing, but many students don’t systematically brainstorm. Instead, they move directly into drafting their sentences which may, in turn, be disjointed and not effectively communicated (Evmenova & Regan, 2019 ). Students, particularly those with high-incidence disabilities may not produce text or compose limited text, struggling with content generation, vocabulary, and the organization of ideas (Chung et al., 2020 ). While multilinguism is an asset, we have observed similar challenges with writing among English Language Learners in our research (Hutchison et al., 2024 ). The cognitive demands needed for drafting a response leave many students at no capacity to then edit or revise their work (Graham et al., 2017 ). Therefore, teachers should provide scaffolds to break down the complex process of writing so that it is sequential and manageable, progressing from simple to more complex concepts and skills.

Instruction for struggling writers is typically characterized as systematic and explicit (Archer & Hughes, 2011 ; Hughes et al., 2018 ). In order to provide explicit instruction, teachers should be guided by ongoing student data. Specifically, special and general education teachers of writing should collaboratively, systematically, and continuously monitor and responsively adjust instruction based on student progress (Graham et al., 2014 ). Formative assessments of writing inform feedback that a teacher provides a learner. McLeskey et al., ( 2017 ) describes:

Effective feedback must be strategically delivered, and goal directed; feedback is most effective when the learner has a goal, and the feedback informs the learner regarding areas needing improvement and ways to improve performance… Teachers should provide ongoing feedback until learners reach their established learning goals. (p. 25)

Various formative assessments are available to guide feedback in writing, with rubrics being one frequently used method, which we will explore in the following section.

Supporting Writing by Struggling Writers

School-aged students are required to show progress towards mastery of writing independently in order to be successful at school, future work, and in their personal lives (Graham, 2019 ). Thus, educators continuously look for tools to increase and support learner agency and independence including in writing (Edyburn, 2021 ). Over the past decade, the authors have developed a digital tool to support learner autonomy, access, and independence during essay composition as part of a federally funded, design-based research project referred to as WEGO: Writing Effectively with Graphic Organizers (Evmenova et al., 2018–2023 ). This tool is a technology-based graphic organizer (or TBGO) that embeds numerous evidence-based strategies and universally designed supports for students as well as an analytic rubric for teachers to evaluate student products and providing feedback. A detailed description of the tool can be found elsewhere (students’ features: Evmenova et al., 2020a ; teachers’ features: Regan et al., 2021 ).

The TBGO was developed to support upper elementary and middle school students with and without high-incidence disabilities to compose multiple genres of writing including persuasive (Evmenova et al., 2016 ), argumentative (Boykin et al., 2019 ), and/or personal narrative writing (Rana, 2018 ). The TBGO has also been effectively used by English Language Learners (Day et al., 2023 ; Boykin et al., 2019 ). In addition, it includes a dashboard that allows a teacher or caregiver to personalize instruction: assign prompts and support features embedded in the TBGO. After the student has an opportunity to write independently, the teacher can engage in what we refer to as data-driven decision making (or DDDM; Park & Datnow, 2017 ; Reeves and Chiang, 2018 ).

Teachers’ DDDM

A common formative assessment of writing used in classrooms is a rubric. In order to facilitate the DDDM process within the TBGO, various data are collected by the tool and provided to teachers including final writing product, total time spent actively using the tool, video views and duration, text-to-speech use and duration, audio comments use and duration, transition words use, total number of words, number of attempts to finish. A teacher first evaluates those data as well as student’s writing using a 5-point rubric embedded in the teacher dashboard of the TBGO (a specific rubric is available at  https://wego.gmu.edu ). Based on the rubric, a teacher identifies an area of need organized by phases of the writing process: Planning (select a prompt; select essay goal; select personal writing goal; brainstorm); Writing (identify your opinion, determine reasons, explain why or say more, add transition words, summarize, check your work); and Reviewing: Revise and Edit (word choice, grammar/spelling, punctuation, capitalization, evaluate). Then, a teacher provides specific instructional suggestions when the students’ score does not meet a threshold (e.g., content video models, modeling, specific practice activities). Once teachers select a targeted instructional move that is responsive to the identified area on the writing rubric, they record their instructional decision in the TBGO dashboard. The student’s work, the completed rubric, and the instructional decision is stored within the teacher dashboard. Recent investigations report that teachers positively perceive the ease and usability of the integrated digital rubric in the TBGO (see Regan et al., 2023a ; b ). Although promising, the teachers in those studies used DDDM with only a few students in their inclusive classes.

Efficient and Effective DDDM

The current version of the TBGO relies on teachers or caregivers to score student writing using an embedded rubric and to subsequently provide the student(s) with instructional feedback. In a classroom of twenty or more students, scoring individual essays and personalizing the next instructional move for each student is time consuming, and teachers may not regularly assess or interpret students’ writing abilities in the upper grades, especially (Graham et al., 2014 ; Kiuhara et al., 2009 ). Generative AI or chatbots are arguably leading candidates to consider when providing students with instructional feedback in a more time efficient manner (Office of Educational Technology, 2023 ). For example, automated essay scoring (AES) provides a holistic and analytic writing quality score of students’ writing and a description as to how the student can improve their writing. Recent research on classroom-based implementation of AES suggests its potential; but questions have been raised as to how teachers and students perceive the scores, and how it is used in classroom contexts (Li et al., 2015 ; Wilson et al., 2022 ). Other investigations remark on the efficiency and reliability among AES systems (Wilson & Andrada, 2016 ) and the consistency of scores with human raters (Shermis, 2014 ). More recently, a large-language model (specifically, GPT-3.5 version of ChatGPT) was prompted to rate secondary students’ argumentative essays and chatbot’s responses were compared to humans across five measures of feedback quality (see Steiss et al., 2023 ). Although GPT-3.5 included some inaccuracies in the feedback and the authors concluded that humans performed better than ChatGPT, the comparisons were remarkably close.

A greater understanding of what generative AI tools can do to support classroom teachers is needed. First, leveraging technology, with the use of automated systems, or logistical tools, can potentially improve working conditions for both general and special education teachers (Billingsley & Bettini, 2017 ; Johnson et al., 2012 ). Also, although educators see the benefits of AI and how it can be used to enhance educational services, there is urgent concern about the policies needed around its use and how it is ever evolving. For example, when writing this manuscript, GPT-4 evolved, but at a cost, this latter version may not be widely accessible for educators or students. With the fast adoption of AI, the Office of Educational Technology states that “it is imperative to address AI in education now to realize and mitigate emergent risks and tackle unintended consequences” (U.S. Department of Education, 2023 , p. 3). A first step in addressing AI in education is to understand what AI can do, and how its use supports or hinders student learning and teacher instruction. In this case, we focus on teachers’ writing instruction and feedback.

As we learn more about AI tools, it becomes obvious that AI literacy skills will need to be developed as part of digital skills by both teachers and students (Cohen, 2023 ). The importance of how we use chatbots, how we prompt them, and what parameters we use to direct the responses of chatbots becomes paramount.

Thus, the purpose of this study was to explore feedback and instructional suggestions generated by different AI tools when using prompts providing varying specificity (e.g., a generic 0–4 rating vs. analytic rubric provided) to help guide teachers of writing in their use of these tools. The purpose of including two versions of ChatGPT was not to criticize one and promote the other; but rather to understand and leverage their similarities and differences, given the same prompt. The research questions were:

RQ1: What is the difference between responses generated by GPT-3.5 and GPT-4 given prompts which provide varying specificity about students’ essays?

RQ2: What is the nature of the instructional suggestions provided by ChatGPT for students with and without disabilities and/or ELLs (aka struggling writers)?

RQ3: How does the formative feedback provided by GPT-3.5 and GPT-4 compare to the feedback provided by teachers when given the same rubric?

Data for this study were selected from a large intervention research study (led by the same authors) for a secondary data analysis. Specifically, while previous studies focused on the improvements in students’ writing outcomes (e.g., both quantity and quality of written essays) as well as explored how teachers provide feedback on students’ writing, the unique focus of this paper was on the use of AI to provide writing feedback (something we have not done before). The data included 34 persuasive student essays, a teacher’s completed analytic rubric evaluating the essay, and a teacher’s data-driven decisions with instructional feedback in the area of Writing and Reviewing (essays with the teachers’ DDDM in the area of Planning were excluded). We purposefully selected essays completed by students with various abilities and needs in different grade levels who struggled with writing and needed the TBGO intervention.

Participants

The 34 essays used in this study were written by 21 girls and 13 boys. Students ranged in age 8–13 and were in grades 3–7. The majority (59%) were White, 21% were Hispanic, 3% were African American, and 17% were other. Among the students, 41% were identified with high-incidence disabilities (learning disabilities, ADHD); 24% were English language learners (with a variety of primary languages); and 35% were struggling writers as reported by teachers. Teachers identified struggling writers as those who consistently demonstrated writing performance below grade level expectations (e.g., needing extra support with writing mechanics, cohesive and well-organized ideas).

Study Context

The data used in this study were collected in two separate settings: two inclusive classrooms in a suburban, private day school and an after-school program in a community center serving economically disadvantaged families. The same essay writing procedures were used in both settings. All students were first asked to write a persuasive opinion-based essay in response to one of two prompts validated by previous research (Regan et al., 2023b ). Examples of the prompts included:

Some students go to school on Saturday. Write an essay on whether or not students should go to school on Saturdays.

Some people believe kids your age should not have cell phones. Using specific details and examples to persuade someone of your opinion, argue whether or not kids your age should have cell phones.

After the pretest, students were introduced to the technology-based graphic organizer (TBGO) with embedded evidence-based strategies and supports. The instruction lasted 5–6 lessons. Then students were asked to use the TBGO to practice independent essay writing without any help from the teachers. As the TBGO is a Chrome-based web application and works on any device with a Chrome browser installed, each student used their own device/laptop and individual login credentials to access the TBGO. After completing the independent writing, teachers reviewed students’ products and completed the analytic rubric built into the TBGO’s teacher dashboard. They identified one primary area of need and determined an instructional decision that should take place in order to address the existing area of need. The instructional decisions included whole- and small-group activities (especially in those cases when multiple students demonstrated the same area of need); independent activities (including watching video models embedded within the TBGO); as well as individual teacher-student check-ins to discuss the area of need and future steps. A posttest with the TBGO and a delayed posttest without the TBGO were later administered. The essays used in the current study were from an independent writing phase since those included teachers’ DDDM. On average, essays had 133.44 ( SD  = 57.21; range 32–224) total words written. The vast majority included such important persuasive essay elements such as a topic sentence introducing the opinion, distinct reasons, examples to explain the reasons, summary sentence, and transition words. While this provides some important context, the quantity and quality of students’ writing products is not the focus of the current study and is reported elsewhere (Boykin et al., 2019 ; Day et al., 2023 ; Evmenova et al., 2016 , 2020b ; Regan et al., 2018 , 2023b ).

Data Sources

The existing 34 essays were imported into two different versions of the ChatGPT generative AI: GPT-3.5 version of ChatGPT (free version) and GPT-4 (subscription version). Four different prompts were used in both ChatGPT versions (see Table  1 ). As can be seen in Table  1 , the different prompts included (1) using a specific analytic rubric (when a rubric from the TBGO was uploaded to ChatGPT); (2) asking for a generic 0–4 rating (without any additional specifics regarding scoring); (3) no rubric (asking to identify the area of need without any rubric); (4) no information (asking to provide generic feedback without any information about the student in the prompt). Each prompt type constituted its own GPT chat. Thus, eight sets of responses (or eight different chats) were generated by ChatGPT. A prompt tailored to include the student’s essay as well as the specific student characteristics and the essay topic when applicable (according to the prompt samples presented in Table  1 ) was pasted into the chat. After GPT had a chance to react and provide feedback, the next prompt was pasted into the same chat. Thus, each chat included a total of 34 prompts and 34 GPT outputs. Each chat was then saved and analyzed.

Data Analysis and Credibility

Inductive thematic analysis was used to explore how generative AI can be used to provide writing feedback and guide writing instruction for struggling writers (Guest et al., 2011 ). First, each set of ChatGPT responses (or each GPT chat) was analyzed individually, and reoccurring codes across responses were grouped into categories. The four members of the research team were randomly assigned to analyze two GPT sets each. Each member generated a list of codes and categories within a chat that were the shared with the team and discussed. During those discussions, the patterns within categories were compared across different sets to develop overarching themes in response to RQ1 and RQ2. The trustworthiness of findings was established by data triangulation across 34 writing samples and eight sets of feedback. Also, peer debriefing was used throughout the data analysis (Brantlinger et al., 2005 ).

To answer RQ3, frequencies were used to compare teachers’ and ChatGPT scores on the analytic rubric and suggested instructional decisions. First, two researchers independently compared teachers’ and ChatGPT scores and suggestions. Since the same language from the rubric was used to identify the area of need, the comparisons were rated as 0 = no match; 1 = match. For instructional suggestions, the scale was 0 = no match; 1 = match in concept, but not in specifics; and 2 = perfect match. Over 50% of comparisons were completed by two independent researchers. Interrater reliability was established using point-by-point agreement formula dividing the number of agreements by the total number of agreements plus disagreements and yielding 100% agreement.

RQ1: Differences in AI Responses

In effort to answer RQ1 and explore the differences between responses generated by different ChatGPT versions when given prompts with varying specificity, we analyzed eight sets of responses. While the purpose was not to compare the sets in effort to find which one is better, several patterns have been observed that can guide teachers in using ChatGPT as the starting point for generating writing feedback to their struggling writers. The following are the six overarching themes that emerged from this analysis.

Predictable Pattern of Response

As can be seen in Table  2 , all sets generated excessive amounts of feedback (average length: M  = 383; SD  = 109.7; range 258–581 words) and followed a consistent, formulaic, and predictable pattern of responses across all the writing samples. While the layout and headers used to organize the responses differed across different ChatGPT versions and prompts, the layout and headers were consistent within each set. That said, it was also observed in all ChatGPT sets that the organization and headings found in a response changed slightly towards the end of the run for the 34 writing samples. It is unclear whether this pattern change may happen after a certain number of entries (or writing samples in our case) were entered into the ChatGPT run or if this shift in pattern occurs randomly. Similarly, we also observed that the later responses seemed to be more concise and lacked details which were observed earlier in the same set.

Specific Analytic Rubric

Both GPT-3.5 and GPT-4 provided responses organized into nine categories matching those included in the uploaded rubric. Each category included 1–2 sentences of feedback along with a numerical rating on a 0–4 scale. An overall holistic score was also calculated at the end along with a summary of the student’s overall strengths and weaknesses.

Generic 0–4 Rating

For each writing sample, GPT-3.5 consistently included an evaluation of student writing using four criteria-based categories: Content, Organization, Language Use (punctuation, spelling, and grammar), and Development of Ideas. Two to three bullet points of feedback were listed under each category along with a numeric rating on a 0–4 scale for each. The scale was not defined or explained. An overall holistic score was totaled at the end along with a summary of feedback presented in a bulleted list.

GPT-4’s response to the first writing sample included a definition of what each point on the scale meant (e.g., 4 = writing is clear, well-organized, well-developed, with effectively chosen details and examples presented logically, and few to no errors in conventions). In all consecutive responses, an introductory paragraph identified an overall bold-faced score (0–4) and an overview of what the student did and did not demonstrate in the writing. The following areas of writing were discussed across essays: Organization, Development, Main Idea, Reasons, Examples, Coherence, and Grammar.

Each response in GPT-3.5 began with “One area of need is…” followed by two sentences including how to address the need. Areas of need for instruction identified by ChatGPT included a high frequency of subject-verb agreement as parts of sentence structure (topic sentence and supporting details), followed by transition words or phrases, spelling and grammar conventions, spelling and word choice, capitalization, and punctuation. The second part of the response, titled Instructional Suggestion, provided an instructional strategy for a teacher to use, followed by a model of a ‘revised’ essay using ideas from the student’s response.

GPT-4 provided four consistent parts. First, the response opened with a statement about what the student wrote, a positive affirmation, and an instructional area of writing that could be improved upon. Next, under a header of Instructional Suggestion was a brief description as to what the teacher should do. The third part was a bold-faced, numbered list of steps for implementing that suggestion with bulleted cues underneath. The final part of the response was a ‘revised’ paragraph using the student’s initial writing and addressing the area of need.

GPT-3.5 provided feedback organized in 9 to 11 bolded categories. The sections that were identical for every writing sample included Proofreading; Revising and Editing; Encourage Creativity; and Positive Reinforcement. The sections that were consistent but individualized for each writing sample were Clarity and Organization (including a topic/introductory sentence); Supporting Details; Sentence Structure and Grammar (primarily focus on sentence fragments, punctuation, and capitalization); Conclusion; Vocabulary and Word Choice. Feedback on spelling and transition words/phrases was offered either as separate categories or subsumed under others.

GPT-4’s response could be organized in 3 overarching groups: Positive Reinforcement (including specific praise, affirmation, and creativity); Areas for Improvement (content feedback including idea development; details; coherence; clarity and focus; concluding sentence; and technical feedback including sentence structure; punctuation; grammar; word choice); as well as Instructional Suggestions. A sample revised paragraph was offered at the end with an explanation as to how it showcased the offered suggestions.

Using Specific Language from the Rubric

Both Specific Analytic Rubric sets (using GPT-3.5 and GPT-4) referred exclusively to the uploaded rubric and provided feedback using specific language from the rubric. This included feedback across the nine categories built into the rubric (e.g., the writer clearly identified an opinion, the writer has determined three reasons that support his/her opinion, etc.). Also, both ChatGPT versions used descriptors from the rubric (0 = Try again; 1 = Keep trying; 2 = Almost there; 3 = Good job; 4 = Got it). However, GPT-3.5 did not use any explicit examples from the student’s writing within the feedback and used broad and general statements. GPT-4 referred to the specific content from the students’ writing samples and was more tailored, or individualized (e.g., There are some grammatical and spelling errors present, e.g., "are" instead of "our").

Identifying General, Broad Areas of Need

Feedback in all GPT-3.5 sets (regardless of the prompt) was characterized as using common phrases representing broad areas of need. These phrases were not specifically targeted or explicit. For example, the Generic Rating GPT-3.5 set included such common phrases as “The essay presents ideas and supports them with reasonable detail, but there's room for more depth and elaboration.” or “The content is well-structured and effectively conveys the main points.” Similarly, the No Rubric GPT-3.5 set identified instructional areas of need that were only broadly relevant to the students’ writing. For example, in several instances, our review questioned the prioritization of the writing area identified and if ChatGPT was overgeneralizing areas in need of improvement. Specifically, does two instances of using lowercase when it should be uppercase mean that capitalization should be prioritized over other essential features of writing? Finally, the No Info GPT-3.5 set also used common phrases to describe areas for improvement regardless of the writing sample. For example, there were no difference in ChatGPT’s feedback for a writing essay with eight complete, robust, well-written sentences vs. an incomplete paragraph with just two sentences indicating the lack of targeted and specific feedback.

No Rubric GPT-4 set would start with identifying a broad area of need (e.g., coherence, grammar, development, organization/development of ideas, attention to detail) followed by a more individualized and specific instructional suggestion (as discussed below). The authors acknowledge that this might be explained by the prompt language to identify one area of need.

Focusing on an Individualized, Specific Areas of Need

Like the Specific Analytic Rubric GPT-4 set, the Generic 0–4 Rating GPT-4 set and the No Info GPT-4 sets were observed to include more guidance for the student, drawing on specific areas of an essay to provide corrective feedback. For example, Generic Rating GPT-4 feedback noted, “We should also try to provide more specific examples or explanations for each reason. For example, you mentioned that students get tired – maybe you can explain more about how having some recess can help them feel less tired.” In turn, No Info GPT-4 included detailed feedback focused on specific areas of need such as encouraging more details and clarifications, cohesion and flow, capitalization, spelling, homophones, and punctuation (including avoiding run-on sentences and properly using commas). Word choice, contractions, and conjunctions were often mentioned offering specific revisions. Varying the length and structure of sentences was sometimes suggested for making the writing more engaging and readable.

Misaligned Feedback

While there were some occasional discrepancies in GPT-4 sets, all GPT-3.5 sets appeared to generate feedback that was more misaligned with writing samples. For example, in the Specific Analytic Rubric GPT-3.5 set, a “Good Job” score of 3 was given for the Summary sentence that read, “Moreover, …” and was not a complete sentence. Also, the Generic Rating GPT-3.5 set did not mention any misuse of capitalization despite numerous cases of such misuse. Subject-verb agreement was erroneously mentioned as an area of need for some writing samples for the No Rubric GPT-3.5 set, and then, not mentioned for those students’ writing in which this feedback would be relevant. In the No Info GPT-3.5 set, the topic or introductory sentence was always noted as a suggested area of improvement and a revised sentence was always provided. This was true for cases when a student:

was missing an opinion that aligned with the prompt

had an opinion but did not start it with words “I believe …” (e.g., “Kids should get more recess time.”); and

already had a strong introductory sentence (e.g., “I believe that school starts too early and should begin later in the morning.”).

Starting with Specific Praise/Positive Affirmation

While most ChatGPT feedback included some general praise and affirmation, Generic Rating GPT-4, No Rubric GPT-4, and No Info GPT-4 sets always started with specific positive reinforcement. Unique elements in each essay were praised including conveying personal experiences, having a clear stance or position, and including a variety of reasons, etc.

RQ2: Instructional Suggestions

Instructional suggestions based on the evaluation of student writing was a focus of RQ2. Although we expected the responses from prompts that included specific student characteristics to differentiate the instructional suggestions in some way, this was not the case. In fact, none of the sets provided explicit instructional suggestions aligned with students’ characteristics (e.g., grade, disability, ELL). First, the suggestions for improving the writing of a 3rd grader’s essay were not distinct from those suggestions provided in response to a 7th grader’s writing (in Generic Rating GPT-3.5 and No Rubric GPT-3.5 sets). Also, there were no remarkable differences in the vocabulary used in the feedback for a 3rd grader vs. a 7th grader (in Generic Rating GPT-4 set). Only one set (Generic Rating GPT-4) offered a personalized message in a student-friendly format (without any additional prompting to do so).

Second, student characteristics were merely acknowledged in some sets. For example, Specific Analytic Rubric GPT-3.5 and GPT-4 only noted those characteristics in the summary section at the end of the feedback (e.g., “This is a well-written persuasive essay by your 7th-grade student with ADHD”). This was also observed in responses from the Generic Rating GPT-4 set, as well. For example, “This feedback emphasizes both the strengths of the student’s writing and the areas where improvement can be made, offering encouragement and guidance that is particularly important for a student with ADHD.” Finally, the No Rubric GPT-4 set also gave a mere nod to the additional context (e.g., Given student characteristics…). Although rare, connecting student characteristics with instruction was observed here: “Students with ADHD often struggle with organizing their thoughts in a coherent manner, and the flow of ideas in this student’s paragraph seems a bit disjointed….” Students’ characteristics were not mentioned in any other sets in which student information was included in the prompt (Generic Rating GPT-3.5 and No Rubric GPT-3.5).

Below is the description of how specific, broad, or no instructional suggestions were included in the ChatGPT sets (see Table  2 ).

Specific Suggestions

Specific instructional suggestions were mentioned in Generic Rating GPT-4, No Rubric GPT-4, and No Info GPT-4 sets. At the end of responses for the Generic Rating GPT-4 set, ChatGPT encouraged the teacher to use self-regulatory instructional strategies with students, such as goal setting or self-evaluation. For example, “By involving the student in the refinement of their work and setting goals, you empower them to take ownership of their learning and progression.”

No Rubric GPT-4 responses used such headings as modeling, guided practice, feedback, and independent practice with bulleted ideas under each. The specific suggestions included practice, mini-instructional lessons, engaging activities, peer review, explicit instruction, sentence-building activities, peer review sentence starters, technology such as word processing and online games, the five W’s and How strategy (i.e., a writing strategy that helps students remember to include the answers to “who,” “what,” “where,” “when,” “why,” and “how” in their writing to make their writing complete and clear), a mnemonic referred to as PEE (i.e., Point, Explain, Elaborate; this mnemonic helps students ensure their writing is focused, well-supported, and thoroughly developed), a personal dictionary, interactive editing, and a graphic organizer or outline. When the latter was suggested to support the “coherence” or “development of ideas,” ChatGPT’s response sometimes provided a backwards planning model of what the student’s ideas would look like in an outline format.

Responses of the No Info GPT-4 set included specific and varied instructional suggestions organized by categories: Writing Exercises; Focused Practice; and Revision Work. Suggestions included mini lessons on sentence structure, transition workshops, details workshops, personal experience illustrations, developing ideas workshops, worksheets, grammar lessons, spelling activities, sentence expansion or completion, and editing practice.

Broad Instructional Suggestions

Primarily broad instructional suggestions were offered in the Generic Rating GPT-3.5 and No Rubric GPT-3.5 sets. For example, Generic Rating GPT-3.5 responses had a section with a bulleted list of actionable, instructional suggestions. Each began with a verb (i.e., Work on…; Encourage the student to…; Practice…). It was also not clear if these suggestions were presented in any order of instructional priority. Also, the items included broad ideas that aligned with the student essays but may or may not have aligned with the lowest rated category of writing. Examples of largely vague and broad instructional suggestions recycled throughout the responses in the No Rubric GPT-3.5 set including: “use different types of sentences,” “teach basic spelling rules,” and “use appropriate punctuation.”

Revised Essay

The following three ChatGPT sets included responses with a revised student essay along with a brief explanation of how it was better (even though a revision was not requested in the prompt): No Rubric GPT-3.5, No Rubric GPT-4, and No Info GPT-4 sets. We considered that a model of writing, revised for improvement, was a broad instructional strategy. This is one of many excellent strategies for teaching writing, however, the revisions were often characterized by sophisticated vocabulary and complex elaborations. For example, a student wrote, “To illustrate, when students are hungry it’s hard for them to listen.” And ChatGPT elevated the sentence with, “To illustrate, when students are hungry, it's hard for them to listen because their minds may be preoccupied with thoughts of food.” Whereas the latter sentence is a well-crafted model for the student, this revision arguably loses the student’s voice and tone.

No Instructional Suggestions

No explicit instructional suggestions were included in the responses for Specific Analytic Rubric GPT-3.5, No Info GPT-3.5, and Specific Analytic Rubric GPT-4 sets. The reader was only reminded to provide feedback in a constructive and supportive manner and encourage students to ask questions and seek clarifications on any offered suggestions. While this is logical for both Specific Analytic rubric sets (not asking for instructional suggestions in the prompt), it is surprising for the No Info GPT-3.5 set (which asked for feedback and instructional suggestions).

RQ3: Comparisons Between Teachers and ChatGPT

In response to RQ3, we compared a real teachers’ data-based decision-making (DDDM), including the score and the instructional decision, to the scores generated in the Specific Analytic Rubric GPT-3.5 and Specific Analytic Rubric GPT-4 sets for students’ essays ( N  = 34). The first rubric category scored with a 2 or below was considered the area of need for writing instruction.

GPT-3.5 matched the teacher’s recommendation for the area of writing need 17.6% of the time. For example, the teacher identified Word Selection as the area of need (e.g., high use of repeated words and lacking sensory words) and GPT-3.5 noted the same area of need (e.g., there is some repetition and awkward phrasing). When comparing teacher versus ChatGPTs instructional decisions, there was no perfect match; however, 26.5% were coded as a partial match. For example, both the teacher and GPT-3.5 suggested an instructional activity of modeling how to write a summary sentence.

GPT-4 matched the teacher’s recommendation for the area of writing need 23.5% of the time. Similarly, when comparing the teacher versus ChatGPT’s instructional decisions, 47.1% were coded as a partial match for instruction.

Discussion and Practical Implications

Since the end of 2022 when it debuted, school leaders and teachers of writing have been grappling with what ChatGPT means for writing instruction. Its ability to generate essays from a simple request or to correct writing samples is making an impact on the classroom experience for students with and without disabilities and it is reshaping how teachers assess student writing (Marino et al., 2023 ; Trust et al., 2023 ; Wilson, 2023 ). However, teachers may have limited knowledge of how AI works and poor self-efficacy for using AI in the classroom to support their pedagogical decision making (Chiu et al., 2023 ). It is imperative to ensure that teachers receive professional development to facilitate the effective and efficient use of AI. There are more questions than answers currently, especially for its application by students struggling with academics.

The purpose of this investigation was to explore the application of ChatGPT chatbot for teachers of writing. Specifically, we used different versions of ChatGPT (GPT-3.5 – free and GPT-4 – subscription) and purposefully different types of prompts, providing limited or more information about the student characteristics and the topic of their writing. Essentially, we asked ChatGPT to evaluate an authentic student’s writing, identify the area(s) of need, and provide instructional suggestion(s) for addressing the problematic area(s) in that individual writing sample. We then compared AI-generated feedback to that completed by humans.

The findings indicate the possibilities and limitations of ChatGPT for evaluating student writing, interpreting a teacher-developed rubric, and providing instructional strategies.

Our finding is that, generally, ChatGPT can follow purposeful prompts, interpret and score using a criterion-based rubric when provided, create its own criteria for evaluating student writing, effectively revise student essay writing, celebrate what students do well in their writing, paraphrase student essay ideas, draft outlines of a student’s completed essay, and provide formative feedback in broad and specific areas along different stages of the writing process. Moreover, the response is immediate. These findings are consistent with previous investigations of ChatGPT and the assessment of student writing (Steiss et al., 2023 ). However, teachers need to consider the following points before relying on ChatGPT to provide feedback to their struggling writers.

In the ChatGPT sets which included no contextual information, the responses included more feedback.

All sets generated excessive amounts of feedback about student writing with no delineation of the next clear instructional move a teacher should attend to. So, ChatGPT may work as a great starting point, but teachers will need to go through the response to prioritize and design their instruction. Sifting through information for relevance can be time consuming and may even warrant a teacher verifying the content further.

Additionally, if students relied directly on ChatGPT, without any vetting from a teacher about the content, they too may be overwhelmed by the amount of feedback given to modify their writing or they may even be provided with erroneous feedback.

All GPT-3.5 sets identified broad areas of writing that needed improvement and frequently used common phrases such as grammar, organization/development of ideas, and attention to detail. In addition, this feedback was more often misaligned with students’ writing. This observation is worrisome since GPT-3.5 version of ChatGPT is free and highly accessible, making it likely the preferred AI tool for classroom educators.

Most GPT-4 sets (except one) generated more specific and individualized feedback about student writing. The specific feedback included in the generated outputs were much lengthier and would take much more time for a teacher to review than GPT-3.5 responses.

All sets identified multiple areas of need and when included in the responses, there were multiple instructional suggestions. Even the No Rubric sets, which explicitly prompted ChatGPT to focus on just one area of instructional need and one suggestion, included much more in the responses. This finding reiterates that we are still learning about AI literacy and the language we need to use to communicate effectively.

Both GPT-3.5 and GPT-4 allowed the upload of a researcher-developed analytic rubric and moreover, interpreted the performance criteria, rating scale, and indicators. ChatGPT also used the rubric’s specific language when providing its evaluation of the student writing.

No tailored feedback or specific suggestions were contextualized when prompts included varying ages, grade levels, or various student abilities and needs. Further research is needed to determine the types of AI literacy prompts or the contextual information that ChatGPT needs to address the particular needs of an individual child. Specially designed instruction, the heart of special education, should be tailored to a particular student (Sayeski et al., 2023 ).

Low agreement reported between the rubric scores and instructional suggestions made by teachers and those generated by ChatGPT does not necessarily mean that ChatGPT’s feedback is incorrect. One explanation for the difference may be that teachers provide targeted and individualized instruction using multiple forms of data and critical information to make instructional decisions. This includes their own professional judgement and knowledge about how each students’ backgrounds, culture, and language may influence student performance (McLeskey et al., 2017 ).

Limitations

This study is an initial exploration. There are several limitations that need to be taken into consideration. First and foremost, the four prompts were designed to present the chatbots with varying levels of details and student information to consider when providing feedback about a student’s writing sample. For example, Specific Analytic Rubric prompt asked the chatbot to assess students’ writing using an uploaded rubric, while No Rubric prompt asked to identify one area of need for the student’s writing and offer one instructional suggestion to address it. In addition to providing the chatbots with varying information, we also used varying language throughout the prompts when seeking feedback and suggestions (e.g., “Identify areas of need for this student’s writing”; “Identify one area of need … and offer one instructional suggestion”; “what feedback and instructional suggestions…”). Chatbots are clearly sensitive to the word choices made; thus, a consistency of the language in prompts should be considered for any future investigations that aim at prompt comparison. The purpose of this work was not to compare the four prompts in effort to find the best possible one. We also were not looking specifically for the feedback that could be shared with students as is (even though some versions generated such feedback without additional prompting). Instead, we were trying to explore how the output might differ depending on the prompts with differing level of detail. So, some of the reported difference are logical. We also did not prompt the ChatGPT any further, which would most likely result in refined feedback and/or suggestions. There is an infinite number of prompts that we could have used in this analysis. In fact, a new field of prompt engineering is emerging right in front of our eyes as we learn to design inputs for generative AI tools that would produce optimal outputs. Further investigations of various prompts to feed ChatGPT are needed. Our hope is that this paper will inspire teacher to spend some time exploring different tools and prompts in effort to find the most appropriate output depending on their context and their students’ needs.

Also, there was a limited numbers of essays from each specific group of learners (e.g., certain age/grade, specific disability categories and other characteristics). While we reported meaningful findings for this initial exploratory analysis, future research should include writing products from more homogeneous groups. Finally, teachers’ DDDM was accomplished by evaluating a completed graphic organizer, while ChatGPT feedback was provided based on the final student essay copied and pasted from the TBGO. Future research should consider new features of generative AI tools (e.g., Chat GPT’s new image analysis feature) where an image of a completed graphic organizer can be uploaded and analyzed.

This study offers examples for how to potentially incorporate AI effectively and efficiently into writing instruction. High quality special education teachers are reflective about their practice, use a variety of assistive and instructional technologies to promote student learning, and regularly monitor student progress with individualized assessment strategies. It seems very likely that teachers will adopt the capabilities of generative AI tools. With ongoing development and enhancements, AI technology is certain to become an integral component of classroom instruction. However, given the limitations of ChatGPT identified in this study, teacher-led instruction and decision making is still needed to personalize and individualize specialized instruction. Engaging with the technology more and building familiarity of what it can do to improve student learning and teacher practice is warranted.

Archer, A. L., & Hughes, C. A. (2011). Explicit instruction: Effective and efficient teaching . Guilford press.

Google Scholar  

Barbetta, P. M. (2023). Remedial and compensatory writing technologies for middle school students with learning disabilities and their classmates in inclusive classrooms. Preventing School Failure: Alternative Education for Children and Youth . https://doi.org/10.1080/1045988X.2023.2259837

Article   Google Scholar  

Boykin, A., Evmenova, A. S., Regan, K., & Mastropieri, M. (2019). The impact of a computer-based graphic organizer with embedded self-regulated learning strategies on the argumentative writing of students in inclusive cross-curricula settings. Computers & Education, 137 , 78–90. https://doi.org/10.1016/j.compedu.2019.03.008

Billingsley, B., & Bettini, E. (2017). Improving special education teacher quality and effectiveness. In J. M. Kauffman, D. P. Hallahan, & P. C. Pullen (Eds.), Handbook of special education (2nd ed., pp. 501-520). Boston: Taylor & Francis.

Brantlinger, E., Jimenez, R., Klinger, J., Pugach, M., & Richardson, V. (2005). Qualitative studies in special education. Exceptional Children, 71 (2), 195–207. https://doi.org/10.1177/001440290507100205

Garg, S., & Sharma, S. (2020). Impact of artificial intelligence in special need education to promote inclusive pedagogy. International Journal of Information and Education Technology, 10 (7), 523–527. https://doi.org/10.18178/ijiet.2020.10.7.1418

Chiu, T. K. F., Xia, Q., Zhou, X., Chai, C. S., & Cheng, M. (2023). Systematic literature review on opportunities, challenges, and future research recommendations. Computers and Education: Artificial Intelligence, 4 , 1–15. https://doi.org/10.1010/j.caeai.2022.100118

Chung, P. J., Patel, D. R., & Nizami, I. (2020). Disorder of written expression and dysgraphia: Definition, diagnosis, and management. Translational Pediatrics, 9 (1), 46–54. https://doi.org/10.21037/tp.2019.11.01

Cohen, Z. (2023). Moving beyond Google: Why ChatGPT is the search engine of the future [Blog Post]. Retrieved from https://thecorecollaborative.com/moving-beyond-google-why-chatgpt-is-the-search-engine-of-the-future/ . Accessed 1 Nov 2023

Day, J., Regan, K., Evmenova, A. S., Verbiest, C., Hutchison, A., & Gafurov, B. (2023). The resilience of students and teachers using a virtual writing intervention during COVID-19. Reading & Writing Quarterly, 39 (5), 390–412. https://doi.org/10.1080/10573569.2022.2124562

Edyburn, D. (2021). Universal usability and Universal Design for Learning. Intervention in School and Clinic, 56 (5), 310–315. https://doi.org/10.1177/1053451220963082

Evmenova, A. S., & Regan, K. (2019). Supporting the writing process with technology for students with disabilities. Intervention in School and Clinic, 55 (2), 78–87. https://doi.org/10.1177/1053451219837636

Evmenova, A. S., Regan, K., Boykin, A., Good, K., Hughes, M. D., MacVittie, N. P., Sacco, D., Ahn, S. Y., & Chirinos, D. S. (2016). Emphasizing planning for essay writing with a computer-based graphic organizer. Exceptional Children, 82 (2), 170–191. https://doi.org/10.1177/0014402915591697

Evmenova, A. S., Regan, K., & Hutchison, A. (2018-2023). W EGO RIITE: Writing ef iciently with graphic organizers – responsive instruction while implementing technology ef ectively (Project No. H327S180004) [Grant]. Technology and media services for individuals with disabilities: Stepping-up technology implementation grant, office of special education.

Evmenova, A. S., Regan, K., & Hutchison, A. (2020a). AT for writing: Technology-based graphic organizers with embedded supports.  TEACHING Exceptional Children , 52 (4), 266–269. https://doi.org/10.1177/0040059920907571

Evmenova, A. S., Regan, K., Ahn, S. Y., & Good, K. (2020b). Teacher implementation of a technology-based intervention for writing. Learning Disabilities: A Contemporary Journal, 18 (1), 27–47. https://www.ldw-ldcj.org/

Golinkoff, R. M., & Wilson, J. (2023). ChatGPT is a wake-up call to revamp how we teach writing. [Opinion]. Retrieved from https://www.inquirer.com/opinion/commentary/chatgpt-ban-ai-education-writing-critical-thinking-20230202.html . Accessed 1 Nov 2023

Graham, S. (2019). Changing how writing is taught. Review of Research in Education, 43 (1), 277–303. https://doi.org/10.3102/0091732X18821125

Graham, S., Capizzi, A., Harris, K. R., Hebert, M., & Morphy, P. (2014). Teaching writing to middle school students: A national survey. Reading and Writing, 27 , 1015–1042. https://doi.org/10.1007/s11145-013-9495-7

Graham, S., Collins, A. A., & Rigby-Wills, H. (2017). Writing characteristics of students with learning disabilities and typically achieving peers: A meta-analysis. Exceptional Children, 83 (2), 199–218. https://doi.org/10.1177/001440291666407

Guest, G., MacQueen, K. M., & Namey, E. E. (2011). Applied thematic analysis . SAGE Publications.

Hughes C. A., Riccomini P. J., & Morris J. R. (2018). Use explicit instruction. In  High leverage practices for inclusive classrooms  (pp. 215–236). Routledge. https://doi.org/10.4324/9781315176093 .

Hutchison, A., Evmenova, A. S., Regan, K., & Gafurov, B. (2024). Click, see, do: Using digital scaffolding to support persuasive writing instruction for emerging bilingual learners. Reading Teacher . https://doi.org/10.1002/trtr.2310

Johnson S. M., Kraft M. A., & Papay J. P. (2012). How context matters in high-need schools: The effects of teachers’ working conditions on their professional satisfaction and their students’ achievement. Teachers College Record , 114 , 1–39.

Kiuhara, S. A., Graham, S., & Hawken, L. S. (2009). Teaching writing to highschool students. Journal of Educational Psychology, 101 (1), 136–160. https://doi.org/10.1037/a0013097

Li, J., Link, S., & Hegelheimer, V. (2015). Rethinking the role of automated writing evaluation (AWE) feedback in ESL writing instruction. Journal of Second Language Writing, 27 , 1–18. https://doi.org/10.1016/j.jslw.2014.10.004

Marino, M. T., Vasquez, E., Dieker, L., Basham, J., & Blackorby, J. (2023). The future of artificial intelligence in special education technology. Journal of Special Education Technology, 38 (3), 404–416. https://doi.org/10.1177/01626434231165977

McLeskey, J., Barringer, M.-D., Billingsley, B., Brownell, M., Jackson, D., Kennedy, M., Lewis, T., Maheady, L., Rodriguez, J., Scheeler, M. C., Winn, J., & Ziegler, D. (2017). High-leverage practices in special education . Council for Exceptional Children & CEEDAR Center.

Office of Educational Technology (2023). Artificial intelligence and the future of teaching and learning: Insights and recommendations. Retrieved from https://tech.ed.gov/files/2023/05/ai-future-of-teaching-and-learning-report.pdf . Accessed 1 Nov 2023

Park, V., & Datnow, A. (2017). Ability grouping and differentiated instruction in an era of data-driven decision making. American Journal of Education, 123 (2), 281–306.

Rana, S. (2018). The impact of a computer-based graphic organizer with embedded technology features on the personal narrative writing of upper elementary students with high-incidence disabilities (Publication No. 13420322) [Doctoral dissertation, George Mason University]. ProQuest Dissertation Publishing.

Reeves, T. D., & Chiang, J.-L. (2018). Online interventions to promote teacher data-driven decision making: Optimizing design to maximize impact. Studies in Educational Evaluation, 59 , 256–269. https://doi.org/10.1016/j.stueduc.2018.09.006

Regan, K., Evmenova, A. S., Good, K., Leggit, A, Ahn, S., Gafurov, G., & Mastropieri, M. (2018). Persuasive writing with mobile-based graphic organizers in inclusive classrooms across the curriculum. Journal of Special Education Technology, 33 (1), 3–14. https://doi.org/10.1177/0162643417727292

Regan, K., Evmenova, A. S., Hutchison, A., Day, J., Stephens, M., Verbiest, C., & Gufarov, B. (2021). Steps for success: Making instructional decisions for students’ essay writing. TEACHING Exceptional Children , 54 (3), 202–212. https://doi.org/10.1177/00400599211001085

Regan, K., Evmenova, A. S., & Hutchison, A. (2023a). Specially designed assessment of writing to individualize instruction for students. In K. L. Write, & T. S. Hodges (Eds.), Assessing disciplinary writing in both research and practice (pp. 29–56). IGI Global. https://doi.org/10.4018/978-1-6684-8262-9

Regan, K., Evmenova, A. S., Mergen, R., Verbiest, C., Hutchison, A., Murnan, R., Field, S., & Gafurov, B. (2023b). Exploring the feasibility of virtual professional development to support teachers in making data-based decisions for improving student writing. Learning Disabilities Research & Practice , 38 (1), 40–56. https://doi.org/10.1111/ldrp.12301

Sayeski, K. L., Reno, E. A., & Thoele, J. M. (2023). Specially designed instruction: Operationalizing the delivery of special education services. Exceptionality, 31 (3), 198–210. https://doi.org/10.1080/09362835.2022.2158087

Shermis, M. D. (2014). State-of-the-art automated essay scoring: Competition, results, and future directions from a United States demonstration. Assessing Writing, 20 , 53–76. https://doi.org/10.1016/j.asw.2013.04.001

Steiss, J., Tate, T., Graham, S., Cruz, J., Hevert, M., Wang, J., Moon, Y., Tseng, W., & Warschauer, M. (2023). Comparing the quality of human and ChatGPT feedback on students’ writing. Retrieved from https://osf.io/preprints/edarxiv/ty3em/ . Accessed 1 Nov 2023

Trust, T., Whalen, J., & Mouza, C. (2023). Editorial: ChatGPT: Challenges, opportunities, and implications for teacher education. Contemporary Issues in Technology and Teacher Education, 23 (1), 1–23.

U.S. Department of Education. (2023). Office of Educational Technology, Artificial Intelligence and Future of Teaching and Learning: Insights and Recommendations , Washington, DC.

Wilson, J., & Andrada, G. N. (2016). Using automated feedback to improve writing quality: Opportunities and challenges. In Y. Rosen, S. Ferrara, & M. Mosharraf (Eds.), Handbook of research on technology tools for real-world skill development (pp. 678–703). IGI Global.

Wilson, J., Myers, M. C., & Potter, A. (2022). Investigating the promise of automated writing evaluation for supporting formative writing assessment at scale. Assessment in Education: Principles, Policy, & Practice, 29 (1), 1–17. https://doi.org/10.1080/0969594X.2022.2025762

Wilson, J. (2023). Writing without thinking? There’s a place for ChatGPT – if used properly [Guest Commentary]. Retrieved from https://www.baltimoresun.com/opinion/op-ed/bs-ed-op-0206-chatgpt-tool-20230203-mydxfitujjegndnjwwen4s4x7m-story.html . Accessed 1 Nov 2023

Zdravkova, K. (2022). The potential of artificial intelligence for assistive technology in education. In M. Ivanović, A. Klašnja-Milićević, L. C. Jain (Eds) Handbook on intelligent techniques in the educational process. Learning and analytics in intelligent systems (vol 29). Springer. https://doi.org/10.1007/978-3-031-04662-9_4 .

Download references

This research was supported by the U.S. Department of Education, Office of Special Education Programs [award number: H327S180004]. The views expressed herein do not necessarily represent the positions or policies of the Department of Education. No official endorsement by the U.S. Department of Education of any product, commodity, service, or enterprise mentioned in this publication is intended or should be inferred.

Author information

Authors and affiliations.

George Mason University, 4400 University Dr. MS 1F2, Fairfax, VA, 22030, USA

Anya S. Evmenova, Kelley Regan, Reagan Mergen & Roba Hrisseh

You can also search for this author in PubMed   Google Scholar

Corresponding author

Correspondence to Anya S. Evmenova .

Ethics declarations

Research involving human participants.

All the procedures in this study were evaluated and approved by the Institutional Research Board. All authors have complied with the ethical standards in the treatment of our participants.

Informed Consent

Informed parental consent and student assent were obtained for all individual participants in the study.

Conflict of Interest

There is no known conflict of interest to declare.

Additional information

Publisher's note.

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ .

Reprints and permissions

About this article

Evmenova, A.S., Regan, K., Mergen, R. et al. Improving Writing Feedback for Struggling Writers: Generative AI to the Rescue?. TechTrends (2024). https://doi.org/10.1007/s11528-024-00965-y

Download citation

Accepted : 19 April 2024

Published : 14 May 2024

DOI : https://doi.org/10.1007/s11528-024-00965-y

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

  • Data-based decision-making
  • Generative AI
  • Essay feedback
  • Struggling writers
  • Students with and without high-incidence disabilities
  • Writing instruction
  • Find a journal
  • Publish with us
  • Track your research

The Hechinger Report

Covering Innovation & Inequality in Education

PROOF POINTS: AI essay grading is already as ‘good as an overburdened’ teacher, but researchers say it needs more work

Avatar photo

Share this:

  • Click to share on LinkedIn (Opens in new window)
  • Click to share on Pinterest (Opens in new window)
  • Click to share on Reddit (Opens in new window)
  • Click to share on WhatsApp (Opens in new window)
  • Click to email a link to a friend (Opens in new window)

The Hechinger Report is a national nonprofit newsroom that reports on one topic: education. Sign up for our  weekly newsletters  to get stories like this delivered directly to your inbox. Consider supporting our stories and becoming  a member  today.

essay writing rubric esl

Get important education news and analysis delivered straight to your inbox

  • Weekly Update
  • Future of Learning
  • Higher Education
  • Early Childhood
  • Proof Points

Grading papers is hard work. “I hate it,” a teacher friend confessed to me. And that’s a major reason why middle and high school teachers don’t assign more writing to their students. Even an efficient high school English teacher who can read and evaluate an essay in 20 minutes would spend 3,000 minutes, or 50 hours, grading if she’s teaching six classes of 25 students each. There aren’t enough hours in the day. 

Website for Mind/Shift

Could ChatGPT relieve teachers of some of the burden of grading papers? Early research is finding that the new artificial intelligence of large language models, also known as generative AI, is approaching the accuracy of a human in scoring essays and is likely to become even better soon. But we still don’t know whether offloading essay grading to ChatGPT will ultimately improve or harm student writing.

Tamara Tate, a researcher at University California, Irvine, and an associate director of her university’s Digital Learning Lab, is studying how teachers might use ChatGPT to improve writing instruction. Most recently, Tate and her seven-member research team, which includes writing expert Steve Graham at Arizona State University, compared how ChatGPT stacked up against humans in scoring 1,800 history and English essays written by middle and high school students. 

Tate said ChatGPT was “roughly speaking, probably as good as an average busy teacher” and “certainly as good as an overburdened below-average teacher.” But, she said, ChatGPT isn’t yet accurate enough to be used on a high-stakes test or on an essay that would affect a final grade in a class.

Tate presented her study on ChatGPT essay scoring at the 2024 annual meeting of the American Educational Research Association in Philadelphia in April. (The paper is under peer review for publication and is still undergoing revision.) 

Most remarkably, the researchers obtained these fairly decent essay scores from ChatGPT without training it first with sample essays. That means it is possible for any teacher to use it to grade any essay instantly with minimal expense and effort. “Teachers might have more bandwidth to assign more writing,” said Tate. “You have to be careful how you say that because you never want to take teachers out of the loop.” 

Writing instruction could ultimately suffer, Tate warned, if teachers delegate too much grading to ChatGPT. Seeing students’ incremental progress and common mistakes remain important for deciding what to teach next, she said. For example, seeing loads of run-on sentences in your students’ papers might prompt a lesson on how to break them up. But if you don’t see them, you might not think to teach it. 

In the study, Tate and her research team calculated that ChatGPT’s essay scores were in “fair” to “moderate” agreement with those of well-trained human evaluators. In one batch of 943 essays, ChatGPT was within a point of the human grader 89 percent of the time. On a six-point grading scale that researchers used in the study, ChatGPT often gave an essay a 2 when an expert human evaluator thought it was really a 1. But this level of agreement – within one point – dropped to 83 percent of the time in another batch of 344 English papers and slid even farther to 76 percent of the time in a third batch of 493 history essays.  That means there were more instances where ChatGPT gave an essay a 4, for example, when a teacher marked it a 6. And that’s why Tate says these ChatGPT grades should only be used for low-stakes purposes in a classroom, such as a preliminary grade on a first draft.

ChatGPT scored an essay within one point of a human grader 89 percent of the time in one batch of essays

essay writing rubric esl

Still, this level of accuracy was impressive because even teachers disagree on how to score an essay and one-point discrepancies are common. Exact agreement, which only happens half the time between human raters, was worse for AI, which matched the human score exactly only about 40 percent of the time. Humans were far more likely to give a top grade of a 6 or a bottom grade of a 1. ChatGPT tended to cluster grades more in the middle, between 2 and 5. 

Tate set up ChatGPT for a tough challenge, competing against teachers and experts with PhDs who had received three hours of training in how to properly evaluate essays. “Teachers generally receive very little training in secondary school writing and they’re not going to be this accurate,” said Tate. “This is a gold-standard human evaluator we have here.”

The raters had been paid to score these 1,800 essays as part of three earlier studies on student writing. Researchers fed these same student essays – ungraded –  into ChatGPT and asked ChatGPT to score them cold. ChatGPT hadn’t been given any graded examples to calibrate its scores. All the researchers did was copy and paste an excerpt of the same scoring guidelines that the humans used, called a grading rubric, into ChatGPT and told it to “pretend” it was a teacher and score the essays on a scale of 1 to 6. 

Older robo graders

Earlier versions of automated essay graders have had higher rates of accuracy . But they were expensive and time-consuming to create because scientists had to train the computer with hundreds of human-graded essays for each essay question. That’s economically feasible only in limited situations, such as for a standardized test, where thousands of students answer the same essay question. 

Earlier robo graders could also be gamed, once a student understood the features that the computer system was grading for. In some cases, nonsense essays received high marks if fancy vocabulary words were sprinkled in them. ChatGPT isn’t grading for particular hallmarks, but is analyzing patterns in massive datasets of language. Tate says she hasn’t yet seen ChatGPT give a high score to a nonsense essay. 

Tate expects ChatGPT’s grading accuracy to improve rapidly as new versions are released. Already, the research team has detected that the newer 4.0 version, which requires a paid subscription, is scoring more accurately than the free 3.5 version. Tate suspects that small tweaks to the grading instructions, or prompts, given to ChatGPT could improve existing versions. She is interested in testing whether ChatGPT’s scoring could become more reliable if a teacher trained it with just a few, perhaps five, sample essays that she has already graded. “Your average teacher might be willing to do that,” said Tate.

Many ed tech startups, and even well-known vendors of educational materials, are now marketing new AI essay robo graders to schools. Many of them are powered under the hood by ChatGPT or another large language model and I learned from this study that accuracy rates can be reported in ways that can make the new AI graders seem more accurate than they are. Tate’s team calculated that, on a population level, there was no difference between human and AI scores. ChatGPT can already reliably tell you the average essay score in a school or, say, in the state of California. 

Questions for AI vendors

At this point, it is not as accurate in scoring an individual student. And a teacher wants to know exactly how each student is doing. Tate advises teachers and school leaders who are considering using an AI essay grader to ask specific questions about accuracy rates on the student level:   What is the rate of exact agreement between the AI grader and a human rater on each essay? How often are they within one-point of each other?

The next step in Tate’s research is to study whether student writing improves after having an essay graded by ChatGPT. She’d like teachers to try using ChatGPT to score a first draft and then see if it encourages revisions, which are critical for improving writing. Tate thinks teachers could make it “almost like a game: how do I get my score up?” 

Of course, it’s unclear if grades alone, without concrete feedback or suggestions for improvement, will motivate students to make revisions. Students may be discouraged by a low score from ChatGPT and give up. Many students might ignore a machine grade and only want to deal with a human they know. Still, Tate says some students are too scared to show their writing to a teacher until it’s in decent shape, and seeing their score improve on ChatGPT might be just the kind of positive feedback they need. 

“We know that a lot of students aren’t doing any revision,” said Tate. “If we can get them to look at their paper again, that is already a win.”

That does give me hope, but I’m also worried that kids will just ask ChatGPT to write the whole essay for them in the first place.

This story about  AI essay scoring was written by Jill Barshay and produced by  The Hechinger Report , a nonprofit, independent news organization focused on inequality and innovation in education. Sign up for  Proof Points   and other  Hechinger newsletters .

Related articles

The Hechinger Report provides in-depth, fact-based, unbiased reporting on education that is free to all readers. But that doesn't mean it's free to produce. Our work keeps educators and the public informed about pressing issues at schools and on campuses throughout the country. We tell the whole story, even when the details are inconvenient. Help us keep doing that.

Join us today.

Jill Barshay SENIOR REPORTER

(212)... More by Jill Barshay

Letters to the Editor

At The Hechinger Report, we publish thoughtful letters from readers that contribute to the ongoing discussion about the education topics we cover. Please read our guidelines for more information. We will not consider letters that do not contain a full name and valid email address. You may submit news tips or ideas here without a full name, but not letters.

By submitting your name, you grant us permission to publish it with your letter. We will never publish your email address. You must fill out all fields to submit a letter.

Your email address will not be published. Required fields are marked *

Save my name, email, and website in this browser for the next time I comment.

Sign me up for the newsletter!

essay writing rubric esl

COMMENTS

  1. ESL Essay Writing Rubric for Teachers

    Scoring essays written by English learners can at times be difficult due to the challenging task of writing larger structures in English. ESL / EFL teachers should expect errors in each area and make appropriate concessions in their scoring. Rubrics should be based on a keen understanding of English learner communicative levels.This essay writing rubric provides a scoring system which is more ...

  2. PDF Essay Rubric

    Essay Rubric Directions: Your essay will be graded based on this rubric. Consequently, use this rubric as a guide when writing your essay and check it again before you submit your essay. Traits 4 3 2 1 Focus & Details There is one clear, well-focused topic. Main ideas are clear and are well supported by detailed and accurate information.

  3. Essay Rubric

    Grading rubrics can be of great benefit to both you and your students. For you, a rubric saves time and decreases subjectivity. Specific criteria are explicitly stated, facilitating the grading process and increasing your objectivity. For students, the use of grading rubrics helps them to meet or exceed expectations, to view the grading process ...

  4. Two Customizable ESL Writing Rubric Templates

    An ESL writing rubric can make the process significantly easier for you and transparent for your students. In this blog post, we will show you two ESL writing rubric templates that you can use to grade your student's writing. But first, let's understand what is an ESL writing rubric. Want to save hours of repetitive typing for free?

  5. ESL Writing Rubrics

    ESL Writing Rubrics. Erin is currently teaching English at a university in South Korea and has a Master's degree in Education. There are many benefits to using a rubric in your ESL writing classes ...

  6. ESL Writing: A Simple Scoring Rubric

    This ESL writing rubric uses a 1 to 4 scoring system. A well-structured paragraph (clear topic sentence, good support and conclusion). Grammar and spelling errors are present though the meaning of the passage is clear. A good topic sentence though the paragraph lacks unity (the main idea is not well supported) or there are too many grammar or ...

  7. PDF B1 Preliminary for Schools Assessing writing for Cambridge English

    Teachers around the world spend many hours evaluating learners' writing. This guide is for you. With lots of practical tips and real examples, it will help you to develop and . assess learners' writing skills in preparation for the Cambridge English B1 Preliminary for . Schools exam. About B1 Preliminary for Schools

  8. PDF B2 First for Schools Assessing writing for Cambridge English

    The B2 First for Schools Writing paper has two parts and lasts for 1 hour and 20 minutes in total. Candidates have to show that they can write diferent types of text in English. Part 1: Writing an essay. Part 2: Writing an article, email, review or a story.

  9. PDF Writing Assessment and Evaluation Rubrics

    Holistic scoring is a quick method of evaluating a composition based on the reader's general impression of the overall quality of the writing—you can generally read a student's composition and assign a score to it in two or three minutes. Holistic scoring is usually based on a scale of 0-4, 0-5, or 0-6.

  10. iRubric: ESL Writing Assessment (Intermediate) rubric

    ESL Writing Assessment (Intermediate) General ESL Writing assessment. Intermediate level. 100-120 words. - None of the writing is about the topic. - The essay does not explicitly answer the question. - The writing is disorganized, having only a body paragraph. - No logical progression of ideas, no use of transitions between paragraphs.

  11. Rubric Best Practices, Examples, and Templates

    A rubric is a scoring tool that identifies the different criteria relevant to an assignment, assessment, or learning outcome and states the possible levels of achievement in a specific, clear, and objective way. Use rubrics to assess project-based student work including essays, group projects, creative endeavors, and oral presentations.

  12. ESL Writing Rubric: Evaluate ESL Students' Writing Easily & Fairly

    ESL Writing Grading Rubric. My solution to ensure fair grading and also appear this way to my students is to use this ESL Writing Grading Rubric. I use it, even though within reading the first couple sentences of a 5-paragraph, 500-word essay, I usually know what grade the student will get.This is especially true if it's going to be an A, or D/F.

  13. ESL 123 Final Essay Scoring Rubric

    ESL 123 Final Essay Scoring Rubric. Scoring Guide for ESL 123 Final Essays. The scoring guide contains 5 categories: content, organization, vocabulary, grammar, and spelling and mechanics. The categories in each essay may have different ratings, but the majority should be level 3 or above for a passing composition. Clear Pass (4)

  14. PDF Five-Paragraph Essay Writing Rubric

    You use some of the target vocabulary from the subunit. Vocabulary/word choice is impressive. All words are used appropriately. Vocabulary/word choice is adequate, but could be improved. More adjectives, adverbs, and descriptive words are needed. Vocabulary/word choice is clearly limited, affecting written communication.

  15. Academic essay rubric

    Academic essay rubric. This is a grading rubric an instructor uses to assess students' work on this type of assignment. It is a sample rubric that needs to be edited to reflect the specifics of a particular assignment. Download this file. Page. /. 3. Download this file [63.33 KB] Back to Resources Page.

  16. Effective Essay Writing Rubrics for Academic Success

    When it comes to effective essay writing, understanding the key elements is essential for academic success. Here are some important elements to keep in mind: Thesis Statement: A clear and concise thesis statement that presents the main idea of your essay. Structure: A well-organized structure with introduction, body paragraphs, and a conclusion.

  17. PDF Argumentative essay rubric

    Arrangement of essay is unclear and illogical. The writing lacks a clear sense of direction. Ideas, details or events seem strung together in a loose or random fashion; there is no identifiable internal structure and readers have trouble following the writer's line of thought. Few, forced transitions in the essay or no transitions are present.

  18. SAT Essay Rubric: Full Analysis and Writing Strategies

    Writing. Your Writing score on the SAT essay is not just a reflection of your grasp of the conventions of written English (although it is that as well). You'll also need to be focused, organized, and precise. The response demonstrates little or no cohesion and inadequate skill in the use and control of language.

  19. 15 Helpful Scoring Rubric Examples for All Grades and Subjects

    Try this rubric to make student expectations clear and end-of-project assessment easier. Learn more: Free Technology for Teachers. 100-Point Essay Rubric. Need an easy way to convert a scoring rubric to a letter grade? This example for essay writing earns students a final score out of 100 points. Learn more: Learn for Your Life. Drama ...

  20. ACT Writing Rubric: Full Analysis and Essay Strategies

    1. demonstrate little or no skill in writing an argumentative essay. The writer fails to generate an argument that responds intelligibly to the task. The writer's intentions are difficult to discern. Attempts at analysis are unclear or irrelevant. Ideas lack development, and claims lack support.

  21. English Essay (Business

    Cheap Business Essay Writing Services. Before being accepted into our company, we underwent extensive background checks. Check their credentials to confirm that they have been writing professionally for some time. If they are members of professional associations, check, for instance. Some students may have difficulty completing their research ...

  22. Improving Writing Feedback for Struggling Writers: Generative AI to the

    Generative AI has the potential to support teachers with writing instruction and feedback. The purpose of this study was to explore and compare feedback and data-based instructional suggestions from teachers and those generated by different AI tools. Essays from students with and without disabilities who struggled with writing and needed a technology-based writing intervention were analyzed ...

  23. PROOF POINTS: AI essay grading is already as 'good as an overburdened

    And that's a major reason why middle and high school teachers don't assign more writing to their students. Even an efficient high school English teacher who can read and evaluate an essay in 20 minutes would spend 3,000 minutes, or 50 hours, grading if she's teaching six classes of 25 students each. There aren't enough hours in the day.