Across the Disciplines wac.colostate.edu/atd
A Journal of Language, Learning and Academic Writing ISSN 554-8244
DOI: https://doi.org/10.37514/ATD-J.2006.3.1.02
Across the Disciplines is an open-access, peer-review scholarly journal published on the WAC
Clearinghouse and supported by Colorado State University and Georgia Southern University. Articles
are published under a Creative Commons BY-NC-ND license (Attribution-NonCommercial-NoDerivs)
ISSN 1554-8244. Copyright © 1997-2017 The WAC Clearinghouse and/or the site's authors,
developers, and contributors. Some material is used with permission.
The Complexities of Responding to Student Writing; or,
Looking for Shortcuts via the Road of Excess
Richard Haswell, Haas Professor Emeritus, Texas A&M University, Corpus Christi
Abstract: In all academic disciplines college teachers respond to student writing
with shortcutschecksheets, correction symbols, computer style checkers, etc. But
while these methods save teachers time, do they help students improve their
writing? A survey of research into teacher commentary, conceived of as a contextual
discourse activity, initially questions the efficiency of many shortcuts because it
finds complexities in all activity areas, in regulation (criteria, rules of genre and
mode, disciplinary styles, and standards), consumption, production, representation,
and identity. The research, however, also recommends particular shortcuts and
methods of revising them for better efficiency and effect. It especially recommends
restricting the volume of teacher commentary in ways that are task, discipline, and
learner specific.
If a way to the Better there be, it exacts a full look at the Worst.
-- Thomas Hardy
The road of excess leads to the palace of wisdom
-- William Blake
The more complex tasks are, the greater the temptation to simplify them. It is a temptation that
faculty responding to student writing have not always resistedhence prêt à porter devices such
as checksheets, correction symbols, mnemonic acronyms, weighted criteria scales, peer-critique
handouts, and computer feedback software. Therein lie two paradoxes. Simplifying a complex task
may take the challenge and fascination out of it and end up making it laborious in a new way, boring
and repetitious for both teacher and student, on the order of vacuuming the house. The other paradox
has more serious implications and is the major concern of the present essay. Latent in a complex task
may exist labor-saving methods, but how can they be extracted except through a full look at the task's
true complexity? That takes labor, precisely the commodity that people who need shortcuts have in
short supply. Sticking to current shortcuts postpones that full look, yet without it there is no
guarantee they are the most efficient.
Perhaps it is time for an examination of the problematics entailed when teachers, anywhere in the
university, respond to student writing. By response, I should emphasize, I do not mean the mere
grading or evaluating of a piece of writing but further the recommending of ways to improve it. How
can that formative act of response be made more workable, for teacher and for student? My answer
can be encapsulated by rewriting Thomas Hardy's famous line in defense of "meliorism": If a way to
Haswell 2
the Efficient there be, it exacts a full look at the Complex. When it comes to the obligation to put words
to a student's words, teachers in all disciplines, those draft horses of instruction, might well choose
William Blake's road of excessexcess of understanding.
Obviously my argument is not that shortcuts in response signal poor teaching. Among other things
that argument would be hypocritical on my part, since one such labor-saving method, called by its
author "minimal marking" and still in wide-spread use after nearly a quarter of a century, was
devised by myself (Haswell, 1983). My argument here is that if we want shortcuts (and who doesn't?)
then we should find those that are doubly workableboth labor-effective for the teacher and
learning-efficient for the studentand that a quick start to finding them is through an exploration of
the true complexity of teacher response to student writing. With luck the pay-off will be directions
for new solutions and for improvement, if not trashing, of old stopgaps.
My plan is to survey the major components of the activity of teachers responding to student writing,
with an eye toward complexities that have been found there by researchers. In doing so I run a risk.
That is of alienating some faculty in a way I do not wish. As we all know, teachers in all academic
fields put in their hours gracing the margins of student writing with commentary. Chemistry lab
reports and health-care case studies and mathematical proofs, when written by students, beg for
writerly advice no less than first-year English-composition essays. But academic genres, with
rhetorical aims and strategies peculiar to the discipline, sometimes beget distinct instructional
response. Differences in the reception of disciplinary discourse, as we will see, are a major source of
some of the complexities under review. Further, I am afraid that the urge to simplify may be stronger
when teachers think about their own discipline's genres and teaching practices. Familiarity breeds
temptationthe temptation to cut corners. It is true that researchers have sometimes found faculty
across academic divisions, including English, ranking or otherwise sorting the criteria for evaluating
writing in rather similar ways (e.g., Spooner-Smith 1973; Michael et al., 1980; Eblen, 1983; McCauley
et al., 1993; Smith, 2003), but we will see that as often researchers have found sharp differences in
understanding and application of the twists and turns of response. The risk is teachers imagining that
when research is cited deconstructing their distinctive response to writing, their discipline is being
put down. I can only trust everyone will soon notice, as the findings accrue, that the even gaze of
research has observed pretty much the same curiosity in every field, that the ecology of response
its full human, social, and institutional contextis more complex than the customary practice of
response seems to warrant.
A Model, Sufficiently Complex, For Surveying Response
Why is instructional response to writing so complicated, so problematic? Not surprisingly, there are
any number of answers to this question, depending upon the questioner's angle of approach.
Response integrates many activities, among them a reader trying to interpret and understand a text
(Kucer, 1989), a mentor encouraging a sometimes resistant apprentice (Young, 1997), and a person
socially and contextually sizing up another person (Bazerman, 1989), interacting with that person
(Anson, 1989), and trying to collaborate together on a task (Martin & Rothery, 1986). As new
approaches take center stage, new answers no doubt will accumulate.
For the purpose of this essay, my own answer derives from theories of discourse acts. I want to look
at instructional response to writing as an activity supported by and supporting normal discourse
practices, an activity no different and no less complex than recommending a book or answering an e-
mail. Response, of course, has long been understood in terms of its ecology or its embeddedness in
group, institutional, and social practices and constructions (Lucas, 1988; Gottshalk, 2003). Discourse
activity theory has added the complication that as all human practices, teacher-student interactions,
The Complexities of Responding to Student Writing
3
including teacher response to student writing, are mediated by cultural tools, especially language but
not exclusively language (Russell, 1995; Evans, 2003). The advantage to a discourse activity
understanding of response is that it locates and embraces a breadth of interacting factors. It will
respect, for instance, the student's perspective as well as the teacher's; will consider outcomes
equally in terms of the learner, the teacher, the course, the academic department, or society at large;
will treat time-on-task on a par with self-image; will accept technologies of response that arose with
the post-Civil War pencil eraser and the post-Vietnam War style checker. The analysis of cultural
discourse activity assumes a field dynamics, and a vision of teacher response as interactive and
intermediating with everything in its surround is what I need for this examination of research
findings and its main argument that such response is anything but simple.
For my survey, let me borrow a recent model of cultural discourse activity that seems comprehensive
yet neither too detailed nor too simplistic, Paul du Gay's "circuit of culture" (1997). Indeed it was
originally designed for college students, organizing at least three volumes of a series of cultural-
studies textbooks published by London's Open University. In this flow chart (Fig. 1), "through which
any analysis of a cultural text or artifact must pass if it is to be adequately studied" (du Gay, 1997, 3),
language practices interrelate dynamically through five activity
nodes: consumption, identity, representation, regulation, and production. Application of the model
does not require specialist knowledge. As a test illustration, take an 14-year-old who reads and then
buys a t-shirt that says "SCHOOL SUCKS" (consumption) because it appeals to her vision of herself
(identity) and next day wears it to class to make a statement about herself and perhaps to please her
friends (representation). The principal soon hears about it from her teachers and with hardly a word
sends her home to change clothes (regulation). It is not long before the company making the t-shirt
becomes alarmed at public reaction and shelves that particular slogan (production). A cooked-up
example, certainly, but not that far removed from the actual discursive chain of events jump-started
last year by a college junior who submitted a paper to me with borrowed language that I readily
traced with the help of Google.com. At every node in du Gay's "circuit of culture," response to
language may operate. As may efficiency in response to language. The 14-year old probably admires
the economy of the two-word sentence in expressing her resistance to school. The principal probably
prefers just sending her to her parents rather than giving her a lecture.
Haswell 4
Figure 1: The Circuit of Culture, from Du Gay, 1997, Production of Culture/Cultures of Production
Regulation, the Epicenter of Instructional Response to Writing
I start with the factor of regulation, as it is perhaps the basic motive for teachers and their response
to student writing. The term "regulation" may seem too Draconian for a service that teachers in all
disciplines view as beneficial and perhaps even freeing for students. But beneficial for what
purposes? Freed in what directions? Response is regulative because it hopes to move novices and
their writing toward some more mature psychological, professional, social, or cultural ends. This is
obvious with commentary that is corrective. "Your point is unclear" does its small part to keep
obscurity out of public discourse. But even response that indicates no more than approval or liking
(Elbow, 1993) reinforces some endin that case an end already achievedand thus helps regulate
discourse around it. Back of all instructional response lie at least four muscular regulative
apparatuses, intricately fused: criteria, or teacherly notions of good writing; rules of genre and mode,
or parameters erected by the established history of particular discourse forms; disciplinary styles, or
parameters erected by professional groups; and standards, or levels of writing achievement set by
social groups. Each is elaborate. Interplay, constraints, and contradictions among them are rife.
Criteria provide the stock replies every teacher can call up at will, often cannot help but call up.
Criteria comprise nothing less than a vast armory of lore that ranges from knowing the correct
spelling of "existence," to feeling a break in flow for lack of a "this," to judging that a summary
paragraph too much repeats the abstract and will offend the reader in the field. The size of the store
can be judged by the thickness of usage handbooks, scores of which are printed, reprinted, and
revised every year. Even taxonomies of criteria are so elaborate they look biological. Mellon (1978)
identified 34 "compositional competencies," each of which can be judged lacking in a piece of student
writing. An international cross-cultural study of school writing located 28 main criteria used by
teachers (Purves & Takala, 1982); Grabe and Kaplan (1996) reduced second-language competencies
to 20 main categories encompassing 47 distinct traits. Broad (2003) found writing teachers
evaluating first-year writing portfolios with 46 textual criteria, 22 contextual criteria, and 21 others.
Although, as I have said, faculty across the campus tend to agree on overarching readability criteria
The Complexities of Responding to Student Writing
5
(unity, focus, clarity, coherence, etc.), there can be substantial differences in they way the terms are
defined and applied and in the way other criteria are prioritized. Science teachers are more willing
than humanities teachers to excuse stylistic infelicities if the content is accurate -- a finding of long
standing (e.g., Halpern, Spreitzer, & Givens, 1978; McCarthy, 1987). Arts faculty like the term
"creative," humanities faculty "eloquent," science faculty "analytical" (Zerger, 1997), predilections
that can underlie rater disagreement in cross-campus writing-assessment programs. The same
problems occur within academic units. Professors of business writing agreed that "wordiness" was
an important criterion in judging business letters yet gave differing grades to the same letter, in part
because they did not agree on a definition of the term (Dulek & Shelby, 1981).
Regardless of discipline-specific preferences, good writing traits hang over the head of teachers like
a mort main. Criteria are long lived ("paragraph unity" has spanned four centuries now) and wield
the force of etiquette. The Harbrace College Handbook, into its 57th year and 13th edition, has an
appendix of 73 abbreviations for teachers to use in responding to student writing. The need to keep
the abbreviations the same, in fact, kept the editors from updating the criteria (Hawhee, 1999).
Whether teachers rely on Harbrace or some disciplinary style manual, most find themselves using
correction symbols and abbreviations¶ for "You need to start a new paragraph here," "SF" for
"You have committed a sentence fragment," "ref?" for "These are facts are not common knowledge
and you need to cite your source for them." The custom is evidence of the multiplicity of criteria,
which cover the field so thoroughly that any passage of student writing will scare up a flock of them.
The custom also betrays the plight of responders, who hope that the makeshift of a limited correction
lexicon, often ill understood by students, will control this flock that persists in coming back, season
after season.
Genre and mode rules complicate matters because they mediate or undercut the assumed universality
of criteria. This is a fact no amount of pleading for global writing traits will dispel. Murray (1982a)
may argue that six qualities bridge all discourse modes, including fiction and nonfiction: meaning,
authority, voice, development, design [organization], and clarity. But this is a position easy to counter,
for instance, the universality of "voice" with the voicelessness or impersonalization of technical
documentation, or the universality of "meaning" with the deliberate non-sense of surrealist poetry.
Besides, all of Murray's qualities are mediated quite differently by different genres and modes. In a
personal journal "authority" may be furthered by emotive language, in a travelogue by proper names,
in a professional article by intertextual reference, in a public notice by organizational affiliation.
College teachers do not have trouble recognizing these differences. But their students do (McCarthy,
1987), in part because there are deeply rooted differences in the kind of genres assigned in different
departments, even with genres that are supposedly cross disciplinary and WAC legitimized. Hobson
& Lerner (1999), for instance, found write-to-learn principles cited as a purpose for course
assignments by 91% of their arts and science faculty, 44% by their pharmacy practice faculty. Also
teachers often do not have a ready synopsis of the rules for the particular genre they are asking
students to write. Further trouble lies in teachers applying response that is presumably consistent
since it is regulatory, when their writing assignments include different genres with different
discourse rules.
Disciplinary style adds to the trouble. All disciplines have their own Harbraces, both literally in the
form of published style manuals and intuitively in the heads of their practitioners, and the rules
governing rhetorical acceptability differ from one field to the next as much as the Council of Biological
Editors citation style differs from that of the Chicago or the American Psychologist Association. From
usage as minute as the way the colon is used in titles (the absence of a colon in my own title follows
a slightly antiquated convention in the humanities) to usage as pervasive as the way evidence is
gathered, respected, and presented, disciplines regulate every aspect of discourse in a bewildering
Haswell 6
variety of ways. For instance how should a teacher in a general-education course respond to the
abundance of unqualified generalities common in first-year college writing? Political-science
discourse seems to recommend such certainty (Ober, Zhao, David, & Alexander, 1999), philosophical
argumentation encourages more formal qualification (Jason, 1988), science reports expect more
hedging in the presentation of data (Luukka & Markkanen, 1997), historical papers more hedging in
the claims (Meyer, 1997).
One could argue that first-year college students, most of whom have hardly imagined a major for
themselves even if they have declared one, are unready for disciplinary styles, but then by what style
are they to be regulated? There is no ready, or efficient, answer to this complication. Needless to say,
in their first-year content courses students often find disciplinary regulation dispensed anyway,
piecemeal and ragtag. Once in a study of the effect of first-year English composition courses upon
essay-examination performance in other courses, my co-researchers and I were startled to see our
ratings of the students' introductions in an environmental science course drop sharply from the first
to the second examination. Between the two tests English-department writing teachers had trained
the students to compose introductory paragraphs in essay-exam questions, while at the same time,
the environmental-science teacher told students as he handed back the first test that when he saw
an introduction in an examination essay he assumed the student didn't know the answer and was
bluffing. He wanted no introduction at all. Here a reasonable response shortcut, advising students
about the rules before they do the writing, was not as simple for the students to absorb as the teacher
may have thought.
Standards, as this incident shows, mediates regulation and response with its own conflicting rules.
Were the English teachers now obligated to teach and judge introductions in a way contrary to their
tastes and beliefs? Couldn't the environmental-science teacher have allowed a sentence or two by
way of introduction? How was I supposed to respond one time when my advanced-composition
students, and I'm not making this up, told me that their journalism teacher had said no sentence could
have more than one comma? Standards have a way either of over-regulating or under-regulating
teacher response to writing. They over-regulate when they force teachers to focus their response on
a limited number of writing traits, excluding traits the teachers know are important and highlighting
traits they know are inconsequential or even vitiating. This bind is not trivial. Daily thousands of
teachers, in and beyond the schools, are penning or voicing response to student writing under the
pressure of state and institutional mandated tests, an effect justly given the bodily term "washback."
Teachers who ten years ago were disapproving of the five-paragraph theme, an essay organization
even high-school juniors recognize as juvenile, today are praising students for accomplishing it.
On the other hand, standards under-regulate response because they tend to be overgeneralized (the
environmental-science teacher's standard for introductions is not typical). Take the Writing Program
Administrators' "Outcome Statement for First-Year Composition," a list of 34 writing competencies
students should have mastered before they are sophomores (Writing Program Administrators,
2000). (Some of the WPA competencies, it goes without saying, were not mentioned by Mellon, 1978
in his own list of 34 some twenty-two years earlier.) Among other skills, the WPA-certified student
will be able to "adopt an appropriate voice, tone, or formality." But rhetorically voice, tone, and
formality can be achieved in a hundred different ways, including frequency and kind of hedging.
"Appropriate" just brings back the confounding of genre, mode, and discipline, and adds the further
confounding element of level of accomplishment, confounding because unspecified. At what point
does a student's performance become inappropriate? It depends on a raft of factors, including
discipline and teacher. Beason (1993) found the instructor of a writing-enriched course in business
law devoting 32% of paper commentary to the criterion of focus, yet in three other writing-enriched
The Complexities of Responding to Student Writing
7
courses, in journalism, dental hygiene, and psychology, the instructors reacted to focus as
inappropriate in their students' papers less than 10% of the time.
Nascent in writing standards is the concept of "acceptable level" or "irritation score," also called
"error gravity" by ESL researchers. It leads to some of the most complex of writing response
interactions. For instance native English-speaking instructors thought second-language English
errors with Greek students writing more serious when the mistakes impeded comprehension, while
native Greek-speaking instructors thought the errors were more serious when they were "basic" or
mistakes that writers should know how to avoid (Hughes & Lascaratou, 1982). Plenty of other factors
than native language mediate error gravity, of course, including sex of teacher (Hairston, 1981),
perceived responsibility for the student (Janopoulos, 1992), academic field (Kantz & Yates, 1994),
and rank or years teaching (Kantz & Yates, 1994)as well as personal idiosyncrasy. An
undergraduate literature teacher of mine entered only one kind of mark on my papers, Gargantuan
commas that cut down through five lines of my text. Now that's grave. Workable shortcuts for
teachers responding to error gravity and other issues of standards are not going to be found in the
WPA Outcomes, or anywhere else that I know of.
In short, regulation leaves too much open, gives the teacher options too numerous with boundaries
too flexible. This is where the paradox of response and complexity becomes clear. The more complex
writing response is seen to beand regulation is only one of five main factorsthe less chance, it
would seem, of finding simplified methods that will let teachers respond to students in productive
and efficient ways.
Consumption, the Functionality of Response
Inevitably the terms "productive" and "efficient" route us toward a new node in the circuit of culture
(Figure 1). There is nothing in the nature of regulation ensuring that the rules and conventions will
be followed or even understood by students. Consumption, at least in terms of discourse practices,
operates within the realm of functional discourse processing, working through focus on language
medium and on reader comprehension, assimilation, recall, and application. In terms of teacher
response, consumption asks some hard questions. Is the communication channel between teacher
and student viable? Are there better channels? Does instructional response actually lead to its
purposeful end, improvement in the student's writing?
To answer, researchers have watched teachers sit in conference with their marginal comments on
the paper between them and the students, have watched them read second drafts of first drafts that
they have previously responded to, have queried the students over the meaning they draw from
teacher comments, have charted the student's writing improvement, or lack of it, paper after paper,
week after week, in relation to the marginal commentary. Their findings are often discouraging. To
begin with, although students, especially second-language students, may pay more attention to
comments on drafts than on final papers (Ferris, 1995), overall they don't consume teacher response
very well. Research into student perceptions have demonstrated this again and again. Students are
avid for commentary (though they may first look at the grade), but when forced to explain their
teachers' comments, they misinterpret a shocking portion of it. When forced to revise, they
assiduously follow the teacher's surface emendations and disregard the deeper suggestions
regarding content and argumentation. They prefer global, non-directive, and positive comments but
make changes mainly to surface, directive, and negative ones. In sum, they want lots and certain kinds
of response, but have trouble doing much with what they ask for (Hayes & Daiker, 1984; Ziv, 1984;
Straub, 1997; Winter, Neal, & Waner, 1996; Underwood & Tregidgo, 2006). Of course, some of the
blame rests on teachers, who often think they are positively emphasizing central qualities such as
Haswell 8
reasoning, genre form, and reader awareness while in fact the bulk of their commentary dwells
negatively on surface mistakes and infelicities of syntax and word choice (Kline, Jr., 1976; Harris,
1977; Connors & Lunsford, 1988).
To make matters worse, the problems lie not so much with students or with teachers but with the
interaction between them. Some of the most disturbing investigation into response seems to show
that students and teachers operate under very different evaluative sets. Whether it is due to age,
gender, experience, expertise, social position (Evans, 1997), or classroom dynamics (O'Neill & Fife,
1999), students and teachers tend to consume writing quite differently. They have trouble speaking
the same language of response because their responses to the writing itself are so far apart. Students
tend to read to comprehend, teachers to judge (Gere, 1977). Students interpret detail realistically,
teachers emblematically and literarily (Hull & Rose, 1990). Students prefer an orally produced style,
teachers a professionally written style (Means, Sonnenschein & Baker, 1980). Students think of
writing as a maze or an obligation over which they have no control, teachers as an activity creatively
in the hands of the writer (Tobin, 1989). Students look on writing options as right or wrong or else
an endless shelf of choices with no way to choose, teachers as rhetorical choices leading to the best
strategy (Anson, 1989). Students take a first interpretation as the only one, teachers as a first position
to be revised upon consideration of other interpretations (Earthman, 1992). Students place the most
importance on vocabulary, teachers on substance (Yorio, 1989). In his study of student and English
teacher reading predilections, Newkirk (1984) observed that students preferred the obvious and
familiar, teachers the unusual and new; students overt meanings, teachers unstated; students explicit
organization with headings, teachers implicit organization; students a bland tone, teachers an
emotionally enhanced tone; students a high register, teachers a middle register.
If students and teachers conflict so much in preference for style and attitude toward the author's task,
then we may have a partial reason for another discouraging set of research findings. Little consistent
association between writing improvement and volume or kind of response has been documented. It
seems that the time-honored effort of inscribing commentary on student papers is sometimes
honored with a proven pay-off and sometimes not, at least within the time span of the course. In a
characteristic pre-post study, Marzano & Arthur (1977) compared the effect of three kinds of
commentary, one that only marks error, one that only suggests revision, and one that only fosters
critical-thinking. They found minimal effect on the students' writing with all three types (for
somewhat dated reviews of such studies, see Knoblauch & Brannon, 1981; Fearing, 1980). The type
of response that holds the best record seems to be praise, as opposed to commentary that finds only
things to correct (e.g., Denman, 1975), but even that record is inconsistent, and students rightfully
are leery of praise when it is general and gives little direction. All this is no reason to return to the
practice, universally condemned by most teachers, of restricting response to a grade, or worse of
hoping that sheer volume of writing without any teacher feedback will improve student skills and
not merely ingrain bad habits (the negative research findings on that hope are reviewed by Sherwin,
1969, pp. 156-167).
Obviously the circuit of response connects consumption with a new node, production, asking if
problems with student use of commentary couldn't be solved with teachers spending more time
producing it. Some modes of response must be more effective than other modes. But the past erose
research record on effectiveness does help explain a history of teacher restlessness in response
practices, usually reflecting a search for methods that will work better with the student and save the
teacher time. There have been historical upswings, for instance, toward audiotaped response, one-
on-one conferencing, hypertext commenting, and peer evaluation. The last has been and remains
especially popular, in part because it helps reduce the paper load for teachers (see below), but also
because it seems a way to translate critique into a language style and preference set that students
The Complexities of Responding to Student Writing
9
can understand, e.g., from their own peers (Spear, 1988). But even with peer-response methods,
disillusionment has been setting in, though perhaps less with teachers than with students who are
sick of so much of it.
Students, after all, are the main consumers of instructional response to writing. Not unreasonably
they want their money's worth. Most students yearn for teacher commentary, even when they don't
make the best use of it. In an old but well controlled semester study (Stiff, 1967), experimental
students found their papers with marginal comments only, or with end comments only, or with both
marginal and end comments, and although the writing of the three groups were not significantly
different at the end of the semester, the first two groups complained that they had been neglected.
But can teacher shortcutting avoid student short-shrifting? All across campus students write papers,
and they all expect a reaction beyond a grade or a number. The question is how can teachers best
produce it.
Production, the Bane of Response
The conventional wisdom is that formative response to student writingnot just corrected and
graded but with recommendations for improvementis produced by teachers only with intense
labor. Teachers feel this to the bone, and although deans seem to ask for new proof of this labor nearly
every biennium, teachers can show that it has long been documented. As early as 1913, in a summary
of reports from 700 teachers, the Modern Language Association calculated that the average college
teacher spent 5.2 minutes per page, if the writing was "carefully criticized" (6). In 1955, based on the
reports of 600 teachers, Dusel calculated 7.7 minutes a page. In 1965 Freyman calculated that "lay
readers"ancillary help for overworked high-school teacherscould do a full reading at 8.1 minutes
a page. The increase in response time may be a shift toward more words per page (hand-written
giving way to type-written), or of writing commentary from marking only for surface mistakes
toward commenting on more substantive points, which takes much longer. Student papers also have
grown longer. Before the 1960s, the model was the "500-word theme," or less than two typed pages.
Today first-year college essays (not counting the full-blown "research paper") probably average over
four pages. Connors and Lunsford (1988) cite their own and others' research indicating an average
paper length of 162 words in 1917, 231 words in 1930, and 422 words in 1986and Andrea
Lunsford (personal communication) says that paper length now averages more than a thousand
words! Currently the National Council of Teachers of English assumes a minimum response time of
20 minutes per paper (2002). We are talking about an huge expenditure of teacher effort. At a
conservative 4 pages per essay, 7 minutes per page, and 25 students, the English or rhetoric
department composition teacher is spending between eleven to twelve hourspure labor, no
breaksbent over an initial response to just one set of papers. That leaves out of the total the time
devoted to second and third drafts.
Long hours marking papers is the mark of the composition teacherthe profession's mark of Cain,
some would say. It is no wonder otherwise gentle colleagues in English departments will fight tooth
and nail to keep out of the writing classroom. Nor is it any surprise to find peppered throughout
composition bibliographies titles such as "When It's One Down and 1,249 Themes To Go," "Avoiding
Martyrdom in Teaching Writing," "Minimal Marking," "Plus, I'm Tired," "Staying Alive," and (my
favorite) "Practical Proposal to Take the Drudgery Out of the Teaching of Freshman Composition and
to Restore to the Teacher His Pristine Pleasure in Teaching." Of the many volumes in the National
Council of Teachers of English series Classroom Practices in Teaching English, by far the oldest still
in print is called "How to Handle the Paper Load" (Stanford, 1979) -- and as I write NCTE has
announced a companion volume to it in the works. Both volumes will be of interest, of course, to
Haswell 10
many teachers outside of English departments, who assign writing in classes with enrollments in the
hundreds.
No wonder that new ways to handle the paper load, advances in efficiency in the production of
response, have a long history in the teaching of college composition. The use of lay readers (called
"reading assistants" at Vassar before they were phased out in 1908) may be one of the earliest, but it
was only a harbinger. Here is a short list of shortcuts, with date of earliest record I can find in the
post-WWII literature.
Mark only the presence of problem, leaving it up to the student to locate and correct it
(1940)
Use a projector to respond to student writing in class (1942)
Use a checklist, or rubberstamped scale of criteria (1950)
Hold one-on-one conferences to respond (1946)
Have fellow students read and respond to papers (1950)
Hold one-on-two or one-on-three conferences to respond (1956)
Record comments on audiotape (1958)
Respond only to praiseworthy accomplishments (1964)
Have students evaluate their own essays (1964)
Respond only to a limited number of criteria (1965)
Have students use computerized grammar, spelling, or style checkers (1981)
Add comments to the student's digital text with word-processing footnotes or hypertext
frames (1983)
The last two items elevate us into electronic space, with promise of a quantum leap in efficiency. The
idea that computer programs might be devised to do the work of humans in evaluating student
writing was proposed in the 1950s (e.g., Wonnberger, 1955) and first effected in the late 1960san
enterprise that ran about a decade behind machine translation and text-corpus analysis for linguistic
and literary purposes. In 1968, Page & Paulus described a program with 30 machine-scorable,
statistically regressed measures that achieved .63 correlation on ratings of student essays by
humans, as good as the human raters themselves could intercorrelate. With advances in
computational algorithms, measuring semantic and topical structures as well as lexical and
grammatical ones, businesses dealing in computerized essay evaluation currently announce much
higher correlations. Currently the College Board's WritePlacer, the Educational Testing Service's e-
rater, and ACT's e-Write have already been used to admit undergraduates and place them into writing
courses. Feedback programs such as ETS's Criterion and Vantage's MY Access! can analyze classroom
writing in lieu of the teacher. Competing programs are cropping up like Deucalion's soldiers. The
profit motive is irresistible, of course, and the sales are going to increase. So it is important to
The Complexities of Responding to Student Writing
11
understand why, in the mind and response practice of most English teachers today, the role of e-
valuation of student writing is largely inconsequential.
Historically, it did not begin that way. Although Page and Paulus were financed by the College
Entrance Examination Board, their success was represented to the writing-teacher world as a way to
reduce the paper load, "a proper program for correction of essays," helping reduce the "tremendous
amounts of time out of class" teachers have to spend doing "an ideal job in essay analysis," "equalizing
the load of the English teacher with his colleagues in other subjects" (p. 3). Their report even ends
with a program, devised by Dieter Paulus, showing how the machine can "make comments about a
student's essay" and do so "in a real school situation" (p. 150). In another early promotion of
computer grading, Slotnick and Knapp (1971) draw a proto-computer-lab scenario for teachers
"burdened with those ubiquitous sets of themes waiting to be graded" (1971, p. 75). In the laboratory
of the future, students would type their first drafts on a special typewriter and a clerk would feed the
typescript through a "character reader" (i.e., scanner); that night the machine grader would encode
the tape and print out responses; next day the student would have commentary to use to revise for a
second draft. Until then, the teacher hasn't seen the paper. A fantasy world in 1971suitable
scanners were some years away. But the future didn't happen, at least not quite yet. Today, with all
the equipment in place and universities often willing to purchase the software, composition teachers
rarely use it in this way, though teachers in other departments sometime do not show the same
reluctance.
The reason has to do with production. In the activity of reading and diagnosing student writing for
instructional purposes, to progress toward machine scoring is to regress back to the single-grade
mark on a paper. Despite multiple-regression or semantic-map analysis of student texts, the
computer program ends up with a single digit response: the paper is a 3, or a 5. (This is one reason
why the computer can be programmed to correlate so well with testing-firm human raters, who also
are also reducing an essay to a single-digit score as fast as they can; see Haswell, 2006.) Machines can
evaluate a text but they cannot diagnose an author, cannot help lead the writer to better writing. Of
course machines can report the outcomes of their subroutinesaverage word-length of sentences,
or percent of words that are among the 1,500 most often appearing in print. But the relation of such
variables with writing growth is notoriously complex, nor do such variables count for much in the
typical writing teacher's repertoire of criteria relevant to student writing improvement. Of the 34
outcomes listed by the WPA (2000) for first-year writing, none, as yet, can be machine scored. And
of course even though home-computer software will now do grammar and spelling checks, neither
does so with anywhere near the accuracy of the average writing teacher (Haswell, 2005). But in most
comp teacher's criterial top ten, grammar and spelling fall at or near one. As for the computer's
programmed Pollyanna diagnostic advice, which hasn't changed much since Paulus's first effort, even
students laugh:
YOU USED 5 PARAGRAPHS WHY SO MANY Q TRY TO REORGANIZE THE ESSAY SO THAT
YOU DON'T USE SO MANY PARAGRAPHS USUALLY A PARAGRAPH SHOULD CONSIST OF
SEVERAL SENTENCES . . . . I ENJOYED READING YOUR ESSAY (Page & Paulus, 1968, p.
159)
If machines, though, could produce even part of a writing teacher's true response, truly instructional,
many writing teachers would be tempted to buy into them (Brent & Townsend, 2006). It would be a
complex love-hate relationship, and most intense with teachers conducting straight composition
courses. Response to writing lies at the essence of what they do, takes time to do well, yet takes up
so much of their time. On a weekly and sometimes daily basis, it takes away energy they want to
devote to the rest of their instructionpreparing instructive assignments, orchestrating class
Haswell 12
activities, deepening connections between writing and cultural and social ends. Content-course
faculty who assign extended writing feel the same conflict, though their critical reading habits and
sites may differ (Walvoord & McCarthy, 1991). Their way with students words is understudied, but
what we have shows production a major if sometimes tacit factor. Swilky (1991) worked with a
philosophy teacher in a WAC faculty seminar who talked about his teaching only after the seminar
was over, finally confessing to her that he had not incorporated writing-to-learn methods in his
classes because he thought they would increase his workload. Issues of response production may be
worse for English composition faculty because their response activities are often represented
publicly in a contradictory way that they find demeaning but half true: lowly comp teachers chained
at home for the weekend stooped over a stacks of student essays (Figure 2). Meanwhile in the
background they can hear the promotional literature from the software companies offering the
greatest writing-teacher shortcut of all time, computerized "diagnostic" feedback on student essays
"in minutes," "immediately," "in an instant" (Rothermel, 2006).
The Complexities of Responding to Student Writing
13
Figure 2: From Composition Chronicle: Newsletter for Writing Teachers 8 (3), April 1995, p. 11.
Representation and the Sites and Acts of Response
Responding to student papers at home over the week-end! How depressing. Indeed, where and how
response takes places makes a difference. Superficially, it may make a difference in terms of
productivity. Some teachers need their special armchair, their familiar mug of tea, and their favorite
pen before they can start commenting on a stack of paperspre-discourse rituals little different than
the ones they need before they compose papers of their own. Other teachers, such as the one I once
saw marking papers in a grocery-store checkout line, seem free of location constraints. More
profoundly, though, and certainly more complexly, teachers are never free of site and site is never
disconnected from the way teachers are represented socially and culturally.
I'm arguing that the role of the responding teacher and the setting of the act of response combine to
represent the teacher-responder in a way that may have complicated and powerful effects on
students, as well as on others. Three decades ago Stewart (1975) made this point in a provocative
essay contending that teachers should present themselves as aesthetically distanced, as people of
Haswell 14
"sensibility" who can remove themselves from the accidental qualities of a student's work and attend
to the rhetorical essences of it, and to do this it helps to read papers in the right setting. (One ideal
setting, Stewart writes bitterly, would be a home "in which you could work undisturbed by the
pressing needs of families which suffer because you are underpaid," p. 241). Stewart's image of the
teacher as a person of "sensibility" has a quaint ring to it, but the need for undisturbed space is
contemporary enough. He gives us just the first of several role/site representations with which
teachers who work with student writing have often identified themselves. With each, I invent a
characteristic scenario, though not an unrealistic one.
1. Distanced aesthetician or rhetorician (Stewart, 1975). The teacher comments on the work of a
student in another teacher's class, on a signed form to be returned to the student, as part of a
cooperative teaching project.
2. Involved co-creator (Lauer, 1989). The teacher and a student work together to write a piece,
sometimes over coffee, sometimes through e-mail exchanges.
3. Demanding coach (Holaday, 1997). A writing-center tutor advises the student on a second draft
for the student's teacher; or the teacher helps prepare students for a department exit portfolio
reading by discussing departmental standards and evaluation habits of the other teacher-
evaluators.
4. Persuasive motivator (Johnston, 1982). In a workshop session the teacher provides a set of
questions designed to prompt students into further ideas on a draft, handing the heuristic out to
small groups of students. Each group is collaborating on a piece. The teacher then visits groups,
helping solve problems.
5. Experienced modeler (Young, 1997). In class the teacher projects a passage from a student essay
on screen and re-writes it; or in an end-comment to a paper requests that the writer read a
particular essay for ideas before starting the second draft.
6. Prompting dialoguer (Lauer, 1989). In the teacher's office the teacher meets with a group of
two or three students, who have no more than pre-writing notes, and the teacher encourages
questioning and inquiry.
7. Judicious lawgiver (Murray, 1982b). When assigning a new essay, the teacher makes available a
student-written essay that fulfills the assignment at the C level, and explains why it would earn
that grade.
8. Supportive parent (Turbill, 1994). The teacher privately praises the student's lab report after
class in the hallway.
9. Expert reader (White, 1984). In class the teacher shows how a reading assignment, a
professionally written piece, achieves rhetorical effects germane to the students' next writing
assignment.
10. Sharp-eyed editor (Dohrer, 1991). During a workshopping session in a computer classroom, the
teacher vets, one last time, each student's letter to the editor, just before it is e-mailed off.
11. Experienced diagnostician (Krest, 1988). Students know that their paper will be read and
commented on by other students in the class, during a "read-around." The students record their
comment on one margin, the teacher on the other margin, and then the teacher and student
conference, one on one, over the comments.
12. Real reader (Maimon, 1979). To the class the teacher reads unrehearsed through a student's
first draft, thinking aloud, or more exactly responding aloud, all along the way.
The Complexities of Responding to Student Writing
15
Note that the last four roles portray the teacher basically as a reader. On the one hand, the act of
reading is so private that it hardly seems a public role at all. On the other hand, there is no discursive
practice that more distinguishes the activity of teaching writing within a campus faculty. Yet
readingthe necessary first step in response to writingis no less complex than the commentary
that follows. Functionally, we read in many different ways, and each way has its public face. Think
again about written response on a student's paper. The commentary is the teacher's representation,
and different kinds of remarks will represent the teacher in different roles. The teacher can read
setting the student's language against usage, striking out solecisms and replacing them with
conventional expressions, in which case the teacher will appear as grammarian or guardian of the
language. Or the teacher can read with an metadiscursive eye to effects the writing will have on other
readers, circle whole passages and write "Deleteeveryone knows this!" or "Especially persuasive"
in the margin, in which case the teacher will appear as experienced author or editor-in-chief. Or the
teacher can read with a memory of the student's previous writing, start an end-note with the
comment, "Finallya breakthrough," in which case the teacher will appear in the role of caring
physician or observant coach. And if teachers leave no mark except for a grade (as in Stiff's 1967
experiment), students may image them as distanced and uncaring dispensers of the law.
This dramatis personae of teacherly roles may sound as though teachers, when they act as
responders to student writing, are just so many authors in search of a character. But one represents
one's self willy-nilly, and a teacher can and does switch to other roles in the blink of an eye, perhaps
to that of content expert setting some subject matter forth to students as the scholar-of-the-lecture-
hall or the director-of-the-research-lab. The point is that writing-responder roles alone are subtle,
unstated, multiplex, and numerous, and they can't but mediate the teacher's commentary to the
student. In interviews with students, O'Neill & Fife (1999) found that they constantly interpreted
teacher comments through their construction of teacher personalities and roles: "‘She's more like
friend than a teacher'" (41). Gordon Harvey (2003), in a piece published after I had compiled the
above list of responder roles, offers his own list, including "introducer" (to academic ways), "midwife-
therapist," "representative of an institution," "one's actual self," and other roles not identical to ones
on my list. His point is that choosing and maintaining these roles adds to the stressful job of
responding to student writing, a job finally that is "impossible" given the numbers of students and
papers, a job that leads to teacher discouragement and burnout. That is not my conclusion, but it does
challenge the hope that out of this problematic multiplicity of role, task, and site representations
effective and sustainable methods of response can be extracted.
Identity and Resistance, the Major Roadblock to Response
This is not to say, of course, that both writing students and writing teachers can't resist these
representations. They often do. Resistance, in fact, may be a constant factor in instructional response
to writing. And it is more than the obvious fact that psychologically students often resist teachers'
advice, or teachers often resist commenting on papers (who hasn't procrastinated on that job?).
Resistance is part of the human response to representation itself. How else can a person escape the
suffocating anonymity of public representations except by resisting them? How else respond to
representation except by developing a counter-representation that images oneself, even if only in
one's head, as somehow different, somehow not representative? That activating image of oneself (or
one's self), identity, brings our circuit of cultural response practices back around to regulation, for
identity can also be defined as the image we use to regulate our own practices in resistance to the
cultural practices foisted upon us.
However all this may be, resistance inheres in response to writing (Paine, 1999; Chase, 1988).
Students are reluctant to change their rhetorically inept ways because their old ways have stood
Haswell 16
them academically in sufficient stead (second-language specialists call this "error resistance"), and
because they don't like the image of themselves as inepts. Room for disbelief is especially large in
straight composition classes, and an argument for WID. It's easier for the student to shrug off the
comment "These last two arguments are far too simplistic to be convincing" as just the English
teacher's opinion than the history teacher's comment "You've listed only two of the five main political
outcomes of the Great Depression." Students resist the marginal question because it's not an answer
(Smith, 1989), the corrective commentary because it's depressing (Reed & Burton, 1986), the pithy
teacher-efficient mark because it lacks explanation (Land & Evans, 1987), the lengthy end-comment
because it seems too critical and generalized (Hillocks, 1982), the recommendation for improvement
too removed from their current habits because it seems too risky (Onore, 1989), the critique from
peers because it comes from novices (Odell, 1989), the critique from teachers because it comes from
an instructor they sense holds a marginal or tenuous academic position (Whichard, et al., 1992). They
resist not only for such questionable reasons but also for a good one, because they are more and more
seeing themselves as independent and ultimately free to use language as they wish. They resist
because they are students. As William Perry put it in a very early piece on resistance and response
(two decades before publishing the book for which he is famous, Forms of Intellectual and Ethical
Development), "the young will resist even your efforts to reduce their resistance" (1953, p. 459).
That doesn't make response any easier for teachers. The self-contradictions in the above panoply of
student resistance just encourages teachers to identify themselves, among other things, as skeptics.
To student papers teachers have to bring their own resistance. Perhaps the least defensible reason
for it is the assumption that student writing is abnormal, apprenticeship work, and therefore worthy
of skepticism, at least a kind of skepticism teachers don't bring to real-world writing (Lawson & Ryan,
1989). The most defensible reason for their resistance is the thought that the effort they are about to
spend as a teacher-responder may well not be worth the effect it will have on the writer. It is a sad
reason that either begs for response shortcuts or pleads to bag the whole operation.
Some Directions to Shortcuts
We are back, full circuit, to the response shortcuts I mentioned at the beginning: checksheets,
correction symbols, mnemonic acronyms, weighted criteria scales, peer-critique handouts, computer
feedback software, and many more. Do they really produce more effect with less effort? Can they be
revised to do so? As I said, my intention with this paper is not to offer new and better ones, an
enterprise which would require a paper at least as long the current one, but to suggest some
directions for review, revamp, or creation. The activity nodes in Du Gay's "circuit of culture" will help
map these suggestions.
Regulation Research clearly documents that for students writing criteria, genre and mode rules,
disciplinary styles, and standards are heavily mediated through the particular academic course. It is
a mediation that changes the critical terms and adds immeasurably to their multiplicity, in practically
every classroom they enter. Teacher feedback can largely be a waste of time, for both teacher and
student, unless the critical language is grounded in the specific rhetoric of the field under study. The
caution applies to every kind of shortcut, assignment desiderata ("Be sure your paper is well
organized"), marginal comments ("Be less wordy here"), and analytic scales ("Content: 15 out of 20
points"). Explaining discipline-specific terminology sounds like more work for teachers, but a
handout early in the course will save much time later. For genre traits, teachers can provide a model
paper, with important aspects marked and explained. Selection is another route to pursue, although
it goes against the cover-all philosophy of many academic responders. Teachers can choose a few of
the most important and most accessible characteristics of disciplinary style and prepare a sheet
explaining and illustrating them. Let the rest go. The same principles of showing and selecting apply
The Complexities of Responding to Student Writing
17
to paper commentary. If a paper is wordy, the teacher can rewrite one sentence, showing how it is
done, and ask the student to redo the entire paper. The redo option also makes sense in dealing with
standards. Student can be told that papers below a certain level of writing will just be handed back
without marks, to be resubmitted within one week. Setting exact critical levels ("More than five
misspellings and the paper will receive a redo") seems a denial of the contextuality of error gravity,
although fair warning of errors that score high on the teacher's irritation scale will save everybody
time. For years my practice has been to stop reading and marking as soon as I sense a paper that is
below a reasonable level, and to draw a line across the page and write "Redo" at that point. Papers
usually come back substantially improved with no more effort on my part.
Consumption clearly is heavily interdependent on regulation and production. The communication
gap between teacher response and student understanding of it may be the most difficult of all these
interactions to shortcut. But avenues toward efficiency are there. Attention to students and how they
are receiving methods of teacher response requires extra time at first but may save time in the long
run. Illustrative is a method developed by Hayes, Simpson, & Stahl (1994) by which students on their
own can prepare for essay exams in content courses. An acronym helps make the system accessible
for students, PORPE: Predict potential essay questions, Organize key ideas to answers using your
own terms and structures, Rehearse key ideas, Practice by writing trial essays, Evaluate the essays in
terms of "completeness, accuracy, and appropriateness." The system certainly saves the teacher time
since all the work is done by the student and performance on the actual examination essays may need
less teacher response. It is a good example of a universal labor-cutting technique more teachers could
avail themselves of, in which shrewd preparation of students for a writing assignment saves the
teacher time in responding to it. But if the teacher does not keep an eye on the process, it can well
turn out counter-productive. With PORPE, how good are the students' self-chosen terms and
structures? How well can they judge their own writing, especially with critical language as discipline-
tailored as "accuracy" and "appropriateness"? In regards the communication channel between
teacher and student, in the long run the best shortcut is to make sure critical terms and response
methods are interpretable, whether they comprise test preparation, assignment direction,
composing heuristic, self- and peer-evaluation, to-do checklist, paper commentary, grade
breakdown. Even the most obvious term may be misinterpreted. Leon Coburn tells a story of the time
he kept writing "cliché" in the margin while the student kept producing more and more clichés.
Finally Coburn asked what was going on, and the puzzled student responded, "‘I thought you kept
marking them because you liked them'" (Haswell & Lu, 2000, 40). Amusing, but the story assumes a
parable-like shape when we consider that students often prefer clichés and frequently don't even
recognize them as such (Olson, 1982).
Production If regulation lies at the center of the purpose of academic writing response, production
lies at the center of its method. Consequently production prompts and guides most teacher shortcuts.
To the formal ones listed above (praise only, peer-evaluation, self-evaluation, audiotaping of
comments, etc.), every teacher has added their own: assign fewer papers, respond only to the worst
papers and OK the others, comment on only one aspect of the paper (as the assignment forewarns),
assign only short papers that demand concise writing, talk with the student over a batch or portfolio
of papers in an office conference, comment only if the student asks for comments, read papers only
if a student brings one in complaining about the commentary or the grade given by a teaching
assistantan endless list. Some are once-only, spur of the moment, and ethically questionable, like
the time a graduate friend of mine, a teaching assistant in political science, told his class he had gone
hunting and forgotten that he had put their papers in the trunk of his car and he had to throw them
away because they got covered with quail blood; or the time an English teacher (who understandably
prefers to remain anonymous) got desperate for time, gave a set of papers unread back to her
Haswell 18
students, and told them they were the worst essays she had ever read and that they were to re-write
them, replacing the next assigned paper (Haswell & Lu, 2000, 51). The teacher said that the re-written
papers turned out to be the best she received all semester, but if other teachers now seize what
appears to be an amazing shortcutmuch student improvement, zero teacher timeagain I
recommend they appraise the results carefully. A major hazard of production shortcuts is that they
are constructed almost solely for the teacher and usually judged solely by the teacher. As with any
workplace production, the situations begs for quality control. The ultimate issue is how well the
methods work, gain of student learning set against gain of teacher time. As in the marketplace,
popularity of a product is no guarantee of its worth, and research into writing response questions
some of the most widespread of production shortcuts. Peer-evaluation, for instance, can generate
superficial revisions (Dochy, Segers, & Sluijsmans, 1999), and automated grammar-checkers and
non-automated English teachers identify the same mistakes only 10% of the time (Brock, 1995). On
the other hand, labor-intensive response methods do not necessarily equate with learning gains
witness the hours teachers have devoted to time-consuming correction of every mistake in student
papers with, it seems, only minor influence on subsequent student writing.
Representation Does the image itself of a teacher swotting over student writings help improve
student writing? For students does that image serve as a model, a goad, a caution, a laughing stock?
As activity theory posits, based upon the way teachers present themselves students will construe
them and act upon that construction. The advice then is to know that, among the many roles of
responder to student writing (distanced aesthetician, involved co-creator, demanding coach, etc.)
there is a choice and some roles are more efficient and efficacious than others. With a doable class
size of twenty-five students, occasionally I have had students bring their newly written paper to me
for a ten-minute office conference. I simply start reading the paper, voicing my reactions as I go. The
student can see me read, rather than operate from some fanciful image of me as a reader. I find things
I like, wonder aloud where the piece is going, hesitate over unclear sentences, and discover what
qualities of writing the student should be working on next. Usually I don't have time to finish the
paper, but students rarely mind, and I imagine I get more useful critique accomplished in ten minutes
than in the twenty minutes it normally takes me to comment in writing on a five-page student essay.
That, I think, also leaves the student with an image of me as a careful and engaged reader, which in
turn encourages them to write better papers for me later. At least students have praised the method
in their course evaluations.
Representations can go wrong, of course. Take my commenting technique of minimal marking
(Haswell, 1983, though not entirely mine, since earlier versions go back at least another quarter
century, e.g., Halvorson, 1940). There are two steps to it. The first has the teacher put only an X in the
margin next to a line with any surface mistake in it. Students have to discover the mistake and correct
it on their own, which places the task of learning where it belongs, on the student. The second step is
for students to have their emendations vetted by the teacher. Although they catch around two-thirds
of their original errors, often (39% of the time in my first-year writing classes) they correct mistakes
incorrectly, correct non-mistakes, or simply can't find the error. Unfortunately the second step
subtracts some of the time teachers gain with the first step, and teachers I have known have omitted
ita major mistake, I think, not only because important student learning takes place in the second
step but also because neglecting it represents the teacher as lazy, not caring, perhaps a little cruel.
Identity / resistance. Or unprofessional. Once a colleague about twenty years senior came into my
office, beefy red in the face, thrusting a student paper at me. It turned out to be one a student of mine
had written and I had responded to. Among my marginal comments were X's of my minimal marking
system. The student had not yet tried to find and correct the errors, or I obviously to vet her
corrections, but that made no difference to my colleague. He told me I had abjured my duty as an
The Complexities of Responding to Student Writing
19
English teacher to correct student writing and that I had better correct my ways. Identity is the
representation one constructs of oneself, in part out of the representations others make of oneself.
Professionally, it is the self one has to live with, on and in the job. In this case I chose to resist my
colleague's notion of myself, and carried on with a response technique I found cheap, effective, and
satisfying. When it comes to one's identity as a responder to student writing, the research offers little
direction, except the self-evident one to avoid self-destructive identitiesMiss Grundy, never
satisfied, or Bartleby the Scrivener, burnt out to the point of inaction. I suppose the main caution is
how easily teachers can forget that the end is students learning and that if productive response to
their writing, despite all the shortcuts we can contrive, still is laborious, well, that is what we are paid
to do.
There is one hard-earned identity, however, that some teachers build for themselves and that I can
especially recommend. It is based on an understanding of the very complexity of response I have
been describing, and on the pragmatic knowledge of response practices that are both sustainable and
functional, sustainable because they are efficient for the teacher and functional because they are lucid
and focused enough for students to put into use. This responder identity might be described as craft
wise. Like experienced fixers in all trades, craft wise responders have learned to home in on the main
problem and attend to it with as little effort as possible.
Their wisdom boils down to eschewing the traditional cover-all-bases approach to writing response
and adopting a smaller task-specific, problem-specific, and learner-specific method. The techniques
and advice have been around for a long time. Know your student and limit your response, especially
your advice for further work, to no more than one or two things your student can do. It is response
easy in the making but not easy to make. It requires sifting through the options and contradictions of
regulation, navigating the constraints of consumption, weighing the cost of production, critiquing the
masks of representation, and working with and not against student and teacher identities. It is no
surprise that the advice often emerges from faculty who have intensively studied student response
to teacher response: Diogenes, Roen, and Moneyhun (1986) and their work with English instructors;
MacAllister (1982), Mallonee and Breihan (1985), and Lunsford (1997) and their work with campus
faculty; Zak (1990) and her conversations with students over her own comments; above all Harris
(1979) and her years of experience conferencing with writing-center students of other teachers. Long
of birth, the advice emerges from a teacher-identity again born of resistancebut in this case a
resistance to the stubbornly rooted professional position of full correction. Yet it is advice that will
meet another resistance, the student's, whose own development so often prefers the safe and known
and is therefore reluctant to pull the one switch that might further that development. Clear in its
directions, easy to consume for the student, easy to produce for the teacher, this particular response
practice seems to have everything going for it. But it doesn't come with guarantees.
The Ever-Frustrating, Perpetually Fascinating Issue of Writing
Response
With instructional response to writing, it seems, skepticism rarely fails to play a part. The poet
William Stafford once told me (it was, in fact, on that lonely, twisty highway between Lewiston, Idaho
and Pullman, Washington) that his job as a teacher responding to student writing often consisted of
no more than "leaving tracks"a trail of marginal dots and checks and squiggleswhich at least
showed the writers that he had cared enough to read and comprehend. Maybe he was right, and the
final worth of writing-teacher commentary is only a kind of passing the torch, keeping the students
assured that they will always have words people will read. But skepticism itself can't last. Don't we
all remember a teacher's comment on our writingvoiced or inscribedthat pulled some sort of
Haswell 20
switch, that rankled forever or precipitated a critical moment changing our writing forever? So
teachers carry on. As regards truly instructional response, it seems teachers will find themselves
nowhere else but on the road of excess, always making the circuit from certainty to complexity to
skepticism and round again.
Wherever they are on that road, however, there is always a via media. Complexity may tempt the
teacher to self-serving simplicity, but it is worth remembering that in an important sense every act
of evaluation of student writing involves simplification. Grades (plus and minus) sort into no more
than twelve categories, "Good, clear examples" expresses the comprehension of only one reader,
"Your ending is flat" omits mention of scores of other writing accomplishments and traits. With the
peculiar and situation-specific act of evaluation we call "response," simplification, of course, can
serve for good or bad. I've never forgotten my sense of indignation when at the end of a sophomore
history course I picked up four assigned book reviews I had written and discovered only a at the
top of each first page. Nor have I forgotten the look on my student's facehe was in the first class of
sophomores I ever taughtand the loss-of-face sound in his voice when he picked up his paper, saw
my fulsome comments, and said, "You wrote more than I did." With response to student writing, it's
uplifting to know that less can sometimes be better and that simplification is unavoidable. It is also
enlightening to know that shortcuts can turn into culs-de-sac. In the final analysis, the best route for
teachers is to keep checking the response to their response and, of what is workable, to go with what
they know works.
A Bibliographic Post Scriptum
What researchers know works in instructional response is considerable, much more than I've been
able to cover. On evaluation of writing within disciplines other than composition, there are three fine
collections of essays, Yancey & Huot (1997), Sorcinelli & Elbow (1997), and Strenski (1986). Anson,
Schwiebert & Williamson (1993) is an extensive annotated bibliography of writing across the
curriculum, where practical writing-response techniques can be found devised by teachers in
accounting, biology, business, computer science, economics, engineering, history, literature,
mathematics, nursing, philosophy, psychology, and elsewhere. Speck (1998), an invaluable
bibliography of writing evaluation, has over 1,000 entries, solidly annotated. Recently Underwood
and Tregidgo (2006) review 21 research articles, 14 of them at the college level, looking much as I
have for what is effective. The research database is much more extensive than they suggest, however.
The online index of post-secondary scholarship into writing, CompPile (Haswell & Blalock), good for
1939-1999, currently locates around 420 data-gathering studies involving teacher response.
References
Anson, Chris M. (1989). Response styles and ways of knowing. In Anson, Chris M. (Ed.), Writing and response:
Theory, practice, and research (pp. 332-366). Urbana, IL: National Council of Teachers of English.
Anson, Chris M. (Ed.). Writing and response: Theory, practice, and research. Urbana, IL: National Council of
Teachers of English.
Anson, Chris M., Schwiebert, John E., & Williamson, Michael M. (1993). Writing across the curriculum: An
annotated bibliography. Westport, CN: Greenwood Press.
Bazerman, Charles. (1989). Reading student texts: Proteus grabbing Proteus. In Lawson, Bruce, Ryan, Susan
S., & Winterowd, W. Ross (Eds.), Encountering student texts: Interpretive issues in reading student
writing (pp. 139-146). Urbana, IL: National Council of Teachers of English.
Beason, Larry. (1993). Feedback and revision in writing across the curriculum classes. Research in the
Teaching of English, 27(4), 395-422.
The Complexities of Responding to Student Writing
21
Brent, Edward, and Townsend, Martha. (2006). Automated essay-grading in the sociology classroom: Finding
common ground. In Ericsson, Patricia Freitag, & Haswell, Richard (Eds.), Machine scoring of student
essays: Truth and consequences (pp. 117-198). Logan, UT: Utah State University Press.
Broad, Broad. (2003). What we really value: Beyond rubrics in teaching and assessing writing. Logan, UT: Utah
State University Press.
Brock, Mark N. (1995). Computerized text analysis: Roots and research. Computer Assisted Language
Learning, 8(2-3), 227-258.
Chase, Geoffrey. (1988). Accommodation, resistance and the politics of student writing. College Composition
and Communication, 39(1), 13-22.
Connors, Robert J., & Lunsford, Andrea A. (1988). Frequency of formal errors in current college writing, or Ma
and Pa Kettle do research. College Composition and Communication, 39(4), 395-409.
Cross, Geoffrey, and Wills, Katherine. (2001). Using Bloom to bridge the WAC/WID divide. ERIC Document
Reproduction Service, ED 464 337.
Denman, Mary E. (1975). I got this here hang-up: Non-cognitive processes for facilitating writing. College
Composition and Communication, 26(3), 305-309.
Diogenes, Marvin, Roen, Duane H., & Moneyhun, Clyde. (1986). Transactional evaluation: The right question
at the right time. Journal of Teaching Writing, 5(1), 59-70.
Dochy, Filip, Segers, Mien, & Shuijsmans, Dominique. (1999). The use of self-, peer and co-assessment in
higher education: A review. Studies in Higher Education, 24(3), 331-350.
Dohrer, Gary. (1991). Do teachers' comments on students' papers help?. College Teaching, 39(2), 48-54.
Du Gay, Paul (Ed.). (1997). Doing cultural studies: The story of the Sony Walkman. London: Sages; The Open
University.
Dulek, Ron, & Shelby, Annette. (1981). Varying evaluative criteria: A factor in differential grading. Journal of
Business Communiation, 181(2), 41-50.
Dusel, William J. (1955). Determining an efficient teaching load in English. Illinois English Bulletin (October),
1-19.
Earthman, Elise A. (1992). Creating the virtual work: Readers' processes in understanding literary
texts. Research in the Teaching of English, 26(4), 351-384.
Eblen, Charlene. (1983). Writing across-the-curriculum: A survey of a university faculty's views and
classroom practices. Research in the Teaching of Writing, 17(4), 343-348.
Elbow, Peter. (1993). Ranking, evaluating, and liking: Sorting out three forms of judging. College English,
55(2), 187-206.
Evans, Kathryn. (2003). Accounting for conflicting mental models of communication in student-teacher
interaction: An activity theory analysis. In Charles Bazerman & David R. Russell (Eds.), Writing selves /
writing societies: Research from activity perspectives. Fort Collins, CO: WAC Clearinghouse,
https://wac.colostate.edu/books/selves_societies/.
Evans, Kathryn A. (1997). Writing, response, and contexts of production or, why it just wouldn't work to write
about those bratty kids. Language and Learning Across the Disciplines, 2(1), 5-20.
Fearing, Bertie E. (1980). When it's one down and 1,249 themes to go, it's time to establish a realistic teaching
load for English teachers. ADE Bulletin (no. 63), 26-32.
Ferris, Dana R. (1995). Student reactions to teacher response in multiple-draft composition
classrooms. TESOL Quarterly, 29(1), 33-53.
Freyman, Leonard. (1965). Appendix: Marking of 250-word themes. In Arno Jewett & Charles E. Bish (Eds.),
Improving English composition: NEA-Dean Langmuir Project on Improving English Composition (p. 99).
Washington, D. C.: National Education Association.
Gere, Anne Ruggles. (1977). Writing and writing. English Journal, 66(8), 60-64.
Gottschalk, Katherine K. (2003). The ecology of response to student essays. ADE Bulletin (no. 134-135), 49-56.
Grabe, William, & Kaplan, Robert B. (1996). Theory and practice of writing: An applied linguistic
perspective. London; New York: Longman.
Haswell 22
Hairston, Maxine. (1981). Not all errors are created equal: Nonacademic readers in the professions respond
to lapses in usage. College English, 43(8), 794-806.
Halpern, Sheldon, Spreitzer, Elmer, & Givens, Stuart. (1978). Who can evaluate writing? College Composition
and Communication, 29(4), 396-397.
Halvorson, Nelius O. (1940). Two methods of indicating errors in themes. College English, 2(3), 277-279.
Harris, Muriel. (1979). The overgraded paper: Another case of more is less. In Stanford, G. (Ed.), How to
handle the paper load: Classroom practices in teaching English 1979-1980 (pp. 91-94). Urbana, IL:
National Council of Teachers of English.
Harris, Winifred H. (1977). Teacher response to student writing: A study of the response patterns of high
school English teachers to determine the basis for teacher judgment of student writing. Research in the
Teaching of English, 11(2), 175-185.
Harvey, Gordon. (2003). Repetitive strain: The injuries of respnding to student writing. ADE Bulletin, Nos.
134/135, 43-48.
Haswell, Richard H. (2005). Automated text-checkers: A chronology and a bibliography of
commentary. Computers and Composition Online. http://www.bgsu.edu/cconline/haswell/haswell.htm.
Haswell, Richard H. (2006). Automatons and automated scoring: Drudges, black boxes, and dei ex machina. In
Ericsson, Patricia Freitag, & Haswell, Richard H. (Eds.). Machine scoring of student essays: Truth and
consequences (pp. 57-78). Logan, UT: Utah State University Press.
Haswell, Richard H. (1983). Minimal marking. College English, 45(6), 600-604.
Haswell, Richard H., & Blalock, Glenn. (2006). CompPile: An ongoing inventory of publications in post-
secondary composition, rhetoric, ESL, technical writing, and discourse studies.
http://comppile.tamucc.edu./
Haswell, Richard H., & Lu, Min-Zhan. (2000). Comp tales: An introduction to college composition throught its
stories. New York: Longman.
Hawhee, Debra. (1999). Composition history and the Harbrace College Handbook. College Composition and
Communication, 50(3), 504-523.
Hayes, Christopher G., Simpson, Michele L., & Stahl, Norman A. (1994). The effects of extended writing on
students' understanding of content-area concepts. Research and Teaching in Developmental Education,
10(2), 13-34.
Hayes, Mary F., & Daiker, Donald A. (1984). Using protocol analysis in evaluating responses to student
writing. Freshman English News, 13(2), 1-4, 10.
Hillocks, George, Jr. (1982). Getting the most out of time spent marking compositions. English Journal 71(6),
80-83.
Hobson, Eric, & Lerner, Neal. (1999). Writing centers / WAC in pharmacy education: A changing prescription.
In Barnett, Robert W., & Blummer, Jacob W. (Eds.). Writing centers and writing across the curriculum
programs: Building interdisciplinary partnerships (pp. 155-176). Westport, CT: Greenwood Press.
Holaday, Lynn. (1997). Writing students need coaches, not judges. In Tchudi, Stephen (Ed.), Alternatives to
grading student writing (pp. 35-46). Urbana, IL: National Council of Teachers of English.
Hughes, Arthur, & Lascaratou, Chryssoula. (1982). Competing criteria for error gravity. ELT Journal, 36(3),
175-182.
Hull, Glynda, & Rose, Mike. (1990). 'This wooden shack place': The logic of an unconventional reading'. College
Composition and Communication, 41(3), 287-298.
Janopoulos, Mihael. (1992). University faculty tolerance of NS and NNS writing errors: A comparison. Journal
of Second Language Writing, 1(2), 109-121.
Jason, Gary. (1988). Hedging as a fallacy of language. Informal Logic: Reasoning and Argumentation in Theory
and Practice, 10(3), 169-75.
Johnston, Brian. (1982). Non-judgmental responses to students' writing. English Journal, 71(4), 50-53.
Kantz, Margaret, & Yates, Robert. (1994). Whose judgments? A survey of faculty responses to common and
highly irritating writing errors. In National Council of Teachers of English Assembly for the Teaching of
The Complexities of Responding to Student Writing
23
English Grammar (Ed.). Partial Proceedings of the 5th Annual Connference of the Assembly for the
Teaching of English Grammar. ERIC Document Reproduction Service, ED 397 435.
Kline, Charles R., Jr. (1976). I know you think you know what I said. College English, 37(7), 661-662.
Knoblauch, C. H., & Brannon, Lil. (1981). Teacher commentary on student writing: The state of the
art. Freshman English News, 10(2), 1-4.
Krest, Margie. (1988). Monitoring student writing: How not to avoid the draft. Journal of Teaching
Writing, 7(1), 27-39.
Kucer, Stephen B. (1989). Reading a text: Does the author make a difference?. In Lawson, Bruce, Ryan, Susan
S., & Winterowd, W. Ross (Eds.), Encountering student texts: Interpretive issues in reading student
writing (pp. 159-170). Urbana, IL: National Council of Teachers of English.
Land, Robert E., Jr., & Evans, Sandra (1987). What our students taught us about paper marking. English
Journal, 76(2), 33-38.
Lauer, Janice M. (1989). Interpreting student writing. In Lawson, Bruce, Ryan, Susan S., & Winterowd, W. Ross
(Eds.), Encountering student texts: Interpretive issues in reading student writing (pp. 121-130). Urbana,
IL: National Council of Teachers of English.
Lawson, Bruce, Ryan, Susan S., & Winterowd, W. Ross (Eds.). Encountering student texts: Interpretive issues in
reading student writing. Urbana, IL: National Council of Teachers of English.
Lawson, Bruce, & Sterr, Susan R. (1989). Introduction. In Lawson, Bruce, Ryan, Susan S., & Winterowd, W.
Ross (Eds.), Encountering student texts: Interpretive issues in reading student writing (pp. vii-xvii).
Urbana, IL: National Council of Teachers of English.
Lucas, Catherine. (1988). Toward ecological evaluation. Quarterly of the National Writing Project and the
Center for the Study of Writing, 10 (1), 1-3, 12-14 [Part I]; 10 (2), 4-10 [Part II]
Lunsford, Ronald F. (1997). When less is more: Principles for responding in the disciplines. In Sorcinelli, Mary
D., Elbow, Peter (Eds.), Writing to learn: Strategies for assigning and responding to writing across the
disciplines (New directions for teaching the learning, No. 69) (pp. 91-104). San Francisco, CA: Jossey-
Bass.
Luukka, Minna-Riitta, & Markkanen, Raija. (1997). Impersonalization as a form of hedging. In Markkanen, R, &
Schröder, Hartmut (Eds.), Hedging and discourse: Approaches to the analysis of a pragmatic
phenomenon in academic texts (pp. 168-187). New York: Walter de Gruyter.
MacAllister, Joyce. (1982). Responding to student writing. In Griffin, C. W. (Ed.), Teaching writing in all
disciplines (New Directions for Teaching and Learning No. 12) (pp. 59-66). San Francisco, CA: Jossey-
Bass Inc., Publishers.
Maimon, Elaine P. (1979). Talking to strangers. College Composition and Communication, 30(4), 364-369.
Mallonee, Barbara, & Breihan, John R. (1985). Responding to students' drafts: Interdisciplinary
consensus. College Composition and Communication, 36(2), 213-231.
Martin, J. R., & Rothery, Joan. (1986). What a functional approach to the writing task can show teachers about
'good writing'. In Couture, Barbara (Ed.), Functional approaches to writing research perspectives (pp.
241-262). Norwood, NJ: Ablex.
Marzano, Robert J., & Arthur, Sandra. (1977). Teacher comments on students essays: It doesn't matter what you
say. ERIC Document Reproduction Service, ED 147 864.
McCarthy, Lucille Parkinson. (1987). A stranger in strange lands: A college student writing across the
curriculum. Research in the Teaching of English, 21(3), 233-265.
McCauley, Lynne, tenHarmsel, Larry, & VanderKogi, Lambert. (1993). Writing at Western Michigan
University: A study of faculty and student perceptions. ERIC Document Reproduction Service, ED 363
152.
Means, Barbara, Sonnenschein, Susan, & Baker, Linda. (1980). Organization and content in writing in
speech. ERIC Document Reproduction Service, ED 208 403.
Haswell 24
Mellon, John C. (1976). A taxonomy of compositional competencies. In Beach, Richard, & Pearson, P. David
(Eds.), Perspectives on literacy: Proceedings on the 1977 Perspectives on Literacy Conference (pp. 247-
272). Minneapolis, MN: University of Minnesota. College of Education.
Meyer, Paul G. (1997). Hedging strategies in written academic discourse: Strengthening the argument by
weakening the claim. In Markkanen, Raija, & Schroder, Hartmut (Eds.), Hedging and discourse:
Approaches to the analysis of a pragmatic phenomenon in academic texts (pp. 21-41). New York: Walter
de Gruyter.
Michael, William B., Cooper, Terri, Shaffer, Phyllis, & Wallis, Earl. (1980). A comparison of the reliability and
vaidity of ratings of student performance on essay examinations by professors of English and by
professors in other disciplines. Educational and Psychological Meaurement, 40(1), 183-195.
Modern Language Association of America, & National Council of Teachers of English. (1913). Report on the
cost and labor of English teaching. Lawrence, KS: Modern Language Association.
Murray, Donald M. (1982a). The qualities of good writing. In Murray, D. M., Learning by teaching: Selected
articles on writing and teaching (pp. 66-67). Montclair, NJ: Boyton/Cook.
Murray, Donald M. (1982b). What can you say besides AWK. In Murray, Donald M., Learning by teaching:
Selected articles on writing and teaching (pp. 151-156). Montclair, NJ: Boyton/Cook.
National Council of Teachers of English. (2002). More than a number: Why class size
matters. http://www.ncte.org/positions/more_than_a_number.shtml.
Newkirk, Thomas.(1984). How students read student papers: An exploratory study. Written
Communication, 1(3), 283-305.
Ober, Scot, Zhao, Jensen J., David, Rod, & Alexander, Melody. (1999). Telling it like it is: The use of certainty in
public business discourse. Journal of Business Communication, 36(3), 280-300.
Odell, Lee. (1989). Responding to responses: Good news, bad news, and unanswered questions. In Lawson,
Bruce, Ryan, Susan S., & Winterowd, W. Ross (Eds.), Encountering student texts: Interpretive issues in
reading student writing (pp. 221-234). Urbana, IL: National Council of Teachers of English.
Olson, Gary A. (1982). Clichés: Error recognition or subjective reality?. College English, 44(2), 190-194.
O'Neill, Peggy, & Fife, Jane M. (1999). Listening to students: Contextualizing response to student
writing. Composition Studies, 27(2), 39-51.
Onore, Cynthia. (1989). The student, the teacher, and the text: Negotiating meanings through response and
revision. In Anson, Chris M. (Ed.), Writing and response: Theory, practice, and research (pp. 231-260).
Urbana, IL: National Council of Teachers of English.
Page, Ellis B., & Paulus, Dieter. (1968). The analysis of essays by computer (Final report, Project No. 6-
1318). Washington, D. C.: Department of Health, Education, and Welfare; Office of Education; Bureau of
Research.
Paine, Charles. (1999). The resistant writer: Rhetoric as immunity, 1850 to the present. Albany, NY: State
University of New York Press.
Perry, William G., Jr. (1953). The 600-word theme and human dignity. College English, 14(8), 454-460.
Purves, Alan C., & Takala, Sauli (Eds.). (1982). An international perspective on the evaluation of written
composition. New York: Pergamon Press.
Reed, W. Michael, & Burton, John K. (1986). Effective and ineffective evaluation of essays: Perceptions of
college freshmen. Journal of Teaching Writing, 4(2), 270-283.
Rothermel, Beth Ann. (2006). Automated writing instruction: Computer-assisted or computer-driven
pedagogies? In Ericsson, Patricia Freitag, & Haswell, Richard H. (Eds.), Machine scoring of student essays:
Truth and consequences (pp. 199-210). Logan, UT: Utah State University Press.
Russell, David. (1995). Activity theory and its implications for writing instruction. In Thomas Kent (Ed.), Post-
process theory: Beyond the writing-process paradigm (pp. 80-95). Carbondale, IL: Southern Illinois
University Press.
Sherwin, J. Stephen. (1969). Four problems in teaching English: A critique of research. Scranton, PA:
International Textbook Co.; National Council of Teachers of English.
The Complexities of Responding to Student Writing
25
Slotnick, Henry B., & Knapp, John V. (1971). Essay grading by computer: A laboratory phenomenon?. English
Journal, 60(1), 75-81.
Smith, Ernest. (1989). "It doesn't bother me, but sometimes it's discouraging": Students respond to teachers'
written responses. Journal of Teaching Writing,8(1), 255-266.
Smith, Summer. (2003). What is "good" technical communication? A comparison of the standards of writing
and engineering students. Technical Communication Quarterly, 12(1), 7-24.
Sorcinelli, Mary D., & Elbow, Peter (Eds.). (1997). Writing to learn: Strategies for assigning and responding to
writing across the disciplines. San Francisco, CA: Jossey-Bass.
Spear, Karen I. (1988). Sharing writing: Peer response groups in English classes. Portsmouth, NH:
Boynton/Cook Publishers.
Speck, Bruce W. (1998). Grading student writing: An annotated bibliography. Westport, CT: Greenwood Press.
Spooner-Smith, Laura. (1976). Applying empirical needs assessment procedures to university-level
instruction in composition. ERIC Document Reproduction Service, ED 122 294.
Stanford, Gene. (Ed.). (1979). How to handle the paper load. Classroom practices in teaching English [vol.
17]. Urbana IL: National Council of Teachers of English.
Stewart, Donald C. (1975). Aesthetic distance and the composition teacher. College Composition and
Communication, 26(3), 238-243.
Stiff, Robert. (1967). The effect upon student composition of particular correction techniques. Research in the
Teaching of English, 1(1), 54-75.
Straub, Richard. (1997). Students' reactions to teacher comments: An exploratory study. Research in the
Teaching of English, 31(1), 91-119.
Strenski, Ellen (Ed.). (1986). Possibilities: Scenarios and scripts to help teaching assistants respond to student
writing in all disciplines. ERIC Document Reproduction Service, ED 315 768.
Swilky, Jody. (1991). Cross-curricular writing instruction: Can writing instructors resist institutional
resistance? ERIC Document Reproduction Service, ED 331 066.
Tobin, Lad. (1989). Bridging gaps: Analyzing our students' metaphors for composing. College Composition and
Communication, 40(4), 444-458.
Turbill, Jan. (1994). Getting a nut and bolt with the same thread: Matching evluation theory with language
learning theory. In Cambourne, Brian, & Jan Turbill (Eds.), Responsive evaluation: Making valid
judgments about student literacy (pp. 8-15). Portsmouth, NH: Heinemann.
Underwood, Jody S., & Tregidgo, Alyson P. (2006). Improving student writing thrugh effective feedback: Best
practices and recommendations. Journal of Teaching Writing, 22(2), 73-97.
Walvoord, Barbara F., & McCarthy, Lucille P. (1991). Thinking and writing in college: A naturalistic study of
students in four disciplines. Urbana, IL: National Council of Teachers of English.
Whichard, Nancy W., Gamber, Cayo, Lester, Valerie, Leighton, Gordon, Carlbery, Judith, & Whitaker, William.
(1992). Life in the margin: The hidden agenda in commenting on student writing. Journal of Teaching
Writing, 11(1), 51-64.
White, Edward M. (1984). Post-structural literary criticism and the response to student writing. College
Composition and Communication, 35(2), 186-195.
Winter, Janet K., Neal, Joan C., & Waner, Karen K. (1996). Student and instructor use of comments on business
communication papers. Business Communication Quarterly, 59(4), 56-68.
Wonnberger, Carl G. (1955). Judging compositionsmachine method. English Journal, 44(8), 473-475.
Writing Program Administrators. (2000). WPA Outcomes Statement for First-Year
Composition. http://www.english.ilstu. edu/hesse/outcomes.html.
Yancey, Kathy B., & Huot, Brian A. (Eds.). (1997). Assessing writing across the curriculum: Diverse approaches
and practices. Greenwich, CT: Ablex Pub. Corp.
Yorio, Carlos. (1989). The other side of the looking glass. Journal of Basic Writing, 8(1), 32-45.
Haswell 26
Young, Art. (1997). Mentoring, modeling, monitoring, motivating: Response to students' ungraded writing as
academic conversation. In Sorcinelli, Mary D., Elbow, Peter (Eds.), Writing to learn: Strategies for
assigning and responding to writing across the disciplines (New directions for teaching the learning, No.
69) (pp. 27-40). San Francisco, CA: Jossey-Bass.
Zak, Frances. (1990). Exclusively positive responses to student writing. Journal of Basic Writing, 9(2), 40 -53.
Zerger, Sandra. (1997). "This is chemistry, not literature": Faculty perceptions of student writing. ERIC
Document Reproduction Service, ED 411 515.
Ziv, Nina D. (1984). The effect of teacher comments on the writing of four college freshmen. In Beach, Richard,
& Bridwell, Lillian S. (Eds.), New directions in composition research (pp. 362-380). New York: Guilford
Press.
Contact information:
Ricard Haswell
Texas A&M University, Corpus Christi
rhaswell@grandecom.net
Complete APA Citation
Haswell, Richard. (2006, November 9). The complexities of responding to student writing; or,
looking for shortcuts via the road of excess. Across the Disciplines, 3. Retrieved from
https://wac.colostate.edu/docs/atd/articles/haswell2006.pdf