Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a...

63
Writing Differences in Teacher Performance Assessments: Use of African American Language September 2003 RR-03-26 Research Report E. Caroline Wylie Michelle Y. Szpara Research & Development

Transcript of Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a...

Page 1: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Writing Differences in Teacher Performance Assessments: Use of African American Language

September 2003 RR-03-26

ResearchReport

E. Caroline Wylie

Michelle Y. Szpara

Research & Development

Page 2: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.
Page 3: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Writing Differences in Teacher Performance Assessments:

Use of African American Language

E. Caroline Wylie, Educational Testing Service, Princeton, NJ

Michelle Y. Szpara, Long Island University, NY

September 2003

Page 4: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Listening. Learning. Leading.

Research Reports provide preliminary and limited dissemination of ETS research prior to publication. They are available without charge from:

Research Publications Office Mail Stop 7-R Educational Testing Service Princeton, NJ 08541

Page 5: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Abstract

This study investigated potential writing differences between African American and European

American candidates in a National Board for Professional Teaching Standards (NBPTS)

portfolio entry from the Middle Childhood/Generalist Certificate. The use of features other than

Edited American English was examined as a possible source of construct-irrelevant variance in

assessors’ scoring judgments. Thirty-two written commentaries, 18 from African American

candidates and 14 from European American candidates, were coded for grammatical, lexical, and

discourse features. The coding frame identified features of African American Language (AAL)

and Speech Code Errors (SCE). Instances of AAL were fewer than instances of SCE and

clustered according to potential users of AAL.

Key words: African American Language, Speech Code Errors, National Board of Professional

Teaching Standards, NBPTS, construct-irrelevant variance, differential performance

i

Page 6: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Acknowledgments

The authors wish to extend thanks to Dr. Carolyn Adger of the Center for Applied Linguistics,

Dr. Judy Diamondstone of Clark University in Worchester, MA, and the other members of the

national expert panel who guided the development of the coding frame and reviewed early drafts

of this report.

ii

Page 7: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Table of Contents

Page

Introduction..................................................................................................................................... 1

Differential Performance of African American Test Takers .............................................. 1

Focus for the Research Study ............................................................................................. 2

Terminology........................................................................................................................ 4

Literature Review............................................................................................................................ 5

Southern White English ...................................................................................................... 7

Impact of Non-edited American English ............................................................................ 8

Research Question ........................................................................................................................ 10

Methodology................................................................................................................................. 10

Developing the Coding Schema........................................................................................ 11

Refinement of Coding Schema ......................................................................................... 12

Identification and Training of Coders............................................................................... 13

Assignment of Written Commentaries to Coders ............................................................. 15

Triangulation of Results Across Coders ........................................................................... 16

Data Analysis .................................................................................................................... 17

Results........................................................................................................................................... 19

Coding Examples .............................................................................................................. 20

Results by Language Category ......................................................................................... 24

Nouns .................................................................................................................... 27

Verbs ..................................................................................................................... 30

Lexical Features: Morphemes, Prepositions, and Connectors .............................. 34

Combined Non-edited American English Usages at the Lexical Level ............... 36

Hypercorrection and Oral Language..................................................................... 39

Discussion and Conclusions ......................................................................................................... 41

References..................................................................................................................................... 45

Notes ............................................................................................................................................. 50

Appendix....................................................................................................................................... 51

iii

Page 8: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

List of Tables

Page

Table 1. Breakdown of the 32 Candidates by Race and Entry Score ...................................... 18

Table 2. Examples From Candidates’ Written Commentaries of the AAL and SCE Codes... 21

Table 3. Examples From Candidates’ Written Commentaries of the AAL and SCE Codes... 22

Table 4. Examples of Lexical Features.................................................................................... 23

Table 5. Examples of Hypercorrection and Oral Language .................................................... 24

Table 6. Summary of Coding Results...................................................................................... 25

Table 7. Instances of AAL Noun Features .............................................................................. 28

Table 8. Instances of SCE Noun Features ............................................................................... 28

Table 9. Instances of Non-EAE Noun Features (Collapsed Across AAL and SCE)............... 29

Table 10. Instances of Dropped “–ed” in Past Tense Verbs...................................................... 30

Table 11. Instances of SCE in Irregular Past Tense Verbs........................................................ 31

Table 12. Instances of Non-EAE Form (AAL and SCE) of Regular and Irregular Past Tense

Verbs.......................................................................................................................... 31

Table 13. Instances of Dropped “–s” in the Third Person Singular Present Tense of Regular

Verbs.......................................................................................................................... 32

Table 14. Subject-Verb Discord ................................................................................................ 33

Table 15. Other Non-EAE Verb Forms ..................................................................................... 34

Table 16. Presence of Missing or Incorrect Morphemes, Prepositions, and Connectors .......... 35

Table 17. Instances of Lexical Features..................................................................................... 36

Table 18. AAL and SCE Codes Combined ............................................................................... 37

Table 19. Total Number of Instances of AAL Features ............................................................ 38

Table 20. Total Number of Instances of SCE Features ............................................................. 38

Table 21. Total Number of Instances of Non-EAE (AAL and SCE) Features.......................... 39

Table 22. Presence of Hypercorrection and Oral Language...................................................... 40

iv

Page 9: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Introduction

Due to increased scrutiny on student achievement and teacher assessment, there has been a

greater emphasis on teacher assessment at all stages of a teacher’s career—from pre-service

teaching to master teaching. These assessments cover basic skills, subject matter, and pedagogy,

and utilize multiple-choice, constructed-response, observations, and portfolios to measure what it

is that teachers know and are able to do.

While the debates over the advantages and disadvantages of each type of assessment are

numerous, they share one attribute—the differential performance of minority teachers on all forms

of teacher assessments. Differential performance results when a specific population subgroup

achieves a passing rate that is significantly lower than that of the normative reference group.

African Americans do less well, in general, on all types of assessments, including constructed-

response tests; all users of test score data and test-makers, themselves, must address this issue

(Ladson-Billings, 1998; Lam, 1995). The present study is concerned specifically with the writing

styles of African American test takers on constructed-response tests, when compared with

European American test takers.

In the case of the portfolio assessment under study, qualitative evidence (Onafowora, 1998)

suggested that some candidates might be using forms of English other than Edited American

English (including African American Language and Speech Code Errors). As part of the intensive

training for assessors who score the portfolio assessment, there are multiple training activities

focused on bias reduction, including specific attention to the influence of writing style. Assessors,

like all individuals, hold preferences for specific types and qualities in candidates’ writing.

However, it is still possible that, without conscious attention to the demands of the scoring rubric,

assessors could allow perceptions of a candidate’s writing style to influence their overall

evaluation of the content. This necessitates the examination of the style and form of the

candidates’ writing in order to explore this potential source for the observed differential

performance.

Differential Performance of African American Test Takers

The differential performance of African American test takers is seen across all types of

constructed-response tests, including short answer, extended response, and portfolio entries

(Jencks & Phillips, 1998), both within the field of educational assessment (Whitaker, 1989) and in

1

Page 10: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

other professional assessments (Hartigan & Wigdor, 1989; Gatewood & Perloff, 1990; Ogbu,

2002). While the gap has been decreasing over time (Jones, 1984), it is by no means closed and

may be increasing once more (Lee, 2002). Differential performance may be due to factors within

the assessment itself, within the scoring design, or with the scorers of constructed-response and

performance-based assessments (Bond, 1998a, 1998b). The causes also may lie outside the

assessment, in the sample of test-takers, specifically in the educational opportunities provided to

test-takers in grade school, college, and post-graduation professional development. Equity of

access to resources such as test preparation materials, release time from work to prepare, and

mentoring also may play a role (Ladson-Billings, 1998). Finally, researchers have explored

interactional factors in differential performance, including stereotype threat (Steele, 1997) and use

of culturally and linguistically distinct modes of expression in assessments (Irvine & Fraser, 1998;

Ball, 1997; Kochman, 1981).

Focus for the Research Study

This study employs a portfolio-entry assessment as a means to examine more closely the

writing of African American candidates—specifically language choice, rhetorical style, and

organization. If significant differences are found in the writing of one candidate subgroup versus

another, then further analyses can be done to explore the relationship between writing differences

and scores achieved on the assessment.

This study uses a portfolio entry from the National Board for Professional Teaching

Standards (NBPTS) Middle Childhood/Generalist (MC/Gen) certificate. Candidates submit 10

components including portfolio entries, which focus on various aspects of classroom performance,

and accomplishments outside the classroom such as working with students’ families and other

colleagues and professionals. The validity of the MC/Gen assessment has been documented

through a content-related validity study (Benson & Impara, 1987), demonstrating the

appropriateness of the assessment for identifying evidence of accomplished teaching. Studies also

have documented the value of NBPTS certification in recognizing the skills of accomplished

teachers, and specifically the added value these teachers bring to the classroom (Bond, Smith,

Baker, & Hattie, 2000). In previous studies attempting to examine the sources for the differential

performance of minority candidates on this assessment, there has been no evidence of discernable

2

Page 11: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

biases within the content of the assessment, the test directions, or the test questions (Bond, 1998a,

1998b).

The researchers selected the MC/Gen certificate from the array of offered certificates,

because it has the greatest number of African American candidates and so provided a larger pool

from which to select responses. The specific entry, Writing: Thinking through the process, was

selected, as it was one of two portfolio entries that did not require a videotape excerpt of the

candidate and students. It was desirable to select an entry where a candidate’s race would not

necessarily be immediately obvious to an assessor. This particular entry requires a candidate to

select two writing assignments/prompts to demonstrate how the individual uses writing to develop

students’ thinking and writing skills for different audiences or purposes. The entry also requires the

candidate to select two students from the class and to provide their responses to both writing

assignments/prompts. Accompanying the student work is a 12-page written commentary in which

the candidate describes his or her teaching context, the nature of the two assignments, and how

they fit into the teacher’s overall writing instruction for the year. In addition, the candidate is

required to analyze the writing of the two students and reflect on the success of using both

assignments/prompts. For this study, only the candidate’s writing was analyzed and coded; the

students’ responses to the two assignments/prompts were not considered.

Portfolio responses are scored holistically by practicing teacher-peers, who have

successfully completed a rigorous four-day training. The training given to these assessors in

preparation for scoring includes emphasis on recognizing and screening out personal and societal

biases. One of the four bias-reduction exercises that assessors complete focuses on the potential

impact of writing on scores, and how to minimize that impact by recognizing and documenting

personal triggers. Much of the training is devoted to understanding and practicing application of

the four-point score scale (1-4). Each of the four points on the scale can be modified with a plus

(+) or a minus (–) that increases or decreases the whole number score by 0.25, so that the score

scale is as follows:

1- 1 1+ 2- 4- 2 2+ 3- 3 3+ 4 4+

1.25 1.00 0.75 1.75 3.75 2.00 2.25 2.75 3.00 3.25 4.00 4.25

3

Page 12: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Candidates who are not certified on their first attempt have an additional two years in

which to resubmit responses. Candidates can redo any entry that received a score of less than 2.75.

For that reason, the score 2.75 represents a critical point on the score scale.

Terminology

The terminology for describing the language used by some African Americans has changed

rapidly over the last three decades. While recognition of the differences in language between

African Americans and other English speakers dates back to the late 18th century in early writings

about the language of African slaves, the term “Ebonics” is relatively young (Smitherman, 2000).

It was first coined at a conference on language and African American children in 1973. From

there, just as descriptors for African Americans have evolved over time, so too have the terms used

to refer to Ebonics. Following the writing of one eminent researcher in this area whose work spans

several decades, one can observe these shifts in terminology. Smitherman (1973), in one of her

earlier published works, uses the term Black English. In a 1997 article (Smitherman, 1997a), she

charts how Black English changed to Black Vernacular English, and then to African American

English. In a later article that year (Smitherman, 1997b), the term experienced a further shift to

African American Language (AAL).

In consultation with a national panel of linguists experienced in the study of African

American Language, it was agreed that the term AAL would be used for the purpose of this

research. Similarly, it was agreed that equanimity in naming would best be achieved through the

use of “African American” and “European American,” rather than “Black” and “White.” To

indicate cultural ways of knowing, believing, and doing that are specific to a particular racial

group, the terms African American Cultural Indicators (AACI) and European American Cultural

Indicators (EACI) were developed. Examples of these in a teaching context might be the use of

directives from teacher to student to achieve a desired behavior (AACI), versus the use of

interrogatives to achieve the same behavioral goal (EACI). This difference in the form of request

made by a teacher to their students was documented in Heath (1994). Culturally relevant pedagogy

(Ladson-Billings, 1998) is a term that encompasses a set of pedagogical methods that are uniquely

appropriate to the cultural and linguistic styles of a particular racial or ethnic group. The term was

originally developed with specific reference to best practices for teaching African American

students; culturally relevant pedagogy methods for African American students would include

4

Page 13: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

approaches such as incorporating real-life experiences from home and community into the

standard curriculum.

The final set of terms involves the linguistic categories into which the different language

patterns were divided. These included African American Language (AAL), Edited American

English (EAE), Southern White English (SWE), and Speech Code Errors (SCE). The use of Edited

American English indicates emphasis on the edited nature of formal prose, written under the

constraints of the assessment directions. The term “standard American English” was avoided, in

part, because it privileged American English over other forms of English, including African

American Language. Southern White English refers to a dialect of English specific to the

Southeastern United States, which developed in parallel with African American Language (Feagin,

1997). The roots of SWE are discussed in more detail in the literature section. Speech Code Errors

encompasses errors that are made by either users of African American Language or Edited

American English that do not conform to consistent grammatical structures. These categories will

be discussed further in the methodology section.

Literature Review

The existence of AAL as an oral language is well documented (Baugh, 1983; Feagin 1997;

Linguistic Society of America, 1997). Writings by authors such as Geneva Smitherman and Toni

Morrison, to name only two, incorporate AAL in their works, but overall less research has been

conducted on written forms of AAL versus spoken forms. Exceptions to this are studies of AAL in

hip-hop and rap music (Smitherman, 1997b) and some work in the area of student writing

discussed below.

Qualitative evidence has suggested the presence of some features of AAL and the use of

African American Cultural Indicators (Onafowora, 1998) in a subset of written assessments

produced by African American teachers. The use of different linguistic codes, specifically African

American Language, has been well documented in the writing of grade school and college students

of color (Blumenthal & Hildenbrand, 1993; Ball, 1992). This use of linguistic codes and cultural

references distinct from Edited American English and majority-group cultural references (White or

European American and middle class) has not been explored extensively in performance-based

teacher assessments.

5

Page 14: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

It is important to note that, in general, there has been observed a low frequency of speech-

to-writing transference for African American Language features. There are distinct patterns of

usage of AAL within texts that consist of predominantly Edited American English, but the patterns

exist at a very low frequency (Smitherman, 1992; Chaplin, 1987). While the use of AAL may be

minimal, it still holds the potential to impact the reader and have the undesired effect of producing

stigmatization of the writer; the phenomenon of stigmatization will be discussed in more detail

below. Even if a reader is intentionally attempting to overlook grammatical and syntactical

features inconsistent with Edited American English, the presence of such features may cause the

reader to devalue the content of the text (Santos, 1988). The following sections examine the

literature that documents the existence of AAL as a language form with consistent grammatical

and syntactic rules. Research in support of the presence of Southern White English and studies on

the stigmatization of non-standard English usage also are discussed.

Smitherman (1992) conducted a longitudinal study of African American student writing

using the National Assessment of Progress (NAEP) assessment. Among the range of features that

she identified were the absence of the “ed” morpheme in past tense main verbs, perfect tense main

verbs, verbal adjectives, and passives; the absence of the “s” morpheme in plural nouns, possessive

nouns, and third person singular verbs; hypercorrection; dropped copula; subject-verb agreement;

multiple negation; and pronominal apposition. These features formed the basis of the present

study’s coding frame, later supplemented and refined by the national expert panel.

Fogel and Ehri (2000) conducted a study in which they focused student learning on six

syntactic features that differed according to the form of spoken English used by the students. In

this study the authors focused on syntactic features since “unlike phonological features,

nonstandard syntactic forms tend to stand out and, as a result, can be stigmatizing (Burling,

1973).” In the study, the six standard syntactic features were possessive “s,” past tense “ed,” third-

person present-tense singular “s,” plural “s,” indefinite article, and subject-verb agreement.

Ball (1998) analyzed a student paper written by a high school sophomore who spoke

African American Vernacular English (AAVE). In this writing, she identified a number of features

common to many AAVE speakers, such as not using the third-person singular present tense “s,”

and a range of semantic and phonological patterns. One pattern of interest was the student’s use of

the word “willn’t” instead of “won’t.” Ball noted this as potentially an example of what Wolfman

and Whiteman (1971) identified as hypercorrection.

6

Page 15: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Hypercorrection results from the writer’s attempt to produce academic patterns with

which he or she is not completely familiar. Because of this unfamiliarity, the writer

incorporates items not only where they are appropriate, but in inappropriate places

as well. This phenomenon appears in spelling, vocabulary choices and syntactic

structures, and is not exclusive to AAVE speakers. (Ball, 1998, p. 235)

Smitherman (1994) and Ball (1995), among others, draw a distinction between syntactic

and discourse writing features, both of which are present in the writings of African American

students. At the discourse level, typical features are “the use of rhythmic language, anecdotes,

parables, and patterns of repetition and call and response” (Ball, 1995, p. 259). Within the context

of the portfolio assessment under study, given that the format is highly structured, the candidates

tended to present fewer of these features. Syntactic features such as these were included in the

coding frame to gauge the degree of their presence in the writing and to provide a more complete

picture of any features that might contribute to construct-irrelevant variance in the assessors’

scoring.

Smitherman extended her work on the NAEP study by drawing parallels with similar

studies done by other researchers. In her 1994 article, Smitherman discussed research by Chaplin

(1987) who also used African American student writing from NAEP. In that study she noted that

some students used what she described as “conversational tone” which she identified as an essay

that read like “recorded language or a conversation” (Chaplin, 1987, p. 37). This was identified as

“use of oral language in written form” in the coding frame for the present study.

Southern White English

AAL is spoken by a subset of the African American population who has been exposed to it

in the home and/or community. Smitherman (2000) estimates that approximately “90% of the

African American community uses one or more aspects of the language some of the time.” AAL is

distributed across the United States and is not limited by socioeconomic boundaries. Southern

White English, a localized dialect of American English, is largely constrained to the southeastern

United States (Bailey, 1997; Feagin, 1997). SWE has been extensively documented in references

such as the Annotated Bibliography of Southern American English (McMillan & Montgomery,

1989) and the Dictionary of American Regional English (Cassidy & Hall, 1996), and demonstrates

7

Page 16: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

distinctive patterns in grammar, pronunciation, vocabulary, naming practices, and wordplay. SWE

shares many features in common with African American Language.

While the specific developmental history of both SWE and AAL are still hotly debated

among historical linguists, it is generally agreed upon that the two languages/dialects developed in

parallel due to cross-exposure between speakers of each. Specifically, as part of maintaining

European American dominance during the period of slavery, enslaved Africans were deliberately

placed in mixed groups with other Africans from dissimilar language backgrounds. In addition, it

was illegal for the enslaved Africans to learn to read or write. Since most of the enslaved

individuals had limited contact with English speakers, they were forced to develop a working

communication system referred to as a pidgin (Schumann 1978; Richards, Platt, & Platt 1993.)

This pidginized language was a combination of West African languages and English. Over time

this language was passed down to the children of the slaves, became creolized,1 and has been

maintained as part of the cultural identity of African Americans still today.

It is hypothesized (Schumann 1978; Richards, Platt, & Platt 1993) that those enslaved

individuals who cared for the young children of the European American plantation owners had a

strong effect on the children’s initial language development, bringing the influence of the nascent

African American Language to bear on the dialect of Southern White English speakers. Evidence

also exists in today’s post-slavery times that the intermixing of Blacks and Whites can lead in

certain cases to the altering of what is defined as Standard American English.

Support for the idea of cross-fertilization of language between African American and

European Americans comes from a study conducted by Botan and Smitherman (1991), which

looked at the degree to which White American industrial workers had familiarity with Black

English. The study showed that while White workers varied in their degree of familiarity with

Black English depending on employment group, African American workers did not. This study

lends support to the notion that while African Americans might learn Black English from a variety

of sources (family, peers, work, church), White workers who are not in contact with African

Americans speaking Black English are much less likely to do so.

Impact of Non-edited American English

Many linguistic researchers such as Ball, Smitherman, Labov, and others, have contributed

to the literature on AAL, describing the historical roots, linguistic structures, and documenting its

8

Page 17: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

use by different groups and in different contexts. Less work has been done on written forms of

AAL than oral forms, although the area of student writing has received considerable attention.

Little specific work has been conducted regarding impact on the writer, but the fact that many

writers have been striving to have AAL recognized as a language and not just as “poor grammar”

indicates an underlying assumption that there are negative connotations with its use.

Smitherman (1992) states that “the question of the relationship between writing scores and

distributions of BEV (Black English Vernacular) in writing is a logical issue given assumed

negative teacher attitudes and the stigma attached to features of Black English in writing.” Ball

(1998) discusses strategies for teachers to employ when working with students who use AAVE

(African American Vernacular English), which infers that they do not know how to appropriately

work with those students. This lack of understanding is not limited to just European American

teachers, as Ball notes that “most teachers, including many African American teachers, are

unaware of the subtle features that characterize AAVE as a distinct linguistic system.” Erickson

(1984) describes it as a “low-prestige” language, and Anderson (1981) noted that employers would

use non-mainstream English as a criterion for eliminating otherwise well-qualified job candidates

who were applying for positions that did not require proficiency in mainstream or academic

English.

The field of research in English as a Second Language (ESL) provides more specific

insights into the stigmatization of non-standard English writing. Research findings are divided on

whether readers can, with training, overlook language errors and focus on content alone.

According to Santos (1988) and Leki (1991b), faculty readers are able to overlook local errors—

those that do not interfere with comprehension. However, this “allowance” may not be readily

extended to native speakers’ writing; errors were perceived as indications of “lack of maturity of

thought and of rhetorical style.” Another insight from ESL research involves the importance

placed on errors according to their frequency. Albrechtson, Henriksen, and Faerch (1980) found

that the specific type of error was of little consequence; what counted was the number of errors.

Two of the most irritating and least acceptable errors were double negatives and subject/verb

agreement, both of which have parallels in AAL. ESL research has shown that, due to lack of

awareness about the second-language learning processes, faculty readers may mistakenly assume

that some common types of errors are “careless errors,” and therefore they may feel justified in

blaming students for making them (Kroll 1991; Vann, Meyer, & Lorenz 1984). Since many

9

Page 18: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

teachers are unaware of the existence of AAL, they also may assume that AAL language patterns

are, in fact, poor English grammar.

Research Question

The research presented in this paper addresses the question of whether there is systematic

evidence of the use of African American Language in a subset of analytic commentaries of

teaching practice (known as written commentaries) prepared by African American teachers. The

research sought to systematically look for language patterns by candidates’ race and by score. That

is, do language patterns differ for African American and European American candidates, and are

there patterns by high and low scores on the entry?

The present study is the first phase of a larger project examining differences in the writing

of African American and European American candidates. The focus represented in this report is on

grammatical and syntactic features. The second phase will examine discourse level features and

cultural indicators, such as use of culturally relevant pedagogy (Szpara & Wylie, in progress).

It is important to note that this study alone cannot answer the question of whether a causal

relationship exists between the presence of language features and scores. In other words, even if

AAL language features do cluster for low scoring African American candidates, it would not be

possible to identify whether the low scores were awarded because of rubric-based decisions

(evidence of poor pedagogy) or because the writing inappropriately influenced the assessors.

For the research presented in this paper, a set of written commentaries from both African

American and European American candidates was coded for specific linguistic features. The

following is a discussion of the methodology and results.

Methodology

To develop a coding frame for systematically documenting the presence or absence of

African American Language features, a national panel of experts in linguistics and AAL were

assembled. Seven experts, including five who were African American, agreed to participate in an

initial review of MC/Gen candidate responses to the selected entry.

10

Page 19: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Developing the Coding Schema

As a primer to meeting to develop the coding scheme, each expert received a “homework”

task to review a small sample of race-blind candidate responses, and to consider the language

features present in each written commentary. The experts were asked to identify “any writing

features that were examples of non-standard American English.” Features could be on the micro-

level, such as grammar or syntax, or on the macro-level, such as organization or tone. Having

reviewed each written commentary in this fashion, the second task was to “hazard a guess” as to

the race of the candidate. The third task was to identify any written commentaries that they thought

represented a typical African American style. At this early stage no other specific guidance was

offered so as not to limit the thinking of the panel. These written commentaries then formed the

primary materials used in the development of the coding frame.

For the homework task, the seven experts reviewed a total of seventy candidate responses,

42 of which were from African American candidates and 28 from European American candidates.

One expert only reviewed each written commentary, as this was not a test of their judgments, but

rather an opportunity to explore the notion of distinct writing patterns. The experts felt that they

could tell the likely race of a candidate for approximately one third of the responses they read. The

experts identified the race of 45% (n=19) of the African American candidates 57% (n=16) of the

European American candidates. Only 11% (n=3) of the European American candidates’ responses

were considered to be mostly likely written by an African American writer, compared to 26%

(n=11) of the African American candidates’ responses which were considered to be most likely

written by an European American writer. It is interesting to note that the three European American

candidates who were identified as mostly likely being African American were all from southern

states. While this occurrence was noted during the early development of the coding frame, no

further attention was paid to the matter until the preliminary analysis was conducted on the full set

of coded written commentaries. Upon receipt of the panelists’ homework-task responses, written

commentaries from African American candidates, which had writing features identified as

examples of AAL, were copied. When the panel of experts met, this subset of written

commentaries was shared with the entire group and formed the basis for their discussions. Then

working together, the panel of experts assembled an exhaustive list of all features of the written

11

Page 20: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

commentaries that were distinct from edited American English patterns. This list was arranged into

three major categories:

African American Language •

African American Cultural Indicators

Speech Code Errors

The AAL category included many of the features common to the oral variety of AAL, such

as dropped “–s” when plural is marked elsewhere in the sentence, dropped “–ed” when the aspect

is indicated by other syntactic features of the sentence, and dropped “–s” on third-person singular

verbs when meaning was clear from context. More salient features of spoken AAL, including

double negatives and the dropped copula (“to be”), were rare or absent in the written texts.

The African American Cultural Indictors (AACI) included features such as use of

culturally relevant pedagogy (Ladson-Billings, 1998), socially conscious curriculum, and “high

talk” (Smitherman, 1977). While these features also might be identified in the writing of non-

African American candidates, the panel felt that these features, alone or combined with African

American Language features, might trigger biases in assessors who were less familiar with African

American cultural styles of teaching and communicating. The AACI codes are not presented in this

paper, but will be discussed in a separate report.

Speech Code Errors (SCE) was a category that developed unexpectedly from the panel’s

findings. SCE included errors in grammar and spelling; these errors could be found in the writing

of both African American and European American candidates. This category of codes had not been

anticipated, but it became apparent that many written commentaries had examples of SCE. It was

decided to investigate whether there is a relationship between increased presence of SCE and

lower scores across all candidates, regardless of race.

Refinement of Coding Schema

From the findings of the national panel, a coding frame was developed and subsequently

reviewed by a subset of the panel members. Each panelist received the coding frame, definitions

for each code along with multiple examples from written commentaries that they studied at the

meeting, and additional written commentaries that had not been the focus of discussion at the

meeting. The panelists were asked to code these new written commentaries using the coding

frame. Comparing across the panelists’ work, a number of codes were identified as being applied

12

Page 21: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

inconsistently. In addition, the panelists’ written comments provided further advice for clarifying

both the codes and coding procedures. With this information, the coding frame was modified

again. Added were a number of codes to address speech code errors and AAL features, as well as

the inclusion of relevant edited American English (EAE) features into the coding frame. The

coding of EAE allowed for the calculation of the relative frequency of a feature of AAL in relation

to the frequency of use for the related EAE feature (Hatch & Lazaraton, 1991). The appendix

provides the subset of the coding frame discussed in the results section.

Identification and Training of Coders

The set of written commentaries included both African American and European American

passing and non-passing candidates. To identify the language features, the group of coders to had

be familiar with basic linguistics, English grammar rules, African American Language rules, and

African American culture, specifically the cultural styles of African American teachers. For this

reason, professors were approached at a Historically Black University in the southern United

States, which has a Masters-level speech and language department within its School of Education.

The professors recruited students from their speech and language courses. Prospective coders

completed a biographical data form, which included questions on educational experience,

education courses taken, and language and linguistic courses taken. From the respondents a group

of 13 individuals was selected to complete a training session, with the expectation that not all of

them would become coders for the study.

The majority of the coders were in the second year of a Masters program in

Communication Disorders. Their years of experience in classroom settings ranged from one to

seven years, with nine coders having two or more years of experience. Their background in

education coursework included the following: Educational Psychology, Clinical Observation,

Scientific Methods of Educational Research, Racial and Cultural Minorities, Educational

Foundations, Language Arts Instruction, and Human Growth and Development. Their collective

background in language and linguistics included: African American Literature I and II, Speech and

Dialect, Grammar, Introduction to Phonetics, Child Language Disorders, Multicultural Issues in

Speech-Language Pathology, Introduction to Communication Disorders, Introduction to

Linguistics, Early Childhood Language Development, Speech Science, English as a Second

Language, and Normal Language Development. Six of the coders self-identified as African

13

Page 22: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

American, four as European American, one as Native American, and two as multi-racial/multi-

ethnic (African American/European, American/Native American).

The training of coders consisted of three days of intensive review of African American

Language (AAL) and Edited American English (EAE) grammar, syntax, and rhetorical features,

with two major goals: (1) creating a shared understanding among coders of the definition of each

feature, and (2) developing coders’ proficiency in recognizing and identifying features of the

coding frame in written commentaries.

To achieve these goals, the training followed a consistent pattern for each block of codes.

The initial discussion focused on a meaningful group of codes to ensure a common understanding

of what would and would not be included in each code category. For example, all the codes

relating to nouns—plurals and possessives—were taken together with part of the discussion

focusing on what correct and incorrect use of an apostrophe would be, and reinforcing that non-

standard plural nouns such as children (that did not add an “–s” to indicate plurality) were not to be

coded. Next, some short sentences, which exemplified each code, were coded and discussed by the

group. A longer passage then provided a practice opportunity to code for more than one code type.

This process was then repeated for the next set of codes, with each longer passage focusing on all

codes considered up to that point, to reinforce previous lessons. At the end of each day, the longer

passages coded by each student were reviewed by the two researchers. Individualized feedback

was written for each student to review the next day. The final passage that students coded at the

end of the third day was longer than previous passages and included examples of codes from the

entire coding frame. The researchers reviewed each student’s codes and again provided written

feedback to each student to clarify any misunderstandings that were evident. In some cases, the

feedback included grammatical explanations (such as tips on distinguishing between various forms

of “–ed” endings on verbs); in other instances it included tips for coding to avoid overlooking

blocks of codes.

Upon completion of the training, two of the students decided not to proceed any further

with the project, but the other eleven coders completed an independent coding of a single set of

written commentaries. This served as a second validation of the coding frame, in support of the

findings of the national panel, and also provided the researchers with further evidence of coding

proficiency. Within the categories of grammatical and syntactic-level features, coders

demonstrated strong agreement in coding. As the section on triangulation across coders illustrates,

14

Page 23: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

the final counts for each code category were not dependent on any one coder, nor were simple

averages across multiple coders taken. This approach accommodated occasional inconsistencies in

coding. For example, one coder identified three instances of the zero copula in this final practice

passage. Had this written commentary been part of the actual data set the researchers would have

located the three instances identified by this coder and seen that she appeared to have applied a

consistent rationale to the identification, albeit an incorrect one. The excerpt below indicates where

the coder had identified the three instances of the zero copula:

“The kind of writing my students do during the course of the day: …

• Dialogue journals that my students use for writing to one another and/or the teacher.

• Double entry journals that my students use to summarize what they learn and react to

the content or learning process. …

• Free writing [is] where my student practice steadily writing without stopping …

• Learning Log [is] where my students record what they have learned

• Sustained [is] writing where my students steadily write without interruption for a

period of time.”

All three instances identified by this coder followed a pattern where the coder deemed that

the word “is” should have been inserted before the word “where.” Considering the distance

between the three bullet points and the introductory sentence that governs the subsequent list the

over-application of the coding frame is understandable. This misapplication of the code would

have been caught through the triangulation process. The dual process of using three coders to code

each written commentary, followed by review from the researchers, resulted in corrections being

made to either misapplications of the coding frame or failure to identify all examples of a

particular code. Applying the triangulation process to the results of the coders on this final coding

passage resulted in the identification of all critical codes in the passage.

Assignment of Written Commentaries to Coders

For the initial review of written commentaries, development of the coding frame, and

training of the coders, candidate responses from the 1998-99 cohort were used. For the study itself,

candidate responses were selected from the 1999-2000 cohort. The entry directions did not change

from the 1998-99 to the 1999-2000 administration. All categorizations by race were made on the

basis of candidate self-identification.

15

Page 24: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Ninety-four written commentaries were selected, consisting of 64 written commentaries

written by African American candidates and 30 written commentaries written by European

American candidates, matched on gender, geographic locale, and the range of scores received.

Written commentaries by African American and European American candidates were selected in

approximately a two-to-one ratio. The focus of the study was on language usage of African

American candidates, but it was also important to have written commentaries from European

American candidates to provide a basis for comparison.

The written commentaries were randomly assigned to the 11 coders who performed

consistently during training and successfully completed the post-training coding sample. There

were three copies of each of the 94 written commentaries, distributed that each coder received 26

written commentaries to code. The coders were arranged in a way that for each commentary read,

the three coders were not all African American or European American.

Of these 11, two withdrew from the study for personal reasons before they coded any

written commentaries; their material was redistributed among the remaining coders.

Triangulation of Results Across Coders

Each written commentary was coded independently by three coders to provide a

triangulation of results for each feature in the coding frame. Rather than rely on the coding of any

one individual or straight averages of two or more individual coders, the following resolution

process was applied. Once the three sets of codes were completed for a particular written

commentary, the data was entered into a database and reviewed for outliers or discrepancies in the

three coders’ results.

For codes with low frequency occurrences (less than ten), all instances of the code were

investigated. The majority of codes discussed in this report fell in the low frequency category. All

instances of AAL codes or SCE found by the three coders were identified by one researcher and

verified by the second. For example, all three coders might have each identified two instances of a

dropped “–s” on a plural noun, but in fact between them the three coders identified a total of four

examples. In this instance, all four codes having been reviewed and verified by both researchers,

the final count for that code was four.

16

Page 25: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

For codes with a moderate (ten to thirty) to high (more than thirty) occurrence rate, a

reasonable degree of discrepancy was tolerated: for moderate frequency codes, discrepancies up to

four and for high frequencies discrepancies up to eight were judged to be reasonable. When the

disagreement was low among the three coders, the average of the codes across the coders was

used. For example, the instances of plural nouns with “–s” (Edited American English) in a

particular candidate’s response were coded as 327, 340, and 337, respectively across three coders.

Given the high level of occurrence, the average of the three numbers was used—335.

If a large discrepancy was noted, even at high levels of occurrence, all instances of the

language feature were checked across multiple coders. When a discrepancy was clearly the result

of a misapplied code, a correction was made to the data set. Occasionally the discrepancy

highlighted an important ambiguity in the coding, which was then resolved through consensus

among the researchers. This resolution process served to reduce instances of miscoding while still

maintaining the integrity of the triangulated data. In no instance did a single coder determine the

nature of a particular linguistic feature in the data set.

Data Analysis

Due to time demands on the student coders, some written commentaries were not coded by

three people. Those written commentaries had only two sets of codes, thereby reducing the degree

of assurance provided regarding the quality of the coding results. Therefore only the thirty-two

written commentaries that had three sets of codes were used for this analysis. Once the

triangulation of the data (described above) was completed, the data set from the 32 commentaries

was cross-checked for accuracy once more.

Of the 32 candidates, 18 self-identified as African American, and the remaining 14 as

European American. Two African American candidates and one European American were male.

Table 1 provides the breakdown by race and entry score of the candidates in the dataset.

17

Page 26: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Table 1

Breakdown of the 32 Candidates by Race and Entry Score Candidates Entry score Number of candidates in

group < 2.75 10 African American ≥ 2.75 8

< 2.75 9 European American ≥ 2.75 5

As indicated earlier, the score scale using for scoring responses to all NBPTS entries spans

from 0.75 to 4.25. The score of 2.75 represents the lowest level of an accomplished performance

(National Board for Professional Teaching Standards, 2003.). For the purpose of this research, the

score scale was divided in half; low scores fall in the range of 0.75 to 2.74, and high scores in the

range of 2.75 to 4.25. It is important to recognize that while this division is somewhat artificial and

that performances with adjacent scores will not be qualitatively different from each other, the

National Board policy for unsuccessful candidates on retaking individual entries/exercises uses

2.75 as the cut-off point. The mean exercise score for both the low scoring African American and

low scoring European American candidates was 1.9, while for both high scoring groups of

candidates it was 3.3. The diagram below shows the score distribution for both groups of

candidates, relative to the 2.75 dividing point.

AA

EAEAEA

EA EA

EA

AA AA

AAAA

AA AA

AAAAAA

AAAA AA

AA AA

AA AA AA

4.25 4.0

3.75

EA

3.25 2.75

3.0

EA

2.25 2.0 EA

1.75 EA

0.75

1.0

EA EA

1.25

EA EA

The initial data analysis involved a series of cross-tabulations of the frequency of usage of

each code for African American and European American candidates. This analysis resulted in data

anomalies that were difficult to explain. In particular, there were several cases of European

American candidates using language features that had been considered primarily AAL features.

18

Page 27: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Considering overall patterns of usage, four European American candidates in particular

were identified. These four European American candidates were all teaching in Southern states

(Mississippi, North Carolina, and South Carolina). To shed light on the patterns in the data, these

four candidates were designated as potential users of Southern White English (SWE), which as

discussed in the literature has substantial overlap with AAL. It is important to note, that just as all

African Americans do not use AAL, not everyone from the southern states is a speaker of SWE,

and there are five other European American candidates from the same states who did not exhibit

SWE language features. For the purposes of this study, these codes will continue to be referred to

as AAL since the linguistic literature supports their identification as such.

Having identified these four candidates as users of SWE, the data was then disaggregated

by candidates’ self-identified race (African American and European American), and the category

of European American was further subdivided to separate the users of SWE. The data in the results

section will separate these candidates from the rest of the European American candidates for the

purpose of understanding the data more clearly. However, care should be taken not to over

interpret results for this group since only two candidates in the SWE group scored below 2.75 and

two scored at or above 2.75.

Results

The primary research question is to investigate whether there was systematic evidence of

the use of African American Language in a subset of written commentaries prepared by African

American teachers, and the degree to which all candidates exhibit speech code errors. Since the

research focus is on whether language features pattern by race or by entry score, the data are

disaggregated by race (African American and European American, with European American

further sub-divided by those who exhibit patterns of SWE use) and by entry score (divided into

less than 2.75 and greater than or equal to 2.75), recognizing that responses immediately on either

side of the 2.75 may not be qualitatively different.

Listed below are the study’s key findings. Following this list are examples from

candidates’ written commentaries organized by the major coding categories: nouns, verbs, lexical

items, and discourse level features. Finally, the full results are presented in disaggregated form, by

race and language use and by entry score.

19

Page 28: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

• In keeping with other findings in the linguistic literature (Smitherman, 2000; Chaplin, 1987),

usage of African American Language (AAL) forms in formal, written, academic prose is

relatively low.

• Evidence for usage of AAL forms appeared consistently across both high-scoring and low-

scoring African American candidates.

• The four European American candidates identified as not systematically using Edited

American English showed linguistic patterns similar to AAL, here hypothesized to be use

of Southern White English (SWE).

• Evidence for Speech Code Errors (SCE) was low across all groups, as can be expected

given the constraints of the formal academic assessment.

• African American candidates demonstrated greater evidence of SCE than European

American who employed Edited American English.

These findings are tempered by the limitations of small sample size and the challenges of

working with human coders. In the latter case, multiple checks and balances were used to minimize

the possibility of one coder missing some of the potential codes in an essay. These included their

background screening, training and post-training assessment, continued monitoring and feedback

during the coding process, resolution of discrepancies through a standardized process, and

triangulation of final results.

Coding Examples

Before discussing the coding results, the following tables present examples of each of the

codes taken from candidates’ responses. Tables 2, 3, 4, and 5 present the results for nouns, verbs,

lexical features, and other discourse level features respectively.

Nouns were coded for features typical to speakers of AAL, as well as common SCE

identified by the panel of experts. Table 2 (below) presents an example of each code. Plural nouns

with a dropped “–s” and possessive nouns with a dropped “–s” and no apostrophe were coded as

possible examples of AAL. Plural nouns with an unnecessary apostrophe, and possessive nouns

with no apostrophe or an incorrect apostrophe were coded as examples of SCE.

20

Page 29: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Table 2

Examples From Candidates’ Written Commentaries of the AAL and SCE Codes

noun, regular plural “–s” has been dropped

“I may touch on sentences that go on and on as well as looking at various way_ to end of piece.” “They were both able to produce a complete paper that met the goal_ that were set”

noun, regular plural, unnecessary apostrophe is added

“Good writer’s research their topic” “the area’s of reading”

noun, singular or plural, possessive, both “–s” and apostrophe have been dropped

“On the personal narrative, Nickie__ work demonstrated similar abilities.” “Whitney_ responsibility was to copy her letter with the correction...”

noun, singular or plural, possessive, “–s” is used but apostrophe has been dropped

“students ability.…students experiences….persons life….students understanding”

noun, singular or plural, possessive, “–s” is used but apostrophe not used according to EAE rules

“The one girls’ coat got stolen last night.” “Twenty boy’s permission slips arrived today in the mail.”

Verb features identified by the expert panel as significant either for AAL or SCE were

dropped “–ed” in past tense verbs, non-Edited American English (non-EAE) forms of past tenses

for irregular verbs, dropped “–s” in the third person singular present tense, and subject-verb

discords. A feature common in spoken AAL is indicating past tense for regular verbs without the

Edited American English “–ed.” The dropped “–ed” feature was examined in simple past tense

verbs, perfect tense, verbal adjectives, and passive voice. The literature on spoken AAL does not

identify a specific pattern for formation of irregular verbs. (Labov, 1972; Rickford, 1999; Rickford

& Rickford, 2000). For the purpose of the analysis, irregular verbs not conjugated in the EAE form

are designated as speech code errors. As for the regular verbs, this occurrence could be found for

irregular past tense, perfects, verbal adjectives, and passive verbs. The subject-verb discord

specifically marked the dropped “–s” on the third person singular verb in the present tense as being

a potential indicator of AAL use when the verb and subject when separated by no more than one or

two words. Lengthy separation between the subject and verb could lead to lack of noticing a

subject-verb discord; therefore these were counted as SCE.

21

Page 30: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Table 3

Examples From Candidates’ Written Commentaries of the AAL and SCE Codes

regular verb, simple past tense, “–ed” ending has been dropped

“Step two, she design__ her drafting her story (composing), Step three, she used editing to refine what she had written.”

irregular verb, simple past tense, EAE form is not used

“Yesterday the students run out of the classroom, despite the teacher’s request not to. “[ran]

regular verb, present/past/future perfect tense “–ed” ending has been dropped

“We have revise_ the models we wrote together as a class to model for my students how to go back and dress up the ideas they were able to put down in their first drafts.”

irregular verb, present/past/future perfect tense EAE form is not used

“I wish Brian could have showed supporting details …”

regular verbal adjective (verb in participle form used as an adjective) “–ed” ending has been dropped

“When David is asked to add more detail, he gets discourage__ because he does not know what to write...”

irregular verbal adjective (verb in participle form used as an adjective) EAE form is not used

“I asked Matt if I could see the re-wrote paper from yesterday.” [re-written]

regular verb, passive voice “–ed” ending has been dropped [be + V – ed]

“The students are expect__ to articulate the message that they are trying to relay, who do they expect to read this piece when it is finish__ and question themselves on the clarity of what they have written.”

irregular verb, passive voice EAE form is not used [be + V – EAE irregular form]

“The paper was wrote by a pair of students.” [written]

regular verb, present tense, third person singular “–s” ending is not used

“Administration usually keep_ the second graders with the same teacher for third grade.”

irregular verb, present tense, third person singular “–s” ending is not used

“He don’t understand how to do this problem.”

Other non-EAE verb forms, excluding S-V concord

“He wants to go straight to writing without given some thought about what he is going to write.” “I realize he is a writer that has being responsible for his own learning.”

22

Page 31: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

A series of codes were collected together under the composite heading of lexical features.

Five codes exist in this category: dropped final letters; other spelling errors; missing or incorrect

prepositions; other missing words; and incorrect use of connecting words. Table 4 presents

examples of these codes.

Table 4

Examples of Lexical Features

Deletion of final letter or letters, e.g., “an(d)” “a(n)” “the(y)” “the(m)” “glad(ly)”

“10 Caucasian, 3 Attention Deficit Hyperactivity Disorder, one on medication an_ 1 who suffers from moderate hearing loss.” “a_ ongoing experience” “Having the_ think about and expressed how the slaves felt, would bring to them a better understanding of what the slaves went through.”

Other spelling errors “The students have been practicing exposity writing for the past several weeks.”

Missing or incorrect preposition

“I had been planning to update my lesson plans to align with the new state standards of students.”

Other missing words “A noun is a ___ used to name a person, place, thing, or idea.” Incorrect use of connecting words (yet, but, and, so, therefore, because, since, while, etc.)

“Shannon chose to become the prime number one hundred, one. She wrote that she became a movie title and the grade on a students paper. Shannon did therefore forgot to state the difference between a prime and a composite number in her paper which was a requirement of the assignment.”

The expert panel identified three features that existed at the discourse level and focused on

the “over-application” of particular EAE grammatical rules, resulting in errors or awkward

phrasing, and oral language. Hypercorrection (error producing) occurs when the writer over-

applies a grammatical rule to a situation where it is not required, leading to an error.

Hypercorrection (without errors) occurs when a writer is overly prescriptive with grammatical

rules, but is technically accurate, resulting in text that is disfluent. The final feature identified in

this set is defined as “oral language.” This occurs when a sentence or phrase is not grammatically

correct in a formal writing situation, and yet if read aloud, the meaning of the phrase or sentence is

completely clear. Smitherman (1994) refers to this as “conversational tone.” Many instances of

oral language could be classified as the use of colloquialisms, run-ons, or quotations of dialogue

that lacked punctuation. Table 5 presents examples of these three codes.

23

Page 32: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Table 5

Examples of Hypercorrection and Oral Language

Hypercorrection (error producing)

“I would have each student generated his/her list by themselves.” “both students are gifted and abled writers” “It was graded by two kids and myself”

Hypercorrection (without errors)

“primary difficulties, seem to relate to reading and _ lack of the capacity to organize …” “I also wanted the students to use successfully the steps to the writing process…. Allowing children to write daily and make the connections in their reading to real life, help them to become independent writers with a purpose…. This presents a challenge to me because I have to remind constantly him of detail.”

Use of oral language in written form

“Keisha’s writing required few revisions or edits; however, it was difficult to impossible to convince her to re-write anything unless is was using the computer.” “She is a very sensitive girl, creative, and talented.”

Tables 2 through 5 presented examples of the codes taken from candidates’ written

commentaries. The tables in the following section present the frequencies with which these codes

occurred across candidates’ essays, disaggregated by race and language use, and by entry score.

Results by Language Category

Table 6 (following) presents the full results in summary form, organized by lexical

category—nouns, verbs, and other rhetorical features. The candidate data is disaggregated by race

and language use (African American and European American, with European American further

sub-divided by those who exhibit patterns of SWE use) and by entry score (divided into less than

2.75 and greater than or equal to 2.75). Although there are only two candidates who were classified

as users of SWE who scored less than 2.75 and two who scored greater than or equal to 2.75, their

data will be presented in Table 6 and following, since their written commentaries were part of the

original set of written commentaries by European American candidates. The decision to group

these four candidates separately was taken to make sense of the anomalies in the data from the

European American candidates. However, comparisons between the users of SWE and other

candidates will be limited as it is unwise to draw conclusions from such small sample sizes.

24

Page 33: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Table 6

Summary of Coding Results Codes African American European American - SWE European American < 2.75 (n = 10)

≥ 2.75 (n = 8) N

< 2.75 (n = 2) N

≥ 2.75 (n = 2) N

< 2.75 (n = 7) N

≥ 2.75 (n = 3) N N

p % Ni p % Ni p % Ni p % Ni p % Ni p % Ni

Nouns Plurals

Dropped “–s” 6 60% 14 5 63% 24 2 100% 15 2 100% 9 4 57% 9 0 0% 0Unnecessary apostrophe 3 30% 3 2 25% 8 0 0% 0 2 100%

3 1 14% 1 0 0% 0

Possessives No apostrophe, no “–s”

1 10% 1 1 13% 1 0 0% 0 1 50% 1 0 0% 0 0 0% 0

No apostrophe 6 60% 10 3 38% 5 2 100% 10

1 50% 10 2 29% 2 0 0% 0Incorrect apostrophe 3 30% 7 7 88% 17 2 100%

2 0 0% 0 2 29% 8 1 33% 1

Verbs Regular verbs, dropped “–ed”

Simple past

3 30% 4 1 13% 2 2 100% 2 2 100% 4 0 0% 0 0 0% 0Perfects 0 0% 0 0 0% 0 1 50% 1 1 50% 1 0 0% 0 0 0% 0Verbal adjective

3 30% 3 1 13% 1 0 0% 0 2 100% 2 0 0% 0 0 0% 0

Passive 2 20% 2 0 0% 0 0 0% 0 0 0% 0 0 0% 0 0 0% 0Irregular verbs, non-EAE form

Simple past

2 20% 2 0 0% 0 1 50% 2 1 50% 4 3 43% 4 0 0% 0Perfects 0 0% 0 0 0% 0 0 0% 0 0 0% 0 0 0% 0 1 33% 1Verbal adjective

0 0% 0 0 0% 0 0 0% 0 0 0% 0 0 0% 0 0 0% 0

Passive 0 0% 0 0 0% 0 0 0% 0 0 0% 0 1 14% 1 0 0% 0Third person singular, present tense

Dropped “–s” 7 70% 10 3 38% 9 2 100%

4 2 100% 2 0 0% 0 0 0% 0Non-EAE form 1 10% 1 0 0% 0 0 0% 0 0 0% 0 0 0% 0 0 0% 0

25

(Table continues)

Page 34: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Table 6 (continued) Codes African American European American - SWE European American < 2.75 (n = 10)

≥ 2.75 (n =8)

N < 2.75 (n = 2) N

≥ 2.75 (n = 2) N

< 2.75 (n = 7) N

≥ 2.75 (n = 3) N Np % Ni

p % Ni p % Ni p % Ni p % Ni p % Ni

Subject-verb discord Copula 3

30% 6 3 38% 7 0 0% 0 1 50% 1 1 14% 1 1 33% 2Non-copula 7 70%

12 4 50% 4 0 0% 0 1 50% 1 0 0% 0 0 0% 0

Other Other non-EAE verb forms 2 20% 2 2 25% 5 1 50%

5 1 50% 4 2 29% 2 1 33% 1

Lexical Features Dropped final letters 6 60% 7 4 50% 6 1 50% 2 0 0% 0 0 0% 0 0 0% 0Other spelling errors 8 80% 19 4 50% 16 2 100% 21 2 100% 15 4 57% 10 2 67% 2Missing or incorrect preposition 6 60% 11 6 75% 13 2 100% 3 2 100% 7 3 43% 6 1 33% 1Other missing words 7 70% 11 4 50% 16 2 100% 5 2 100% 4 5 71% 11 2 67% 2Incorrect use of connecting words 5 50% 8 1 13% 2 0 0% 0 0 0% 0 1 14% 1 0 0% 0

Hypercorrection and oral language

Hypercorrection (with errors) 1 10% 2 7 88% 12 2 100% 7 1 50% 0 1 14% 1 1 33% 1Hypercorrection (without errors) 5 50% 11 6 75% 18 1 50% 2 0 0% 1 1 14% 1 1 33% 1

Oral language 10 100% 48 6 75% 83 2 100% 14 2 100% 15 6 86% 28 3 100% 6

26

Note. n = total number of candidates in a race/language use group; Np=number of candidates in a race/language use group who exhibited the language feature one or more times; %=percentage of candidates out of all candidates in a race/language use group; Ni=number of instances that a language feature was used by candidates in a race/language use group.

Page 35: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Referring back to Table 6 for the individual code level data, in keeping with the overall

findings, the use of AAL noun features clustered among African American candidates, both high

and low scoring, and among European American users of Southern White English. High-scoring

European American candidates who used Edited American English showed no evidence of use of

features that could be categorized as AAL. Specific findings include the following: when indicating

plurality of nouns, the use of the AAL pattern of dropped “–s” was found among more than half

of the African American candidates, both high- and low-scoring. When indicating possession,

however, the AAL feature of dropped apostrophe/dropped “–s” was rarely seen—used by only

one African American candidate in each group (high and low-scoring). Speech Code Errors of

pluralization with an unnecessary apostrophe and possession with no/incorrect apostrophe were

seen among African American candidates, European American candidates who showed SWE

patterns, and low-scoring European American candidates who used Edited American English. In

only one instance did a high-scoring European American candidate exhibit use of SCE in nouns.

Nouns

Table 6 presented data for five separate codes for plural and possessive nouns. Tables 7

and 8 below group multiple codes together, showing the number of instances of AAL codes

(plural nouns with a dropped “–s” and possessive nouns with a dropped “–s” and no apostrophe)

and SCE codes (plural nouns with an unnecessary apostrophe, and possessive nouns with no

apostrophe or an incorrect apostrophe) according to entry score and race/language use classification.

For each group of candidates, the table presents the total number of instances that the codes were

used, the number of candidates who exhibited the features, and the range of instances per candidate.

27

Page 36: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Table 7

Instances of AAL Noun Features

Candidates Entry score

Number of instances of AAL codes in nouns

Number in group

Range

< 2.75 15 7 of 10 1-4 African American ≥ 2.75 25 5 of 8 1-13

< 2.75 15 2 of 2 4, 11 European American

– SWE ≥ 2.75 10 2 of 2 2, 8

< 2.75 9 4 of 7 1-4 European American ≥ 2.75 0 0 of 3 –

Table 8

Instances of SCE Noun Features

Candidates Entry score

Number of instances of SCE codes in nouns

Number in group

Range

< 2.75 20 7 of 10 1-7 African American ≥ 2.75 30 8 of 8 1-10

< 2.75 12 2 of 2 4, 8 European American

– SWE ≥ 2.75 13 2 of 2 1, 12

< 2.75 11 4 of 7 1-8 European American ≥ 2.75 1 1 of 3 1

The European American users of SWE and the African American candidates both used

multiple examples of AAL features as they relate to possessive and plural nouns; only one feature in

this group of codes was used once by a high-scoring European American candidate. The pattern

of AAL features does not strongly cluster by high or low entry scores for African American. The

consistent patterned use of the dropped “–s” for plural nouns by the four European American

candidates contributed to their ultimate identification as users of SWE.

SCE are writing features that neither conform to the grammatical structure of Edited

American English, nor of African American Language. Both African American and European

28

Page 37: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

American candidates exhibit such features. However, the rate at which the European American

candidates display these features is lower than for the African American candidates.

To measure how frequently candidates use non-edited American English (a combination of AAL

and SCE) noun forms, the data were collapsed across all five noun features. Table 9 shows the

number of instances of combined noun codes according to entry score and race classification.

Table 9

Instances of Non-EAE Noun Features (Collapsed Across AAL and SCE)

Candidates Entry score

Number of instances of non-EAE nouns codes

Number in group

Range

< 2.75 35 8 of 10 1-7 African American ≥ 2.75 55 8 of 8 1-18

< 2.75 27 2 of 2 8, 19 European American

– SWE ≥ 2.75 23 2 of 2 3, 20

< 2.75 20 4 of 7 2-12 European American ≥ 2.75 1 1 of 3 1

Combining all the codes that exhibit non-edited American English usage for plural and

possessive nouns shows some clustering of features among African American candidates in

general and in particular among those candidates with high entry scores.

Two African American and seven European American candidates had no non-EAE noun

features. Of the remaining 23 who displayed some of these features, the majority had fewer than

10 instances each of non-EAE noun usage in their written commentaries. The five candidates who

had 10 or more instances of non-EAE features included two high-scoring African American

candidates, one high-scoring and one low-scoring European American candidate who demonstrated

use of SWE, and one low-scoring European American candidate who used Edited American English.

Overall, there were consistent patterns of non-Edited American English usage in these

written commentaries for the majority of African American candidates, and a small group of the

low-scoring European American candidates who used predominantly Edited American English.

Similar patterns are seen in the usage of non-EAE verb forms.

29

Page 38: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Verbs

Similar to patterns observed for nouns above, the presence of AAL and SCE among verbs

was low across all written commentaries. The patterned use of the dropped “–ed” in regular past

tense verbs clustered among African American candidates, with half of the low-scoring candidates

demonstrating some use of the dropped “–ed,” and one-fourth of the high-scoring candidates

demonstrating use of dropped “–ed.” The consistent patterned use of the dropped “–ed” by the

four European American candidates contributed to their identification as users of SWE. No

evidence of dropped “–ed” in verbs was seen among European American candidates who used

Edited American English. These results are summarized in Table 10 below.

Table 10

Instances of Dropped “–ed” in Past Tense Verbs

Candidates Entry score

Number of instances of AAL codes in verbs

Number in group

Range

< 2.75 9 5 of 10 1-4 African American ≥ 2.75 3 2 of 8 1, 2

< 2.75 3 2 of 2 1, 2 European American

– SWE ≥ 2.75 7 2 of 2 3, 4

< 2.75 0 0 of 7 – European American ≥ 2.75 0 0 of 3 –

It is of interest to note that for the African American candidates, the use of the AAL verb

forms was much less frequent that the use of AAL noun forms. Sixteen of the 18 African

American candidates used AAL noun forms compared to seven of 18 for the verbs.

The use of non-Edited American English forms of irregular past tense verbs was lower

than the instances of dropped “–ed.” The pattern of usage differed somewhat from the AAL

usage. In this case, speech code errors among irregular verb forms were actually more common

(although still very low) among the European American candidates—four out of 10 have at least

one SCE compared to two out of the 18 African American candidates. These results are

summarized in Table 11.

30

Page 39: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Table 11

Instances of SCE in Irregular Past Tense Verbs

Candidates Entry score

Number of instances of SCE codes in verbs

Number in group

Range

< 2.75 2 2 of 10 1 African American ≥ 2.75 0 0 of 8 –

< 2.75 2 1 of 2 2 European American

– SWE ≥ 2.75 4 1 of 2 4

< 2.75 5 3 of 7 1-2 European American ≥ 2.75 1 1 of 3 1

Although Tables 10 and 11 distinguish between regular and irregular verb forms

and between AAL and SCE instances in each, an assessor reading a candidate response

may not make such linguistic distinctions, but rather perceive them all as errors. Table 12

below shows the combined total number of instances of AAL codes (dropped “–ed” for regular

verbs in simple past tenses, present/past/future perfect tenses, verbal adjectives and passive

voice) according to entry score and race/language use classification. For each group of

candidates, the table presents the total number of instances that the codes were used, the number

of candidates who exhibited the features and the range of instances per candidate.

Table 12

Instances of Non-EAE Form (AAL and SCE) of Regular and Irregular Past Tense Verbs Candidates Entry

score Number of instances of non-EAE codes

in verbs

Number in group

Range

< 2.75 11 5 of 10 1-5 African American ≥ 2.75 3 2 of 8 1, 2

< 2.75 5 2 of 2 1, 4 European American

– SWE ≥ 2.75 11 2 of 2 3, 8

< 2.75 5 2 of 7 1, 4 European American ≥ 2.75 1 1 of 3 1

31

Page 40: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

When indicating past tense in the written commentaries, low-scoring African American

candidates exhibited more non-EAE verb usage in their responses than high-scoring African

American candidates—one-half of the low-scoring African American candidates versus one-

fourth of the high-scoring African American candidates.

Dropped “–s” in third person singular present tense verbs. An additional verb form that

was coded on the basis of being a common occurrence in spoken AAL was the presence of a

dropped “–s” in third person singular present tense for regular verbs. In addition, non-EAE forms

of the third person singular present tense for irregular verbs was coded; only one African

American candidate had an instance of this form. The use of the dropped “–s” in AAL verb

conjugation was found predominantly among African American candidates—seven 7 of the 10

low-scoring candidates and three of the eight high-scoring candidates. All four of the candidates,

who were identified as users of SWE, had at least one instance of the dropped “–s” features in

verbs. No European American candidates who used Edited American English showed the

dropped “–s” feature in their written commentaries. Table 13 below provides the breakdown by

race/language use and score for the regular third person singular present tense AAL forms that

were coded.

Table 13

Instances of Dropped “–s” in the Third Person Singular Present Tense of Regular Verbs Candidates Entry

score Instances of AAL form of regular third person singular present tense

Number in group

Range

< 2.75 10 7 of 10 1-3 African American ≥ 2.75 9 3 of 8 1-6

< 2.75 4 2 of 2 1, 3 European American

– SWE ≥ 2.75 2 2 of 2 1, 1

< 2.75 0 0 of 7 – European American ≥ 2.75 0 0 of 3 –

Subject-verb discord. Two codes were created to identify subject-verb discords—in

copula and non-copula verb forms. These features were of interest because it is a pattern of AAL

to identify number from context; adding endings to the verb to indicate number would be

32

Page 41: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

redundant. However, it is also a common Speech Code Error to produce subject/verb discords

particularly when the verb is separated from the subject by multiple prepositional phrases. For

the purposes of this study, both forms of subject-verb discord were coded together, since the

assessor would not be likely to make linguistic distinctions between the two types. In keeping

with earlier patterns observed for both nouns and verbs, the use of this feature clustered mostly

among African American candidates, both high- and low-scoring. Table 14 provides the number

of instances of both forms of discord.

Table 14

Subject-Verb Discord

Candidates Entry score

Number of instances

Number in group Range

< 2.75 18 8 of 10 1-3 African American ≥ 2.75 11 4 of 8 1-6

< 2.75 0 0 of 2 – European American

– SWE ≥ 2.75 2 2 of 2 1

< 2.75 1 1 of 7 1 European American ≥ 2.75 2 1 of 3 2

Other irregular verb forms. A final code that was included was for “other non-EAE verb

forms”—that is, other irregular uses of verbs that were not accounted for by any of the earlier

codes. These included speech code errors such as incorrect verb endings, “without given,” and

miscombined tenses, such as “has being.” No evidence exists in the linguistic literature to suggest

that these verb usages pattern as part of AAL, and so they are designated as SCE. Since this

category of SCE includes a wide range of possible entries, the results show a similarly dispersed

distribution of usage patterns. In each group of candidates, only one or two individuals included

any instances of “other non-EAE verb forms” in their commentaries. Table 15 (below) shows the

instances of these verb forms. For each group of candidates, the table presents the total number

of instances that the codes were used, the number of candidates who exhibited the features, and

the range of instances per candidate.

33

Page 42: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Table 15

Other Non-EAE Verb Forms

Candidates Entry score

Instances of other non-EAE verb forms

Number in group

Range

< 2.75 2 2 of 10 1 African American ≥ 2.75 5 2 of 8 2, 3

< 2.75 5 1 of 2 5 European American

– SWE ≥ 2.75 4 1 of 2 4

< 2.75 2 2 of 7 1 European American ≥ 2.75 1 1 of 3 1

In reviewing overall use of AAL verb forms, instances were few, but consistent

clustering patterns were seen among African American candidates. Both African American and

European American candidates displayed forms of verb usage categorized as Speech Code Errors.

Lexical Features: Morphemes, Prepositions, and Connectors

Five codes were included in the category of lexical features: dropped final letters; other

spelling errors; missing or incorrect prepositions; other missing words; and incorrect use of

connecting words. For each candidate group, Table 16 on the next page presents the total number

of instances that each code was used and the number of candidates who exhibited the feature.

34

Page 43: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Table 16

Presence of Missing or Incorrect Morphemes, Prepositions, and Connectors

African American European American-SWE European American

< 2.75 ≥ 2.75 < 2.75 ≥ 2.75 < 2.75 ≥ 2.75

NI NP NI NP NI NP NI NP NI NP NI NP

Dropped final letters 7 6 of 10 6 4 of 8 2 1 of 2 0 0 of 2 0 0 of 7 0 0 of 3

Other spelling errors 19 8 of 10 16 4 of 8 21 2 of 2 15 2 of 2 10 4 of 7 2 2 of 3

Missing/incorrect preposition 11 6 of 10 13 6 of 8 3 2 of 2 7 2 of 2 6 3 of 7 1 1 of 3

Other missing words 11 7 of 10 16 4 of 8 5 2 of 2 4 2 of 2 11 5 of 7 2 2 of 3

Incorrect use of connecting words 8 5 of 10 2 1 of 8 0 0 of 2 0 0 of 2 1 1 of 7 0 0 of 3

Note. NI = number of instances that a language feature was used; NP = number of candidates in the group who exhibited the language feature one or more times.

The pattern of dropping final letters from words in spoken AAL is identified in the

linguistic literature as an additional form of reducing the language to its essential communicative

elements. It was therefore anticipated that this reductive pattern would be specific to users of

AAL, if it appeared at all in written form. On the other hand, both spelling errors and missing

words could be expected to appear evenly distributed across all racial groups.

35

Page 44: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

As seen in Table 16 on the previous page, the percentage of candidates with both of these

SCE is fairly constant across groups. The incorrect use of connecting words is found

predominantly in the African American candidates’ writing, and the missing or incorrect

prepositions also seem to occur more strongly among African American candidates. Neither of

these two features has been identified in the AAL literature, but the patterned presence of these

features in this data set may warrant future study.

Table 17 combines the two SCEs (spelling errors and other missing words) together and

the two other codes (missing/incorrect prepositions, incorrect connectors).

Table 17

Instances of Lexical Features

Spelling errors and other missing words

Missing/incorrect prepositions, incorrect connectors

Candidates Entry score

Instances Number in group

Range Instances Number in group

Range

< 2.75 30 10 of 10 1-5 19 7 of 10 1-3 African American ≥ 2.75 32 5 of 8 1-18 15 5 of 8 1-8

< 2.75 26 2 of 2 10, 16 3 2 of 2 1, 2 European American

– SWE ≥ 2.75 19 2 of 2 7, 12 7 2 of 2 3, 4

< 2.75 21 5 of 7 1-7 7 3 of 7 1-3 European American ≥ 2.75 4 3 of 3 1-2 1 1 of 3 1

Combined Non-edited American English Usages at the Lexical Level

While losing the specific characteristics of the individual codes, it is useful to combine

codes across all categories of nouns, verbs, and sentence level features, to get a measure of the

overall effect of non-EAE instances within a given set of written commentaries. Table 18

presents the codes that were combined for the purposes of this analysis:

36

Page 45: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Table 18

AAL and SCE Codes Combined

AAL codes SCE codes Dropped “–s” in plural nouns Unnecessary apostrophe in plurals Dropped “–s,” dropped apostrophe in possessive

No apostrophe in possessives

Regular verb, simple past tense, dropped “–ed”

Incorrect apostrophe in possessives

Regular verb, present/past/future perfect tense, dropped “–ed”

Irregular verb, simple past tense, non-EAE form

Regular verbal adjective, dropped “–ed” Irregular verb, present/past/future perfect tense, non-EAE form

Regular passive voice, dropped “–ed” Irregular verbal adjective, non-EAE form Regular third person singular present tense, dropped “–s”

Irregular passive voice, non-EAE form

Zero copular Irregular third person singular present tense, non-EAE form

Aspectual or invariant be Conjugated be, non-EAE form Dropped final letters Subject-verb inversion Subject-verb discord Other non-EAE verb forms Other spelling errors Missing or incorrect preposition Other missing words Incorrect use of connecting words Noun-pronoun discord Pronominal apposition Homophones

While many of the codes had low occurrence rates, combined with others, they contribute

to the reader’s overall impression. Several of these, specifically noun-pronoun discord,

pronominal apposition, homophones, non-EAE forms of copula and subject-verb inversion were

not discussed individually above, due to their relative scarcity in the data pool. Tables 19 and 20

present total number of instances of AAL and SCE features. Only features that were fully

identified in the literature as aspects of AAL were included in the composite AAL code. Table

37

Page 46: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

21 presents the total number of AAL and SCE features across all the written commentaries, with

range and means (instances/number in group).

Table 19

Total Number of Instances of AAL Features

Candidates Entry score

Number of instances of AAL codes

Number in group

Range Mean

< 2.75 43 10 of 10 1-9 4 African American ≥ 2.75 46 8 of 8 1-24 6

< 2.75 24 2 of 2 8, 16 12 European American

– SWE ≥ 2.75 19 2 of 2 6, 13 10

< 2.75 13 4 of 7 2-4 3 European American ≥ 2.75 0 3 of 3 – –

Table 20

Total Number of Instances of SCE Features

Candidates Entry score

Number of instances of SCE codes

Number in group

Range Mean

< 2.75 101 10 of 10 4-21 10 African American ≥ 2.75 106 8 of 8 2-40 13

< 2.75 50 2 of 2 22, 28 25 European American

– SWE ≥ 2.75 56 2 of 2 13, 43 28

< 2.75 50 5 of 7 2-17 10 European American ≥ 2.75 10 3 of 3 2-5 3

38

Page 47: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Table 21

Total Number of Instances of Non-EAE (AAL and SCE) Features

Candidates Entry score

Number of instances of all non-EAE codes

Number in group

Range Mean

< 2.75 144 10 of 10 5-24 14 African American ≥ 2.75 152 8 of 8 2-64 19

< 2.75 74 2 of 2 30, 44 37 European American

– SWE ≥ 2.75 75 2 of 2 19,56 38

< 2.75 63 5 of 7 2-21 13 European American ≥ 2.75 10 3 of 3 2-5 3

Given the results from the earlier tables, it is not surprising on review of the composite

results to note that there is clustering of codes for the African American candidates. The

frequency at which they use both AAL and SCE features is greater than for the European

American candidates. It is important to note that clustering is not associated with high or low

scoring responses.

Hypercorrection and Oral Language

The expert panel identified three discourse-level features: hypercorrection (both with and

without error) and oral language. For each candidate group, Table 22 presents the total number

of instances that each code was used and the number of candidates who exhibited the feature.

39

Page 48: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Table 22

Presence of Hypercorrection and Oral Language

African American European American-SWE European American < 2.75 ≥ 2.75 < 2.75 ≥ 2.75 < 2.75 ≥ 2.75

NI NP NI NP NI NP NI NP NI NP NI NP Hypercorrection With error 2 1 of 10 12 7 of 8 7 2 of 2 1 1 of 2 1 1 of 7 1 1 of 3Without error 11 5 of 10 18 6 of 8 2 1 of 2 0 0 of 2 1 1 of 7 1 1 of 3 Oral language 48 10 of 10 83 6 of 8 14 2 of 2 15 2 of 2 28 6 of 7 6 3 of 3

Note. NI = number of instances that a language feature was used; NP = number of candidates in the group who exhibited the language feature one or more times.

Although instances of hypercorrection with error were very few, seven of the eight high-

scoring African American candidates had one or more instances of this feature. The related

feature of being overly prescriptive (hypercorrection without errors) also clustered strongly in the

group of African American candidates, both high and low scoring. Similarly, the use of phrases

and sentences that were appropriate in oral but not written format clustered strongly in the

African American candidate group. Of the 32 candidates, both African American and European

American, five had more than 10 instances of this feature; four of those were high-scoring

African American candidates.

The use of oral language is a possible indication that the individual is an inexperienced

writer. The patterns of oral language use closely mirror the total number of noun, verb, and

discord errors presented earlier in the analysis (a correlation of 0.68). The two candidates with

the greatest number of oral language instances were ranked first and third for greatest number of

grammatical errors—both high scoring African American candidates.

In all of the analyses in this section, candidates have been divided into those who scored

equal to or above 2.75 and those who scored below 2.75, following the NBPTS policy on the re-

submission of specific entries. It would have been preferable to have had additional candidates in

each group whose scores did not lie close to the 2.75 cut-off. However, given the limited number

of written commentaries available, it was deemed inappropriate to divide up the responses

further beyond the four groups (high and low scoring African American and European American

40

Page 49: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

candidates). While the number of candidates in the data set is low, it is instructive to briefly

examine those African American candidates at the two extremes of the score scale to consider

whether the patterns seen for the entire group hold. There are four African American candidates

who scored 3.625 or above (averaged across two assessors) and four African American

candidates who scored 1.75 or below (averaged across two assessors).

The average number of AAL instances for all African American candidates was five,

with an average number of SCE of 12. Considering the two extreme groups separately, the four

candidates at each extreme of the score scale had a slightly lower average of both AAL features

and SCE. Examining the group of 18 African American candidates by high and low scores has

suggested that the presence of non-EAE forms of writing does not cluster by entry score. Even

when including only the most extreme cases (lowest and highest scores), the candidates do not

look particularly different when considering patterns of AAL usage and SCE, thereby lending

support to the conclusion that score and language are not strongly related.

Discussion and Conclusions

This study provides new insight to the research on African American Language, by

focusing on written usage, rather than spoken. It specifically examines the presence of AAL in

the context of a formal, academic development activity for teachers—a portfolio entry in the

National Board for Professional Teaching Standards assessments. This research contributes to

the continuing debate over what factors lead to the significant differential performance seen

between African American and European American candidates in all forms of assessments by

examining writing differences. The study lays the foundation for examining the relationship

between AAL and differential performance outcomes. This is done by providing evidence for the

existence of AAL in a context where it has the potential to affect assessors’ judgments, in the

form of construct-irrelevant variance in the portfolio scoring process.

This study examined a small sample of portfolio entries written by African American and

European American candidates. Evidence was found for the limited use of AAL features by a

subgroup of African American candidates, which aligns with Smitherman’s research and others:

up to 90% of African Americans use aspects of AAL some of the time. A pattern of features

similar to AAL was also found among a subgroup of European American candidates, who

identified themselves as being from the Southeastern United States; it was postulated that this

41

Page 50: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

might be evidence of Southern White English use. The European American candidates who use

American English (AE) tend not to have employed any AAL features. All candidates, to varying

degrees, demonstrated evidence of Speech Code Errors. High scoring European American

candidates who used American English had the fewest Speech Code Errors.

Evidence for the use of AAL features was found among African American candidates at

all score levels, which lends some support to the effectiveness of the NBPTS bias-reduction

exercises conducted with assessors before scoring. That is, the existence of AAL features at all

score levels suggests that assessors may have experienced some success in screening out personal

prejudices against this writing form. The same could be said for the evidence that SCE were

found at all score levels. The NBPTS assessor training explicitly addresses the need for assessors

to identify and monitor writing biases, and not to factor writing style into their score judgments.

The study described here did not attempt to take the race of the assessor into

consideration, nor were judgments made about the quality of each candidate’s work. Rather the

NBPTS assigned scores were taken as given. In other words, attempts were not made to identify

whether there were responses awarded low scores in spite of high quality work or responses

awarded high scores in spite of deficiencies. If AAL usage had been concentrated in the low-

scoring responses, then the content of the responses would have been examined as well.

However, research with English as a second language and various American dialects, and

some limited studies on AAL, indicate a real potential for negative evaluations by readers of

texts containing non-Edited American English features. Given that both high- and low-scoring

African American responses demonstrated evidence of AAL usage, it is reasonable to ask

whether the presence of AAL features has a small negative impact on assessors’ judgments

across the full spectrum of the score range, causing a possible dampening effect at all score

levels. This also might hold true for the presence of SCE, which has been shown to cause reader

irritation. For the cohort of candidates from which the data set came, calculating the mean scores

on this entry for African American and European American candidates, approximately a half-

point difference is observed between the two groups. This is consistent with results from

previous cohorts. It is possible that some of this differential performance could be due to the

construct-irrelevant effect of the writing features used.

A follow-up study has been proposed to explore this possibility. Selected written

commentaries that display African American Language features would be “re-written” to employ

42

Page 51: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

only Edited American English. Similarly, selected commentaries written in EAE would be re-

written with the help of linguists experienced in the production of AAL, to employ some African

American Language features. These candidate responses would be re-scored by trained

assessors, who would not be aware of the changes made to the texts. Scores from the altered

responses would be compared to the scores earned for each original response.

Additional questions stemming from the results of this study include why the occurrence

of SCE is less frequent for the European American candidates than for African American

candidates. One possible hypothesis is that greater training and experience in formal academic

writing for European American candidates could be a contributing factor. It has been

hypothesized in the literature that a large number of African American (and Latino) children may

be more likely to receive a less adequate K-12 education, given that many are more likely to

attend under-funded urban and rural schools and may be more likely to be taught by teachers

teaching outside their certification area (National Center for Educational Statistics, 1999).

To further explore the hypothesis that the African American candidates who displayed

more non-EAE language features had less experience in formal academic writing, a reasonable

correlate was sought among the limited background data available. Candidates provide

information on highest degree earned. An assumption could be made that candidates with a

Bachelors degree would have had less formal, academic writing experience than candidates who

had a Masters degree or higher. Furthermore, it was hypothesized that those candidates with only

Bachelors degrees would be the ones with greater non-EAE features in this particular entry.

However, when the data were examined, twelve of the 18 African American candidates have a

Masters level or higher qualification, compared to seven of the 14 European American

candidates. In particular, for the African American candidates who had high scores on this

particular entry, and yet who had significant numbers of SCE, seven out of the eight have

Masters degrees.

Looking at highest degree earned, therefore, as a measure of formal writing experience

does not explain the data. Another avenue for exploration could involve the effect of learning

two languages (American English and African American Language or Southern White English),

mediated by access to high quality language education and consistent opportunities for practice

in both target languages. More background data would need to be collected from candidates to

pursue this avenue.

43

Page 52: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

There also may be an interaction between audience and purpose of the writing and the

presence of non-EAE features. The pressure of preparing the portfolio responses may result in

increased language production errors and decreased attention to the elimination of such errors.

The process of language production becomes more complex for bidialectal speakers. A further

avenue for study might be to examine candidates’ production of academic writing under less

pressured circumstances.

A final consideration in understanding the results involves the demands on candidates’

time versus formal support for preparing the portfolio. In other words, those candidates who have

access to district and school support, mentors, or other National Board certified teachers may

perform better than those who do not have such support. The limited background data available

in this study did not allow for exploration of this hypothesis.

Differential performance for African American test takers has been documented across

all assessment types for several decades. “There are urgent needs in this country for a systematic

cultural-linguistic review of testing and assessment devices to be used with African-

American[s]” (Hilliard, 2002). While the differential performance gap is slowly decreasing,

more work must be done to uncover sources of this difference and to further reduce their effects.

This includes including research into areas of language use, cultural dissonance between

candidates and assessors, educational preparedness of candidates, access to preparation courses

and mentors, types of schools and student bodies where teachers practice, and the stereotype

threat phenomenon.

44

Page 53: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

References

Albrechtson, D., Henriksen, B., & Faerch, C. (1980). Native speaker reactions to learners’

spoken interlanguage. Language-Learning, 30(2), 365-396.

Anderson, E. (1981). Language and success. College English, 43, 807-817

Bailey, G. (1997). Southern American English: A prospective. In C. Bernstein, T. Nunnally, &

R. Sabino (Eds.), Language variety in the South revisited (pp. 21-32). Tuscaloosa: The

University of Alabama Press.

Ball, A. F. (1992). Cultural preference and the expository writing of African American students.

Written Communication, 9(4), 501-532.

Ball, A. F. (1995). Text design patterns in the writing of urban African American students:

Teaching to the cultural strengths of students in multicultural settings. Urban Education

30(3), 253-289.

Ball, A. F. (1997). Expanding the dialogue on culture as a critical component when assessing

writing. Assessing Writing, 4(2), 169-202.

Ball, A. F. (1998). Evaluating the writing of culturally and linguistically diverse students: The

case of the African American vernacular English speaker. In C. R. Cooper & L. Odell

(Eds.), Evaluating writing: The role of teachers’ knowledge about text, learning, and

culture (pp. 225-248). Urbana, IL: National Council of Teachers of English Press.

Baugh, J. (1983). Black street speech: Its history, structure and survival. Austin: University of

Texas Press.

Benson, J., & Impara, J. (1987). Content validation of the National Board for Professional

Teaching Standards, Middle Childhood/Generalist, July 1997.

Blumenthal, A., & Hildenbrand, J. (1993, December). Tone and role: From “participant” stance

to the “spectator” stance in the writing of African American college composition

students (ERIC Document Reproduction Service No. ED374418). Paper presented at the

Annual Meeting of the Modern Language Association..

Bond, L. (1998a). Culturally responsive pedagogy and the assessment of accomplished teaching.

Journal of Negro Education, 67(3), 242-254.

Bond, L. (1998b). Disparate impact and teacher certification. Journal of Personnel Evaluation in

Education, 12(2), 211-220.

45

Page 54: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Bond, L., Smith, T., Baker, W. K, & Hattie, J. (2000). The certification system of the National

Board for Professional Teaching Standards: A construct and consequential validity

study. Center for Educational Research and Evaluation, University of North Carolina at

Greensboro.

Botan, C., & Smitherman, G. (1991). Black English in the integrated workplace. Journal of

Black Studies, 22(2), 168-186.

Burling, R. (1973). English in black and white. New York: Holt, Rinehart & Winston.

Cassidy, F. G., & Hall, J. H. (1996). Dictionary of American Regional English. Cambridge, MA:

Belknap Press of Harvard University Press.

Chaplin, M. (1987). An analysis of writing features found in the written commentaries of

students in the National Assessment of Educational Progress and the New Jersey high

school proficiency test. Unpublished manuscript, Rutgers University, Department of

English, Camden.

Erickson, F. (1984). Rhetoric, anecdote, and rhapsody: Coherence strategies in a conversation

among African-American American adolescents. In D. Tannen (Ed.), Coherence in

spoken and written discourse, (pp. 81-154). Norwood, NJ: Ablex.

Feagin, C. (1997). The African contribution to southern states English. In C. Bernstein, T.

Nunnally, and R. Sabino (Eds.), Language variety in the South revisited (123-139),

Tuscaloosa: The University of Alabama Press.

Fogel, H., & Ehri, L. C. (2000). Teaching elementary students who speak Black English

Vernacular to write in standard English: Effects of dialect transformation practice.

Contemporary Educational Psychology, 23, 232-235

Gatewood, R., & Perloff, R. (1990). Testing and industrial application. In G. Goldstein & M.

Hersen (Eds.), Handbook of psychological assessment, (pp. 486-501). New York:

Pergamon Press.

Hartigan, J. A., & Wigdor, A. K. (Eds.). (1989). Fairness in employment testing: Validity,

generalization, minority issues and the general aptitude test battery. Washington, D.C.:

National Academy Press.

Hatch, E., & Lazaraton, A. (1991). The research manual: Design and statistics for applied

linguistics. Boston: Heinle & Heinle.

46

Page 55: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Heath, S. B. (1983/1994). Ways with words: Language, life, and work in communities and

classrooms. Cambridge, UK: Cambridge University Press.

Hilliard, A. G., III. (2002). Language, culture, and the assessment of African American children.

In L. Delpit & J.K. Dowdy (Eds.), The skin that we speak, (pp. 87-106). New York: The

New Press.

Irvine, J. J., & Fraser, J. W. (1998, May 13). Warm demanders. Education Week.

Jencks, C., & Phillips, M. (Eds.). (1998). The Black-White test score gap. Washington, DC:

Brookings Institution Press.

Jones, L. V. (1984). White-Black achievement differences: The narrowing gap. American

Psychologist, 39, 1207-1213.

Kochman, T. (1981). Black and White styles in conflict. Chicago: University of Chicago Press.

Kroll, B. (1991). Teaching writing in the ESL context. In M. Celce-Murcia (Ed.), Teaching

English as a second or foreign language (2nd ed., pp. 245-263). New York: Newbury

House.

Labov, W. (1972). Language in the inner city: Studies in the Black English vernacular.

Philadelphia: University of Pennsylvania Press.

Ladson-Billings, G. (1998). Teaching in dangerous times: Culturally relevant approaches to

teacher assessment. Journal of Negro Education, 67(3), 255-267.

Lam, T. C. M. (1995). Fairness in performance assessment, ERIC Digest (ERIC Document

Reproduction Service No. ED391982). Retrieved September 2003, from

http://searcheric.org/digests/ed391982.html

Lee, J. (2002). Racial and ethnic achievement gap trends: Reversing the process toward equity.

Educational Researcher, 31(1), 3-12.

Leki, I. (1991). The preferences of ESL students for error correction in college-level writing

classes. Foreign Language Annals, 24, 203-218.

Linguistic Society of America. (1997, January). LSA resolution on the Oakland “Ebonics” issue.

Retrieved September 2003, from http://www.lsadc.org/web2/resolutionsfr.htm

McMillan, J. B., & Montgomery, M. B. (1989). Annotated bibliography of Southern American

English. Tuscaloosa: University of Alabama Press.

47

Page 56: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

National Board for Professional Teaching Standards. (2003, April). Candidate resource center:

Score banking & retakes Q&A. Retrieved September 2003, from

http://www.nbpts.org/candidates/scorebankretake.cfm

National Center for Educational Statistics. (1999). The condition of education. Washington,

D.C.: U. S. Department of Education, Office of Educational Research and Improvement.

Ogbu, J. U. (2002). Cultural amplifiers of intelligence: IQ and minority status in cross-cultural

perspective. In J. M. Fish, (Ed.), Race and intelligence: Separating science from myth,

(pp. 241-278). Mahwah, NJ: Lawrence Erlbaum Associates.

Onafowora, L. (1998). Measurement consequences reconsidered: African American

performance and communication styles in the assessment for National Teacher

Certification. Unpublished doctoral dissertation, The University of North Carolina at

Greensboro.

Oxford English Dictionary (2nd ed.). (2003). New York: Oxford University Press. Retrieved

August 2003, from http://dictionary.oed.com

Richards, J. C., Platt, J.,& Platt, H. (1993). Longman Dictionary of language teaching and

applied linguistics (2nd ed.). Singapore: Longman Singapore Publishers.

Rickford, J. R. (1999). African American Vernacular English. Malden, MA: Blackwell.

Rickford, J. R., & Rickford, R. J. (2000). Spoken soul: The story of Black English. New York:

John Wiley & Sons.

Santos, T. (1988, March). Professors’ reactions to the academic writing of nonnative-speaking

students. TESOL Quarterly, 22(1), 69-90.

Schumann, J. H. (1978). The pidginization process: A model for second language acquisition.

Rowley: Newbury House.

Smitherman, G. N. (1973, May-June). White English in blackface or, who do I be? The Black

Scholar, 32-39.

Smitherman, G. N. (1977). “How I got ovuh:” African world view and Afro-American oral

tradition. In Talkin that talk: Language culture and education in African America. NY:

Routledge.

Smitherman, G. N. (1992). Black English, diverging or converging? The view from the National

Assessment of Education Progress. Language Problems and Language Planning, 6(1).

48

Page 57: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Smitherman, G. N. (1994). The blacker the berry the sweeter the juice: African American student

writers. In A. H. Dyson & C. Genishi (Eds.), The need for story: Cultural diversity in

classroom and community (pp. 80-101). Urbana, IL: National Council of Teachers of

English.

Smitherman, G. N. (1997a). Black language and the education of Black children: One mo once.

Black Scholar, 27(1), 28-36.

Smitherman, G. N. (1997b). The chain remain the same. Journal of Black Studies, 28(1).

Smitherman, G. N. (2000). Introduction to ebonics. In Talkin that talk: Language, culture and

education in African America. NY: Routledge.

Steele, C. (1997). A threat in the air: How stereotypes shape intellectual identity and

performance. American Psychologist, 52(6), 613-629.

Szpara, M. Y. & Wylie, E. C. (2003). Differences in teacher performance assessments: Evidence

of African American teaching styles. Manuscript in preparation.

Whitaker, C. (1989, January). The disappearing Black teacher. Ebony, pp. 122, 126.

Wolfman, W. A., & Whiteman, M. F. (1971, Fall). The role of dialect interference in

composition. Florida FL Reporter, 9(1/2), 34-38, 59.

49

Page 58: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Notes 1. Creolized is used here as the linguistic term meaning “a language that has developed from that

of a dominant group, first being used as a second language, then becoming the usual language

of a subject group, its sounds, grammar, and vocabulary being modified in the process”

(Oxford English Dictionary, 2003).

50

Page 59: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Appendix

Subset of Coding Frame

Table A1

Nouns

1A noun, regular plural, “–s” has been dropped [N – s]

1B noun, regular plural “–s” is used [N + s]

1C noun, regular plural, unnecessary apostrophe is added [N + ‘s]

2A noun, singular or plural, possessive, both “–s” and apostrophe have been dropped [N – ‘s] [N – s’]

2B noun, singular or plural, possessive, “–s” is used but apostrophe has been dropped [N + s] [N + s]

2C noun, singular or plural, possessive, “–s” is used but apostrophe has been placed incorrectly according to EAE rules [N + s] [N + s]

2D noun, singular or plural, possessive “–’s” or “–s’” is used according to EAE rules [N + ‘s] [N + s’]

3 Noun/pronoun discord [the pronoun does not agree in number with its corresponding noun]

4 Pronominal apposition (a noun or noun phrase placed with a pronoun as an explanatory equivalent, both functioning in the same role)

5 Homophonic form substitution (words that sound the same but are spelled differently)

51

Page 60: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Table A2

Verbs

6A regular verb, simple past tense, “–ed” ending has been dropped [V – ed]

6B regular verb, simple past tense, “–ed” ending used [V + ed]

6C irregular verb, simple past tense, EAE form is not used [V – EAE irregular form]

6D irregular verb, simple past tense, EAE form is used [V + EAE irregular form]

7A regular verb, present/past/future perfect tense “–ed” ending dropped [have + V – ed], [had + V – ed] [will have + V – ed]

7B regular verb, present/past/future perfect tense, “–ed” ending used [have + V + ed] [had + V + ed] [will have + V + ed]

7C irregular verb, present/past/future perfect tense, EAE form is not used [have + V – EAE irregular form] [had + V – EAE irregular form] [will have + V – EAE irregular form]

7D irregular verb, present/past/future perfect tense, EAE form is used [[have + V + EAE irregular form] [had + V + EAE irregular form] [will have + V + EAE irregular form]

8A regular verbal adjective (verb in participle form used as an adjective) “–ed” ending has been dropped [V – ed]

8B regular verbal adjective (verb in participle form used as an adjective) -ed ending used [V + ed]

8C irregular verbal adjective (verb in participle form used as an adjective) EAE form is not used [V – EAE irregular form]

(Table continues)

52

Page 61: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Table A2 (continued) 8D irregular verbal adjective

(verb in participle form used as an adjective) EAE form is used [V + EAE irregular form]

9A regular verb, passive voice -ed ending has been dropped [be + V – ed] [Note: use feature 12 to code the copula “be”]

9B regular verb, passive voice “–ed” ending is used [be + V + ed] [Note: use feature 12 to code the copula “be”]

9C irregular verb, passive voice EAE form is not used [be + V – EAE irregular form] [Note: use feature 12 to code the copula “be”]

9D irregular verb, passive voice EAE form is used [be + V + EAE irregular form] [Note: use feature 12 to code the copula “be”]

10A regular verb, present tense, third person singular “–s” ending is not used [V – s]

10B regular verb, present tense, third person singular “–s” ending is used [V + s]

10C irregular verb, present tense, third person singular “–s” ending is not used [V – s]

(Table continues)

53

Page 62: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Table A2 (continued) 10D irregular verb, present tense,

third person singular “–s” ending is used [V + s]

11A zero copula [“be” is dropped altogether]

11B aspectual or invariant be

11C conjugated be [use of “be” in any non-EAE form]

11D conjugated be [use of “be” in any EAE form]

12 Subject-Verb inversion in indirect questions and statements

13 Use of double negatives

14 Other non-EAE verb forms, excluding S-V concord

Table A3

Sentence Level Features

15A Subject/verb discord for copula verb forms [subject-verb disagreement for “to be” verbs]

15B Subject/verb discord for non-copula verb forms [subject-verb disagreement for other verbs]

16A Deletion of final letter or letters e.g., an(d), a(n), the(y), the(m), glad(ly)

16B Other spelling errors

16C Missing or incorrect preposition

16D Other missing words

16E Incorrect use of connecting words (yet, but, and, so, therefore, because, since, while, etc.)

54

Page 63: Writing Differences in Teacher Performance Assessments ...Edited American English was examined as a possible source of construct-irrelevant variance in assessors’ scoring judgments.

Table A4

Rhetorical Features

24A Overapplication of Edited American English (EAE) rules, resulting in errors e.g., Double-marking of a verb as past tense (would have + –ed); Use of possessive instead of objective case after a preposition

24B Use of highly prescriptive EAE, resulting in rhetorical dis-fluency (does not result in errors) e.g., Overuse of the definite article “the”; Highly restrictive adverb-placement rules; Avoidance of splitting an infinitive

26 Use of oral language in written form, resulting in dis-fluency or errors (text sounds fluent when read aloud, but is incorrect according to rules for written academic English)

55