construct valitity

Upload: joseph-kennedy

Post on 13-Apr-2018

233 views

Category:

Documents


0 download

TRANSCRIPT

  • 7/27/2019 Construct Valitity

    1/35http://journals.cambridge.org Downloaded: 26 Mar 2009 IP address: 194.80.32.9

    provides an account of the development of changingviews in language testing on validity, and languagetesters have come to accept that there is no onesingle answer to the question What does our testmeasure? or Does this test measure what it is sup-posed to measure?. Messick argues that the questionshould be rephrased along the lines of: What is theevidence that supports particular interpretations anduses of scores on this test?.Validity is not a character-istic of a test, but a feature of the inferences made onthe basis of test scores and the uses to which a test isput.One validates not a test,but a principle for mak-

    ing inferences (Cronbach & Meehl, 1955:297).Thisconcern with score interpretations and uses necessar-ily raises the issue of test consequences,and in educa-tional measurement, as well as in language testingspecifically, debates continue about the legitimacy ofincorporating test consequences into validation.Whether test developers can be held responsible fortest use and misuse is controversial, especially in lightof what we increasingly know about issues like testwashback (see Part One of this review).A new termhas been invented consequential validity but itis far from clear whether this is a legitimate area ofconcern or a political posture.

    Messick (1989:20) presents what he calls a pro-gressive matrix (see Figure 1), where the columnsrepresent the outcomes of testing and the rows rep-resent the types of arguments that should be used tojustify testing outcomes. Each of the cells containsconstruct validity, but new facets are added as onegoes through the matrix from top left to bottomright.

    As a result of this unified perspective, validation isnow seen as ongoing, as the continuous monitoringand updating of relevant information, indeed as aprocess that is never complete.

    Bachman and Palmer (1996) build on this newunified perspective by articulating a theory of testusefulness, which they see as the most importantcriterion by which tests should be judged. In sodoing, they explicitly incorporate Messicks unified

    In Part 1 of this two-part review article (Alderson &Banerjee, 2001), we first addressed issues of wash-back, ethics, politics and standards.After a discussionof trends in testing on a national level and in testingfor specific purposes, we surveyed developments incomputer-based testing and then finally examinedself-assessment, alternative assessment and the assess-ment of young learners.

    In this second part, we begin by discussing recenttheories of construct validity and the theories oflanguage use that help define the constructs thatwe wish to measure through language tests. The

    main sections of the second part concentrate onsummarising recent research into the constructsthemselves, in turn addressing reading, listening,grammatical and lexical abilities, speaking andwriting. Finally we discuss a number of outstandingissues in the field.

    Construct validity and theories oflanguage use

    Traditionally, testers have distinguished differenttypes of validity: content, predictive, concurrent,construct and even face validity. In a number ofpublications, Messick has challenged this view (forexample 1989, 1994, 1996), and argued that con-struct validity is a multifaceted but unified and over-arching concept, which can be researched from anumber of different perspectives. Chapelle (1999)

    Lang.Teach.35,79113. DOI: 10.1017/S0261444802001751 Printed in the United Kingdom 2002 Cambridge University Press 79

    J Charles Alderson is Professor of Linguistics andEnglish Language Education at Lancaster University.He holds an MA in German and French from OxfordUniversity and a PhD in Applied Linguistics fromEdinburgh University. He is co-editor of the journal

    Language Testing (Edward Arnold), and co-editor ofthe Cambridge Language Assessment Series(C.U.P.), and has published many books and articles onlanguage testing, reading in a foreign language, andevaluation of language education.

    Jayanti Banerjee is a PhD student in theDepartment of Linguistics and Modern EnglishLanguage at Lancaster University. She has beeninvolved in a number of test development and researchprojects and has taught on introductory testing courses.She has also been involved in teaching English forAcademic Purposes (EAP) at Lancaster University. Herresearch interests include the teaching and assessment ofEAP as well as qualitative research methods. She is par-ticularly interested in issues related to the interpretationand use of test scores.

    State-of-the-Art Review

    Language testing and assessment (Part 2)

    J Charles Alderson and Jayanti Banerjee Lancaster University, UK

    Figure 1. Messicks progressive matrix (cited in Chapelle, 1999:259).

    http://journals.cambridge.org/http://journals.cambridge.org/
  • 7/27/2019 Construct Valitity

    2/35http://journals.cambridge.org Downloaded: 26 Mar 2009 IP address: 194.80.32.9

    view of construct validity, but also add dimensionsthat affect test development in the real world. Testusefulness, in their view, consists of six major compo-nents, or what they call critical qualities of languagetests, namely construct validity, reliability, conse-quences, interactiveness, authenticity and practicality.Interestingly, Shohamy (1990b) had already arguedthat a test validation research agenda should be definedin terms of utility (to what extent a test serves thepractical information needs of a given audience),fea-sibility (ease of administration in different contexts)and fairness (whether tests are based on materialwhich test takers are expected to know). We willaddress the issue of authenticity and interactivenessbelow, but the appearance of practicality as a princi-pled consideration in assessing the quality of a testshould be noted.What is less clear in the Bachmanand Palmer account of test usefulness is how these

    various qualities should be measured and weightedin relation to each other.

    What is particularly useful about the reconceptu-alisation of construct validity is that it places the testsconstruct in the centre of focus, and readjusts tradi-tional concerns with test reliability. An emphasis onthe centrality of constructs what we are trying tomeasure requires testers to consider what is knownabout language knowledge and ability, and ability touse the language. Language testing involves not onlythe psychometric and technical skills required toconstruct and analyse a test but also knowledge

    about language: testers need to be applied linguists,aware of the latest and most accepted theories of lan-guage description, of language acquisition and lan-guage use.They also need to know how these can beoperationalised:how they can be turned into ways ofeliciting a persons language and language use.Language testing is not confined to a knowledge ofhow to write test items that will discriminatebetween the strong and the weak. Central to test-ing is an understanding of what language is, and whatit takes to learn and use language, which thenbecomes the basis for establishing ways of assessingpeoples abilities.A new series of books on languagetesting, the Cambridge Language AssessmentSeries (Cambridge University Press, edited byAlderson and Bachman), has the intention of com-bining insights from applied linguistic enquiry withinsights gained from language assessment, in order toshow how these insights can be incorporated intoassessment tools and procedures, for the benefit ofthe test developer and the classroom teacher.

    Alderson and Clapham (1992) report an attemptto find a model of language proficiency on whichthe revised ELTS test the IELTS test could bebased.The authors did not find any consensus among

    the applied linguists they surveyed, and report a deci-sion to be eclectic in calling upon theory in order todevelop the IELTS (International English LanguageTesting System) specifications. If that survey were to

    be repeated in the early 21st century we believethere would be much more agreement, at leastamong language testers, as to what the most appro-priate model should be. Bachman (1991) puts for-ward the view that a significant advance in languagetesting is the development of a theory that considerslanguage ability to be multi-componential, andwhich acknowledges the influence of the testmethod and test taker characteristics on test perfor-mance. He describes what he calls an interactionalmodel of language test performance that includestwo major components, language ability and testmethod, where language ability consists of languageknowledge and metacognitive strategies and testmethod includes characteristics of the environment,rubric, input, expected response and the relationshipbetween input and expected response. This hasbecome known as the Bachman model, as described

    in Bachman (1990) and Bachman and Palmer (1996)and it has become an influential point of reference,being increasingly incorporated into views of theconstructs of reading, listening,vocabulary and so on.The model is a development of applied linguisticthinking by Hymes (1972) and Canale and Swain(1980), and by research, e.g.,by Bachman and Palmer(1996) and by the Canadian Immersion studies(Harley et al., 1990) and it has developed as it hasbeen scrutinised and tested. It remains very useful asthe basis for test construction, and for its account oftest method facets and task characteristics.

    Chalhoub-Deville (1997) disagrees with thisassessment of the usefulness of the Bachman model.She reviews several theoretical models of languageproficiency but considers that there is a degree oflack of congruence between theoretical models onthe one hand and operational assessment frame-works,which necessarily define a construct in partic-ular contexts, on the other. She argues that althoughtheoretical models are useful, there is an urgent needfor an empirically based, contextualised approach tothe development of assessment frameworks.

    Nevertheless, we believe that one significant con-tribution of the Bachman model is that it not onlybrings testing closer to applied linguistic theory, butalso to task research in second language acquisition(SLA), one of whose aims is to untangle the variouscritical features of language learning tasks. TheBachman and Palmer model of the characteristics oftest tasks shows how much more advanced testingtheory and thinking has become over the years, astesters have agonised over their test methods.Yet SLAresearchers and other applied linguists often usetechniques for data elicitation that have long beenquestioned in testing. One example is the use ofcloze tests to measure gain in immersion studies;

    another is the use of picture descriptions in studies oforal performance and task design. Klein Gunnewiek(1997) critically examines the validity of instrumentsused in SLA to measure aspects of language acquisi-

    Language testing and assessment (Part 2)

    80

    http://journals.cambridge.org/http://journals.cambridge.org/
  • 7/27/2019 Construct Valitity

    3/35http://journals.cambridge.org Downloaded: 26 Mar 2009 IP address: 194.80.32.9

    He draws up a possible research agenda that wouldflow from the inclusion of a social perspective and in-deed such a research agenda has already borne fruit inseveral studies of the nature of the interaction in oraltests (Porter,1991a,1991b;OSullivan,2000a,2000b).

    As a result, we now have a somewhat betterunderstanding of the complexity of oral perfor-mance, as reflected in the original McNamara model(1995:173), seen in Figure 2 below.

    We can confidently predict that the implicationsof this model of the social dimensions of languageproficiency will be a fruitful area for research forsome time to come.

    Validation research

    Recent testing literature has shown a continuedinterest in studies which compare performances ondifferent tests, in particular the major influential testsof proficiency in English as a Foreign Language.(Unfortunately, there is much less research usingtests of languages other than English.) Perhaps thebest known study, the Cambridge-TOEFL study(Bachman et al., 1988; Davidson & Bachman, 1990;Kunnan, 1995;Bachman et al., 1995; Bachman et al.,1996), was expected to reveal differences betweenthe Cambridge suite of exams and the TOEFL.Interestingly, however, the study showed more simi-larities than differences. The study also revealedproblems of low reliability and a lack of parallelism insome of the tests studied, and this research led to sig-nificant improvements in those tests and in develop-ment and validation procedures.This comparabilitystudy was also a useful testing ground for analysesbased on the Bachman model, and the various testsexamined were subject to scrutiny using that modelsframework of content characteristics and testmethod facets.

    In a similar vein, comparisons of the EnglishProficiency Test Battery (EPTB), the English

    Language Battery (ELBA) and ELTS, as part of theELTS Validation Study (Criper & Davies, 1988),proved useful for the resulting insights into Englishfor Specific Purposes (ESP) testing, and the some-

    Language testing and assessment (Part 2)

    81

    tion. Such critical reviews emphasise the need forapplied linguists and SLA researchers to familiarisethemselves with the testing literature, lest they over-look potential weaknesses in their methodologies.

    Perkins and Gass (1996) argue that, since pro-ficiency is multidimensional, it does not always de-velop at the same rate in all domains. Therefore,models that posit a single continuum of proficiencyare theoretically flawed.They report research whichtested the hypothesis that there is no linear relation-ship between increasing competence in differentlinguistic domains and growth in linguistic pro-ficiency. They conclude with a discussion of theimplications of discontinuous learning patterns forthe measurement of language proficiency develop-ment, and put forward some assessment models thatcan accommodate discontinuous patterns of growthin language. In similar vein, Danon-Boileau (1997)

    argues that, since language development is complex,assessment of language acquisition needs to considerdifferent aspects of that process: not only proficiencyat one point in time, or even how far students haveprogressed, but also what they are capable of learn-ing, in the light of their progress and achievement.

    We have earlier claimed that there is no longer anissue about which model to use to underpin testspecifications and as the basis for testing research.Indeed, we have argued (see Part One) that theCouncil of Europes Common European Frameworkwill be influential in the years to come in language

    education generally, and one aspect of its useful-ness will be its exposition of a model of language,language use and language learning often explicitlybased on the Bachman model. For example, theDIALANG project referred to in Part One based thespecifications of its diagnostic tests on the CommonEuropean Framework.At present, probably the mostfamiliar aspects of the Framework are the variousscales, developed by North and Schneider (1998) andothers, because they have obvious value in measuringand assessing learning and achievement.

    However,McNamara (1995; McNamara & Lumley,1997) challenges the Bachman model. McNamaraargues that the model ignores the social dimension oflanguage proficiency, since the model is, in his opin-ion, based on psychological rather than social psy-chological or social theories of language use. Heurges language testers to acknowledge the socialnature of language performance and to examinemuch more carefully its interactional i.e., social aspects.He points out that in oral tests, for example,acandidates performance may be affected by how theinterlocutor performs, or by the person with whomthe candidate is paired.The raters perception of theperformance, and that persons use (or misuse) of

    rating scales, are other potential influences on the testscore,and he asks the important question:Whose per-formance are we assessing?This he calls the PandorasBoxof language testing (McNamara,1995).

    Figure 2. Proficiency and its relation to perfor-mance (McNamara,1995)

    http://journals.cambridge.org/http://journals.cambridge.org/
  • 7/27/2019 Construct Valitity

    4/35http://journals.cambridge.org Downloaded: 26 Mar 2009 IP address: 194.80.32.9

    what surprising stability of constructs over verydifferent tests. Many validation studies have beenconducted on TOEFL (some recent examples areFreedle & Kostin, 1993, 1999 and Brown, 1999), theTest of Spoken English (TSE) (for example, Powers etal., 1999) and comparisons have been made betweenTOEFL, TSE and the Test of Written English(TWE) (for example, DeMauro, 1992), comparisonsof IELTS and TOEFL (Geranpayeh,1994;Anh, 1997;Al-Musawi & Al-Ansari, 1999)), and of TOEFLwith newly developed tests (Des Brisay, 1994).

    Alderson (1988) claims to have developed newprocedures for validating ESP tests, using the exam-ple of the IELTS test development project. Fulcher(1999a) criticises traditional approaches to the use ofcontent validity in testing English for AcademicPurposes (EAP), in light of the new Messick frame-work and recent research into content specificity.

    Although testing researchers remain interested inthe validity of large-scale international proficiencytests, the literature contains quite a few accountsof smaller scale test development and validation.Evaluative studies of placement tests from a varietyof perspectives are reported by Brown (1989),Bradshaw (1990),Wall et al. (1994), Blais and Laurier(1995), Heller et al. (1995), and Fulcher (1997a,1999b). Lynch (1994) reports on the validation ofThe University of Edinburgh Test of English atMatriculation (comparing the new test to the estab-lished English Proficiency Test Battery). Laurier and

    Des Brisay (1991) show how different statistical andjudgemental approaches can be integrated in small-scale test development. Ghonsooly (1993) describeshow an objective translation test was developed andvalidated, and Zeidner and Bensoussan (1988) andBrown (1993) describe the value and use of test takerattitudes and feedback on tests in development orrevision. OLoughlin (1991) describes the develop-ment of assessment procedures in a distance learningprogramme, and Pollard (1998) describes the historyof development of a computer-resourced Englishproficiency test, arguing that there should be morepublications of research-in-development, to showhow test development and testing research can gohand-in-hand.

    A recent doctoral dissertation (Luoma, 2001)looked at theories of test development and constructvalidation in order to explore how the two can berelated, but the research revealed the lack of pub-lished studies that could throw light on problems intest development (the one exception was IELTS).Most published studies of language test developmentare somewhat censored accounts, which stress thepositive features of the tests rather than addressingproblems in development or construct definition or

    acknowledging the limitations of the publishedresults. It is very much to be hoped that accounts willbe published in future by test developers (along thelines of Peirce, 1992 or Alderson et al., 2000),

    describing critically how their language tests weredeveloped, and how the constructs were identified,operationalised, tested and revised. Such accountswould represent a valuable contribution to appliedlinguistics by helping researchers, not only test devel-opers, to understand the constructs and the issuesinvolved in their operationalisation.

    A potentially valuable contribution to our knowl-edge of what constitutes and influences test perfor-mance and the measurement of language proficiencyis the work being carried out by the the Universityof Cambridge Local Examinations Syndicate(UCLES), by, for example, the administration ofquestionnaires which elicit information on culturalbackground, previous instruction, strategy, cognitivestyle and motivation (Kunnan, 1994, 1995; Purpura,1997, 1998, 1999). Such research, provided it ispublished warts and all, has enormous potential. In

    this respect, the new Cambridge Studies inLanguage Testing series (Cambridge UniversityPress, edited by Milanovic and Weir) is a valuableaddition to the language testing literature and to ourunderstanding of test constructs and research meth-ods, complementing the Research Reports from ETS.It is hoped that other testing agencies and examina-tion boards will also publish details of their research,and that reports of the development of nationalexams, for example, will contribute to our under-standing of test performance and learner characteris-tics, not just for English,but also for other languages.

    Language testers continue to use statistical meansof test validation, and recent literature has reportedthe use of a variety of techniques, reviewed byBachman and Eignor (1997). Perhaps the best knowninnovation and most frequently reported methodof test analysis has been the application of ItemResponse Theory (IRT). Studies early in the periodof this review include de Jong and Glas (1989), whoreport the use of the Rasch model to analyse items ina listening comprehension test, and McNamara(1990, 1991) who uses IRT to validate an ESP listen-ing test. Hudson (1991) explores the relative meritsof one- and two-parameter IRT models and tradi-tional bi-serial correlations as measures of itemdiscrimination in criterion-referenced testing.Although he shows close relationships between thethree measures,he argues that, wherever possible, it ismost appropriate to use the two-parameter model,since it explicitly takes account of item discrimina-tion. He later (Hudson, 1993) develops three differ-ent indices of item discrimination which can be usedin criterion-referenced testing situations where IRTmodels are not appropriate.

    One claimed drawback of the use of IRT modelsis that they require that the tests on which they are

    used be unidimensional. Henning et al. (1985) andHenning (1988) study the effects of violating theassumption of unidimensionality and show thatdistorted estimates of person ability result from such

    Language testing and assessment (Part 2)

    82

    http://journals.cambridge.org/http://journals.cambridge.org/
  • 7/27/2019 Construct Valitity

    5/35http://journals.cambridge.org Downloaded: 26 Mar 2009 IP address: 194.80.32.9

    violations. Buck (1994) questions the value in lan-guage testing of the notion of unidimensionality,arguing that language proficiency is necessarily amultidimensional construct. However, others, forexample Henning (1992a), distinguish between psy-chological unidimensionality, which most languagetests cannot meet, and psychometric unidimension-ality, which they very often can. McNamara andLumley (1997) explore the use of multi-facetedRasch measurement to examine the effect of suchfacets of spoken language assessment as interlocutorvariables, rapport between candidate and interlocu-tor, and the quality of audiotape recordings of per-formances. Since the results revealed the effect ofinterlocutor variability and audiotape quality, theauthors conclude that multi-faceted Rasch measure-ment is a valuable additional tool in test validation.

    Other studies investigate the value of multidimen-

    sional scaling in developing diagnostic interpreta-tions of TOEFL subscores (Oltman & Stricker,1990), and of a new technique known as Rule-Space Methodology to explore the cognitive andlinguistic attributes that underlie test performanceon a listening test (Buck & Tatsuoka, 1998) and areading test (Buck et al., 1997). The authors arguethat the Rule-Space Methodology can explainperformance on complex verbal tasks and providediagnostic scores to test takers. Kunnan (1998) andPurpura (1999) provide introductions to structuralequation modelling in language testing research, and

    a number of articles in a special issue of the journalLanguage Testing demonstrate the value of suchapproaches in test validation (Bae & Bachman, 1998;Purpura,1998).

    A relatively common theme in the exploration ofstatistical techniques is the comparison of differenttechniques to achieve the same ends. Bachman et al.(1995) investigate the use of generalisability theoryand multi-faceted Rasch measurement to estimatethe relative contribution of variation in test tasks andrater judgements to variation in test scores on aperformance test of Spanish speaking ability amongundergraduate students intending to study abroad.Similarly, Lynch and McNamara (1998) investigategeneralisability theory (using GENOVA) and multi-faceted Rasch measurement (using FACETS) toanalyse performance test data and they compare theirrelative advantages and roles. Kunnan (1992) com-pares three procedures (G-theory, factor and clusteranalyses) to investigate the dependability and validityof a cr iterion-referenced test and shows how the dif-ferent methods reveal different aspects of the testsusefulness. Lynch (1988) investigates differential itemfunctioning (DIF a form of item bias) as a result ofperson dimensionality and Sasaki (1991) compares

    two methods for identifying differential item func-tioning when IRT models are inappropriate.Henning (1989) presents several different methodsfor testing for local independence of test items.

    In general, the conclusions of such studies are thatdifferent methods have different advantages and dis-advantages, and users of such statistical techniquesneed to be aware of these.

    However, quantitative methods are not the only

    techniques used in validation studies, and languagetesting has diversified in the methods used to exploretest validity, as Banerjee and Luoma (1997) haveshown. Qualitative research techniques like intro-spection and retrospection by test takers (e.g.,Alderson, 1990b; Storey, 1994; Storey, 1997; Green,1998) are now widely used in test validation.Discourse analysis of student output, in oral as well aswritten performance, has also proved to be useful(Shohamy, 1990a; Lazaraton, 1992; Ross, 1992; Ross& Berwick, 1992; Young, 1995; Lazaraton, 1996;Young & He, 1998). Increasingly, there is also tr ian-gulation of research methods across the so-called

    qualitative/quantitative divide (see Anderson et al.,1991, for an early example of such triangulation) inorder better to understand what constructs areindeed being measured.

    We address the use of such qualitative techniquesas well as triangulations of different methods in thefollowing sections, which deal with recent researchinto the various constructs of language ability.

    Assessing reading

    How the ability to read text in a foreign language

    might best be assessed has long interested languagetesting researchers. There is a vast literature andresearch tradition in the field of reading in ones firstlanguage (L1), and this has had an influence on boththe theory of reading in a foreign language and onresearch into foreign language testing. Indeed, theissue of whether reading in a foreign language is alanguage problem or a reading problem (Alderson,1984) is still a current research topic (Bernhardt,1991; Bernhardt & Kamil, 1995; Bernhardt, 1999).Theorists and practitioners in the 1980s arguedabout whether L1 reading skills transferred to read-ing in a foreign language. One practical implicationof transfer might be that one should first teach stu-dents to read accurately and appropriately in the firstlanguage before expecting them to do so in a foreignlanguage.A second implication for assessment mightbe that tests of reading may actually be tests of oneslinguistic abilities and proficiency.

    However, consensus has slowly emerged that theshort-circuit hypothesis (Clarke, 1979, 1988) isessentially correct. This hypothesis posits that onesfirst language reading skills can only transfer to theforeign language once one has reached a thresholdlevel of competence in that language.Whilst this may

    seem perfectly obvious at one level how can youpossibly read in a foreign language without knowinganything of that language? the real question is atwhat point the short circuit ceases to operate.The

    Language testing and assessment (Part 2)

    83

    http://journals.cambridge.org/http://journals.cambridge.org/
  • 7/27/2019 Construct Valitity

    6/35http://journals.cambridge.org Downloaded: 26 Mar 2009 IP address: 194.80.32.9

    issue is not quite so simple as deciding that a givenlevel of language proficiency is the threshold level.Reading is an interaction between a reader with allthat the reader brings with him/her backgroundknowledge, affect, reading purpose, intelligence, firstlanguage abilities and more and the text, whosecharacteristics include topic, genre, structure, lan-guage (organisation, syntax, vocabulary, cohesion)and so on.Thus a readers first language reading skillsmay transfer at a lower level of foreign languageproficiency, on a text on a familiar topic, written ineasy language,with a clear structure, than they wouldon a less familiar topic, with much less clearly struc-tured organisation,with difficult language.

    One of the goals of testing research is to explorethe nature of difficulty of tests, of test tasks anditems and the causes of such difficulty.As indicatedabove, difficulty is relative to readers, and thus the

    readers ability and other characteristics have to betaken into account. Perkins and Brutten (1988a)report what they call a behavioural anchoring analy-sis of three foreign language reading tests. For eachtest, they identified items that clearly discriminatedbetween different levels of reading ability, andanalysed the items in terms of their relation to thestructure of the texts, the readers backgroundknowledge and the cognitive processes supposedlyrequired to answer the questions. They claim thathigher-level students could comprehend micro-propositions and questions whose sources of infor-

    mation were implicit, whereas lower-level studentscould not. Both students with higher reading abilityand those with lower reading ability showed com-petence with linguistic structures that related partsof the text, regardless of their language proficiency.This again raises the question of the relationshipbetween reading ability and linguistic proficiency.

    Perkins et al. (1995) showed that an artificial neur-al network was an interesting technique for investi-gating empirically what might bring about item leveldifficulty, and Buck et al. (1997) have developedthe Rule-Space Methodology to explore causes ofitem difficulty.Whilst to date these techniques havemerely identified variables in item design relation-ship between words in the items and words in thetext, wording of distractors, and so on futureresearch might well be able to throw light on theconstructs that underlie reading tests, and thusenhance our understanding of what reading ability ina foreign language consists of (but see Hill & Parry,1992, for a sceptical approach to the validity of anytraditional test of reading).

    Alderson (2000) suggests that views of difficulty inreading in a foreign language could be explored bylooking at how test developers have specified their

    tests at different levels of proficiency, and refers to theCambridge suite of tests, the Council of EuropeFramework, and the frameworks used in nationalassessments of foreign language ability. One such

    characterisation is contained in the ACTFL(American Council on the Teaching of ForeignLanguages) Guidelines for Reading (Child, 1987).ACTFL divides reading proficiency into three areas content, function and accuracy. Two parallel hier-archies are posited, one of text types and the otherof reading skills, which are cross-sectioned to definedevelopmental levels (here developmental level isheld to indicate relative difficulty and ease). TheACTFL Guidelines have been very influential in theUSA in foreign language education and assessment,but have only rarely been subject to empirical scru-tiny. Lee and Musumeci (1988) examined studentscompleting tests at five different levels of ACTFL dif-ficulty, where questions were based on four differenttypes of reading skill.They found no evidence for theproposed hierarchy of reading skills and text types:the performances of readers from all levels were

    remarkably similar, and distinct from the hypothe-sised model.Further confirmation of this finding wasmade by Allen et al. (1988). Debate continues, how-ever, as Edwards (1996) criticises the design of pre-vious studies, and claims to have shown that whensuitably trained raters select an adequate sample ofpassages at each level, and a variety of test methodsare employed, then the ACTFL text hierarchy mayindeed provide a sound basis for the development offoreign language reading tests. Shohamy (1990b)criticises the ACTFL Guidelines for being simplistic,unidimensional and inadequate for the description of

    context-sensitive, unpredictable language use. Sheargues that it is important that the construction oflanguage tests be based on a more expanded andelaborated view of language.

    A common belief in foreign language readingis that there is a hierarchy of skills, as posited byACTFL, and asserted by Benjamin BloomsTaxonomy of Educational Objectives (Bloom et al.,1956).This is an example of theory in first languageeducation being transferred,often uncritically, to for-eign language education. This hierarchy has oftenbeen characterised as consisting of higher-order andlower-order skills, where understanding explicitlystated facts is regarded as lower-order and appreci-ating the style of a text or distinguishing betweenmain ideas and supporting detail is held to behigher-order. In foreign language reading assess-ment, it has often been held that it is important totest more than mere lower-order skills, and theinference from such beliefs has been that somehowhigher-order skills are not only more valuable, butalso more difficult for foreign language readers.Alderson and Lukmani (1989) and Alderson (1990a)critically examine this assumption, and show that,firstly, expert judges do not agree on whether test

    questions are assessing higher- or lower-order skills,and secondly, that even for those items where expertsagree on the level of skill being tested, there is nocorrelation between level of skill and item difficulty.

    Language testing and assessment (Part 2)

    84

    http://journals.cambridge.org/http://journals.cambridge.org/
  • 7/27/2019 Construct Valitity

    7/35http://journals.cambridge.org Downloaded: 26 Mar 2009 IP address: 194.80.32.9

    Alderson concludes that item difficulty does notnecessarily relate to level of skill, and that no impli-cational scale exists such that students have to acquirelower-order skills before they can acquire higher-order skills.

    This conclusion has proved controversial andLumley (1993) shows that, once teachers have beentrained to agree on a definition of skills, and providedany disagreements are discussed at length, substantialagreements on matching subskills to individual testitems can be reached. Alderson (1991a) argues thatall that this research and similar findings by Bachmanet al. (1989) and Bachman et al. (1995) prove is thatraters can be trained to agree. That does not, heclaims, mean that individual skills can be tested sepa-rately by individual test items.Alderson (1990b) andLi (1992) show that students completing tests pur-porting to assess individual sub-skills in individual

    items can get answers correct for the wrongreason i.e., without displaying the skill intended and they can get an item wrong for the rightreason that is, showing evidence of the skill inquestion.Reves and Levine (1992), after examining amastery reading test, argue that enabling readingskills are subsumed within the overall mastery ofreading comprehension and therefore need not bespecified in the objectives of reading tests. Alderson(2000) concludes that individuals responding to testitems do so in a complex and interacting variety ofdifferent ways, that experts judging test items are not

    well placed to predict how learners, whose languageproficiency is quite different from that of the experts,might actually respond to test items, and that there-fore generalisations about what skills reading testitems might be testing are fatally flawed. This issomething of a problem for test developers andresearchers.

    Anderson et al. (1991) offer an interestingmethodological perspective on this issue, by ex-ploring the use of think-aloud protocols, contentanalyses and empirical item performances in orderto triangulate data on construct validity. Findingson the nature of what is being tested in a reading testremain inconclusive, but such triangulated method-ologies will be imperative for future such studies(and see above on construct validation).

    One perennial area of concern in reading tests isthe effect of readers background knowledge and thetext topic on any measure of reading ability. Perkinsand Brutten (1988b) examine different types of read-ing questions: textually explicit (which can beanswered directly from the text), textually implicit(which require inferences) and scriptally implicit(which can only be answered with backgroundknowledge).They show significant differences in dif-

    ficulty and discriminability and conclude that testingresearchers (and by implication test developers) mustcontrol for background knowledge in readingtests. Hale (1988) examines this by assuming that a

    students academic discipline (a crude measure ofbackground knowledge) will interact with test con-tent in determining performance, and he shows thatstudents in humanities/social sciences and in the bio-logical/physical sciences perform better on passagesrelated to their disciplines than on other passages.However, although significant, the differences weresmall and had no practical effect in terms of theTOEFL scale perhaps, he concludes, becauseTOEFL passages are taken from general readingsrather than from specialised textbooks. Peretz andShoham (1990) show that, although students ratetexts related to their fields of study as more compre-hensible than texts related to other topics, their sub-jective evaluations of difficulty are not a reliablepredictor of their actual performance on readingtests.Clapham (1996) also looked at students ratingsof their familiarity with, and their background

    knowledge of, the content of specific texts in theIELTS test battery. In addition she compared the stu-dents academic discipline with their performance ontests based on texts within those disciplines.She con-firmed earlier results by Alderson and Urquhart(1985) that showed that students do not necessarilyperform better on tests in their subject area. Sometexts appear to be too specific for given fields andothers appear to be so general that they can beanswered correctly by students outside the discipline.Clapham argues that EAP testing, based on theassumption that students will be advantaged by

    taking tests in their subject area where they havebackground knowledge, is not necessarily justifiedand she later concludes (Clapham, 2000) that sub-ject-specific reading tests should be replaced by testsof academic aptitude and grammatical knowledge.

    Interestingly, Alderson (1993) also concludes, onthe basis of a study of pilot versions of the IELTStest, that it is difficult to distinguish between tests ofacademic reading and contextualised tests of gram-matical ability. Furthermore, with clear implicationsfor the short-circuit hypothesis mentioned above,Clapham (1996) shows that students scoring less than60% on a test of grammatical knowledge appear tobe unable to apply their background knowledge tounderstanding a text, whereas students scoring above80% have sufficient linguistic proficiency to be ableto overcome deficits in background knowledgewhen understanding texts. The suggestion is thatthere must be two thresholds, and that only studentsbetween, in this case, scores of 60% and 80%, are ableto use their background knowledge to compensatefor lack of linguistic proficiency.Again, these findingshave clear implications for what it is that reading testsmeasure.

    A rather different tradition of research into reading

    tests is one that looks at the nature of the text and itsimpact on test difficulty. Perkins (1992), for example,studies the effect of passage structure on test per-formance, and concludes that when questions are

    Language testing and assessment (Part 2)

    85

    http://journals.cambridge.org/http://journals.cambridge.org/
  • 7/27/2019 Construct Valitity

    8/35http://journals.cambridge.org Downloaded: 26 Mar 2009 IP address: 194.80.32.9

    derived from sentences where given informationprecedes new information, and where relevant infor-mation occurs in subject position, they are easierthan questions derived from sentences with otherkinds of topical structure. However, Salager-Meyer(1991) shows that matters are not so simple. Sheinvestigated the effect of text structure across differ-ent levels of language competence and topicalknowledge, and different degrees of passage familiar-ity. Students familiarity with passages affected testperformance more than text structure, and wherestudents were less familiar with passages, changes intext structure affected only weaker students, notstronger students. Where passages are completelyunfamiliar, neither strong nor weak students areaffected by high degrees of text structuring. Thus,text structure as a variable in test difficulty must beinvestigated in relation to its interaction with other

    text and reader characteristics, and not in isolation.Finally, one ongoing tradition in testing research is

    the investigation of test method effects, and this isoften conducted in the area of reading tests. Recentresearch has both continued the investigation of tra-ditional methods like multiple-choice, short-answerquestions, gap-filling and C-tests, but has also gonebeyond these (see, for example, Jafarpur, 1987, 1995,1999a,1999b).Chapelle (1988) reports research indi-cating that field independence may be a variableresponsible for systematic error in test scores andshows that there are different relationships between

    measures of field independence and cloze, multiplechoice and dictation tests. Grotjahn (1995) criticisesstandard multiple-choice methods and proposes pos-sible alternatives, like the cloze procedure, C-testsand immediate recall. However,Wolf (1993a, 1993b)claims that immediate recall may only assess theretrieval of low-level detail. Wolf concludes thatlearners ability to demonstrate their comprehensiondepends on the task, and the language of the testquestions. She claims that selected response (mul-tiple-choice) and constructed response (cloze, short-answer) questions measure different abilities (as alsoclaimed by Grotjahn, 1995), but that both mayencourage bottom-up low-levelprocessing.She alsosuggests that questions in the first language ratherthan the target language may be more appropriate formeasuring comprehension rather than production (seealso the debate in the Netherlands reported in PartOne about the use of questions in Dutch to measurereading ability in English Welling-Slootmaekers,1999, van Elmpt & Loonen,1998 and Bhgel & Leijn,1999). Translation has also occasionally been re-searched as a test method for assessing reading ability(Buck, 1992a), showing surprisingly good validityindices, but the call for more research into this test

    method has not yet been taken up by the testingcommunity.

    Recall protocols are increasingly being used as ameasure of foreign language comprehension. Deville

    and Chalhoub-Deville (1993) caution against uncrit-ical use of such techniques, and show that only whenrecall scoring procedures are subjected to item andreliability analyses can they be considered an alterna-tive to other measures of comprehension. Riley andLee (1996) compare recall and summary protocols asmethods of testing understanding and conclude thatthere are significant qualitative differences in the twomethods.The summaries contained more main ideasthan the recall protocols and the recalls contained ahigher percentage of details than main ideas.Different methods would appear to be appropriatefor testing different aspects of understanding.

    Testing research will doubtless continue to addressthe issue of test method, sometimes repeating theresearch and debates of the 1970s and 1980s into theuse of cloze and C-test procedures, and sometimesmaking claims for other testing methods.The con-

    sensus, nevertheless, is that it is essential to use morethan one test method when attempting to measure aconstruct like reading comprehension.

    Quite what the construct of reading comprehen-sion is has been addressed in the above debates.However, one area that has received relatively littleattention is: how do we know when somebody hascomprehended a text? This question is implicit in thediscussion of multiple-choice questions or recall pro-tocols: is the ability to cope with such test methodsequivalent to understanding a text? Research bySarig (1989) addresses head-on the problem of vari-

    able text meaning,pointing out that different readersmay indeed construct different meanings of a textand yet be correct.This is partly accounted for byschema theory, but still presents problems for decid-ing when a reader has correctly interpreted a text.She offers a methodology for arriving at a consensusview of text meaning,by analysing model answers toquestions from samples of readers from diverse back-grounds and levels of expertise. She recommendsthat the Meaning Consensus Criterion Answer beused as a basis for item scoring (and arguably also forscoring summary and recall protocols).

    Hill and Parry (1992), however,offer a much moreradical perspective. Following Street (1984), theycontrast two models of literacy the autonomousand the pragmatic and claim that traditional tests ofreading assume that texts have meaning, and viewtext, reader and the skill of reading itself asautonomous entities.They offer an alternative viewof literacy, namely that it is socially constructed.Theysay that the skill of reading goes beyond decodingmeaning to the socially conditioned negotiation ofmeaning, where readers are seen as having social, notjust individual, identities.They claim that reading andwriting are inseparable,and that their view of literacy

    requires an alternative approach to the assessment ofliteracy, one which includes a social dimension.Theimplications of this have yet to be worked out in anydetail, and that will doubtless be the focus of work in

    Language testing and assessment (Part 2)

    86

    http://journals.cambridge.org/http://journals.cambridge.org/
  • 7/27/2019 Construct Valitity

    9/35http://journals.cambridge.org Downloaded: 26 Mar 2009 IP address: 194.80.32.9

    the 21st century. Alderson (2000) gives examples ofalternative assessment procedures which could besubject to detailed validation studies.

    Assessing listening

    In comparison to the other language skills, theassessment of listening has received little attention,possibly reflecting (as Brindley, 1998, argues) thedifficulties involved in identifying relevant features ofwhat is essentially an invisible cognitive operation.Recent discussions of the construct of listening andhow the listening trait might be isolated and mea-sured (Buck, 1990, 1991, 1992b, 1994; Dunkel et al.,1993) suggest that there is a separate listening traitbut that it is not necessarily operationalised by oralinput alone.

    Buck has extended this claim (see Buck, 1997,

    2001), explaining that, while listening comprehen-sion might primarily be viewed as a process of con-structing meaning from auditory input, that processinvolves more than the auditory signal alone.Listening comprehension is seen as an inferentialprocess involving the interaction between both lin-guistic and non-linguistic knowledge. Buck (2001)explains that listening comprehension involvesknowledge of discrete elements of language such asphonology, vocabulary and syntax but it goes beyondthis because listening also involves interpretation.Listening must be done automatically in real time

    (listeners rarely get a second chance to hear exactlythe same text), involves background knowledge andlistener-specific variables (such as purpose for listen-ing) and is a very individual process, implying thatthe more complex a text is the more varied the pos-sible interpretations. It also has unique characteristicssuch as the variable nature of the acoustic input.Listening input is characterised by features such aselision and the placement of stress and intonation.Ideas are not necessarily expressed in a linear gram-matical manner and often contain redundancy andhesitation.

    All these features raise the question of what is thebest approach to assessing listening. Recent researchinto test methods has included research into the useof dictation (see Kaga, 1991 and Coniam, 1998) andsummary translation (see Stansfield et al.,1990,1997;Scott et al., 1996). While dictation has often beenused as a measure of language proficiency in Frenchand English as second languages, it has been arguedthat it is not as effective a measure when the targetlanguage has a very close relationship between itspronunciation and orthography. Kaga (1991) consid-ers the use of graduated dictation (a form of modi-fied dictation) to assess the listening comprehension

    of adult learners of Japanese in a university context.Her results indicate that the graduated dictation isan effective measure of language proficiency even inthe case of Japanese where, arguably, the pronuncia-

    tion and orthography in the target language areclosely related. Coniam (1998) also discusses the useof dictation to assess listening comprehension, in hiscase a computer-based listening test the TextDictation. He argues that this type of test is moreappropriate as a test of listening than short fragmentswhere the required responses are in the form oftrue/false questions, gap-fill etc., because the text ismore coherent and provides more context. Hisresults indicate that the Text Dictation procedurediscriminates well between students of different pro-ficiency levels.

    Summary translation tests, such as the ListeningSummary Translation Exam (LSTE) developed byStansfield et al. (1990, 1997, 2000) and Scott et al.(1996), first provide an instructional phase in whichthe test takers are taught the informational and lin-guistic characteristics of a good summary.Test takers

    are then presented with three summary translationtasks.The input consists of conversations in the targetlanguage (Spanish or Taiwanese). These vary inlength from one to three minutes. In each case thetest takers hear the input twice and are permitted totake notes.They then have to write a summary ofwhat they have heard, in English. Interestingly, thistest method not only assesses listening, but alsowriting and the developers report that listeningperformance in the target language has an inverserelationship with writing performance in English. Itis also clear from Kagas and Coniams research that

    the target of the assessment is general language pro-ficiency rather than the isolation of a specific listen-ing trait.This confirms Bucks (1994) suggestion thatthere are two types of listening test, the first beingorally presented tests of general language compre-hension and the second tests of the listening traitproper. Indeed, one of the challenges of assessinglistening is that it is well nigh impossible to constructa pure test of listening that does not require the useof another language skill. In addition to listeningto aural input, test takers are likely to have to readwritten task instructions and/or questions.They alsohave to provide either oral or written responses tothe questions.Consequently,what might be intendedas a listening test could also be assessing anotherlanguage skill.

    To complicate matters further, other test factorssuch as the test method and test taker characteristicssuch as memory capacity (Henning, 1991) could alsocontribute to the test score. Admittedly though,research into the effects of these factors has beensomewhat inconclusive. Hale and Courtney (1994)examine the effect of note-taking on test taker per-formance on the listening section of the TOEFL.They report that allowing test takers to make notes

    had little effect on their test scores while activelyurging them to take notes significantly impairedtheir performance. This finding perhaps says moreabout the students note-taking experiences and

    Language testing and assessment (Part 2)

    87

    http://journals.cambridge.org/http://journals.cambridge.org/
  • 7/27/2019 Construct Valitity

    10/35http://journals.cambridge.org Downloaded: 26 Mar 2009 IP address: 194.80.32.9

    habits than about the value of note-taking in thecontext of the TOEFL listening test.

    Sherman (1997) considers the effect of questionpreview. Subjects took listening tests in four differentversions:questions before the listening exercise, sand-wiched between two hearings of the listening text,after the text, or no questions at all. She found thatthe test takers had a strong affective preference forpreviewed questions but previewing did not neces-sarily result in more correct answers. In fact, theversion that produced significantly more correctanswers was the one in which the test takers heardthe passage twice (with the questions presentedbetween the two hearings). It is not clear, in this case,whether the enhanced scores were due to the oppor-tunity to preview the questions or the fact that thetext was played twice, or indeed a combination ofthe two.

    In an effort to disentangle method from trait,Yian(1998) employed an immediate retrospective verbalreport procedure to investigate the effect of a mul-tiple-choice format on listening test performance.Her results, apart from providing evidence ofthe explanatory power of qualitative approaches inassessment research (see also Banerjee & Luoma,1997), show how test takers activate both theirlinguistic and non-linguistic knowledge in order toprocess input.Yian argues that language proficiencyand background knowledge interact and that non-linguistic knowledge can either compensate for

    deficiencies in linguistic knowledge or can facilitatelinguistic processing.The former is more likely in thecase of less able listeners who are only partiallysuccessful in their linguistic processing. More com-petent and advanced listeners are more likely to usetheir non-linguistic knowledge to facilitate linguisticprocessing. However, the use of non-linguisticknowledge to compensate for linguistic deficienciesdoes not guarantee success in the item.Yians resultsalso indicate that the multiple-choice format dis-advantages less able listeners and allows uninformedguessing. It also results in test takers getting an itemcorrect for the wrong reasons.

    Other research into the testing of listening haslooked at text and task characteristics that affect diffi-culty (see Buck, 2001: 149-151, for a comprehensivesummary). Task characteristics that affect listeningtest difficulty include those related to the informa-tion that needs to be processed, what the test taker isrequired to do with the information and how quick-ly a response is required.Text characteristics that caninfluence test difficulty include the phonologicalqualities of the text and the vocabulary, grammar anddiscourse features. Apart from purely linguisticcharacteristics, text difficulty is also affected by

    the degree of explicitness in the presentation of theideas, the order of presentation of the ideas and theamount of redundancy.

    Shohamy and Inbar (1991) investigated the effect

    of both texts and question types on test takers scores,using three texts with various oral features. Theirresults indicate that texts located at different pointson the oral/written continuum result in different testscores, the texts with more oral features being easier.They also report that, regardless of the topic, texttype or the level of the test takers language profi-ciency, questions that refer to local cues are easierthan those that refer to global cues. Freedle andKostin (1999) examined 337 TOEFL multiple-choice listening items in order to identify character-istics that contribute to item difficulty.Their findingsindicate that the topic and rhetorical structure of theinput text affect item difficulty but that the two mostimportant determinants of item difficulty are thelocation of the information necessary for the answerand the degree of lexical overlap between the textand the correct answer.When the necessary informa-

    tion comes at the beginning of the listening text theitem is always easier (regardless of the rhetoricalstructure of the text) than if it comes later. In ad-dition, when words used in the listening passage arerepeated in the correct option, the item is easier thanwhen words found in the listening passage are usedin the distractors. In fact, lexical overlap between thepassage and item distractors is the best predictor ofdifficult items.

    Long (1990) and Jensen and Hansen (1995) lookat the effect of background/prior knowledge on lis-tening test performance. Jensen and Hansen postu-

    late that listening proficiency level will affect theextent to which prior knowledge of a topic can beaccessed and used, hypothesising that test takers willneed a high proficiency level in order to activatetheir prior knowledge. Their findings, however, donot support this hypothesis. Instead, they concludethat the benefit of prior knowledge is more likely tomanifest itself if the input text is technical in nature.Long (1990) used two Spanish listening texts, oneabout a well-known pop group and the other abouta gold rush in Ecuador. She reports that the betterthe Spanish proficiency of the test takers the bettertheir score on both texts.However, for the text aboutthe pop group, there was no significant difference inscores between more and less proficient test takers.Her interpretation of this result is that backgroundknowledge of a topic can compensate for linguisticdeficiencies. However, she warns that schematicknowledge can also have the reverse effect. Some testtakers actually performed badly on the gold rush textbecause they had applied their knowledge of a differ-ent gold rush.

    Long does not attempt to explain this finding(beyond calling for further study of the interactionbetween schematic knowledge, language level and

    text variables). However, Tsui and Fullilove (1998)suggest that language proficiency level and textschema can interact with text processing as a dis-criminator of test performance. They identify two

    Language testing and assessment (Part 2)

    88

    http://journals.cambridge.org/http://journals.cambridge.org/
  • 7/27/2019 Construct Valitity

    11/35http://journals.cambridge.org Downloaded: 26 Mar 2009 IP address: 194.80.32.9

    types of text, one in which the schema activated bythe first part of the text is not congruent with thesubsequent linguistic input (non-matching) and theother in which the schema is uniform throughout(matching). They argue that non-matching textsdemand that test takers process the incoming lin-guistic cues quickly and accurately, adjusting theirschema when necessary. On the other hand, testtakers were able to rely on top-down processing tounderstand matching texts. Their findings indicatethat, regardless of question type, the more proficienttest takers performed better on non-matching textsthan did the less proficient. They conclude thatbottom-up processing is more important than top-down processing in distinguishing between differentproficiency levels.

    With the increasing use of technology (such asmulti-media and computers) in testing, researchers

    have begun to address the issue of how visual infor-mation affects listening comprehension and test per-formance. Gruba (1997) discusses the role of videomedia in listening assessment, considering how theprovision of visual information influences the defi-nition and purpose of the assessment instrument.Ginther (forthcoming) in a study of the listeningsection of the computer-based TOEFL (CBT), hasestablished that listening texts tend to be easier ifaccompanied by visual support that complementsthe content, although the converse is true of visualsupport that provides context. The latter seems to

    have the effect of slightly distracting the listener fromthe text, an effect that is more pronounced with lowproficiency test takers. Though her results are indi-cative rather than conclusive, this finding has ledGinther to suggest that high proficiency candidatesare better able to overcome the presence of contextvisuals.

    The increasing use of computer technology is alsomanifest in the use of corpora to develop listeningprototypes, bringing with it new concerns. Douglasand Nissan (2001) explain how a corpus of NorthAmerican academic discourse is providing the basisfor the development of prototype test tasks as part ofa major revision of the TOEFL. Far from providingthe revision project team with recorded data thatcould be directly incorporated into test materials,inspection of the corpus foregrounded a number ofconcerns related to the use of authentic materials.These include considerations such as whetherspeakers refer to visual material that has not beenrecorded in the corpus text, whether the excerptmeets fairness and sensitivity guidelines or requiresculture-specific knowledge.The researchers empha-sise that not all authentic texts drawn from corporaare suitable for listening tests, since input texts need

    to be clearly recorded and the input needs to bedelivered at an appropriate speed.A crucial issue is how best to operationalise the

    construct of listening for a particular testing context.

    Dunkel et al. (1993) propose a model for test specifi-cation and development that specifies the person,competence, text and item domains and compo-nents. Coombe et al. (1998) attempt to provide cri-teria by which English for Academic Purposespractitioners can evaluate the listening tests they cur-rently use, and they provide micro-skill taxonomiesdistinguishing general and academic listening.Theyalso highlight the significance of factors such ascultural and background knowledge and discuss theimplications of using test methods that mimicreal-life authentic communicative situations ratherthan indirect, discrete-point testing. Buck (2001)encourages us to think of the construct both interms of the underlying competences and the natureof the tasks that listeners have to perform in the realworld. Based on his own research (see Buck, 1994),he warns that items typically require a variety of

    skills for successful performance and that these candiffer between test takers. He argues, therefore, that itis difficult to target particular constructs with anysingle task and that it is important to have a range oftask types to reflect the construct.

    Clearly there is a need for test developers to con-sider their own testing situation and to establishwhich construct and what operationalisation of thatconstruct is best for them.The foregoing discussion,like other overviews of listening assessment (Buck,1997, Brindley, 1998), draws particular attention tothe challenges of assessing listening, in particular the

    limits of our understanding of the nature of the con-struct. Indeed, as Alderson and Bachman commentin Buck (2001):

    The assessment of listening abilities is one of the least under-stood, least developed and yet one of the most important areas oflanguage testing and assessment.(2001:x series editorspreface)

    Assessing grammar and vocabulary

    Grammar and vocabulary have enjoyed rather differ-ent fortunes recently.The direct testing of grammarhas largely fallen out of favour, with little researchtaking place (exceptions are Brown & Iwashita, 1996and Chung,1997),while research into the assessmentof vocabulary has flourished. Rea-Dickins (1997,2001) attributes the decline in the direct testing ofgrammar to the interest in communicative languageteaching, which has led to a diminished role forgrammar in teaching and consequently testing. Shealso suggests that changes in the characterisation oflanguage proficiency for testing purposes have con-tributed to the shift of focus away from grammar.Instead, assessment tasks have been developed thatreflect the target language use situation; such was thecase of the English as a Second Language Placement

    Examination (ESLPE) which, when it was revised,was designed to assess the test takers ability tounderstand and use language in academic contexts(see Weigle & Lynch, 1995). Furthermore, even in

    Language testing and assessment (Part 2)

    89

    http://journals.cambridge.org/http://journals.cambridge.org/
  • 7/27/2019 Construct Valitity

    12/35http://journals.cambridge.org Downloaded: 26 Mar 2009 IP address: 194.80.32.9

    cases where a conscious effort has been made todevelop a grammar test, for example when theIELTS test was being developed, a grammar test hasnot been found to contribute much additional infor-mation about the test takers language proficiency.Indeed, Alderson (1993) found that the proposedIELTS grammar sub-test correlated to some degreewith all of the other sub-tests and correlated particu-larly highly with the reading sub-test.The grammartest was therefore dropped,to save test taker time.

    This overlap could be explained by the fact thatgrammar is assessed implicitly in any assessment oflanguage skills e.g., when raters assign a mark foraccuracy when assessing writing. Moreover, as Rea-Dickins (2001) explains, grammar testing has evolvedsince its introduction in the 1960s. It now encom-passes the understanding of cohesion and rhetoricalorganisation as well as the accuracy and appropriacy

    of language for the task. Task types also vary morewidely and include gap-filling and matching exer-cises, modified cloze and guided summary tasks.As aconsequence, both the focus and the method ofassessment are very similar to those used in theassessment of reading, which suggests that Aldersons(1993) findings are not,in retrospect,surprising.

    Certainly (as was the case with the IELTS test), ifdropping the grammar sub-test does not impact sig-nificantly on the reliability of the overall test, itwould make sense to drop it. Eliminating one sub-test would also make sense from the point of view of

    practicality, because it would shorten the time takento administer the full test battery. And, as Rea-Dickins (2001) has demonstrated, the lack of explicitgrammar testing does not imply that grammar willnot be tested.Grammatical accuracy is usually one ofthe criteria used in the assessment of speaking andwriting, and grammatical knowledge is also requiredin order successfully to complete reading items thatare intended to measure test takers grasp of details,ofcohesion and of rhetorical organisation.

    The assessment of vocabulary has been a moreactive field recently than the assessment of grammar.For example, the DIALANG diagnostic testingsystem, mentioned in Part One, uses a vocabularysize test as part of its procedure for estimating testtakers proficiency level in order to identify theappropriate level of test to administer to a test taker,and in addition offers a separate module testing vari-ous aspects of vocabulary knowledge.

    An active area of research has been the develop-ment of vocabulary size tests. These tests arepremised on the belief that learners need a certainamount of vocabulary in order to be able to operateindependently in a particular context. Two differentkinds of vocabulary size test have been developed.

    The Vocabulary Levels Test, first developed by Nation(1990), requires test takers to match a word with itsdefinition, presented in multiple-choice format inthe form of a synonym or a short phrase.Words are

    ranked into five levels according to their frequencyof occurrence and each test contains 36 words ateach level. The other type of vocabulary size testemploys a different approach. Sometimes called theYes/No vocabulary test, it requires test takers simplyto say which of the words in a list they know. Thewords are sampled according to their frequency ofoccurrence and a certain proportion of the items arenot real words in the target language.These pseudo-words are used to identify when a test taker might beover-reporting their vocabulary knowledge (seeMeara & Buxton,1987 and Meara,1996).Versions ofthis test have been written for learners of differenttarget languages such as French learners of Dutch(Beeckmans et al., 2001) and learners of Russian andGerman (Kempe & MacWhinney, 1996), andDIALANG has developed Yes/No tests in 14European languages.

    Since these two kinds of test were first developed,they have been the subject of numerous modificationsand validation. Knowing that Nations VocabularyLevels tests (Nation, 1990) had not undergone thor-ough statistical analysis, Beglar and Hunt (1999)thought it unlikely that the different forms of thesetests would be equivalent.Therefore, they took fourforms of each of the 2000 Word Level and UniversityWord Level tests, combining them to create a 72-item 2000 Word Level test and a 72-item UniversityWord Level test which they then administered to496 Japanese students. On the basis of their results,

    they produced and validated revised versions of thesetwo tests.Schmitt et al. (2001) also sought to validatethe four forms of the Vocabulary Levels Test (three ofwhich had been written by the main author). In avariation on Beglar and Hunts (1999) study, theyattempted to gather as diverse a sample of test takersas possible. The four forms were combined intotwo versions and these were administered to 801students. The authors conclude that both versionsprovide valid results and produce similar, if notequivalent, scores.

    When examining the Yes/No vocabulary test,Beeckmans et al. (2001) were motivated by a con-cern that many test takers selected a high number ofpseudo-words and that these high false alarm rateswere not restricted to weak students. There alsoseemed to be an inverse relationship between identi-fication of real words and rejection of pseudo-words. This is counter-intuitive, since one wouldexpect a test taker who is able to identify real wordsconsistently also to be able to reject most or allpseudo-words. They administered three forms of aYes/No test of Dutch vocabulary (the forms differedin item order only) to 488 test takers.Their findingsled them to conclude that the Yes/No format is

    insufficiently reliable and that the correction pro-cedures cannot cope with the presence of bias inthe test-taking population.

    What is not clear from this research, however, is

    Language testing and assessment (Part 2)

    90

    http://journals.cambridge.org/http://journals.cambridge.org/
  • 7/27/2019 Construct Valitity

    13/35http://journals.cambridge.org Downloaded: 26 Mar 2009 IP address: 194.80.32.9

    what counts as a word and what vocabulary size isenough.Although nobody would claim that vocabu-lary size is the key to learners language needs (theyalso need other skills such as a grasp of the structureof the language), there is general agreement thatthere is a threshold below which learners are likelyto struggle to decode the input they receive.However, there is little agreement on the nature ofthis threshold. For instance, Nation (1990) andLaufer (1992, 1997) argue that learners at universitylevel need to know at least 3000 word families.YetNurweni and Read (1999) estimate that universitylevel students only know about 1200 word families.

    Apart from concerns about how to measurevocabulary size and how much vocabulary isenough, researchers have also investigated the depthof learners word knowledge i.e., how well words areknown.Traditionally, this has been studied through

    individual interviews where learners provide expla-nations of words which are then rated by theresearcher. One example of this approach is a studyby Verhallen and Schoonen (1993) of both Dutchmonolingual and bilingual Turkish immigrant chil-dren in the Netherlands. They elicited as manyaspects as possible of the meaning of six Dutch wordsand report that the monolingual children were ableto provide more extensive and varied meanings thanthe bilingual children.However, such a methodologyis time-consuming and restricts researchers to smallsample sizes. Results are also susceptible to bias as a

    result of the interview process and so otherapproaches to gathering such information have beenattempted. Read (1993) reports on one alternativeapproach. He devised a written test in which testitems comprise a target word plus eight others.Thetest takers task is to identify which of the eightwords are semantically related to the target word(four in each item). This approach is, however, sus-ceptible to guessing and Read suggests that a betteralternative might be to require test takers to supplythe alternatives.

    This challenge has been taken up by Laufer et al.(2001), who address three concerns. The first iswhether it is enough merely to establish how manywords are known.The second is how to accurately(and practically) measure depth of vocabularyknowledge. The third concern is how to measureboth receptive and productive dimensions of vocab-ulary knowledge. Laufer et al. designed a test ofvocabulary size and strength. Size was operationalisedas the number of words known from various wordfrequency levels. Strength was defined according to ahierarchy of depth of word-knowledge beginningfrom the easiest receptive recognition (test takersidentify words they know) and proceeding to the

    most difficult productive recall (test takers have toproduce target words).They piloted the test on 200adult test takers, paying attention to two questionsthat they considered key to establishing the validity

    of the procedure.The first had to do with whetherthe assumed hierarchy of depth of word-knowledgeis valid, i.e., does recall presuppose recognition anddoes production presuppose reception? Second,howmany items does a test taker need to answer correctlyat any level before they can be considered to haveadequate vocabulary knowledge at that level? Theyconsider the results so far to be extremely promisingand hope eventually to deliver the test in computer-adaptive form.

    But such studies involve the isolation of vocabu-lary knowledge in order to measure it in detail.Therehave also been efforts to assess vocabulary knowledgemore globally, including the attempts of Laufer andNation (1999) to develop and validate a vocabulary-size test of controlled productive ability. They tookvocabulary from the five frequency levels identifiedby Nation (1990) and constructed completion item

    types in which a truncated form of the word is pre-sented in a short sentence.Test takers are expected tobe able to use the context of the sentence to com-plete the word.The format bears some resemblanceto the C-test (see Grotjahn, 1995) but has two keydifferences. The words are presented in single sen-tences rather than paragraphs and instead of alwayspresenting the first half of the word being tested,Laufer and Nation include the minimal number ofletters required to disambiguate the cue.They arguethat this productive procedure allows researchers tolook more effectively at breadth of vocabulary

    knowledge and is a useful complement to receptivemeasures of vocabulary size and strength.Other approaches to measuring vocabulary glob-

    ally have involved calculating the lexical richness oftest takers production (primarily writing), usingcomputerised analyses, typically of type/token ratios(see Richards & Malvern, 1997 for an annotated bib-liography of research in this area). However,Vermeer(2000), in a study of the spontaneous speech data offirst and second language children learning Dutch,argues that these measures have limitations. His datashows that at early stages of vocabulary acquisitionmeasures of lexical richness seem satisfactory.However, as learners progress beyond 3000 words,the methods currently available produce lower esti-mations of vocabulary growth. He suggests, there-fore, that researchers should look not only at thedistribution of and relation between types and tokensused but also at the level of difficulty of the wordsused.

    Research has also been conducted into the vocab-ulary sections in test batteries. For instance, Schmitt(1999) carried out an exploratory study of a smallnumber of TOEFL vocabulary items, administeringthem to 30 pre-university students and then ques-

    tioning the students about their knowledge of thetarget words associations, grammatical properties,collocations and meaning senses. His results suggestthat the items are not particularly robust as measures

    Language testing and assessment (Part 2)

    91

    http://journals.cambridge.org/http://journals.cambridge.org/
  • 7/27/2019 Construct Valitity

    14/35http://journals.cambridge.org Downloaded: 26 Mar 2009 IP address: 194.80.32.9

    of association, word class and collocational knowl-edge of the target words. On the basis of thisresearch, Schmitt argues for more investigation ofwhat vocabulary items in test batteries are reallymeasuring. Takala and Kaftandjieva (2000) adopt adifferent focus, looking at whether gender bias couldaccount for differential item functioning (DIF) in thevocabulary section of the Finnish Foreign LanguageCertificate Examination.They report that the test asa whole is not gender-biased but that particularitems do indicate DIF in favour of either males orfemales. They argue that these results have implica-tions for item-banking and call for more research,particularly into whether it is practically (as opposedto theoretically) possible for a test containing DIFitems to be bias-free.

    However, all the work we have reported focuseson vocabulary knowledge (whether it be size,

    strength or depth),which is, as Chapelle (1994,1998)has argued, a rather narrow remit. She calls for abroader construct, proposing a model that takes intoaccount the context of vocabulary use, vocabularyknowledge and metacognitive strategies for vocabu-lary use. Thus, in addition to refining our under-standing of vocabulary knowledge and theinstruments we use to measure it (see Meara, 1992and 1996 for research into psycholinguistic measuresof vocabulary) it is necessary to explore ways ofassessing vocabulary under contextual constraintsthat, Read and Chapelle (2001:1) argue,are relevant

    to the inferences to be made about lexical ability.

    Assessing speaking

    The testing of speaking has a long history (Spolsky,1990, 1995, 2001) but it was not until the 1980s thatthe direct testing of L2 oral proficiency becamecommonplace,due, in no small measure, to the inter-est at the time in communicative language teaching.Oral interviews, of the sort developed by theForeign Service Institute (FSI) and associated USGovernment agencies (and now known as OPIs Oral Proficiency Interviews) were long hailed asvalid direct tests of speaking ability. Recently, how-ever, there has been a spate of criticisms of oral inter-views, which have in their turn generated a numberof research studies. Discourse, conversation and con-tent analyses show clearly that the oral proficiencyinterview is only one of the many possible genres oforal test tasks, and the language elicited by OPIs isnot the same as that elicited by other types of task,which involve different sorts of power relations andsocial interaction among interactants.

    Some researchers have attempted to reassure scep-tics about the capacity of oral tests to sample sufficient

    language for accurate judgements of proficiency(Hall, 1993) and research has continued into indirectmeasures of speaking (e.g., Norris, 1991). In addi-tion, a number of studies document the development

    of large-scale oral testing systems in school and uni-versity settings (Gonzalez Pino, 1989; Harlow &Caminero, 1990; Walker, 1990; Lindblad, 1992;Robinson, 1992).An influential set of guidelines forthe assessment of oral language proficiency waspublished in 1986 by the American Council on theTeaching of Foreign Languages (the ACTFL guide-lines ACTFL, 1986). This was followed by theintroduction of the widely influential ACTFL OralProficiency Interview (ACTFL OPI).

    This increase in the testing of speaking wasaccompanied by a corresponding expansion ofresearch into how speaking might best be assessed.The first major subject of this research was, notsurprisingly, the ACTFL OPI and there have been anumber of studies investigating the construct validityof the test (e.g., Raffaldini, 1988; Valdes, 1989;Dandonoli & Henning, 1990; Henning, 1992b;

    Alonso, 1997); the validity of the scores and ratingscale (e.g., Meredith, 1990; Halleck, 1992; Huebner& Jensen, 1992; Reed, 1992; Marisi, 1994; Glisan &Foltz, 1998); and rater behaviour and performance(Barnwell, 1989; Thompson, 1995). Conclusionshave varied, with some researchers arguing for theusefulness and validity of the OPI and its accompa-nying rating scale and others criticising it for its lackof theoretical and empirical support (e.g., Bachman& Savignon, 1986;Shohamy, 1990b; Salaberry, 2000).Fulcher (1997b: 75) argues that speaking tests areparticularly problematic from the point of view of

    reliability, validity, practicality and generalisability.Indeed, underlying the debate about the ACTFLOPI are precisely these concerns.

    Questions about the nature of oral proficiency,about the best way of eliciting it, and about the eval-uation of oral performances have motivated muchresearch in this area during the last decade.The mostrecent manifestation of this interest has been a jointsymposium between the Language Testing ResearchColloquium (LTRC) and the American Associationof Applied Linguistics (AAAL) held in February2001 which was devoted to the definition and assess-ment of speaking ability.The LTRC/AAAL sympo-sium encompassed a range of perspectives onspeaking, looking at the mechanical aspects of speak-ing (de Bot, 2001), the sociolinguistic and strategicfeatures of speaking ability (Bachman, 2001; Conrad,2001;Selinker, 2001;Swain, 2001b;Young, 2001) andthe implications for task design of the context-dependent nature of speaking performance (Liskin-Gasparro,2001).

    The most common mode of delivery of oralproficiency tests is the face-to-face oral proficiencyinterview (such as the ACTFL OPI). Until the 1990sthis took the form of a one-to-one interaction

    between a test taker and an interlocutor/examiner.However, this format has been criticised because theasymmetrical relationship between the participantsresults in reduced or no opportunities for genuine

    Language testing and assessment (Part 2)

    92

    http://journals.cambridge.org/http://journals.cambridge.org/
  • 7/27/2019 Construct Valitity

    15/35http://journals.cambridge.org Downloaded: 26 Mar 2009 IP address: 194.80.32.9

    conversational interaction to occur. Discussions ofthe nature of oral proficiency in the early 1990s (vanLier, 1989 and Lazaraton, 1992) focused on therelationship between OPIs and non-test discourse.Questioning the popular belief that an oral profi-ciency interview (OPI) is a structured conversationalexchange, van Lier asked two crucial questions: first,how similar is test taker performance in an OPI tonon-interview discourse (conversation), and second,should OPIs strive to approximate conversations?His view was that OPIs frequently do not result indiscourse resembling conversational exchange (aview shared by Lazaraton, 1992, Chambers &Richards, 1995 and Kormos, 1999) and thatresearchers need to think carefully about whetheroral proficiency is best displayed through conversa-tion. Johnson and Tyler (1998) take a single OPI thatis used to train OPI testers and analyse it for features

    of naturally occurring conversation.They report thatthis OPI lacks features typical of normal conversa-tion. For instance, the turn-taking is more structuredand predictable with longer turns always being takenby the test taker. Features of topic nomination andnegotiation differ as well and the test taker has nocontrol over the selection of the next speaker. Finally,the tester tends not to react to/add to the test takerscontributions, and this contributes to the lack ofcommunicative involvement observed. Egbert(1998), in her study of German OPIs, also suggeststhat repair is managed differently. The repair strat-

    egies expected are more cumbersome and formalthan would occur in normal German native-speakerconversation. However, Moder and Halleck (1998)challenge the relevance of this, asking whether thelack of resemblance between an OPI and naturallyoccurring conversation is important. They suggestinstead that it should be viewed as a type of in-terview, arguing that this is an equally relevant com-municative speech event.

    Researchers have sought to understand the natureof the OPI as a communicative speech event from anumber of different perspectives. Picking up onresearch into the linguistic features of the question-answer pair (Lazaraton, 1992; Ross & Berwick,1992), He (1998) looks at answers in the question-answer pair. She focuses specifically on a failingperformance, looking for evidence in the test takersanswers that were construed as indicating a limitedlanguage proficiency. She identifies a number of fea-tures including an unwillingness to elaborate, pausesfollowing questions, and wrong and undecipherableresponses.Yoshida-Morise (1998) investigates the useof communication strategies by Japanese learners ofEnglish and the relationship between the strategiesused and the learners level of proficiency. She reports

    that the number and nature of communicationstrategies used varies according to the proficiency ofthe learner. Davies (1998),Kim and Suh (1998), Ross(1998), and Young and Halleck (1998) look at the

    OPI as a cross-cultural encounter.They discuss theconstruction of identity and maintenance of face, theeffect of cultural assumptions on test taker behaviourand how this might in turn be interpreted by theinterlocutor/examiner.

    Researchers have also been concerned about theeffect of the interlocutor on the test experiencesince any variation in the way tasks are presented totest takers might impact on their subsequent perfor-mance (Ross, 1992; Katona, 1996; Lazaraton, 1996;Brown & Hill,1998;OLoughlin, 2000), as might thefailure of an interlocutor to exploit the full rangeof a test takers ability (Reed & Halleck, 1997).Addressing the concern that the interlocutor mighthave an effect on the amount of language that candi-dates actually produce, a study by Merrylees andMcDowell (1999) found that the interviewer typi-cally speaks far more than the test taker.Taking the

    view that the interview format obscures differencesin the conversational competence of the candidates,Kormos (1999) found that the conversational inter-action was more symmetrical in a guided role-playactivity.Yet even the guided role-play is dependenton the enthusiasm with which the interlocutorembraces the spirit of the role-play, as research byBrown and Lumley (1997) suggests. As a result oftheir work on the Occupational English Test (OET)they report that the more the interlocutor identifieswith the role presented (rather than with the testtaker), i.e., the more genuinely s/he plays the part

    required by the role play, the more challenging theinteraction becomes for the test taker.Katona (1998)has studied meaning negotiation, considering theeffect of familiarity with the interlocutor on the wayin which meaning is negotiated between the partici-pants. She concludes that both the frequency andtype of negotiation differ according to whether theinterlocutor is known or unfamiliar to the test taker:if the interlocutor is unknown to the test taker, this ismore likely to result in misunderstandings and thediscourse is more artificial and formal.

    Other variations on the format of the OPI includethe assessment of pair and group activities, as well asthe inclusion of more than one examiner. For exam-ple, the oral components of the Cambridge mainsuite of tests have gradually been revised to adopt apaired format with two examiners (Saville &Hargreaves, 1999). It is argued that the paired formatallows for a variety of patterns of interaction (i.e.,examiner-examinee(s), examinee(s)-examiner, andexaminee-examinee) and that assessment is fairerwith two examiners (typically a holistic judgementfrom the interlocutor/examiner and an analytic scorefrom the silent assessor). Ikeda (199