1. THE TESOL-ILTA JOINT SESSION The Design and Development of the Web-Based Online Oxford Placement Exam James E. Purpura, Teachers College, Columbia University [email_address] TESOL, Boston March 2010 (Thanks to Yoko Ameriks, Simon Beeston, Jee Wha Dakin, Kristin DiGennaro, & Kirby Grabowski) (For support thanks to: Oxford University Press, TESOL International & The International Language Testing Association (ILTA—www.iltaonline.com)
2.
3.
4. Initial Design Decisions • The revised OOPT would have 5 sections: a language knowledge section (i.e., grammatical & pragmatic knowledge) & 4 language use (i.e., skills) sections. • The decision to measure language knowledge separately stemmed from research showing that grammatical knowledge (i.e., form & meaning) is a critical resource (Purpura, 1999, 2006) & a strong predictor of the ability to communicate in meaningful & pragmatically appropriate ways (Ameriks, 2009; Chang, 2002, 2004; Grabowski, 2009; Kim, 2009; Liao, 2009). • The language knowledge section would consist of 4 tasks: (1) MC grammar (F&M); (2) MC Meaning (Literal & Intended); (3) Cloze (F&M); & (4) MC Pragmatic Meaning.
5. Initial Design Decisions • The grammar section was developed first since it would be used to determine what CEFR-level-specific items the examinees would receive in the rest of the adaptive test. This is a computer-adaptive test.
6. How Does the Computer Adaptive System Work? 1. Candidate views test demo info, then registers & selects level description e.g. B1 4. Listening items (15 items) selected on the basis of level at end of the Task 3. 2. Candidate starts test with B1 level item from Task 1 (MC F&M). Person ability is calculated & the next item selected & so on, until 10 items are administered—at the point the person may be identified at B1 or other level Item level adaptive across the CEFR levels 3. Task 2 (Meaning) testlet is selected from database based on the examinee’s level at end of Task 1. Task 2 then administered. The same procedures occurs for Task 3 (F&M). Pre-constructed CEFR level tests The score is derived from 2 levels combined into one “block” score showing the examinee’s relative position each. This is a probabilistic model based in IRT. 5. Score generated: etc 48 47 46 45 etc. Listening Average Grammar
7. How the tests work - adaptively B1--1st item CORRECT Easier 7th item CORRECT Easier 5th item CORRECT Harder 6th item INCORRECT Harder 4th item INCORRECT Easier 9th item CORRECT Harder 3rd item CORRECT Harder 2nd item CORRECT Test focuses on student’s ability & adapts to provide items targeted for that person’s ability level – final test is shorter but still accurate. Harder 8th item INCORRECT
8. Initial Design Decisions • The listening section was developed last. Listening ability is defined in terms of understanding a range of literal, intended & implied meanings. This section consists of 3 tasks.
10. The Language Knowledge Section Purpose: to measure the grammatical & pragmatic knowledge resources that examinees need to communicate at six levels of CEFR proficiency . Information needed • How the structures & associated meanings in the OUP course books related to the scaled descriptors of grammatical knowledge in the CEFR. — Could we get this from the CEFR itself? • How grammatical knowledge was defined & operationalized in the 6 CEFR levels — Could we get this from the CEFR itself?
11.
12.
13. The Oxford Curricula (aligned with CEFR) Grammar covered at A1/A2 level • countable & uncountable nouns • adjectives & adverbs of frequency • how much/many & other quantifiers • present & past simple/present perfect • present continuous, going to future • can/can’t , like + ing, would like to • some/any; there is/are/was/were • subject & object pronouns • possessive adjectives/pronouns • demonstrative adjectives/pronouns • prepositions of time & place • comparative/superlative adj. • Saxon genitive • telling the time Vocabulary & Topic Areas • classroom expressions • personal information • daily routines, days of the week • family, holidays, • shopping, food & drink, Instances of Language Use • greetings (formal/informal) • dinner invitations • making small talk • past experiences, • making reservations
14. Grammar from the CEFR B1 > B2 Intermediate [Independent user-Vantage] • Uses accurately & appropriately vocabulary adequate for understanding and responding to a range of familiar & some unfamiliar topics & can identify appropriate lexical items . • Can handle with reasonable accuracy simple & some more complex grammatical forms . Can produce longer relevant utterances in spoken & written discourse using a variety of cohesive devices . • Can read & understand different kinds of texts. Can identify overall meaning & specific detail attitudes & opinions & can understand text structure. • Can understand standard spoken language on both familiar & unfamiliar topics relating to personal social academic life or work. • Can communicate in most situations ask & respond to questions appropriately & participate in interactions without resorting to repair strategies.
18. We decided to use Purpura’s (2004) model of language ability as a basis for test development at 6 CEFR levels & as a basis for validation. Models of Communicative Language Ability as an Organizing Framework Models of language ability: Canale & Swain, 1980; Bachman, 1990; Bachman & Palmer, 1996; Chapelle, 1998; Purpura, 2004, etc.).
19. Forms Meanings Grammatical Knowledge Based on Jaszczolt (2002) & Wedgwood (2005) The model assumes that grammatical forms are the primary resource for conveying a range of literal & intended meanings. What is Language Knowledge? Knowledge of forms at the sentence & discourse levels Logical representation of semantic meanings conveyed by grammatical forms Use in Context Pragmatic Meanings Production of implicatures, irony, stance, affect, identity, power, metaphor, politeness, etc. This serves to convey a range of implied ( pragmatic) meanings in context. And this determines overall communicative effectiveness.
20. Model of CLA Graphological or Phonological Meanings Graphological or Phonological Forms Lexical Meanings Lexical Forms Morphosyntactic Meanings Morphosyntactic Forms Cohesive Meanings Cohesive Forms Info. Management Meanings Info. Management Forms Interactional Meanings Interactional Forms Sentence Level Discourse Level Language Ability Grammatical Knowledge Grammatical Form Grammatical Meaning
21. The OOPT Language Knowledge Section • Item distractors (especially in task 1) were written to reflect partial knowledge of the form and/or meaning being tested, when possible. • Specs for each task were drawn up, one for each CEFR level tested (yes, 24!). Items were designed “theoretically” to target the different levels. • Designed to measure grammatical & pragmatic knowledge. Tasks Task 1: Testing knowledge of forms & meanings 10 items (out of 300) Task 2: Testing meanings (literal & intended) 10 items (out of 142) Task 3: Testing forms & meanings 10 items (out of 91) Task 4: Testing pragmatic meanings (implied) To be determined
33. Item Descriptors for Grammatical Form with the Item Focus ‘Tense & Aspect’ Within the A2 CEFR Level for Task 1 A2 Key Item Difficulty Hypoth. CEFR Future progressive we'll be packing 24.74 B2 Negation with “do” aux I don't 25.51 A2 "will" used with offers I’ll take 28.64 B1 Future time in question form will you do 32.46 B1 Future tense I’ll do/I’m going to do 33.50 B2 Present continuous to express future time am going 34.24 A1 Present tense of verb "go" go 34.28 A1 Simple past we went 34.36 A2 Present continuous in fixed expression “Who shall I say is calling?” is calling 35.38 B1 Simple past felt 35.45 A2 Simple past in interrogative did 37.18 A2
34. Item Descriptors for Grammatical Form with the Item Focus ‘Tense & Aspect’ Within the A2 CEFR Level for Task 1 A2 Key Item Difficulty Hypoth. CEFR Future progressive we'll be packing 24.74 B2 Negation with “do” aux I don't 25.51 A2 "will" used with offers I’ll take 28.64 B1 Future time in question form will you do 32.46 B1 Future tense I’ll do/I’m going to do 33.50 B2 Present continuous to express future time am going 34.24 A1 Present tense of verb "go" go 34.28 A1 Simple past we went 34.36 A2 Present continuous in fixed expression “Who shall I say is calling?” is calling 35.38 B1 Simple past felt 35.45 A2 Simple past in interrogative did 37.18 A2
35. Item Descriptors for Grammatical Form with the Item Focus ‘Tense & Aspect’ Within the A2 CEFR Level for Task 1 A2 Key Item Difficulty Hypoth. CEFR Future progressive we'll be packing 24.74 B2 Negation with “do” aux I don't 25.51 A2 "will" used with offers I’ll take 28.64 B1 Future time in question form will you do 32.46 B1 Future tense I’ll do/I’m going to do 33.50 B2 Present continuous to express future time am going 34.24 A1 Present tense of verb "go" go 34.28 A1 Simple past we went 34.36 A2 Present continuous in fixed expression “Who shall I say is calling?” is calling 35.38 B1 Simple past felt 35.45 A2 Simple past in interrogative did 37.18 A2
36. Item Descriptors for Grammatical Form with the Item Focus ‘Tense & Aspect’ Within the A2 CEFR Level for Task 1 Easy Hard A2 Key Item Difficulty Hypoth. CEFR Future progressive we'll be packing 24.74 B2 Negation with “do” aux I don't 25.51 A2 "will" used with offers I’ll take 28.64 B1 Future time in question form will you do 32.46 B1 Future tense I’ll do/I’m going to do 33.50 B2 Present continuous to express future time am going 34.24 A1 Present tense of verb "go" go 34.28 A1 Simple past we went 34.36 A2 Present continuous in fixed expression “Who shall I say is calling?” is calling 35.38 B1 Simple past felt 35.45 A2 Simple past in interrogative did 37.18 A2
37. Item Descriptors for Grammatical Form with the Item Focus ‘Tense & Aspect’ Within the A2 CEFR Level for Task 1 Easy Hard A2 Key Item Difficulty Hypoth. CEFR Future progressive we'll be packing 24.74 B2 Negation with “do” aux I don't 25.51 A2 "will" used with offers I’ll take 28.64 B1 Future time in question form will you do 32.46 B1 Future tense I’ll do/I’m going to do 33.50 B2 Present continuous to express future time am going 34.24 A1 Present tense of verb "go" go 34.28 A1 Simple past we went 34.36 A2 Present continuous in fixed expression “Who shall I say is calling?” is calling 35.38 B1 Simple past felt 35.45 A2 Simple past in interrogative did 37.18 A2
38. Item Descriptors for Grammatical Form with the Item Focus ‘Tense & Aspect’ Within the A2 CEFR Level for Task 1 Easy Hard A2 Key Item Difficulty Hypoth. CEFR Future progressive we'll be packing 24.74 B2 Negation with “do” aux I don't 25.51 A2 "will" used with offers I’ll take 28.64 B1 Future time in question form will you do 32.46 B1 Future tense I’ll do/I’m going to do 33.50 B2 Present continuous to express future time am going 34.24 A1 Present tense of verb "go" go 34.28 A1 Simple past we went 34.36 A2 Present continuous in fixed expression “Who shall I say is calling?” is calling 35.38 B1 Simple past felt 35.45 A2 Simple past in interrogative did 37.18 A2
39. Item Descriptors for Grammatical Form with the Item Focus ‘Tense & Aspect’ Within the A2 CEFR Level for Task 1 Easy Hard A2 Key Item Difficulty Hypoth. CEFR Future progressive we'll be packing 24.74 B2 Negation with “do” aux I don't 25.51 A2 "will" used with offers I’ll take 28.64 B1 Future time in question form will you do 32.46 B1 Future tense I’ll do/I’m going to do 33.50 B2 Present continuous to express future time am going 34.24 A1 Present tense of verb "go" go 34.28 A1 Simple past we went 34.36 A2 Present continuous in fixed expression “Who shall I say is calling?” is calling 35.38 B1 Simple past felt 35.45 A2 Simple past in interrogative did 37.18 A2
40. A2 Level-- Hypothesized at B2 ID: 0104 Select a word or phrase to complete the conversation shown below. 0 0 1 0 Example: Dichotomous Scoring Woman: What are you thinking about? Man: I was just thinking-- this time next week _____ our suitcases to go on holiday. a. we’ve packed b. we’d be packing c. we’ll be packing d. we’ve been packing
41.
42.
43. Example: Partial Credit Scoring C2 Level ID: 0034 Select a word or phrase to complete the conversation shown below. 2 3 1 0 Woman: It’s time I _____ a new car. Man: Why? What’s wrong with the one you’ve got? a. buy b. bought c. should buy d. had bought
45. Example: Partial Credit Scoring A1 Level ID: 0008 Select a word or phrase to complete the conversation shown below. 3 1 0 1 12.85 16.30 18.48 40.15 ??? Man: What do you usually _____ in the evenings? Woman: I like to watch television. a. do b. look c. have d. make
55. Model of CLA Graphological or Phonological Meanings Graphological or Phonological Forms Lexical Meanings Lexical Forms Morphosyntactic Meanings Morphosyntactic Forms Cohesive Meanings Cohesive Forms Info. Management Meanings Info. Management Forms Interactional Meanings Interactional Forms Sentence Level Discourse Level Language Ability Grammatical Knowledge Grammatical Form Grammatical Meaning
56. Task 2: Example Item Literal meaning of the elliptical form “not really.”
57. Task 2: Classical Item Analyses The percentage of students who answered correctly The percentage of students who answered A & C The percentage of students who selected each option in the bottom scoring and the top scoring group. The discrimination index; top group – bottom group, so .96 -.63 = .33 The point biserial gives an overall index of discrimination
59. Model of CLA Graphological or Phonological Meanings Graphological or Phonological Forms Lexical Meanings Lexical Forms Morphosyntactic Meanings Morphosyntactic Forms Cohesive Meanings Cohesive Forms Info. Management Meanings Info. Management Forms Interactional Meanings Interactional Forms Sentence Level Discourse Level Language Ability Grammatical Knowledge Grammatical Form Grammatical Meaning
62. Model of CLA Graphological or Phonological Meanings Graphological or Phonological Forms Lexical Meanings Lexical Forms Morphosyntactic Meanings Morphosyntactic Forms Cohesive Meanings Cohesive Forms Info. Management Meanings Info. Management Forms Interactional Meanings Interactional Forms Sentence Level Discourse Level Language Ability Grammatical Knowledge Grammatical Form Grammatical Meaning
63. Situation: A conversation between a customer & waiter at a restaurant about the menu. Task 4 Item: CEFR Level A1 (contextual, sociolinguistic & psychological meanings) Administered to all levels of the Community English Program .
64. Situation: A conversation between a two close friend . Task 4: CEFR Level B1 (contextual, sociolinguistic & psychological meanings) Circle the best answer.
67. Listening Section Purpose: • To measure a range of literal, intended, and implied(pragmatic) meanings. More specifically… • To measure literal meanings with endophoric reference (“to understand the lines”—literal paraphrase) — The answer explicitly stated in the text.
68. Listening Section • To measure implied meanings with endophoric reference (“to understand between the lines”) — The answer is implied by what is stated in the text. it is implied by or through the text. • To measure implied meanings with exophoric reference (“to understand beyond the lines”) — The answer is neither explicitly nor implicitly stated in the text; rather it is implied by or through the text.
69. Listening for Literal Mng (Endophoric Ref.) “ I think I’ll just have a coffee” = “ The woman want some coffee.”
70. Listening for Implied Meaning (Endo. Ref.) By asking for the building location, she is implying that she wants to go there. The man understands her implied meaning, & gives her the address of the art gallery. The meaning is implicitly stated in the text (i.e., endophoric).
72. Listening for Implied Meaning (Exophoric. Ref.) Nothing in the text explicitly states the woman is addressing a group of students. This can be surmised, however, by the type of activities & locations the woman describes and by the shared knowledge that students are often required to take tests when they arrive at a new school. The woman’s implied audience is beyond what is stated in the text (i.e., exophoric reference). Therefore, the focus of the item is on implied contextual or sociocultural meaning, rather than the literal meaning. And the reference to this information is exophoric because meaning can only be derived from outside the spoken text.
English language students in programs around the world need to be placed quickly and efficiently into class levels at the beginning of a course. Oxford University Press has provided a placement exam for many years, but recently they decided to revise it. To get information one what stakeholders wanted, the Assessment Team at Oxford surveyed some 300 stakeholders. Based on this research, the revised exam would… Measure language knowledge and use (e.g., listening); Make score-based placement decisions aligned with Common European Framework of Reference (CEFR); Report scores in relation to the CEFR; Provide detailed feedback to constituents; Be short, easy to administer & inexpensive. 03/27/10
03/27/10
• The new placement exam would have five sections: a language knowledge section (i.e., grammatical & pragmatic knowledge) & four language skills sections. • The grammar section was developed first, and would be used determine what CEFR-level-specific test items would follow. • The decision to measure language knowledge separately stems from research showing that grammatical knowledge (i.e., form & meaning) is a critical component & strong predictor of the ability to communicate in meaningful & pragmatically appropriate ways. Thus, the grammar section would be used determine what CEFR-level-specific test items would follow. The grammar section of the Oxford English placement Test (OOPT) will be the focus of today’s talk. 03/27/10
KEY SALES MESSAGE: Because the test is computer adaptive (CAT), it is able to provide tailored test for each person so the test can be shorter yet still reliable i.e. measure a test taker with a known and acceptable level of error = TEST RELIABILITY NB: The placement test item bank has around 1200 items across the CEF levels so each test of approximately 45 items can be expected to be somewhat different to any other test. But it is impossible for tests to be unique! Can the item bank be learnt? If you keep retaking the test e.g. if you buy test access as an individual, you could keep moving up the ability continuum and getting higher scores but you would need to make a huge effort to do this and would arguably be learning more and more English as you did this – i.e. the system would select items of increasing difficulty. But getting a higher score by learning more English is fine! TESOL 2010 03/27/10 Purpura & Beeston
• The new placement exam would have five sections: a language knowledge section (i.e., grammatical & pragmatic knowledge) & four language skills sections. • The grammar section was developed first, and would be used determine what CEFR-level-specific test items would follow. • The decision to measure language knowledge separately stems from research showing that grammatical knowledge (i.e., form & meaning) is a critical component & strong predictor of the ability to communicate in meaningful & pragmatically appropriate ways. Thus, the grammar section would be used determine what CEFR-level-specific test items would follow. The grammar section of the Oxford English placement Test (OOPT) will be the focus of today’s talk. 03/27/10
03/27/10
03/27/10
03/27/10
We drew on a number of sources of information for defining grammatical knowledge in the OOPT. TESOL 2010 Purpura & Beeston 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. Over the years, the Common European Framework of Reference has been increasingly used for the development of language curricula, textbooks, and assessments. This framework claims to provide a comprehensive description of (1) what language learners have to learn to do in order to use a language for communication and (2) what knowledge and skills they have to develop so as to be able to act effectively. The CEFR also claims to provide language proficiency descriptors that can be used to portray language growth on a vertical scale across different levels of achievement. In other words, the CEFR describes what learners might be expected to do with the language at these levels, but it does not actually specify what knowledge components learners need to have acquired at different proficiency levels! 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. Over the years, the Common European Framework of Reference has been increasingly used for the development of language curricula, textbooks, and assessments. This framework claims to provide a comprehensive description of (1) what language learners have to learn to do in order to use a language for communication and (2) what knowledge and skills they have to develop so as to be able to act effectively. The CEFR also claims to provide language proficiency descriptors that can be used to portray language growth on a vertical scale across different levels of achievement. In other words, the CEFR describes what learners might be expected to do with the language at these levels, but it does not actually specify what knowledge components learners need to have acquired at different proficiency levels! 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. Over the years, the Common European Framework of Reference has been increasingly used for the development of language curricula, textbooks, and assessments. This framework claims to provide a comprehensive description of (1) what language learners have to learn to do in order to use a language for communication and (2) what knowledge and skills they have to develop so as to be able to act effectively. The CEFR also claims to provide language proficiency descriptors that can be used to portray language growth on a vertical scale across different levels of achievement. In other words, the CEFR describes what learners might be expected to do with the language at these levels, but it does not actually specify what knowledge components learners need to have acquired at different proficiency levels! 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. Over the years, the Common European Framework of Reference has been increasingly used for the development of language curricula, textbooks, and assessments. This framework claims to provide a comprehensive description of (1) what language learners have to learn to do in order to use a language for communication and (2) what knowledge and skills they have to develop so as to be able to act effectively. The CEFR also claims to provide language proficiency descriptors that can be used to portray language growth on a vertical scale across different levels of achievement. In other words, the CEFR describes what learners might be expected to do with the language at these levels, but it does not actually specify what knowledge components learners need to have acquired at different proficiency levels! 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. Over the years, the Common European Framework of Reference has been increasingly used for the development of language curricula, textbooks, and assessments. This framework claims to provide a comprehensive description of (1) what language learners have to learn to do in order to use a language for communication and (2) what knowledge and skills they have to develop so as to be able to act effectively. The CEFR also claims to provide language proficiency descriptors that can be used to portray language growth on a vertical scale across different levels of achievement. In other words, the CEFR describes what learners might be expected to do with the language at these levels, but it does not actually specify what knowledge components learners need to have acquired at different proficiency levels! 03/27/10
TESOL 2010 Purpura & Beeston Grammatical knowledge is defined in terms of knowledge of grammatical forms and meanings…this includes a variety of forms and meanings at the sub-sentential, sentential, and discourse levels. Pragmatic knowledge is also part of langage abilty in this model, but task 1 is only explicitly measuring grammatical knowledge. Lexical forms at the advanced levels: Lexical meanings at the advanced levels • Noun compounding, • denotation, connotation, collocation • co-dependence restrictions (depend on) Morphosyntactic forms :Morphosyntactic meanings • inflectional & derivational affixes, • notions of time, space,, • syntactic structure, tense, • passivization • aspect, mood, complex/compound • interrogation sentence types Cohesive forms: Cohesive meanings • referential forms, ellipsis, • logical connectors Information management forms Information management meanings • given/new information, • emphasis, foregrounding • word order, parallelism 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. 03/27/10
03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. We coded the items for a number of variables. We will discuss a few here. We first coded the item of hypothesized CEFR level. Then we coded the items grammatical knowledge—the possibilities were F, M or F&M After that, the items were coded for the type of grammatical form or meaning according to Purpura (2004). Item 191 was a MS form item designed to measure the past simple. Then, we coded the item for the specific grammatical focus. Item 191 was a tense & aspect item. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. We coded the items for a number of variables. We will discuss a few here. We first coded the item of hypothesized CEFR level. Then we coded the items grammatical knowledge—the possibilities were F, M or F&M After that, the items were coded for the type of grammatical form or meaning according to Purpura (2004). Item 191 was a MS form item designed to measure the past simple. Then, we coded the item for the specific grammatical focus. Item 191 was a tense & aspect item. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. We coded the items for a number of variables. We will discuss a few here. We first coded the item of hypothesized CEFR level. Then we coded the items grammatical knowledge—the possibilities were F, M or F&M After that, the items were coded for the type of grammatical form or meaning according to Purpura (2004). Item 191 was a MS form item designed to measure the past simple. Then, we coded the item for the specific grammatical focus. Item 191 was a tense & aspect item. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. We coded the items for a number of variables. We will discuss a few here. We first coded the item of hypothesized CEFR level. Then we coded the items grammatical knowledge—the possibilities were F, M or F&M After that, the items were coded for the type of grammatical form or meaning according to Purpura (2004). Item 191 was a MS form item designed to measure the past simple. Then, we coded the item for the specific grammatical focus. Item 191 was a tense & aspect item. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. We coded the items for a number of variables. We will discuss a few here. We first coded the item of hypothesized CEFR level. Then we coded the items grammatical knowledge—the possibilities were F, M or F&M After that, the items were coded for the type of grammatical form or meaning according to Purpura (2004). Item 191 was a MS form item designed to measure the past simple. Then, we coded the item for the specific grammatical focus. Item 191 was a tense & aspect item. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. KIRBY STARTS HERE Once we had a definition of grammatical knowledge, we prepared test specs and commissioned items to be written to specs. 03/27/10
03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. We are focusing on Task 1 for this paper. Task 1 – Testing knowledge of grammatical forms It is designed to measure the test takers’ knowledge of grammar. In this task, test takers are asked to read a short gapped dialogue and then complete the dialogue by selecting one of four options. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. The levels identified by the coders often differed by more than one level from the codings generated by Rasch. Given the number of discrepancies, this raised questions about which what set of item codings to believe as a basis for specifying what it means to have grammatical knowledge at what CEFR level—perhaps both. Even though item writers were asked to write developmentally sequenced distractors, the items were scored dichotomously by means of the simple Rasch model. This raises the question of the correspondence between the measurement model and the scoring method. Partial credit? An example of a discrepancy is when an item is judged to be a XX item on the CEFR scale, but the Rasch based coding rooted to item difficulty places the item at the XX level. The problem with using item difficulties is that difficulty is often a function of the key AND the options, but the analysis is base only on they key—partial credit?. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. The levels identified by the coders often differed by more than one level from the codings generated by Rasch. Given the number of discrepancies, this raised questions about which what set of item codings to believe as a basis for specifying what it means to have grammatical knowledge at what CEFR level—perhaps both. Even though item writers were asked to write developmentally sequenced distractors, the items were scored dichotomously by means of the simple Rasch model. This raises the question of the correspondence between the measurement model and the scoring method. Partial credit? An example of a discrepancy is when an item is judeged to be a XX item on the CEFR scale, but the Rasch based coding rooted to item difficulty places the item at the XX level. The problem with using item difficulties is that difficulty is often a function of the key AND the options, but the analysis is base only on they key—partial credit?. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. JIM STARTS AGAIN HERE The levels identified by the coders often differed by more than one level from the codings generated by Rasch. Given the number of discrepancies, this raised questions about which what set of item codings to believe as a basis for specifying what it means to have grammatical knowledge at what CEFR level—perhaps both. Even though item writers were asked to write developmentally sequenced distractors, the items were scored dichotomously by means of the simple Rasch model. This raises the question of the correspondence between the measurement model and the scoring method. Partial credit? An example of a discrepancy is when an item is judeged to be a XX item on the CEFR scale, but the Rasch based coding rooted to item difficulty places the item at the XX level. The problem with using item difficulties is that difficulty is often a function of the key AND the options, but the analysis is base only on they key—partial credit?. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. The levels identified by the coders often differed by more than one level from the codings generated by Rasch. Given the number of discrepancies, this raised questions about which what set of item codings to believe as a basis for specifying what it means to have grammatical knowledge at what CEFR level—perhaps both. Even though item writers were asked to write developmentally sequenced distractors, the items were scored dichotomously by means of the simple Rasch model. This raises the question of the correspondence between the measurement model and the scoring method. Partial credit? An example of a discrepancy is when an item is judeged to be a XX item on the CEFR scale, but the Rasch based coding rooted to item difficulty places the item at the XX level. The problem with using item difficulties is that difficulty is often a function of the key AND the options, but the analysis is base only on they key—partial credit?. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. The levels identified by the coders often differed by more than one level from the codings generated by Rasch. Given the number of discrepancies, this raised questions about which what set of item codings to believe as a basis for specifying what it means to have grammatical knowledge at what CEFR level—perhaps both. Even though item writers were asked to write developmentally sequenced distractors, the items were scored dichotomously by means of the simple Rasch model. This raises the question of the correspondence between the measurement model and the scoring method. Partial credit? An example of a discrepancy is when an item is judeged to be a XX item on the CEFR scale, but the Rasch based coding rooted to item difficulty places the item at the XX level. The problem with using item difficulties is that difficulty is often a function of the key AND the options, but the analysis is base only on they key—partial credit?. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. The levels identified by the coders often differed by more than one level from the codings generated by Rasch. Given the number of discrepancies, this raised questions about which what set of item codings to believe as a basis for specifying what it means to have grammatical knowledge at what CEFR level—perhaps both. Even though item writers were asked to write developmentally sequenced distractors, the items were scored dichotomously by means of the simple Rasch model. This raises the question of the correspondence between the measurement model and the scoring method. Partial credit? An example of a discrepancy is when an item is judeged to be a XX item on the CEFR scale, but the Rasch based coding rooted to item difficulty places the item at the XX level. The problem with using item difficulties is that difficulty is often a function of the key AND the options, but the analysis is base only on they key—partial credit?. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. The levels identified by the coders often differed by more than one level from the codings generated by Rasch. Given the number of discrepancies, this raised questions about which what set of item codings to believe as a basis for specifying what it means to have grammatical knowledge at what CEFR level—perhaps both. Even though item writers were asked to write developmentally sequenced distractors, the items were scored dichotomously by means of the simple Rasch model. This raises the question of the correspondence between the measurement model and the scoring method. Partial credit? An example of a discrepancy is when an item is judeged to be a XX item on the CEFR scale, but the Rasch based coding rooted to item difficulty places the item at the XX level. The problem with using item difficulties is that difficulty is often a function of the key AND the options, but the analysis is base only on they key—partial credit?. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. The levels identified by the coders often differed by more than one level from the codings generated by Rasch. Given the number of discrepancies, this raised questions about which what set of item codings to believe as a basis for specifying what it means to have grammatical knowledge at what CEFR level—perhaps both. Even though item writers were asked to write developmentally sequenced distractors, the items were scored dichotomously by means of the simple Rasch model. This raises the question of the correspondence between the measurement model and the scoring method. Partial credit? An example of a discrepancy is when an item is judeged to be a XX item on the CEFR scale, but the Rasch based coding rooted to item difficulty places the item at the XX level. The problem with using item difficulties is that difficulty is often a function of the key AND the options, but the analysis is base only on they key—partial credit?. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. The levels identified by the coders often differed by more than one level from the codings generated by Rasch. Given the number of discrepancies, this raised questions about which what set of item codings to believe as a basis for specifying what it means to have grammatical knowledge at what CEFR level—perhaps both. Even though item writers were asked to write developmentally sequenced distractors, the items were scored dichotomously by means of the simple Rasch model. This raises the question of the correspondence between the measurement model and the scoring method. Partial credit? An example of a discrepancy is when an item is judeged to be a XX item on the CEFR scale, but the Rasch based coding rooted to item difficulty places the item at the XX level. The problem with using item difficulties is that difficulty is often a function of the key AND the options, but the analysis is base only on they key—partial credit?. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. The levels identified by the coders often differed by more than one level from the codings generated by Rasch. Given the number of discrepancies, this raised questions about which what set of item codings to believe as a basis for specifying what it means to have grammatical knowledge at what CEFR level—perhaps both. Even though item writers were asked to write developmentally sequenced distractors, the items were scored dichotomously by means of the simple Rasch model. This raises the question of the correspondence between the measurement model and the scoring method. Partial credit? An example of a discrepancy is when an item is judeged to be a XX item on the CEFR scale, but the Rasch based coding rooted to item difficulty places the item at the XX level. The problem with using item difficulties is that difficulty is often a function of the key AND the options, but the analysis is base only on they key—partial credit?. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. The levels identified by the coders often differed by more than one level from the codings generated by Rasch. Given the number of discrepancies, this raised questions about which what set of item codings to believe as a basis for specifying what it means to have grammatical knowledge at what CEFR level—perhaps both. Even though item writers were asked to write developmentally sequenced distractors, the items were scored dichotomously by means of the simple Rasch model. This raises the question of the correspondence between the measurement model and the scoring method. Partial credit? An example of a discrepancy is when an item is judeged to be a XX item on the CEFR scale, but the Rasch based coding rooted to item difficulty places the item at the XX level. The problem with using item difficulties is that difficulty is often a function of the key AND the options, but the analysis is base only on they key—partial credit?. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. The levels identified by the coders often differed by more than one level from the codings generated by Rasch. Given the number of discrepancies, this raised questions about which what set of item codings to believe as a basis for specifying what it means to have grammatical knowledge at what CEFR level—perhaps both. Even though item writers were asked to write developmentally sequenced distractors, the items were scored dichotomously by means of the simple Rasch model. This raises the question of the correspondence between the measurement model and the scoring method. Partial credit? An example of a discrepancy is when an item is judeged to be a XX item on the CEFR scale, but the Rasch based coding rooted to item difficulty places the item at the XX level. The problem with using item difficulties is that difficulty is often a function of the key AND the options, but the analysis is base only on they key—partial credit?. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. The levels identified by the coders often differed by more than one level from the codings generated by Rasch. Given the number of discrepancies, this raised questions about which what set of item codings to believe as a basis for specifying what it means to have grammatical knowledge at what CEFR level—perhaps both. Even though item writers were asked to write developmentally sequenced distractors, the items were scored dichotomously by means of the simple Rasch model. This raises the question of the correspondence between the measurement model and the scoring method. Partial credit? An example of a discrepancy is when an item is judeged to be a XX item on the CEFR scale, but the Rasch based coding rooted to item difficulty places the item at the XX level. The problem with using item difficulties is that difficulty is often a function of the key AND the options, but the analysis is base only on they key—partial credit?. 03/27/10
03/27/10
TESOL 2010 Purpura & Beeston Grammatical knowledge is defined in terms of knowledge of grammatical forms and meanings…this includes a variety of forms and meanings at the sub-sentential, sentential, and discourse levels. Pragmatic knowledge is also part of langage abilty in this model, but task 1 is only explicitly measuring grammatical knowledge. Lexical forms at the advanced levels: Lexical meanings at the advanced levels • Noun compounding, • denotation, connotation, collocation • co-dependence restrictions (depend on) Morphosyntactic forms :Morphosyntactic meanings • inflectional & derivational affixes, • notions of time, space,, • syntactic structure, tense, • passivization • aspect, mood, complex/compound • interrogation sentence types Cohesive forms: Cohesive meanings • referential forms, ellipsis, • logical connectors Information management forms Information management meanings • given/new information, • emphasis, foregrounding • word order, parallelism 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. 03/27/10
03/27/10
TESOL 2010 Purpura & Beeston Grammatical knowledge is defined in terms of knowledge of grammatical forms and meanings…this includes a variety of forms and meanings at the sub-sentential, sentential, and discourse levels. Pragmatic knowledge is also part of langage abilty in this model, but task 1 is only explicitly measuring grammatical knowledge. Lexical forms at the advanced levels: Lexical meanings at the advanced levels • Noun compounding, • denotation, connotation, collocation • co-dependence restrictions (depend on) Morphosyntactic forms :Morphosyntactic meanings • inflectional & derivational affixes, • notions of time, space,, • syntactic structure, tense, • passivization • aspect, mood, complex/compound • interrogation sentence types Cohesive forms: Cohesive meanings • referential forms, ellipsis, • logical connectors Information management forms Information management meanings • given/new information, • emphasis, foregrounding • word order, parallelism 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. 03/27/10
03/27/10
TESOL 2010 Purpura & Beeston Grammatical knowledge is defined in terms of knowledge of grammatical forms and meanings…this includes a variety of forms and meanings at the sub-sentential, sentential, and discourse levels. Pragmatic knowledge is also part of langage abilty in this model, but task 1 is only explicitly measuring grammatical knowledge. Lexical forms at the advanced levels: Lexical meanings at the advanced levels • Noun compounding, • denotation, connotation, collocation • co-dependence restrictions (depend on) Morphosyntactic forms :Morphosyntactic meanings • inflectional & derivational affixes, • notions of time, space,, • syntactic structure, tense, • passivization • aspect, mood, complex/compound • interrogation sentence types Cohesive forms: Cohesive meanings • referential forms, ellipsis, • logical connectors Information management forms Information management meanings • given/new information, • emphasis, foregrounding • word order, parallelism 03/27/10
We scored these responses on multiple dimensions. TESOL 2010 Purpura, Grabowski, Dakin, Ameriks, Beeston
We scored these responses on multiple dimensions. TESOL 2010 Purpura, Grabowski, Dakin, Ameriks, Beeston
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. We coded the items for a number of variables. We will discuss a few here. We first coded the item of hypothesized CEFR level. Then we coded the items grammatical knowledge—the possibilities were F, M or F&M After that, the items were coded for the type of grammatical form or meaning according to Purpura (2004). Item 191 was a MS form item designed to measure the past simple. Then, we coded the item for the specific grammatical focus. Item 191 was a tense & aspect item.
03/27/10
A) Performance: Test Takers can successfully… • Use grammatical resources to understand the literal and intended meanings explicitly stated within the spoken text (i.e., literal & intended meaning). • Use grammatical resources to understand a range of implied or pragmatic meanings implicit within and beyond the spoken text (implied meaning). More specifically, depending on the proficiency level of the test, listeners must be able to infer: contextual meanings (e.g., Where does this conversation take place? What can you conclude from the conversation? What’s the main point of the lecture? ) sociolinguistic meanings (e.g., What is the relationship between the two speakers? ) sociocultural meanings (e.g., What does the speaker mean by “the Big Apple?” ) psychological meanings (e.g., What is the woman’s attitude toward the man? How do you think the man feels? What is the tone of the conversation? rhetorical meanings (e.g., What is the purpose of this announcement? How does the discussion end? What is the woman trying to do in this lecture? ) • The interpretation of meaning can be derived from explicit information within the spoken text. This is referred to as “endophoric reference” (Halliday & Hasan, 1989). For example, a traditional “detail” item like Where does the man live? might be coded as “literal meaning/endophoric reference” if the place the man lived is explicitly mentioned in the text. Similarly, a “gist” item, where the main idea is explicitly stated in the spoken text (e.g., “ The purpose of my talk today is to…” ) would be coded as “literal meaning/endophoric reference.” • The implied meaning of a spoken text can be derived from information implicit within the spoken text. This is called endophoric reference (Halliday & Hasan, 1989). For example, consider the “gist’ item: What is the speaker’s main message? If the main idea of the speaker’s message is not explicitly stated in the text, but can be inferred through implicit contextualization cues, then this item would be coded “implied (contextual) meaning/endophoric reference.” Consider also a lexical expression item such as: What does the man mean when he says “Suppose you had to be there.”? If the answer to this question is not explicitly stated in the spoken text, but can be inferred from the surrounding text, this item would also be coded “implied (contextual) meaning/endophoric reference.” Finally, consider an “inference” item, where the test taker is asked where the conversation seems to be taking place. If the location is not explicitly stated, but can be inferred from lexical cues in the co-text, this item would be coded “implied (contextual) meaning/endophoric reference.” • The implied meaning of a spoken text can also be derived implicitly from information that lies outside the spoken text itself. This is called exophoric reference (Halliday & Hasan, 1989). These extensions of meaning draw on (1) the overall context of the situation, (2) the norms, assumption, expectations and presuppositions associated with the situational context, and (3) the background knowledge of the listener in relation to that context. For example, consider the following “tone” items: What is the tone of the conversation? or How do you think [person X]) feels about the situation?” If the answer to these questions are not explicitly stated in the spoken text, but can be inferred, not so much from the surrounding text, but from the context of the situation coupled with the norms and expectations of interaction in this situation, this item would be coded “implied (psychological) meaning/exophoric reference.” In this case, the pragmatic meaning clearly lies outside the spoken text itself and in the context of the situation. • Note that sociocultural references (e.g., the Big Apple, the Windy City, the Angel of the North, the Bull Ring) should be avoided if they are deemed outside the reference of the Oxford Test candidature. Nonetheless, sociocultural references are a part of the living language, and on some level, it is impossible to avoid them for certain populations (e.g., some students might not be familiar with train or airplane timetables). If the reference is clearly outside the reference, dialogues should be written so that the meaning can be derived from within the text, perhaps as a repair strategy. For example: A: I’m off to the Big Apple tomorrow. B: What? A: You know… New York City. B: Oh, I see. How long are you going for? B) Focus: Test Takers need to have knowledge of … • Grammatical Forms & Meanings : This task requires examinees to know a range of grammatical forms so their literal and intended meaning s can be understood from spoken text (See Purpura, 2004). • Pragmatic Meanings: This task also requires examinees to be able to use grammatical resources to understand a range of pragmatic (implied) meanings (See Purpura, 2004). More specifically, depending on the proficiency level, this task may require the ability to interpret: contextual meanings (e.g., situational features) sociolinguistic meanings (e.g., social identity markers referring to gender, age, status; social meanings pointing to power, politeness and formality; and social norms, preferences, and situational expectations) sociocultural meanings (e.g., cultural references, figurative meanings, metaphor) psychological meanings (e.g., affective stance, point of view, tone) rhetorical meanings (e.g., coherence and genre)
A) Performance: Test Takers can successfully… • Use grammatical resources to understand the literal and intended meanings explicitly stated within the spoken text (i.e., literal & intended meaning). • Use grammatical resources to understand a range of implied or pragmatic meanings implicit within and beyond the spoken text (implied meaning). More specifically, depending on the proficiency level of the test, listeners must be able to infer: contextual meanings (e.g., Where does this conversation take place? What can you conclude from the conversation? What’s the main point of the lecture? ) sociolinguistic meanings (e.g., What is the relationship between the two speakers? ) sociocultural meanings (e.g., What does the speaker mean by “the Big Apple?” ) psychological meanings (e.g., What is the woman’s attitude toward the man? How do you think the man feels? What is the tone of the conversation? rhetorical meanings (e.g., What is the purpose of this announcement? How does the discussion end? What is the woman trying to do in this lecture? ) • The interpretation of meaning can be derived from explicit information within the spoken text. This is referred to as “endophoric reference” (Halliday & Hasan, 1989). For example, a traditional “detail” item like Where does the man live? might be coded as “literal meaning/endophoric reference” if the place the man lived is explicitly mentioned in the text. Similarly, a “gist” item, where the main idea is explicitly stated in the spoken text (e.g., “ The purpose of my talk today is to…” ) would be coded as “literal meaning/endophoric reference.” • The implied meaning of a spoken text can be derived from information implicit within the spoken text. This is called endophoric reference (Halliday & Hasan, 1989). For example, consider the “gist’ item: What is the speaker’s main message? If the main idea of the speaker’s message is not explicitly stated in the text, but can be inferred through implicit contextualization cues, then this item would be coded “implied (contextual) meaning/endophoric reference.” Consider also a lexical expression item such as: What does the man mean when he says “Suppose you had to be there.”? If the answer to this question is not explicitly stated in the spoken text, but can be inferred from the surrounding text, this item would also be coded “implied (contextual) meaning/endophoric reference.” Finally, consider an “inference” item, where the test taker is asked where the conversation seems to be taking place. If the location is not explicitly stated, but can be inferred from lexical cues in the co-text, this item would be coded “implied (contextual) meaning/endophoric reference.” • The implied meaning of a spoken text can also be derived implicitly from information that lies outside the spoken text itself. This is called exophoric reference (Halliday & Hasan, 1989). These extensions of meaning draw on (1) the overall context of the situation, (2) the norms, assumption, expectations and presuppositions associated with the situational context, and (3) the background knowledge of the listener in relation to that context. For example, consider the following “tone” items: What is the tone of the conversation? or How do you think [person X]) feels about the situation?” If the answer to these questions are not explicitly stated in the spoken text, but can be inferred, not so much from the surrounding text, but from the context of the situation coupled with the norms and expectations of interaction in this situation, this item would be coded “implied (psychological) meaning/exophoric reference.” In this case, the pragmatic meaning clearly lies outside the spoken text itself and in the context of the situation. • Note that sociocultural references (e.g., the Big Apple, the Windy City, the Angel of the North, the Bull Ring) should be avoided if they are deemed outside the reference of the Oxford Test candidature. Nonetheless, sociocultural references are a part of the living language, and on some level, it is impossible to avoid them for certain populations (e.g., some students might not be familiar with train or airplane timetables). If the reference is clearly outside the reference, dialogues should be written so that the meaning can be derived from within the text, perhaps as a repair strategy. For example: A: I’m off to the Big Apple tomorrow. B: What? A: You know… New York City. B: Oh, I see. How long are you going for? B) Focus: Test Takers need to have knowledge of … • Grammatical Forms & Meanings : This task requires examinees to know a range of grammatical forms so their literal and intended meaning s can be understood from spoken text (See Purpura, 2004). • Pragmatic Meanings: This task also requires examinees to be able to use grammatical resources to understand a range of pragmatic (implied) meanings (See Purpura, 2004). More specifically, depending on the proficiency level, this task may require the ability to interpret: contextual meanings (e.g., situational features) sociolinguistic meanings (e.g., social identity markers referring to gender, age, status; social meanings pointing to power, politeness and formality; and social norms, preferences, and situational expectations) sociocultural meanings (e.g., cultural references, figurative meanings, metaphor) psychological meanings (e.g., affective stance, point of view, tone) rhetorical meanings (e.g., coherence and genre)
A) Performance: Test Takers can successfully… • Use grammatical resources to understand the literal and intended meanings explicitly stated within the spoken text (i.e., literal & intended meaning). • Use grammatical resources to understand a range of implied or pragmatic meanings implicit within and beyond the spoken text (implied meaning). More specifically, depending on the proficiency level of the test, listeners must be able to infer: contextual meanings (e.g., Where does this conversation take place? What can you conclude from the conversation? What’s the main point of the lecture? ) sociolinguistic meanings (e.g., What is the relationship between the two speakers? ) sociocultural meanings (e.g., What does the speaker mean by “the Big Apple?” ) psychological meanings (e.g., What is the woman’s attitude toward the man? How do you think the man feels? What is the tone of the conversation? rhetorical meanings (e.g., What is the purpose of this announcement? How does the discussion end? What is the woman trying to do in this lecture? ) • The interpretation of meaning can be derived from explicit information within the spoken text. This is referred to as “endophoric reference” (Halliday & Hasan, 1989). For example, a traditional “detail” item like Where does the man live? might be coded as “literal meaning/endophoric reference” if the place the man lived is explicitly mentioned in the text. Similarly, a “gist” item, where the main idea is explicitly stated in the spoken text (e.g., “ The purpose of my talk today is to…” ) would be coded as “literal meaning/endophoric reference.” • The implied meaning of a spoken text can be derived from information implicit within the spoken text. This is called endophoric reference (Halliday & Hasan, 1989). For example, consider the “gist’ item: What is the speaker’s main message? If the main idea of the speaker’s message is not explicitly stated in the text, but can be inferred through implicit contextualization cues, then this item would be coded “implied (contextual) meaning/endophoric reference.” Consider also a lexical expression item such as: What does the man mean when he says “Suppose you had to be there.”? If the answer to this question is not explicitly stated in the spoken text, but can be inferred from the surrounding text, this item would also be coded “implied (contextual) meaning/endophoric reference.” Finally, consider an “inference” item, where the test taker is asked where the conversation seems to be taking place. If the location is not explicitly stated, but can be inferred from lexical cues in the co-text, this item would be coded “implied (contextual) meaning/endophoric reference.” • The implied meaning of a spoken text can also be derived implicitly from information that lies outside the spoken text itself. This is called exophoric reference (Halliday & Hasan, 1989). These extensions of meaning draw on (1) the overall context of the situation, (2) the norms, assumption, expectations and presuppositions associated with the situational context, and (3) the background knowledge of the listener in relation to that context. For example, consider the following “tone” items: What is the tone of the conversation? or How do you think [person X]) feels about the situation?” If the answer to these questions are not explicitly stated in the spoken text, but can be inferred, not so much from the surrounding text, but from the context of the situation coupled with the norms and expectations of interaction in this situation, this item would be coded “implied (psychological) meaning/exophoric reference.” In this case, the pragmatic meaning clearly lies outside the spoken text itself and in the context of the situation. • Note that sociocultural references (e.g., the Big Apple, the Windy City, the Angel of the North, the Bull Ring) should be avoided if they are deemed outside the reference of the Oxford Test candidature. Nonetheless, sociocultural references are a part of the living language, and on some level, it is impossible to avoid them for certain populations (e.g., some students might not be familiar with train or airplane timetables). If the reference is clearly outside the reference, dialogues should be written so that the meaning can be derived from within the text, perhaps as a repair strategy. For example: A: I’m off to the Big Apple tomorrow. B: What? A: You know… New York City. B: Oh, I see. How long are you going for? B) Focus: Test Takers need to have knowledge of … • Grammatical Forms & Meanings : This task requires examinees to know a range of grammatical forms so their literal and intended meaning s can be understood from spoken text (See Purpura, 2004). • Pragmatic Meanings: This task also requires examinees to be able to use grammatical resources to understand a range of pragmatic (implied) meanings (See Purpura, 2004). More specifically, depending on the proficiency level, this task may require the ability to interpret: contextual meanings (e.g., situational features) sociolinguistic meanings (e.g., social identity markers referring to gender, age, status; social meanings pointing to power, politeness and formality; and social norms, preferences, and situational expectations) sociocultural meanings (e.g., cultural references, figurative meanings, metaphor) psychological meanings (e.g., affective stance, point of view, tone) rhetorical meanings (e.g., coherence and genre)
A) Performance: Test Takers can successfully… • Use grammatical resources to understand the literal and intended meanings explicitly stated within the spoken text (i.e., literal & intended meaning). • Use grammatical resources to understand a range of implied or pragmatic meanings implicit within and beyond the spoken text (implied meaning). More specifically, depending on the proficiency level of the test, listeners must be able to infer: contextual meanings (e.g., Where does this conversation take place? What can you conclude from the conversation? What’s the main point of the lecture? ) sociolinguistic meanings (e.g., What is the relationship between the two speakers? ) sociocultural meanings (e.g., What does the speaker mean by “the Big Apple?” ) psychological meanings (e.g., What is the woman’s attitude toward the man? How do you think the man feels? What is the tone of the conversation? rhetorical meanings (e.g., What is the purpose of this announcement? How does the discussion end? What is the woman trying to do in this lecture? ) • The interpretation of meaning can be derived from explicit information within the spoken text. This is referred to as “endophoric reference” (Halliday & Hasan, 1989). For example, a traditional “detail” item like Where does the man live? might be coded as “literal meaning/endophoric reference” if the place the man lived is explicitly mentioned in the text. Similarly, a “gist” item, where the main idea is explicitly stated in the spoken text (e.g., “ The purpose of my talk today is to…” ) would be coded as “literal meaning/endophoric reference.” • The implied meaning of a spoken text can be derived from information implicit within the spoken text. This is called endophoric reference (Halliday & Hasan, 1989). For example, consider the “gist’ item: What is the speaker’s main message? If the main idea of the speaker’s message is not explicitly stated in the text, but can be inferred through implicit contextualization cues, then this item would be coded “implied (contextual) meaning/endophoric reference.” Consider also a lexical expression item such as: What does the man mean when he says “Suppose you had to be there.”? If the answer to this question is not explicitly stated in the spoken text, but can be inferred from the surrounding text, this item would also be coded “implied (contextual) meaning/endophoric reference.” Finally, consider an “inference” item, where the test taker is asked where the conversation seems to be taking place. If the location is not explicitly stated, but can be inferred from lexical cues in the co-text, this item would be coded “implied (contextual) meaning/endophoric reference.” • The implied meaning of a spoken text can also be derived implicitly from information that lies outside the spoken text itself. This is called exophoric reference (Halliday & Hasan, 1989). These extensions of meaning draw on (1) the overall context of the situation, (2) the norms, assumption, expectations and presuppositions associated with the situational context, and (3) the background knowledge of the listener in relation to that context. For example, consider the following “tone” items: What is the tone of the conversation? or How do you think [person X]) feels about the situation?” If the answer to these questions are not explicitly stated in the spoken text, but can be inferred, not so much from the surrounding text, but from the context of the situation coupled with the norms and expectations of interaction in this situation, this item would be coded “implied (psychological) meaning/exophoric reference.” In this case, the pragmatic meaning clearly lies outside the spoken text itself and in the context of the situation. • Note that sociocultural references (e.g., the Big Apple, the Windy City, the Angel of the North, the Bull Ring) should be avoided if they are deemed outside the reference of the Oxford Test candidature. Nonetheless, sociocultural references are a part of the living language, and on some level, it is impossible to avoid them for certain populations (e.g., some students might not be familiar with train or airplane timetables). If the reference is clearly outside the reference, dialogues should be written so that the meaning can be derived from within the text, perhaps as a repair strategy. For example: A: I’m off to the Big Apple tomorrow. B: What? A: You know… New York City. B: Oh, I see. How long are you going for? B) Focus: Test Takers need to have knowledge of … • Grammatical Forms & Meanings : This task requires examinees to know a range of grammatical forms so their literal and intended meaning s can be understood from spoken text (See Purpura, 2004). • Pragmatic Meanings: This task also requires examinees to be able to use grammatical resources to understand a range of pragmatic (implied) meanings (See Purpura, 2004). More specifically, depending on the proficiency level, this task may require the ability to interpret: contextual meanings (e.g., situational features) sociolinguistic meanings (e.g., social identity markers referring to gender, age, status; social meanings pointing to power, politeness and formality; and social norms, preferences, and situational expectations) sociocultural meanings (e.g., cultural references, figurative meanings, metaphor) psychological meanings (e.g., affective stance, point of view, tone) rhetorical meanings (e.g., coherence and genre)
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. 03/27/10
ILACA 2009 03/27/10 James E Purpura, Teachers College, Columbia Univ. The levels identified by the coders often differed by more than one level from the codings generated by Rasch. Given the number of discrepancies, this raised questions about which what set of item codings to believe as a basis for specifying what it means to have grammatical knowledge at what CEFR level—perhaps both. Even though item writers were asked to write developmentally sequenced distractors, the items were scored dichotomously by means of the simple Rasch model. This raises the question of the correspondence between the measurement model and the scoring method. Partial credit? An example of a discrepancy is when an item is judeged to be a XX item on the CEFR scale, but the Rasch based coding rooted to item difficulty places the item at the XX level. The problem with using item difficulties is that difficulty is often a function of the key AND the options, but the analysis is base only on they key—partial credit?. 03/27/10