|
Job Discussion Forums "The Internet's Meeting Place for ESL/EFL Students and Teachers from Around the World!"
|
| View previous topic :: View next topic |
| Author |
Message |
LongShiKong
Joined: 28 May 2007 Posts: 1082 Location: China
|
Posted: Tue Apr 17, 2012 9:31 am Post subject: |
|
|
| Sashadroogie wrote: |
| Is this a matter of test reliability or validity? |
If reliability refers to the extent to which assessments are consistent then if the test is not clearly written then one assessor may give an 'A' while another may give a 'C'. I've encountered such ambiguous tests more than once where they're too subjective to be reliable in terms of how to score a response. |
|
| Back to top |
|
 |
Sashadroogie

Joined: 17 Apr 2007 Posts: 11061 Location: Moskva, The Workers' Paradise
|
Posted: Tue Apr 17, 2012 9:42 am Post subject: |
|
|
If the student is giving 'the right answer' then it seems to me that issues of validity are involved, i.e. is the test testing what it is supposed to be?
In any case, it seems to be more a case of examiners not correctly applying the criteria than a problem with test design. Which tests are you referring to by the way? |
|
| Back to top |
|
 |
LongShiKong
Joined: 28 May 2007 Posts: 1082 Location: China
|
Posted: Tue Apr 17, 2012 2:43 pm Post subject: |
|
|
| Sashadroogie wrote: |
| I am not sure what relationship you see between doing the Delta and IELTS examining, or the exact point being made. |
I guess IELTS came to mind as a shining example of what your DELTA covered--reliability and validity in ELT testing although a few weeks ago I came across an opinion poll regarding consistency differences between IELTS examiners/test centers. I'd like to think IELTS and TOEFL set a standard for reliability and validity.
| Sashadroogie wrote: |
| If the student is giving 'the right answer' then it seems to me that issues of validity are involved, i.e. is the test testing what it is supposed to be? |
Yes, you're right. The exam instruction was: "Any factually correct answer is acceptable." but I personally chose to award only half marks for single word answers or complete grammatically accurate but incorrect answers.
Test material, along with placement tests as I'd mentioned, is routinely pulled directly from coursebooks without any change (both from private schools but also from industry publications)--this doesn't test language competency but text familiarity. I remember one instance where the question accompanying a set of activity illustrations surprisingly differed from the coursebook. I think it was simple past on the test but past continuous in their books. Needless to say, a certain percent of students didn't listen to the question grammar and answered according to the book. |
|
| Back to top |
|
 |
spiral78

Joined: 05 Apr 2004 Posts: 11534 Location: On a Short Leash
|
Posted: Tue Apr 17, 2012 4:02 pm Post subject: |
|
|
| Quote: |
| but I personally chose to award only half marks for single word answers or complete grammatically accurate but incorrect answers. |
Whether the single word answers were acceptable would really depend on if the test is measuring comprehension (receptive skill) or writing/other production skills. On comprehension tests, I've seen some single-word answers that were really perfectly acceptable in terms of indicating a students' sufficient comprehension of the meaning of a text, and, as you point out, grammatically perfect but incorrect answers are not acceptable. |
|
| Back to top |
|
 |
LongShiKong
Joined: 28 May 2007 Posts: 1082 Location: China
|
Posted: Tue Apr 17, 2012 4:28 pm Post subject: |
|
|
| This was an oral test. With no other instruction, I chose to interpret the 'acceptable' as meaning 'passable' a 50%, not 'perfect'. For many of these tests, the syllabus explicitly mapped out such responses as falling within the domain of expressive language. |
|
| Back to top |
|
 |
Sashadroogie

Joined: 17 Apr 2007 Posts: 11061 Location: Moskva, The Workers' Paradise
|
Posted: Tue Apr 17, 2012 6:36 pm Post subject: |
|
|
If you are in a situation where you can/have to choose to interpret what the criteria are or mean than the entire test seems to be lacking everything: all types of validity, but especially construct validity. What skills was the test supposed to assess exactly? You have my sympathies.
Who made these tests? Professional materials designers? Or some unqualified teachers rushing around a staff room trying to cobble something together fast? |
|
| Back to top |
|
 |
johnslat

Joined: 21 Jan 2003 Posts: 13859 Location: Santa Fe, New Mexico, USA
|
Posted: Tue Apr 17, 2012 6:43 pm Post subject: |
|
|
Dear Sasha,
"Professional materials designers? Or some unqualified teachers rushing around a staff room trying to cobble something together fast?"
Aren't they the same people?
Regards,
John |
|
| Back to top |
|
 |
Sashadroogie

Joined: 17 Apr 2007 Posts: 11061 Location: Moskva, The Workers' Paradise
|
Posted: Tue Apr 17, 2012 6:52 pm Post subject: |
|
|
Dear Johnslat
Certainly not! At least not in the hallowed halls of Cambridge. Perhaps in whichever place in China poor Longshikong is working.
Off-his-face-validity Sasha |
|
| Back to top |
|
 |
timothypfox
Joined: 20 Feb 2008 Posts: 492
|
Posted: Wed Apr 18, 2012 12:11 am Post subject: |
|
|
Yes, my MS TESOL did. But, I grew very disillusioned with assessment when I worked for public school in NYC. My job was collecting data on ESL high school students with moderate disabilities. Invariable the assessments confused measuring the students language level with their various disabilities such as language delays, ADHD, lower level cognitive functions, and speech and language deficits.
And, don't get me started on standardized testing! I am of the generation that had none, and did just fine!
I now work with EFL high school students at a Japanese high school. Assessment is done through written tests either writing or listening. The written tests cover a topic they practiced in class. During classes their work is edited using a marking guide indicating the type of errors, so that they learn to correct their own mistakes.
I also have junior high classes that I give speaking tests one by one too. Sections include memorizing a part of 3 dialogues that they perform with me, an oral spelling test, proper oral response to question structures we practiced, and reading a short passage aloud with correct pronunciation. Each section is marked using a 5 level rubric (each level has a description of achievement), and the total is then added for a final score...
Having said all this - Japanese schools still teach heavily with the grammar translation approach even though US programs adhere more to the communicative approach to instruction and assessment. US schools are so slanted towards collaborative group work and student empowerment that high level student grow bored as beaurocrats push "no child left behind...." Love it , or lump it! |
|
| Back to top |
|
 |
LongShiKong
Joined: 28 May 2007 Posts: 1082 Location: China
|
Posted: Wed Apr 18, 2012 2:46 am Post subject: |
|
|
| Sashadroogie wrote: |
| If you are in a situation where you can/have to choose to interpret what the criteria are or mean .... |
Among the 70+ franchises who used these oral tests (there were no written tests), I'm sure I was the only one who took the liberty of interpreting 'acceptable' in light of the syllabus expectation. Everyone else awarded either a '0' or a '1'. Surprisingly, my overall class scores exceeded the average for all schools sometimes by a factor of twice as many 'A's.
| timothypfox wrote: |
| I now work with EFL high school students at a Japanese high school. Assessment is done through written tests either writing or listening. The written tests cover a topic they practiced in class. During classes their work is edited using a marking guide indicating the type of errors, so that they learn to correct their own mistakes. |
Would you say the writing test scoring system and the marking guides encourage or discourage risk-taking in their writing? No one wants to make mistakes (lose face) but making mistakes and learning from them are an integral part of learning.
As for memorizing dialogues to perfection or displaying mastery of pronunciation, what's the point? English isn't martial art, it's just a language. |
|
| Back to top |
|
 |
Sashadroogie

Joined: 17 Apr 2007 Posts: 11061 Location: Moskva, The Workers' Paradise
|
Posted: Wed Apr 18, 2012 3:48 am Post subject: |
|
|
Well, if this is a ready-made placement test, then I'm not sure that using it 'in light of syllabus expectation' is in keeping with the test's original purpose. And again, if not all examiners are applying the criteria consistently, whatever they may be, then expect skewed results.
There is always some degree of subjectivity in assessing written and spoken language. But what you are describing seems to transgress many basic norms in language testing. Have a closer look at Cambridge and IELTS exams and tests, and refer to the band descriptors. There are public versions available on the net. As you'll see, the criteria are quite well-defined, in most cases, and so long as examiners actually stick to the criteria they are supposed to apply, then reliability will remain high. (I wouldn't place much credence in whatever 'opinion polls' you've come across either...)
Can you tell us what test you and your franchise are using? Which industry publications? It may help to understand your situation a little better. |
|
| Back to top |
|
 |
LongShiKong
Joined: 28 May 2007 Posts: 1082 Location: China
|
Posted: Thu Apr 19, 2012 1:50 am Post subject: |
|
|
| Sashadroogie wrote: |
| Well, if this is a ready-made placement test, then I'm not sure that using it 'in light of syllabus expectation' is in keeping with the test's original purpose. |
No, this was the mid and final oral test instruction accompanying a series of tests not just for preschoolers but for children ranging from 6-12 yr olds. We were explicitly told not to vary from the way the test was written. Surprisingly, the test instruction was actually revised for all tests one term to also reflect grammatical accuracy (without any explanation from head office). However, subsequent tests, for whatever reason, reverted back to 'Any factually correct answer is acceptable.' The 'negative backwash' from such tests is blatantly obvious.
| Sashadroogie wrote: |
| There is always some degree of subjectivity in assessing written and spoken language. But what you are describing seems to transgress many basic norms in language testing. |
Go back to pg 1 and read my opening quote from Hughes. This transgression is the norm--tests such as IELTS are the exception. I'm now working for another such franchise system where there is no explicit instruction on how to score a factually correct response vs a complete, accurate one. And, once again, the validity of the test is in question given the focus of instruction. At least at Longman Schools, there were no 'standardized tests'; teachers were expected to come up with their own--I doubt many did. |
|
| Back to top |
|
 |
Sashadroogie

Joined: 17 Apr 2007 Posts: 11061 Location: Moskva, The Workers' Paradise
|
Posted: Thu Apr 19, 2012 5:57 am Post subject: |
|
|
I'm not sure that Hughes has said that - at least not in your quotation. Phrases like 'too often' are not the same as poor testing is the norm.
In your specific situation, this seems to boil down to a case where nobody who is designing/selecting the test materials is qualified to do so. It may even come down to lack of examiner training also. This would seem to me to be because teachers (and school administration) are not the same manner of being as test materials designers. Nor should they be. It's a specialised area - which is why it is probably not covered very much in most TEFL-related courses.
Teachers often confuse listening exercises designed to test comprehension with listening skills lessons designed to teach specific listening skills. However, it is also true that we can confuse teacher-training courses with qualifications in testing and assessment. Of course it is helpful for all teachers to be familiar with the basic principles of testing, and even to be able to put together a reasonable test themselves, for their own students etc. But designing a battery of test items for a whole school, where the stakes may be very high indeed... this is a very different matter entirely.
Being able to drop in technical terms like 'concurrent validity' into staff room conversations is not the same as being fully competent in a very technical area. G-theory? Classical true Score theory? Most teachers have not got any real clue about the intricacies of all this - and I include myself in that category. So, it is hardly surprising that teachers, according to Hughes, have a mistrust of testing: they have not ever experienced real testing by real testers. Instead they have had to deal with what you have described, with the predictable confusion arising therein.
Teaching and testing. Different areas. |
|
| Back to top |
|
 |
Sashadroogie

Joined: 17 Apr 2007 Posts: 11061 Location: Moskva, The Workers' Paradise
|
Posted: Thu Apr 19, 2012 6:02 am Post subject: |
|
|
| LongShiKong wrote: |
No, this was the mid and final oral test instruction accompanying a series of tests not just for preschoolers but for children ranging from 6-12 yr olds. We were explicitly told not to vary from the way the test was written. Surprisingly, the test instruction was actually revised for all tests one term to also reflect grammatical accuracy (without any explanation from head office). However, subsequent tests, for whatever reason, reverted back to 'Any factually correct answer is acceptable.' The 'negative backwash' from such tests is blatantly obvious.
|
For testing young learners, getting any verbal response at all is usually a bonus. Cambridge tests for very young learners merely require some sort of gesture, e.g. pointing at a picture.
'Factually correct answer' still sounds so wrong to me though. Can you post up an example from the tests of what you mean? |
|
| Back to top |
|
 |
LongShiKong
Joined: 28 May 2007 Posts: 1082 Location: China
|
Posted: Fri Apr 20, 2012 2:16 am Post subject: |
|
|
| Sashadroogie wrote: |
| I'm not sure that Hughes has said that - at least not in your quotation. Phrases like 'too often' are not the same as poor testing is the norm. |
How do others feel about this? Are well-written tests the exception or the norm?
| Sashadroogie wrote: |
| ...teachers (and school administration) are not the same manner of being as test materials designers. Nor should they be. It's a specialised area - which is why it is probably not covered very much in most TEFL-related courses. |
It may be specialized if you're writing tests for immigrant screening or post secondary admission but public school teachers write their own tests without such training. What my B.Ed covered on assessment could've been condensed to less than a 30 pg handout. Furthermore, on practicum, I was being evaluated for my assessment prior to even beginning the assessment module on campus.
In the preface, Hughes starts by saying: "The simple objective of this book is to help language teachers write better tests." Nowhere does he deplore the lack of trained test writers in the industry. Furthermore, he cites teachers, not test writers, as being directly responsible for the improvement of TOEFL:
| Quote: |
For the reader who doubts that teachers can influence the large testing institutions, let this chapter end with a further reference to the testing of writing through multiple choice items. This was the practice followed by those responsible for TOEFL, the test taken by most non-native speakers of English applying to North American universities. Over a period of many years they maintained that it was simply not possible to test the writing ability of hundreds of thousands of candidates by means of a composition: it was impracticable and the results, anyhow, would be unreliable. Yet in 1986 a writing test (Test of Written English), in which candidates actually have to write for thirty minutes, was introduced as a supplement to TOEFL, and already many colleges in the United States are requiring applicants to take this test in addition to TOEFL. The principal reason given for this change was pressure from English language teachers who had finally convinced those responsible for the TOEFL of the overriding need for a writing task which would provide beneficial backwash.
|
----------------------------------------------------------
| Sashadroogie wrote: |
| For testing young learners, getting any verbal response at all is usually a bonus. Cambridge tests for very young learners merely require some sort of gesture, e.g. pointing at a picture. |
Sewer skewer!!! If they're old enough to speak their language, they're old enough to speak English. The problem here isn't with assessment, it's with teaching. If you lower the bar and treat them as Pavlovian dogs, they'll respond that way. 'Any factually correct answer is acceptable' lowers the same bar to maintain a significant 'success' rate to convince parents they're child is learning so they'll sign up for more classes. As an example, if you ask: What are they?, What do monkeys like? or What are they eating? and the chlld responds 'Banana'. You award them a complete point.
Keep in mind that the majority of us (?) teach children and teens but none of us spend more than an hour or 2 of our cert training on how to teach them. The 'A' in DELTA and CELTA stands for adults. When's the last time you'd seen a job offer require a CELTA-YL? Is there even such a thing as a DELTA-YL? |
|
| Back to top |
|
 |
|
|
You cannot post new topics in this forum You cannot reply to topics in this forum You cannot edit your posts in this forum You cannot delete your posts in this forum You cannot vote in polls in this forum
|
This page is maintained by the one and only Dave Sperling. Contact Dave's ESL Cafe
Copyright © 2018 Dave Sperling. All Rights Reserved.
Powered by phpBB © 2001, 2002 phpBB Group
|