Beruflich Dokumente
Kultur Dokumente
David D. Qian
The Hong Kong Polytechnic University
Hung Hom. Kowloon. Hong Kong
David Qian@pol\>u. edu. hk
Abstract • Iti order to tnotivate university students to improve their English profi-
ciency, the Hong Kong govemtnent decided to adopt a common exit English language
test for all graduating students. In the process ot selecting a suitable measure for this
purpose, an empirical study with a sample of over 240 students was conducted to
compare two English proficiency tests, the English Test ofthe Graduating Students'
Language Proficiency Assessment (GSLPA) and the Academic Version ofthe Inter-
national English Language Testing System (IELTS). The comparisons covered both
speaking and writing components ofthe two tests, based on information provided by
the two tests as well as performances ofthe participating candidates on the compo-
nents under study. Results indieate: (1) that GSLPA writing and speaking scores
distinguish candidates" abilities more clearly than the corresponding scores on the
IELTS but IELTS overall scores, generated from writing, speaking, reading and listen-
ing sub-scores, have a discriminating power similar to that of GSLPA; (2) that the
GSLPA and IELTS writing subtests basically measure different skills; and (3) that
the constructs ofthe GSLPA and IELTS speaking subtests, while having their own
distinctive feattires, overlap by about 48%. This paper recommends options for Improv-
ing the current assessment framework for graduating university sttidents and discusses
the possibie impact of such a test on English language teaching and learning.
Keywords • English proficieney test, GSLPA, Hong Kong, IELTS, perfonnance test,
speaking assesment, writing test, university exit test.
Background
In Hong Kong, English language proficiency is an important factor for
new university graduates in securing employment (Qian 2005). When
applying for a job, new graduates are often required to show some evi-
GSLPA
First implemented in the 1999/2000 academic year, the GSLPA has been
a formal English language exit test at HKPU for three years. During the
three-year period of 1999/2000 and 2001/2002, over 6,500 HKPU final-
year students sat the test, which was also made available to students from
Lingnan University in Hong Kong.
The GSLPA test aims to provide infonnation for prospective employers
and otber interested parties about the English language proficiency of stu-
dents around the time of graduation. Its content, therefore, looks forward
to the context of professional employment rather than backwards at the
context of using language for academic purposes (Lumley and Qian 2003).
The GSLPA was initially developed with UGC research funds at HKPU.
Its continuous development has received expert input from language testing
specialistsfrom the UK, USA, Canada, Australia and other countries. It is a
wholly workplace-oriented, task-based perfonnance test mainly consisting
of two components, writing and speaking.
The writing component, consisting of three tasks, lasts 105 minutes.
The first two tasks (A and B) require the candidate to write a memo or
professional letter. Task A is intended to be straightforward, addressing a
relatively unproblematic situation, for which most necessary infonnation
is supplied as writing prompts. Task B, which carries more weight for
scoring, is designed to be more cognitively and linguistically challenging.
The task requires some kind of problem solving or argument on the part of
the candidate in a workplace context, with careful attention to audience,
register and communicative strategies. Candidates are given sufficient
time (90 minutes for both tasks) and space to draft their responses before
writing the fmal version for submission. Task C takes the fonn of a proof-
reading and error correction task, based on a passage supplied, on a theme
related to modem world workplaces. This task contributes only in a minor
way to the final band score of writing.
21
Assessing University Students
IELTS
The IELTS test is jointly provided by three organizations: the British
Council, IDP: IELTS Australia atid Cambridge ESOL, with its develop-
ment and validation unit based in Cambridge, UK.
There are two versions ofthe IELTS test: the Academic Modules and
the General Modules. Both versions contain four components: listening,
speaking, reading and writing. According to the IELTS handbook, the Aca-
demic Modules ofthe IELTS are designed to assess a candidate's English
language proficiency for academic studies at the undergraduate or post-
22
Regional Language Centre Journal 38.1
graduate level, whereas the General Modules are developed to assess test
candidates who intend to go to English-speaking countries to complete
their secondary education or undertake work experience or training pro-
grammes at below degree level. People who need to demonstrate their
English proficiency in order to immigrate to Australia or New Zealand are
also required to sit the General Modules (British Council, IDP: IELTS
Australia and University of Cambridge ESOL Examinations 2005).
The listening subtest, which lasts 30 minutes, is composed of four sec-
tions including two monologues and two conversations. Forty questions in
a variety of item types are asked in this subtest, which include multiple
choice, short-answer questions, sentence completion, table/chart comple-
tion, and matching.
The reading subtest, lasting for 60 minutes, contains 40 questions based
on three passages with a total length of 2,000-2,750 words. The question
formats include multiple choice, short-answer questions, cloze summary,
sentence completion, table/chart completion, matching lists or phrases,
supplying headings for paragraphs, and so on.
The writing component, also lasting for 60 minutes, is made up of two
academic-oriented tasks. The first task asks the candidate to describe in
about 150 words a chart, a graphic or a table which they might encounter
during their study at university. The second task usually requires the can-
didate to write an argumentative essay of about 250 words based on a
controversial topic supplied in the question paper.
The speaking subtest, which lasts 11-14 minutes, is conducted in the
form of a one-to-one interview with an examiner. The interview is re-
corded on an audiotape for fiature quality assurance (see below). The sub-
test comprises three parts. In the first part, the examiner interviews the
candidate with questions of familiar nature which are selected from a
pool; In the second part, the candidate makes a short presentation on a
topic supplied at the beginning ofthe session; and the third part is in the
form of a discussion between the examiner and the candidate, based on
questions linked to the main theme of Part 2 (British Council, IDP: IELTS
Australia and University of Cambridge ESOL Examinations 2005).
According to the most recent information available, the Cronbach's
Alpha reliabilities for IELTS listening and reading subtests in 2003 were
0.89 and 0.88 respectively. IELTS candidates' writing and speaking per-
formances are rated by a single certified rater at local test centres for the
purpose of score reporting. Reliabilities of writing and speaking assess-
ment is ensured 'through a sample monitoring process', where a sub-
23
Assessing University Students
Methodology
The data were collected from a voluntary sample of 243 final-year stu-
dents from 17 academic departments at HKPU, who sat for both the
GSLPA and the Academic Modules ofthe IELTS within a month. The
IELTS was administered to them directly by IDP Australia and the British
Council. With consent from the participating students, the test results were
provided to the research team as the data for the present study.
To achieve the two purposes ofthe study, statistical analyses were per-
formed using SPSS Version 14. Three procedures were applied to the data:
computation of Descriptive Statistics mainly for comparing the distri-
butions of obtained scores on the corresponding writing and speaking
components ofthe IELTS and GSLPA, in order to examine the discrimi-
nating power of each test, and calculation of Pearson Product Moment
Correlations and Shared Variances in order to determine whether or not
these components ofthe two different tests measure the same areas of
language knowledge and skills.
6.0 67 27.6
6.5 71 29.2
7.0 43 17.7
7.5 16 6.6
8.0 2 0.8
8.5 0 0
9.0 0 0
Total 243 100
50.0
1 I
% 30,0
— | - 11
1 ^^ I
I
1M-
1i) 2.0 2.5 3.0 3.5 4.0 4.6 6.0 5.S 8JQ
0.0 4-
8,0 9.0
It can be observed frotn these tables and figures that while there are
nine bands available for the IELTS to describe candidates' overall, and
writing and speaking proficiency, only about half of these bands were
actually used in describing the writing and speaking proficiency levels of
this group of candidates. The scores on the IELTS writing are distributed
over five bands (4-8). The speaking scores span the same range.
On the other hand, the GSLPA scores are spread over a wider range.
For example., the scores on the GSLPA writing are distributed over seven
bands (2.5-5.5), including four half bands, which were reported in the
score certificates. The GSLPA speaking scores are distributed over 10
28
Regional Language Centre Journal 38.1
bands (1.5-6), which again include five half bands as they were reported
in the certificates.
It can also be observed from Tables 2-5 that GSLPA writing and
speaking scores are more evenly distributed than their IELTS counterparts.
Most notably, as shown in Tables 3 and 5, on the IELTS writing subtest,
125 (51.4%) out ofthe 243 candidates are bunched up on Band 6, and 116
candidates (47.9%) cluster on Band 6 for the IELTS speaking subtest. We
can note that, as shown in Table 6, the distribution ofthe IELTS overall
scores looks much better than the distributions of its writing and speaking
scores: This is because the IELTS listening and reading scores also con-
tributed equally to the production ofthe IELTS overall scores. With a
potential range of 17 bands/half bands, and spreading over 11 bands (4—9)
and 10 bands (4.5-9) respectively in this dataset, the listening and reading
scores are much more widely distributed.
Intercorrelations
Table 7 reports the intercorrelations between all components ofthe two
tests. As indicated in the table, the overall scores ofthe two tests are
highly correlated. The correlation between the scores on the GSLPA and
IELTS speaking components is also fairly strong. However, there is only
a weak, albeit statistically significant, correlation between the writing
components ofthe two tests. It is also interesting to note that the GSLPA
speaking component is relatively strongly associated with the IELTS
listening component. That is probably because some ofthe GSLPA speak-
ing tasks, such as Task 1, also require listening comprehension.
Based on these correlation coefficients, the following shared variances
(R^) were obtained (see Table 8). The magnitude of an R' indicates the
proportion of overlap between two datasets, in this case two tests the cor-
relation (r) of which was used to produce the R". Because all the candi-
dates took both tests it is safe to assume that the variance is not caused by
differences in test populations, but by some differences in the tests. In the
light of these shared variances, the following observations can be made.
The low value of 0.21 indicates that there is only 21% overlap ofthe
variance between the scores on the writing components ofthe GSLPA and
IELTS, which suggests that these two writing subtests basically measure
different skills.
The R^ value of 0.48 indicates overlap of almost half of the variance
between the scores on the speaking components ofthe two tests. Never-
theless, more than half of the construct (52%) of each test is distinctly
different from the other.
29
Assessing University Students
The 53% of overlap ofthe variance of overall scores on the two tests,
based on the R" value of 0.53 in Table 8, suggests that the GSLPA and
IELTS share about 53% ofthe test construct but there is still 47% ofthe
variance in each test that cannot be explained by the variance in the other
test. In other words, although the two tests seem to be measuring similar
areas of knowledge about half the time, almost half the time they seem to
be measuring different constructs or aspects of constructs.
p<.01 (2-tailed)
by design the GSLPA writing and speaking subtests, being the only com-
ponents ofthe test, measure candidates' perfonnance in these areas in a
more intensive way than the corresponding components ofthe IELTS, as
indicated by the different lengths ofthe test time allowed and the number
of tasks assigned to each subtest. it is therefore critical that IELTS make
full use of its receptive components (reading and listening) in compen-
sating for what its productive components (writing and speaking) are
unable to achieve, that is to say. it is desirable that the lELTS should better
utilize the available proficiency bands in reporting test scores. This com-
pensation appears to have been made through a careful grading of reading
and listening performances. However, the IELTS has the potential of
being more discriminatory among mid-level candidates if its mid-range
score bands ofthe writing and speaking subtests can be further refined and
applied more meticulously by utilizing half bands.
Test Components. The current exit test used by UGC, IELTS, appears to
be more comprehensive than the GSLPA as it contains components cov-
ering all four commonly recognized language skills, listening, speaking,
reading and writing. This comprehensive fonnal may make the test look
more credible to the general public, especially to some potential test users.
In contrast, the GSLPA contains only two main components, writing and
speaking, which, to some test users, may sound insufficient in temis of
skill coverage, although this format was drawn up based on a large-scale
survey of an important test user group, namely, senior executive officers
of business companies, who pointed out that professional writing and
speaking were the most essential communicative skills in workplaces.
REFERENCES
Writing Bands
W6: Can produce clear, convincing, well argued texts, using suitable tone
and style. Vocabulary is precise and effective, and grammar is well
controlled. A highly proficient writer.
* W5+
W5: Cati produce well organized texts that communicate successfully and
clearly on required tasks. Has generally good control of tone, styie,
vocabulary and sentence structure, despite some inaccuracies. A
clearly competent writer.
W4+ - •
W4: Can produce relevant, interpretable and generally well organized texts
that address task requirements. Vocabulary is generally adequate, and
grammatical errors do not obscure communication. A generally
competent writer.
W3-t- • • • • •
W3: Can produce generally relevant and interpretable texts that show basic
ability to organise content appropriately. Vocabulary is adequate to
convey basic meanings and grammatical errors rarely prevent
communication. A basic writer.
W2+
W2: Can produce texts with some relevance to the task and some sense of
organisation. Vocabulary and grammar are inconsistent, but allow
meaning to be conveyed. A limited writer.
WI +
WI: Has some grasp of basic forms of written English, although these may
not be applied relevantly, appropriately or consistently. An elementary
writer.
* Grades which include + indicate that the student's performance falls between
two levels (e.g. 3+ means the student's performance is between a 3 and a 4).
Speaking Bands
S6: Can speak clearly, precisely and confidently on a range of tasks, using
complex language when necessary. Speech is effortless for listener to
follow.
S5+
S5: Can communicate successfully, clearly and with confidence on a range
of speaking tasks, generally using precise language. Sense is easy to fol-
low throughout.
36
Regional Language Centre Journal 38.1
S4+
S4: Can convey meaning successfully on a range of speaking tasks,
despite inaccuracies or limitations in vocabulary. Although
organization sometimes lacks clarity or fluency, message can always
be followed.
S3: ' Can convey meaning on a limited range of speaking tasks, but is
sometimes hesitant. Despite inaccuracies, unevenness in pronun-
ciation and/or limitations in vocabulary, message is mostly
comprehensible.
S2: Can convey basic meaning in some situations, although there are
frequent pauses and hesitations, and ideas may lack organization.
Speech can be hard to follow.
S1+
SI: Has some grasp of basic forms of spoken English, although speech is
disjointed and very hard to follow.
(Adapted from The Hong Kong Polytechnic University 2007)
9—Expert user
Has ftilly operational command ofthe language: appropriate, acctirate and fluent
with complete understanding.
8—Very good user
Has fully operational command ofthe language with only occasional unsystematic
inaccuracies and inappropriacies. Misunderstandings may occur in unfamiliar
situations. Handles complex detailed argumentation well.
7-—Good user
Has operational command ofthe language, though with occasional inaccuracies,
inappropriacies and misunderstandings in some situations. Generally handles com-
plex language well and understands detailed reasoning.
6—Competent user
Has generally effective command of the language despite some inaccuracies,
inappropriacies and misunderstandings. Can use and understand fairly complex
language, particularly in familiar situations.
5—Modest user
Has partial command ofthe language, coping with overall meaning in most situa-
tions, though is likely to make many mistakes. Should be able to handle basic
communication in own field.
4—Limited user
Basic competence is limited to familiar situations. Has frequent problems in under-
standing and expression. Is not able to use complex language.
37
Assessing University Students