Beruflich Dokumente
Kultur Dokumente
Construct definition
The construct being assessed with this test is concerned with both receptive and
productive uses of the target grammatical structures. Examinees will be able to choose forms of
stative verbs, parallel structure, and use of articles. Examinees will also be able to choose the
meaning of comparatives with adverbs, present perfect progressive, reduced adjective clauses,
and used to/would with past habits. Choosing both the forms and meanings of these grammatical
structures demonstrates students grammatical knowledge in these areas.
In contrast, students must demonstrate grammatical ability in several tasks and content
areas. Examinees will be able to create sentences and evaluate the grammaticality of items in a
paragraph, then revise them with correct comparative adverbs, adverb clauses, and conditionals.
Examinees ability to speak and to listen will not be tested. Ability to correctly use and identify
simple present/past, progressive present/past and the present perfect is assumed.
Description of Tables of Specifications
A common method of test planning and construction is to create a Table of Specifications
(TOS), as shown in Appendix A, which helps organize topics covered and general instructional
objectives (Miller, Linn, & Gronlund, 2009). This method was used to organize content and
objectives in the development of this test. The content areas covered during the weeks since the
last achievement test are listed on the far left hand side of the table, and the instructional
objectives are listed in the top two rows. Each cell in the center of the table has the number of
items total, followed directly by the numbers of the actual items in the test. We met with the
curriculum supervisor for grammar classes at INTO CSU and discussed the content areas that
were being taught, and how these needed to be assessed. At INTO CSU, the terms grammatical
knowledge and grammatical ability are used. Here, teachers define grammatical knowledge as
the assessment of the capacity to recognize meaning and form of grammatical structures.
Grammatical ability is the capacity to produce limited or extended responses.
As seen in the Table of Specifications, the biggest focus for grammar instruction in the
week 10 test was on adverb clauses for cause/effect and time (43%), followed by comparatives
(21%), and then followed by conditionals (16%). Less time (3% each) was spent on several other
structures: present perfect progressive, reduced adjective clauses, parallel structure, stative verbs,
used to/would for past activities, and articles (a/an, the and ). The latter three structures were
items originally covered in their previous class, but were reviewed briefly in the past few weeks.
It should be noted that item number 2D.10 could be correctly answered with either a conditional
or an adverb clause of time. Because we feel that students will be more likely to answer it as a
conditional, it is listed under this content area in the TOS.
Description of Test Tasks
For our test, there are six different types of test tasks. We ordered our tasks from easiest
to most difficult and verified this order of task with the curriculum supervisor. Therefore, each
task will take more time for students to complete than the prior one. We have not included
examples for students of how to perform the tasks for every part because they will work through
a practice test one day before the actual test. The practice test will enable them to be familiar
with the types of tasks and reduce the amount of time students need to understand the task. We
also ordered our tasks starting with grammatical knowledge and then grammatical ability. An
example of grammatical knowledge would be the ability to recognize the correct form; students
could be given the following sentence: I do not enjoy running, swimming, nor (played, playing,
plays). When students choose the correct form of the parallel structure, this demonstrates
grammatical knowledge. In this assessment, there are four main types of limited or extended
response items which are used to assess students grammatical ability: constructed response,
sentence combination, paragraph editing, and paragraph writing, which will be further described
below.
The first task, Multiple Choice, which tests grammatical knowledge, consists of
multiple choice questions. Instructions for the task state: Circle the letter of the best answer; for
directions on all tasks, we attempted to avoid using technical vocabulary and to be written
between a 6th-8th grade reading level (Becker, 2016). The use of these multiple choice items
allow us to measure a variety of learning outcomes (Miller et al., 2009, p. 196). Because there
is only one single criterion for correctness for these questions, the students receive one point
for the correct response (14 points total) (Bachman & Palmer, 2010). The expected response
would then be a circle around the correct answer of the letter of the question; these questions test
students knowledge of parallel structure, present perfect progressive, articles and other
grammatical knowledge as previously mentioned. Due to the similarity between the expected
response and the input, there is a direct relationship with a narrow scope (Bachman & Palmer,
2010). Because of the strategies of ordering tasks, we expect the students to take the least amount
of time per question in this section. We predict that low-leveled students take a maximum of 1315 minutes, while higher-leveled students will take 9-10 minutes. A multiple choice item for
non-native speakers may take as long as 1-2 minutes to answer; however, that may vary
according to the length of the stem and options (T. Becker, personal correspondence, March 29,
2016). Most of the items in this part are short, so we have reduced the allotted time accordingly.
Our second test task, Completion, asks students to Fill in the blank with the correct
from the verbs provided in parentheses. The expected response of this limited response task is to
have students provide the correct verb tense/aspect based on the rules of the conditional forms
(types 0-2). There is a direct relationship between the input and expected response, while there is
a narrow scope of the relationship (Bachman & Palmer, 2010). Because students only fill in one
blank, we expect that for low-level students the task would take a maximum of 10-12 minutes,
while higher-leveled students would take a maximum of 7-10 minutes. Additionally, each blank
is worth one point for the correct verb tense/aspect (7 points total). Having students provide an
answer, rather than just selecting one, reduces the chances that students will guess the correct
answer (Miller et al., 2009, p. 175).
For the third task, Sentence Combining. there are two sections within this task. The first
asks students to, Using the given words/phrases in the parentheses, combine ideas from a and
b to create a new sentence. Keep the meaning the same. Under a and b two sentences are
given in each. The a set of sentences gives a reason, while b sentences show the result of
these reasons. The expected response for this task is a complete sentence that uses the adverb,
preposition or transition word accurately in both form and meaning. Due to this, there is a direct
relationship between the input and expected response and a narrow scope of the relationship
(Bachman & Palmer, 2010). Since students are creating new sentences, we anticipate that for
low-level students the task would take a maximum of 10-12 minutes, while higher-leveled
students would take a maximum of 7-10 minutes. Each sentence constructed is worth two points,
one point for correct form and one point for correct meaning (10 points total). The use of this
partial-credit scoring allows teachers to capture more precise information from test-taker
responses (Purpura, 2004, p. 117).
The second part of the task has different directions than the first part; the directions state
Create new sentence(s) by combining the sentences using the word given in the parentheses.
Keep the meaning the same. Students are expected to write a new sentence with an adverb
would take a maximum of 14-16 minutes, while higher-level students would take a maximum of
10-12 minutes.
The final task for our test is Short Answer, where students are to Write a paragraph (34 sentences) to answer the following question. This tasks asks students to answer What causes
students to think school is boring? What effects does this feeling (boring) have on students in
school? The expected response from this question is to show multiple ways of reason/result (or
cause/effect); the various ways that these can be shown are adverb clauses, prepositional phrases
and transition words, which are all classified as current grammar. Because students are
expected to create an extended response, there is a narrow scope with an indirect relationship
(Bachman & Palmer, 2010). For this task, an analytic rating scale is the most appropriate option
(Purpura, 2004). Therefore, students responses are worth five points total: 1 point for a complete
answer, 2 points for previous grammar (i.e. punctuation, subject/verb agreement, etc.), and 2
points for current grammar, as described above. Even though this task is worth five points, we
predict that students will take the most time completing this task. We foresee low-level students
to take 18-20 minutes to complete, while high-level students would take 16-18 minutes.
Development of Criteria
In order to write the items for the test, we focused on the TLU domain, the construct, and
the course objectives as laid out by the Table of Specifications (Appendix A). We created the
items for the different parts of the test while maintaining a progression from easier to more
difficult items. After having written the items, we got feedback from experienced grammar
instructors at INTO CSU, the curriculum supervisor, and our professor. For example, initially on
the paragraph editing task (Part 2D), some grammar mistakes occupied two different lines of the
text. In other words, journalism isnt as difficult engineering was separated in the middle so that
it would be much more difficult for students to see the mistake. After receiving input from these
teachers about this potential problem, we made the necessary changes.
After all of the items were written and had been edited, we developed rubrics for items
2C through 2E. All of these items were worth at least 2 points. In most cases, the rubric was
simply one point for meaning, one point for form. However, Part 2E needed a special rubric as
can be seen on the test (Appendix C). Finally, after administering the test, some elements needed
to have other answers added to the key that would also be acceptable, which is further elaborated
on in the Discussion section.
Pilot Test Procedure
This grammar test was administered during a regular class period, meaning that
participants had 90 minutes to complete it. The 15 participants in this study were considered by
INTO CSU to be of mid-intermediate English proficiency. The students L1s are Arabic (10),
Chinese (4), and Thai (1). Of the 10 Arabic participants, 7 were female and 3 were male, while
there was 1 female Chinese speaker, and the rest of the participants were male.These students
were in the second level of the program; thus, they had been in the program prior to this
semester. Because of this experience in the program, these students were familiar with the testing
methods that were implemented. This was the second major achievement test in this course, but
the types of tasks that were included on this test are tasks that were also included on the last test
in this class. The grammatical structure content was different from the first achievement
assessment. The paragraph editing was the only new type of task. As a review before the test, a
practice test was given in exactly the same format as the actual test, and the grammar content
was similar. The instructions of each part were read and the test takers were able to ask
questions. The test takers completed most parts of the practice test, and answers were discussed.
10
In order to enable students to keep track of time, a digital clock was displayed on the
board where everyone could see it. Before distributing the test, students were asked to do the
following: distance themselves from other students, put their phones on silent and in their bags,
go to the restroom if need be, and if they had a question to raise their hand. Students were also
told that once they left the room, then they must turn in their test and not come back in the
classroom; only in the case of emergencies would students be allowed to leave and come back.
Before the test was administered, the test key was created (Appendix D). Within this key,
each section was assigned certain points as laid out by particular criteria. These criteria allowed
us to score the test accurately and fairly. After the students had taken the test, they were graded
by one instructor due to practicality reasons and because there was a deadline to have the grades
into the administration. Even though only one person was grading the tests, when questions arose
the grammar curriculum supervisor was able to provide valuable input.
Through this input, we were able to have conversations about adding more potential
answers to the key. For two questions we expanded the possible correct answers, like under Part
2A #4, which states:
If people live in the city, they (have) _______ more possibilities for shopping.
And the initial answer that was included on the key was:
If people live in the city, they (have) _have_ more possibilities for shopping.
When reviewing students answers, this came up frequently:
If people live in the city, they (have) _will have_ more possibilities for shopping.
Since it is possible to use the future in the independent clause in the first conditional, the
curriculum supervisor and I decided to add it to the key. We also saw a similar trend with 2C #3,
and add the same as as a potential answer.
11
Test Results
After the administration of the test, the instructor graded the tests. Following this, the
tests were scanned, copied and the corrections made by the instructor were whited out. Then, the
other researcher graded all the students tests individually as well. The names of the students had
also been replaced by numbers so that there was no chance for bias. The raters then met up to
compare the correction results. The items on each test were compared and the raters discovered 4
main areas of difference. One mistake in correcting was simply overlooked. Five mistakes of the
second rater were due to a lack of knowledge of level appropriate grammar. Lastly, minor
scoring differences and misinterpretations of the rubrics caused 6 mistakes in grading. These
items were discussed and it was discovered that only one of the mistakes out of 62 points on 15
tests was made by the original instructor (i.e. Rater 1).
After these items had been discussed we analyzed which students were masters or nonmasters. As the cut score was 74.5%, all 15 students in the class passed the test, so in general,
they mastered the material from the construct established.
Table 1
Summary of Rater 1 Scores for Pilot Study
Scoring Criteria
Min
Max
Mean
SD
15
11
14
13.01
1.03
Ability: Conditionals
15
5.73
0.88
15
15
18
17.01
1.16
Ability: Comparisons
15
7.33
0.98
15
10
8.23
1.03
15
4.53
0.39
Rater 1
12
Table 2
Summary of Rater 2 Scores for Pilot Study
Scoring Criteria
Min
Max
Mean
SD
15
10
14
12.73
1.28
Ability: Conditionals
15
5.73
0.88
15
16
18
17.4
0.63
Ability: Comparisons
15
7.2
1.21
15
10
8.4
0.91
15
3.5
4.5
0.5
Rater 2
As shown by Table 1 and Table 2, both raters, at first glance, appeared to grade similarly
throughout all six parts to the test. This was confirmed for some parts of the test by running
interrater reliability using Pearsons r. Displayed in Table 3 below, there was a very high
correlation (1) between raters on Part 2A Ability: Conditionals. Despite this, the correlation for
Part 2C was rather low with 0.36; our hypothesis of why this correlation was so weak is included
in the Discussion.
After calculating the variance on the mean scores for all of the parts, the Cronbachs
Alpha was calculated for all six tasks and was 0.50. This means that there was a medium-low
correlation that all six tasks were measuring the same grammatical ability. Then the Standard
Error of Measurement was 2.01. Therefore, a students score should be (+/-) 2 points of the final
score calculated.
13
Table 3
Interrater Reliability by parts
Test
Section
Part 1
0.88
Part 2A
1.0
Part 2B
0.54
Part 2C
0.36
Part 2D
0.71
Part 2E
0.45
Item Analysis
With the multiple choice items in Part 1, item analysis was performed to see how difficult
each item was and how much each item discriminated between the test takers (Miller, Linn &
Gronlund, 2009). Taken globally, the average difficulty of the items was 93%, and the
discrimination was 14%. These calculations were performed by separating students into 2
groups: those who had done well, and those who, in comparison, had done poorly. Then we
looked at each item to see how these two groups performed. The distractors were analyzed to see
how effective they were. Six of the items had distractors which were chosen at least once, and
two of the items had distractors where both of the distractors were chosen. Eight of the items
were correct for all of the students, so they had a difficulty level of 1.00 and a discriminatory
power of zero. The other 6 items had a difficulty level ranging from .75 to .92 and a
discriminatory power ranging from 0.17 to 0.5.
14
Discussion
Test revisions made
Since this was the first time this test was piloted, there are several revisions that have
been made. In Part 1, some of the items had individual directions in addition to the general
directions at the beginning. We made changes to keep only the general directions. Furthermore,
the answer to Part 1.9 was too easy and has been changed. In Part 2D, the paragraph editing
activity, one of the items to be corrected was missed by most students in the class. After
discussing this with the curriculum supervisor, this item was not counted in students final
scores. Although Part 2D is an authentic activity from the TLU domain, it needs more
scaffolding for students at the intermediate level, so it has been changed to include a list of
possible mistakes that students could look for based on the grammar that has been covered
during that part of the semester. A sample of how corrections could be made has also been
included.
Suggested changes to answer key
After correcting the tests, some issues were raised about the scoring methods used on the
rubrics in the answer key, and there are several suggested changes to the answer key. In Part 2A,
two answers that were also correct have been added to the answer key as mentioned in the test
piloting section. Furthermore, in Part 2D, we added acceptable answers to the key for item
numbers 5 and 7. For example, for item number 5, we accepted ,so because it also shows result.
If this test were to be used again in the future, several changes need to be made to the
answer key for the short and extended response test items. For example, in Part 2E, it would be
helpful to list the main items that belong to both previous grammar and current grammar. We
would also add a comment that no points should be deducted for spelling except where the
15
mistakes occurred in the grammar, like verb endings. Another possibility for this item is to
design a holistic rubric with general guidelines for the main elements in the task.
Critique of item performance
As mentioned in the results section, a simplified item analysis (Miller, Linn & Gronlund,
2009) was done for the multiple choice section. This kind of item analysis enables the inspection
of each multiple choice item to look at item difficulty and discriminating power of each item.
Since this is not a norm-referenced test, our goal is not to differentiate among students, so the
fact that over half of the items in this part were answered correctly does not mean they are not
effective items (Miller, Linn & Gronlund, 2009). As mentioned earlier, this test follows good
assessment practices of having the easier items towards the beginning of the test, so it is to be
expected that students would do well on this first section.
Reliability
In order to evaluate the reliability of our test, two raters scored the two tests, based upon
the answer key. After computing interrater reliability, it seems as though we were fairly
inconsistent. As previously mentioned, we had a very high correlation on Part 2A, and Part 1 and
2D were also fairly high. Even though half of the parts had a high correlation, the other three had
a medium-low to a very low correlation, specifically Part 2C, with a 0.36. Our reasoning as to
why there was a low correlation to that part was because one of the raters was not sufficiently
trained enough in the grammatical structures acceptable. The same goes for Part 2E, which was
the second lowest correlation between raters. If both raters had been trained to the same
standards, then reliability could have been higher.
16
17
a fairly good indicator that we can interpret the week 10 achievement test for students overall
grammatical ability, as represented by the objectives and outcomes in the syllabus.
Consistency among raters is also a criteria for validation of the use of scores. Since some
parts of the test did not have high inter-rater reliability, this would be an argument against
validity. At the same time, it should be noted that the first rater was in constant communication
with the curriculum supervisor who confirmed the actual scores that the students received.
Therefore the second raters scores did not directly affect the students final scores on the test.
Self-report methods can also be used to show the correct use of scores. When students
were asked about the test, they commented that it was a fair assessment of their abilities in the
areas of instruction based on the standards in the syllabus.
Authenticity
The authenticity of the assessment is measured by the degree of correspondence between
the test tasks and the TLU domain. Part 2D was designed and analyzed through the use of a TLU
domain table (Appendix B), and was judged to be fairly authentic by the instructor. It was only
fairly authentic because the kinds of mistakes were of limited scope. On the original test, there
were no instructions about what kind of mistakes test takers should look for. However, this may
have been too challenging for the students, and we decided to list the kinds of mistakes test
takers should look for as can be seen on the revised test (Appendix C).
Impact and Practically
Since parts of this test were also used by other instructors, their input about the relevance
and similarity to previous tasks were obtained. Two out of the four instructors who taught similar
sections of the course felt as though this test assessed students fairly due to the fact that students
had practiced with similar types of tasks throughout the semester. Their students also felt as
18
though that this test had been fair because the tasks and knowledge being tested had been
thoroughly covered prior. One of these instructors did make revisions, as to be tailored to their
particular class, which is why those results are not included.
The other two instructors did not necessarily feel the same way, and made major
revisions prior to the administration of the test. They did not give more details about this, but it
can be assumed that they did not use similar types of tasks during their course; because this test
was not standardized, this was acceptable by INTO. In order to determine the impact on students,
and the applicability to their future studies, an administrator would need to evaluate the different
tasks utilized.
In terms of practicality, this test was administered during normal class time and did not
require extraneous resources nor training to administer. That being said, some training would
have been necessary in order to properly score the test, especially with Part 2E, even though
there was a guiding rubric. Additionally, it did take some time to score all of the tests, and some
discussion about certain items were discussed with the grammar curriculum supervisor, in order
to properly record students grades. The more authentic parts 2D and 2E took the longest to
correct, but this is a tradeoff we were willing to make in order to have a high degree of
correspondence with real-life tasks.
Overall estimate of test purpose
Altogether, it seemed as though the test achieved its purpose in assessing students in their
knowledge of their study on material from weeks 6-10, as stated by the syllabus. There was
evidence for construct validity, but in order to gather more data about the reliability and
effectiveness of this test, it would have been beneficial to have obtained results from all four
sections of intermediate grammar.
19
Reflection
One of the major weaknesses seen in this project is not directly an aspect of the test, but
of the way it was rated. Since I, Brian, had not taught the class nor was I familiar with the topics
being assessed, there was a need for more training before I corrected the test. A way this could
have been resolved is by correcting 3-4 tests from another teacher who gave the test and then
discussing the corrections in detail. Furthermore, we could have put more detail into the rubrics
in parts 2B, 2C, and 2E so that some of these errors would have been avoided. For example, in
Part 2E, clearer instruction could have been given concerning the method of awarding or taking
away points. This is an aspect of the test that could be developed if it is given again in the future,
as INTO typically reuses tests if the same constructs are being tested.
The use of the TLU domain table was very helpful in analyzing every aspect of the
editing paragraph. However, in hindsight, Part 2E was also a very authentic task. Accordingly,
we would need to create a TLU domain table for Part 2E to really analyze it in detail for
authenticity; notwithstanding, it does seem very close to what students are asked to do at the
university level. We had discussed putting more detail into the directions so that future students
would have to respond with the grammar that we were trying to elicit. When students are given
an assignment in one of their future classes, they are free to use the grammar structures that are
most appropriate for the response. The actual tests answers to that part showed that test takers
answered using cause/effect and conditional grammar, which comprised a major part of the
grammatical structures which were taught.
From my perspective, I, Kathleen, felt very similarly to Brian that the training for him
couldve been more substantial so that mistakes couldve been avoided. One way that I couldve
have better prepared him for those mistakes wouldve been to update the answer key as needed.
20
Instead, I wrote down errors (not correct answers) that I needed to discuss with the curriculum
supervisor, which ended up confusing him, and understandably so. Nonetheless, so much of my
time was consumed with copying, and removing indicators from the marks that I had made on
students tests, that I hadnt thought about the notes I had made that couldve been interpreted
wrong.
Another point that I wish couldve gone better was the collaboration between other
grammar teachers. Even though we had been receiving comments from them throughout the
process of creating this test, one instructor had decided to not share all suggestions they had.
Instead, they informed us the day of the exam, specifically an hour before it was to be
administered, that they had made enough changes to the test. If we were informed about this
earlier, we could have had a discussion about these potential changes, so that we could have had
more data to provide for the reliability and effectiveness of our test. This was a learning
experience for me that both input from others can be highly beneficial, and that others might also
not want to truly be a part of that collaborative process, which at times can be both
understandable and frustrating, depending on your perspective in the matter.
21
References
Bachman, L., and Palmer, A. (2010). Language assessment in practice. New York, NY: Oxford
University Press.
Becker, T. (2016). Developing classroom assessment [Powerpoint slides]. Retrieved from the
class Canvas page.
Cummin, A (2013). In C. A. Chapelle (Ed.), The Encyclopedia of Applied Linguistics (pp. 1-10).
New York, NY: Blackwell Publishing Ltd.
Miller, M. D., Linn, R. L., and Gronlund, N. E. (2009). Measurement and assessment in
teaching. Upper Saddle River, New Jersey: Pearson Education.
OMalley, J.M., & Pierce, L.V. (1996) Authentic assessment for English language learners:
Practical approaches for teachers. Boston, MA: Addison-Wesley Publishing Company.
Purpura, J. E. (2004). Assessing grammar. New York, NY: Cambridge University Press.
Purpura, J. E. (2013). Assessing grammar. In C. A. Chapelle (Ed.), The Encyclopedia of Applied
Linguistics (pp. 1-10). New York, NY: Blackwell Publishing Ltd. DOI:
10.1002/9781405198431.wbeal004
22
Appendix A
Table of Specifications
Content Areas
Wk 6: comparatives with
adverbs, ch. 13
# of
items/#
of points
% of
points
(2)
1.1 + 1.14
9/13
21 %
(2)
1.12 + 1.13
2/2
3%
(7)
2B.1-5 + 2B.2.7 +
2B.2.9
12/23
37 %
(2)
2B.2.6 + 2B.2.8
2/4
6%
10/10
16 %
2/2
3%
Para. editing
Sentence combining
(3)
2D.1 + 2D 2 + 2D.9
(4)
2C 1-4
Completion
Grammatical knowledge. Ss
recognize meaning and form.
# of items
M/C choose
M/C choose
form
Meaning
(1)
2E.1
(4)
2D. 3-5 + 2D.8
(3)
2D.6,7 + 2D.10
(7)
2A 1-7
(2)
1.9 + 1.10
(2)
1.3 + 1. 6
2/2
3%
(2)
1.2 + 1.7
2/2
3%
2/2
3%
2/2
3%
(2)
1.8 + 1.11
1/5
10/10
13/26
7/7
(2)
1.4 + 1.5
6/6
8%
16 %
41 %
11 %
10 %
8/8
45/62
13 %
100%
100 %
Appendix B
Target Language Use task
23
At the university, students will numerous papers to hand in. They will need to edit their own writing for grammatical errors at the
discourse level of these assignments.
TLU Task Characteristics
Characteristics of the setting
physical characteristics
library or study setting
participants
student alone
time of task
morning, afternoon, evening
Characteristics of the scoring rubric
instructions
language
channel
specification of procedures and tasks
structure
time allotment
scoring method
criteria for correctness
procedures for scoring the response
explicitness of criteria and procedures
Characteristics of the input
format
channel
visual
form
language
language
target language
length
from 1 to 30 pages
type
input for interpretation
degree of speededness
less speeded than native speaker
vehicle
live
language of the input
language characteristics
organizational characteristics
grammatical
syntax, vocabulary, graphology
textual
academic or technical
pragmatic characteristics
functional
ideational, heuristic
sociolinguistic
genre, register, natural/idiomatic,
topical characteristics
personal
Characteristics of the expected response
format
channel
visual
form
language
language
target
length
1-30 pages
type
extended production
degree of speededness
not speeded as students have time to revise
language of the expected response
language characteristics
organizational characteristics
grammatical
syntax, vocabulary, phonology, graphology
textual
rhetorical and cohesion
pragmatic characteristics
functional
ideational, heuristic
sociolinguistic
genre, register, natural/idiomatic, cultural/figures of
speech
topical characteristics
technical/academic
Relationship between input and response
reactivity
non-reciprocal
scope of relationship
broad scope
directness of the relationship
indirect
24
25
Appendix C
AEIN8213: Week 10 Achievement Test
The purpose of this test is to assess your knowledge and ability to use grammatical
structures from level 2 texts. You have 90 minutes to complete this test and all answers
should be written on this test. Each section has directions, please read them carefully.
Unless otherwise noted, each question on the test is worth 1 point. There are a total of 62
points.
CSU Honor Pledge- Write this statement on the lines below and initial.
I will not give, receive, or use any unauthorized assistance.
_________________________________________________________________Initials:______
Part 1: Knowledge
1) Multiple Choice (14 points)
Directions: Read each item and circle the letter that best answers the question or completes the
sentence.
1. John isnt as happy as Mary is.
Who is happier?
a.
b.
c.
Mary
John
Neither
played
playing
plays
is hating
hates
is hated
26
an / The / the
an / The / an
the / An / the
a / The /
The / / a
The / An / the
A / An /
/ A / the
want
am wanted
am wanting
Yangping had her pens, got her notebooks, and her bookbag.
Yangping had her pens, her notebooks, and her bookbag.
Yangping wanted her pens, notebooks, and remembered her bookbag.
Yangping had one pen, notebooks, and many bookbags.
Past activity
Finished action before now
Recent activity
27
My mother
My father
Neither
We dont know
Part 2: Ability
2A) Completion (7 points)
Directions: Fill in the blank with the correct form of the verbs provided in parentheses.
1.
2.
3.
4.
If people live in the city, they (have) ___________________ more possibilities for
shopping.
5.
If I (find) ________________$1 million on the street, I would not give it to the police.
6.
If you (not, study) ______________ conditionals, you will have problems with grammar.
7.
If I were you, I (not, be) _________________ late to class for the exam next Friday.
28
29
Example: Due to him being a bad cook, I dont enjoy the meals he makes.
1. (although)__________________________________________________________________
2. (because)___________________________________________________________________
3. (thus)_______________________________________________________________________
4. (because of)__________________________________________________________________
5. (so) _______________________________________________________________________
2B.2) Adverb Clauses (8 points)
Directions: Create new sentence(s) by combining the sentences using the word given in the
parentheses. Keep the meaning the same.
6. I bought food at the grocery store.
Then, my dad cooked me dinner.
(after)_________________________________________________________________________
7.
30
4.
31
32
2E) Short Answer (5 points; 1pts completion, 2pts previous grammar, 2pts current grammar)
Directions: Write a paragraph (3-4 sentences) to answer the following questions.
What causes students to think school is boring? What effects does this feeling (boring) have on
students in school?
______________________________________________________________________________
______________________________________________________________________________
______________________________________________________________________________
______________________________________________________________________________
______________________________________________________________________________
______________________________________________________________________________
______________________________________________________________________________
______________________________________________________________________________
______________________________________________________________________________
______________________________________________________________________________
______________________________________________________________________________
______________________________________________________________________________
______________________________________________________________________________
______________________________________________________________________________
______________________________________________________________________________
______________________________________________________________________________
______________________________________________________________________________
______________________________________________________________________________
Appendix D
Achievement Test Answer Key
Name _____________________________________
33
TOTAL: ______/62pts
Mary
b. John
c.
Neither
2. Circle the correct form to complete the parallel structure in the sentence.
I do not enjoy running, swimming nor ____________.
a.
played
b. playing
c.
plays
is hating
b. hates
c.
is hated
34
an / The / the
b.
an / The / an
c.
the / An / the
d.
a / The /
The / An / the
j.
A / An /
k. / A / the
6. Circle the correct form of the verb.
I __________________your wifes wonderful spaghetti sauce.
a. want
b. am wanted
c.
am wanting
Yangping had her pens, got her notebooks, and her bookbag.
b.
c.
Mary smoked in the past, but she does not smoke now.
b.
Past habit
c.
Conditional
Past activity
35
36
My mother
b.
My father
c.
Neither
d.
We dont know
Part 2: Ability
2A) Completion (7 points)
Directions: Fill in the blank with the correct form of the verbs provided in parentheses.
Contractions are also possible where applicable. Other acceptable answers are in parenthesis
after the item.
1. If tomorrow (be) _is____ a nice day, we might go swimming.
2. People (travel) _would travel___ more in the U.S. if it were less expensive.
3. Fatimah will be sad if her friends (not write) __do not write__ to her. (cant write)
4. People who live in the city, (have) _______have_______ more possibilities for shopping.
(will have or might have)
5. If I (find) _____found__$1 million on the street, I would not give it to the police. (could
find).
6. If you (not, study) __do not study___ conditionals, you will have problems with grammar.
7. If I were you, I (not, be) ____would/might not be___ late to class for the exam next Friday.
37
(or)
1. (although) Although my dad is a good cook, I dont enjoy the meals he makes/ I dont enjoy
the meals my dad makes although he is a good cook. OR
always enjoy the meals he makes/ I always enjoy the meals my dad makes although he is a bad
cook.
2. (because) Because my dad is a good cook, I always enjoy the meals he makes/ I always
enjoy the meals my dad makes because he is a good cook. OR Because my dad is a bad cook, I
dont enjoy the meals he makes/ I dont enjoy the meals my dad makes because he is a bad cook.
3. (thus) My dad is a good cook. Thus, I always enjoy the meals he makes. OR My dad is a
bad cook. Thus, I dont enjoy the meals he makes.
4. (because of) Because of my dads good cooking skills/abilities, I always enjoy the meals he
makes. / I always enjoy the meals my dad makes because of his good cooking skills/abilities.
Because of my dads bad cooking skills/abilities, I dont enjoy the meals he makes. / I dont
enjoy the meals my dad makes because of his bad cooking skills/abilities.
5. (so) My dad is a good cook, so I always enjoy the meals he makes./ My dad is a bad cook, so
I dont enjoy the meals he make.
38
39
40
in many different countries. I want to see the world. I know that if I 7will stay in one place for a
long time, I will get bored. Due to8 the fact that I want to travel, I need to work hard in school.
Recently, I have gotten the 9highest grades in my classes, so my father is very proud. 10Whenever
I graduate, we will have a party.
Paragraph editing correction notes
Do not take off points if students correct parts of the paragraph that are not incorrect.
In addition to the correct answers highlighted in the paragraph, here are some other answers that
should be considered correct.
2. More difficult, more challenging etc.
3. Since my dad forced me
My dad forced me to study engineering, so now I am studying it.
5. ...good at writing; therefore, I was/ good at writing, so, I was
7. ...if I stayed in one place...I would get bored
8. Since or Because I want to travel,
10 After or other time phrase
41
42
Appendix E
Score Reporting Form
Week 10 Grammar Achievement Assessment
Name of Student_________________________________
Part of test
points/points
Percentage
correct
74.5 percent
/14
/7
/18
/8
/10
/5
/62