Admin, Forster
Admin, Forster
Educational Technology
2005, 21(1), 1-18
Introduction
Since the 1970s various information and communication technologies (ICT)
have been appropriated for teaching and learning, including digital
cameras, projection technologies, and the world wide web. Education
authorities, internationally, are promoting the use of ICT. For example, an
expected outcome in the UK is that:
…pupils become increasingly independent users of ICT tools and
information sources. They have a better understanding of how ICT can help
their work in other subjects and develop their ability to judge when and how
to use ICT and where it has limitations. They think about the quality and
reliability of information, and access and combine increasing amounts of
information. They become more focused, efficient and rigorous in their use
of ICT, and carry out a range of increasingly complex tasks. (Department for
Education and Employment, 1999, p. 20)
The second phase of the study took place in 2003 with a new student cohort
(n=42) in the science education unit. Consistent with the action research
paradigm (Carson & Sumara, 1997), the purpose of the second phase was
building on what had been learnt in the first phase in order to develop
further the ICT curriculum. As part of the second phase, a questionnaire
was designed to measure students’ perceptions of preparedness to teach science
with ICT. The questionnaire was administered early in the unit and again
after instruction and project work on ICT use had concluded. The principal
purpose for the questionnaire was that comparison of the responses from
the two implementations would assist unit evaluation: the comparison
would allow us to identify the domains in which students perceived they
were more prepared to teach with ICT after completing the unit, and the
domains in which they perceived they were less prepared. A second
purpose was that the first implementation would raise students’ awareness
of the scope of ICT use.
The main topic of this paper is the questionnaire. First, we place its design
in context by describing the theoretical views on learning that guided
instruction and the scope of the ICT curriculum. Sections on construction of
the questionnaire and the analysis of responses using Rasch modelling
follow. The Rasch model is suited to analysing a single trait, which in the
case of the questionnaire was ‘preparedness to teach science with ICT’.
Forster, Dawson and Reid 3
Finally, we discuss the Rasch outputs in relation to the conduct of the unit,
and discuss future refinement of the questionnaire and development of the
unit.
For assessment of the ICT component of the unit, each student was asked
to develop a curriculum resource that would require lower secondary
science students to use ICT and they were given the option of having their
resource placed on a CD that would be made available to all students
enrolled in the unit. It was expected that students would develop their own
ICT skills through completing the assessment requirements and, although
they were required to individually produce a resource, they were
encouraged to work collaboratively and seek technical assistance. Because
of the diversity of technical skills, they were given a choice of presentation
and submission formats. For example, the ICT resource could comprise a
web quest worksheet, a web site, an online assessment or an interactive
Powerpoint. The assignment could be submitted online, by email
attachment, CD, floppy disc, hard copy or a combination. Students were
asked to: provide instructions on how the resource could be used; discuss
the advantages and disadvantages when compared to traditional resources;
and describe how their resource was cognisant of current theories of
teaching and learning.
The assignment contributed 20% towards the unit mark. The allocation of
marks was based primarily on creativity, accuracy of content and
pedagogy rather than ICT skills, so as not to discourage and disadvantage
Forster, Dawson and Reid 5
The questionnaire
Construction of the questionnaire for measuring students’ preparedness to
teach science with ICT commenced with writing a bank of items and
designing suitable Likert scales. Then, items were chosen from the bank,
the interpretative validity of them was checked, and the instrument was
finalised.
Drawing on the experience of the first phase of the study and our other
experience in teaching with ICT, we created a bank of 72 questionnaire
items. They addressed: skills to do with Word documents, spreadsheets,
Powerpoint, the world wide web, digital cameras and other equipment; and
knowledge of ICT resources, pedagogy with ICT, and schemes for assessing
students’ performance and for assessing the merit of resources. Items were
intentionally made short and addressed one topic, to minimise ambiguity.
Negatively worded items were avoided, as they are more open to multiple
interpretations (Peck, 2000), responses to them are likely to be anomalous
(Taylor, Fraser, & Fisher, 1997), and responses can favour the lower end of
Likert scales (Andrich & van Schoubroek, 1989).
After deciding the scales, 36 items were selected from the item bank for use
in the questionnaire. They were chosen to cover ICT use that would be
treated explicitly in the science education unit. We grouped the items
under four major headings: Skills, Resources, Pedagogy and Critique, and
used subheadings to distinguish skill and resource types. The subheadings
for Skills were World Wide Web, Powerpoint, and Computer related
equipment. Subheadings for Resources were World Wide Web and CDs.
6 Australasian Journal of Educational Technology, 2005, 21(1)
Skills
Resources
More
WORLD WIDE WEB None One Two
than two
I can name web sites with
3. -graphics for learning science. n n n n
4. -information for preparing my n n n n
teaching.
Analysis of responses
Rasch modelling
The Rasch model for ordered response categories (Andrich, de Jong &
Sheridan, 1997) and the rack and stack approach described by Wright
(1996, 2003) were used to analyse the data, with the RUMM2010 software
(Andrich, Sheridan & Luo, 2000). The Rasch model locates persons and
items on a scale, where locations reflect person ability (i.e., the extent to
which persons possess the trait being examined) and the difficulty of items
(high difficulty is synonymous with few ‘highly competent’ responses).
More precisely, the difficulty value for an item equals the natural logarithm
of the odds of being successful on the item (Cavanagh & Romanoski, 2004).
Being successful is responding in the top half of the Likert categories. If a
person and item have the same location (ie., the same value on the scale),
then the probability that he or she was successful on the item is 50%. The
probability that persons were successful on items below them on the scale
is greater than 50%, and the probability they were successful on items
above them is less than 50%.
Racking and stacking (Wright, 2003) allows the treatment of pre-test and
post-test data. With racking, the items in the pre-test and post-test are
considered to be different items. The rationale is that the same students are
involved in each test but their knowledge about the items changes - in
other words, the items change for them. The purpose of racking is the
examination of the difficulty of individual items, in other words, the
tendency of students to respond at the left hand end or right hand end of
the Likert scale on each item in the pre-test and on each item in the post-
test.
Stacking (Wright, 2003) involves treating students who completed the pre-
test and post-test as different students, each of whom answered the same
items. The rationale for stacking is that students change as a result of a
course of instruction, so they are effectively different students. The purpose
of stacking is the examination of the ability of individual students across all
items in the pre-test and across all items in the post-test.
an item is the ability of persons for whom scoring 0 and 1 is equally likely;
the second threshold is the ability of persons for whom scoring 1 and 2 is
equally likely; and so on for the third and other thresholds. Reversed
thresholds, where the thresholds are out of order (eg., the third threshold is
between the first and second thresholds) indicate the Likert scale does not
work well for the item. The thresholds should indicate a monotonic (one
direction) response process (Hagquist & Andrich, 2004). Reversed
thresholds can be addressed by rescoring items to combine adjacent
categories of response, or problematic items can be deleted from the
analysis.
Two final statistics that measure reliability were checked for the racked and
stacked data: the separation index, which indicates how well item
thresholds cover the range of person abilities; and the chi-square
probability for the item-trait interaction, which indicates how well students
agree on difficulties of items across the scale and whether the assumption
of unidimensionality (analysis of a single trait) is reasonable (Cavanagh &
Romanoski, 2004).
We used two class intervals only for the analysis of the ICCs because the
data set was small (n=33). Visual inspection of the ICCs indicated that the
most grossly misfitting item was in the pre-test (Item 23). This item had the
lowest chi-square probability of all items (p=0.005). However, (a)
discrimination was in the correct direction (actual mean scores for students
in the first class interval were lower than the actual mean scores for
students in the second class interval), (b) data for the corresponding item in
the post-test fitted the model well (p=0.646), and (c) inspection of the
wording did not indicate any reason to reject the item. Therefore it was
retained. Three other items had chi-square probabilities of less than 0.05
(p= 0.021, 0.022 and 0.042) and were retained on the same basis. The
separation index for the final set of racked data was 0.962, and the chi-
square probability for item-trait interaction was 0.464, which indicated
high reliability and mapping onto a single trait.
The analysis of stacked data, where the students in the pre-test and post-
test were treated as (2x33) different students who had each answered the
same (34) items, proceeded as follows. The scoring of responses for racked
data was retained, so responses on most items were scored 0, 1, 1, 2. No
items showed reversed thresholds, and discrimination between groups was
appropriate for all items, according to ICCs with three class intervals and
the chi-square probabilities, which were greater than 0.05 for all items.
Three class intervals instead of two were chosen because of the larger
number of data points (66 instead of 33). The separation index was 0.939
10 Australasian Journal of Educational Technology, 2005, 21(1)
and the chi-square probability for item-trait interaction was 0.803, which
indicated high reliability.
The items with low negative locations attracted high numbers of responses
at the right hand end of the Likert scales. Negative changes in location
indicated that students as a group knew more at the end of the unit than at
the beginning about the topic addressed by the that item (whether due to
their experience in the unit or outside it). For example, pre-test Item 1 (I
know how to search the web) was located at –2.68, so attracted a relatively
high number of ‘highly competent’ responses. The same item in the post-
test was located at -3.50, so attracted an increased number of responses at
the ‘highly competent’ end of the scale. The change in location was
negative (-0.82) and indicated that, overall, students’ competence in
searching the web was greater at the conclusion of the unit than before it.
The distribution of responses for the item ‘I know how to search the web’ is
shown in Figure 1.
f f
25 25 0 no experience
20 20 1 novice
2 competent
15 15 3 highly competent
10 10
5 5
Figure 1: Pre and post test responses on ‘I know how to search the web’.
Items with high locations (see Table 1) attracted relatively low numbers of
responses at the right end of the scales, and positive changes in location
between corresponding items indicated students as a group perceived their
competence/ knowledge was less at the end of the unit than at the
beginning, in regards to the topics addressed by the items. For example,
pre-test Item 4 (I know how to import video into a web page) was located
Forster, Dawson and Reid 11
at 2.85 and the same item in the post-test was located at 4.78. The change in
location was 1.93. So, students’ responses indicated that, as a group, they
started knowing little about importing video into a web page, and
perceived they knew less by the end of the unit. Their awareness of limited
knowledge was raised over the duration of the unit and the unit apparently
did not successfully address knowledge in this domain.
The change in location for most items was negative (see Table 1), implying
gains in knowledge/competence. The result is highlighted in Figure 2,
which shows the plot of post-test location against pre-test location for all
items. Points below the line belong to items for which the change in
location was negative. Points close to the line belong to items for which the
12 Australasian Journal of Educational Technology, 2005, 21(1)
Post-test 6
Post-test
locations
locations
5 Item 4 Skills
4
Resources
3
2 Pedagogy
1
Critique
0
-5 -4 -3 -2 -1-1 0 1 2 3 4 5 6 Serie
Pre-test locations
Pre-test
-2
s5
locations
-3
Item 1
-4
-5
-6
The post-test results, in particular, had implications for the conduct of the
unit and future units for the student cohort. We divided the items into
three groups according to post-test locations - the eleven items with highest
locations (the points that were highest in the vertical direction on the
graph), the eleven items with the lowest locations, and a middle group of
12 items.
The results to do with techniques for creating web pages (Items 2-5),
Powerpoint (Item 10), and digital cameras (Items 13-14) were not a major
concern, for the purpose of the ICT intervention was not to increase ICT
skills per se, and students can enrol in an ‘Introduction to Computing’ unit
for the development of such skills. Rather, the ICT focus in the unit was
enhancing knowledge about ICT pedagogy and critical awareness of ICT
resources within the context of the science education. However, the
assessment will be modified for future use, to include guidelines on basic
ICT skills, and technical assistance will be more widely available in relation
to the assessment. Only those students who chose in their authentic
assessment to construct a web page or a Powerpoint resource, and use a
digital camera were likely to have developed the associated skills.
The eleven items with lowest locations, which indicated relatively high
levels of knowledge/competence after the unit, came from all categories.
They were skills Item 1; resources Items 17-18, 21; pedagogy Items 25-26,
28; and critique Items 31-34. These items pointed to domains where
students’ knowledge/competence more adequately prepared them for
teaching science with ICT and, by implication, pointed to domains of
knowledge that were treated adequately by the unit. Hence, the
questionnaire results indicated that the emphasis on pedagogy and critique
in the unit had impacted favourably on students’ learning.
students. They were skills Items 7-9, 12; resource Items 16, 19, 23-24;
pedagogy Items 27, 29; and critique Item 30.
The issue of interest with the stacked data was the pre-test and post-test
locations of each student. The locations are graphed in Figure 3. High
locations indicate students chose relatively high numbers of responses in
the right half of the Likert scales. Therefore, high locations signify
relatively high preparedness to teach science with ICT (as judged by the
students). Points above the line y=x indicate that post-test locations were
greater than pre-test locations. Hence, the majority of students chose more
responses to the right on the Likert scales in the post-test than in the pre-
test, which implies they felt more prepared to teach science with ICT after
the unit than before it.
Post-test
locations
5
(high preparedness)
0 Pre-test locations
-4 -3 -2 -1 0 1 2 3 4
(low preparedness) -1 (high preparedness)
-2
-3
-4
(low preparedness)
Points on the graph (Figure 3) that are close to the line y=x and above it
indicate marginal gains, whereas points distant from the line indicate
greater gains. Furthermore, judging by the gap between the least squares
regression line and the line y=x, gains for students with low pre-test
Forster, Dawson and Reid 15
locations tended to be greater than gains for students with higher pre-test
locations. Therefore, students who started the unit with relatively low
levels of preparedness to teach with ICT may have gained more from the
unit than some of the students who started with higher levels of
preparedness. Even so, the points in the third quadrant indicate low post-
test locations, and indicate students for whom special attention may be
warranted in future units, in regards to moving them towards using ICT
for teaching science.
Conclusion
In regards to the instrument, finding a single Likert scale that would
address skills and knowledge to do with preparedness to teach science with
ICT proved impossible. As a result, we included two scales in the
questionnaire. Rasch analysis of the response data after rescoring (to 0, 1, 1,
2) indicated the items and scales were directed sufficiently at a single trait
(‘preparedness to teach science with ICT’).
The outcomes of the questionnaire also indicated that students from the
whole range of ICT backgrounds felt more prepared to teach with ICT after
the unit than before it, and therefore offered evidence that the principles
underpinning instruction were sound. Hence, the science education unit
will continue to include and expand on opportunities for the instructor to
model the use of ICT, and for students to (a) increase their awareness of
ICT resources through using them and (b) increase their understanding of
effective ICT pedagogy through discussion and project work.
Acknowledgement
The study of which this paper is part was supported by a grant from Edith
Cowan University.
References
Andrich, D., de Jong, J. H. A. L. & Sheridan, B. E. (1997). Diagnostic opportunities
with the Rasch model for ordered response categories. In J. Rost & R.
Langeheine (Eds), Applications of latent trait and latent class models in the social
sciences (pp. 59-70). New York, NY: Waxmann.
Andrich, D., Sheridan, B. & Luo, G. (2000). RUMM2010: A Windows interactive
program for analysing data with Rasch unidimensional models for measurement. Perth
Western Australia: RUMM Laboratory.
Andrich, D. & van Schoubroek, L. (1989). The general health questionnaire: A
psychometric analysis using latent trait theory. Psychological Medicine, 19, 469-
485.
Biotechnology Australia (2001). Biotechnology Online School Resource. Canberra, ACT:
Commonwealth of Australia. [verified 23 Oct 2004]
http://www.biotechnology.gov.au/biotechnologyOnline/Resource/Resource.htm
British Educational Communications and Technology Agency (2003). What research
says about using ICT in science. Report to the BECTA ICT Research network. [15 Feb
2004, verified 23 Oct 2004] http://www.becta.org.uk/research/ictrn/
Cahill, J. & Spence, R. (1999). Heinemann science for Western Australia 2: An outcomes
based approach. Post Melbourne, Vic: Reed International.
Forster, Dawson and Reid 17
Carson, T. R. & Sumara, D. (1997). Action research as a living practice. New York, NY:
Peter Lang.
Cavanagh, R. F. & Romanoski, J. T. (2004). Application of the Rasch model to
develop a measure of classroom information and communication technology
learning culture. Paper presented at the Second International Rasch Conference,
Murdoch University, Western Australia, January.
Curriculum Council of WA (1998). Curriculum Framework for Kindergarten to Year 12
Education in Western Australia, Perth, Western Australia: Curriculum Council of
WA.
Dawson, V. M. & Reid, D. (2003). Preparing preservice science teachers for ICT rich
secondary classrooms. Paper presented at the 34th annual conference of the
Australasian Science Education Research Association (ASERA), Melbourne,
July.
Dawson, V., Forster, P. & Reid, D. (2004). Evaluating preservice teachers’ ICT
pedagogy skills. Paper presented at the annual conference of the Australasian
Science Education Research Association (ASERA), Armidale, New South Wales,
July.
Dawson, V. & Taylor, P. C. (1998). Establishing open and critical discourses in the
science classroom: Reflecting on initial difficulties. Research in Science Education,
28(3), 317-336.
Department for Education and Employment (1999). Information and communication
technology: National curriculum for England. London: Department for Education
and Employment and Qualifications and Curriculum Authority.
Dorling Kindersley (1996). The Ultimate Human Body 2.0. London: Author.
Hagquist, C. & Andrich, D. (2004). Is the sense of coherence-instrument applicable
on adolescents? A latent trait analysis using Rasch-modelling. Personality and
Individual Differences, 36, 955-968.
Leamnson, R. (2001). Does technology present a new way of learning? Educational
Technology & Society, 4(1), 75-79. [verified 5 Oct 2004]
http://www.ifets.info/journals/4_1/leamnson.html
Linacre, J. M. (2003). Size vs. significance: Standardized chi-square fit statistic. Rasch
Measurement Transactions, 17(1), 918. [2 Feb 2004, verified 23 Oct 2004]
http://www.rasch.org/rmt/rmt171n.htm
Taylor, P. C. (1998). Constructivism: Value added. In B. Fraser & K. Tobin (Eds), The
international book of science education (pp. 1111-1123). Dordrecht: Kluwer
Academic.
Taylor, P. C., Fraser, B. J. & Fisher, D. L. (1997). Monitoring constructivist learning
environments. International Journal of Educational Research, 27, 293-302.
Tytler, R. (2002). Teaching for understanding in science: Student conceptions,
research, and changing views of science. Australian Science Teachers’ Journal,
48(3), 14-21.
Wang, Y.-M. (2002). When technology meets beliefs: Preservice teachers' perception
of the teacher's role in the classroom with computers. Journal of Research on
Technology in Education, 35(1), 150-161.
Wildy, H., Forster, P., Louden, W. & Wallace, J. (2004). The international study of
leadership in education: Monitoring decision making by school leaders. Journal
of Educational Administration, 42(4), 416-430.
Wright, B. D. (1996). Time 1 to time 2 comparison. Rasch Measurement Transactions,
10(1), 478-479. [2 Feb 2004, verified 23 Oct 2004]
http://www.rasch.org/rmt/rmt101f.htm
Wright, B. D. (2003). Rack and stack: Time 1 vs. time 2. Rasch Measurement
Transactions, 17(1), 905-906. [2 Feb 2004, verified 23 Oct 2004]
http://www.rasch.org/rmt/rmt171a.htm