Internet Learning Volume 3, Number 1, Spring 2014 | Page 39
Internet Learning
Data Coding and Analysis
To satisfy the first and second objectives
of this study, data were collected from
the three online courses and analyzed separately
with Winsteps--a Windows-based
software that assists with several Rasch
model applications--particularly in the areas
of educational testing, attitude surveys
and rating scale analysis (Linacre, J. M.,
2009).
To address the third objective of
this project the resulting data were treated.
Students’ results were converted into a
measure that is comparable to the reviewers’
rating. Student responses of to a great
extent “4” or to a very great extent “5” are
used as at or above 85% level and coded as
“1”. Student responses of to a moderate extent
“3”, to some extent “2” and to little or no
extent “1” are used as below 85% level and
coded as “0”. According to the majority
rule principle, if 2/3 of the students selects
to a great extent “4” or to a very great extent
“5” for an item in the survey then it is
determined that the course meets that specific
standard from a student’s perspective.
See Tables 1, 2, and 3.
Three QM certified peer reviewers
reviewed the three courses according to
QM standards and input their scores into a
spreadsheet. If a standard was met, “1” was
recorded for the standard. If a standard was
not met, “0” was recorded for the standard.
If two (2/3) of the peer reviewers assigned a
score to a specific standard, then it was determined
that the course met the standard
from a peer reviewer’s perspective. See Tables
1, 2, and 3.
The data were treated in a spreadsheet
and analyzed with SPSS. A nonparametric
Mann-Whitney U test (2 independent
samples) was used to evaluate median
differences between the two groups (students
and peer reviewers). Although the
peer reviewers were asked to take a student’s
point of view, the two groups were
independent.
Results
Course A
Thirty-five out of the 44 students
completed the course design evaluation
survey with a response rate of 79.55%. The
person reliability was 0.83 and the item reliability
was 0.48.
The item statistics indicate that
Item 1 (MNSQ = 3.31) will need to be revised
and Item 16 (MNSQ = 3.13) will need
to be revised or dropped if the instrument
is used in the future.
The Item Map (Fig. 1) indicates
that there are three tiers regarding course
quality from a student’s persepctive. Tier
I contains the items that students strongly
agreed with, thus indicating the course
met Standards 2.1 (Item 4), 3.1 (Item 14),
3.2 (Item 15), 6.1 (Item 20), 3.3 (Item 7) to
a great extent. Tier II contains items that
students agreed with, which indicates that
course met those standatds to a moderate
extent. Tier III contains items that students
agreed with to some extent, thus indicating
that the course did not meet Standards 7.1
(Item 24), 5.1 (Item 13), and 8.1 (Item 25).
Course B
Eighteen out of the 38 students
completed the course design evaluation
survey with a response rate of 47.37%. The
person reliability was 0.95 and the item reliability
is 0.63.
The item statistics indicate that
Item 14 (MNSQ = 2.29) needs to be revised
if the instrument is used in the future.
The Item Map (Fig. 2) indicates
that there are three tiers regarding course
38