Internet Learning Volume 3, Number 2, Fall 2014 | Page 60
Using Early Warning Signs to Predict Academic Risk in Interactive, Blended Teaching Environments
by one of the authors) that permits the instructor
to pose non-multiple-choice ConcepTests
to students (e.g., sketch a graph)
and then can use student responses to automatically
group students for discussion.
Peer Instruction Self-Efficacy Instrument
To measure self-efficacy, both in general
and in a Peer Instruction environment,
we developed a set of
25 Likert-scale items aimed at measuring
various qualities related to self-efficacy,
including qualities that we believed would
be unique to a Peer Instruction environment.
These items were based on Fencl
and Scheel’s (2003) Sources of Self-efficacy
in Science COurses (SOSESC). The
statements, such as “When I come across a
tough physics problem, I work at it until I
solve it” were designed to gather data about
students’ self-reported beliefs their abilities
in physics and in a Peer Instruction environment.
As this was the first time the instrument
was used, this study simultaneously
served as an opportunity to use measurements
from this instrument as covariates
as well as an opportunity to gather some
initial validation data from the study. We
later extracted two subscales that we used
as variables in the study. The first subscale
was a seven-item set that conceptually covered
general self-efficacy; Cronbach’s coefficient
alpha reliability for this subscale was
0.85 when the scale was administered at the
beginning of the semester (pretest) and 0.83
when it was administered at the end of the
semester (posttest). The second subscale
was a six-item set that conceptualized our
notion of “Peer Instruction self-efficacy.”
Unsurprisingly, since the notion of self-efficacy
in a Peer Instruction environment is
a new concept, this subscale proved to be
somewhat less reliable, with coefficient alpha
values of 0.53 for the pretest and 0.68
for the posttest. The fifteen items used in
these subscales (as well as the other ten
items that were ultimately not used in this
analysis) appear in Appendix A.
Data Set
Our data set included all performance
data for students over the
semester, including:
• Summative assessment data collected
over the semester, including scores on
problem sets (eight over the course of
the semester), three midterm exams,
and the final exam.
• Pre and post-test scores on the Conceptual
Survey of Electricity and
Magnetism (Maloney, O’Kuma,
Hieggelke, & Van Heuvelen, 2001), a
conceptual inventory measuring understanding
of fundamental concepts
in electricity and magnetism.
• Pretest and posttest data from a noncognitive
assessment, developed by
the authors, to measure students’
self-efficacy in a Peer Instruction
environment as well as attitudes towards
science and education. Seven
items measured general self-efficacy;
Cronbach’s coefficient alpha reliability
for this subscale was 0.85 for the
pretest and 0.83 for the posttest. Eight
items measured Peer Instruction
self-efficacy; this subscale proved to
be somewhat less reliable, with coefficient
alpha values of 0.66 for the
pretest and 0.73 for the posttest.
• Formative assessment data consisting
of student responses to ConcepTests
asked by the instructor in class. These
59