zmWmQs
zmWmQs
zmWmQs
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
Designing Video for Massive Open Online-Education:<br />
Conceptual Challenges from a Learner-Centered Perspective<br />
Carmen Zahn, Karsten Krauskopf, Jonas Kiener, Friedrich W. Hesse<br />
Table 1. Measures grouped into three levels .<br />
Level Variable Measure<br />
1: Cognitive learning<br />
outcome<br />
History content knowledge<br />
acquisition<br />
Factual Knowledge Test<br />
Picture Recognition Test<br />
2: Surface level effects on<br />
collaboration and learning<br />
Performance, collaboration<br />
and learning<br />
Number of panels created in<br />
partnership<br />
Number of comments<br />
Length of comments<br />
Collaboration index<br />
3: Deeper level effects on<br />
collaboration and learning<br />
Performance, collaboration<br />
and learning quality<br />
Number of panels referring to<br />
details<br />
Number of utterances in<br />
comments addressing historical<br />
content<br />
Number of utterances in<br />
comments addressing filmic style<br />
Number of utterances in<br />
comments<br />
integrating aspects of historical<br />
content and filmic style<br />
Results<br />
(M = 1.8, SD = 0.7)<br />
Comparability of the Conditions. To ensure comparability<br />
of the two conditions a number of control variables<br />
were investigated prior to analysis. Participants in both<br />
conditions did not differ with regard to age, expertise in<br />
technological and film/media production, interest in the<br />
historical content, or factual knowledge of the historical<br />
context (all p > .10). Similarly, chi-square tests on gender<br />
and gender composition of dyads (same gender vs. mixed)<br />
did not yield significance (p > .10). Thus, conditions were<br />
considered comparable.<br />
Furthermore, we examined the named control variables,<br />
to investigate whether participants on average exhibited<br />
medium level values as we expected. With respect<br />
to domain-specific knowledge, dyads’ pre-questionnaire<br />
scores were above average with a mean of M = 9.0 (SD<br />
= 1.3) correct answers out of 12. Similarly, participants’<br />
interests in the historical content (M = 3.5, SD = 0.4, theoretical<br />
maximum = 5) and participants’ prior computer<br />
experience (M = 3.5, SD = 0.5, theoretical maximum = 5)<br />
were also higher than average. Their self-reported expertise<br />
in film and media production, however, was very low<br />
Knowledge Acquisition. We compared the conditions with<br />
regard to participants’ performance in the post-experimental<br />
factual-knowledge test on the historical content. Test<br />
performance was in general above average (M = 33.1, SD =<br />
2.4, theoretical maximum = 45). We did not find a significant<br />
difference between conditions, t(34) = -0.80, p = .43. This<br />
indicates that in both conditions, participants were equally<br />
successful in understanding the content of the video and<br />
the historical material. Yet, marginally significant differences<br />
between the conditions were found with regard to<br />
the visual recognition test. A t-test revealed a marginally<br />
significant effect, t(34) = 1.79, p = .08, d = 0.60, showing a<br />
better performance in the picture recognition test for the<br />
dyads in the design condition (M = 25.7, SD = 1.3 compared<br />
to the discussion condition, M = 24.9, SD = 1.4, respectively).<br />
Although just a trend, this might indicate that the students<br />
with a design task instruction, seemed to have paid more<br />
joint attention to visual information than the students discussing<br />
the visual content.<br />
Table 2: Between group comparisons of indicators for cognitive<br />
learning outcome.<br />
Research Track |164