Go home now Header Background Image
Search
Submission Procedure
share: |
 
Follow us
 
 
 
 
Volume 4 / Issue 3 / Abstract

available in:   PDF (216 kB) PS (456 kB)
 
get:  
Similar Docs BibTeX   Write a comment
  
get:  
Links into Future
 
DOI:   10.3217/jucs-004-03-0273

Evaluating Educational Multimedia in the Context of Use

Steven McGee, Ph.D.
NASA Classroom of the Future, Wheeling Jesuit University
United States
mcgee@cotf.edu

Bruce Howard, Ed.D.
NASA Classroom of the Future, Wheeling Jesuit University
United States
howard@cotf.edu

Abstract: Researchers at the NASA Classroom of the Future have been using the design experiment framework to conduct evaluations of multimedia curricula. This method stands in contrast to more traditional, controlled experimental methods of evaluating curricular reforms. The methodology presented here is integrated with Walter Doyle's [1983] notion of using academic tasks to describe how classroom activities impact student learning. We will report the results from a design experiment with a multimedia program developed at the NASA Classroom of the Future, and we will examine the methodologies that were used in the evaluation.

Key Words: Multimedia, Evaluation, Curriculum, Design Experiment, Methodology, Computer Uses in Education, Science Education

Categories: K.3.1, K.3.2, K.3.m

1 Introduction

Traditionally, the evaluation of curricula has been conducted using the "horse race" method. In this method, evaluators challenge a traditional curriculum "horse" to a race using an experimental curriculum "horse." The horses are placed in the starting gate by giving students a pretest that covers the learning objectives. The teachers in each condition run the horses, and a posttest serves as the finish line. Whichever curriculum horse achieves the greatest student-learning gains from pretest to posttest is declared the winner. If the experimental curriculum wins enough races across a variety of contexts, then reformers proclaim it superior to the traditional curriculum. Although the horse race method has been widely used, there are two fundamental assumptions associated with the this method of curriculum evaluation that have recently been called into question.

The first assumption that has been questioned regarding the horse race method is the feasibility of creating comparable experimental conditions and isolating the variable of experimental interest [Brown 1992][Collins 1990]. Both notions are derived from laboratory-based research and both translate poorly to an actual classroom setting. In a laboratory, researchers randomly assign subjects to either treatment or control

Page 273

groups to ensure that subjects in each group are comparable. Subjects in each condition are treated the same, with the exception of the variable of experimental interest. By ensuring comparability and isolating the variable of experimental interest, researchers can be confident that the superiority of one group over the other is attributable only to the treatment. The horse race method of evaluation uses principles of quasi-experimental research [see Cook and Campbell 1979] to apply experimental constraints to classroom settings.

When moving from the laboratory to the classroom, however, it is not feasible to meet the stringent requirements of ensuring comparability between groups or to isolate the variable of experimental interest. It is virtually impossible to create comparable groups using entire classrooms as a group, since each classroom is a unique learning context that is affected by factors such as state and local education policies, and the classroom teacher. Quasi-experimental researchers overcome this limitation by assuming that the entirety of instruction within each group was implemented identically, thus concealing differences in how teachers implemented the experimental curriculum. In addition, it is difficult to isolate only the variable of experimental interest, since reformers must often enact changes to multiple variables at once in order for the reform itself to be successful. Thus, the frenetic environment of real classrooms makes the laboratory style, horse race method of evaluation unrealistic for testing treatment effectiveness in real classrooms [Brown 1992].

The second assumption that has been questioned is that reform will occur through gradual acceptance of curricula deemed "superior" by the horse race method. The horse race evaluation method has been used repeatedly over the last century to promote reform. In science education reform, the goal has consistently been to help students engage in scientific inquiry and cease being passive receptors of information [McGee 1996]. In every era of reform, there have been pilot projects that have successfully engaged students in scientific inquiry. In those pilot projects, the horse race evaluation method has been used to demonstrate that students learned more and developed healthier attitudes toward science than did students in traditional science classrooms [e.g., Collings 1923][Bredderman 1983][Shymansky, Kyle, and Alport 1983]. However, contrary to the assumptions of the reformers, an investigation of today s science classroom indicates that teachers have not uniformly adopted curriculum materials that engage students in scientific inquiry [Bybee and DeBoer 1994], even though these materials were judged superior across a variety of horse races.

The horse race method overlooks an important reality: that teachers always need to adapt the materials they are given [Cuban 1993]. To accommodate the often conflicting demands of school boards, administrators, parents, and students, teachers must make decisions about what features of the reform materials to preserve and what features to adapt. However, the horse race method does not provide information about which features are most critical to the success of the implementation. Teachers are left with making an educated guess, about how to adapt the reform materials. Teachers can easily make the wrong adaptations, by neglecting or distorting critical elements of

Page 274

the innovation. Therefore, the horse race method fails to provide long-term support that would help teachers make prudent adaptations.

In order to use multimedia as a tool for supporting educational reform, it will be important that the evaluation of multimedia provide richer insights into performance than those of the horse race method. At the NASA Classroom of the Future (COTF) program, researchers have been pursuing ways of evaluating educational multimedia. The strength of their methodology lies in its ability to produce reliable evidence of superiority despite the frenetic environments of disparate classroom settings. Their evaluative techniques result in "implementation frameworks," which prescribe guidelines on how to implement the multimedia within classroom settings.

2 Conceptual Framework for COTF Evaluation Models

In order to successfully implement multimedia software that has been judged effective by the horse race evaluation method, teachers need to consider differences between their schools and the schools that participated in the evaluation. For example, if schools in the evaluation had a wealth of technology available, students could use the program individually. However, if a teacher's school owns only a fraction of the technology used in the evaluation, they would have to make adjustments to implement the new multimedia software. Teachers would have to consider whether it would be as effective to place students at computers in pairs. Dozens of such considerations are inevitable during actual implementation. Without tools for comparing model implementations with their own, teachers may unknowingly alter a critical reform element when adapting software, and expected benefits may be attenuated. The once "superior" new curriculum may be reduced to mediocrity and relegated to the bookshelf.

It is crucial that teachers be empowered with ways to successfully adapt multimedia, since nonuniformity of implementation across real classrooms is inevitable. Sections 2.1 and 2.2 describe two alternative evaluation models that underpin the evaluation methods being developed at COTF.

2.1 Academic Tasks: Solving the Problem of Multiple Curricular Levels

It is a well-documented observation that there are three levels of curricula and that the topics addressed within each differ significantly [Cuban 1993][Schmidt et al. 1996]. The curriculum as intended represents the content and processes outlined in official curriculum documents, such as the National Science Educational Standards [National Research Council 1996], as well as curriculum materials that implement the recommendations of the official documents [see Tab. 1]. The curriculum as implemented represents teachers' and students' experiences with the curriculum. Teachers adapt curriculum materials based on personal content goals, pedagogical beliefs, and the constraints imposed by their local schools [Schmidt et al. 1996]. In addition, students bring their own goals to the classroom and have varying

Page 275

experiences with the curriculum materials, resulting in learning outcomes that are unique to each student. The curriculum as measured represents the content and processes that appear on tests for measuring the amount of learning gain. The curriculum as measured is distinct from the other two levels because assessment materials contain only questions about content and processes for which it is possible to ask valid and reliable questions [Crocker and Algina 1986]. By focusing evaluation efforts on the curriculum as intended and measured, the horse race evaluation method fails to reveal the diverse ways in which teachers can implement the curriculum.

Curriculum Level Focus of Evaluation Other Research in This Area
1.Curriculum as Intended Examine software itself, and its learning objectives
  • "Official" Curriculum [Cuban 1993]
  • "Intended" Curriculum [TIMSS- Schmidt et al. 1996]
2.Curriculum as Implemented Examine classroom use, without experimental constraints
  • "Taught" and "Learned" Curricula [Cuban 1993]
  • "Implemented" Curriculum [TIMSS- Schmidt et al. 1996]
3.Curriculum as Measured Examine measures used to assess desired learning objectives
  • "Assessed" Curriculum [Cuban 1993]
  • "Attained" Curriculum [TIMSS- Schmidt et al. 1996]

Table 1: How levels of curricula may affect evaluation

Examining curricula from a variety of levels is not new. It has been used in evaluations previously, albeit with various terminology [see Tab. 1 Column 3]. For example, [Cuban 1993] argues that curriculum reform has not been sustainable because reformers have failed to recognize that multiple curriculum levels exist. He refers to these levels as the "official," "taught," "learned," and "assessed." The Third International Mathematics and Science Study (TIMSS) also recognizes the existence of multiple curriculum levels. In the design of international assessment instruments, the TIMSS staff capture information about all levels of the curriculum so that assessment results can be interpreted in light of students experiences with the curriculum as implemented [Schmidt et al. 1996]. They refer to these levels as "intended," "implemented," and "attained."

We concur with these researchers in believing that each level of the curriculum must be considered before conclusions are drawn about overall program effectiveness. At the level of the curriculum as intended, evaluators should consider improvements to be made to the software and should examine potential mismatches between expected learning objectives and actual learning outcomes. At the level of the curriculum as implemented, evaluators should examine how the software is used, without

Page 276

experimental constraints, and the degree to which this use mirrors what was designed. At the level of the curriculum as it is measured, evaluators should consider the validity of assessment instruments used to measure desired learning objectives.

Academic task research reveals the diverse ways in which teachers implement the curriculum (level 2). Academic task research was initiated by Walter [Doyle 1983] as a method of investigating the implemented curriculum by breaking down classroom procedures into individual, measurable units called academic tasks. Doyle defined academic tasks as composed of (a) the products that students are expected to produce, (b) the resources that are available to students while fulfilling the task, and (c) the operations that students are expected to perform to turn resources into the assigned products. Given the direct influence of academic tasks on student behavior, [Doyle 1983] proposed that focusing on academic tasks, as opposed to focusing on the official curriculum, would provide better explanations for how to support student learning [Doyle and Carter 1984]. He noted, "Tasks influence learners by directing their attention to particular aspects of content and by specifying ways of processing information" [Doyle 1983, p. 161].

[Stein, Grover, and Henningsen 1996] provide an excellent example of how academic task research can be used to investigate the curriculum as implemented. They examined how mathematical tasks evolved as teachers introduce them to students and how those tasks evolved further as students completed them. Stein et al. analyzed the task goals and task operations that the teachers assigned, and then investigated the conditions under which students were not able to meet the teacher s expectations. The results indicated that students often had great difficulty in meeting the demands of those tasks that most closely matched the [National Council of Teachers of Mathematics 1989] curriculum standards. In those cases, teachers often modified the task to make it more conventional, and in many cases, students completed the task in a rote fashion. Although more difficult, when students were successful at engaging in mathematical problem solving according to the NCTM standards, it led to greater improvements in learning outcomes [Stein and Lane, in press].

By focusing on the actual tasks that teachers assign, as opposed to the tasks suggested by software and curriculum materials, academic task researchers can get a more accurate account of how instruction influences student learning in classroom contexts. Academic task research also reveals differences among the various levels of the curriculum, whereas the horse race approach conceals differences. Through academic task research, evaluators can analyze the full spectrum of the curriculum in order to identify particular components that are essential for achieving desired learning outcomes.

The COTF model of evaluation includes two facets of academic task research that have been linked to student achievement: task completion and the level of task demand [Hiebert and Wearne 1993][Stein and Lane in press]. That is, we measure the degree to which students are able to successfully accomplish the task goal, and the degree to which such tasks present an appropriate level of cognitive demand. In

Page 277

addition, we report on the development of an indicator for task completion that can be used to monitor implementation of reform.

2.2 Design Experiments as an Alternative to Quasi-experiments

Design experiment research complements academic task research so that evaluators can be relieved of the quasi-experimental requirements of ensuring comparability between groups and isolating the variable of experimental interest. Through academic task research, it is possible to identify variables that are relevant to curriculum implementation. Through design experiment research, teachers can compare the efficacy of multiple variables at once using their own curriculum implementations as comparable groups. This comparison is important because teachers do not have the luxury of comparing their own teaching to that of other teachers. However, it is possible to compare their own improvements from one semester to the next. Thus, design experiments are a way for teachers to systematically test instructional manipulations and revisions, so as to adapt the implemented curriculum in a manner that is consistent with the curriculum designers' reform principles. Design experiments used in conjunction with academic task research can provide the direction that is necessary for sustaining reform at the classroom level.

Instead of comparing a treatment to a control group, teachers in a design experiment compare successive designs of their own curriculum [Brown 1992]. Teachers begin a design experiment by using their knowledge of content and pedagogy to design the best possible instruction for meeting the learning objectives or design goals. The resulting design becomes a working hypothesis that instantiates a teachers' ideas about how learning can take place through educational multimedia [Tanner 1997]. After implementing the instruction, design experiment teachers reflect on how well the outcomes match the design goals and then redesign their procedures by developing new working hypotheses. The next time the instruction is implemented, the teacher can compare how closely the results of the new implementation match the design goals as compared to the results of previous implementations. This design experiment cycle can continue indefinitely.

The success of the design experiment process rests on the soundness of the measures used to determine effectiveness. In the case of educational multimedia, research on academic tasks can provide objective measures of effectiveness. If there is overall improvement on objective measures from one year to the next, then it can be argued that the instructional modifications improved the quality of the instruction.

Throughout the history of reform, there are several examples of successful design experiments. John Dewey provides an early example of what would later come to be called a design experiment [Tanner 1997]. From 1896-1904, Dewey ran the University of Chicago Laboratory School. Dewey s fundamental belief was that students learn best when investigating problems of personal interest [Dewey 1916]. The lab school began as way for he and the lab school teachers to test this idea through successive designs of an elementary school curriculum. Dewey met with the

Page 278

lab school teachers on a weekly basis to help teachers articulate design goals for instruction and to help them reflect on the outcomes of instruction relative to the design goals. Through this iterative process, Dewey and the lab school teachers successfully developed authentic problems that balanced the learning objectives with student interest [Tanner 1997].

[Polman 1997] and [Linn and Muilenburg 1996] provide more recent examples of design experiments. Polman documented the curricular changes of one science teacher over a two-year period as he made a transition from traditional science instruction to a completely project-based approach. His students conducted three to four extended science projects per year. After each project cycle, the teacher modified the instructional supports based on the areas of difficulty for the students. Over time, the student projects more closely resembled the process of scientific inquiry. [Linn and Muilenburg 1996] describe the Computers as Learning Partner (CLP) project. The goal of the project was to teach students to use pragmatic models for investigating the distinction between heat and temperature. Over a ten-year period, the CLP group worked with the same middle school teacher to conduct a design experiment on a set of semester-long activities. At the end of each semester, the teacher assessed the percent of students who could accurately describe the distinction.

In each design experiment example, the teachers and researchers simultaneously manipulated multiple variables related to the curriculum. Working hypotheses were developed explicating how the design would achieve the desired outcome goal. These working hypotheses were tested through the process of implementation. Through multiple iterations of the same curriculum and a set of objective outcome measures, it was possible for the teachers and researchers to monitor whether the curriculum adjustments enhanced the overall quality of the instruction.

The design experiment and academic task research models described above served as the basis for methods used to evaluate a multimedia program developed at COTF called Astronomy Village: Investigating the Universe. As part of the design experiment, researchers conducted three studies across three semesters. The design experiment methodology allowed us to manipulate multiple variables related to the implementation, yet still compare the results of each implementation to determine whether the changes to the curriculum were enhancing the instruction. The comparability of the results was established using the academic task research methodology. In the next section we describe the Astronomy Village software. The remainder of the paper will describe the methodology we used to evaluate Astronomy Village. For a complete description of the results of the evaluation, see [McGee, Howard, and Hong 1998].

3 Astronomy Village: The Multimedia Evaluation Target

The NASA Classroom of the Future program (COTF) is a NASA-funded research and development center that specializes in the development and testing of educational multimedia for math, science, and technology education. In March 1996, COTF

Page 279

published a CD-ROM called Astronomy Village: Investigating the Universe for use as a curriculum supplement in high school science classrooms. It has been distributed to over 11,000 teachers, educators, and resource centers, and won Technology and Learning magazine's Science Software of the Year Award for 1996 [Technology and Learning 1996]. Astronomy Village uses the metaphor of living and working at a mountain-top observatory [the village] as the primary interface from which students investigate contemporary problems in astronomy [see Pompea and Blurton 1995][see www.cotf.edu]. Academic activities are designed to promote learning of both astronomical concepts and processes related to scientific inquiry. Students join a "research team" and choose one of ten "investigations" to complete. In the Stellar Nursery investigation, for example, students investigate how stars are born. For each investigation, students progress through five phases: background research, data collection, data analysis, data interpretation and presentation of results. For any given phase, there are from three to seven content-related activities to be completed before proceeding to another phase. The primary means of tracking progress through an investigation is the Research Path Diagram - a chart that shows each phase of the investigation and icons to represent activities within each phase [see Fig. 1]. Students also have access to an electronic LogBook for recording their scientific notes and observations.

Figure 1: Stellar Nursery research path diagram

4 Method and Data Sources

In conjunction with academic task research techniques, researchers at COTF recently completed a design experiment using Astronomy Village. The design experiment spanned three semesters across two school years. In each study of the design

Page 280

experiment, two area schools used the COTF facilities, and students conducted astronomy projects using the software.

4.1 Design Experiment Populations

All three studies of the design experiment involved schools from a rural community with a population of approximately 35,000. The demographics for each study varied [see Table 2]. In the first study, thirteen students from the ninth grade class of a girls' academy (college preparatory) participated. In the second study, nine students from the eighth-grade class of the same academy participated. In the third study, twelve students from the tenth and eleventh grade of a large public high school participated. Students from the third study were from an at-risk population. In each case, students attended class daily for approximately four weeks at the COTF facility in lieu of their science class. Sessions using Astronomy Village were co-taught by the students' classroom teacher and the first author.

  Study 1 Study 2 Study 3
Time Frame Apr-May 1996 Oct-Nov 1996 Mar-Apr 1997
Number of Students 13 9 12
Gender All Female All Female F = 4; M = 8
Grade Level 9th 8th 10th-11th
Type of Students college preparatory college preparatory at-risk

Table 2: Demographics of Design Experiment Populations

It should be noted that the students in study 2 were younger than students in study 1, and in study 3 the population shifted from college preparatory students in a private school to at-risk youth in a public school setting. Based on previous research [Blank and Gruebel 1995], the population characteristics of each study would predict that the ninth grade students from the college preparatory academy should perform the best with Astronomy Village.

4.2 Data Sources

There were three sources of data for this investigation that enabled academic task research to support the design experiment. The first source was a compilation of documents in the students' electronic notebooks. In the first two studies, students used the electronic notebook embedded within Astronomy Village. This notebook is called the LogBook. In the third study, students used an electronic notebook called the Collaboratory Notebook developed as part of the Learning Through Collaborative Visualization Project (CoVis) Project at Northwestern University [Edelson and O'Neill 1994]. The second source of data was videotapes of student interactions while

Page 281

using the software. The third source of data was field notes and classroom observations by the teachers and first author.

For each academic task in Astronomy Village, students were expected to produce written responses in their electronic notebooks. Examples of notebook entries include activity summaries, answers to "press conference" questions, answers to teacher-posed questions, and reflections on "thought experiments." Activity summaries consisted of a brief description of the activity, a statement of how the activity related to the main research question, and any new questions that arose from the activity. Press conference questions were posed to students by members of a simulated press corps to which they had to respond with answers from their investigation. Thought experiments were designed to prompt students to integrate path activities. Other collected virtual "data" included images, snippets of articles they had read, or student reflections.

4.3 Dependent Measures

As indicated from prior research on academic tasks, it is essential that students be able to complete academic tasks in a cognitively demanding way in order for them to learn from the tasks [Hiebert and Wearne 1993][Stein and Lane in press]. The primary dependent measure for each study within the design experiment was an indicator of task completion called the task completion rate. This was an objective measure of the extent to which students were able to complete the task goals for the assigned academic tasks. Task completion rate was defined as the number of activity summaries that students completed divided by the number of academic tasks that the mentor suggested.

The task completion rate was mediated by the level of cognitive task demand associated with each academic task. In utilizing academic tasks, teachers must consider the cognitive demands of the task in the context of learning objectives. In order for learning to take place, there needs to be an appropriate amount of cognitive demand. If an academic task is too demanding, students will encounter a cognitive overload and will have difficulty completing the task [Sweller and Chandler 1994]. If an academic task is not demanding enough, students won't learn from the task. Through an examination of the task completion rate, it became possible to identify academic tasks that might be too demanding for the students. In those cases where students were systematically not completing a task, the teachers provided instructional supports that would lessen the cognitive overload associated with those tasks. Student notebooks, teacher field notes and video tapes provided the data to make judgments about task demands.

5 Results of A Design Experiment with Astronomy Village

In brief, the design experiment began with instructional procedures in line with the Astronomy Village curriculum as it was intended, and examined the effects of the curriculum as it was consequently implemented. Each proceeding study then refined

Page 282

the procedures for more effective outcomes. In the description of the first study, there is a complete discussion of the instructional procedures used and the impact of that instruction on the task completion rate. In the descriptions of the later studies, there is a discussion of how instructional procedures were subsequently modified. The section concludes with a discussion of planned modifications to the Astronomy Village software based on results from the design experiment.

5.1 Study 1

In the first study (20 days in April-May 1996), the overall goal for the students was to complete all of the academic tasks as suggested by the virtual mentor. Our purpose in this study was to implement the curriculum as closely as possible to the curriculum as intended by the software designers. Students completed activities related to background research, data collection, data analysis, data interpretation, and presentation of results. As intended by the developers, students began by watching a videotape introduction to the software. Next, they worked within their project teams to complete a tutorial on using the software. After completing the tutorial, students proceeded to the virtual Conference Center in Astronomy Village, where the virtual mentors were available to describe the different investigations. The students selected different mentors to hear a description of the research that each mentor was conducting at the Village. The students chose an investigation that interested them and "signed-up" to join an investigation "team." Next, students were given access to the Research Path Diagram, to show them the resources on the Astronomy Village CD-ROM that related specifically to their investigation. By following the Research Path Diagram, students should have been able to complete the steps necessary to learn the appropriate concepts and conduct the level of problem solving needed for their investigation. Each recommended activity was defined as one academic task for purposes of this evaluation.

Within the first week of software use, the interplay between task completion and task demand became apparent. Students were expected to record the results of each academic task in their LogBook. However, the software did not provide any templates to help the students create LogBook entries. The students were either copying and pasting entire articles in their LogBook, taking detailed notes, or not writing anything in their LogBook. The teachers felt that none of these strategies were going to be effective for the students to be able to synthesize all of the investigation activities. Therefore, the teachers created an activity summary worksheet that students used to summarize each activity. For each summary, the students were asked to include a brief description of the activity, a statement indicating how the activity fit into their investigation, and any new questions that arose during the activity.

The activity summary template provided an instructional support to lessen the cognitive overload associated with the investigation. Activity summary worksheets made task completion easier, although completion rates were still below teacher expectations. The average task completion rate for this study was 42%. This value indicates that over the four-week period, students completed less than half of the

Page 283

activities that the mentor suggested. Further analysis of the task completion rate within each phase of research (i.e., background research - 55%, data collection - 75%, data analysis - 35%, data interpretation - 20%, and presentation - 27%) revealed that the task completion rate was much lower during later phases of research [see Figure 2]. This analysis of the task completion rate and task demand indicates that the Research Path Diagram and virtual mentors were not sufficient to guide students in complex problem-solving. If students are not able to complete the tasks in the path, they will not achieve the desired learning outcomes.

Figure 2: Task Completion Rate by Research Investigation Phase for Study 1

In examining additional data from the four weeks of instruction, there were several noteworthy observations. First, students spent most of their time in the background research phase of Astronomy Village, which left little time for the remaining activities. This was most likely due to the high task demand of background research activities, which hindered students' attention to time management issues. That is, students were observed to be engrossed in these early activities, to the exclusion of process-related discussions or questions that indicated that they were tracking individual activities in the context of the larger investigation. Second, students averaged 3.5 class periods to complete the tutorials and select their investigation. Student comments and questions indicated that what was learned during the advance tutorials was not remembered very well, if at all, when the time came to apply such knowledge. It was concluded that such tutorials, if needed at all, should be completed on an as-needed basis, thus lessening cognitive overload during the earlier phases, where demand was highest. In the next study, the teacher and first author used these results to guide a redesign of the task demands to meet the needs of the students.

5.2 Study 2

Based on the apparently poor performance of the students in Study 1, the teachers for Study 2 (20 days in October-November 1996) modified the curriculum slightly from

Page 284

the procedures intended by the designers of the Astronomy Village. The overall goal of the modifications was to relieve cognitive overload so that task completion could be improved. The modifications took three forms: target dates for phase completion, more time for task completion, and more contextualized training.

First, students were given target dates for each of the phases of research so that they would have sufficient time to complete later phases of research. This instructional support was intended to lessen cognitive overload during the phase of background research. Second, the teachers eliminated activities that did not seem to benefit the students. That is, students selected a pathway from a list of abstracts prior to beginning the study, and the tutorial was eliminated. These two changes resulted in approximately 17% more time for completion of academic tasks (3.5 instructional days out of 20). Third, training was given in the context of use, which was intended to lessen cognitive overload incurred by needing to remember various software procedures that would be used in later phases of research. Instead, the training was done at the beginning of each phase of research by the teacher demonstrating software features that would be needed.

The average task completion rate for this study was 85% [See Figure 3]. This value indicates that over the four-week period, students completed twice as many activities as the students in the first study even though these students were a year younger. The task completion rate within each phase of research was as follows, background research - 78%, data collection - 100%, data analysis - 83%, data interpretation - 62%, and presentation - 100%. The students in this study had more opportunity to achieve the desired learning outcomes than students in the previous study. Through an objective measure such as the task completion rate, it is possible to begin to investigate the factors that contribute toward students' abilities to complete the assigned academic tasks.

Figure 3: Comparison of task completion rate across Studies 1 and 2

Page 285

At the end of the second study, researchers at COTF had the opportunity to review the videotapes, logbooks and field notes from both of the studies. Researchers used this analysis to begin to develop indicators of student learning. This review identified two important issues related to students engaging in scientific inquiry. First, within the activity summary, students were capable of developing brief descriptions of the tasks. However, they were not developing good statements of how the activity related to the overall research path. Second, there was no evidence that students were activating prior knowledge about the research question in order to build connections to these new experiences. The emphasis of the third study was on helping students build connections between prior knowledge, academic tasks and the research question.

5.3 Study 3

In Study 3 (19 days in March-April 1997), the researchers and teachers initiated several major modifications. The purpose of all modifications was to provide more structure to support the activation of prior knowledge and provide opportunities for students to complete multiple investigations in four weeks. Rather than have students select the paths to work on, the teacher and first author selected the pathways that students would work on, and all student teams investigated the same pathways.

In order to accomplish the goal of completing investigations more quickly, research phases were revised and truncated into five alternative phases: the motivating question phase, background research, background review, data analysis, and reflection. In the motivating question phase, the teacher posed the main investigation question and the students individually typed responses in their electronic notebooks. Next, the teacher showed the students the data that they would be analyzing and asked them to record observations. These two activities were meant to activate students' prior knowledge and connect it to the activities of the investigation. This approach is similar to Minstrell's benchmark lessons [Bruer 1993]. In the background research phase, the teacher selected the three most relevant articles from the pathways, and students each read one of the three articles and developed an activity summary. In the background review phase, students used their activity summaries to answer questions as a team that would prompt students to integrate across the readings that were done individually. Since each student was an expert on only one of the articles, students would be forced to discuss the readings with each other in order to answer the question. In the data analysis phase, students completed analysis worksheets as a team. And finally, in the reflection phase, students responded to integrative, teacher-posed questions in their notebooks. Using the redesigned pathways, it was possible to fit three investigations paths into the four-week period.

The average task completion rate for Study 3 was 74%. This rate is comparable to the task completion rate from Study 2, even though this group consisted of students who were at-risk. The task completion rate for each phase was as follows, motivating question - 82%, background research - 85%, background review - 62%, reflection - 73% [see Figure 4]. The instruction for this study involved more explicit prompts for students to reflect on their background knowledge and relate that knowledge to the

Page 286

tasks in the research path. These expanded prompts increased the likelihood that students would assimilate the new information from Astronomy Village into existing knowledge structures.

Figure 4: Comparison of task completion rate across all three studies

5.4 Design Changes to Astronomy Village

This design experiment resulted in planned changes to the Astronomy Village interface in subsequent revisions to the software. Currently, the logbook pages are unstructured. When a student requests a new page, they are given a blank textbox. This research has indicated that it is difficult for students to relate the content in Astronomy Village to their background knowledge. It is also difficult for students to use the logbook as a progress monitoring device. Through the use of writing prompts and questions, the teachers were able to increase students' ability to plan their project as well as reflect on how the content relates to their own experiences. The design experiment context has provided hypotheses for the types of prompts that might work the best. These hypotheses will be tested in more detail through laboratory and quasi-experimental research before being introduced into the software.

Another mechanism that was introduced to increase students' connections between their background knowledge and the content in Astronomy Village was the motivating question phase. This new phase allowed students the opportunity to explore the basic question of the path and come to understand why it was a problem. The success of the motivating question phase has resulted in the planned addition of an exploration phase to the research path diagram.

Finally, it was determined that a single tutorial at the beginning of the project was not sufficient for guiding students on how to use the Astronomy Village interface. In Studies 2 and 3, the teachers provided modeling and coaching on how to use the interface features at the beginning of each phase of research. This assistance made the

Page 287

instruction more immediate to when students were going to be using that knowledge, and it required students to hold less information in memory at the same time. In subsequent versions, the tutorials will be distributed within each of the Village buildings so that students will learn about the interface for that building upon first entering the room.

6 Conclusion

6.1 Summary

The main findings of this study indicate that students need structure in processing individual activities in the context of a multi-week curriculum centered around researching a singular scientific question. By providing this support, along with deadlines for task completion, students were able to finish more and understand more. In the third study, by scaffolding conceptual development through building links to prior knowledge, the teacher was able to maintain task completion rate while altering the level of task demand and cover more investigations with a more difficult population. Based on these findings, it is recommended that future implementations of this curriculum use the model of building a structure for scaffolding and fading (like activity summaries), activating prior knowledge (like asking motivating questions) and directing coaching of procedural knowledge, as needed for students to complete investigations.

6.2 Benefits of the Design Experiment Method

The results and conclusions of the design experiment would not have been possible using the horse race evaluation method. Through the horse race evaluation method, reformers can conclude whether use of an educational multimedia program will enhance student learning in particular contexts. However, every classroom presents a unique context that may or may not match the evaluatory contexts. Teachers must decide, therefore, how to adapt the multimedia for their classroom situation. The design experiment techniques being developed at the classroom of the future offer promise for providing teachers with the tools for evaluating their instruction.

The design experiment allowed the curriculum as it was intended to be studied so specific recommendations could be made for subsequent software revisions. In addition, the curriculum as it was implemented was examined, and teacher adaptations, such as the need to provide a notebook template, were evaluated in vivo. Further, the design experiment allowed for examination of the curriculum as measured. New measures such as task completion rate and judgments of task demand were explored and found useful. Through the research techniques demonstrated here, many limitations inherent in the horse race method were overcome.

Academic task research provides a uniform structure for describing the activities in which students engage in the classroom. Using the structure, it is possible to identify

Page 288

the goals that are assigned by the teacher and the extent to which the students were able to accomplish the goal. This is the task completion rate. This construct can be used to identify activities that are difficult for students to implement. The teacher can then design instructional supports that will help students to be able to accomplish the tasks.

Reform will only take place when teachers are able to explore the implications of a new curriculum within the context of their own classrooms. The fact that an educational multimedia program was used successfully in a neighboring school district does not mean that it will be effective in a new teacher's classroom. Design experiments are especially powerful if attempts are made based on design principles. Creating partnerships between teachers and researchers is useful for enhancing reform by giving teachers tools for evaluating their own implementations and providing researchers a means to compare implementations across different contexts.

6.3 Future Directions

When designing instruction, teachers need to balance the level of cognitive challenge of an academic task with students' abilities to complete the task. The most cognitively demanding tasks are those that are most difficult for students to complete. With the task completion rate indicator, teachers can monitor the extent to which students are capable of completing tasks and they can identify those parts of an academic task that students are having the most difficulty with. In the case of Astronomy Village, the task completion rate indicator served to alert the teacher and researcher to the fact that students were not engaging in efficient planning of their solution, preventing them from completing academic tasks that came later in the project cycle. Having been alerted to the problem, the teacher made adjustments in Study 2 that increased the task completion rate, without negatively affecting the cognitive demand of the task. Future research at COTF will focus on developing efficient indicators for monitoring cognitive task demand.

Just as the results of each individual study resulted in new hypotheses and promoted instructional redesign, the results of the design experiment resulted in working hypotheses for the next experiment. The results of this study will be used in further curriculum evaluations conducted by COTF to design models of appropriate task demands and to develop assessment instruments to examine the curriculum as intended and implemented.

Acknowledgments

This research was supported in part by grants from the National Aeronautics and Space Administration (NCCW-0012) and by the National Science Foundation (ESI-9617857). We would like to thank the teachers and students who participated in the design experiment. We would like to thank Brigitte Gegg for her assistance in conducting the studies and processing the data. We would like to thank John Hornyak and Steven Croft for answering our endless questions about astronomy content. We

Page 289

would like to thank all of our Wheeling Jesuit University student interns who helped in analyzing the data. We would like to thank Dorothy Frew and Pat Carlson for their helpful comments on earlier drafts of the paper.

References

[Blank and Gruebel 1995] Blank, R. K., and Gruebel, D.: "State Indicators of Science and Mathematics Education 1995"; Washington, DC: Council of Chief State School Officers. (1995).

[Brown 1992] Brown, A. L.: "Design experiments: Theoretical and methodological challenges in creating complex interventions in classroom settings"; The Journal of the Learning Sciences, 2, 2, (1992), 141-178.

[Bredderman 1983] Bredderman, T.; "Effects of activity-based elementary science on student outcomes: A quantitative synthesis"; Review of Educational Research, 53, 4, (1983), 499-518.

[Bruer 1993] Bruer, J. T.: "Schools for thought: A science of learning in the classroom"; Cambridge, MA: The MIT Press (1993).

[Bybee and DeBoer 1994] Bybee, R. W., and DeBoer, G. E.: "Research goals for the science curriculum"; In D. L. Gabel [Ed.], Handbook of Research on Science Teaching; New York: Macmillan Publishing Company (1994)

[Collings 1923] Collings, E.: "An experiment with a project curriculum"; New York: The Macmillan Company (1923).

[Collins 1990] Collins, A.: "Toward a design science of education" [Technical Report #1]; Cambridge, MA: Bolt Beranek and Newman (1990, January).

[Cook and Campbell 1979] Cook, T. D. and Campbell, D. T.: "Quasi-experimentation: Design and analysis issues for field settings"; Boston: Houghton Mifflin Company (1979).

[Crocker and Algina 1986] Crocker, L. and Algina, J.: "Introduction to classical and modern test theory"; Orlando: Harcourt Brace Jovanovich (1986).

[Cuban 1993] Cuban, L.: "The lure of curricular reform and its pitiful history"; Phi Delta Kappan, (1993), October, 182-185.

[Dewey 1916] Dewey, J.: "Democracy and education"; New York: Free Press (1916).

[Doyle 1983] Doyle, W.: "Academic work"; Review of Educational Research, 53, 2, (1983), 159-199.

[Doyle and Carter 1984] Doyle, W., and Carter, K.: "Academic tasks in the classrooms"; Curriculum Inquiry, 14, 3, (1984), 129-149.

[Edelson and O'Neill 1994] Edelson, D. C., and O'Neill, D.K.: "The CoVis Collaboratory Notebook: Supporting collaborative scientific inquiry"; In A. Best [Ed.], Proceedings of the 1994 National Educational Computing Conference (pp. 146-152). Eugene, OR: International Society for Technology in Education in cooperation with the National Education Computing Association, (1994).

Page 290

[Hiebert and Wearne 1993] Hiebert, J., and Wearne, D.: "Instructional tasks, classroom discourse, and students' learning in second-grade arithmetic"; American Educational Research Journal, 30, 2, (1993), 393-425.

[Linn and Muilenburg 1996] Linn, M., and Muilenburg, L.: "Creating lifelong science learners: What models form a firm foundation?"; Educational Researcher, 25, 5, (1996), 18-24.

[McGee 1996] McGee, S.: "Designing curricula based on science communities of practice", Dissertation, Northwestern University (1996).

[McGee, Howard, and Hong 1998] McGee, S., Howard, B., and Hong, N.: "Evolution of academic tasks in a design experiment of scientific inquiry"; Paper to be presented at the American Educational Research Association. San Diego, CA (1998).

[National Council of Teachers of Mathematics 1989] National Council of Teachers of Mathematics: "Curriculum and evaluation standards for school mathematics"; Reston, VA: National Council of Teachers of Mathematics (1989).

[National Research Council 1996] National Research Council: "National science education standards"; Washington, DC: National Academy Press (1996).

[Polman 1997] Polman, J.: "Guided science expeditions: The design of a learning environment for project-based science"; Dissertation, Northwestern University (1997).

[Pompea and Blurton 1995] Pompea, S. M., and Blurton, C.: "A walk through the Astronomy Village"; Mercury, (1995 Jan-Feb) 32-33.

[Schmidt et al. 1996] Schmidt, W. H., Jorde, D., Cogan, L. S., Barrier, E., Gonzalo, I., Moser, U., Shimizu, K., Sawada, T., Valverde, G. A., McKnight, C., Prawat, R. S., Wiley, D. E., Raizen, S. A., Britton, E. D., and Wolfe, R. G.: "Characterizing pedagogical flow"; Dordrecht: Kluwer Academic Publishers (1996).

[Shymansky, Kyle, and Alport 1983] Shymansky, J. A., Kyle, W. C., and Alport, J.: "The effects of new science curricula on student performance"; Journal of Research in Science Teaching, 20, 5, (1983), 387-404.

[Stein, Grover, and Henningsen 1996] Stein, M. K., Grover, B. W., and Henningsen, M.: "Building student capacity for mathematical thinking and reasoning: An analysis of mathematical tasks used in reform classrooms"; American Educational Research Journal, 33, 2, (1996), 455-488.

[Stein and Lane in press] Stein, M. K., and Lane, S.: "Instructional tasks and the development of student capacity to think and reason: An analysis of the relationship between teaching and learning in a reform mathematics project"; Educational Research and Evaluation, in press.

[Sweller and Chandler 1994] Sweller, J. and Chandler, P.: "Why Some Material is Difficult to Learn"; Cognition and Instruction, 12, 3, (1994), 185-233.

[Tanner 1997] Tanner, L. N.: "Dewey's Laboratory School: Lessons for Today"; New York: Teachers College Press, (1997).

[Technology and Learning 1996] Technology and Learning, 17, 3, 1996.

Page 291