Go home now Header Background Image
Search
Submission Procedure
share: |
 
Follow us
 
 
 
 
Volume 4 / Issue 4 / Abstract

available in:   PDF (185 kB) PS (281 kB)
 
get:  
Similar Docs BibTeX   Write a comment
  
get:  
Links into Future
 
DOI:   10.3217/jucs-004-04-0429

A Study of User Model Based Link Annotation in Educational Hypermedia

Peter Brusilovsky
(Human-Computer Interaction Institute, Carnegie Mellon University, Pittsburgh PA 15213, USA
plb@cs.cmu.edu)

John Eklund
(Learning Systems Research and Development Group, Faculty of Education, The University of Technology, Sydney NSW Australia
j.eklund@uts.edu.au)

Abstract: Adaptive link annotation is a new direction within the field of user-model based interfaces. It is a specific technique in Adaptive Navigation Support (ANS) whose aim is to help users find an appropriate path in a learning and information space by adapting link presentation to the goals, knowledge, and other characteristics of an individual user. More specifically, ANS has been implemented on the WWW in the InterBook system as link annotation indicating several states such as visited, ready to be learned, or not ready to be learned. These states represent an expert's suggested path for an individual user through a learning space according to both a history-based (tracking where the user has been), and a pre-requisite based (indexing of content as a set of domain model concepts) annotation. This particular process has been more fully described elsewhere [Brusilovsky, Eklund & Schwarz 1998].

This paper details results from an investigation to determine the effectiveness of user-model based link annotation, in a real-world teaching and learning context, on learning outcomes for a group of twenty-five second year education students in their study of databases and spreadsheets. Using sections of a textbook on ClarisWorks databases and spreadsheets, which had been authored into the InterBook program, students received sections of the text both with and without the adaptive link annotation. Through the use of audit trails, questionnaires and test results, we show that while this particular form of ANS implemented in InterBook initially had a negative effect on learning of the group, it appears to have been beneficial to the learning of those particular students who tended to accept the navigation advice, particularly initially when they were unfamiliar with a complex interface. We also show that ANS provided learners with the confidence to adopt less sequential paths through the learning space. Considering ANS tools comprised a minimal part of the interface in the experiment, we show that they functioned reliably well. Discussion and suggestions for further research are provided.

Keywords: Hypertext, Adaptive Hypermedia, Navigation support, Evaluation, Navigation, User Model, WWW

Page 429

1 Introduction

Adaptivity is one of the ways of increasing the functionality of hypermedia. Adaptive hypermedia systems build a model of the goals, preferences, and knowledge of the individual user and use this throughout the interaction for accommodating the individual needs of the particular user. Adaptive hypermedia can be useful in any situation when the system is expected to be used by people with different goals and knowledge, where the hyperspace is reasonably big, or where the system can successfully guide the user in his or her work [Brusilovsky 1996]. Education is one of the most promising application areas for adaptive hypermedia, as it can be applied to adapt the presented information to the current knowledge level of the student, to provide navigation support, and to guide the student in the learning process without being too prescriptive and directive.

There are two general methods of implementing adaptation in adaptive hypermedia: adaptive presentation (or content-level adaptation) and adaptive navigation support (or link-level adaptation). In adaptive presentation the content of a hypermedia page is generated or assembled from pieces according to the user's background and knowledge state. Generally, qualified users receive more detailed and deep information, while novices receive more additional explanation. By adaptive navigation support (ANS) we mean all the methods of altering visible links to support hyperspace navigation.

Adaptive annotation of links is a promising technique for ANS ineducational hypermedia. This technique was suggested in [Brusilovsky, Pesin & Zyryanov 1993; de La Passardiere & Dufresne 1992]. The idea of adaptive annotation technology is to augment the links with some form of comments which can tell the user more about the current state of the nodes behind the annotated links. These annotations can be provided in textual form [Zhao, O'Shea & Fung 1993] or in the form of visual cues using, for example, different icons [Brusilovsky, Schwarz & Weber 1996; de La Passardiere & Dufresne 1992], colours [Brusilovsky & Pesin 1994], font sizes [Hohl, Bscker & GunzenhSuser 1996], or font types [Brusilovsky, Schwarz & Weber 1996]. Annotation seems to be a very relevant way of adaptive navigation support in educational hypermedia. Annotation can be naturally used with all possible forms of links in hypertext and hypermedia. This technique supports stable order of links and avoids problems with incorrect mental maps.

Our position is that adaptive navigation support can be successfully applied in educational hypermedia in a real world teaching and learning context and that adaptive annotation is a relevant technique for that purpose. However at present there are very few instructional systems with adaptive navigation support and there are very few experimental studies which can test how useful adaptive navigation support can be for educational application. [Weber & Specht 1997] used the ELM-ART [Brusilovsky, Schwarz & Weber 1996] system to count the number of navigation steps for those with and without ANS and found no significant difference with a relatively small (n=16) group of novice learners. [Brusilovsky & Pesin 1998] investigated ISIS-Tutor system [Brusilovsky & Pesin 1994] and reported a significant decrease of the number of navigation steps and the number of repeated visits to the same node for a group with ANS. We focus instead on the impact of ANS on learning outcomes and user paths. In the following sections we describe the InterBook system

Page 430

which demonstrates a particular implementation of adaptive navigation support, and we report the results of an empirical study using InterBook.

2 Adaptive Navigation Supportin InterBook

InterBook is a system for authoring and delivering adaptive electronic textbooks on the WWW. Electronic textbooks reside on an InterBook server and can be accessed with any frame-enabled Web browser. The InterBook interface [Fig. 1] divides the screen into four sections, the largest window being the textbook window in which the content in the form of text, hypertext and graphics appears.On the top-right is the toolbar, in which the links to a table of content,a glossary, and a search interface, and a help button appear. The window at the top left is called the navigation bar, and this provides the learner with a navigable hierarchy of surrounding nodes. The window at the bottom right is called the concept bar, and this lists the pre-requisite and outcome concepts for the section presented in the textbook window.

Figure 1: Text window of InterBook with adaptive link annotation. Green bullet means recommended, red bullet means" not ready to be learned", white bullet means "nothing new", while a checked bullet means "visited".

Page 431

The key to adaptivity in InterBook is what we call "knowledge behind pages". InterBook uses a structured domain model represented as a network of domain concepts. Domain concepts are important terms of the domain. They designate atomic pieces of knowledge about the domain. A special part of an electronic textbook, the Glossary provides some descriptions of domain model concepts.A description of each concept is individually accessible as a glossary entry [Fig. 2].

Figure 2: Glossary window of InterBook showing a glossary entry for the concept "Database operating mode"

All sections of an electronic textbook are indexed with domain model concepts. For each section, a list of concepts related with this section is provided (this list is

Page 432

called the spectrum of the section). The spectrum can also represent the role of a concept in the section (either an outcome concept or a background concept). A concept is included in the spectrum as an outcome concept if some part of this section presents the piece of knowledge designated by the concept. A concept is included in the spectrum as a prerequisite concept if a student has to know this concept to understand the content of the section.

The knowledge about the domain and about the textbook content is used by InterBook to serve a well-structured hyperspace. In particular, InterBook generates links between the glossary and the textbook. Links are provided from each textbook section to corresponding glossary entries for each involved background or outcome concept. Similarly, for each glossary entry describing a concept InterBook provides links to all textbook units that can be used to learn this concept. This means that an InterBook glossary integrates features of an index and a glossary.

InterBook uses coloured bullets and different fonts to provide adaptive navigation support [Fig. 1]. Wherever a link appears on InterBook pages (in the table of contents, in the glossary or on a regular page), its font and the colour of its bullet will inform the user about the status of the node behind that link. Green bullet and bold font means "ready and recommended", i.e., the node is ready-to-be-learned but still not learned and contains some new material. A red bullet and an italic font warn about a not-ready-to-be-learned node. A white bullet means "clear, nothing new" (i.e., all concepts presented on a node are known to the user). A check mark is added for already visited nodes. InterBook integrates all three methods of annotation: history-based (on the basis of where the user has been), prerequisite-based (on the basis of what prerequisite nodes the user has visited, and knowledge-based (on the basis of the user's demonstrated knowledge).

The user model in InterBook represents levels of user's knowledge of every domain concept. It is initialized from the registration page via a stereotype model, and is modified as the user moves through the information space. The user model for each user is stored in a file on the server.

3 Experimental Overview

In a study involving 25 undergraduate teacher education students in an educational computing elective at the University of Technology, Sydney, students were exposed to two chapters of a textbook [Rubin 1996] about ClarisWorks databases and spreadsheets, and used the InterBook system both with [Fig. 1] and without [Fig. 3] adaptive link annotation. The experiment was created to be in a real-world teaching and learning context, with the use of InterBook as an integral part of a university subject as described inthe previous section.

Page 433

Figure 3: Text window of the non-adaptive version of InterBook used for the experiment. Adaptively chosen bullets and fonts are replaced by green bullets and regular font.

The goal of this experiment was to assess what impact, if any, user-model based link annotation would have on students' learning and on their paths through the learning space, in this realistic situation. The experiment was aimed to investigate both the effect of link annotation on learning and the effect of link annotation on user paths. The hypothesis was that adaptive link annotation would provide students with a more efficient path through the knowledge space with improved learning outcomes. Tests of knowledge were carried out, audit trails and questionnaires were gathered and the results analyzed.

The experiment took place over a four-week period. In the first two-hour session, students were introduced to InterBook and its features explained to them. They used the system for an hour, and answered a questionnaire about its features. This questionnaire showed that almost all students were familiar with what each of the buttons and annotations meant. They were then free to use the system at any time during the following week.

In the second session, students were randomly divided into two groups of equal size, one group receiving the link annotation, while the other group did not [Tab. 1]. They were allowed access to the chapter of the textbook on databases which had been

Page 434

authored into InterBook and they completed a questionnaire. Students had access to the database chapter for the following week.

Group 1

Group 2

Database chapter WITH adaptive link annotation n= 12 Database chapter WITHOUT adaptive link annotationn=13
Spreadsheet chapter WITHOUT adaptive link annotation n=12 Spreadsheet chapter WITH adaptive link annotation n=13

Table 1: Allocation of adaptive link annotation to groups

In the third session, students took a multiple choice test on the database section of the textbook. They were then allowed access to the spreadsheet section of the textbook in InterBook which they could access for the following week. This time, the group that did have the adaptive link annotation for the database section now did not receive it, and vice-versa for the other group [see Tab. 1]. In the final session, students took a multiple choice test on the spreadsheet section and completed a questionnaire. The audit trails from the sessions were extracted, and analyzed along with the test results and the questionnaire responses. After the second session the students were asked to rate their use of the various features of the interface, apart from the link annotation which some of them had not been receiving. These ratings are shown in [Tab. 2]. The purpose of this was to determine if any feature was not well received by the students.

4 Experimental Results - Interface

A questionnaire was used to assess the functionality of each of the key interface features of InterBook, the results showing that all the features were working as expected, quite uniformly across the group.

Page 435

Question

Level of response

Mean

Standard
Deviation

drth The Multiple windows were 1=useless
5=useful
3.5 1.1
I used the hot links in the Text 1=never
5=often
3.6 1.0
The 'you are/were here feature' in the table of content was 1=useless
5=useful
3.9 1.2
The Navigation was rather 1=hard
5=easy
3.3 1.2
I used the search feature 1=never
5=often
3.4 1.2
I took into account the checked balls 1=never
5=often
4.0 1.1
The search feature was 1=useless
5=useful
3.5 1.1
The local overview in the table of contents was: 1=useless
5=useful
3.2 0.9
The list of related pages in glossary was 1=useless
5=useful
4.1 1.0

Table 2: User's rating of the interface features of InterBook (n=25)

5 Experimental results - Test Scores

5.1 Procedure

The students' test scores were then used as a measure of their learning of the material in each of the sessions. There was a reasonable margin for error in using this variable, as students' prior knowledge of the domain, and their learning of it from other sources such as the actual text held in the library's closed reserve, could not be determined. It was particularly important that students' scores in each test were a reliable measure of their learning, considering each test was rather short and they were not standard experimental instruments. Experimental effects which were not totally accounted for were minimized through two methods. Firstly the tests, consisting of ten and twelve multi-choice questions taken directly from the material in InterBook, were validated on another small group of fifteen students. A Cronbach Alpha for the database test was calculated at 0.58 and for the spreadsheet test an initial Cronbach Alphavalue of 0.23 was obtained, and these unsatisfactory results were improved by modifying each test. Adequate reliability and performance of all test items were established by discarding some of the test questions or individual

Page 436

distracters. In this way, each of the test questions was constructed to be an adequate predictor of how astudent would score in the overall test, as is desirable in a norm-referenced test. At the conclusion of the experiment, alphas of 0.75 and 0.82 for thirty-two students were obtained for the database and spreadsheet tests respectively, and these very acceptable values were interpreted as establishing adequate reliability for the tests to be used in this and any subsequent experiment.

Secondly, other measures of performance were examined for the groups. To avoid the possibility that the randomly chosen groups consisted of a disproportionate number of less able or more able students, means and standard deviations of test results for each group (i.e., the group that was to receive the annotations and the group that was not), taken from other aspects of the course were compared, and these were found to be very similar. In other words the random selection of students provided two groups with very similar academic ability.

5.2 Results

The results of the students' knowledge tests are shown in table 3:

Group

Test Result Database

Test Result Spreadsheet

1 ANS on database only 6.41 7.77
2 ANS on spreadsheet only 7.12 8.10

Table 3: Test results for groups with and without ANS

A two-sample t-Test was performed on the results. The t value of -0.3667 shows that link annotation had a statistically significant negative effect at the p < 0.05 level on the database session (the first session), and no effect on the spreadsheet session (the second session). This unexpected initial result suggested that further investigation was required.

A careful analysis of the audit trails revealed two factors. First, some of the subjects apparently learned about ClarisWorks from other sources since they were able to obtain good test results after hitting only 5 to 15 sections of the electronic textbook (the Databasepart alone contains about 100 sections). Second, for most of the students ANS appears to be a minor factor because about 80% of all navigation steps were made with Continue and Back buttons which were not annotated in the experimental version of InterBook [see Chapter 6 for more details]. In this situation we had touse some more elaborate techniques to find a relationship between ANS and test performance.

To exclude students who learned less from the system than from other sources in the study, two subgroups were introduced, the first based on spending a 'reasonable time' with the system. This consisted of those students who spent a reasonable time using InterBook over both sessions, as it became clear from the audit trails that a number of students relied heavily on either their previous knowledge of the content, or on the printed version of the ClarisWorks book. For both the database and

Page 437

spreadsheet sections, two-sample t-Testsshowed that there was no significant difference at the 0.05 level in the test means for those with ANS and those without ANS.

The second (overlapping) subgroup was intended to eliminate those students who made very few (fewer than 15) hits on the system.

  Mean Std.Dev. Std.Error Count Minimum Maximum #Missing
Score, Total 7.9 1.2 .3 17 6.0 10.0 0
Score, yes 7.9 1.0 .4 8 7.0 9.0 0
Score, no 8.0 1.5 .5 9 6.0 10.0 0

Table 4: Database Subgroup which excludes those with fewer than 15 hits on the system

Table 4 shows that of the 17 students who made greater than 15 hits in the database section, 8 students received ANS and 9 students did not. There is no statistically significant difference in the test performance of those that did and those that did not receive ANS.

This result is natural taking into account that the average number of navigation steps (with annotated links) made by the "adaptive" group was too small to affect their performance. However, users appeared to be very different in their navigation behavior. Some of them almost never used annotated links, some of them used it reasonably often. We decided to investigate the performance of users who did use annotated links.

5.3 The Value of ANS for Those Who Use It

Separate audit trails for each of the two time periods were generated, to examine how users navigated through InterBook with and without ANS. For each user these trails showed the number of times they selected a link with a green ball and also a red ball, as well as their use of all the other features of the interface. Certain unexpected behaviour was immediately apparent for a small group of students, who were purposefully and continually selecting nodes which were not recommended. More generally, it was noted that just because link annotation was evident in the interface for one group, individual students within that group were accepting it to varying extents. Just because a student was offered link annotation does not mean that they were accepting or making profitable use of it. A measure of the students' acceptance of navigational advice was calculated from the audit trails, taken as the number of green ball hitsminus the number of red ball hits divided by the total hits. This measure of acceptance of this particular interface functionality is a more important variable than the fact that they were provided with it.

agreement rate = (Ngreenballs - Nredballs)/(Ngreenballs + Nredballs + Nwhiteballs)

Where Ngreenballs, Nredballs, and Nwhiteballs is the number of times the user hits a link which should be annotated with green, red, or white bullet. Note that for the

Page 438

students of a group with ANS these links were really annotated with bullets of different colours, while the students of anon-ANS group were shown the same green bullet regardless of the state.

For students who always follow green balls the agreement rate is 1, for those who always follow red balls it is -1. Then four distinct groups for each of with and without ANS were established depending on this agreement rate:

High-positive rate > 0.5
Low positive 0 < rate <=0.5
Low negative -0.5 < rate <= 0
High-negative rate <= -0.5

(a)

(a)

(b)

Figure 4: Clustering acceptance of link annotation and database scores for the group with (a) and without (b) ANS. For the ANS group a better agreement rate results in generally better test results. Columns and keyshave the same order, i.e., the top key corresponds to the most left column and the bottom key corresponds to the most right column.

Page 439

A clear correlation (R=0.670) was found between the agreement rate and score in the database tests: the more students agree with system's suggestion, the better is the score - for the group receiving link annotation [Fig. 4a]. Moreover, in the group with link annotation there were no high-negative students at all. For the non-ANS group [Fig. 4b] there is a mild negative correlation (R=-0.383) and one high-negative student. This is natural because they have not seen the annotations, but it is also an argument for ANS - without annotation the students cannot recognize and use the state of the page.

The above calculations were repeated with a modified formula, namely:

agreement rate = (Ngreenballs - Nredballs)/All-Hits

This formula may provide a more reliable measure of agreement for the users who almost never hit annotated links (i.e., when Ngreenballs + Nredballs + Nwhiteballs is very small). A similar correlation of 0.618 was obtained for the group with the ANS and -0.176 for those without ANS.

This positive correlation in the first session on databases suggests that while link annotation is a distracting complication to an interface, it is helpful to those that choose to follow it in terms of improving their knowledge of the content.

6 Student's Use of Individual Navigation Tools

One of the major problems in determining if ANS was effective or not was that approximately 80% of the available navigation tools that were used in the experimental version of InterBook were non-adaptable. An analysis of the proportion of use of different navigation tools [Tab. 6] shows that the non-annotated Continue button (pCONTINUE) is used more than all other navigation tools combined. The bullet-annotated pCONTENT, pINTRODUCING, pREQUIRING, pSEARCH, pREGISTER,pPATH, pHELP [see Tab. 5] are used on less than 20% of hits (for those users who received annotation). Additionally the checkmark-annotated pPREREC and pOUTCOME are almost not used at all with only 1-2% of hits. This implies that the majority of navigation choices made by students were made without annotations.

Page 440

  Mean Std. Dev. Std. Error Count Minimum Maximum #Missing
pCONTINUE 15.4 8.6 2.1 17 2.0 30.0 0
pCONTENT 3.9 3.8 .9 17 1.0 16.0 0
pTEXT 2.9 3.1 .8 17 0.0 11.0 0
pBOOKSEL 1.5 .7 .2 17 1.0 3.0 0
pINTRODUCING 1.1 1.4 .3 17 0.0 4.0 0
pBOOKTITLE .8 .8 .2 17 0.0 2.0 0
pTOOLBAR .4 .9 .2 17 0.0 3.0 0
pOUTCOME .4 1.0 .2 17 0.0 4.0 0
pREQUIRING .3 .7 .2 17 0.0 2.0 0
pHELP .1 .2 .1 17 0.0 1.0 0
pPATH .1 .2 .1 17 0.0 1.0 0
pPREREQ .1 .2 .1 17 0.0 1.0 0
Not annotated 21.9 7.8 1.9 17 8.0 40.0 0
Balls 5.6 4.4 1.1 17 1.0 19.0 0
Checkmarks .4 1.1 .3 17 0.0 4.0 0

Table 5: The number of times different navigation tools were used [see Tab. 6]

Not annotated in both versions
Annotated in ANS version
- pCONTINUE - using continue button - pCONTENT - link from the separate or embedded table of contents
- pBACK - using back button -pINTRODUCING - link from a glossary page to a page introducing aconcept
- pTEXT - hypertext reference from one page to another - pREQUIRING -link from a glossary page to a page which requires a concept
- pBOOKSEL - link to a book from book list (top of the table of content page) -pHELP - link to one of the helpful pages from background help page
- pBOOKTITLE - using the link to the book title in navigation centre -pPATH - link to an higher level section from the navigation center
- pTOOLBAR - using buttons on theToolbar - pPREREC, pOUTCOME - links from the concept bar to glossary annotated with checkmarks

Table 6: Description of most often used navigation tools

Using sequential navigation (i.e., continue-back) vs. non-sequential navigation is known behaviour exposed by novices in hyperspaces. However, we were able to show that ANS encourages the novices to use annotated non-sequential tools more often. This was achieved using the count of hits on annotated links such as table of contents links versus non-annotated links such as Continue button.

Page 441

Figure 5: ANS and the use of sequential and non-sequential navigation tools (average number of hits per user). The non-shaded "yes" columns are for the ANS group; the shaded "no" columns are for the non-ANS group

[Fig. 5] and [Fig. 6] show that those students who did not receive ANS (the shaded "no" columns) used more of the non-annotatable and sequential navigation features (pCONTINUE, pBACK). At the same time those who did receive ANS (the non-shaded "yes" column) used more of the annotatable navigation features (pCONTENT,pINTRODUCING, pREQUIRING). Even the use of pTEXT which is a non-annotatable but non-sequential navigation tool is slightly smaller for the ANS group. This implies that ANS provides the learner with a non-linear guide through the learning space, and learners are more likely to use non-sequential paths with adaptive link annotation. It again reflects the student's trust in the annotations - ANS provides some security for those users who would like to follow non-linear paths but might be afraid of becoming lost.

Page 442

Figure 6: ANS and the proportion of use of annotatable (left) and sequential (right) navigation tools. The presence of ANS encourages the subjects of the ANS group (white "yes" bars) to use annotatable and non-sequential links more often.

7 The Role of The Page State

If the number of hits on pages of various 'states' [Tab. 7] is examined, it is clear that students prefer to visit ready-to-be-learned pages (s2) than those which are annotated as "no information" (s1). Students spent approximately twice as much time reading ready-to-be-learned pages than reading all other pages (s1 and s3) combined. Thus, the data shows that a green and unchecked page is one that students read most. This is naturally due to the fact that the Continue button was used most of the time, bringing the user into a page with a "ready to be learned" status.

  Mean Std.Dev. Std.Error Count Minimum Maximum #Missing
s1 4.294 2.687 .652 17 1.000 9.000 0
s2 15.882 6.343 1.538 17 6.000 30.000 0
s3 7.706 4.413 1.070 17 1.000 15.000 0
s1t 207.882 178.126 43.202 17 23.000 592.000 0
s2t 1220.765 515.682 125.071 17 451.000 2242.000 0
s3t 749.941 427.570 103.701 17 25.000 1476.000 0
r0 24.176 8.118 1.969 17 15.000 45.000 0
r1 3.706 2.779 .674 17 0.000 9.000 0
r0t 1944.294 518.402 125.731 17 1108.000 2935.000 0
r1t 234,824 207.745 50.386 17 0.000 748.000 0

Table 7: Number of hits on pages of various states [see Tab. 8]

Page 443

Time

Hits

s1t = time on ready but not suggested (while ball) s1 = hits on pages ready but not suggested (while ball)
s2t = time on ready and suggested (green ball) s2 = hits on pages ready and suggested (green ball)
s3t = time on not ready (red ball) s3 = hits on pages not ready (red ball)
r1t = time on pages that have been read before (checkmark over ball) r0 = hits on pages that have not been read before (no checkmark overball)
r0t = time on pages that have not been read before (no checkmark over ball) r1 = hits on pages that have been read before (checkmark over ball)

Table 8: Key to states of pages

Moreover, it can be seen from [Fig. 7] showing the average time students spent on different types of pages, that "nothing new", "not ready" and "ready" pages are very different. The average time spent on a not-ready page is much larger than the time for a ready page, which is close to the average time per hit. Also, the average time to spent on a "nothing new"page is much less than average.

Figure 7: Average time per page for the case when the student navigate to a page using a non-annotated link .

What we observe on the [Fig. 7] isa "real value" of page state. Here students navigate to a selected page with a non-annotated link and without any warning about a page state. This data shows that the mechanism which determines different classes of pages works quite well. A page classified as "nothing new" can be read much faster (or just passed over) because it has no new information and a page classified as "not ready" is the most hard to understand because some background can be missed.

Page 444

The data are quite different for the minority of pages (about 20% of hits) selected with annotatable link [Fig. 8]. What we observe on [Fig. 8] is a mixture of two effects: an effect of page state and an effect of annotation. This means that students of the ANS group noticed the annotations, may decide apriori to spend less time on "nothing new" pages and more on those annotated as "not ready". The effect of annotation clearly dominates in the case of "not-ready" pages. Those rare users who selected a page with full understanding that this page is not ready are willing to allocate significantly more time (ANOVA, p=0.012) for reading this page. Again, they understood how the system worked and trusted the integrity of the annotations.

Figure 8: Average time per page for the case when the student navigate to a page using an annotatable link. Thenon-shaded "yes" bars are for the ANS group; the shaded "no" bars are for the non-ANS group

8 Discussion and Conclusion

These results suggest that ANS is a feature which is initially useful in improving comprehension for those new to a complex interface who are prepared to accept it. However, it adds another option to an interface: a cognitive overhead which may distract users from the content. This was reflected in the fact that the overall group who received ANS initially performed significantly worse in the knowledge tests. User model based link annotation seems to be of value to those that agree with it, those that accept and follow the annotations. Those learners who follow the annotated links are essentially in agreement with the cognitive model of the knowledge that the author, as expert, has placed on the content [Eklund 1995]. They take advantage of the fact that the content has been examined and structured for them, and they make use of both the implicit structure of the knowledge that the courseware embodies in its static form, as well as the individual link annotations which hint at the domain structure relative to the current path of the learner through the user model.

The experiment offers firm evidence that adaptive link annotation has an effect on student learning in an educational system. However, once users with minimal hits on the system were excluded, numbers in each of the two groups with and without annotations was a mere 8 and 9 respectively, and this is one of the severest limitations

Page 445

of the experiment. This study also suggests that the existence of ANS in the interface has an effect on the linearity of a user's path through the learning space, with those users experiencing link annotation being prepared to use those annotatable links more often than those who received no link annotation. As a result, their paths through the material were less linear, more exploratory, as they selected more "real links" and exhibited less use of the Continue button. In some ways this is hardly surprising, and again reflects the learner's trust of the system's annotations. If a link is annotated, a user has more confidence about the relevance of the material behind it than under a non-annotated link. In a non-annotated InterBook interface, the safest option for users was the repeated use of the Continue button.

A difficulty with the experimental procedure that was briefly described is that the majority of the navigation features were not annotated, so the difference in the interface for those students who did receive ANS was marginal. Another factor to be considered was the indexing of content and the subsequent authoring of the electronic textbook. This involves the allocation of a set of attributes for each domain node as both pre-requisite and outcome concepts, as described earlier. This indexing implies an ideal order in which to view content, represented by the continue button, as well as optional ways to view it, either recommended by green annotations or not recommended by those nodes annotated with red bullets. In the non-hypertext world, it is easy to flip the pages of a book in order, pausing only for those pages of interest. That way a reader may be sure that all the content has been examined, even if briefly. Similarly, it appears that one popular strategy was to follow the non-annotated continue link, and immediately move forward to the next node if the material on the current node is already known or of little interest. Students clearly felt that the continue link was the simplest and quickest way to review the material, thus following the domain structure imposed by both the textbook and the authoring process. This is a favourable outcome for adaptive curriculum sequencing, but not necessarily for adaptive link annotation.

More generally, it can also be argued that the lack of detail in the artificial world of the user-model somewhat trivialises the broad range of human responses and motives possible in learning. Since the start of the 1990s, this has been identified as an "intractable problem" [Self 1990], although more recent approaches give the learner access to this information as a goal-planning aid and learning-reflective device [Kay1997]. Is it possible to make a reasonable suggestion of the next best link for a user to follow with such a paucity of information about that individual? This revisits the problem of building practical intelligent tutors as discussed in the AI literature since the early 1980s. Even in well defined, highly organised and simplified domains, implementing a reliable system to account for user preferences, knowledge and idiosyncratic behaviour is highly problematic. In terms of the ramifications for informing the design of this experiment, to obtain a measurable favourable result for adaptive link annotation which confirms some of the interpretations of the study as presented above will require a simplification of the domain and a greater control of the variables earlier described. This may be difficult to achieve with an experiment situated in a real world teaching and learning context.

Page 446

Acknowledgments

Many thanks to Elmar Schwarz of Tellux, Germany, for his advice about this research and for making available InterBook as an experimental vehicle, as well as Associate Professor Ken Sinclair and Dr. Mike Bailey of the Faculty of Education, University of Sydney, for their advice in the design and implementation of components of the empirical study discussed in this paper.

References

[Brusilovsky 1996] Brusilovsky, P.:"Methods and techniques of adaptive hypermedia"; User Modeling and User-Adapted Interaction, 6, 2-3 (1996) 87-129.

[Brusilovsky, Eklund & Schwarz 1998] Brusilovsky, P., Eklund, J., &Schwarz, E.: "Web-based education for all: A tool for developing adaptivecourseware"; Computer Networks and ISDN Systems, 30, 1-7 (1998),291-300.

[Brusilovsky & Pesin 1994] Brusilovsky, P., & Pesin, L.: "An intelligent learning environment for CDS/ISIS users"; Proc. The interdisciplinary workshop on complex learningin computer environments (CLCE94), EIC, Joensuu, Finland (1994) 29-33, also apeared as electronic version, http://cs.joensuu.fi/~mtuki/www_clce.270296/Brusilov.html.

[Brusilovsky & Pesin 1998] Brusilovsky, P., & Pesin, L.: "Adaptive navigation support in educational hypermedia: An evaluation of the ISIS-Tutor"; Journal of Computing and Information Technology, (1998).

[Brusilovsky, Pesin & Zyryanov 1993] Brusilovsky, P., Pesin, L., &Zyryanov, M.: " Towards an adaptive hypermedia component for an intelligent learning environment"; In Bass, L.J., Gornostaev, J., & Unger, C. (Eds.),Human-Computer Interaction, Springer-Verlag, Berlin (1993) 348-358.

[Brusilovsky, Schwarz & Weber 1996] Brusilovsky, P., Schwarz, E., & Weber, G.: "ELM-ART: An intelligent tutoring system on World Wide Web"; InFrasson, C., Gauthier, G., & Lesgold, A. (Eds.), Intelligent Tutoring Systems, Springer Verlag, Berlin (1996) 261-269.

[de La Passardiere & Dufresne 1992] de La Passardiere, B., & Dufresne, A.: "Adaptive navigational tools for educational hypermedia"; In Tomek, I.(Ed.) Computer Assisted Learning, Springer-Verlag, Berlin (1992)555-567.

[Eklund 1995] Eklund, J.: "Cognitive models for structuring hypermedia and implications for learning from the world-wide web"; Proc. Ausweb95: The First Australian World-Wide Web Conference, Southern Cross University Press, Ballina, Australia (1995) 111-116, also apeared as electronic version, http://elmo.scu.edu.au/sponsored/ausweb/ausweb95/papers/hypertext/eklund/index.html.

[Hohl, Bscker & GunzenhSuser 1996] Hohl, H., Bscker, H.-D., & GunzenhSuser, R.: "Hypadapter: An adaptive hypertext system forexploratory learning and programming"; User Modeling and User-Adapted Interaction, 6, 2-3 (1996) 131-156.

[Kay 1997] Kay, J.: "Learner know thyself: Student models to give learner control and responsibility"; Proc. ICCE97, International Conference on Computers in Education, Malasia, Kuching, Sarawak (1997) 18-26.

[Rubin 1996] Rubin, C.: "Macintosh Bible Guide to CarisWorks 4"; Peachpit Press, CA (1996)

[Self 1990] Self, J.A.: "Bypassing the intractable problem of student modelling"; In Frasson, C., & Gauthier, G. (Eds.), Intelligent Tutoring Systems: At the crossroads of artificial intelligence and education, Ablex Publishing, Norwood (1990) 107-123.

Page 447

[Weber & Specht 1997] Weber, G., & Specht, M.: "User modeling and adaptive navigation support in WWW-based tutoring systems"; In Jameson, A., Paris,C., & Tasso, C. (Eds.), User Modeling, Springer-Verlag, Wien (1997) 289-300.

[Zhao, O'Shea & Fung 1993] Zhao, Z., O'Shea, T., & Fung, P.: "Visualization of semantic relations in hypertext systems"; Proc. ED-MEDIA'93, World conference on educational multimedia and hypermedia, AACE, Orlando, FL(1993) 556-564.

Page 448