ITS 2002 WORKSHOP
Individual and Group Modelling Methods that
help Learners Understand Themselves
Chairs:
Paul Brna, Northumbria University, UK
Vania Dimitrova, University of Leeds, UK
Workshop Web Site:
http://www.cbl.leeds.ac.uk/~paul/its2002workshop
http://computing.unn.ac.uk/staff/cgpb4/its2002workshop
Programme Committee:
Fabio Akhras, University of Sao Paulo, Brazil (fabio.akhras@poli.usp.br)
Paul Brna, Northumbria University, UK (Paul.Brna@unn.ac.uk)
Susan Bull, University of Birmingham, UK (s.bull@bham.ac.uk)
Vania Dimitrova, University of Leeds, UK (vania@comp.leeds.ac.uk)
Jim Greer, University of Saskatchewan, Canada (greer@cs.usask.ca)
Judy Kay, University of Sydney, Australia (judy@cs.su.oz.au)
Gord McCalla, University of Saskatchewan, Canada (mccalla@cs.usask.ca)
Tanja Mitrovic, University of Canterbury, New Zealand (tanja@cosc.canterbury.ac.nz)
Rafael Morales Gamboa, Instituto di Investigaciones Electricas, Mexico (rmorales@iie.org.mx)
Patricia Tedesco, Universidade Federal de Pernambuco, Brazil
(ptedesco@elogica.com.br)
Julita Vassileva, University of Saskatchewan, Canada (jiv@cs.usask.ca)
Preface
Adaptive systems have a great deal of potential to encourage learners to engage in a variety of reflective
activities.
Developments in open learner modelling allow for the possibility of the system reflecting back the
contents of the learner model to the learner.
Developments in dialogue management allows for the system to
take the learner’s expressed intentions into account through the use of intentional interfaces.
Developments in
analysing interactions with the system or with other users allows for the possibility of feeding the system’s
interpretation of events back to the learner (or learners) so that they can take note in order to decide what to do.
Developments in the use of agents allows for the learner model play an important role in obtaining assistance.
Developments in affective computing allow for the learner’s affective state to be modelled and that model may
have an increased role in helping learners understand themselves.
Various architectures for individual and group modelling that help learners understand themselves have been
explored.
The common theme is to provide learners with a feedback mechanism which facilitates a variety of
metacognitive activities.
There is an increased recognition of the need to engage the learner in an interactive
process with the possibility of the learner model being revised by either the learner or the system.
This
interactive process depends on formal models of belief revision as well as models of interaction such as those
advocated by Baker, Pilkington, Suthers amongst others.
Additionally the importance of work which seeks to open the learner (or group learner) model is underlined by
the increasing need to respect various stringent data protection laws which become highly relevant once adaptive
Web-based learning environments are deployed.
1
The notion of models of learners that can be examined and possibly revised has received increased attention
since the Workshop on “Open, interactive and other overt approaches to learner modelling” held at the 9
Conference on Artificial Intelligence in Education at Le Mans, France
th
World
in July 1999.
(see http://cbl.leeds.ac.uk/ijaied/abstracts/Vol_10/modelling.html)
The analysis of interactions allows for the detection of patterns which can be presented to learners to encourage
a change in the pattern of interaction to improve learning.
This includes analysing an individual’s interactions
with a computer system and reasoning about the process of learning followed by changing the environment in
ways that create possibilities for processes of interaction that increase learning over time.
More recently, group
interactions have been monitored in order to detect aspects of group behaviour which can be used for provoking
reflection and articulation in group problem solving.
The
purpose
of
this
workshop
is
to
explore
various
issues
concerned
with
the
development
of
novel
computational architectures for individual and group modelling that help learners understand themselves.
Empirical investigations and experience with existing architectures that suggest advantages, indicate potential
problems, and propose further extensions will be encouraged.
Different perspectives on developing learning
environments that build models of learners and support meta-cognition will be discussed.
These may include
educational theories that justify the advantages of such approaches as well as formal methods that support the
application of the architectures in a variety of learning situations.
The papers that have been collected together in this set of proceedings provide a basis for discussion at the
workshop. The topics of these papers span issues in giving learners control in diagnosis, interactive cognitive
modelling, interactive modelling of affect, maintaining distributed student models e.g. to help the learner
examine and use multiple models of themselves, modelling groups of learners to increase group performance
and improve learning, and using student and group modelling methods to promote meta-cognitive skills.
Bull and Nghiem are taking a pragmatic approach to explore the issues in system-learner and learner-learner
interactions based on simple domain independent models of learners. This work discusses the role of inspectable
student models to promote meta-cognition in various classroom situations: the student model is inspectable by
the student and promotes reflection on his/her learning; students can view models of their peers and compare
their own progress against that of others; and finally, instructors may view the learner models of their students
and use this to support the students in their learning.
The paper by Andrade and her colleagues seeks to provide a diagnostic agent that advises other agents about
what to do at a task level or at a group recombination level conditioned by the generalised notion that a group of
learners has a ZPD. The work combines cognition with some affective and motivational elements within an
agent-based approach.
They seek to use learner provided information and allow learners to modify this if
required.
Dufresne and Hudon seek to use humour in interactions with learners and make the learner's preferences about
humour visible and changeable.
In these two contexts, the learner's preferences are considered to be accurate
and learners have ultimate control - unlike the systems developed by Dimitrova and Bull which are both based
on the more generalised notion that the system can argue with the learner on some matters. Dufresne and Hudon
follow a principle that people draw upon their experience in real interactions when participating in computer
mediate interactions They argue that affective tutors (tutors that exploit humour are discussed here) may provide
a more effective learning environment. The paper prompts an interesting research issue of how to employ learner
control in the design of affective interactive tutors.
There
are
many
opportunities
to
encourage
reflection
within
the
context
of
working
with
simulations.
Grigoriadou, Samarakou, Mitropoulos and Panagiotou provide a simulation environment that seeks to diagnose
student misunderstandings and misconceptions and show this diagnosis in some form to the learners.
Chesher,
Kay and King provide a simulation environment encouraging professionals to manage multiple cases and
consider how well they are doing.
These two somewhat different environments (learning elementary physics
and learning within a professional context) both offer excellent test beds for whether seeing the system's
(possibly erroneous) judgement has any important positive effect on learning.
2
Finally, the paper by Dimitrova focuses on the student-computer interaction in inspectable student models. It
seeks to provide intelligent support to develop an extended dialogue with learners where the learner and the
system discuss the content of the learner model and it is through this interaction that reflective learning may take
place. The paper discusses how far a combination of knowledge based systems and an improved model of
dialogue can be used to help the learner argue about, and reflect on, the learner's own beliefs as reflected to them
by the system presenting its model of the learner's beliefs.
Paul Brna and Vania Dimitrova
3
Index
Helping Learners to Understand Themselves with a Learner Model Open to
5
Students, Peers and Instructors
Susan Bull, Theson Nghiem
A Diagnostic Agent based on ZPD approach to improve Group Learning
14
Adja F. de Andrade, Paul Brna, Rosa Maria Vicari
A Web-based Medical Case Simulation for Continuing Professional Education.
26
Douglas Chesher, Judy Kay, Nicholas King
Knowledge-based Fuzzy Evaluation of Learners in Intelligent Educational Systems
32
Maria Grigoriadou, Maria Samarakou, Dionissis Mitropoulos, Michael Panagiotou
Modeling the learner preferences for embodied agents : experimenting with the
43
control of Humor.
Aude Dufresne, Martin Hudon
Interactive Cognitive Modelling Agents - Potential and Challenges
52
Vania Dimitrova
4
Helping Learners to Understand Themselves with a
Learner Model Open to Students, Peers and Instructors
Susan Bull & Theson Nghiem
Electronic, Electrical and Computer Engineering,
University of Birmingham, Edgbaston, Birmingham, B15 2TT, U.K.
s.bull@bham.ac.uk
Abstract: This paper introduces work in progress on an open student model designed to help learners to better
understand their learning. The aim of the system is to investigate issues relevant to open learner models in a
large-scale, real use context. This will be achieved initially through the deployment of a simple, domainindependent system. The student model is inspectable by the student it represents, to help focus reflection on
their learning. Students may also view the models of their peers, to enable them to compare their own progress
against that of others. Furthermore, instructors may view the learner models of their students, to help them
support students in their learning. The initial version of the system and the learner model are very simple, to
enable early deployment in a variety of contexts. (Subsequent investigations may lead to more detailed
recommendations
for
specific
domains.)
Planned
extensions
to
enable
student-system
and
peer-peer
collaboration with reference to student models (based on existing work on other systems), are also discussed.
Key words: open learner models, learner reflection.
1.
Introduction
There has recently been growing interest in opening the learner model to the individual it represents, with
several systems demonstrating this approach (e.g. Bull & Pain, 1995; de Buen et al., 1999; Dimitrova et al.,
2001; Kay, 1997; Morales et al., 2000; Silva et al., 2001; Specht et al., 1997; Weber et al., 2000). An important
reason for rendering the learner model accessible is to help the student to better understand their learning opening the learner model to the modellee offers a source of information about their relationship with the target
domain which is otherwise unavailable, encouraging them to reflect on their beliefs and on the learning process.
Student reflection on their learning has also been encouraged in collaborative learning situations, where pairs of
students view their own and each other's learner models, in order to provide a focus for collaborative interaction
(Bull & Broady, 1997). Making peer models available more generally could also be beneficial as it would enable
students to compare their progress against that of their peers, as proposed by Kay (1997).
In addition, suggestions have been made to open learner models to instructors, allowing tutors to access
information about those they teach. This can help instructors to adapt their teaching to the individual or to the
group (Grigoriadou et al., 2001; Zapata-Rivera & Greer, 2001); or enable them to suggest suitable peer helpers,
or organise learning groups (Mühlenbrock et al., 1998).
Each of the above approaches to open learner modelling (self-access, peer access and instructor access to an
individual's student model) can be used to help learners to better understand their learning of a target domain.
However, one of the difficulties in effecting learner reflection in this manner on a wider scale, is that typically
systems which have inspectable student models as a focus for encouraging reflection are quite complex, with the
system requiring an understanding of the target domain in order to model in some detail, a student's knowledge,
and infer their misconceptions, in that area. While this results in potentially very effective methods of promoting
5
learner reflection, it also renders the systems relatively expensive to implement, and often restricted to a single
or limited set of domains.
To complement such approaches, we suggest employing a very simple student model in a system that can be
easily deployed in a variety of course types, with the aim of investigating the potential of open learner modelling
in a range of realistic settings, with large numbers of users. While this more straightforward approach to student
modelling will not allow the system to adapt its tutoring to specific misconceptions held by specific individuals,
it will nevertheless allow investigations of broader questions such as whether students will pay attention to their
learner models in a variety of disciplines. Some investigations have suggested that even when they know about
the inspectability of their own student model, learners do not necessarily attempt to view it (Barnard &
Sandberg, 1996; Kay, 1995). Nevertheless, small-scale studies have suggested that students might indeed use an
open student model if it were available, in contexts where they may discuss the contents of their model with the
system (Bull & Pain, 1995; Dimitrova et al., 2001). It is worthwhile, therefore, investigating whether the
disinterested reactions from some students in the former situations are typical, or whether, as suggested in some
other studies, students might find open learner models helpful to their learning at least in some contexts.
Furthermore, it would be useful to discover whether instructors in a range of subjects find open learner models a
useful way of helping them to recognise student difficulties, enabling them to respond to specific student
populations or individuals in appropriate ways. With this as one of the system aims, the limitations in its ability
to offer fine-grained adaptive interaction to an individual student, based on the student model, are less crucial.
Hence the system is intended more as a practice environment, than a tutoring system.
Another issue that can be investigated with simple open learner models is the means of externalising these to
students, peers and instructors. As argued by Morales et al. (2000), presentation modality might be important for
the comprehensibility of student models at least in some domains. Similarly, different individuals might work
better with different model representations, thus even within a single domain it may be useful to offer alternative
or complimentary learner model representations (e.g. Bull, 1998). Furthermore, Zapata-Rivera & Greer (2001)
add that the goals of the system may influence the manner in which student model data should be presented.
This paper describes work in progress on a learner model open to the student it represents, their peers and their
instructor. A student inspecting their own student model can benefit through reflection on its contents; viewing
peer models can enable learners to understand their progress against the context of the advancement of their
peers; and allowing teachers to view individuals' learner models can help them to help those specific learners
individually, and also adapt their teaching to the group's difficulties, where common problems arise.
The learner model presented here is very simple - in its basic version containing representations only of a
student's level of knowledge of various topics. While necessarily restricting the system's ability to adapt to an
individual's misconceptions, it does allow it to be straightforwardly deployed more widely (as a practice
environment) and, as stated above, the involvement of the teacher (or peers) can compensate for the system's
inability to interact with reference to specific misconceptions. Moreover, multiple choice exercises can be
created in text files, further allowing the system to be easily deployed in a variety of subjects, by a variety of
tutors. The simplicity of both the student model and the method of exercise creation permits investigation of the
potential of open student models to help learners to better understand themselves, enabling some of the general
questions pertinent to inspectable student models to be investigated in a variety of contexts.
Section 2 introduces the system. Section 3 describes the inspectable learner models from three perspectives:
inspection by the student (modellee); inspection of the models of peers; and inspection of learner models by the
instructor.
In
each
case,
the
potential
for
helping
learners
to
better
understand
themselves,
is
discussed.
Conclusions are presented in Section 4.
2.
The General Learning Environment
The system is suitable for use in domains where multiple choice questions are appropriate as a means of
practising the target material. The system presents ten questions at a time, each with a drop-down box where the
correct answer should be selected. Figure 1 illustrates the interface for the domain of Japanese particles, where
the grammatically correct particle has to be chosen for each of the sentences.
6
Figure 1: Multiple choice practice questions
Tutors can advise students to use the system to practise material presented during classroom sessions, using
textbooks or course notes for reference. Alternatively, instructors can add their own instructional materials, to be
accessed from within the system.
Questions for each concept (in this example, Japanese particles) are selected randomly. Candidate concepts are
selected according to the student's previous performance. The aim is to provide students with opportunities to
practise the areas with which they are experiencing most difficulty, thus each individual will receive exercises
targeted in particular at a restricted set of concepts which are represented as least understood in their student
model. (During the initial interaction, questions on all concepts are presented.) Standard questions can be loaded
for specific domains. New questions can also be added by instructors, thus for particularly complex concepts
many practice questions can be made available. These can be divided into various levels of difficulty, to enable
the same instantiation of the system to be used by students at a range of levels. It also allows for the same
concepts to be practised in different contexts within the same general domain - for example, Japanese particles in
business or conversational Japanese. Moreover, it allows the system to be deployed in a range of subjects - not
only different languages, but any domain for which appropriate multiple choice practice questions can be
created. New questions are simply added to existing text files, or for new concepts (in the same or a new
domain), new text files can be created. Thus the system can be easily deployed in a variety of courses, to
investigate the research questions.
3.
The Open Learner Model
The basic student model is a very simple, numerical model. The total number of questions attempted for each
concept is stored, as is the proportion of correct versus incorrect attempts. Greater weighting should be awarded
to later answers to enable the system to use this information more accurately to infer where the student's current
difficulties lie. (The weighting algorithm for the student model has not yet been implemented.) As described
7
below, these student models can be extended in particular cases, but for the simplest version of the system, the
student model is a straightforward representation of knowledge levels in the various areas.
The learner model is open to the individual student, to encourage them to reflect on their knowledge and
misconceptions. Students can also access the models of peers, to help them further gauge their progress.
Instructors can view the various learner models to enable them to better help individuals; to adapt their teaching
for a particular group; or to help them set up optimal peer learning or tutoring groups.
3i.
A Learner Model Open to the Student
As described above, the student model contains representations of the student's performance on sets of questions
related to different concepts in some domain. Remaining with the example of Japanese particles, Figure 2
illustrates the way in which the model is externalised to the student.
Figure 2: The open learner model
Currently the student model can be presented to the learner in one or both of two forms: tabular and graphical.
The graphical representation is similar to the skill or knowledge meters of APT (Corbett & Bhatnagar, 1997),
ADI (Specht et al., 1997) and ELM-ART (ELM Research Group, 1998), which display bars where the filled
portions represent the student's attained knowledge or skill against the expert level; and OWL (Linton &
Schaefer, 2000), which displays in bar form a user's knowledge level against the combined knowledge of other
user groups. In contrast to the above, our system illustrates a comparison of correct versus incorrect attempts at
questions. While, as pointed out by Linton and Schaefer, the skill meter approach focuses on achievements
rather than shortcomings, our system aims to raise learner awareness of their performance in general - i.e. not
only the degree to which they have mastered any particular aspect of the domain, but also concepts with which
they are having difficulty (in contrast to areas they have not yet attempted).
Other external representations for the student model will be considered at a later date. Nevertheless, as it is
knowledge level that is represented, rather than specific concepts and misconceptions, these representations will
likewise not be complex.
8
At present, a score of 1 is awarded for each correct answer for a given concept, and 1 is subtracted for an
incorrect response. This can be seen in the tabular representation in Figure 2. (Representations are updated when
students complete a set of ten questions.) The overall score, illustrated in both the tabular and the graphical
representation
implementation
of
of
the
the
student
model,
mechanism
to
indicates
award
the
greater
student's
weighting
overall
to
later
performance.
answers
is
(As
still
stated
above,
required.
Once
implemented, the student model will better reflect the student's actual current knowledge levels.)
The student can use the externalised learner model to see at a glance areas in which they are strongest and
weakest, and use these as recommendations of what should be studied further, within or outside the system. If
instructors choose to add domain content or explanations of common misconceptions to a particular instantiation
of the system, these can be linked to the student model, and examined by the student.
It can be seen that the information shown to students about themselves, is limited. An extension to the system,
allowing the learner to negotiate the contents of their student model, is planned. This is based on the MR
COLLINS approach of collaborative student modelling (Bull & Pain, 1995), where the student can argue with the
system if they disagree with the contents of their learner model. Negotiating the content of the learner model is
designed in part to enhance learner awareness of their learning. In the case of the new system, such negotiation
will be less complex, focusing on the system offering the student a quick test if they claim to know something
that they have not yet adequately demonstrated (with the system accepting the student's claim if they can
demonstrate their knowledge in the test); and allowing the learner to inform the system if they believe its
representations of their knowledge level of any concept is too high (for example, if they had guessed the answer
to some of the multiple choice questions, and by chance got them correct; or if they have simply forgotten
previously known material). This will enable the student to influence their learner model, and thereby also
influence the subsequent selection of practice material.
3ii.
A Learner Model Open to Peers
In addition to inspecting their own learner model, a student can compare their performance to that of peers in
their group (this occurs anonymously unless students choose to reveal their usernames to others). This enables
learners to appreciate how their developing understanding of the target concepts compares to that of other
learners, as suggested by Kay (1997). Students can retrieve the student models of good peers, to appreciate the
level of knowledge for which they could aim; and also of weaker peers, which in some cases could help learners
to realise that they are performing better than they had realised. Figures 3 and 4 show a comparison of the
student models of different individuals.
Figure 3: Comparing student models (graphical)
9
Figure 4: Comparing student models (tabular)
A planned extension to the current system will also allow learners to compare their student model against the
'average' student model of all other learners (or a subset), in their cohort.
The possibility of viewing peer models, in addition to enabling students to gauge their progress against other
learners, also allows a student to collaborate with another learner, using their respective student models as a
focus for discussion. In 2SM (Bull & Broady, 1997), displaying two student models to their co-present owners
was designed to prompt peer tutoring, with each student individually completing an exercise, and then coming
together to repeat the same exercise in the presence of their two inspectable learner models. In the new system, a
pair (or small group) of learners can also come together, viewing their own and their partner's student models,
and learning from collaborative interaction. In contrast to 2SM, each learner will have experienced a different set
of questions tailored to their own needs, and the combined exercise will again be different. The pair will
commence a fresh session, building a new student model reflecting their joint performance, drawing on the
representations in their individual learner models to recognise whose explanation is more likely to be correct, in
cases where they disagree about an answer when collaborating about their joint answers. (The system provides
correct answers for student consultation after an exercise has been completed. Thus, should the pair agree on an
incorrect answer, they are made aware of any inaccurate selection, and can see what the correct choice should
have been.) The collaboration and peer tutoring expected to occur with the comparison of learner models (see
Bull & Broady, 1997), will focus students' attention more directly on their knowledge and misconceptions. An
interesting observation will be how students decide on the representation type to view, if their preferences differ.
A planned extension to the current version of the system is to allow it to use the two student models of a
collaborating pair of learners, to adapt the joint exercise to best utilise the relative strengths and weaknesses of
the pair.
10
With the extensive sets of questions possible, students can potentially work with a variety of peers on a set of
concepts if they prefer to learn with others. They can use the student models of others to help them select
suitable collaborative partners or helpers who complement their own abilities. An extension to the system could
even enable it to suggest suitable learning partners based on the contents of the various participants' learner
models (see e.g. Collins et al., 1997).
3iii.
A Learner Model Open to the Tutor
The final use for the open student model is to help tutors to better understand their students. In the same way that
students can compare their learner model to the models of others, teachers can do likewise with their students'
learner models. This can allow instructors to help individuals with particular problems, or enable them to better
target their teaching to the general difficulties of specific groups. The possibility of viewing a combined
'average' student model (see 3i) would facilitate the latter process. This is almost the reverse of the original aim
of intelligent tutoring systems - to teach students as a human tutor might, using knowledge about a student
contained in their student model. Here the student model is providing information for the human teacher, helping
them to adapt their teaching appropriately.
Although not part of the initial planned investigations, teachers could also view models in order to form learning
groups
(as
in
Mühlenbrock
et
al.,
1998).
Teachers
might
ultimately
be
able
to
provide
some
additional
information to learner models based on, for example, student performance on assignments, and engage in
system-mediated discussion with a student, about their learner model (see Bull, 1997).
In the first deployment three instructors will use the system in three different courses: Japanese, physics and
interactive systems.
4.
Summary
This paper has described a system with an open learner model designed to be viewed by the student modelled,
their peers, and their tutor. The student model therefore has an even more central role than in the traditional
intelligent tutoring system - as a learning resource for the student, to help them reflect on their beliefs in a
student-system setting; as a means to prompt collaboration and peer tutoring; and as a source of information for
instructors to aid the human teaching process. The system is currently in the early stages of development. While
it could now be deployed in its most simple state, a few extensions will increase its utility as a focus for
promoting learner reflection on their understanding. These extensions include a simple form of collaborative
student modelling, and taking into account pair models to further support peer collaboration. Most importantly,
weightings
will
be
applied
to
the
most
recent
answers
given,
in
order
to
ensure
that
the
student
model
appropriately updates as the student learns.
Even with these extensions, the system will remain straightforward. It is a practice environment rather than a
tutoring environment - though the instructor may provide links to course materials. Its primary aim is to
investigate some of the questions relevant to the use of inspectable learner models in a range of contexts - both
different subject areas, and different students. Such questions include the method by which learners prefer to
view (and later interact with) their learner model - the individual's preferred method of accessing the learner
model is likely to be more relevant than, for example, the influence of the domain, as it is knowledge
level
(rather than actual knowledge or misconceptions), that is represented. The goal here is to raise learner awareness
of where their strengths and weaknesses lie, thus representations similar to those illustrated, could be usefully
considered. Other issues include how much attention students pay to their own and to peer models; whether
interaction with the learner model occurs with some students more than others; with some domain types more
than others; or whether there are differences in the same student's acknowledgement of their learner model in
different domains. Also relevant are instructors' use of their students' models.
It is acknowledged that a more complex student model is ultimately likely to be most helpful, in particular in the
context of the individual student viewing their own learner model. However, to investigate questions such as the
above, it is useful to employ a simple learner model in an easy to deploy system. Of course, more complex
11
student models will have additional requirements for their externalisation - these can be investigated in parallel,
or once some of the simpler questions have been addressed. An obvious question is how to show larger models the current type of display, while clear for small domains, may need to be modified if many interrelated concepts
were to be included in a single exercise set. Nevertheless, even with our initial straightforward approach, it is
intended that students will be able to benefit from interactions with the system, and that some initial guidelines
about the effective externalisation of learner models can be obtained.
References
Barnard, Y.F. & Sandberg, J.A.C. (1996). Self-Explanations, do we get them from our students?, in P. Brna, A.
Paiva & J. Self (eds), Proceedings of European Conference on Artificial Intelligence in Education, Lisbon,
115-121.
Bull, S. (1997). See Yourself Write: A Simple Student Model to Make Students Think, in A. Jameson, C. Paris
& C. Tasso (eds), User Modeling: Proceedings of the Sixth International Conference, Springer, Wien New
York, 315-326.
Also available from: http://w5.cs.uni-sb.de/~UM97/abstracts/BullS.html
Bull, S. (1998). 'Do It Yourself' Student Models for Collaborative Student Modelling and Peer Interaction, in
B.P. Goettl, H.M. Halff, C.L. Redfield & V.J. Shute (eds), Intelligent Tutoring Systems, Springer-Verlag,
Berlin, Heidelberg, 176-185.
Bull,
S.
&
Broady,
E.
(1997).
Spontaneous
Peer
Tutoring
from
Sharing
Student
Models
in B. du Boulay & R. Mizoguchi (eds), Artificial Intelligence in Education, IOS Press, Amsterdam, 1997,
143 - 150.
Also available from: http://www.eee.bham.ac.uk/bull/papers/AIED97.htm
Bull, S. & Pain, H. (1995). 'Did I say what I think I said, and do you agree with me?': Inspecting and
Questioning the Student Model, in J. Greer (ed), Proceedings of World Conference on Artificial Intelligence
in Education, Association for the Advancement of Computing in Education (AACE), Charlottesville, VA,
1995, 501-508.
Also available from: http://www.eee.bham.ac.uk/bull/papers/AIED95.htm
Collins, J., Greer, J.E., Kumar, V.S., McCalla, G.I., Meagher, P. & Tkatch, R. (1997). Inspectable User Models
for
Just-In-Time
Workplace
Training,
in
A.
Jameson,
C.
Paris
&
C.
Tasso
(eds), User
Modeling:
Proceedings of the Sixth International Conference, Springer Wien New York, 327-337.
Also available from: http://w5.cs.uni-sb.de/~UM97//abstracts/CollinsJA.html
Corbett,
A.T.
&
Bhatnagar,
A.
(1997).
Student
Modeling
in
the
ACT
Programming
Tutor:
Adjusting
a
Procedural Learning Model with Declarative Knowledge, in A. Jameson, C. Paris & C. Tasso (eds), User
Modeling: Proceedings of the Sixth International Conference, Springer Wien New York, 243-254.
Also available from: http://w5.cs.uni-sb.de/~UM97/abstracts/CorbettA.html
de Buen, P.R., Vadera, S. & Morales, E.F. (1999). A Collaborative Approach to User Modeling within a MultiFunctional Architecture, in J. Kay (ed), UM99: User Modeling, Proceedings of the Seventh International
Conference, Springer Wien New York, 291-293.
Also available from: http://www.cs.usask.ca/UM99/Proc/short/BuenRodriguez_032414.pdf
Dimitrova, V., Self, J. & Brna, P. (2001). Applying Interactive Open Learner Models to Learning Technical
Terminology, in M. Bauer, P.J. Gmytrasiewicz & J. Vassileva (eds), User Modeling 2001: 8th International
Conference, Springer-Verlag, Berlin Heidelberg, 148-157.
ELM Research Group (1998). ELM-ART Lisp Course.
http://www.psychologie.uni-trier.de:8000/elmart
Grigoriadou, M., Papanikolaou, K., kornilakis, H. & Magoulas, G. (2001). INSPIRE: An Intelligent System for
Personalized Instruction in a Remote Environment, in P. De Bra, P. Brusilovsky & A. Kobsa (eds), PreWorkshop Proceedings: Third Workshop on Adaptive Hypertext and Hypermedia, User Modeling 2001, 3140.
Also available from: http://wwwis.win.tue.nl/ah2001/papers/papanikolaou.pdf
Kay, J. (1995). The UM Toolkit for Cooperative User Modelling, User Modeling and User Adapted Interaction
4, 149-196.
Kay, J. (1997). Learner Know Thyself: Student Models to Give Learner Control and Responsibility, in Z. Halim,
T.
Ottomann & Z. Razak (eds), Proceedings of International Conference on Computers in Education,
Association for the Advancement of Computing in Education (AACE), 17-24.
12
Also available from: http://www.cs.su.oz.au/~judy/Research_um/index.html
Linton, F. & Schaefer, H-P. (2000). Recommender Systems for Learning: Building User and Expert Models
through Long-Term Observation of Application Use, User Modeling and User-Adapted Interaction 10, 181207.
Morales, R., Pain, H. & Conlon, T. (2000). Understandable Learner Models for a Sensorimotor Control Task, in
G.
Gauthier,
C.
Frasson
&
K.
VanLehn
(eds),
Intelligent
Tutoring
Systems:
Proceedings
of
the
5
th
International Conference, Springer-Verlag, Berlin, Heidelberg, 222-231.
Mühlenbrock, M., Tewissen, F. & Hoppe, H.U. (1998). A Framework System for Intelligent Support in Open
Distributed Learning Environments, International Journal of Artificial Intelligence in Education 9(3-4),
256-274.
Silva, A., Vale, Z.A. & Ramos, C. (2001). Inspectability and User Controlled Revision on Long Term User
Models, in M. Bauer, P.J. Gmytrasiewicz & J. Vassileva (eds), User Modeling 2001: 8th International
Conference, Springer-Verlag, Berlin Heidelberg, 254-256.
Specht, M., Weber, G. & Schoech, V. (1997). ADI: Ein adaptiver Informations- und Lehragent im WWW, in R.
Schaefer & M. Bauer (eds), Proceedings of ABIS'97: Workshop Adaptivitaet und Benutzermodellierung in
interaktiven Softwaresystemen, Universitaet des Saarlandes, Saarbruecken, 53-60.
Weber,
G.,
Kuhl,
H-C.
& Weibelzahl,
S.
(2001).
Developing
Adaptive
Internet
Based
Courses
with
the
Authoring System NetCoach, in P. De Bra, P. Brusilovsky & A. Kobsa (eds), Pre-Workshop Proceedings:
Third Workshop on Adaptive Hypertext and Hypermedia, User Modeling 2001, 41-53.
Also available from: http://wwwis.win.tue.nl/ah2001/papers/GWeber-UM01.pdf
Zapata-Rivera, J-D. & Greer, J.E. (2001). Externalising Learner Modelling Representations, Proceedings of
Workshop
on
External
Representations
of
AIED:
Multiple
Forms
and
Multiple
Roles,
International
Conference on Artificial Intelligence in Education 2001, 71-76.
13
A Diagnostic Agent based on ZPD approach to improve
Group Learning
1
2
Adja F. de Andrade
3
Paul Brna
School of Computing
Rosa Maria Vicari
PGIE- UFRGS
School of Computing and Mathematics
University of Leeds
Caixa Postal: 5071
University of Northumbria, NEI 8ST - UK
LS2 9JT - UK
Zip Code: 90041-970
Paul.Brna@unn.ac.uk
adja@cbl.leeds.ac.uk
Porto Alegre-RS
rosa@inf.ufrgs.br
Abstract: This paper explores the question of cognitive diagnosis based on the analysis of ZPD (Zone
of Proximal Development) based on a Vygotskian perspective and Core concepts. The main aim is to
present a new proposal of architecture, to model the diagnostic agent’s behaviour, describing some
“scaffold tactics”. The paper also presents a scenario example to illustrate this proposal. This research
demonstrates that the possibility of using skills, competencies and difficulties to parameterise the
formation of a group can become an interesting approach for cognitive diagnosis. Although, we are
conscious of the fact that to form groups does not guarantee that the learning will be better, we believe
that it permits social interaction and exchange of knowledge, important pre-requisites in the group
learning.
Keywords: Diagnosis, Modelling Group, Zone of Proximal Development.
Introduction
If we don’t know each other, how can we interact? If we don’t know our own skills and difficulties,
how can we share them, request or offer helping? The solution is not in building the “perfect” learning
environment. But, perhaps, an environment able to describe our cognitive model and share this model
with our peers. Perhaps, we are inserting a high responsibility for the technology. This certainly seems
an ambitious goal regarding the actual “pedagogical role” of the technology.
Motivated for this challenge, this paper presents the first sketches of the cognitive diagnosis using a
Vygotskian
perspective.
The
main
aim
is
to
present
a
new
proposal
of
the
architecture
for
the
diagnostic agent and to model his behaviour.
To explain this proposal, we begin classifying two important parameters: ZPD and Core. The first
parameter is based on Vygotsky’s theory. This concept defines what we are not able to do alone but we
can do with some scaffold. The second parameter identifies what we know, that is, which are our skills
-Core level. The concept of core was firstly found in the work of Hansen (Hansen, 1999). Recent
research (Kay, 2001b) has shown a tendency towards using the ZPD concept to build the learner
model.
In our approach, we will use the conception of ZPD to parameterise the formation of group
model and not only a model of learner. About the kind of model utilised, we are investigating how to
build an open learner model, where the student will be able to intervene in his own diagnostic process.
The group model is inspired by some attributes presented in Paiva’s work (Paiva, 1997). Based on
Luckin’s work (Luckin, 1996, 1999) we describe the use of the ZPD.
1
PhD Student in PGIE- Federal University of Rio Grande do Sul (UFRGS). Lecturer at the Pontifical
Catholic
University
of
Rio
Grande
do
Sul
Brasil
(PUCRS).
Currently
in
School
of
Computing
at
University of Leeds.
2
3
Supervisor in Leeds-UK.
Supervisor in Graduate Program in Computer Science in Education-PGIE, Brasil.
14
The research problem refers to investigate the diagnosis process able to detect the zone of proximal
development of the learner in a distance learning context. The motivation to use this approach is
founded in Vygotsky theory. He defends that learners can improve their learning through mediated
interaction with more capable peers.
However, to detect the skills that needs support, firstly, we have
to diagnostic them, and furthermore to identify the possible “scaffold tactics” adequate to the level of
the learner.
The approach adopted is based on the analysis of the beliefs and observation of the task done by the
group. This task is monitored by a mediator agent and sent to the diagnostic agent. The proposal
analyses the self-confidence model of the learner, which represents a kind of self-evaluation made by
the learner. The evaluation of the diagnosis will be a dynamic evaluation, with focus in zone of
proximal
development
of
the
learner.
The
first
step
in
the
evaluation
process
is
to
analyse
the
individual performance of the task, the second moment this evaluation is compared to the performance
of the group, which involves mediation with other learners more able in this knowledge domain. The
focus of our analyses privileges the process of interaction and not only the result (product) of task
realised.
One of the big difficulties to describe the diagnostic agent is due to need to interact with other agents
(mediator, semiotic, collaboration), exchange message, register and evaluate the behaviour of the group
to determine the diagnosis and take some support decision.
This is a distributed task, that request a
complex architecture involving communications protocols, access to group models, task performance
module and still maintaining the individual characteristics of the learners.
The actual stage of this PhD research is the formalisation and modelling of the behaviour of the
diagnostic
agent
and
definition
of
the
diagnosis
process.
After
this
stage,
we
will
start
the
implementation and the experimental study case. For the implementation of the diagnosis process, we
are
investigating
several
approaches
based
on
model,
mentalist,
stereotype,
overlay,
bugs,
and
misconception approaches. Until the present moment the intentional approach based on the mental state
seems to be the most adequate for our work.
Although, the ZPD approach is a consolidated concept presented in Vygotsky’s theory and recently has
been used by several researchers as a diagnostic function. It is important to highlight, even thought, we
still do not have any concrete result, which would make a pedagogical validation possible to underline
the significance of this approach.
The Society of Pedagogical Agents
Pedagogical
Agents
(Gürer,
1998)
are
described
as
intelligent
agents
that
have
an
educational
or
pedagogical role to facilitate or improve learning. These agents can be modelled as personal and animated
agents that interact with the user, or as cooperative agents who work in one background as part of the
architecture of the educational system.
The
society
of
agents,
a
multiagent
system
is
formed
by
four
agents:
Diagnostic, Mediator,
Collaboration and Semiotic. This architecture is part of the project “A Computational Model of
Distance Learning Based in the Socio-Cultural Approaches” (Andrade, 2001)(Jaques, 2002). The
Mediating
Agent is the interface agent of the society. The
Diagnostic
Agent has the function of
diagnosis and updates the information in the group model, besides sending “scaffold tactics” to the
Mediating Agent. The Semiotic Agent is a service agent that has the function to search for content in
the knowledge base and send to the Mediating Agent to be shown to the student. The Collaboration
Agent is an animated character agent who has the role of searching partners capable of assisting one
student
in
his/her
learning
and
to
mediate
the
interaction
between
students
using
collaborative
communication tools. The architecture of the multiagent system is shown in figure 1 below.
15
Semiotic
Agent
Request Pedagogical Content
WWW
Exercises
Access Knowledge
Examples
Base
Request
Pedagogical
Agents that aid in Student ZPD
Content
Pedagogical Tactic
Student Behavior
Diagnostic
...
Mediating
Agent
Agent
Cognitive
Group
Profile
Model
Diagnostic
Mediating
Agent
Student
Agent
Model
Actions
Affective
tool)
Students Group (Collaboration
Mediating
Agent
Profile
User
Pedagogical
Content
Pedagogical Tactic
Pedagogical Content
Student Message
Message to the student
Collaboration
Group
Group
Model
Agent
Message to
the student
Student
Student
Profile
Figure 1. Multiagent architecture (Andrade et al, 2001)
Diagnostic Agent
A Diagnostic Agent has the function of describing the cognitive diagnosis, modelling the group and
suggesting “scaffold tactics”. Initially, when the task is proposed for the group, the diagnostic agent
must create a mechanism to evaluate which skills are in the core region and which are in the ZPD
region.
To start the diagnosis process, the diagnostic agent must propose some task to the learner or group
varying the degree of difficulty. This is represented as an action that is translated to communication
speech act (FIPA, 1997):
act
(propose-task
(TID,
task(name(-))))
where task is associated to a definition as task
(name(-),
[skill-list]).
The response expected for this message is a message of the sort:
task-outcome
(TID,
[[core-skill],
[ZPD-skill]])
where TID is a task identification. A task may have many instances, i.e., it may occur several
times.
These messages are generates by diagnostic agent and sent to the mediator agent, which provide
adequate assistance for skills in ZPD. In figure 2, we can see the internal architecture of the Diagnostic
Agent.
16
BDI Model
Desires, Intentions
and plans
Beliefs
Practical Reasoning
Theoretical Reasoning
Open
Learner
model
Cognitive
diagnosis
Group
Model
Open
Mediator
Diagnosis
Learner
Agent
model
Validation of diagnosis
Knowledge
Update
Module
Open
Learner
Tactic of scaffold
Scaffold
model
Module
Message of Learner behaviour
Teacher or
Sensor
Learner behaviour
Module
Other
Level
Mediator
of
support
Tactics of Scaffold
Message of tactic of scaffold
Effec
tor
diagnose
Colla
borat
Figure 2. Diagnostic agent architecture
The Sensor Module is an internal part of the diagnostic agent. He has the role of communicating with
mediator/collaboration agents. This module has the objective to interpret the message that arrives from
the mediator and collaboration agents. The messages are described in ACL format (Fipa, 1997) and
represent the input of the diagnosis process. The ACL is a communication language that permits one
agent to send a demand to another agent to perform some task. The messages are received through the
sensor and sent to the communication module for the disencapsulation or separation of messages in
blocks. The next phase is the treatment or interpretation of this message. When the diagnostic agent
needs send a new message, he has to set the encapsulation status to re-start the process again. This
process
is
needed
because
the
diagnostic
agent
needs
to
communicate
with
other
agents
in
the
architecture, and he is not able to make any decision about the diagnostic without this interaction. The
choice of ACL format is because the message need to be stored in an adequate structure that becomes
possible the agent to understand your content. One example of this message is outline follows:
(inform
:sender Mediator
:receiver Diagnostic
:content <create-web_page>
:ontolgy <exercise>
:reply-with <profile_interaction>
)
17
The
Effector
Module
collaboration agents.
is
responsible
to
send
messages
with
“scaffold
tactics”
to
mediator
and
Besides, this module informs the diagnostic of the learners for the other agents
for that they can facilitate the learning.
The Open Learner Model is formed by the cognitive and emotional profile of the learner. The cognitive
profile stores the information about beliefs, skills, difficulties and assistance. The emotional profile
contains information about the personality of the learner, like introvert, extrovert, if he/she likes to
work in group, and his/her level of motivation.
These parameters are used as first reference, however
they can change during the interaction. This model is considerate as an open learner model because it is
inspectable by the learner (Dimitrova, 2001) (Kay, 2001b), which he/she can analyse and agree or
disagree about his/her diagnosis.
The Group Model is formed during the evaluation of the learner’s ZPD.
When the diagnostic agent
discovers that the learner has some skill in ZPD, he suggests forming a group with some expert that has
knowledge in that domain area and can help that student. It maintains not only a cognitive status, but
also an affective profile of the group sent by the collaboration agent. The Knowledge Update Module
updates the agent beliefs about the learner’s performance, goals and skills.
The Scaffold Module is formed by description of task, list of skill in ZPD, level of support (low,
moderate, advanced) and the tactics suggested.
The tactics of scaffold (for instance modelling, start
solution, give clues) before to be applied must be observed the level of knowledge of the learner.
The
tactics have the role to help the students to perform some task, which skills are in ZPD level. These
tactics are sent through the diagnostic agent to the mediator agent and the collaboration agent (Jaques,
2002) that interact directly with the learner.
Diagnosis is the main module of the architecture.
Its function is to diagnose what is in ZPD or Core
level. The diagnosis starts when the diagnostic agent suggests some task to be performed by the group
without any support. After, the group accomplished this task, the agent assess the task performance
model and self-confidence model, which describe the level of knowledge and confidence of the learner
to realise the given task. In function of this analysis the agent determine the ZPD_skills, in other words,
the
skills
that
the
learner can
not
carry
out
alone
and
need
some
“tactics
of
support”.
For
the
pedagogical validation of the diagnosis, the agent must communicate with the teacher. The diagnosis
must be updated also in the group model.
The
Theoretical-Reasoning
Module
represents
the
agent’s
beliefs.
These
beliefs
modelling
the
knowledge about the domain (based on the BDI description to be described in the next section).
However, Beliefs and desires are not enough to implement the behaviour of agent. They need of a
“plan of actions” to achieve the goal and desires. The Practical Reasoning Module represents the
planning module, in other words, this module describes the agent’s reasoning about what it should do.
Diagnostic Agent Modelled with BDI Architecture
When we desire to describe the human behaviour is common to use terms like “believe”, “want”,
“desire”, “need”, etc. These terms are used by beeing human to explain the observable proprieties of
the mind. With the objective to fundament these intentional explanations, also denominated intentional
systems, we describe one set of mental states, which will represent the behaviour of the diagnostic
agent.
In the literature, a number of different approaches have emerged to modelling agent-oriented learning
environment. The mentalist approach (Bratman, 1990) allows view the system as a rational or cognitive
agent having certain mental attitudes of Belief, Desire and Intention (BDI), representing respectively,
the information, motivational and deliberative state of the agent. Our goal is to model the diagnostic
agent using a human metaphor. In the same way, the knowledge of learners also will be modelled
through your beliefs and intentions.
The choice to using the BDI approach is to facilitate the modelling of the diagnostic agent.
Several
intelligent tutoring systems have adopted this approach. But other approaches also could be used for
18
this function. BDI is only one form to describe the behaviour of diagnostic agent and your knowledge.
It enables us to look at the importance of the social context and the believes, desires and intentions of
the diagnostic agent involved. Nevertheless, the focus this research is not the formalisation in BDI, but
the idea of ZPD to describe the diagnosis process.
The beliefs represent the knowledge of the world. They are some way of representing the state of the
world. BEL (a, p) means that the belief p has been ascribed to the agent a by some learner. The mental
state Belief is visualised as the most basic epistemic concept usually considered in the form of
preposition. Desires (DES) are commonly thought like other essential component of system state. They
represent the motivational state of the system. Intentions are representations of possible actions (or
chosen course of action) that the system may take to achieve its goals. The intentions capture the
deliberative component of the agent.
The diagnostic agent has two main desires: diagnostic and support.
The first desire has the objective to
identify the skills learned (skill-cor) and the skills that need support (skill-zpd). To achieve this desire the
agent has the belief that the learner’s skills will be in ZPD level when the learner request help or the task is
not performed with success.
The second desire is to support the group using some strategy of managing of group.
The agent will use
the merge strategy when he desires join one learner (L1), expert in some knowledge domain (skill-cor),
with some learner novice (L2) that has difficulties and his/her skills are in ZPD. One example of the beliefs
and desire bases is outlined follows:
DES (Diag, diagnostic_skills)
BEL(Diag, skill_ZPD) if
BEL(Diag, skill_cor) if
BEL (Diag, diagnostic_skills) if
BEL (Diag, next (BEL (group,
BEL (Diag, next (BEL (group,
BEL (Diag, skills_ZPD) or
request_ajuda))
not-request_ajuda)),
BEL (Diag, skills_cor)
BEL (Diag, next (BEL
BEL
(group,competence_task_is
competence_task_is_successful)
unsuccessful))
BEL (Diag, next (BEL (group,
(Diag, next
(BEL
(group,
confident_knowledge))
DES (Diag, support_group)
DES (Diag, support_group)
DES (Diag, support_group)
BEL (Diag, group_merge) if
BEL (Diag, group_split) if
BEL (Diag, group_cor) if
BEL (Diag, skill_ZPD_L1)
BEL (Diag, evaluate_subgroup),
BEL (Diag, skill_cor_G1)
Overlap with BEL (Diag,
BEL (Diag, skill_cor_L1)
overlap with (Diag,
skill_cor_L2)
BEL (Diag, skill_cor_L2)
AND
skill_cor_G2)
Table 1. Beliefs and Desire base
Formalisation of Core and ZPD
The notion of the zone of proximal development (ZPD) was proposed by Vygotsky- as "the distance
between the actual developmental level as determined by independent problem solving and the level of
potential development as determined through problem solving under adult guidance or in collaboration
with more capable peers" (Vygotsky, 1978).
The concept of core is mentioned in Lewis’ work (Lewis,
2000), where he mentions that “the knowledge of an individual has a central core that is “owned” by the
individual who is able to use that knowledge in the autonomous performance of tasks”.
The core i s a subset of the domain, which represents the knowledge internalised (learned) by the
learner or group.
Definition 4.1 (Domain): The domain is a structure < D, pre>, where:
•
D is a set of sentences in the form skill(name(…),
type,
pedagogical_goal,
[support_list],context) and Pre is a relation D x D stating that a given skill is a
pre-requisite of some other skill.
19
Definition 4.2 (core): Given a Knowledge Domain <D, pre>, the core of a learner is a structure
<Cor,L> where:
Cor
•
•
⊆
D.
Cj ∈ Cor ∧ (Ci , Cj ) ∈ pre then Ci ∈ Cor
If
Ci, Cj are elements of core.
The ZPD is also a subset of domain that describes skills that are not internalised, i.e., skills that the
learner does not have yet, but that he/she can perform with some support or scaffold. Hansen defines
ZPD as: “Surrounding that core is in ZPD region, in which the individual has some knowledge but not
the full structure of capacities required, and thus needs help in performing tasks that depend upon that
knowledge”(Hansen, 1999).
Definition 4.3 (ZPD): Given a knowledge Domain <D,
<ZPD,L>
pre>, the ZPD of a learner L is a structure
where: ZPD is a subset of D.
The definition of core and ZPD (Dillenbourg, 1992b) are practically the same. In fact, the nature of the
propositional content of the core and ZPD are very similar, although dynamic. In a given moment a
skill
can
be
in
the
core
and
in
another
can
be
is
ZPD.
The
skill
can
move
from
ZPD
to
core
(internalisation) or from core to ZPD (externalisation).
core knowledge
zone of proximal
development
individuals
groups
Figure 3. Core knowledge and zones of proximal development for
individual and group (from Hansen et al, 1999)
Figure 3 above shows the representation of core and ZPD. When a community or group is considered,
some parts of each person’s core knowledge overlap with that of others. Besides this, that one person’s
ZPD may overlap with the core of others as well.
From this model, one might conclude that the
collective core, union of cores, is greater than that of an individual; but also that each person can
support cognitive development in the group by providing scaffolding for others (Wood, D., Bruner,
J.C., & Ross, G, 1976).
Group Diagnosis
The Student Model is one of the important component in a traditional ITS. It is a collection of data about
the knowledge’s of the student and is used by other components of ITS (in our case by others agents) to
planning some sequence of instruction, feedback, explanation or scaffold. Sometimes the term student
20
model is called student diagnosis in the literature. For a literature review of cognitive diagnosis we suggest
(Self, 1994) (Dillenbourg and Self, 1992a) and (Ragnemalm, 1996).
Student diagnosis (Ragnemalm, 1996) is defined as “the abstract process of gathering information about
the students and turning that information into the basis for instructional decisions”. In this paper we use the
term group diagnosis for the process of cognitive modelling of the group, because we are interesting in the
behaviour of the group and not only of the learner. As well as the student, group diagnosis has a similar
purpose on how to adapt the tutoring, how to provide explanation or clues to provide the right level of
coaching in the group.
The group model is inspired by Paiva’s work (Paiva, 1997), where the notions of belief, action and group
skills are discussed. This model is formed by a collection of data about the knowledge of the student and is
used by other components of the environment (in our case by other agents) to plan some sequence of
instruction, feedback or scaffold (support). The group model is
also inspired by the work of Luckin
(Luckin, 1996). The model known as VIS- The Vygotskian Instructional System - selects some factors that
combined represent the learner’s ZPD. This model can be extended as for individual as groups, considering
individual as instance of groups. The main attributes proposed are described as follows: group beliefs,
social context of interaction, group skills, motivational and emotional characteristics, group difficulties and
group relationship (i.e. Assistance required and offered).
The approach used to evaluate the learner’s performance is based on Kay’s work (Kay, 2001a) through
the use of stereotypes like novice, intermediate and expert. In her work, the author describes the
essential elements of a stereotype, namely triggers, inferences and retraction. The stereotype approach
seems an appropriate way to establish the initial model of the learner, before the diagnosis agent starts
the interaction activity. It is important to notice that the use of these stereotypes will not be static. On
the contrary, we are proposing a dynamic and ever changing configuration, where one learner can be in
a given moment a novice at one task and in another task have an expert status.
Continuing to follow the notion of stereotype, we adopt the community notion presented in Kay’s work.
This notion adds some significant advantages to our proposal of group modelling because the data
collected for a large number of users becomes feasible for the learners to classify their own stereotype
inside the community. This community notion states that the learner can share some model of
preferences and helps to facilitate the performance assistance in the learning process. The user-adapted
interaction allows the system to “be able to cater for a diverse user population, with different users
having different needs, preferences, interests, background, and so on” (Kay, 2001a).
Some characteristics of the learner/group can also show some means or measure for evaluation if the skills
are or
are not
in
ZPD
level. Some
measures
chosen
for our
evaluation
are confidence, competence,
motivation and help parameters of the learner in their interaction with learning environment (see section
3.1). The confidence level of the student refers to a belief of the learner about some knowledge of their
level of performance in doing some task. The competence means the ability of the learner or group to
perform
some
task
successfully.
The motivation
is
an
important parameter
in
an
interaction
context.
Sometimes, the learner can not perform some task or skill because he/she is not motivated, or there is no
empathy between the members of your group. Finally, the capacity of helping means the capacity of the
learner to accept and offer support. If a learner ignores help, and still cannot perform at the higher level of
the ZPD as expected, then the agent needs to rethink the support. Perhaps the skill is outside this learner's
ZPD, or the assistance provided is not useful and should be modified.
The Scaffolding Process
The “scaffolding” term was coined by Bruner (Wood, D., Bruner, J.C., & Ross, G, 1976) to specify
types of assistance that make it possible for learners to function at the ZPD level.
“Scaffolding” is
currently used to describe how a more able mediator (other student, teacher, computational agent, etc)
can facilitate the learner’s transition from assisted to independent performance.
The Support (or scaffold) is a kind of assistance offered to the learner to perform some task that is at
the ZPD level.
This support is applied according to the level of the learner with relation to a given
domain knowledge. The selection of support is based on the notions of stereotype and community
21
described by Kay (Kay, 2001a). Analysis of the group level, the agent that will perform the diagnosis
detects which learners can be classified as novice, intermediate or expert. The support is associated
with activities and actions.
Each support must have a name, parameters, level and tactics and is
represented as:
support(name(…),level,tactic)
where level is either low, moderate, or advanced, and
tactic is one of the pedagogical strategies of support.
When a task is selected is offered assistance to achieve a solution.
This assistance also change with
parameters of ZDP. Luckin (Luckin, 1996) argues that the “ZPD metric is the deciding factor in the
choice of type and amount of teaching input”.
We can notice that the ZPD is something dynamic that
has been used for pedagogical decisions and scaffold and must be updated with the learner’s
performance.
In the perspective of assistance, the task is categorised initially into non-scaffold and scaffold approach.
When a task is selected is offered assistance to achieve the solution.
The assistance also change with the
level of the learner (see table 2).
a)
The non-scaffold approach.
The learner must realise the activity without any support. The goal is to
identify, which abilities of the learner are at the ZPD level. When the agent starts the mediation
process he has to consult the model of the learner to know about his background and competence in
this domain. If he starts the scaffold approach without consulting this model, he can propose some
task too hard and destroy the confidence of the learner, or to suggest some task very easy what can
frustrate him/her. So, the support must be effectuated gradually of accord of learner’s level.
b)
The scaffold approach includes using some “scaffold tactics” to support the activity and discussions
about how and when providing some help. We can use different tactics in several levels (see table
2).
The
assistance
can
be also
interpreted
as
scaffold
tactics,
a “step-by-step”
formation,
where through
mediation activity the gradual transfer of responsibility is transferred from the mediator for the learner. In
the modelling of our interface, the own learner or group can help the diagnostic agent in the task of
diagnosis. For this was planning a self-explanation space used by the learner to express their private speech
about the interaction (Moll, 1990) or to share the understanding of the group, one kind of summary of
group beliefs.
There are three different levels of support. The low level is adequate for the group that needs maximum
assistance, generally in the start of the activity. The moderate level is usually suggested in the middle of the
learning process, during the performing of task. The advanced level is more used when the student has a
high level of confidence of their knowledge and is able to express or explain their reasoning. The table
below
shows
the
group’s
support
level
and
the
respective
relation
with
some
tactics
for
assisting
performance.
Self-explanation
Level of support
Tactics
Private speech (individual
Low
Modeling
commentaries)
Start the solution
Give clues
Group beliefs
Moderate
Explanatory
(commentaries of the group)
Advanced
Cognitive structuring
Questioning
Table 2. Ontology of tactics
The agent is responsible to apply the tactics, but the learner or group must be able to choose which tactics
they prefer as support. The modeling tactic is the process of offering behaviour for imitation or analogy, it
is used when the group needs some example, tutorial or some demonstration before starts the activity. The
start solution is used in the begin of the task, when the learner needs some initial support. To give clues is
22
an interactive approach, applied with the goal to guide a learner to improvement in performance on the next
try. It must be used when the group has some difficulties and requests some information, the agent can
send
some
clues
to
help,
these
tactics
permit
the
learners
fill
gaps
in
their
reasoning
process.
The
explanatory tactic is employed when the group needs some explanation more textual or explanatory and
not only tips. Cognitive structuring is more used at an advanced level, when the agent requests the group to
structure their thinking or action, to summarise concepts, schemas or describe some mental operation.
The
questioning tactic is generally used to inquire mental operation and calls for a cognitive response, this tactic
evaluate if the skill’s group is really in core level (learned). For a full exposition of the means of assisting
performance see Moll (Moll, 1990).
The group model described in this paper is being built using two perspectives: focus on the group
beliefs and focus on the nature of assistance. This model looks at the interactions of the learners with
the environment and the evolution in time, meaning that the beliefs and knowledge are subject to more
frequent changes, where different tactics can be used as well as several support levels to improve the
collaborative learning process.
A Scenario Example
In this section, we describe a scenario example to illustrate the process of diagnosis and support. We
desire that the reader builds a mental picture of the system “in action” that helps to understand our
proposal. The scenario is focusing on the role of the diagnostic agent and his interaction with other
agents in the multiagent society. The context is distance learning, the task is to create a WebPage and
the domain is the Internet.
Firstly, one teacher asks the student or a group of students to create a webPage (task) in some
appropriate software. The group begins to explore the software, without any scaffold (support). Some
functions are known, others not, this will show us that only some skills are in the Core level. The
mediator agent will be observing the behaviour and beliefs of this group, asking questions about the
task.
For
example,
if
the
group
knows
how
to
insert
pictures
in
the
page.
Analysing
the
task
performance and the level of confidence/motivation of the learner (see parameters in section 5), the
mediator agent is able to identify which are the group’ difficulties and which part of the process they
need some support. After the mediator agent monitor the task, he must send a message with the
performance of group trough some protocol message ACL to the diagnostic agent. This message
contains the identification of the group, task, assistance required, offered, overall performance, skills
are
not
performed, cause
of
error,
misconceptions,
profile
of
interaction. In
this
moment,
diagnostic agent analyses the difficulties of the group and he requests to the semiotic
the
agent some
content (examples, concepts, etc) that can help the group.
Continuing the process, the diagnostic agent sends the diagnosis result to the mediator agent informing
that the group is in ZPD level and suggest some tactic of scaffolding, for instance, provide examples,
clues, request some content about how to insert hyperlinks for example. In the same moment, an other
agent, the collaborative agent is looking in the Internet for partners, who are more capable peers to
support the group in this task.
How does he do this task? The Collaborative agent communicate with
the personal agents (diagnostic agent of the others students) to know about the cognitive model of other
learners. Before the application the scaffold tactic, the diagnosis agent must communicate with the
teacher for the final validation. After the implementation of the scaffold strategy (show example how to
insert hyperlinks), the mediator agent must inform the diagnostic agent about the self-confidence
evaluation of the group.
Subsequently the diagnostic agent must update the group model with respect to the behaviour for this
task.
The model of the group will be formed for several attributes: group identification, group beliefs
(general
knowledge
about
the
domain),
group
relationship
(difficulties
to
interact),
profile
of
interaction, the common interests (community notion), skills-core (abilities consolidate), like for
instance change colours, insert separators, change colour of background. Beside, the skills-ZPD,
abilities that need support like for instance (insert figures and hyperlinks), support required, support
offered (no), cognitive diagnostic hypothesis.
23
Claim
Create a webPage collaboratively
Support
Provide examples, to make questions
Because
Several students can collaborate around the internet
Check-rule
Evaluate the creation the WebPage by task analysis, Evaluate the level of
To give tips or slots
confidence of learner, Evaluate skill_core (learned)
Evaluate skill-ZPD (need support)
Open Issue
Some conflicts between goals individual and collective
Table 3. A scenario example
Conclusions
This research has examined how Intelligent Learning Environments can provide useful diagnostic
information. The arguments presented consider the socio-interactionist approach of Vygotsky, at the
level of pedagogic model, and using the agents' paradigm, at the level of computational model. This
paper has provided the first sketches about cognitive diagnosis using the concepts of the ZPD and core.
The status of this work is that the diagnostic process has been defined and is being formalised in a BDI
Model (Bratman, 1990). The next step is the definition of requirements to describe the interface for the
diagnostic agent.
There is a lot of research to explore the diagnosis domain, specially, because this subject involves the
question of
“evaluation”, “support” and modelling of the learner, important discussion point in the ITS
community.
We
parameterise
the
have
seen
formation
that
of
the
the
possibility
group
can
of
be
using
an
skills,
interesting
competencies
approach
to
and
build
difficulties
group
to
model.
Although, we are conscious that to form groups does not guarantee that the learning will be better, we
believe that it permits social interaction and exchange of knowledge, important pre-requisites in the
learning process.
In short, we described the learning environment like an efficient assistance channel. We believe that
the
question
involving
the
designer
of
environments
that
effectively
promote
the
support
and
collaboration between the learners, can be a gap with complexity and relevance to be explored in a
thesis work. As a future result, we are optimists that the ideas presented in this work shows perspective
to increase the group performance and improve learning.
Acknowledgements
This research is supported by Brazilian Government (CNPq/CAPES) and Northumbria University.
Special thanks to Michael C. Mora and anonymous reviewers, all of whom provided many constructive
comments and appointed open questions and problems on previous version of this paper.
References
Andrade, A. F.; Jaques, P. A.; Vicari, R. M.; Bordini, R.; Jung, J. (2001). A Computational Model of
Distance Learning Based on Vygotsky's Socio-Cultural Approach. Proceedings of the MABLE
Workshop (Saint Antonio, Texas, May 19-23). X International Conference on Artificial Intelligence
on Education.
Bratman, M (1990). What is Intention? In: Intentions in Communication. P.R. Cohen et al (Eds). MIT
Press.
Dillenbourg P., Baker M., Blaye A. and O’Malley C. (1994). The evolution of Research on
Collaborative Learning. In Spada and Reimann (Eds.) Learning in Humans and Machines.
Dillenbourg, P. and Self, J. A.(1992a). A Framework for learner modelling. Interactive Learning
Environments, 2(2):111-137.
24
Dillenbourg, P. (1992b). The Language Shift: A Mechanism for Triggering Metacognitive Activities. In
Adaptive Learning Environments Foundations and Frontiers. M. Jones and P. Winne (Eds.)
Springer-Verlag, vol.85, 287-315. ISBN: 3 54055459-9.
Dimitrova, V., Self, J. and Brna, P. (2001). Applying interactive open learner models to learning
technical terminology, Proceedings of the 8th Conference on User Modelling, Berlin:Springer,
LNAI.
FIPA Specification. (1997). http://www.fipa.org/
Giraffa, L. M. M., Vicari, R. M.(1998). The use of agent techniques on Intelligent Tutoring Systems. In:
18th International Conference of Chilean Computer Science Society.
Gürer, D. (1998). The Use of Distributed Agents in Intelligent Tutoring. Proceedings of Workshop On
Pedagogical Agents, ITS’98 (San Antonio) p.20-25.
Hansen, T; Holmfeld, L; Lewis, R.; Rugelj, J (1999). Using Telematics for Collaboration Knowledge
Construction. In Collaborative Learning: cognitive and computational approaches. P. Dillenbourg
(Eds.) Elsevier Science Ltd., Oxford, pp. 169-196.
Jaques, A. P. Andrade, A. F.; Vicari, R. M.; Bordini, R.; Jung, J (2002). Pedagogical agents to support
collaborative distance learning. Proceedings of Computer Support for Collaborative Learning CSCL2002, Colorado. http://newmedia.colorado.edu/cscl/275.html.
Kay, J (2001a). User Modelling for Adaptation. In User Interfaces for all: concepts, methods and tools.
Constantine Stephanides (Ed.), Lawrence Erlbaum Associates, Inc. EUA, pp. 271-293.
Kay, J (2001b). Learner Control. In User Modelling and User-Adapted Interaction, Kluwer Academic
Publishers, Netherlands, vol. 11, pp. 111-127
Lewis, R. (2000). Human activity in learning societies. Invited Paper. Proceeding of ICCE’00/ICAI,
Taipei, Taiwan.
Luckin, R. (1996). Trivar:Exploring the “zone of proximal development”. Proceedings of European
Conference on Artificial Intelligence in Education. In P.Brna, A. Paiva, J. Self (Eds.) Colibri,
Lisbon, 16-22.
Luckin, R., & du Boulay, B. (1999). Capability, Potential and Collaborative Assistance. In User
Modelling: Proceedings of the Seventh International Conference, UM99 J. Kay (Eds). Springer
Wien New York, pp. 139-148.
Moll, L. C. (1990). Vygotsky and Education: instructional implications and applications of
sociohistorical psychology. Cambridge University Press, pp 177-183.
Mora, M. C. (2000) Um Modelo Formal e Executavel de Agente BDI. PhD thesis, Porto Alegre:
CPGC/UFRGS.
Paiva, A.(1997) “Learner Modelling for Collaborative Learning Environments”, in Proceedings of the
AI-ED 97 World Conference on Artificial Intelligence in Education. B. du Boulay & R. Mizoguchi
(eds.), pp. 215-222.
Ragnemalm, E. L. (1996) Student diagnosis in practice; Bridging a gap. User Modelling and User
Adapted Interaction, 5(2):93-116.
Self, J. A. (1993). Model Based Cognitive Diagnosis In User Modelling and User-Adapted Interaction, 3,
89-106.
Self, J. A. (1994). Formal approaches to student modelling. In Greer, J. E. and McCalla, G. I., editors,
Student Modelling: The Key to Individualised Knowledge-Based Instruction, NATO_ASI Series F,
295-352. Springer-Verlag.
Vygotsky, L.S. (1978) Mind in society: The development of higher psychological processes. (M.Cole,
V.J.Steiner, S.Scribner, & E. Souberman, Eds.) Cambridge, MA: Harvard University Press.
Wood, D., Bruner, J.C., & Ross, G (1976). The role of tutoring in problem solving. Journal of Child
Psycology and Psychiatry, 17, 89-100.
25
A Web-based Medical Case Simulation for Continuing
Professional Education
Douglas Chesher*, Judy Kay#, Nicholas King*
*School of Pathology and #School of Information Technologies
University of Sydney, Australia 2006
dougc@med.usyd.edu.au, judy@it.usyd.edu.au, nickk@pathology.usyd.edu.au
Abstract:
This paper describes the Simprac system, a simulation-based tool to support long-term
learning by medical professionals and students in the area of management of chronic illness. Since it is
intended to support independent learning, one of the important aspects of its design is the inclusion of
several opportunities for learners to reflect on their knowledge and learning.
This paper gives an overview of the system, and the motivation underlying its design. We then describe
the opportunities it provides for learner reflection. One of these supports reflection at the end of each
simulated consultation, with comparison of the learner’s performance against that of experts and peer
groups. Another part supports reflection at the end of a whole case.
Keywords: life-long learning, reflection, medical education
Introduction
In the rapidly changing field of health care, like many other professions, medical practice involves
learning through the whole of working life (Brna P, 2000). Traditionally, this has involved modalities
such as rounds, educational meetings, conferences, refresher courses, programs, seminars, lectures,
workshops, and symposia (Davis D, 1999). However, in their review of the published literature on
continuing medical education, Davis et al (1999) found that traditional formal didactic continuing
medical education (CME) had little influence on physician behaviour and concluded that, "where
performance change is the immediate goal of a CME activity, the exclusively didactic CME modality
has little or no role to play". They did, on the other hand, find some evidence that, "interactive CME
sessions that enhance participant activity and provide the opportunity to practice skills can effect
change".
This has led us to explore the potential role of simulation-based teaching systems to support long-term
learning for medical practitioners. In the design of such system, we were influenced by the evidence
cited by Anderson, Reder and Simon (Anderson JR, 1995). This suggests that optimal learning occurs
when
a
combination
of
abstract
and
situation-specific
training
is
provided
and
that
abstraction
promotes the transfer of knowledge and thus insight from one situation to another. In the context of a
simulation-based learning system, this means that we should take care to go beyond a pure simulation
of one or more cases. We should ensure that the fundamental design of the system includes elements
which encourage the practitioner to reflect on each case, evaluate their performance in meaningful
ways and abstract broadly from the particular learning experience provided by each simulation in order
for knowledge transfer between similar or related cases to occur effectively.
While most computer simulations have been developed with an emphasis on medical diagnosis and
have generally involved a single patient encounter, much patient morbidity is associated with the
diagnosis and management of chronic disease such as diabetes mellitus and cardiovascular disease.
These chronic disorders evolve over time and involve multiple doctor – patient encounters. The
simulation we have been developing explores this chronic disease model by enabling the user / health
professional to review the patient over a number of consultations with the patient outcome for the
following consultation being defined by the management strategy chosen by the practitioner.
26
Within a simulation, there is considerable scope for various styles of tutoring. For example, one might
follow the model of Lajoie, Faremo and Wiseman (2001) in studying expert human tutoring as a
foundation for the design of a teaching system. We have taken a rather different approach in creating a
framework which can capture the diagnostic and management actions of a range of users, from novice
to expert. From this, we can construct models of the expert’s behaviour as well as a variety of group
models for learners at various levels, such as medical students and practicing general practitioners. The
merit of a range of such models is that each can serve different purposes. For example, all learners
would aspire to achieve expert levels of performance (and their patients would support this standard as
the goal!). However, this might be an unrealistic and unfair comparison for medical students who
would more reasonably compare themselves against their fellow medical students. It might also be
valuable for them to compare their actions against the model for general practitioners since senior
students should be aspiring to reach that level in the near term.
Our goal is to design a simulation that includes natural points for reflection. That reflection can be
based on the learner’s analysis of their actions. Importantly, it can also be supported by the presentation
of comparisons between the learner’s actions and those in the expert model or one of the group models.
Related Work
Patient simulations have been used in teaching as well as assessing “clinical competence”. Simulations
involving trained actors can provide the closest approximation to the real patient-doctor relationship.
These have the potential to provide both verbal and non-verbal cues and can be accurately depicted in a
standardised manner.
Actors have been used in research into medical problem solving by clinicians
(Elstein ES, 1978) but this form of simulation is clearly impractical for continuing education purposes.
It seems desirable to build simulation-based systems which can offer some of the advantages of such
human-actor-based simulations.
Computer-based simulations have become increasingly common for both teaching (Hayes KA, 1996;
Bryce DA, 1997) and assessment (Clyman SG, 1990). These are appearing as stand-alone simulations
available on a CD-ROM as well as on the web. To-date, the web-based cases (Hayes KA, 1996) have
been
considerably
(Friedman
CP,
less
1995)
sophisticated
has
provided
than
an
those
excellent
using
CD-ROM
outline
of
the
(Bryce
features
DA,
to
be
1997).
Friedman
considered
when
developing a computer based clinical simulation. Combining this with the description by Melnick
(Melnick DE, 1990) of the system used by the American National Board of Medical Examiners
(NBME), a minimum feature set can be developed.
Below is a compilation of the ideas and description by Friedman and Melnick.
27
Menu vs. Natural
Using natural language creates a higher fidelity simulation and avoids
Language Requests for
cueing the student but requires more sophisticated programming. On the
Data
other hand, some users may become frustrated with trying to make their
requests understood. With recent advances in computing power and
language processing offering a natural language interface is becoming more
practical.
Interpreted vs.
Again this can vary considerably between programs. One can provide the
Uninterpreted Clinical
raw results such as chest x-rays and pathology results without comment or
Information
one call provide a text report. Alternatively there may be some combination
of the two.
Deterministic vs.
With deterministic progression, taking the same action always leads to the
Probabilistic Progression
same clinical result. In the probabilistic approach each action is associated
with a set of probable outcomes and each outcome has a medically realistic
probability of occurring. While the latter approach has the potential for a
much richer simulation environment the initial development of the program
is far more complicated. With computer technology, it is now feasible to
generate cases from a knowledge base that stores information about the
prevalence of disease and the probability of specific findings in the presence
of that disease.
Natural Feedback vs.
Instructional intervention comes in at least two forms. The first is to provide
Instructional Intervention
feedback to the student on their progress through the case compared to some
desired optimum. The second method requires that the student relate
diagnostic hypotheses to clinical findings. This type of feedback has been
criticised for being distracting and imposing a reasoning framework that is
foreign to the student. Natural feedback involves the realistic progression of
the case. If appropriate action is taken the patient's health improves. In
contrast, if inappropriate action is taken the health of the patient will
deteriorate and may necessitate action to restore the patient to good health.
Scoring
Many programs assess participants by; determining the errors of omission
and errors of commission and derive indices from these.
Single vs. Multiple
Most computer simulations have been based on single patient encounters.
Encounters
Table 1 : Features of Computer Based Medical Case Simulations
We will refer to these in the description of the design of our system.
Overview of System
We illustrate the operation of our system in terms of a case we have developed. This is an example of a
reasonably challenging simulated task in diagnosis and management of hyperlipidaemia, which is a
major independent risk factor for the development of cardiovascular disease. An overview of the
consultation sequence for this first case is shown in Figure 1. The case has been chosen to require
management
over
a
period
of
several consultations which would typically run over six months.
Notably, there is a real possibility that plausible but incorrect treatment would cause serious problems
for the patient. This case is also interesting because it involves use of Fibrate, a drug which was once
prescribed only by specialist practitioners. Recent changes mean that it is now widely prescribed by
general practitioners. This is a classic case of the type of long-term workplace-based learning that
arises in the medical profession.
28
Red: No action or cornstarch alone.
Green: Fish oil alone or with cornstarch.
Initial
Blue: Statin alone or in combination with cornstarch or fish oil.
First
Presentation
Purple: Fibrate alone or in combination with any other treatment.
Consultation
Thick line represents preferred path.
First episode
Further
First episode
pancreatitis.
in triglyceride
No change
Predominant
reduction in
reduction in
cholesterol
in triglyceride
Mild reduction
from baseline
[1]
Predominant
Mild reduction
from baseline
[1]
Law suit. [2]
Mild Reduction
No Change
pancreatitis.
pancreatitis +
No Change
from baseline
in triglycerides
Second
Consultation
triglycerides
Predominant
Predominant
reduction in
reduction in
cholesterol
Predominant
Predominant
reduction in
reduction in
cholesterol
triglycerides
Side effect of
fibrate. [3]
triglycerides
Third
Consultation
Fourth
Consultation
[1] Pancreatitis occurs if there have been two consecutive consultations without effective treatment.
[2] Negligence if the patient has suffered pancreatitis and no effective treatment commenced.
[3] Pancytopenia develops if fibrate used on two consecutive consultations.
Figure 1: Overview of four consultation sequence in simulation
This case involves four consultations. In the first, the medical practitioner takes a patient history and
performs initial examinations. From these, a management action is proposed. For example, the optimal
management action calls for a Fibrate-based treatment. The second consultation, illustrated by the four
boxes in the second row, provides the practitioner with the outcome of their first management regime.
As Figure 1 indicates, it is possible at this stage to move from a sub-optimal situation to the optimal
one. (It is also possible to move away from the optimal situation.)
This case exemplifies the tightly
connected influence of the various management options with the ongoing disease outcomes. Most of
the states at each consultation can lead to any of the states in the subsequent consultation, depending
upon the treatment option chosen.
The diagnostic component is supported through the user being initially presented with a short case
vignette after which they have the option of
•
taking a medical history from the patient,
•
performing a physical examination,
•
ordering investigations, or
•
selecting management options.
Questions are asked using free text entry and a series of “matching” questions are returned, that once
selected, will elicit an appropriate response from the patient. The practitioner can perform a virtual
physical examination by selecting a variety of “tools” or actions and applying these to different parts
of the body (
Figure
2).
A
investigations,
wide
as
is
variety
often
of
the
investigations
case
with
can
clinical
be
requested,
practice,
will
not
however,
always
the
be
results
available
of
the
until
a
predetermined time later in the case. For example, the results of a CAT scan will not be available until
the next consultation.
29
Figure 2: Sample examination screen
Since an essential element of the system is to support learner reflection, there are two levels of review.
At the first level, the user is able to review their progress at the end of each consultation. This enables
the user to evaluate the questions they have asked, examinations, investigations and management
options they have chosen. They can simply review this as a basis for reflecting about what they have
done in this consultation. Our system has also been designed so that the user will be able to compare
their own performance with their peers or an expert in the domain. The peer data will be built up over
time as more and more individuals with that professional background attempt the case. The second
level of review occurs once a case is completed and enables the user to assess their treatment path and
patient outcomes through the case, again with the ability to compare their performance with both the
optimum, as defined by the authors of the case, and against the performance of their peers.
Summary and current status
We have described the motivation for our design of a simulation-based learning environment designed
to support the long-term learning of medical practitioners. We have also outlined the main elements of
the current prototype and provided a high level view of one case that has been implemented. There are
several distinctive aspects of the work. First, it tackles the domain of learning of long-term medical
management in the context of chronic disease. This includes the need to perform diagnosis but it also
requires the choice of treatment options which influence subsequent patient presentations. The second
distinctive aspect is the focus on learner reflection, with learners being encouraged to review their
performance and then compare it with various benchmark levels of performance: peer groups, more
expert groups and the approach considered optimal by the designers of the case.
The current prototype encodes the above case. However, it also implements generic elements for
building
further
cases. These include the initial consultation interface, a keyword-based natural
language understanding interface for collecting information from the history of the presenting illness, a
graphical interactive interface for performing examinations, menu-based interfaces for ordering tests,
recording tools for the learner to make notes about the case and aspects thereof, as well as to note
30
hypotheses. We still need to perform user evaluations and collect the data for the peer and expert
comparisons.
The current work represents an exploration of the ways that a simulation-based learning environment
can support long-term learning by medical practitioners. It is also an exploration of ways to support
reflection as a basis for abstracting the learning and for self-assessment and monitoring of learning
progress.
References
Anderson JR, R. L., Simon HA (1995). Applications and misapplications of cognitive psychology to
mathematics education.
Barrows HS (1986). “A taxonomy of problem based learning.” Medical Education 20: 481-486.
Brna P,
I.
P.,
Duncan
D
(2000).
Support
for
Lifelong
Self-Learners
Using
the
Internet.
Leeds,
Computer Based Learning Unit, Leeds University. 2001.
Bryce DA, K. N., Jeremy RW, Myers JH (1997). Facilitating clinical reasoning through computer
based learning and feedback. Australian society for computers in learning in tertiary education.
Clyman SG, O. N. (1990). “Status report on the NBME's computer-based testing.” Academic Medicine
65: 235-240.
Davis D, O. B. M., Freemantle N, Wolf FM, Mazmanian P, Taylor-Vaisy A (1999). “Impact of formal
continuing
medical
education:
Do
conferences,
workshops,
rounds
and
other
traditional
continuing education activities change physician behaviour or health care outcomes?” Journal of
the American Medical Association 282(9): 867-874.
Elstein ES, S. L., Sprafka SS (1978). Medical Problem Solving: An Analysis of Clinical Reasoning.
Cambridge, Harvard University Press.
Elstein ES, S. L. (1990). “Continuing medical education: A ten year retrospective.” Evaluation and the
Health Professions 13(1): 5-36.
Fitzgerald JT, W. F., Davis WK, Barklay ML, Bozynski ME, Chamberlain KR, Clyman SG, Shope
TC,
Woolliscroft
JO, Zelenock
GB
(1994).
“A
preliminary
study
of
the
impact
of
case
specificity on computer-based assessment of medical student clinical performance.” Evaluation
and the Health Professions 17(3): 307-321.
Friedman CP (1995). “Anatomy of the clinical simulation.” Academic Medicine 70(3): 205-209.
Hayes KA, L. C. (1996). “The interactive patient: a multimedia interactive educational tool on the
world wide web.” MD Computing 13(4): 330-334.
Lajoie, SP, S Faremo, and J Wiseman (2002). “Identifying Human Tutoring Strategies for Effective
Instruction in Internal Medicine.” IJAIED, 12, to appear
Melnick DE (1990). “Computer-based clinical simulation: State of the art.” Evaluation and the Health
Professions 13(1): 104-120.
Patel VL, G. G. (1993). “Reasoning and Instruction in Medical Curricula.” Cognition and Instruction
10(4): 335-378.
Patel VL, K. D., Arocha JF (2000). Conceptual Change in the Biomedical and Health Science Domain.
Advances in Instructional Psychology Volume 5. Educational Design and Cognitive Science. G.
R. Mahwah, Lawrence Erlbaum Associates. 5: 329-392.
31
Knowledge-based Fuzzy Evaluation of Learners in
Intelligent Educational Systems
1
2
3
Maria Grigoriadou , Maria Samarakou , Dionissis Mitropoulos , Michael
Panagiotou
1
2
4
Department of Informatics, University of Athens, GR-157.71, Athens, Greece,
E-mail: gregor@di.uoa.gr, Phone: +30 10 7275205, Fax: +30 10 7219561
Department of Energy Technology, Technological Institute of Athens, GR-122.61, Athens,
Greece,
E-mail: marsam@teiath.gr, Phone: +30 10 5385322,397, Fax: +30 10 5385306
3
Department of Informatics, University of Athens, GR-157.71, Athens, Greece
4
01-Pliroforiki S.A., 438 Acharnon Str., GR-111.43, Athens, Greece
Abstract: An empirical approach that makes use of fuzzy logic to evaluate the learners towards greater
learner control in the context of an intelligent educational system is presented. In this paper we propose
a fuzzy logic-based decision making model that is able to store and analyse uncertain information
regarding the various characteristics of the learner and evaluate his knowledge status, skills and
cognitive
abilities
and
reflect
back
the
contents
of
the
learner
model
to
the
learner.
Teachers’
experience is incorporated in the definition of the fuzzy sets and in the final estimation of the learner
characteristics using the knowledge-based system.
In our approach, the evaluation of learner's knowledge level, the existence, or non-existence of a
misconception and the estimation of his cognitive abilities, is based on processing qualitative and
quantitative information. Thus, the proposed evaluation procedure, employing methods from fuzzy
logic, evaluates the learner's progress, strengths and weaknesses by keeping learner performance
parameters. In addition, learner-computer interaction provides the evaluation procedure with several
measurements. The evaluation procedure aims at encouraging learners to engage in a variety of
reflective activities and provides a feedback of events back to the learner so that he can decide if he is
going to change his reasoning.
At the same time, the learner is allowed to inspect the content of his/her
learning model and, either follow the suggestions given, or decide to take responsibility for his/her own
learning.
Prototype educational software was developed in the following domains of Physics: Mechanics,
Reflection-Refraction, Heat, Electricity, Models and Atoms. The fuzzy model has been tested on this
prototype, in the domain of Mechanics (free fall of objects in the air or in vacuum) and the results have
been very satisfactory.
Keywords: Fuzzy model, intelligent system, physics, learner control
Introduction
The development of an Intelligent Tutoring System (ITS) requires the design of several modules that
dynamically interact to provide individualised instruction towards greater learner control over the
learning process (see Figure 1). A very important one is the module that evaluates the learner's
knowledge, his misconceptions and his personal characteristics. The results of this evaluation, along
with any background information, such as the learner's history, form a learner profile pattern upon
which pedagogical decisions could be based (Georgouli, 2001).
A typical criticism of the need for a detailed learner evaluation is addressed by Self (Self, 1990).
However, he suggests that, with realistic aims, intelligent methods for learner evaluation will play a
32
significant role in an ITS provided they are closely linked with tutorial interaction, obtain their input
directly from the learner and are open to teacher or learner inspection.
Figure 1: Schematic diagram of an Intelligent Tutoring System.
Fuzzy logic techniques have been used to improve the performance of an ITS due to their ability to
handle
imprecise
information,
such
as
learner's
actions,
and
to
provide
human
descriptions
of
knowledge and of learner's cognitive abilities (Kandel, 1992; Panagiotou and Grigoriadou, 1995). In
the BSS1 tutoring system a general fuzzy logic engine was designed and implemented to support
development of intelligent features, which can better manage the learner's learning (Warenford and
Tsao
1997).
Uncertainty
of
learner's
performance
in
Sherlock
II
was
managed
with
"fuzzy"
distributions (Katz et al. 1993). A qualitative learner model was designed using fuzzy logic techniques
for a tutoring system in the domain of physics inferring about learner's knowledge level and cognitive
abilities from learner's behaviour (Panagiotou et al. 1994).
In this paper we propose a fuzzy logic-based decision making model that is able to store and analyse
uncertain information regarding the various characteristics of the learner and evaluate his knowledge
status, skills and cognitive abilities and reflect back the contents of the learner model to the learner.
The
proposed
learner
model
combines
ideas
from
cognitive
psychology
with
methods
from
computational intelligence. Teachers’ experience is incorporated in the definition of the fuzzy sets and
in the final estimation of the learner characteristics using the knowledge-based system. The proposed
model has been incorporated in a prototype Intelligent Educational System and evaluated by a group of
teachers and students.
Extracting information for learner’s evaluation
A human tutor usually bases his pedagogical decisions on the information he collects regarding
learner's
knowledge,
capability.
The
same
beliefs,
mistakes,
information
misconceptions,
should
be
acquired
cognitive
and
abilities
analysed
in
an
and
ITS.
problem
To
this
solving
end,
the
educational program should present the theory in different ways and use questions and exercises,
organised into groups that allow us to evaluate the learner for specific knowledge, mistakes and
misconceptions (Mandl and Lesgold, 1988). More specifically, as suggested in
(Bertles, 1994), an ITS
should extract and associate information regarding:
•
The knowledge status: knowledge level, mistakes, misconceptions.
•
The skill status: reading level, audio-visual ability, handling numeric computations.
•
The
cognitive abilities: memory limitations, rate of learning, learning performance, attention,
synthesis, abstraction, and generalisation.
33
•
The meta-cognitive skills: the ability of linking together newly acquired with already existing
knowledge, understanding.
Information is extracted through learner's interrogation and monitoring, and can be used to formulate
performance patterns that contain qualitative and quantitative data. Questions, or groups of questions,
can be related to a subset of the domain knowledge that the learner should acquire and should have a
weight representing their importance or complexity as regards the evaluation of the knowledge and the
abilities of the learner (Nawrocki, 1987). For example, in order to facilitate information extraction in
our prototype implementation we adopted an organisation similar to the one proposed by Mandl and
Lesgold (1988).
For the implementation of the prototype, the following actions took place:
•
Analysis of the knowledge domain: The analysis was done in co-operation with two high-school
teachers (3
rd
and 4
th
rd
grade), and tested in five 3
grade and seven 4
th
grade learners, using the
textbooks of these grades.
•
Decomposition of the knowledge domain, e.g.
“free fall” in distinctive parts: free fall in the air,
free fall in vacuum.
•
Development
of
a
database containing different categories of questions and exercises, to be
presented to the learner according to the evolution of the lesson. Each question is related to all the
possible answers and reasons that a learner can give.
•
Recording possible misconceptions, and questions through which they were detected.
Learner’s model for the level of learning, misconceptions and cognitive skills
In our approach, the evaluation of learner's knowledge level, the existence, or non-existence of a
misconception and the estimation of his cognitive abilities, is based on processing qualitative and
quantitative information. Thus, the proposed evaluation procedure, employing methods from fuzzy
logic, evaluates the learner's progress, strengths and weaknesses by keeping learner performance
parameters. For example, the answers to the questions are compared with the typical answers and
related reasons stored in a database, providing information about the number of correct or incorrect
answers
expressed
as
linguistic
variable.
In
addition,
learner-computer
interaction
provides
the
evaluation procedure with several measurements. The evaluation procedure aims at encouraging
learners to engage in a variety of reflective activities and provides a feedback of events back to the
learner so that he can decide if he is going to change his reasoning. The time spent to answer the
questions and exercises, the number of learner attempts to find the correct answer to the questions and
exercises are analysed to provide information about the learner’s characteristics.
Figure 2: The three stages of the evaluation procedure.
The evaluation procedure is realised in three stages (see Figure 2) using a set of decision-making
systems, each one processing fuzzy information and evaluating a predefined learner's characteristic.
Below, we provide an overview of the proposed approach.
More specifically, a preliminary evaluation of the knowledge level and cognitive abilities is conducted
depending on the learner's answer. To this end, learner's answers are compared with the right answers,
recorded by the teachers in the ‘right answers data base’, and qualitative characterisations are assigned
to the result. Thus, for the learner's answers to a category of questions that are related to his rate of
learning
we
have
used
9
discrete
levels
corresponding
to
‘excellent,
good,
satisfactory,
almost
satisfactory, unknown, almost unsatisfactory, unsatisfactory, bad, very bad’ level of answers. This
34
evaluation
corresponds
to
a
discretisation
of
the
universe
of
discourse
according
to
the
above-
mentioned qualitative terms.
In order to evaluate and extract information about the learner’s knowledge level, misconceptions and
cognitive skills, we defined, for each distinctive part of knowledge, the following:
Ei
The categories of questions related to the distinctive part of knowledge under consideration.
Eij
The questions of the category Ei.
Aij
All the possible answers to the questions of the category Eij.
Ai
The
percentage
of
the
answers
in
the
category
of
questions
Ei,
according
to
certain
characterisations.
_i
The level of knowledge to be examined by each category of questions Ei.
_n
The misconceptions that the system can detect.
Ik
The skills that the system can detect.
M_
The measurements performed by the system.
Wik
The weight of the category of questions Ei, or the measurement M_ related to the level of
knowledge _i, or the misconception _n, or the skill Ik.
U1={y11, y12, …, y1n}
The discretisation of the universe of discourse of the knowledge level.
U2={y21, y22, …, y2n}
The discretisation of the universe of discourse of misconceptions.
U3={y31, y32, …, y3n}
The discretisation of the universe of discourse of the cognitive skills.
Modeling teacher’s expertise in assessing learner’s knowledge, as well as modeling teacher’s personal
way assessing, is based on the following resources:
−
The criteria that the teacher defines in order to assess learner’s knowledge level
−
Teacher’s
estimations
of
the
importance
of
different
types
of
assessment
questions
that
correspond to the above criteria, with respect to the learner’s knowledge level and the type of the
topic under consideration, i.e. a theoretical concept or a procedure.
−
Teacher’s
estimations
of
the
relationship
between
learner’s
correct
answers
and
his/her
proficiency of the topic.
The assessment of the knowledge level _i, or the misconception _ n , or the cognitive skills Ik,
is
achieved via the formation of certain fuzzy sets, which are obtained from the correspondence of the
learner’s answers and the measurements of the system to the experts’ fuzzy sets. The procedure for the
assessment of the learner’s knowledge level is the following:
When the learner answers all the questions E ij of a category Ei, the process of the answers provides a
classification of these answers in certain categories, based on the percentages of the type of the answers
(e.g. 30% sufficient, 50% rather sufficient and 20% insufficient answers). Suppose R the sufficient
answers, AR rather sufficient, AW rather insufficient and W insufficient. Suppose also __ a percentage
of the answers of the learner. Then we define 9 different categories of answers:
1.
R≥_1
2.
R≥_2 AND R+AR≥_3
3.
R≥_4 AND R+AR≥_5
4.
R+AR≥_6 AND R≥AR
5.
R+AR≥_7 AND R≥AR
6.
R+AR≥_8 AND AR≥R
7.
W+AW≤_9 AND AW≥W
8.
W+AW≤_10 AND W≥AW
9.
W+AW≥_11
The knowledge level _i can be represented as a fuzzy set of the universe U1 as follows:
_i = _1/y11 + _2/y12 + ... + _n/y1n
where _i is the membership function, _i ∈ [0,1] and y1i is a characterization.
35
Table 1 shows a table of fuzzy sets for the assessment of the learner’s knowledge level, obtained from
a certain category of questions.
Table 1:
y1
y2
.
yn
1
0
0.3
.
0.1
2
0.8
0.1
.
0
.
.
.
.
.
N
0.3
0.2
.
0.1
Table of fuzzy sets for the assessment of knowledge level for a certain
category of answers, where 1, 2, ..., N are the categories of answers
and yi are the characterisations of the knowledge level
An example of such a table is shown in table 2. If now we suppose that the answers of the learner
satisfy category 2 (R≥_2 AND R+AR≥_3), then the assessment of the knowledge level, for the certain
category of questions, is:
_i
=
0.9/(sufficient)+0.1/(rather
sufficient)+0/(medium)+0/(rather
insufficient)+
0/(insufficient)
For the final assessment of the knowledge level the fuzzy sets obtained from all the categories of
questions are taken into consideration. The assessment of misconceptions and cognitive skills follows
similar procedure.
sufficient
rather sufficient
medium
rather
insufficient
insufficient
1
1
0
0
0
1
2
0.9
0.1
0
0
0
3
0.6
0.3
0.1
0
0
4
0.1
0.6
0.3
0
0
5
0
0.2
0.7
0.1
0
6
0
0
0.3
0.5
0.2
7
0
0
0.1
0.6
0.3
8
0
0
0
0.3
0.7
9
0
0
0
0
1
Table 2: Assessment of knowledge level for a certain category for questions
Importance of different types of questions
For the assessment of the knowledge level, or misconceptions, or the cognitive skills of the learner,
there is a set of questions and measurements with different weights for each assessment. The proper
weight used in each case, is based on the algorithm of T. L. Saaty (Petrushin and Sinitsa, 1993) and a
table is created according to the following rule:
If the importance of the question E i or the measurement Mi with respect to the question E k or the
measurement M k is w then the importance of the question E k or the measurement M k with respect to
the question Ei or the measurement Mi is 1/w.
Description of the prototype
Experiments
have
been
performed
to
evaluate
the
performance
of
the
proposed
approach.
The
development of the hybrid model has been made for a prototype educational system in the following
domains of Physics (M.Grigoriadou, M. Samarakou et al 1999):
36
Heat:
-Expansion of solid materials, liquids, gases
Optics:
-Linear Propagation of Light, Shadow
-Change in physical condition
-Reflection and Refraction of Light
-Analysis and Synthesis of Light
Mechanics:
-Free Fall on the Earth
-Free Fall in vacuum
Electricity:
-Closed Circuits with an Electric Source and one or more Lamps/Resistors
-Electric Current and Voltage measurements
Atoms/Models:
-The closed circuit Electron-based Model
-The expansion Molecule-based Model
-Geometric Models for Shadow, Reflection and Refraction
-Newtonian Model for Free Fall
In this prototype the knowledge domains have been analysed for Greek learners of 3
secondary
education.
In
the
following
we
will
describe
the
actions
that
rd
took
and 4
th
place
grade of
for
the
implementation of the part of the prototype that deals with The Free Fall of objects.
Interface for the Free Fall of objects and recording of the students’ actions
This part aims at helping teachers to teach and students to learn about free fall of objects and the role of
the weight and the resistance of the air in free fall.
Research in the domain of didactics in Science has pointed out that students develop the following
misconceptions and difficulties while they try to study the fall of objects:
“Heavier objects fall faster”. Students that have this idea do not take under consideration the influence
of the resistance of the air to fall. They believe that heavier objects fall faster either in vacuum or in the
atmosphere (Viennot L. 1979).
“If there is no air, then there is no gravity – objects in vacuum have no weight”. This idea has its origin
to the everyday experience that an object doesn’t fall only when it is supported (Whitelock 1991), i.e.
when a force is applied by another object in contact. It seems that the wrong idea that forces are applied
between objects only when they are in contact, comes from this perception. This wrong idea leads
sometimes to the misconception that there are no forces applied in vacuum and the objects have no
weight – the air and the atmospheric pressure are responsible for the gravity force (Mayer 1987). The
misconception mentioned above is even stronger in “space”, i.e. when the object is not very close to the
earth (e.g. a satellite): “Objects in space have no weight”.
The example that follows illustrates the didactic approach and the interaction with the student, by using
the scenarios “Fall of objects” of the relevant software. Each task is carried out in certain steps, where
the activity of the student is recorded, as well as his/her choices. This data is used by the diagnostic
system in order to detect specific cognitive difficulties, misconceptions, wrong ideas or lack of
prerequisite knowledge. Consequently, this detection leads to the activation of an adapted instructional
strategy, e.g. suggestion to read a part of theory or accomplish another task, presentation of some
examples, etc. At the same time, the learner is allowed to inspect the content of his/her learning model
and, either follow the suggestions given, or decide to take responsibility for his/her own learning. The
above procedure is briefly described in the following.
37
Description of the interface:
Figure 3: A typical screen of the educational prototype system.
In the building on the right side of the screen (Fig. 3) there is a boy that can “let” an object free to fall
on the ground. The user can change the altitude, using the slider in the control window seen on the left.
S/he can also select one or more objects to fall (iron sphere, light or heavy ball). There are buttons to
control the motion (start, pause, stop, reset), synchronized with a stopwatch and a distance recorder.
Trails of motion, graphs of distance – speed – acceleration vs. time and the representation of the
relevant vector quantities are also available.
By clicking on the upper left corner of the image the user can “remove” the atmosphere and study the
fall without the resistance of the air.
The student can read in a separate window a description of the activities – tasks that she/he is asked to
carry out.
Detection of cognitive difficulties:
1. The student believes that heavier objects always fall faster
2. The student believes that if there is no atmosphere then there is no gravity force.
Description of a students’ activity:
Each activity is divided in three parts:
−
In the first part, the student reads a description of the problem / phenomenon and the relevant
actions to be done. At this point, the student is asked to predict what is going to happen.
−
In the second part, the student is asked to carry out the simulation of the relevant experiment
−
In the third part, the student is asked to compare his/her prediction to the results of the
and to record the results.
experiment.
38
The answers in the first and third part activate the instructional strategy that proposes to the student
either to revise a part of relevant theory, or to perform another task, or to observe some relevant
simulations/ examples, etc. By the same time, the student can be informed about his/her learner model,
through comments and remarks related to his/her actions and selections, and he/she can either follow
the proposed tasks, or choose to control the progress of his/her activity.
Example:
Objective: Study of the factors that affect the fall of an object.
In this series of activities you will study the fall of an object and the factors that affect it. You can
select different objects and let them fall from various heights, simultaneously, or one at a time. (You
can also “eliminate” the atmosphere from the earth an let the objects fall in vacuum.)
Activity: Free fall of two objects in the air
1. Prediction
If you let an iron sphere and a light ball fall simultaneously from the same height, which one will
reach the ground first? Check the correct answer.
0
The iron sphere (correct)
0
The light ball (wrong)
0
They will reach the ground at the same time (wrong)
Select one reason that justifies your answer.
Depending on the previous answer, possible reasons are presented for each case:
The iron sphere (correct)
0
o
It is heavier
o
The resistance of the air has less effect on the iron sphere (correct)
The light ball (wrong)
0
It is lighter and it moves faster
o
They will reach the ground at the same time (wrong)
0
All objects fall with the same acceleration
o
2. Experiment
Now, select the iron sphere and the light ball and let them fall simultaneously from the maximum
available height. Observe which of them reaches the ground first.
0
The iron sphere
0
The light ball
0
They will reach the ground at the same time
Is the result of the experiment in agreement with your prediction?
0
Yes
0
No
At this point, the following feedback is given to the student, according to his/her learner model:
In case of agreement between prediction and results, and depending on the reason selected in
prediction, the following are proposed:
−
Selected reason “It is heavier”. The student is asked to “remove” the atmosphere and
carry out the related activity.
−
Selected reason “The resistance of the air has less effect on the iron sphere (correct)”.
The student is encouraged to go to the next activity.
39
In case of disagreement between prediction and results, it is proposed to the student to observe some
real life falls (e.g. the fall of a sheet of paper and a book), observe the result and then repeat this
activity.
Representing Expert’s Knowledge
According to the way the program is structured, the expert part contains the following databases:
•
Knowledge database, containing the theory to be presented to the learner.
•
Questions/exercises/activities database, containing the following categories:
E1. Questions related to the knowledge of the relevant theory.
E2. Activities aiming at checking the understanding of theory, the ability of the learner to
solve equations and arithmetic formulae, and the ability to perform math operations.
E3. Questions aiming to test the ability of the learner to associate new knowledge with real life
situations, and to detect possible misconceptions.
•
Answers database, which contains all possible variations of the answers that the learner can
give to the proposed questions/exercises.
•
Messages database, containing messages to the teacher for problematic situations that the
system cannot handle.
•
Instructions/explanations database, containing messages to the learner to be used during the
interaction with the ITS. The selection of the appropriate message depends on the recorded
“qualitative model of learner’s knowledge and cognitive abilities”, or on the choices made by
the teacher, or even on the choices made by the learner - if the teacher permits this.
Exploiting learner’s record
All the evaluations related to the level of knowledge, misconceptions and cognitive characteristics are
recorded every time the learner goes through a distinctive part of knowledge. Furthermore, the topic,
the way of presentation, the number of questions/exercises and the sequence of these questions are also
recorded. These recordings compose the events in this incident of the distinctive part of knowledge,
which belongs to the entire learner’s record, i.e. the knowledge that the system has about the learner.
An example of such a recording of the events of an incident, during the instruction of free-fall is:
12/04/2002, free-fall, presentation of theory, five questions related to the kind of motion and the topics
presented
in
theory,
knowledge
level
of
theory
medium,
four
questions
related
to
possible
misconceptions about free fall, no misconception detected, …., knowledge level rather satisfactory,
ability in maths average, rate of learning normal,
no misconceptions.
Learner’s record, during the presentation and exploitation of a distinctive part of knowledge, is mainly
responsible for the activation of the counteractive rules included in the educational software and the
adaptation of the instructional strategy, during the learner - computer interaction, within an incident.
The recorded information is responsible for the way of initiation, of the program and the handling of
repeated problematic situations. The program also uses the recorded information, so that it can exclude
from further use approaches that were determined to be ineffective.
If now the learner does not work out successfully the presented distinctive part of knowledge, then this
ineffective approach is recorded, and the rules of the pedagogical part are activated. The part is
repeated, until the learner finishes it successfully, or stops the program.
Learner is always informed about the content of his/her record. In case that he/she prefers to deactivate
the adaptation of the system, all the available topics are presented to him/her and he/she becomes
responsible for his/her further learning process.
40
Conclusions
The learners and teachers that participated in the analysis of the knowledge domain evaluated the
prototype (Group 1). Three more teachers and their learners (Group 2) also evaluated the prototype by
using it in two lessons during the instruction of free-fall. Altogether, 5 teachers and 27 learners
evaluated the program.
There has been a growing appreciation of the importance of purposeful learner control of the learners
that need to be able to access and control them (Kay J. 1995, 1997). With regards to detecting learner’s
misconceptions a success of 92% was achieved while the interaction with the system improved their
knowledge. The experiments with the prototype verified that, by evaluating learner interaction it is
possible for the system not only to adapt its educational strategy, but also the degree of learnercomputer
interaction,
provided
of
course
that
this
option
is
supported
by
the
system.
Adaptive
educational multimedia technology offers such possibilities, since it supports sound, video or animation
and virtual laboratories.
Moreover, the system responded to the anticipations of the teachers with respect to the estimations for
their learners. These estimations were related to the evaluation of the level of knowledge and cognitive
abilities of the learner, provided that the same answers were given in the class environment. The
approach presented in this paper is an empirical one: the learner's evaluation depends on the designer's
ability to analyse the knowledge domain suitably, define fuzzy variables and appropriate membership
functions for their fuzzy sets, and relate learner response with appropriate knowledge and cognitive
characteristics.
The evaluation procedure is closely related to the knowledge scheme adopted for the representation of
the
particular
knowledge
and
cognitive
characteristics
of
the
learner
and
can
be
used
for
the
development of a qualitative model of the learner's knowledge and cognitive abilities. The model
response can further be used for deciding about the appropriate teaching strategy. At the same time, the
learner is allowed to inspect the content of his/her learning model and, either follow the suggestions
given, or decide by his/her own to take responsibility for his own learning.
The proposed hybrid approach realises a qualitative model of the learner, which stores and analyses
information about the knowledge status and the cognitive abilities of each learner. We currently
investigate techniques to enhance the percentage of success of our model by exploiting the training and
generalisation capabilities of the neural networks to extract information from learner profiles that can
further extend the applicability of the evaluation procedure.
References
Bertles K. (1994). A dynamic view on cognitive learner modelling in computer programming, Journal
of Artificial Intelligence in Education, 5, 1.
Georgouli
K.
(2001).
Modelling
a
Versatile
Mathematical
Curriculum
for
Low-attainers.
In
Proceedings of the 8th Panhellenic Conference in Informatics, vol 1, Nicosia/Cyprus (2001)
463-472.
Grigoriadou M., Samarakou M., Mitropoulos D., Rigoutsos A., Stavridou E., Solomonidou C. (1999).
Vectors in Physics and Mathematics, Conference on Technology and Education, International
Conference on Technology and Education, ICTE 99, March 1999, Edinburgh, p.71-73.
Kandel A. (1992). Fuzzy expert systems, CRC Press, 1992.
Katz S., Lesgold A., Eggan G. & Gordin M. (1993). Modelling the student in Sherlock II, Journal of
Artificial Intelligence in Education, 3, 4, 495-518.
Kay J. (1995). The UM toolkit for cooperative user modelling. User Modelling and User-Adapted
Interaction, 4, 149-196.
Kay J. (1997). Learner know thyself: a student models to give learner control and responsibility. In
Proceedings of ICCE’97 Kuching, Australia.
King R.E., Magoulas G.D. & Stathaki A.A. (1993). Multivariable fuzzy controller design, Control
Engineering Practice, 2, 431-437.
Mandl H. & Lesgold A. (1988). Learning issues for intelligent tutoring systems, Springer.
41
Nawrocki L. (1987). Maintenance in training. In Kearsley (ed.), Artificial intelligence and instruction:
application and methods, Addisson Wesley.
Panagiotou M. and Grigoriadou M. (1995). An application of fuzzy logic to learner modeling.
In
Proceedings of the IFIP World Conference on Computers in Education, Birmigham.
Panagiotou M., Grigoriadou M. & Philokiprou G. (1994). Knowledge and learning learner model. In
Vosniadou S, Corte E & Mandl H (eds), Technology-Based Learning Environments, NATO
ASI series F., Springer, 243-249.
Petrushin V., Sinitsa K. (1993). Using Probabilistic Techniques for Learning Modelling. Proceedings
of (AI-ED93) World Conference on Artificial Intelligence in Education, Edinburgh 1993.
Saaty T.L. (1978). Exploring the interface between hierarchies, multiple objectives and fuzzy sets.
Fuzzy sets and systems, 1, 57-68.
Self (1990). Bypassing the intractable problem of learner modelling. In Frasson and Gauthier (eds.),
Intelligent tutoring systems: at the crossroads of artificial intelligence and education, Norwood,
NJ: Ablex.
Viennot L. (1979). Spontaneous reasoning in elememtary dynamics. Eur. J. Sci. Ed. 1,205-221 (1979).
Warendorf K. and Tsao S.J. (1997). Application of fuzzy logic techniques in the BSS1 tutoring system.
Journal of Artificial Intelligence in Education, 8, 1, 113-146.
Wenger,
E. (1987)
AI
and
Tutoring
Systems:
Computational
and
Cognitive
Approaches
to
the
Communication Knowledge, California: M. Kaufmann Publishers, Inc.
42
Modeling the Learner Preferences for Embodied Agents:
Experimenting with the Control of Humor
Aude Dufresne, Martin Hudon
Department of Communication, University of Montreal.
C.P. 6128 Succ Centre-Ville, Montreal, Qc, Canada, H3C 3J7,
dufresne@com.umontreal.ca
Abstract:
This
research
presents
a
model
and
an
experiment
on
the
integration
of
personality
preferences in support systems for learning. We will present briefly the context of the research on the
access to learner and group models, than the theoretical background on personalization of interface and
more specifically of the functions of humor in general and how it can be used to integrate affective
dimensions in tutoring interaction. This research stems partly from the Reeves and Nass [1] postulate
that people will react the same way to a computer mediated interaction, then they do to a real
interaction, and thus that their reaction to an humoristic tutor would be generally more positive than to
the non-humoristic tutor: the tutor attracts more attention, the perceived usability, social presence and
personalization are improved. We will present the design and experimentation of an open model
support system, where feedback is given to learners on their progression, but also where preferences
for support may be defined. It was experimented under two conditions “with” and “without humor” and
qualitative attitudes measures where taken. Though results are only preliminary, this study of the
impact of humor suggests various considerations on how personality aspects can be integrated and their
impact studied in ITS systems.
Keywords: Personalization, embodied agents, humor, evaluation, distance education.
Context of the research
This research stems from a preoccupation for interface and communication design in ITS especially in
the context of distance education, in very open domains of learning. To design appropriate interaction
in the context of learning it is necessary to have access to rich tasks and learners models. Thus the
majority of ITS models depend on a highly detailed representation of the domain to be learned, where
knowledge can be tested at micro level and inference can be made amongst knowledge elements. It is
then possible to design natural language understanding, specific explanations or deitic demonstration
for a given problem or scene layout. Our challenge was to generalize the principles of support, to make
them accessible when the models of the domains are more shallow; when the programming of support
has to be done by a professor with no special training. In such context, evaluations cannot be as
detailed and more self assessment is necessary to access the learner’s model.
The support functions were to be part of a course editor. Principles that could be applied to extract
constraints in tasks and to give advice; to present demonstration and suggest content element using a
generic editor. In order to enrich the contextual model and also the means of support, we developed a
dynamic and adaptive interface in which we experimented various dimensions of personalized support.
ExploraGraph a dynamic interface to support the task and learner models
The
ExploraGraph© Navigator (figure 1) was designed to facilitate the visualization of learning
structures, that present task scenarios, knowledge or document structure. A conceptual model, with
typed nodes and typed links, is used to simulate the relationships between the elements and organize
the global representation to serve as a front to the course content. The graphs follow a physical model,
where links express semantical relations among elements, and which reacts with zoom and fish eye
effects when they are explored by the learner (see previous description [2-4]).
43
Figure 1:
ExploraGraph© Navigator, showing the graphical structure of activities with
individual and group completion levels and the Hacker MsAgent contextual help.
As in other kinds of maps, adaptive annotations were used to give the learners feedback on their
progression in the content, how often nodes have been visited. They can also edit their learner model
by editing the degree of "completion" for each node displayed in graphs. The learner's model is thus
presented as an overlay on the task's model, in structures of activities or concepts, giving feedback to
what as been explored (visited), and what the learner consider as being finished. The hierarchy
amongst nodes makes it possible to de propagate, both exploration and completion levels in the learner
model .
Individual learner models are kept both locally and on a database server, so the learner may access his
profile from different computers (e.g. home or university). It is thus possible to compile group models
so a feedback can be given on the activity of the class. Feedback on the group of learners [5] can also
be used to motivate the learner as a passive feedback or to support more active type of feedback (e.g.
“ All the others have finished”). Isolation is a problem in distant learning and graphs can become a
transparent way to provide learners with information on what is happening to the group. Thus, learners
can also display the levels of "visit" and "completion" of fellow learners. The postulate was that this
visibility and easy access to learner models and group models would encourage learners to update
them, thus improving the information the ITS system was using.
In ExploraGraph©, since graphs are dynamically generated, further adaptive functions may be added,
so graphs are contextually arranged in order to cue the learner toward more relevant areas, using zoom
functions (this corresponds to Brusilovsky's [6] principle of maps adaptation). Other support functions
are
included
in
the
environment,
like
MsAgents
avatar
animations
or
messages,
control
of
the
environment when the learner specify intentions. Microsoft Agents were introduced in the environment
partly “to fulfill the need for social context” when no other learner is on-line, but also to serve as the
interface to the advisory system. They are driven by the rule-based support system as defined in the
database.
We evaluated a first version of the system with 9 learners over a 6-weeks course. We used observation,
trace analysis and questionnaires to evaluate the usability of different modalities of help in the
environment [7]. We found, as expected, that the pacing of the help was critical, that physical (force
feedback)
guidance
seemed
to
provide
better
retention,
and
that
prolonged
support
improved
motivation.
44
Comments (video taped observation, focus group and open questionnaires) showed a lot of variation in
the preferences of learners for support: modality, animation of graphs, agents, timing and other
personality aspects of support. It appeared important to introduce more parameters in the learner model
linked to his preferences for support. It was important to investigate the attitudes toward various
personalization factors, that could be taken into account while defining the rules. Those personalization
factors should eventually appears as ways for the learner to control his environment (adaptation) or
eventually for the system to adjust to the learner’s reactions (adaptive). We are planning to give the
learner the possibility to choose from various coaches, with different personalities, to ask more or less
support and finally to choose whether he want humor or not. We wanted to experiment and evaluate the
impact of those personalization factors separately. We describe here an experiment on the attitudes
toward humor used by embodied agents in pedagogical interaction. This was a controlled experiment
were humor was either present or not present, we wanted to investigate how learners reacted and
whether that could be an interesting variable to include in an ITS.
Embodied agent and the personalization of support
Embodied agents are a seen as an interesting paradigm in AIED to support learning, because they offer
a better integration between personalization and the spatial context of explanation [8-11]. As the term
“embodied” suggest, they are more expressive and thus appear to be more effective to simulate real
tutorial interactions. Embodied agents can be used to attract attention, guide and demonstrate using
deitic gestures, suggest emotional context using the expression of emotion.
Studies were made on the integration of affective dimensions in embodied agents. As Elliott, Rickel, &
Lester [12] suggest, the affective dimensions are important, because they make the coach appear to
care about the learner, to be with him, and because it may communicate enthousiasm about the task.
Okonkwo and Vassileva [13] also studied the impact of having coaches express emotions while giving
advices. They found that non-verbal expression of emotions did not influence performance but had a
positive effect on the perception of the help. But it is difficult to make an agent to be emotionally
appropriate in his interaction. To do so the system’s model should be more elaborate to better sense the
learner emotional state, and to better react to it, both at the content and emotional level, “bridging
between sensory input and action generation” [14]. As Cassel and Thórisson [15] have found it is not
as much emotions per se, as the envelope role of non-verbal expressions that accompany dialogue, that
are important to give a lifelike impression and to ensure fluidity in interaction.
But emotion is but one aspect of embodied agents. As André et al. [16] affirmed “the next major step in
the evolution of interfaces is very likely to focus on highly personalized interfaces”.
So the new
undergoing challenge in interface design in general and in ITS is to try to enrich the support model
with
more
dimensions.
human
But
like
the
dimensions,
integration
of
incorporating
personalized
aspects
of
embodied
personality,
agents
poses
affective
the
and
problem
social
of
the
investigation, the simulation and the evaluation of complex dimensions of affective and personality
aspects of learning. Inside the ITS, the personalization of support appears to be an enrichment of the
communication model, that uses the specifications of the context of the activity of the learner, but more
importantly his learning model and preferences, to intervene or shape the communication model.
The integration of personalized support in education appears more and more essential on a pedagogical
point of view, but also to make supportive agents more believable and trustable. As André [8]explains:
“A growing number of research projects in academia and industry have recently started to
develop
lifelike
agents
as
a
new
metaphor
for
highly
personalised
human-machine
communication. A strong argument in favour of using such characters in the interface is the
fact that they make human-computer interaction more enjoyable and allow for communication
styles common in human-human dialogue”. She presents the Presence system which “uses
affect
to
enhance
the
believability
of
a
virtual
character,
and
produce
a
more
natural
conversational manner”.
In this context we thought that humor could be an interesting dimension to integrate [17], first because
of its potential for creating more personalized embodied agents, displaying emotions and social
personalities depending on the context. Just as the simulation of affective reactions might make
45
embodied agents more believable, we postulated that humor might add to the impression of intelligence
and complicity of supportive agents and should make them more acceptable to humans. Humor could
be an efficient communication strategy used to attract attention, diminish stress or stimulate affective
and motivational reactions of learners. But humor is highly tinted with personality aspects and thus is
difficult to integrate in a learning environment.
In fact, the development of personalization dimensions asks for new methodologies of research and
evaluation
in
AIED.
Should
the
ontology
be
defined
theoretically
and
its
usefulness
assessed
empirically ? Should it be extracted from the observation of interaction in comparison to pretest or
posttest of psychological dimensions?
Experimenting with the control of humor
The perception and impact of humor in communication
Humor is similar to a game, it is possible only when the participants are capable of a certain degree of
meta-communication, stating that “this is a game”. As Eco [18] describes it breaks the links between
signs and signification, introducing an incongruity between them and thus a second level meaning. This
signification game mixes the expectations of the receiver, which experience first an interrogation and
then surprise when he is confronted with the unexpected meaning.
Humor has been said to increase the perceived social presence in a medium. According to Lombard
and Ditton [19, p. 9]
“The presence is [20] the perceptual illusion of non-mediation […] occurs when a person fails
to perceive or acknowledge the existence of a media during a technologically mediated
experience”. The definition suggests that this illusive experience is at the perceptual, cognitive
and emotional level of the user interaction. According to them the media is not only perceived
as transparent but it naturally suggests the possibility to support and simulate “real” social
interactions.
According to Short et al. [21] a high degree of social presence is important for specific types of
interactions, for example persuasion and problem solving are difficult when the level of social presence
is low. In a context where a learner has difficulties, and where the advisory system tries to influence
him or to stimulate his motivation, social presence might be especially critical. ¨For Biocca [22] the
perceived “social presence” is linked to the perception of intelligence and of intention expressed
through the mediation of the artefact.
Theories on humor may be grouped under three dimensions: the superiority theory (humor presumes
and places the receptor into an inferior position), the relaxation theory, the incongruity theory. Those
dimensions of humor may each be used for a specific purpose in the context of tele-learning:
to exert authority to bring students into a different behavior;
to unleash tensions associated with learning or with the tutoring interaction;
to destabilize students and provoke new understanding.
But humor is accompanied by a high level of noise in the interaction. The intent meaning might be
unclear, the underlying model of social interaction might be inappropriate to the context, or to the
learner personality. Humor is highly cultural, and its meaning is negotiated as the interaction evolve
between participants. If humor is to be used, some coherent common codes must be developed between
the system and users; some means of communication must be designed so users may understand, learn
and adjust the models that lay behind the system; and so the system can be influenced by the reaction
and preferences users have for humor. It is important to diminish distance and noise in the supportive
interaction.
So
the
messages
the
system
is
giving
are
understood
and
efficient
in
promoting
understanding and efficient learning on the part of the participant.
46
Possible impact of using humor in an ITS
What could be the impact of using humor in ITS ? Research on embodied agents and research on
humor suggest that humor might make embodied agents more believable; that it may make the
computer look more intelligent, since it would seems as though he not only communicate, but also
metacommunicate
about
the
situation.
Humor
might
help
alleviate
tensions
associated
with
the
interaction with a computer, with the isolation and stress of distance education. Finally, in a way, the
noise
in
communication
associated
with
humor
might
be
a
way
to
hide
or
dilute
inappropriate
interventions of the ITS.
Methodology and hypothesis
What could be the impact of humor and how is it possible to experiment using it in the context of ITS ?
As Reeves and Nass [1] proposed, it is possible to experiment interaction with computers the same way
we are evaluating real interactions, in this case having a condition ”with” and a condition “without
humor”, and comparing impact on the usability and attitudes toward the system. But if we generalize
the objectives of the system which were to integrate personalization of the interaction in the system; it
is important to integrate personalization parameters of the support system, which could affect the
support system, and eventually be controlled either with direct adaptation by the learner (I want more
humor) or by adaptive adjustment by the system (He does seem to like humor). Though the system was
designed so the learner could control it using the control panel, in the context of this research the
conditions were fixed by the experimental set up, ie the learner could not change them.
As a first step we transformed the ExploraGraph rule editor so it would be possible to define rules
“with” or “without humor”. We experimented the system in the context of a course on learning the
Flash software with undergraduate students at University of Montreal. The experiment was reduced to
a three hours period, where the students were to explore the content, do some exercises and then pass a
small test. Support rules were designed both with and without humor, using the different dimensions of
humor – incongruity, superiority, relaxing, etc. The design was a split group experiment, where half of
the subjects had first a version with humor and then without, and the other half started without humor
and then with humor; the switch was time based and blinded to them. Twelve subjects participated in
the experiment but only eight filled both questionnaires and were kept for the analysis.
Two questionnaires (after each experimental period) and interviews were used to collect data on
attitudes of learners. As suggested by research on humor, the hypothesis were that humor would bring a
more favorable evaluation of the support system, having more “social presence”; that agents would be
perceived as more sensible, intelligent, more credible.
Results
Though we cannot use the results to confirm our hypothesis, because of the limited number of subjects
and the short duration of the experiment. But they can serve as indication not only of the possible
impact of humor on ITS, but also on the complexity of its interweaving with personality and context.
In general the learners had difficulty getting to know and use the ExploraGraph environment in such a
short period. They found the graphs very different from what they are accustomed to, like regular
hierarchical hypertext pages. In fact even with simple hypertext course, we had found that students
suffered from the “lost in hypertext” syndrome and that they lack directions on how to organize their
learning activity [7, 23]. In a previous experiment ExploraGraph had been used for 6 weeks, and it had
taken some time, before learners got accustomed to it, to understand the semantic of the links and to
organize their activity using it. Also in this experiment, the students had to pass a test at the end, and
this might have influenced their evaluation of the environment and of the agents’ support.
In general the evaluation of the agents was better for the humorous version and the appreciation of the
agents was lower for the second evaluation (see Table 2). In fact, the difference between the first and
the second evaluation was more important when the learners passed from an humorous to an non
humorous condition (3.9 to 3.3 vs 3.6 to 3.5).
47
First
Second
evaluation
evaluation
Group
Humor
No Humor
Mean
Group
Mean
Mean
SD
G1
3.9
G2
3.5
3.7
0.99
G2
3.6
G1
3.3
3.45
1.08
Mean
3.75
3.4
SD
1.03
1.03
Table 1: Mean attitudes and standard deviation toward the agents in the
support system forconditions with or without humor (scale 1 to 6), for the
two groups at the first and second evaluations.
Figure 2 presents the agglomeration of results to the questionnaires grouped by category according to
the hypothesis. In general the attitudes toward the agent were more positive, when the condition was
“with humor”: the agent was found to be more funny, more clear, more intelligent; it was found to
attract more attention, to be more relaxing. As for attributes associated with the “social presence”
dimension, the humorous agent was found more friendly, sensible, expressive, and social; though he
was found less personal (but the meaning of this question might not have been understood clearly
(more personality vs more personalized to learner).
On the contrary, the humorous version of the agent was found not to be as good a support to
motivation, orientation and learning. This might be due first to the stress associated with the test at the
end of the experimentation. It might also be due to the very short experience with the course and the
agent - only three hours, which limited the number of possible support interventions. It would be
interesting to see if the attitudes would be the same for a course lasting many weeks, when the isolation
and motivation might become a problem.
48
toward agents
Believable, intelligent
attract attention
Relaxing
Like humour
friendly
Presence
Attitudes
Social
Clear, appropriate
sensible, expressive
social
Learning
personal
Support motivation
Support learning
Orientation, guiding
0.00
1.00
2.00
3.00
Mean
4.00
5.00
attitudes
Humor
No humor
Figure 2 – Mean attitudes toward the agent and the help provided in the ITS
for conditions with or without humor on a scale from 0 to 6.
Discussion
In general results and interviews showed that the appreciation of humor was much dependent on
context: How difficult is the course and the learning environment? Whether there is an evaluation at
stake. Specific situations where the agent interventions where out of place.
While some students liked
the agent, other expressed reserves on his personality, they did not like some of his remarks which they
found inappropriate. The agent, the “Hacker”, had been designed to address “resistant” students [24];
his humor was found by some students to be aggressive (using a superiority strategy, the agent was
teasing the student), his relaxing and incongruous behavior was also found misplaced, by some
students since there was an evaluation and the students were stressed. Some students expressed the
need to stop the agent at one point or they wished they could have chosen another one, when proposed
so at the interview. So, even though the theory and general empirical results suggest humorous agents
might preferred to non-humorous one, in some conditions and for some students they were found
disturbing. In general students wished they could have more control on supportive interventions.
Another interesting result, was that humorous interventions were perceived as less supportive for the
task, orientation and motivation. Even though this might be linked to the very short experiment, it
might be important when adjusting the degree of humor to take into consideration both task and
personality factors. May be keep humor for when a task has just been completed (reinforcement) or no
task is urgent (beginning of the course).
This experiment on humor is part of a more general research, where students will eventually be able to
choose the personality of the supportive agent. Not only would they be able to choose humor or not,
more
or
less
support,
but
they
will
be
able
to
choose
amongst
a
set
of
coaches
with
different
personalities. Following Martinez [24] research on learning styles, we will offer them four archetypal
coaches designed to match the four kinds of learners – transforming or intentional, performing,
conforming, and resistant. She describes how individuals follow a complex mix of beliefs, desires,
emotions, intentional effort, and cognitive and social styles to learn, which must be taken into account
49
6.00
by the supportive environment. In fact learning styles and humor theories suggest ontologies for
preferences, which may be included in general rules of support systems. For example for intentional or
resistant learner, incongruity type of humor might be more appropriate; for p e r f o r m i n g learner
superiority type of humor might be more supportive. So the apparent personality of the coach can be
used to represent a style of support – timing, parameters of the situation where help will be offered,
content of the advices. We had designed four coaches for the different type of learners, but only one the
Hacker was used in the context of this research. Eventually with multiple coaches, we will experiment
the differential models of personalities of help in relation to style of learners, in order to measure the
spontaneous use of the coaches, and their impact on attitudes toward help (perception and reaction to
help).
Conclusion
Even though research on embodied agents with models of affective reactions are interesting, few
evaluations have been made of their acceptance by learners in real educational context. More so, it
appears interesting to study how we can design general models to use them in context where the
knowledge of the domain is limited to interdependencies in tasks or concepts, like what is described by
Paquette & Tchounikine. [25]. In this direction, the ExploraGraph system was built to externalize the
structure of the task and the learner and group models and to facilitate its access to learners. In it,
support
rules
may
be
designed
to
use
adaptive
interface
and
MsAgent
animations
and
advices.
Parameters were added in the support system to take into account preferences of the learners (humor,
chosen coaches, level of support). We have used it to make this control experiment on the attitudes
toward humor in embodied agents.
Though with only a limited number of subjects, we found that attitudes toward agents displaying
humor was generally more positive, that it makes them appear more intelligent, sensible, believable,
that their social presence is higher. We also found that embodied agents with humor were more
distracting and not perceived as being as good support for learning. As Eco[18] suggests, the learner
must look twice to understand the second level in humorous intervention. This appears to attract his
attention, but also to distract him from what he is doing. The learners reacted negatively to such
disturbance.
Learners comments suggested that their affective reaction is highly dependent on their personality and
that of the agent. More research is needed to describe in more details how humor and personality could
be linked to define situations and actions to support learners based on cognitive or learning styles [24].
It is also important to analyze and model the control and reaction of learners to the help provided and
to compare this to their attitudes toward help and the justification they see for preferring one coach
over the other. An environment where the learner could control the degree and style of coaching would
be interesting to study, but it must rely on strong and generic theoretical models of coaching, linking
the
diagnosis
of
situations
and
the
types
of
support
actions,
to
generic
models
of
tasks
as
in
ExploraGraph [2].
It would be interesting to do observation and to ask learners to characterize the personalities of the
coaches
and
their
interventions
to
precisely
understand
the
reaction
of
learners
to
different
personalities, and types of humor in agents. We may also find gender-based differences in the use and
control of the support system as in Okonkwo and Vassileva [13]. But letting the learner adjust the
support preferences is not enough, since even though it is important to let the learner control his
environment; it might be a tedious task for him. So we intend to integrate in the environment adaptive
features taking into account reactions to support (agent is stopped, advice are not followed). So general
rules could be adjusted to students in general and to a specific learner using learning mechanisms.
References
1.
Reeves, B. and Nass, C. The Media Equation. CSLI Publications, New York, 1996.
2.
Dufresne,
A.,
ExploraGraph
:
Improving
interfaces
to
improve
adaptive
support.
in
AIED'2001, (San Antonio, Texas, 2001), IOS Press: Amsterdam, 306-313.
50
3.
Dufresne, A., Model of an Adaptive Support Interface for Distance Learning. in ITS'2000,
4.
Dufresne, A., Cosmova, V., LeTran, T. and Ramstein, C. An interface to support the learner
(Montréal, 2000), Springer Verlag, 334-343.
with dynamic graphs and multimodal goal driven explanations. in Lajoie, S.P. and Vivet, M.
eds. AIED'99, IOS Press, Amsterdam, 1999, 660- 662.
5.
Dufresne, A. Conception d'une interface adaptée aux activités de l’éducation à distance -
6.
Brusilovsky, P. Methods and Techniques of Adaptive Hypermedia. User Modeling and User-
7.
Dufresne, A., Interface design for Distant Education : Experimenting Different Environments.
ExploraGraph. Sciences et Techniques Éducatives, 8 (3) 2001. 301-320.
Adapted Interaction, 6 1996. 87-129.
in EDICT'2000 International conference on information a communication technologies for
education, (Vienne, Austria, 2000), 363-372.
8.
André, E., Klesen, M., Gebhard, P., Allen, S. and Rist, T., Integrating Models of Personality
and Emotions into Lifelike Characters. in workshop on Affect in Interactions – Towards a new
Generation of Interfaces in conjunction with the 3 rd i3 Annual Conference, (Siena, Italy,
1999), 136-149.
9.
Nkambou, R. and Laporte, Y. (eds.). Producing Non-Verbal Output for an Embodied Agent in
10.
Rickel, J. and Lewis Johnson, W. Task Oriented Collaboration with Embodied Agents in
an Intelligent Tutoring System. Springer-Verlag, Berlin, 2001.
Virtual Worlds. in Justine Cassell, J.S., Scott Prevost and Elizabeth Churchill editors ed.
Embodied conversational agents, MIT Press, Cambridge, Mass., 2000.
11.
Johnson, W.L., Rickel, J.W. and Lester, J.C. Animated Pedagogical Agents: Face-to-Face
Interaction
in
Interactive
Learning
Environments.
International
Journal
of
Artificial
Intelligence in Education, 11 2000. 47-78.
12.
Elliott, C., Rickel, J. and Lester, J.C., Integrating affective computing into animated tutoring
agents. in IJCAI Workshop on Animated Interface Agents: Making Them Intelligent, (Nagoya,
Japan, 1997), 113-121.
13.
Okonkwo, C., Vassileva, J., Affective Pedagogical Agents and User Persuasion. in Universal
Access in Human - Computer Interaction (UAHCI), (New Orleans, USA, 2001), 397-401.
14.
Thórisson,
K.R.
Communicative
Humanoids
-
A
Computational
Model
of
Psychosocial
Dialogue Skills Media Arts & Sciences, School of Architecture & Planning, MIT, Boston,
1996.
15.
Cassell, J. and Thórisson, K.R. The Power of a Nod and a Glance: Envelope vs. Emotional
Feedback in Animated Converstional Agents. Applied Artifical Intelligence, 13 (4-5) 1999.
519-538.
16.
André, E., Muller, J. and Rist, T., The PPP Persona: A Multipurpose Animated Presentation
17.
Hudon, M. Modélisation des stratégies d'intégration de l'humour au sein de l'environnement
18.
Eco, U. Le signe. Labors, Bruxelles, 1988.
Agent. in Advanced Visual Interfaces, (1996), ACM.
ExploraGraph. Département de Communication, Université de Montréal, Montréal, 2002.
19.
Lombard, M. and Ditton, T.B. At the Heart of It All: The Concept of Presence. J. of Computer
Mediated Communication, 3 (2) 1997.
20.
21.
Bertrand, D. Dire et faire au travail. Érès, Toulouse, 1995.
Short, J., William, E. and Christie, B. The social Psychology of Telecommunication. Wiley,
London, 1976.
22.
Biocca, F. Cyborg's dilemna: Progressive embodiment in virtual environments. J. of Computer
Mediated Communication, 3 (2) 1997.
23.
Dufresne, A. Conception d’interfaces pour l’apprentissage à distance. La Revue de l'Éducation
à Distance, XII (1/2) 1997. 177-200.
24.
Martinez,
M.
and
Bunderson,
C.V.
Foundations
for
Personalized
Web
Learning
Environments. ALN Magazine, 4 (2) 2000.
25.
Paquette, G. and Tchounikine, P. Contribution à l'ingénierie des systèmes conseillers : une
approche méthodologique fondée sur l'analyse du modèle de la tâche. Sciences et Techniques
Éducatives, 9 2002. (to be published).
51
Interactive Cognitive Modelling Agents - Potential and
Challenges
Vania Dimitrova
School of Computing, Leeds University, Leeds, LS2 9JT, UK,
vania@comp.leeds.ac.uk
Abstract: Interactive cognitive modelling agents are defined here as diagnostic agents that involve
human learners in diagnostic dialogues and extract a picture of the learner's cognition in terms of
beliefs, misunderstandings, misconceptions, and reasoning. This paper is written both as a reflection on
our recent work on interactive open learner modelling, which is a specific and fairly simplified
interactive cognitive modelling method, and as a proposal for developing a framework for interactive
cognitive modelling agents. We discuss advantages of the approach and outline pitfalls with the initial
architecture suggesting possible research techniques to tackle these problems.
Key words: interactive cognitive modelling, meta-cognition, evaluation.
1
Introduction
Learners expect to be understood when they ask for advice, assistance, explanation, guidance, tutoring,
etc. Effective adaptive learning environments require robust learner models (LMs) (Self, 1999) that
represent learners’ preferences, needs, knowledge, misconceptions, skills.
Diagnosis
is
a
mutual
process – it depends on the diagnosee's involvement and the diagnoser's ability to encourage this
involvement (Dillenbourg, 1996). Learners would expect the computer teachers with whom they are
interacting to be willing to participate in a discussion about their problems rather than providing quick,
incomplete, and incomprehensible responses. The result of such diagnostic interactions is eliciting a
picture of the learner's cognition with the active participation both of the learner and the teacher. Not
only is interactive diagnosis likely to be more accurate and to enable effective personalisation, but
when it does take place in educational situations, it can bring deeper insights for both the teacher in
terms of reflection on their own practice and the learner in terms of promoting important metacognitive skills. Such wealth diagnostic interactions are increasingly needed nowadays in many
advanced learning environments – particularly systems that require sophisticated learner models and
promote meta-cognition to help learners understand themselves what their problems and needs are.
In contrast, traditional computer diagnostic systems seek to infer the reasons for the learners' behaviour
without directly involving the learners. Recently, approaches that involve learners in diagnosis have
been proposed (Bull et al., 1995; Bull & Brna 1999; Kay, 1995; McCalla et al., 2000; Morales et al.,
2000; Paiva & Self, 1995; Zapata-Rivera & Greer, 2001). Most of these methods are concerned with
open learner modelling where the learners are provided with the means to inspect and change the
models the systems build of them. Commonly, these systems externalise the LMs in some viewers and
provide menu options for the learners to change the content of their models. The users can sometimes
ask for explanations and justifications of the computer's opinions. The approach proposed by Bull
(1997) suggests an enhanced method of interaction in a menu-based environment for negotiating the
learner model. When inconsistencies between the student and the computer’s views about the student’s
beliefs are identified, negotiative dialogue is triggered. However, naturally occurring human-human
diagnostic interactions, e.g. peer diagnostic systems (Bull & Brna, 1999), accommodate a richer set of
communicative activities – the dialogue comprises multiple exchanges, no one controls or restricts
what another may say (at the same time some participants, e.g. the teacher, might be attributed a
guidance role), and all parties can change the topic and initiate a new discussion. Such rich interactions
that aim at extracting a model of the learner’s cognition are addressed in this paper.
52
This position paper will discuss the design of a new diagnostic architecture - artificial agents that
understand human users by talking and listening to them. More specifically, we will present Interactive
Cognitive Modelling Agents (ICOMAs) - computer diagnosers capable of involving learners in an
ongoing dialogue that results in eliciting models of the learners’ cognition with the active participation
of the learners. We will present an initial exploration of the approach - interactive open learner
modelling, illustrated in the system STyLE-OLM. Advantages and possible applications of ICOMAs
will then be discussed. Finally, we will sketch out further research towards extending STyLE-OLM to
provide a framework for interactive cognitive modelling. We will outline challenges and will speculate
about
possible
solutions
in
order
to
accommodate
more
advanced
domain
reasoning,
richer
communication means, extended dialogue management, and enhanced learner model maintenance.
2
Interactive Cognitive Modelling Agents
Human teachers follow a variety of diagnostic tactics when they guide diagnostic dialogue. To
illustrate, let us consider an excerpt from the communication between a teacher and a language student
who studies for technical translator and faces problems with some domain terms. Similar dialogues
have been observed in early studies we have conducted with university students in Bulgaria (Dimitrova
& Dicheva, 1998). Note, that the actual dialogue was in Bulgarian, the translation below is done by the
author. The domain concepts mentioned in the dialogue are given in small capital letters.
[1]
Learner.
about
[2]
it.
Is
I
Teacher.
your
We
question.
[3]
Learner.
[4]
Teacher.
OPERATING
[5]
is
[6]
an
a
No.
Teacher.
discuss
of
some
OPERATING
Isn't
example
for
PASCAL
but
I
know
nothing
of
other
terms
before
I
can
answer
SYSTEMS?
this
MSDOS?
an
You
While
with
OPERATING
the
icons.
[9]
Teacher.
to
SYSTEM.
OPERATING
[10]
Learner.
BASIC
at
[11]
Teacher.
meant
SYSTEM
WINDOWS
to
the
WINDOWS
return
You
about
mean
run
example
the
Do
MsDos
have
the
to
your
PROGRAMMING
languages
an
as
you
know
any
other
and
am
confused
now.
it.
maintain
MS D O S
OPERATING
is
the
one
communication
example
of
an
SYSTEM.
MSDOS?
OPERATING
OPERATING
write
a
that
on
Different
with
to
of
same
provides
back
PROGRAMS
that
same.
people
SYSTEM,
also
is
contact
MS D O S
Let’s
COMPUTER
PROGRAMS
another
exactly
in
talk
the
is
mean
Not
OPERATING
are
SYSTEMS
and
people
and
thought
WINDOWS
COMPUTERS.
a
about
SYSTEM?
OPERATING
SYSTEM.
[8]
P ASCAL .
to
so.
one
talking
WINDOWS?
heard
think
I
COMPUTER
Learner.
need
need
you
is
OPERATING
[7]
we
I
MSDOS
Teacher.
means
will
people
to
SYSTEMS?
between
OPERATING
heard
related
Have
Yes,
Learner.
What
have
PASCAL
in
COMMANDS
GRAPHICAL
question.
SYSTEMS
USER
To
provide
and
SYSTEM
to
order
INTERFACE
compare
different
control
to
the
communicate
which
WINDOWS
uses
and
menus
PASCAL,
LANGUAGES.
used
to
write
programs?
My
friends
did
some
school.
That
PROGRAMMING
COMPUTER
with
is
right.
LANGUAGE
the
help
BASIC
allow
of
an
is
an
example
programmers
OPERATING
SYSTEM,
to
of
write
for
a
PROGRAMMING
PROGRAMS ,
example
LANGUAGE.
which
are
So
is
run
on
WINDOWS.
This example shows that diagnostic dialogues are naturally embedded into the whole teaching process.
We propose that these dialogues are conducted by Interactive Cognitive Modelling Agents which aim at
extracting a picture of the learner’s cognition. Such agents will have to share resources with other parts
of an interactive learning environment, for example, domain expertise and learner model. These agents
may serve as main diagnostic components in learning environments. They could also be used together
with other diagnostic methods (e.g. assessing learners’ drill performance) to enhance the quality of the
learner models by addressing aspects that may well have been missed or diagnosed wrongly by the
traditional diagnostic methods. To build models of ICOMAs, we will make the following assumptions:
•
ICOMAs share common communication means with the learners where domain facts are discussed
and models of the learners' conceptual understanding are extracted. The communication language
should allow effective diagnostic interactions where both a diagnosee and a diagnoser participate
actively.
53
•
The interaction comprises a sequence of episodes which span over multiple turns and follow
specific
diagnostic
tactics.
ICOMAs
plan
the
content
of
the
interaction
and
take
diagnostic
decisions based on their domain expertise. These agents are empowered by discourse knowledge
that enables them to lead a coherent interaction aimed at eliciting a picture of the diagnosee’s
conceptual understanding
•
ICOMAs' aim is to elicit a picture of the learner’s conceptual understanding in terms of beliefs (as
in the example above), misunderstandings, misconceptions, and reasoning. These agents must
incorporate
appropriate
reasoning
capabilities
that
enable
the
extraction
of
an
interactively
constructed learner model.
Computational frameworks of interactive cognitive modelling agents will allow understanding the
process and will provide vehicles for building robust computer simulations of interactive teachers
capable of understanding the learners' problems and needs. Moreover, formalisations will aid the
implementation of diagnostic agents in various domains.
3
Initial Exploration – STyLE-OLM
We have examined a specific interactive cognitive modelling method, called interactive open learner
modelling (IOLM), where a learner is provided with the means to inspect and discuss the conceptual
models that computer systems build of them (Dimitrova, 2001). Despite the fact that IOLM agents
focus mainly on discussing the content of the learner model and demonstrate a fairly simplified case of
the interaction discussed above, it has confirmed the feasibility of the assumptions discussed in section
2. Many techniques from this specific method appear fruitful in the more advanced diagnostic model
the ICOMAs address.
A formal framework for interactive open learner modelling has been developed (Dimitrova et al., 1999;
Dimitrova et al., 2000; Dimitrova 2001). It includes distinctive components: a discourse model based
on
an
approach
known
as
dialogue
games
manages
diagnostic
interactions
and
provides
both
a
diagnoser and a diagnosee with a common communication method and symmetrical power in dialogue
maintenance (Dimitrova et al. 1999, Dimitrova et al., 2002) while a formally defined mechanism based
on a belief modal operator adopts nonmonotonic reasoning to maintain a jointly constructed LM
(Dimitrova et al., 2000). The framework has been demonstrated in STyLE-OLM - an IOLM system in
a terminological domain (Dimitrova, 2002). Two instantiations of STyLE-OLM – in Computer Science
and in Finance - have been developed, the example here is from an experimental study with the system
conducted in a Finance domain (topic – Financial Markets). The study involved seven post graduate
students from the author's department. More details about the study are given in Dimitrova (2001).
The excerpt below shows a fragment of an interaction with STyLE-OLM where the system challenges
the robustness of a learner's conceptualisation, which provides for reflection and scaffolding of the
learner’s domain knowledge. Initially, the learner model contains information that the learner knows
some aspects of the terms MONEY MARKET and CAPITAL MARKET but often mixes these terms.
STyLE-
OLM initiates a dialogue to explore the learner's knowledge about these terms and to find possible
reasons for the confusion.
[1]
STyLE-OLM[inquire]
[2]
Learner[agree]
[3]
STyLE-OLM[inquire]
[4]
Learner[inform]
I
DO
YOU
AGREE
I
THINK
that
CAPITAL
THINK
MONEY MARKET
MARKET
that
MONEY MARKET
raises
operates
operates
CAPITAL
CAPITAL MARKET
and
with
with
SHORT TERM INVESTMENTS.
SHORT TERM INVESTMENTS.
operates
operates
with
with
WHAT?
SHORT TERM INVESTMENTS.
The dialogue follows a game tactic that elaborates on the similarities and differences between the two
terms. Move [4], however, reveals an erroneous aspect of a learner's knowledge: the proposition
'CAPITAL MARKET operates with SHORT TERM INVESTMENTS', which the learner is assumed to believe, is
not confirmed by the diagnoser’s domain knowledge. A new dialogue episode is initiated searching for
some explanation for the learner's error.
[5]
STyLE-OLM[challenge]
WHY
DO
YOU
THINK
CAPITAL MARKET
operates
with
SHORT TERM
INVESTMENTS?
[6]
Learner[justify]
are
BECAUSE
CAPITAL MARKET
operates
with
TREASURY BILLS
and
TREASURY BILLS
SHORT TERM INVESTMENTS .
54
The screen shot in Figure 1 depicts the environment at the moment when the learner constructs his
justification.
The system now registers a possible reason for misattribution: the student may believe wrongly that
'CAPITAL MARKET operates with SHORT TERM INVESTMENTS' because he believes that 'CAPITAL MARKET
operates with TREASURY BILLS and TREASURY BILLS are SHORT TERM INVESTMENTS'. Since the last
belief has also not been confirmed by its domain expertise, STyLE-OLM initiates a new dialogue
episode for explaining this error:
[7]
STyLE-OLM[challenge]
and
WHY
TREASURY BILLS
are
DO
YOU
THINK
CAPITAL MARKET
operates
with
TREASURY
BILLS
SHORT TERM INVESTMENTS ?
Buttons for
Graphical
browsing
tools for
through the
constructing
dialogue
CGs
history
Button to
switch to
BROWSE
mode
The learner
constructs and
selects a CG
with the
proposition of
his justification
Textual
To justify the
representation
learner selects
of the system's
a justify move
challenge
Dialogue
history
Figure 1. Learner's justification in STyLE-OLM (move [6]).
At this time, the student is confused. The robustness of his domain beliefs is challenged. He switches to
a BROWSE mode where he browses through the beliefs in the LM and sees the level of correctness the
system assignes to them. Having seen that the belief presented in Figure 2 has been classified as known
wrongly by the system,
TREASURY BILLS
the
learner
realises
that
his
claim
that CAPITAL MARKET operates with
and TREASURY BILLS are SHORT TERM INVESTMENTS might be wrong.
55
Buttons for
browsing
through the
beliefs in the
LM
Button to
switch to
DISCUSS
mode
A belief from
the LM
presented with a
CG
Degree of
correctness the
system assigns
to the belief,
represented
with the CG
History of
the preceding
discussion
Figure 2. The STyLE-OLM environment in BROWSE – the learner browses the beliefs in the
LM provoked by the system's challenges in move [7].
He now needs the system's help to explore facts about these terms. He switches back to DISCUSS mode
to continue the discussion of the LM. He asks a question:
[8]
Learner[inquire]
IS
[9]
STyLE-OLM[inform]
IT
I
TRUE
KNOW
TREASURY BILLS
that
that
TREASURY BILLS
are
are
SHORT TERM INVESTMENTS ?
SHORT TERM INVESTMENTS .
The learner realises that one part of his statement in [6] is correct and challenges the other:
[10]
Learner[inquire]
IS
IT
TRUE
that
CAPITAL MARKET
operates
with
SHORT TERM
INVESTMENTS?
[11]STyLE-OLM[inform]
I
DO
NOT
KNOW
that
CAPITAL MARKET
operates
with
SHORT TERM
INVESTMENTS.
The learner has clarified the part of his beliefs which is wrong. He withdraws his claim that CAPITAL
MARKET
operates with SHORT TERM INVESTMENTS.
[12]Learner[disagree]
I
DISAGREE
that
CAPITAL
MARKET
operates
with
SHORT
TERM
INVESTMENTS?
The interaction continues with the comparison between MONEY MARKET and CAPITAL MARKET. Due to
space restrictions, we will not present the whole dialogue here. A full transcript of this dialogue as well
as some other example interactions with STyLE-OLM are given in Dimitrova (2001). After the
dialogue is terminated, a jointly constructed learner model that takes into account what has been
expressed by the two agents during the interaction is obtained (e.g. the beief 'C APITAL MARKET
operates with SHORT TERM INVESTMENTS' will be deleted from the initial LM).
STyLE-OLM allows inspecting and discussing the learner model in a relatively expressive graphical
manner which fosters the articulation of domain knowledge and can lead to conceptual understanding.
A constructive dialogue guided by the system enables the exploration of aspects of a learner's domain
knowledge and the extension of the scope of beliefs in the learner model. Learners are provided with a
symmetrical role in maintaining the dialogue. A flexible diagnostic mechanism allows the management
of a learner model jointly constructed by the computer system and the learner with the latter being
provided with equal power to influence the diagnosis.
56
3.1
Potential of Interactive Cognitive Modelling Agents
Interactive cognitive modelling agents have a strong potential in advanced learning environments
capable
of
evaluative
tailoring
study
to
with
the
needs
of
STyLE-OLM
the
learners
has
and
demonstrated
promoting
meta-cognitive
advantages
of
the
processes.
approach
in
The
terms
of
improving the quality of the learner model and providing the means for reflective activities (Dimitrova
et al., 2001, Dimitrova, 2002).
Improving the quality of the learner model. We observed fewer inconsistencies in the resulting LM,
•
a larger scope of learner's beliefs, and some explanations of the learner's errors. The obtained LM
included a higher proportion of valid assertions about the learner's knowledge and minimised the
number of not valid assertions about the learner's knowledge.
Providing means for reflective activities. The study allowed us to monitor the following reflective
•
activities with STyLE-OLM: the students rendered statements about their domain beliefs, they went
back to claims about their beliefs and (sometimes) changed these claims, and they investigated
arguments to support their beliefs. While more knowledgeable learners were engaged in reflective
dialogues about the domain, less knowledgeable learners were provoked to inspect their models and
challenge the robustness of these models.
ICOMAs can be embedded in advanced e-learning systems to enable better understanding of the
learners and to help learners understand themselves what their accomplishments and problems are.
There is a growing interest in finding robust and computationally tractable methods for eliciting models
of the users’ cognitive states to aid the development of personalised systems in various domains,
especially in the increasingly popular Internet applications. ICOMAs
developing
sophisticated
personalised
Internet
agents,
for
example
can
be
personal
used
as
a
basis
e-consultants
or
for
e-
metntors.
Personal e-consultants are interactive agents that offer personalised advice tailored to the users'
•
problems
and
needs.
Such
agents
may
be
incorporated
in
modern
e-commerce
or
e-banking
systems. For example, a non expert in financial planning seeking to understand the concept of an
"ISA" may be provided with a personal e-consultant that discusses the domain terminology with the
user, infers a model of the user's conceptual understanding, and offers adaptive explanations
tailored to the user's understanding of the terminology.
E-mentors are agents that act as personal mentors. Mentoring is a relationship in which one person -
•
usually someone more experienced - helps another to discover more about themselves, their
potential and their capability. The mentor's role is to listen, ask questions, probe for facts and
understand its mentee and to act as a source of information, experience, and advice. Artificial
mentors
could
be
embedded
in
new
generation
e-training
systems
to
provide
the
means
to
understand the trainees, offer personalised help, and help the trainees identify themselves what their
needs are.
While the evaluation of STyLE-OLM outlined potentials of the IOLM framework, it also revealed
unsolved aspects that led to pitfalls of the architecture. We will sketch out these aspects next and will
draw speculations about how they may be addressed in a more sophisticated framework for ICOMAs.
4
Challenges
In
this
section,
based
on
examining
learner
interactions
with
STyLE-OLM,
we
outline
further
improvements of the IOLM framework in order to maintain enhanced diagnostic interactions required
in interactive cognitive modelling.
4.1
Exploring advanced domain inference
Interactive cognitive modelling requires high level logic in order to develop appropriate tactics to
reveal reasons for users’ misconceptions. We have experimented with conceptual graphs, which have
been
found
a
suitable
formalisation
for
the
purposes
of
interactive
diagnosis.
However,
some
commonsense reasoners, such as modus tollens, which are often applied by humans have been difficult
to capture. This led sometimes to missing student reasoners and interrupting profitable dialogue
episodes. In order to address negations of domain propositions represented with conceptual graphs, an
57
additional modal operator not shall be considered (Sowa, 1984), a methodology how this can be
implemented in computer applications is discussed in Dau (2000).
ICOMAs need to analyse propositions composed by the users. It is difficult to predict how the learners
will
express
their
propositions.
Even
in
a
highly
structured
graphical
communication
language
exploited in STyLE-OLM (Dimitrova et al., 2002), computational problems with ambiguity of domain
propositions became apparent. Firstly, many relations have overlapping of their meanings in everyday
language. When learner mixed such relations (e.g. "agent" and "actor", see (Sowa, 1984)) StyLE-OLM
assigned erroneous conceptualisation while the learners had simply confused very similar words.
Dealing with mixed relations requires some representation of interdependencies between relations and
suitable
reasoning
to
find
relation
similarities.
For
example,
λ-definitions
of
relations
used
in
conceptual graphs theory (Sowa, 1984) may empower such reasoning.
Secondly, often a proposition is a re-phrase of another, which, if not captured by the domain reasoning
mechanism, may lead to misdiagnosis or obscure dialogue moves such as repetition or inappropriate
challenging.
Extended
comparison
techniques
to
allow
for
different
perspectives
of
the
same
knowledge to be captured are needed. Mechanisms, similar to those presented in (Dieng & Hug, 1998;
Martin, 2000) seem applicable in ICOMAs.
Dealing with ambiguity of domain propositions requires not only discovering potential ambiguous
situations
but
also
addressing
them
in
the
dialogue.
Meta-dialogue
for
dealing
with
mis-
communications and grounding, e.g. (Traum & Dillenbourg, 1996), need to be incorporated in the
dialogue maintenance mechanism.
The study with STyLE-OLM showed the need to handle reasoning under incomplete domain expertise.
When the system did not have information about a domain fact, it simply assumed that this was an
erroneous belief and challenged it, which led to inappropriate system behaviour at times (e.g. a
learner's statement "A bank operates with money" was not confirmed by the system's domain expertise
and challenged "Why do you think a bank operates with money", which frustrated the learner). A less
knowledge-centred behaviour of the diagnostic agent is required. We may envisage that at times the
diagnostic agent behaves as a peer who may extend its competence to incorporate information provided
by the learner depending on its trust in the learner and its own domain reasoning. Planning diagnostic
dialogue when the diagnoser's domain expertise is incomplete seems to relate to decision making under
uncertainty which deals with reasoning that require information not available at the time it is needed.
One way to tackle the problem is to employ some form of defeasible reasoning, i.e. to make some
assumptions from which some conclusions may be drawn and withdraw the assumptions later on if the
assumption is proven invalid (Davis, 1990; Parsons, 2001). In this case, ICOMAs will need to have a
mechanism for dealing with the degree of certainty about the truth of domain propositions. Another
possible method to deal with incomplete domain expertise is argumentation (Krause & Clark, 1993;
Parsons, 2001). For instance, an ICOMA may accept a proposition suggested by a student if it cannot
find a rebuttal for it. In this case, the agent needs to be able to incorporate some kind of argumentative
reasoning in its dialogue planning.
4.2
Providing rich communication means
STyLE-OLM provided a graphical communication medium combining propositions represented as
conceptual graphs and illocutions represented with sentence openers. While, such environment was
found
favourable
for
diagnostic
interactions
(Dimitrova
et
al.,
2002),
some
problems
were
also
identified.
Firstly,
mixing
textual
and
graphical
representations
requires
keeping
the
meaning
of
both
representations coherent. When graphics is utilised for constructing dialogue utterances, a sophisticated
mechanism
for
linguistically
generating
coherent
text
linguistic
that
expressions
represents
the
from
graphics
meaning
of
the
is
needed
graphical
in
order
to
provide
expressions.
When
communication is based on conceptual graphs, natural language processing approaches that generate
text from graphs, e.g. (Angelova & Bontcheva, 1996; Nikolov et al., 1995), may be employed.
Secondly, there is no comprehensive study of the type of operations needed when communication is
done
with
graphics,
for
example
how
to
facilitate
the
construction
of
graphical
utterances,
the
modification of graphical "propositions", the search through graphical expressions, etc. We adopted a
58
rather heuristic approach following conventional operations used in graphical packages but it became
apparent that a more systematic approach is needed to examine the effectiveness and pitfalls of these
operations. In this line, approaches from Human-Computer Interaction seem favourable, for example
(Green & Petre, 1996).
Thirdly, the participants in the evaluation of STyLE-OLM did not agree regarding their choice of
graphics or text for communication. The study was too limited to discuss this issue deeply, and further
exploration is needed. In this line, (Cox, 1999) provides possible directions highlighting the difference
between situations in which a presented external representation is interpreted and situations in which
participants construct external representation (both types of situations are present in communicating
with diagrams). As Bull et al. (2001) argue, differences in the learners' cognitive styles impose a
variety of communication means to be combined in a single system. Since the associations between
cognitive style and presentation format are not straightforward, Bull et al. propose that learners' should
be given the choice of a textual or graphical environment (in domains for which either may be
appropriate) for discussing their cognitive model. Providing text input would require student diagnosis
based on a free text, which is a challenging computational task at present, e.g. the learners' statements
may not make sense according to the system's domain model. Recent research in natural language
processing is addressing relevant aspects (e.g. Ramsay & Seville, 2000) and one would expect in due
course interactive cognitive modelling to accommodate communication in a free natural language. This
would open a new research issue of how to accommodate misunderstanding, repair and grounding (e.g.
Traum & Dillenbourg, 1996) in a dialogue which is aimed at student diagnosis.
4.3
Maintaining a coherent diagnostic dialogue
Maintaining a coherent diagnostic dialogue requires dealing with vague and incomplete information
about possible learner's misconceptions and suitable diagnostic tactics. Reasons for people’s cognitive
errors are generally difficult to define. There is a fair bit of work done on defining theories for concept
learning that deal with the correct application of rules such as generalisation, specification, similarity
(Thagard, 1982). However, very little has been done to precisely define types of rules that people apply
wrongly or fail to apply, which leads to building wrong or incomplete conceptual models. In this line,
in order to deal with vagueness of defining diagnostic tactics, ICOMs would need to appropriately
incorporate methods for reasoning under uncertainty (Krause & Clark, 1993; Parsons, 2001).
An important role of ICOMAs is to scaffold learners' conceptual understanding. This can be achieved
via adopting a variety of strategies to discuss wrong beliefs, e.g. comparison (Milosavlevic, 1997),
explanation (Bontcheva & Wilks, 1999; McCoy, 1989), inquiry dialogue (Wong et al., 1998). In
addition, learners need to be navigated through the dialogue in order to help them understand the
semantic structure of the conversation. At times, the learners in the study with STyLE-OLM felt that
the system was jumping between different propositions without any obvious reason for this, and the
learners failed to follow what was going on. Diagnostic dialogue can be improved if the diagnoser
makes apparent the rhetorical structure of the conversation, for instance what the current dialogue
tactic is, why the system is changing the tactic, when the system goes back to a previously initiated
dialogue episode, etc. To illustrate this, we present below an improvement of the dialogue example
given above with added phrases to explain the rhetorical structure of the conversation.
[1]
STyLE-OLM[inquire]
MARKET.
MARKET
We
It
can
operates
[2]
Learner[agree]
[3]
STyLE-OLM[inquire]
trying
I
to
seems
try
to
with
AGREE
We
of
SHORT
that
just
compare
definition
that
you
compare
MONEY
MARKET
MARKET
MARKET.
the
two
terms
terms
MONEY
here.
DO
MARKET
and
YOU
THINK
CAPITAL
MONEY
INVESTMENTS.
mentioned
MONEY
CAPITAL
TERM
confuse
these
operates
the
and
CAPITAL
CAPITAL
with
definition
SHORT
of
MARKET,
MARKET
so
raises
TERM
MONEY
let
CAPITAL
INVESTMENTS.
MARKET.
us
and
We
are
consider
operates
the
with
WHAT?
[4]
Learner[inform]
[5]
STyLE-OLM[challenge]
operate
that
[6]
I
with
CAPITAL
Learner[justify]
TREASURY
THINK
see,
SHORT
MARKET
IT
BILLS
I
IS
are
that
CAPITAL
you
TERM
think
TRUE
with
BECAUSE
TERM
operates
that
INVESTMENTS.
operates
SHORT
MARKET
Can
SHORT
CAPITAL
both
MONEY
you
TERM
with
SHORT
MARKET
explain
TERM
and
what
INVESTMENTS.
CAPITAL
makes
MARKET
you
think
INVESTMENTS?
MARKET
operates
with
TREASURY
BILLS
and
INVESTMENTS.
59
[7]
STyLE-OLM[challenge]
you
and
to
TREASURY
Learner[inquire]
IS
[9]
STyLE-OLM[inform]
are
IT
Learner[inquire]
IS
[11]
STyLE-OLM[inform]
IT
To
operates
a
SHORT
to
my
I
TERM
CAPITAL
of
TREASURY
know
you
believed.
operates
MARKET
with
Can
I
TREASURY
ask
BILLS
INVESTMENTS?
TREASURY
that
didn't
CAPITAL
BILLS
knowledge,
best
with
that
believe
that
TRUE
the
fact
you
TRUE
Well,
[10]
is
that
BILLS
[8]
MARKET
This
confirm
my
are
TREASURY
MARKET
SHORT
BILLS
TERM
are
operates
knowledge,
I
INVESTMENTS?
SHORT
with
can't
TERM
INVESTMENTS.
TREASURY
confirm
BILLS?
that
CAPITAL
BILLS.
Sessions with STyLE-OLM have revealed that both the diagnoser and the diagnosee may need to
express uncertainty in their dialogue utterances. Fir example, the diagnoser may deal with information
that is not available and may need to make assumptions about the student's knowledge, while the
student may not be completely sure about the validity of their statements. Consequently, diagnostic
dialogue has to accommodate different verbal expressions of uncertainty, such as definite, likely,
possible, unlikely, and impossible (Krause & Clark, 1993).
4.4
Eliciting a learner model under uncertain conditions
The result of a method for student modelling is a model of the student’s cognition. Following the
discussion above, it is apparent that the mechanism for eliciting a resultant student model has to
accommodate reasoning under uncertainty. While some level of uncertainty might be handled via
interaction enabling agents to challenge or withdraw their beliefs and to clarify their statements, a more
elaborated notion of uncertainty would be sensible. For example, representing some strength of beliefs
in the learner model (e.g. 'entirely sure', 'not very sure', 'guessing') and making plausible inferences that
incorporate degree of belief and nonmonotonic reasoning (Parsons, 2001).
The study with STyLE-OLM confirmed that inconsistency is often a case in students' beliefs. Although
clarification dialogue in STyLE-OLM did enable us to handle inconsistency in student beliefs and
avoid
extensive
belief
revision
(Giangrandi
&
Tasso,
1995),
we
found
that
some
contradicting
propositions were left due to limitations of the system's reasoning (e.g. a learner was thought to believe
both ‘CAPITAL
operates
BILLS
MARKET
with
are not LONG
operates
TREASURY
TERM
with
BILLS ’
LONG
TERM
INVESTMENTS’
(correct) and ‘CAPITAL
MARKET
(wrong), which are actually contradictory because TREASURY
INVESTMENTS
b u t SHORT
TERM
INVESTMENTS).
Therefore,
a
reasoning
mechanism that explores deeper all consequences of the student's claims is required. A feasible
approach seems advanced nonmonotonic reasoning (Davis, 1990).
The mechanism for ascribing participants' beliefs from their communicative acts would have to adjust
the belief ascription according to the agents' goals. In STyLE-OLM, when asking questions learners
were diagnosed that they did not know a domain fact. However, a question may not always indicate
missing knowledge, but may sometimes mean that the students seek for confirmations of domain
aspects they know. Further extensions need to take into account theories that deal with the repair of
mistaken ascriptions, e.g. (Lee & Wilks, 1997).
5
Conclusions
This paper is written both as a reflection on our recent work on interactive open learner modelling and
as a proposal for further research on interactive cognitive modelling. Our long term goal is to develop
robust and efficient models of computer agents that can conduct diagnostic dialogues with a learner (or
a group of learners) in order to understand the learners and help themselves understand what their
problems and needs are. One type of such diagnostic agents - Interactive Cognitive Modelling Agents have been discussed in this paper. ICOMAs are interactive diagnostic agents that involve human
learners in diagnostic dialogues and extracts a picture of the learner's cognition in terms of beliefs,
misunderstandings, misconceptions, and reasoning. As an initial exploration of the approach we have
examined a method called interactive open learner modelling where a computer diagnoser enables a
human learner to inspect and discuss the model the diagnoser builds of him/her. STyLE-OLM - the
system we have built to illustrate our IOLM framework - is a rather simplified demonstration of
ICOMA. However, it did allow us to observe some advantages of the approach, which were outlined in
60
the paper. We have also sketched out potential problems with STyLE-OLM and have pointed to
possible
methods
to
tackle
these
problems
in
an
enhanced
architecture
of
interactive
cognitive
modelling agents.
References
Angelova,
G.,
&
Bontcheva,
K.
(1996).
DB-MAT:
Knowledge
acquisition,
processing
and
NL
generation using conceptual graphs. In P.W. Eklund, G. Ellis & G. Mann (Eds.), Conceptual
structures: knowledge representation as interlingua: 4th international conference on conceptual
structures (pp. 115-128). Springer, LNAI 115.
Bontcheva, K., & Wilks, Y. (1999). Generation of adaptive (hyper)text explanations with an agent
In Proceedings of the European Workshop on Language Generation (EWNLG'99),
model.
Toulouse, France.
Bull, S. (1997). Collaborative student modelling in foreign language learning. PhD Thesis. University
of Edinburgh.
Bull, S., & Brna, P. (1999). Enhancing peer interaction in the Solar system. In P. Brna, M. Baker & K.
Stenning (Eds.). Roles of communicative interaction in learning to model in Mathematics and
Science: Proceedings of the C-LEMMAS conference, Ajaccio, Corsica.
Bull, S., Brna, P., & Pain, H. (1995). Extending the scope of student models. User Modeling and UserAdapted Interaction, 5(1), 45-65.
Bull, S., Dimitrova, V., & Brna, P. (2001). Enhancing reflective modelling through communicative
interaction in learning environments. In P. Brna, K. Stenning, M. Baker & A. Tiberghien (Eds.),
The role of communication in learning to model. Mahwah, NJ: Lawrence Erlbaum Associates,
in press.
Cox,
R.
(1999).
Representation
construction,
externalised
cognition
and
individual
differences.
Learning and Instruction, 9, 343-363.
Dau,
F.
(2000).
Negations
Conceptual
in
structures:
simple
conceptual
logical,
linguistic,
graphs.
and
In
B. Ganter
computational
&
G.
issues:
W.
Mineau (Eds.),
8th
International
conference on conceptual structures. Berlin London: Springer, 2000.
Davis, E. (1990). Representations of commonsense knowledge. San Mateo: Morgan Kaufmann.
Davis, E. (1990). Representations of commonsense knowledge. San Mateo: Morgan Kaufmann.
Dieng, R., & Hug, S. (1998). MULTIKAT, a Tool for comparing knowledge of multiple experts. In ML.
Mugnier & M. Chein (Eds.), Conceptual structures: theory, tools, and applications: 6th
International conference on conceptual structures (pp. 139-153). Springer, LNAI 1453.
Dillenbourg,
P.
(1996).
Some
technical
implications
of
distributed
cognition
on
the
design
on
interactive learning environments. Journal of Artificial Intelligence in Education, 7 (2), 161180.
Dimitrova, V. (2001) Interactive open learner modelling. PhD thesis, University of Leeds.
Dimitrova, V. (2002). STyLE-OLM: Interactive open learner modelling. Technical report, School of
Computing, Leeds University (submitted to the International Journal of Artificial Intelligenec in
Education).
Dimitrova, V., & Dicheva, D. (1998). "Who is Who": The Roles in an Intelligent System for Foreign
Dimitrova,
V.,
Brna,
P.,
Self,
J.A.
(2002).
The
design
and
implementation
of
a
graphical
communication medium for interactive open learner modelling. In Proceedings of ITS2002 –
Intelligent Tutoring Systems Conference, Springer, LNCS.
Dimitrova, V., Self, J.A. & Brna, P. (2001). Applying interactive open learner models to learning
technical terminology. In
Proceedings of UM2001 – User Modelling conference.
Springer,
LNCS.
Dimitrova, V., Self, J.A., & Brna, P. (1999). The interactive maintenance of open learner models. In
S.P.Lajoie & M. Vivet (Eds.), Artificial intelligence in education (pp. 405-412). Amsterdam:
IOS Press.
Dimitrova, V., Self, J.A., & Brna, P. (2000). Maintaining a jointly constructed student model. In S.A.
Cerri & D. Dochev (Eds.), Artificial intelligence: Methodology, systems, and applications (pp.
221-231). Berlin: Springer, Lecture Notes of Artificial Intelligence.
Giangrandi, P. & Tasso, P. (1995). Truth maintenance techniques for modeling student’s behavior.
Journal of Artificial Intelligence in Education, 6, 153-202.
Green, T.G., & Petre, M. (1996). Usability analysis of visual programming environments: a 'cognitive
dimensions' framework. J of Visual Languages and Computing, 7, 131-174.
61
Kay, J. (1995). The UM toolkit for cooperative user modelling. User Modeling and User-Adapted
Interaction, 4, 149-196.
Krause,
K.,
&
Clark,
D.
(1993).
Representing
uncertain
knowledge:
An
Artificial
Intelligence
approach. Intellect. Oxford.
Language Terminology Learning. British Journal of Educational Technology, Vol. 29 (1), 47-57.
Martin, Ph. (2000). Conventions and notations for knowledge representation and retrieval. In B. Ganter
& G. W. Mineau (Eds.), Conceptual structures: logical, linguistic, and computational issues:
8th International conference on conceptual structures. Berlin London: Springer.
McCalla, G., Vassileva, J., Greer, J., & Bull, S. (2000). Active learner modelling. In G. Gauthier, C.
Frasson & K. VanLehn (Eds.), Intelligent
tutoring
systems (pp. 53-62). Berlin Heidelberg:
Springer-Verlag.
McCoy, K. (1989). Highlighting a user model to respond to misconceptions. In A. Kobsa & W.
Wahlster (Eds.), User models in dialogue systems (pp. 233-253). Springer-Verlag, Berlin.
Milosavljevic, M. (1997). Augmenting the user’s knowledge via comparison. In A. Jameson, C. Paris
&
C.
Tasso
(Eds.).
Proceedings
of
the
6th
International
conference
on
User
Modeling.
Available on-line from http://um.org.
Morales, R., Pain, H., & Conlon, T. (2000). Understandable learner models for a sensorimotor control
task. In G. Gauthier, C. Frasson & K. VanLehn (Eds), Proceedings of the 5th International
conference on Intelligent Tutoring Systems (pp. 222-231). Springer, LNCS.
Nicolov, N., Mellish, C., & Ritchie, G. (1995). Sentence generation from conceptual graphs. In G. Ellis
et al. (Eds.), Conceptual structures: Applications, implementation and theory: 3rd International
conference on conceptual structures (pp. 74-88). Springer, LNAI 954.
Paiva, A., & Self, J.A. (1995). TAGUS - a user and learner modelling workbench. User Modeling and
User-Adapted Interaction, 4, 197-226.
Parsons, S. (2001). Qualitative methods for reasoning under uncertainty. The MIT Press. Cambridge.
Ramsay, A., & Seville, H. (2000). What did he mean by that?. In S. Cerri & D. Dochev (Eds.),
Artificial
Intelligence:
Methodology,
systems
and
applications,
9th
International
AIMSA
Conference (pp. 199-209). Springer, LNAI 1904.
Self, J. A. (1999). The defining characteristics of intelligent tutoring systems research: ITSs care,
precisely. International Journal of Artificial Intelligence in Education, 10, 350-364.
Sowa, J. (1984). Conceptual structures: Information processing in mind and machine. MA: AddisonWesley.
Thagard, P. (1992). Conceptual revolutions. Princeton University Press.
Traum, D., & Dillenbourg, P. (1996). Miscommunication in multi-modal collaboration. In Working
Notes
of
the
AAAI
Workshop
on
Detecting,
Repairing,
And
Preventing
Human--Machine
Miscommunication.
Walton, D. (1984). Logical dialogue games and fallacies. University Press of America, Lanham.
Wong, L.H., Quek, C., & Looi, C.K. (1998). TAP-2: a framework for an inquiry dialogue based
tutoring system. International Journal of Artificial Intelligence in Education, 9, 88-110.
Zapata-Rivera, J-D., & Greer, J. (2000). Inspecting and visualising distributed Bayesian student
models. In G. Gauthier, C. Frasson & K. VanLehn (Eds), Proceedings of the 5th International
conference on Intelligent Tutoring Systems (pp. 544-553). Springer, LNCS.
62
Authors (in order of their appearance)
Susan Bull
Electronic, Electrical and Computer Engineering, University of Birmingham,
Edgbaston, Birmingham, B15 2TT, UK
Theson Nghiem
Electronic, Electrical and Computer Engineering, University of Birmingham,
Edgbaston, Birmingham, B15 2TT, UK
Douglas Chesher
School of Pathology, University of Sydney, Australia 2006
Judy Kay
School of Information Technologies, University of Sydney, Australia 2006
Nicholas King
School of Pathology, University of Sydney, Australia 2006
Maria Grigoriadou
Department of Informatics, University of Athens, GR-157.71, Athens, Greece
Maria Samarakou
Department
of
Energy
Technology,
Technological
Institute
of
Athens,
GR-
122.61, Athens Greece,
Dionissis Mitropoulos
Department of Informatics, University of Athens, GR-157.71, Athens, Greece
Michael Panagiotou
01-Pliroforiki S.A., 438 Acharnon Str., GR-111.43, Athens, Greece
Adja F. de Andrade
Computer Based Learning Unit, Leeds University, Leeds, LS2 9JT, UK
Paul Brna
School of Computing and Mathematics, Northumbria University, Newcastle
upon Tyne NE1 8ST, UK
Rosa Maria Vicari
PGIE- UFRGS, Caixa Postal: 5071 - Zip Code: 90041-970, Porto Alegre-RS,
Brasil
Aude Dufresne
Département de Communication, Université de Montréal, C.P. 6128 Succ CentreVille, Montréal, Qc, Canada
Martin Hudon
Département de Communication, Université de Montréal, C.P. 6128 Succ Centre-
Vania Dimitrova
School of Computing, Leeds University, Leeds, LS2 9JT, UK
Ville, Montréal, Qc, Canada
63