Book 2987 PDF
Book 2987 PDF
Book 2987 PDF
5-2013
Lam, Derrick A., "Improving Virtual Collaboration: Modeling for Recommendation Systems in a Classroom Wiki Environment"
(2013). Computer Science and Engineering: Theses, Dissertations, and Student Research. 55.
http://digitalcommons.unl.edu/computerscidiss/55
This Article is brought to you for free and open access by the Computer Science and Engineering, Department of at DigitalCommons@University of
Nebraska - Lincoln. It has been accepted for inclusion in Computer Science and Engineering: Theses, Dissertations, and Student Research by an
authorized administrator of DigitalCommons@University of Nebraska - Lincoln.
IMPROVING VIRTUAL COLLABORATION: MODELING FOR RECOMMENDATION
by
Derrick A. Lam
A THESIS
Lincoln, Nebraska
May, 2013
IMPROVING VIRTUAL COLLABORATION: MODELING FOR
emphasis placed on working jointly with others, whether it is in the classroom, in the lab,
in the workplace, or virtually across the world. The wiki is one particular virtual
collaboration tool that is gaining particular prominence in recent years, enabling people –
either in small project groups or as part of the wiki’s entire user base – to socially
construct knowledge asynchronously on a wide variety of topics. However, there are few
This thesis investigates the topic of user and data modeling for recommendation
model uses new metrics designed for the wiki domain, including: active-passive activity
level rating, minimalist-overachiever score, and others. For evaluation, the Biofinity
Intelligent Wiki was designed, developed, and deployed to a classroom environment and
is used for collaborative writing assignments. Post-hoc analysis on the usage data
demonstrates the effects of assignment criteria on student behavior, the value of the new
metrics and their correlation to various student strategies, and the potential for applying
This work provides insights and tools that are beneficial to virtual collaboration.
For example, the active-passive activity level rating provides a quick overview of a
strongly correlates to evaluations that participants have received, and with additional
artifacts that a participant has contributed towards are indicative of the participant’s
Table of Contents
List of Multimedia Objects .............................................................................................................. ix
Chapter 1: Introduction ................................................................................................................... 1
1.1 Problem Declaration ........................................................................................................ 3
1.1.1 Initiation Rate........................................................................................................... 3
1.1.2 Collaboration Quality ............................................................................................... 3
1.1.3 Target Domain ......................................................................................................... 4
1.2 Motivation........................................................................................................................ 6
1.3 What is the State of the Art Lacking? .............................................................................. 8
1.3.1 The Annoki Platform ................................................................................................ 9
1.3.2 Socs ........................................................................................................................ 11
1.3.3 Automated Recommendations .............................................................................. 12
1.4 Our Solution ................................................................................................................... 14
Chapter 2: Related Work and the State of the Art ........................................................................ 16
2.1 Improving Collaboration in Wikis................................................................................... 16
2.1.1 SuggestBot ............................................................................................................. 18
2.1.2 Annoki .................................................................................................................... 23
2.1.3 Socs ........................................................................................................................ 27
2.2 Other Relevant Wiki-Related Work ............................................................................... 31
2.2.1 Article Quality ........................................................................................................ 31
2.2.2 User Expertise Modeling ........................................................................................ 34
2.3 Recommendation Algorithms ........................................................................................ 35
2.3.1 Collaborative Filtering ............................................................................................ 37
2.3.2 Content-Based ....................................................................................................... 38
2.3.3 Hybrid Recommendation Algorithms .................................................................... 40
Chapter 3: Proposed Approach...................................................................................................... 42
3.1 Page Recommendation Algorithm ................................................................................. 42
3.1.1 Page Topics, Tags, Disciplines, and Keywords........................................................ 44
3.1.2 Explicit Page Ratings .............................................................................................. 46
3.1.3 Reputation and Expertise of Author(s) .................................................................. 47
3.1.4 Links Between Pages .............................................................................................. 51
vi
TABLE 5.27: RAIK STUDENT ACTIVITY LEVEL FOR EACH TRACKED ATTRIBUTE. THE NUMBER FOLLOWING
EACH ATTRIBUTE, I.E. (2), INDICATES THE NUMBER OF CLUSTERS USED TO CATEGORIZE THE
STUDENTS. SEE TABLE 5.25 FOR THE SCORE VALUES MAPPED TO EACH ACTIVITY LEVEL. ................173
TABLE 5.28: CLUSTERS AND MEMBERS FOR THE RAIK CLASS, BASED ON MINIMALIST VS. OVERACHIEVER
NET SCORE. COLLABORATION PHASE SCORE IS ALSO LISTED FOR COMPARISON. ASTERISKS DENOTE
USERS WHO DID NOT CONSENT TO THE USE OF ASSIGNMENT GRADE FOR THIS ANALYSIS. ............173
TABLE 5.29: MAS STUDENT ACTIVITY LEVEL FOR EACH TRACKED ATTRIBUTE. THE NUMBER FOLLOWING
EACH ATTRIBUTE, I.E. (2), INDICATES THE NUMBER OF CLUSTERS USED TO CATEGORIZE THE
STUDENTS. SEE TABLE 5.25 FOR THE SCORE VALUES MAPPED TO EACH ACTIVITY LEVEL. ................179
TABLE 5.30: CLUSTERS AND MEMBERS FOR THE MAS CLASS, BASED ON MINIMALIST VS. OVERACHIEVER
NET SCORE. COLLABORATION PHASE SCORE IS ALSO LISTED FOR COMPARISON..............................180
TABLE 5.31: RAIK UNIQUE PAGES COMMENTED/EDITED AND THE OVERLAP BETWEEN THE TWO ...........191
TABLE 5.32: RAIK UNIQUE PAGES COMMENTED/EDITED AND THE OVERLAP BETWEEN THE TWO, FOR
STUDENTS PLACED IN CLUSTER 0 .......................................................................................................192
TABLE 5.33: RAIK UNIQUE PAGES COMMENTED/EDITED AND THE OVERLAP BETWEEN THE TWO, FOR
STUDENTS PLACED IN CLUSTER 1 .......................................................................................................192
TABLE 5.34: RAIK UNIQUE PAGES COMMENTED/EDITED AND THE OVERLAP BETWEEN THE TWO FOR
STUDENTS PLACED IN CLUSTER 2 .......................................................................................................192
TABLE 5.35: MAS UNIQUE PAGES COMMENTED/EDITED AND THE OVERLAP BETWEEN THE TWO ...........196
TABLE 36: MAS UNIQUE PAGES COMMENTED/EDITED AND THE OVERLAP BETWEEN THE TWO, FOR
STUDENTS PLACED IN CLUSTER 0 .......................................................................................................196
TABLE 5.37: MAS UNIQUE PAGES COMMENTED/EDITED AND THE OVERLAP BETWEEN THE TWO, FOR
STUDENTS PLACED IN CLUSTER 1 .......................................................................................................197
TABLE 5.38: DISTINCT CONTRIBUTORS FOR EACH WIKI PAGE IN THE RAIK CLASS .....................................203
TABLE 5.39: RAIK CLIQUE OCCURRENCES AND CLIQUE STRENGTHS FOR GROUPS OF SIZE 2 WITH 2+
OCCURRENCES ....................................................................................................................................204
TABLE 5.40: RAIK CLIQUE OCCURRENCES AND CLIQUE STRENGTHS FOR GROUPS OF SIZE 3 WITH 2+
OCCURRENCES ....................................................................................................................................204
TABLE 5.41: DISTINCT CONTRIBUTORS FOR EACH WIKI PAGE IN THE MAS CLASS .....................................206
TABLE 5.42: POTENTIAL CLIQUES OF SIZE 2 FOR THE MAS CLASS...............................................................207
TABLE 5.43: EFFECTS OF VARIOUS CLASS/INSTRUCTOR-CONTROLLED FACTORS ON CLIQUES ..................210
TABLE 5.44: USER ACTIONS AND ASSOCIATED ATTRIBUTES FOCUSED UPON IN POST-HOC ANALYSES .....212
TABLE 5.45: PRIMARY FINDINGS WITHIN THE VARIOUS CHAPTER 5 ANALYSES .........................................214
TABLE A.1: TIME PERIODS AND CORRESPONDING WEIGHTS USED IN WEIGHTED HARMONIC MEAN
CALCULATIONS ...................................................................................................................................227
TABLE A.2: WEIGHTED COUNTS USED FOR CALCULATING USER INTERESTS. .............................................228
TABLE B.1: CLUSTER MEMBERSHIP FOR RAIK FOUR-CLUSTER RESULTS .....................................................229
TABLE B,2: ATTRIBUTE DETAILS FOR RAIK CLUSTER 0 .................................................................................230
TABLE B.3: ATTRIBUTE DETAILS FOR RAIK CLUSTER 1 .................................................................................230
TABLE B.4: ATTRIBUTE DETAILS FOR RAIK CLUSTER 2 .................................................................................230
TABLE B.5: ATTRIBUTE DETAILS FOR RAIK CLUSTER 3 .................................................................................230
1
Chapter 1: Introduction
Modern society has undergone a paradigm shift. In the past, businesses highly
valued the highly-skilled individual, the “rock stars” who could single-handedly carry
teams and corporations to greater heights through their ability, talent, knowledge, and
insight. But during the time span since then, focus had shifted more towards that of
teamwork and cooperation (Limerick and Cunningham 1993). Today’s society now
places increased emphasis on working jointly with others, whether it is in the classroom,
in the lab, in the workplace, or across the world (Karoly and Panis 2005). These
principles of increased collaboration and information sharing are even reflected in web
(2004). From these developments, it is evident that the technological trend for the near
"…a mutually beneficial and well-defined relationship entered into by two or more
definition of mutual relationships and goals; a jointly developed structure and shared
responsibility; mutual authority and accountability for success; and sharing of resources
and rewards."
ideas off of one another. Of the two, it is by far the older, dating back much further than
the technology needed for virtual collaboration. This form of collaboration is typically
face manner and typically arise offline from specific needs or from casual, unplanned
conversation (Kraut et al. 1988). While traditional collaboration could be augmented with
long-distance collaboration, e.g. via telephone, it wasn’t until relatively recent years that
meeting in person.
to carry out long-distance collaboration over the Internet. Virtual collaboration over this
medium makes use of tools such as audio and video conferencing, e-mail, forums, and
groups in lieu of physical proximity, though geographically localized groups may and do
use the tools to augment traditional collaboration. The new medium offers benefits over
(Kraut et al. 1988). Consequently, it may be considerably harder to build trust and rapport
collaboration for the research and education domains by improving initiation rate and
collaboration quality on wikis. The following paragraphs will describe these two terms in
greater detail before delving into the details of our chosen domain.
contributions throughout the duration of their involvement. While initiation rate can be
to qualify the contributions to distinguish it from “lesser” contributions. That is, we wish
to separate “collaboration” from mere “participation,” where the former requires a greater
degree of commitment and provides further benefit to the group. This distinction is also
made by Katz and Martin (1995), who define collaborators as “those who work together
on the research project throughout its duration or part of it, or who make frequent or
substantial contributions" and exclude “those who make only an occasional or relatively
in the collaboration process, participant satisfaction of the end product, and the quality of
the end product. Meier et al. (2007) further break down participant satisfaction in the
et al. (1997) also include participant perception of group cohesiveness in their evaluation
prominence in recent years, enabling people – either in small project groups or as part of
the wiki’s entire user base – to socially construct knowledge asynchronously on a wide
variety of topics (Forte and Bruckman 2007). Wikis generally consist of a collection of
interlinked pages and include basic functionality such as creating, updating, and deleting
for discussion. Participation is typically voluntary, and wikis are generally open for any
user to edit, although access can be restricted to prevent “vandalism” of pages. We are
particularly interested in targeting wiki usage since wikis embody the principles of
Vygotsky’s social constructivist theory: “the discovery of knowledge stems from its
social construction.” As stated by Cosley et al. (2007), “social science theory suggests
reducing the cost of contribution will increase [users’] motivation to participate.” With
few intelligent support tools for wikis available, the wiki domain holds many
knowledge, and highly specialized expertise.” Additionally, a single person no longer has
narrow area of research (Hara et al. 2003). This is reflected in the increase of the average
number of co-authors per paper during recent decades: Mattessich and Monsey (2001)
cite that the average number of co-authors on a paper rose from 3.9 to 8.3 between 1981
and 2001 in the Proceedings of the National Academy of Sciences of the United States of
America. This number can only continue to grow as technology improves, enabling
initiation rate and quality are both of importance. Regarding initiation rate, researchers
are often interested in discovering new, interesting connections and extensions from their
current work. As such, they often express great interest to new ideas and collaboration
quality is also of interest since it could provide reduced costs, improved productivity, and
opportunities for discovering new knowledge and exchanging information with experts
and peers.
a high degree of collaborative activity. Most, if not all, educational curricula now include
collaborative learning opportunities in the form of group projects and papers or some
other activities requiring students to learn and solve problems jointly. Recent research
shows that in addition to preparing students for the team environment in post-education
work, these activities also provide educational benefits including enhanced critical
6
thinking and increased interest and understanding (Gokhale 1995). To further support the
trend towards the development of collaboration tools, there is recent interest in computer-
supported collaborative learning for the educational setting, where the collaborative
are typically more important than initiation rate. Students may not necessarily be
concerned with initiation rate since their “participation” in the collaborative learning is
often mandated by the requirements of the assignment or activity, and the instructor may
assign particular groups for the exercise. However, collaboration quality may be of
greater concern to the students. Since their grades are at stake, they are motivated to
perform well and seek high-quality collaboration with other similarly-motivated peers.
On the other hand, instructors are interested in both improved initiation rate and
collaboration quality. In addition to the desire to see the students collaborating frequently
with one another, instructors are interested in seeing the students’ learning experiences
1.2 Motivation
Collaboration may be initiated and sought for multiple reasons, including
opportunities for reducing cost, improving productivity, discovering new knowledge, and
exchanging information with experts and peers within and across multiple disciplines and
domains. But on the other hand, poor collaborations often have costly consequences,
including inefficient use of resources, decreased productivity, reduced output quality, and
dissatisfaction amongst the members involved. In the worst cases, the group may lose
emerging as a hot trend, growing in significance as more virtual teams are formed
where team members can work separately on pieces of the “bigger picture.” Although
there may be great physical distances and few dependencies between them, they are still
readily accessible through the various tools for both synchronous and asynchronous
forums, wikis, etc. The importance of virtual collaboration will grow exponentially as
While collaboration is the key core of research, the means through which it
happens is the dissemination of ideas, data, findings, and the resulting discussion. The
this exposure can be seen as a benefit in and of itself, the key value is that it enables
Vygotsky’s oft-cited social constructivist theory on a larger scale (Anderson et al. 1997).
After data is published online, it can then be discovered by other labs, which can augment
the data to their own data sets, analyze and compare the data with their own findings, and
discuss the results with the originating lab. The process of sharing, discovering, and
discussing can lead to “big picture” connections between the seemingly disjoint data sets,
Although there are concerns about the electronic medium encumbering the
that today’s users are more tech- and Internet-savvy than in the past. Herring (2004)
8
states that users – both those growing up in the “Net Generation” and members of the
communication over the years. Having considerable experience and familiarity with the
technology and tools, these users are less hesitant and less reserved about contributing to
communication. In turn, their apparent comfort on the medium may encourage the more-
characteristic will allow for more widespread “buy in” and more effective online
partnership is motivating, people deliver higher-quality work in less time and at a lower
cost, leaving more time and resources for additional pursuits. Other benefits on the
interpersonal front are also available: improved networking and goodwill, and potential
for further future collaboration. In addition to these, Katz and Martin (1997) also mention
the aforementioned sharing and transfer of knowledge, skills, and techniques, potential
and Open Collaboration1 series is one particular avenue dedicated to sharing, discussing,
and advancing research and practice in the area. The most popular topics covered in the
proceedings are: 1) the evaluation of the effectiveness of using a wiki for web
such as the inclusion of semantic web concepts to create semantic wikis (e.g. Schaffert
2006) and trust and reputation aspects (e.g. Suh et al. 2008), and 3) the development of
new tools to enhance the wiki user experience. However, very few of these solutions are
developed to specifically target our goal of improving the wiki collaboration process (e.g.
Coslet et al. 2007, Tansey and Stroulia 2010). We thus widen our scope and also include
state of the art. In particular, we place particular emphasis on the user and data models
used in each approach and summarize their applicability (or lack thereof) to the wiki
setting.
Our work will focus on addressing the holes found in current user and data
modeling approaches for the wiki domain. For further detail on the works mentioned in
provided over traditional wikis include: 1) namespace-based access control and an easy-
1
http://www.wikisym.org
2
http://www.mediawiki.org
10
structure, page content, and user contributions; 4) sentence differencing; and 5) additional
One aspect that most pertains to our work is the tracking and the visualization of
specific user contributions. In addition to tracking insertions, deletions, and internal and
external links made, Annoki uses a sentence ownership mechanism to attribute sentences
of an article to specific authors. That is, the model of each user can be derived from the
actions they perform and their proportion of “ownership” of the entire wiki. Pages can
then be visualized in terms of the actions performed by each contributor to the article.
Another aspect that deserves particular mention is the “wiEGO” graphical page
structure editor and Annoki-specific page templates used to assist users with structuring
ideas and information content to wiki pages. Aside from the apparent benefit of aiding in
enabling classification via page type. Coupled with Annoki’s capability to annotate pages
with tags, this enables pages to be modeled based upon structure and associated keywords.
The models leveraged by the Annoki platform, being based in a wiki setting, are
completely applicable to our work. However, these models are not leveraged to their
fullest extent since the Annoki platform does not provide any intelligent features — for
example, their use is limited to displaying information for users to act upon (should they
choose to). The helpfulness of this functionality, excepting general usage numbers, was
not quantitatively reported. Although Annoki has been deployed and is in use at the time
11
collaboration. We are thus unable to determine how much collaboration initiation rate
1.3.2 Socs
Another approach to improving collaboration in this setting is by improving the
user interface to highlight information that is not readily accessible in traditional wikis.
This is exemplified in the Socs prototype developed by Atzenbeck and Hicks (2008), an
application for Mac OS that “serves as a means to express, store, and communicate social
information about people.” Socs provides social and group awareness in Wikipedia by 1)
providing a visualization of the authors contributing to each wiki page and the social
groups to which they belong, 2) linking authors with their other works (i.e. other pages
they have contributed towards), 3) enabling the user to flag authors of interest, and 4)
integrating with the Apple Address Book. The goal is that this functionality encourages
Since Socs is primarily a social application, its primary modeling occurs on the
user side, associating wiki contributors for a particular page with social groups that
potentially overlap. A visualization of these group associations can be seen in Figure 2.4.
Further, users are also linked to the pages that they’ve contributed towards, enabling easy
Unfortunately, the solution relies heavily on manual user action. For instance,
users manually flag specific authors as noteworthy and must manually request the system
12
to retrieve other pages that flagged authors have contributed towards. These “implicit
recommendations” are neither automated by the system nor ranked by any sort of
relevance.
are unable to determine how well it improves collaboration initiation rate and quality.
already exist, and they span a wide variety of targets, such as resources, products, and
people that may be of interest to the user. Much research has already been performed in
such as collaborative filtering (e.g. in Konstan et al. 1997 and Linden et al. 2003) and k-
nearest neighbors (e.g. Shepitsen et al. 2008). User-to-user recommendations have also
been leveraged for locating expertise and procuring help for specific tasks (e.g. Vassileva
et al.2003), and research for recommending users to users for social networking purposes
(e.g. Chen et al. 2009 and Guy et al. 2009) is also recently gaining traction.
Wikipedia3) have also been developed including expertise location by Demartini (2007)
locate experts in the Wikipedia user base, and user models are created by processing
contributors’ revisions to determine their areas of expertise as well as the level of their
3
http://wikipedia.org
13
While each recommender system builds and leverages user models to derive
recommendations from, these models are often limited in scope to the bare necessities
overlook wiki-specific factors (such as frequency and implicit quality of edits) and
synergy between factors in separate algorithms when applied to the wiki domain without
modifications.
In general, it can be said that these works focus on helping wiki users locate
expertise or interesting items. While these works can be leveraged to improve wiki
collaboration, this improvement is not the focus of the works themselves. Thus, they do
not directly address our targeted problem of improving collaboration rate and quality.
since it is one such recommendation-based tool that suggests wiki pages to contribute
towards. However, it is limited by its recommendation scope, i.e. pages that are: 1) not in
the top 1% of most frequently-edited articles and 2) are explicitly and manually flagged
by users as stubs, needing improvement, etc. It is also limited by a design that favors ease
of implementation over accuracy. The three intelligent algorithms used – based upon text
similarity, explicit links, and co-editing patterns –are combined via random selection.
Their work reports that 2.5 to 4.3 percent – roughly 30 to 40 out of 1150 – of the
recommended pages for each algorithm were followed and edited within two weeks of
Due to the approach taken, the user and page modeling performed, while
only user interests. Profiles are represented as a set of article titles and are implicitly
determined via the edits to article pages. Edits to non-articles and revision edits are
ignored, and each page is counted only once regardless of the number of edits made to it.
Page modeling is also limited, only factoring the article title and direct links to and from
other pages.
article, this tool only offers a modest success rate, though it should be noted that the
results reported are not relative to the users’ activity levels. The problem of collaboration
quality, i.e. the quality of the edits made or of the collaboration between the authoring
model for users and wiki data that includes the following factors from existing
approaches:
Page ratings
to recommend wiki pages. Details for both the models and the recommendation algorithm
implemented a user and data model specific to the wiki domain that leverages factors
used across individual, separate approaches not yet combined in this manner. Second, we
pages to the user. Third, we developed an intelligent wiki within the Biofinity Project
(Scott et al. 2008) – a software framework that unifies biodiversity and genomic data
across multiple, varied sources – to use and gather data for our models. Fourth, our
empirical evaluation of the models provides additional data and insights regarding their
The rest of the thesis is organized in the following manner. We will first review the
current state of the art and related works in Chapter 2 before introducing and detailing
our proposed approach in Chapter 3. Chapter 4 then describes our implementation of the
Biofinity intelligent wiki and of our approach, and Chapter 5 details our experimental
methodology and the results and analysis of the experiments performed, respectively.
Finally, Chapter 6 closes out the thesis with the conclusions drawn from the results and
on the user modeling, data modeling, and recommendation techniques used in the
solutions summarized in Chapter 1.3. We can broadly categorize these solutions into one
of three areas:
(Chapter 2.1)
2.2)
work for the user to perform (e.g. recommendations and task routing), 2) reducing the
cognitive cost of creating and maintaining wiki content (e.g. content management), and
relationship visualization). We have thus identified three research works, one in each of
They were chosen because they: 1) specifically target the goal of improving
collaboration in a wiki setting, and 2) contain some mechanism for users to explore and
discover new information (which may or may not encourage collaboration directly).
pages to contribute towards, and it is considered the most relevant to our work of the
three intelligent algorithms based on text similarity, links between pages, and co-editing
profiles. Consequently, SuggestBot’s user and page models only contain factors relevant
are a few design decisions made which are not applicable to its use in a more general
wiki. First, one of the intelligent algorithms hinges upon the use of a specific database (i.e.
leverages MySQL 4.1’s built-in fulltext search). Second, the pool of candidate articles for
the algorithms is limited to those manually marked by users as needing work via
reasonable size, this is not generalizable to other wikis since they do not follow the same
standards for bots limit the SuggestBot to the use of relatively simple algorithms that
management, and visualization. It contains three features that may indirectly promote
18
collaboration: the tag mechanism and corresponding tag cloud visualization, the
WikiMap, and the Wiki Contribution Analysis. While the tags and WikiMap facilitate
user exploration of the wiki for related elements, this exploration process is entirely
manual, and the Annoki implementation of tags offers little over the basic tagging
functionality commonly used in Web 2.0 applications (i.e. not automated and without any
prioritization of results). The Wiki Contribution Analysis component displays editing and
ownership statistics for each article, which could motivate users to periodically review
Overall, the platform lacks “intelligent” features that could further benefit collaboration.
The mentioned Annoki platform features are described in greater detail in Chapter 2.1.2.
social and group awareness and facilitating communication in wikis via an application for
the Mac OS. Its “social space” and “awareness features” increase the visibility of the
contributions of authors that are noteworthy to the user, and integration with the Apple
Address Book facilitates contact with them when the need for collaboration arises. As
with the Annoki platform, Socs lacks “intelligent” features that could further benefit
collaboration. While the application displays all page contributors for a wiki page and
prioritize authors to contact. The Socs functionality is covered in greater depth in Chapter
2.1.3.
2.1.1 SuggestBot
As described previously, the SuggestBot developed by Cosley et al. (2007) is a
recommendation-based tool that suggests wiki pages to contribute towards. It limits its
19
recommendations to Wikipedia pages manually marked by users with the flags in Table
2.1:
These flags constitute the most common types of work needed on the articles.
Additionally, the authors exclude pages that are already frequently edited (i.e. in the top
recommendation scope to articles that are known to be in need of work. That is, the
SuggestBot does not need to algorithmically determine whether a wiki page needs editing.
of articles in the user’s editing profile into keywords, and 2) using the keywords in a
search against the full text of articles using MySQL 4.1’s built-in fulltext search feature4.
The recommendation set returned is ordered based on the determined relevance from the
search algorithm, which uses a modified version of the term frequency-inverse document
frequency method.
4
http://dev.mysql.com/doc/refman/4.1/en/fulltext-search.html
20
articles in the user’s editing profile, representing Wikipedia pages as nodes and the links
traversal with loops and node revisiting allowed, starting from the articles that the user
has edited. Scores are assigned to pages by counting the number of times they have been
reached when the algorithm ends, and the recommendation set returned is ordered based
Chapter 2.3.1) variant to recommend pages that authors similar to the user have edited.
This version of the algorithm differs from traditional collaborative filtering in a few
aspects. First, it uses editing profiles rather than ratings to calculate similarity between
users since Wikipedia does not use a ratings system. Second, an author is considered as a
“neighbor” of the user if any of the pages in its editing profile is also in the user’s profile.
Third, the algorithm uses Jaccard similarity instead of the more-common similarity
measures, such as cosine similarity and Pearson correlation. The recommendation set
returned is then ordered based on the score calculated for each article.
The recommendation set returned to the user consists of 34 article slots: 19 stubs
and 3 of each of the remaining five flag types. The articles to place in each slot are
chosen in the following manner. First, a recommender is randomly chosen from four
approaches: the three intelligent algorithms and random selection. The slot is then filled
with the first article that: 1) matches the flag type for the slot, 2)has not already been used
in another slot, and 3) is not in the top 1% of frequently edited articles. If the selected
randomly chosen. Figure 2.1 illustrates the interface used to display these
strongly tied to the recommendation algorithms used. The user model in this particular
user’s editing profile. Cosley et al. represent this as the set of titles for the articles that the
user has edited, ignoring minor revisions (e.g. vandalism reverts), edits to non-article
pages, and the number of edits to each page. This user model is leveraged in the text
containing only the article title, body text, and intra-wiki links to and from the article.
22
Due to its targeted deployment to Wikipedia, there are a few design decisions
general wiki for ease of implementation and on-line calculation speed. For instance, the
pool of candidate articles for the algorithms is limited to those not in the top 1% of most
edited articles and to those manually marked by users as needing work via Wikipedia-
specific notation. While this reduces the recommendation scope to a more-reasonable size,
less hesitant to perform, such as tagging articles with the work types in Table 2.1 and
providing preliminary stub content for “red links.” The Wikipedia community standards
for bots also contribute towards SuggestBot’s limited accuracy in that they motivate the
bot’s design focus on simplicity. This focus, coupled with Wikipedia’s limited action
tracking, leads to the selective exclusion of some tracked features, such as excluding edit
counts from users’ editing profiles, which in turn may reduce a recommendation’s
relevance. For this particular example, disregarding edit counts for each article may
As previously mentioned, the results reported by Cosley et al. focus only on what
article. Their work reports that 2.5 to 4.3 percent (roughly 30 to 40 out of 1150) of the
recommendation pages for each algorithm were followed and edited within two weeks of
being presented to the user – a modest success rate. The problem of collaboration quality,
i.e. the quality of the edits made or of the collaboration between the authoring users, was
2.1.2 Annoki
The Annoki platform by Tansey and Stroulia (2010) is a suite of MediaWiki
projects). The functionality that the platform provides over traditional wikis include: 1)
visualizations for overall wiki structure, page content, and user contributions; and 4)
additional features such as calendar extensions and LaTeX export. In short, Annoki
enhancing tools. A few features – tags, WikiMap, and Wiki Contribution Analysis –
improve awareness of peer activity and facilitates information discovery. These are of
Tags in Annoki are largely implemented in a similar manner to tags in Web 2.0
applications. Users may annotate pages with tags to associate them with particular
categories of pages. Each tag has its own wiki page which is automatically populated
with links to all wiki pages marked with the same tag. This is akin to Wikipedia’s
automatically-generated “category pages” for locating other items sharing the same
category. Annoki also features a wiki-level tag cloud which displays all the tags used in
the wiki in varying sizes based on frequency of use. While simple, this mechanism
WikiMap (in Figure 2.2) is a tool that visualizes the structure of a wiki, displaying
how the elements of the wiki (e.g. pages, users, and tags) are related to the particular
centered element via connected nodes. The user may also click on any of the items to
24
navigate to the corresponding page or re-center the map on a new element. The links
connecting the center element are color-coded based upon element type, and the size of
the node reflects the “importance” of the element. For a user node, this corresponds to the
number of edits made; for a tag, this corresponds to the frequency of its use; and for
pages, the user can choose between weighting based on the number of revisions, the
number of contributing authors, the number of page views, and the number of links
Figure 2.2: An example of Annoki’sWikiMap, showing author, page, and category nodes, centered on the page
“Main page” (Tansey and Stroulia, 2010)
contributions of wiki users for particular articles as shown in Figure 2.3. In addition to
displaying statistics on insertions, deletions, and internal and external links made, Annoki
introduces the notion of sentence ownership to attribute sentences to specific authors, and
the number of sentences owned in the article is also displayed. Sentence ownership is
given to a revision author if: 1) the sentence written is not in a previous revision, or 2) the
platform and as base for other systems that require domain-specific features. In the
former use case, ten instances of the platform were installed for “various groups.” The
heaviest use was seen by the Software Engineering Research Lab (SERL) at the
University of Alberta over the course of two years. Table 2.2 lists some usage statistics
While Annoki does not make use of user and data models for intelligent user
interface content, it does have the potential to build such models and apply them in future
extensions. User modeling includes wiki actions tracked (e.g. pages viewed and edited)
as well as specifics of the edits (e.g. sentence ownership, links added/deleted). Such low-
level tracking could be leveraged for user expertise modeling, described later in Chapter
2.2.2. Users’ interests can also be implicitly modeled through their “links” to other wiki
content, as mapped by the WikiMap feature. Data models in Annoki include: links to and
from other wiki pages and users, as shown in the WikiMap; keywords and associated
topics through the tagging functionality; and a particular page “type” based on any
templates or wiEGO graphical page structures used to create the page. Coupled with the
Figure 2.3: Graphical display of wiki page contributions in Annoki (Tansey and Stroulia, 2010)
that could be leveraged to further improve ease of use and collaboration in the system. It
is more a collection of tools that facilitate wiki management and content creation than it
is a tool for directly promoting collaboration between its users, and the qualitative results
reported reflect this. The usefulness of the namespace-based access control mechanism,
the wiEGO visualization, and simplified template creation mechanism in particular were
highlighted over the other features. It should be noted, however, that the template feature
gave rise to a powerful collaboration tool. By creating and making use of a template for
academic papers, SERL was able to circulate interesting or useful papers throughout the
group, using the paper’s corresponding wiki page to share thoughts and identify potential
discussion partners.
27
Now, the three features specifically covered – the tag mechanism and
corresponding tag cloud visualization, the WikiMap, and the Wiki Contribution Analysis
– can all indirectly lead to improved collaboration, but require user initiative and
motivation to do so. In the case of tags and the WikiMap, user exploration and navigation
visualizations. However, there is no distinction or prioritization for pages that may need
component could be used to motivate users to periodically review the pages they’ve
contributed towards or encourage others to increase participation. But again, this requires
human motivation to make use of the information and contact other users.
The performance of the Annoki platform overall was not thoroughly reported
aside from minor quantitative usage data and qualitative descriptions of which features
were particularly helpful. Although it has been deployed and is currently in use for a few
collaboration. We are thus unable to determine how much collaboration initiation rate
2.1.3 Socs
The Socs prototype developed by Atzenbeck and Hicks (2008) is an application
for Mac OS that “serves as a means to express, store, and communicate social
through increased social and group awareness. Socs provides these in Wikipedia by 1)
providing a social space visualization of the authors contributing to each wiki page and
the social groups to which they belong, 2) retrieving information on authors’ activity
28
levels (i.e. how frequently s/he modified the page), 3) enabling the user to flag authors of
interest, and 4) integrating with the Apple Address Book framework. The hope is that the
among wiki page authors, which increases understanding of author intentions and
The cornerstone to the Socs prototype is the social space visualization, which
presents the user’s people and groups of interest in a 2D area as seen in Figure 2.4.
People are represented by markers, and groups are represented by colored rectangles.
person on the social space – rather, groups and people on the space are limited to those
already known by the user (i.e. in the user’s address book). Placement of the markers is
then determined based on the groups that the user has manually associated them with.
The space also utilizes other visual cues such as distance, alignment, color, and size to
convey additional information to the user. The social space integrates with the Apple
Address Book and Wiki (Page) Authors list via drag and drop functionality –people and
groups from the address book and article authors from the authors list can be dropped
into the user’s social space to visualize the relations between them and highlight their
participation on wiki pages. Any changes to the social space (i.e. insertion and deletion of
members and groups) are reflected in the system-wide address book, and contact with
29
clicked.
The second component to the Socs prototype is its awareness features. When the
user navigates to a Wikipedia page (or another compatible website), Socs obtains its
contributors and populates them in a list along with activity level (i.e. number of
revisions made) for easy viewing. If an author is already in the Socs system, it is
indicated in the “Loc” column of the window, and authors that are in the user’s current
social space are also highlighted in the social space window. By highlighting authors in
this manner, the user is: 1) made aware of acquaintances that took part in the article and
the groups to which they belong and 2) provided with a simplified mechanism for
contacting them if needed. The cost of communication is reduced since the article authors
Since Socs is primarily a social application, its primary modeling occurs on the
user side, associating wiki contributors with various social groups that potentially overlap.
Further, users are also linked to the pages that they’ve contributed towards, enabling easy
The proposed solution’s primary shortcoming, just like that described for the
Annoki platform, is its lack of intelligent support. Since the application only highlights
authors that manually marked by the user, the potential benefits of the tool is diminished
since it does not identify, display, or recommend new social relations to groups or people
that the user does not yet have on the social space. That is, there is no guided discovery of
30
authors or social groups. While “strangers” may be manually added to the social space
(and consequently, the user’s address book) from the Wiki Authors window, the user is
component of our approach, which recommends new social relations via suggesting
are unable to determine how well it improves collaboration initiation rate and quality.
Figure 2.3: Screenshot of Socs social space, web browser, wiki authors list, address book (Atzenbeck and Hicks,
2008)
31
improving wiki collaboration, other wiki-related research may be relevant to our goal of
article quality and user expertise is especially relevant to our interests, since their
inclusion in our user and data models may improve the accuracy of our recommendations.
We have included some of these measures and the ideas that they are based on in our own
on Lih’s (2004) “rigor” and the notion of page quality based on contributing users’
approach of improving collaboration since it: 1) helps determine good quality articles to
highlight (e.g. for recommending articles to view) and 2) helps determine which articles
are of poorer quality and need work (e.g. for recommending articles to edit) (Huet
al.2007). The approaches to calculating this can be broadly categorized based on the
metadata-based quality metrics and the article content-based quality models of Hu et al.
(2007).
Based on Metadata
o Diversity (Lih 2004) – the number of unique editors for the page.
contributing authors and the amount each author contributed to the article.
o PeerReview (Hu et al. 2007) – Basic model with text “review”; all
text review; text that is closer to the revision author’s contribution is more
likely to be reviewed.
Lih (2004) proposes two basic methods for benchmarking article quality based
strictly upon metadata, i.e. without analyzing the content of the article: rigor and diversity.
Rigor is the number of edits that the article has undergone, and its importance is based on
the assumption that an article that has been edited more times undergoes a “deeper
treatment of the subject or more scrutiny of the content.” Diversity is the number of
unique authors contributing to the article, and greater diversity for an article is indicative
of “more voices and different points of view” on its subject. Lih proposes finding
benchmark values, i.e. high quality thresholds, for these measures by calculating the
quality as a function of the expertise of its contributing authors: Basic, PeerReview, and
ProbReview. The Basic model is based upon the assumption that higher expertise authors
leads to a better quality article. An article’s quality is then the sum of the expertise of its
contributing authors, with each author’s expertise weighted by the amount s/he has
contributed to the page. However, an author’s expertise is also based on the quality of the
33
pages that s/he contributed towards – thus, the two have a circular relation and reinforce
one another. From this setup, values for quality and expertise are then calculated by first
initializing them to a value and then iteratively computing them until they converge.
PeerReview and ProbReview differ from Basic in that it introduces the notion of
between revisions is considered to be reviewed and implicitly accepted, and the author’s
expertise is factored into the quality of the existing text. In the PeerReview model, it is
assumed that all unmodified text is reviewed by the author. However, this assumption is
not particularly accurate – users who contribute minor changes or contribute changes to
only a specific area of the article. The ProbReview accounts for this by adding a
probabilistic element to the “review” of unmodified text – it assumes that the unmodified
text that is closer to the author’s contributions are more likely to be reviewed than text
These measures of article quality are related to our work since we have
incorporated ideas suggested in both Lih’s and Hu et al.’s works in our data models.
Lih’s rigor measurement (i.e. the number of edits) is used directly in our algorithm when
calculating the recommendation score of the article due to its ease of implementation. We
currently choose to exclude diversity from our data model since collaboration in our
target domain is typically carried out in groups of a fixed size during its primary
development. We also use Hu et al.’s idea of calculating article quality based on authors’
expertise and the proportion of their contributions to the article. However, we use an
alternative to convergence between article quality scores and author expertise, which may
detailed in Chapter 2.2.2 and Chapter 3, and our recommendation algorithm is described
in Chapter 3.
expertise, either through explicit feedback provided by other users or implicitly through
the user’s actions in the environment. For a wiki, this is often derived primarily from the
user’s contributions to wiki pages. In models that account for both page quality and user
expertise, the two reinforce one another: the collective expertise of page authors
contribute towards a page’s quality, and the quality of each page in the user’s editing
calculations of article quality. In the Basic model, user expertise is the sum of the
qualities of the articles that the author has contributed towards, with each contributing
term being discounted by the proportion of the text not authored by the user. That is, the
quality of each article in the author’s editing profile is multiplied by the percentage of the
author’s contribution. In the other two models, the expertise of users who have “reviewed”
Similar to Hu et al.’s models for wiki article quality, the notion of author expertise
is included in our user model and plays a role in our recommendation algorithm. As
that of PeerReview – text of high quality will be left unchanged (i.e. reviewed and
35
al. base this on the expertise of the authors who have “reviewed” the text, and this
time and does not require finding convergence. Contribution longevity and its use within
Research in this area has largely been centered upon its use in e-commerce to suggest
items for the user to purchase (e.g. on commercial websites) or on news and other special
Collaborative filtering leverages information from people similar to the target user
inaccurate recommendations for new users and new items, and inaccurate
items that the user has liked. While it generally lacks the same weaknesses as
collaborative filtering, it has its own distinct limitations, including the need for a large
recommendations. Our approach also contains content-based elements, but is not a pure
content-based algorithm.
algorithms or elements from those algorithms to generate recommendations, with the idea
that the varied strengths of the components compensate for their individual weaknesses.
While recommendations from these algorithms have higher accuracy than their pure
elements from Lih’s rigor (Chapter 2.2.1) and Hu et al.’s page quality and user expertise
(Chapters 2.2.1 and 2.2.2) in addition to elements from other algorithms mentioned in
this subsection.
not been as thoroughly explored as their use in the previously mentioned domains.
approaches, one of which is based on collaborative filtering and another is content based;
and the works of Durao and Dolog (2009) and Sruirai et al. (2009)which utilizes a
The contents of the user and data models leveraged by the recommenders are
generally limited to only the requisite attributes needed to generate the recommendations,
i.e. data used during the computation. Consequently, currently-popular algorithms may
overlook wiki-specific factors (such as frequency and implicit quality of edits) and the
37
synergy between factors in separate algorithms when applied to the wiki domain without
modifications. While specific model contents are dependent upon the approach taken in
hybrid recommenders, general statements can be made of user and data models for
methods as predictions on the utility of an item for a particular user based on ratings that
similar users have given it. That is, it recommends items that users with similar
preferences – “neighbors” – have found favorable. Sarwar et al. (2000) generalize the
collaborative filtering process into three parts: 1) the representation of input data, 2)
the number of users and N is the number of items in the system. Each entry denotes a
user’s affinity (e.g. through rating, number of views, etc.) for the item. The biggest
differences between the various CF-based algorithms then lie in the techniques used in
neighborhood formation and recommendation generation. For example, the popular user-
based top-N variant of CF uses Pearson correlation to determine the k users most similar
to the target user. Predicted ratings are then calculated by taking a weighted average of
the ratings given by these k neighbors, and the top N items with the highest ratings are
recommended to the target user. As another example, one of the recommenders used in
This recommendation approach is driven largely by its user modeling. The most
common ratings-based CF approach uses a model where user interests are represented by
the set of ratings provided throughout their entire history within the system.
“neighborhood score” for items not yet rated by the target user. The “rating” aspect of
representing user interest can be swapped out or augmented with other indicators
available in the application domain, such as item views, edits, etc. Since this approach is
not driven directly by page content, it does not leverage a data model in its computation.
recommendations for new users and new items (i.e. new users and items lack the history
recommendations due to sparsity of ratings (i.e. there is a lack of jointly rated items due
to a very large number of items in the system relative to the number of items rated by
performance when operating on very large and very sparse matrices. Performance can be
relevant to our work since our algorithm contains collaborative filtering-like aspects in
greater detail.
2.3.2 Content-Based
Content-based recommendation methods leverage the features or characteristics
of an item to predict whether the target user would like it, based on how favorably the
39
user has received items with similar features. That is, in contrast to collaborative filtering
between items. The intuition is that users are more likely to enjoy items that have similar
qualities to items that the user already likes. For instance, a person who enjoys the Harry
Potter series may be more likely to enjoy The Lord of the Rings than Lawrence of Arabia
since the former arguably has more in common with it than the latter.
this category of algorithm, including those based on clustering (e.g. Shepitsen et al. 2008)
and Tuzhilin (2005) aggregates the target user’s tastes into a feature vector and finds its
This recommender type utilizes both a user and data model in its calculations.
Here, user interests and data content are represented as a subset of some set of keywords
or tags global to the entire system. Either can be built explicitly through manual listing of
interests and related topics, or implicitly through text analysis of page content viewed.
feature sets that must be known beforehand if automatic feature extraction is not possible
items containing features the user favors with little chance for “serendipitous”
approach is relevant to our work since our algorithm contains content-based aspects in
recommendation algorithms that lack the same weaknesses, e.g. basing the collaborative
(Adomavicius and Tuzhilin 2005). Adomavcius and Tuzhilin (2005) specify three
from multiple algorithms separately and combining their results, 2) adding elements from
model that incorporates elements from multiple algorithms. Cosley et al.’s SuggestBot is
one example, leveraging four different recommenders combined via the first approach
(see Chapter 2.1.1 for more details). Experimental results comparing pure collaborative
confirmed that the performance of the hybrid CF algorithms provides superior accuracy
at the cost of additional computational complexity (Melville et al. 2002, and Han and
Karypis 2005).
This is particularly relevant to our work since we leverage the third hybrid
such as the identification of similar pages via keywords, collaborative filtering elements
41
such as the identification of peers with similar interests, and other elements such as
author expertise, ratings, and other page metadata to suggest pages to view and edit.
Rather than using only one of the “pure” algorithms previously described, we implement
recommendation. We appraise the cost of a false positive in this domain as greater than
the cost of a false positive for recreational browsing due to the increased costs and
potential losses for poor quality collaborations, as outlined in Chapter 1. This places
in a wiki, and thus a single element on its own is not sufficient to provide accurate
recommendations. Chapter 3 justifies our decision and describes our hybrid algorithm
improving wiki collaboration either lack intelligent features that adapt to the users’
profiles or fail to address both collaboration initiation rate and quality. We thus propose
our own hybrid recommendation algorithm that leverages and unifies aspects of directly
and tangentially related works to address these problems. The result is an algorithm that
filtering, 3) links between pages, 4) author reputation and expertise, and 5) the number of
page views and edits to provide recommendations that are relevant to user (and thus
Recommendation feature added to support the use of the Biofinity Intelligent Wiki in a
classroom setting.
a weighted mean of the individual component scores based on commonly used attributes
Links between pages (e.g. Cosley et al. 2010, Page et al. 1998, etc.)
43
Number of page views and edits (e.g. Lih 2004, [older search engines], etc.)
These can be divided into two categories based on whether they contribute
towards determining a page's relevance to the target user and its quality:
o Keywords-/tags-based similarity
o User expertise
Note that ratings-based collaborative filtering can be considered to fall into both
categories – the ratings-based aspect determines page quality whereas the collaborative
Attribute scores: [ ]
3.1. After calculating the page score, the top n highest-scoring pages will then be
Attribute Weight
Page Tags and Keywords 0.25
Explicit Page Ratings 0.25
Author Reputation/Expertise 0.25
Links between Pages 0.15
Number of Page Views 0.05
Number of Page Edits 0.05
Table 3.1: Weights for each attribute in the weighted sum
The following subsections detail the values of these individual weights and the
recommendation algorithms (Shepitsen et al. 2008). These are often utilized in the
following two ways: directly matching the terms to the target user’s interests and
indirectly matching related terms to the user’s interests. Additionally, each topic may
recommendations (2008). It is selected since it is designed for the social tagging domain
and makes use of the ideas mentioned in the previous paragraph. The algorithm is
detailed as follows:
45
1. Calculate the cosine similarity between the user’s interests and each resource:
∑ ( ) ( )
( ) ( ) , where
√∑ ( ) √∑ ( )
u and r are vectors over the set of tags, with u representing the user’s
tf(t,v) is the tag frequency of tag t in vector v - for wiki pages, a tag
It should be noted that T can grow to be fairly large as the system grows, with the
number of tags in the system being orders of magnitude larger than the number of pages
and users. Further, not all tags will be relevant to all pages and users, resulting in
relatively sparse vectors. Since tags that aren’t relevant to the page or user do not figure
into the calculation, we can limit the iterations to the union between the user’s interests
( ) ( )
( ) ( )
( ) ∑ ( ) ( )
46
Here, Tags(i) is defined to be the set of tags that an item i is associated with,
where i is either a resource r or cluster c. Similarly, Interests(u) is the set of tags that a
user u is observed to have interest in. We compute this by counting the tags associated
with the pages that the user created, viewed, edited, positively rated, and discussed.
( ) ( ) ( ) where
Details for how the tag clusters used in Step 2 of the procedure are found, as well
A key assumption that the algorithm had is that the recommendation is generated
for single-tag queries – that is, the vector of user interests u only contains a single tag. It
interests, and simply iterating this process over all user-interested tags may not scale up
well. We thus adapted the algorithm to cover the entire spectrum of the user’s interests.
The weight for this attribute will be initially set to 0.25 due to the relative
importance of topics, etc. in determining whether a page is suitable to the target user.
explicit ratings by users are generally more accurate than those determined through
the explicit page ratings provided by other users. There is a possibility for frustration bias
to factor into the ratings if the system actively and persistently queries the user to obtain
these ratings. However, this bias can be ignored since the ratings are voluntarily provided.
Our system used a binary voting system of “Likes” and “Dislikes.” The net page
rating for this particular system is then a simple difference between the number of Likes
Since a page’s contents change over time as users make revisions, it is possible
that older ratings are not indicative of contemporary opinion towards it. We will thus
weight the raw net page rating according to when it was made relative to the date the
collection of discrete time periods where all ratings in the same period receive the same
weight.
We will use a weighted harmonic mean to calculate the time-adjusted page rating,
𝑊𝐻𝑀( )
The weight for this attribute will be initially set to 0.25 due to the
of a user. This is analogous to trusting and valuing the opinions of domain experts.
48
Similarly, the reputation and expertise of the contributing authors should be considered
We can trace the page content to the users responsible for each contribution by
successively “diff-ing” each revision to the page to determine the changes made with
each one. The page reputation derived from author reputations is then calculated as:
∑ ( ) , where
( ) ( ) ( )
Expertise and Rating functions are defined in chapters 3.1.3.1 and 3.1.3.2.
The weight for this attribute will be initially set to 0.25 due to the perceived
Depending on the target user, similar levels of expertise (i.e. a peer relationship) or
our system, we define expertise to be a quality inherent in the revision contributions that
the users make, distinguishing it from participation which encompasses any sort of action
49
the user takes within the wiki. The implicit indicators of expertise that we consider
include: views per contribution authored, ratings towards contributions authored, and
( ) ∑ ( )
the views and ratings received by the page. This is essentially done by scaling the views
and ratings that a particular revision has received by the proportion of the user’s
defined as:
( ) ∑ ∑ ( ) where
and i-1
the weighted harmonic mean of the page rating between revisions i and i-1
50
surviving subsequent revisions, this weight increases with revisions further away
is the proportion of target user u’s contribution towards revision i (i.e. the
Table 3.2 below details the weights and value used for each attribute. The weights
for each contribution are preliminarily set based upon their perceived importance in
determining expertise.
attr ( )
Views .5 Number of views the page received between revisions i and i-1,
normalized by total number of views page received
Ratings .5 Average rating per user using ratings received between revisions i
and i-1
Table 3.2: Weights and values for attributes used in expertise calculation
on the binary scale of “Likes” and “Dislikes.” The net user rating for this rating scheme is
then a simple difference between the number of Likes and the number of Dislikes
Since opinions on users may change over time as they improve and make
contributions, it is possible that older ratings are not indicative of their current
performance. We will thus weight the raw net user rating according to when it was made,
relative to the date the calculations are performed. Again, we will consider time as a
51
collection of discrete time periods where all ratings in the same period receive the same
weight.
We will use a weighted harmonic mean to calculate the time-adjusted user rating,
𝑊𝐻𝑀( )
Where is the net rating for user , and the WHM function is as defined in Appendix
A.
use of a modified version of the publically available PageRank algorithm (Page et al.
calculating the likelihood that a user browsing at random will reach it. That is, a page’s
importance is proportional to its in-bound links and inversely proportional to its outbound
links. Essentially, we consider a page important if many other pages link to it.
( )
( ) ∑ , where:
evaluated (Page et al. 1998). The post-convergence values in I are the PageRanks
of each .
The weight for this attribute will be initially set to 0.15 , which is lower
than the previous weights assigned thus far. This is due to the fact that the PageRank
algorithm, on its own, does not consider the relevance of the connection between pages.
While this can allow for the serendipitous discovery of strongly-linked concepts that
users are unaware of, a recommendation is more likely to be followed if its relation to
some degree be indicative of the page quality. While its accuracy and precision for
role in identifying which pages may be considered essential reading by the user base.
Like with explicit page ratings, we consider recent page views to be of greater
importance than older page views. We will discount the number of views with time using
the same weighting as for explicit ratings. The weighted harmonic mean is then
𝑊𝐻𝑀( )
53
Where is the number of views for page , and WHM is as defined in Appendix A.
The weight for this attribute will be initially set to 0.05, which is
considerably lower than the previous weights assigned thus far. This is due to the facts
that: 1) the number of page views can be easily manipulated, i.e. artificially inflated, and
2) early search engines using this attribute to return search results were not particularly
successful.
be of higher quality after being refined many times, and the content on a frequently edited
page may arguably be considered “fresher” than those updated less frequently. However,
a high or low edit count may hold negative implications. For instance, a high edit count
may be indicative of less value contributed per edit. Similarly, a low edit count may be
indicative of an abandoned page when instead the page may be relatively “complete.”
We thus use the number of page edits as the recommendation factor. Like with the
number of page views and the explicit page ratings, we consider the more-recent edit
counts to be of greater importance than older edit counts, and we will thus weight the
time-adjusted page edits in a similar manner to views and ratings. The weighted harmonic
𝑊𝐻𝑀( )
Where is the number of edits for page , and the WHM function is as defined in
Appendix A.
54
The weight for this attribute will be initially set to 0.05, which is
considerably lower than the previous weights assigned thus far. This is due to the facts
that: 1) the number of page edits can be easily manipulated, i.e. artificially inflated; 2) the
quality/value added of the edit is not considered, i.e. the edits may primarily be aesthetic
or minor edits; and 3) the attribute has the uncertain implications previously described.
upon keywords located on the page currently viewed or edited. The goal of this feature is
to improve student collaboration via easy access to peripheral information that can: 1)
improve student comprehension of the page and related topics, and 2) aid the student in
contributing content during the early revisions of the page. Due to time constraints in
deploying the feature prior to student use, we opted for a fast, basic approach on this
for each revision of each Biofinity wiki page, and 3) ordering and presenting the results
to the user.
breadth-first traversal of Wikipedia pages, starting from arbitrary root pages. The titles of
each page visited are added to the dictionary with common stop words filtered out.
Duplicates (i.e. pages linked to by more than one page) are added only once, and redirects
55
(i.e. aliases and plural versions of pages) are linked to their target pages before both the
redirect and target pages are added. Separate dictionaries are created for the biology lab
page count towards the redirect target’s keyword/phrase. A keyword/phrase’s score for a
revision is the value of its counter at the end of the scanning and matching process.
ordered by their counts (as detailed in 3.2.2) with tie breakers handled by the word length
of the keyword/phrase.
56
Chapter 4: Implementation
To fulfill our goals of obtaining wiki usage data and creating a framework to
proprietary, full-featured wiki that integrates with the Biofinity Project. The wiki,
currently dubbed as the Biofinity Intelligent Wiki, supports the following basic features:
Page search – indexing pages based on content and retrieving them based upon
user query
File upload and download – adding and retrieving files such as images,
The wiki also supports the following Web 2.0 and social features:
Page tagging – associate “tags” or key words with a wiki page to denote topics,
relevant categories,
Page ratings –express opinions on overall quality of a page’s contents via a 1-5
scale
Page sharing via Facebook, Twitter, and the intra-wiki framework – share
pages to other users within the wiki, or to other social media outlets such as
Page and user modeling – create and maintain a model of each page and user in
monitor and track user activity, with potential to carry out further autonomous
algorithms/techniques
The latter two intelligent features– user tracking and recommendation framework
– are the ones that directly enable the investigation of the thesis topic. In particular, the
user tracking feature is the cornerstone that provides data for our analyses in the next
chapter.
Since the wiki is integrated with the Biofinity Project and is to be running on the
same server, we are constrained in the server and database software used. Specifically,
the Biofinity core ran using Glassfish v3 and a MySQL database, and the wiki was
designed to operate in the same environment. Additionally, it was required that the wiki
be encapsulated as a separate project and be packaged into a separate WAR file for easy
deployment to the Glassfish server. By separating the wiki in this manner, changes to the
wiki would not require changes to the core Biofinity site and vice-versa.
While the core Biofinity Project was written in Scala and leveraged the Lift
framework, we wanted to use a more general and common language for the wiki in order
to ease its development and to ease the implementation of future work for it. Java, HTML,
58
and Javascript are an attractive alternative, since they are among the most common and
popular tools used for web development on both the server and client ends. The decision
is further simplified with the Google Web Toolkit (GWT), which enables the
development of web applications written completely in Java, compiling the source files
into equivalent HTML and Javascript code. These factors, combined with our familiarity
with the language, led us to write the wiki almost completely in Java.
To summarize, the core technologies used by the Biofinity Intelligent Wiki are:
Glassfish v3
MySQL
Java EE 6
The rest of the chapter is arranged in the following manner. First, we will briefly
describe the overall architecture of the Biofinity Intelligent Wiki in Chapter 4.1,
including the interactions between the wiki client, server, and main Biofinity site. We
will then delve into the specific architectures of the client, server, and database sides of
the wiki in Chapters 4.2 through 4.4. Please note that the wiki features and
implementation details discussed in this chapter may have changed after the time it is
written.
As previously mentioned, the wiki exists as a separate project from the Biofinity
core, which is bundled in Biofinity.war in Figure 4.1. The Biofinity core and wiki store
data in separate databases and rarely store or retrieve data from their counterpart, with
few exceptions (e.g. querying user permissions). Further, the wiki front end and back end
BiofinityWikiServer.war, respectively. While the user interacts with the wiki front end
through a frame in the Biofinity front end, the wiki server performs the bulk of the wiki-
related processing.
The wiki’s implementation has a few aspects and features tied to the Biofinity
Biofinity data
First, the wiki’s appearance must match the “look and feel” of the main site. Thus,
the wiki front end is designed to make use of Biofinity’s stylesheets and three-column
page layout. Since one of the columns is reserved for navigation, the wiki page content –
panels and text – is placed in the two remaining columns. Figure 4.2 illustrates this layout
on a deployed wiki. Throughout the remainder of this chapter, we refer to the left sidebar
as the navigation sidebar, the center area as the primary content area, and the left sidebar
Second, the wiki must integrate with Biofinity’s authentication system and search
bar. Biofinity uses the Google OpenID authentication service5, and users are redirected to
a Google login page when attempting to log in to Biofinity. Since the Biofinity and wiki
components have their own execution contexts and, consequently, their own session data
instances for the same consumer, the session data pertaining to the current user needs to
be synchronized across the two domains when the login occurs. That is, the Biofinity core
notifies the Intelligent Wiki back end of the logged-in user via web service upon
successful authentication. The Biofinity core also provides a web service for retrieving
the current user and his/her current “lab”, which the wiki leverages to renew its session
data after it expires. The intelligent wiki also integrates with the search bar provided with
the Biofinity core. While search bar events are handled by the core, wiki-related search
queries are forwarded to the wiki server to process and return results. Details of the
The final integrated feature is the creation of data pages within the wiki. While
viewing data in the Biofinity core, the user can request a wiki page to be created for it.
The corresponding data page is a wiki page with a panel automatically populated with
the corresponding information from the data in the Biofinity DB, and users can then
expound upon the content with the wiki tools provided. Chapter 4.2.1.4 describes the
format and features available for the data page in greater detail.
subset of Java via the Google Web Toolkit, and although the GWT automatically
5
http://code.google.com/apis/accounts/docs/OpenID.html
62
generates them from the Java source, we manually write some HTML and Javascript to
supplement the generated code, e.g. for using the TinyMCE6 rich text editor for edit
mode. However, the amount of HTML and Javascript is relatively minimal – the front
end is thus composed primarily of Java classes for each of the UI elements and their
underlying representations. It interacts with the wiki back end by passing and receiving
HTML messages and XML data through the back end’s public-facing RESTful web
services.
The front end is designed around the ideas of 1) providing distinct presentations
or pages of the content to the user based upon the type of information involved, i.e.
providing an intelligent user interface, and 2) reusing UI elements and features across
terms, pages exist as top level items with subsets of panels as child elements.
Additionally, each page and panel has a corresponding data object that mirrors the
representation used in the wiki back end. When the front end requests and receives XML
data from the server, it immediately parses it into a data object whose values are used to
populate the page or panel. Similarly, changes to the data object due to user interaction
are sent to the server as XML data where it is parsed back into a data object.
To summarize, the Java classes in the wiki front end are categorized into three
main groups, and we delve into their details in the upcoming sub-chapters:
6
http://www.tinymce.com/
63
individual classes
As mentioned before, user tracking, i.e. logging user behavior and activities
during sessions, is the primary source of data for our analyses in Chapter 5. Since every
action and feature in the client requires requesting information from the server, we are
easily able to determine what actions the user is performing, which wiki objects are
involved, and when the action is performed. When applicable, the upcoming sub-chapters
will also detail the tracking performed for each feature, as well as how it applies to our
4.2.1 Pages
Information in the Biofinity Intelligent Wiki is primarily presented via pages that
are equivalent to web pages on a website. While most of these pages exist to display
particular information to the user, pages with user editable content each have two distinct
layouts, one for viewing the information and one for editing the information. The content,
panels, and functionality to be enabled in the view mode and the edit mode are determined
by the type of the page requested. As of the writing of this thesis, the following distinct
Information Pages
o User Page
o Consent Page
o Main Page
o Results Page
While they may differ in terms of content and functionality available, they all
have common events and members that every page should manage. These include
keeping track of panels loaded for the page, requesting the root container to resize page
contents to the Biofinity frame, and providing a means to determine whether a pop-up
dialog box is currently open. We thus encapsulate these common elements into the
WebPage interface and AbstractPage superclass from which all pages inherit from.
Figure 4.6: Class diagram for WebPage interface and AbstractPage superclass
sequence when users request it. First, the user performs the page request, either by
65
clicking a wiki hyperlink or manually entering a URL. These URLs often contain
parameters specific to the page being loaded, including the page ID and particular
revision ID to view. Starting with this information, the wiki client then queries the server
for additional page-related information and processes the initial response with page-
specific event handlers. Depending on the page, additional server requests/responses may
be necessary before the initialization continues. Finally when sufficient information has
been obtained, the page creates its UI elements, and instantiated panels may make their
own server requests and perform their own processing before loading is finished. In a
sense, this enables the asynchronous loading of the page’s components, since the
individual parts may finish loading before others, depending on which receive responses
In terms of tracking, the system generally tracks views and edits made by wiki
users to each of the pages. Broadly speaking, these impact the page’s relevance to the
user by determining topics and keywords of interest (Chapter 3.1.1) and the quality of a
page through number of views/edits (Chapters 3.1.5 and 3.1.6) and through the user’s
expertise (3.1.3). The number of views and edits also impact the modeling and
Wiki pages, as seen in Figure 4.4 above, are the page type most similar to the
typical article on Wikipedia and other wikis on the Internet, with the primary content area
consisting of formatted text and images. The other intelligent wiki features available,
such as basic page control, revision control, page rating, page sharing, tags, attachment
upload/download, and a discussion area, are encapsulated in the panels attached to the
page in both the primary and secondary content areas. That is, users can:
publish the page for viewing by users outside the user’s lab group
set the currently-viewed revision as the “main” one first seen on page load
share the page with other wiki users through the intra-wiki recommendation
system
share the page through other social platforms such as Twitter and Facebook
and it displays suggested related Wikipedia articles based upon the keywords extracted
from the page’s text. Further explanation for each of these functionality and their
Walking through the initialization process, the wiki page is first instantiated with
a particular page ID, which is embedded as a URL parameter to the wiki client. It then
queries the server for the basic page info associated with the ID, i.e. its deletion/lock
status and its main revision (if one isn’t specified in the URL). If the page is flagged as
deleted or the user lacks permissions to view the page, then a corresponding message is
displayed and processing stops. Otherwise, it requests the wiki page content for the
particular page ID and revision. When that information is successfully returned, the wiki
page instantiates its panels with the appropriate known page/revision information.
69
Viewing a wiki page creates a tracking entry for the user, which includes the page
and revision viewed and the timestamp for the action. View actions contribute directly
towards determining the page’s quality (Chapter 3.1.5) and indirectly towards
The edit mode of wiki pages, as seen in Figure 4.6, has considerably fewer
comments panel, and Wikipedia recommendations panel. While the title field is a regular
text box, the content field is a TinyMCE WYSIWYG rich text editing text area, and its
contents are transmitted to the server as plain text HTML upon revision submission. After
a revision is successfully submitted, users are forwarded back to the view mode of the
The attachments panel is the same as the one used in the view mode, and users
may upload additional files to the page. Any additions to the attachments list here are
reflected upon the panel in the view mode. While it may seem unusual for the panel to be
placed in the edit mode, this enables users to embed images within the content body via
URL reference. Further explanation of how the attachments panel functions can be found
in Chapter 4.2.2.2.
Similarly, the comments panel is the same as the one used in the view mode, and
the comments and discussions carried out are also displayed in the comments panel.
Additional comments can be made while in the edit mode, and changes made will be
reflected in the view mode’s comments panel. Although the comments do not
automatically refresh as new ones are made, his enables users to refer to the comments or
carry out further discussion while editing the page by manually clicking the panel’s
“Refresh” link. Further explanation of how the comments panel functions can be found in
Chapter 4.2.2.3.
clicking a link opens the corresponding article in a separate window, which provides a
means for the user to quickly refer to the article to supplement their knowledge of related
topics and prevents the user from losing his/her current work due to navigation. Further
Chapter 4.2.2.
71
Editing a wiki page creates a tracking entry for the user, which includes the page
edited, the particular revision’s ID, and the timestamp for the action. Edit actions
contribute directly towards determining the page’s quality (Chapter 3.1.6) and the user’s
expertise (Chapter 3.1.3), and contribute indirectly towards determining pages relevant to
location, or classification data in the Biofinity database. In terms of page layout and
available features, data pages are nearly identical to wiki pages – the only difference is
the addition of a data panel that displays the corresponding Biofinity data for which the
page was generated. Due to this similarity, data pages are represented in the underlying
72
component of the page is encapsulated in the former whereas the Biofinity data-specific
information, such as the data type and data entity ID are contained in the latter. This
Since they contain different fields, each of the four page types has a different data
panel, although the fields and values for all four are displayed in “<heading> <value>”
format. For more information on the data panels, refer to Chapter 4.2.2.10. Note that the
Biofinit data cannot be edited directly through the wiki interface – the user must navigate
Initialization of data pages also occurs in a manner similar to wiki pages. After
the ViewDataPage class is instantiated with a page ID by the client, the client requests
basic page information from the server and parses the server response for the main
revision number and page type. The process differs slightly at this point – in addition to
requesting the corresponding wiki page information, it also requests data page-specific
information, such as the associated data entity ID and type. Based on these, one of the
four data panel classes (Chapter 4.2.2.10) is instantiated to display the Biofinity data. The
74
remainder of the process, i.e. creating the UI and initializing panels, is the same as for
wiki pages.
Viewing a data page creates a tracking entry for the user, which includes the page
and revision viewed and the timestamp for the action. View actions contribute directly
towards determining the page’s quality (Chapter 3.1.5) and indirectly towards
revision form, the same panels displayed, and the same event handling. That is, no unique
edit functionality or information is introduced to the edit mode for data pages, and
consequently, the edit mode for data pages reuses the EditWikiPage class in its entirety.
Editing a data page creates an entry similar to ones for wiki pages, which includes
the page edited, the particular revision’s ID, and the timestamp for the action. Edit
actions contribute directly towards determining the page’s quality (Chapter 3.1.6) and the
user’s expertise (Chapter 3.1.3), and contribute indirectly towards determining pages
Publication pages, an example of which is shown in Figure 4.9, are laid out in a
manner similar to wiki and data pages, with their main body and comments panel in the
primary content area and with panels for wiki features in the secondary content area. The
main body consists of publication information, such as the publication authors, year of
publication, and venue, as a well as an abstract for the publication. The publication itself
can be uploaded to the page as an attachment or added as a URL in the abstract body.
As for panels attached to the page, publication pages generally have the same
panels as its wiki page and data page counterparts: basic controls, revision control, page
ratings, page sharing, tags, and attachments. Note that it does not have a Wikipedia
recommendation panel.
76
Tracking for viewing a publication page is similar to tracking for the previous
types. An entry for it includes the page and revision viewed and the timestamp for the
action. View actions contribute directly towards determining the page’s quality (Chapter
3.1.5) and indirectly towards determining pages relevant to the user (Chapter 3.1.1).
77
As can be seen in Figure 4.11, publication pages have their body contents split
into multiple, specific fields rather than as one generic content field as with wiki and data
pages. This is also reflected in the underlying model for the page in that each of these is
its own attribute in the PublicationPage data class. Title, Authors, Year, and Venue all
use a plain text box in the revision form. However, the Abstract field uses the TinyMCE
rich text editor. Upon submission, the form’s contents are sent to the server in XML
format.
Editing a publication page creates a tracking entry for the user, which includes the
page edited, the particular revision’s ID, and the timestamp for the action. Edit actions
contribute directly towards determining the page’s quality (Chapter 3.1.6) and the user’s
expertise (Chapter 3.1.3), and contribute indirectly towards determining pages relevant to
User pages, as pictured in Figure 4.12 when directly accessed outside of the
Biofinity UI, are automatically generated whenever a new user is registered with the wiki,
displaying the associated user’s first name, last name, and e-mail address. This
information is obtained from the Biofinity DB, which originally obtains the information
from the associated user’s Google OpenID profile. It should be noted that user pages
have not been updated as frequently as the other page types and are disabled for the
When the user page is first loaded, it requests the user information from the server.
After receiving a successful response, the GetUserHandler then triggers the process of
parsing the wiki user information from XML to an instance of the Wikiuser data class. It
should be noted that while user pages do have this underlying data class, their contents
cannot be directly edited within the wiki. Rather, the user associated with the page will
need to navigate to Account > Manage Account through the Biofinity interface to change
it.
User pages currently have user information, attachments, comments, and user
ratings panels associated with them. The user information panel is intuitive to include
since it displays the main content of the page, as is the user ratings panel since it allows
others to evaluate the user. As for the remaining two panels, the attachments panel is
included for users to upload personal files (e.g. resume/CV), and the comments panel
The Access Error Page is a fixed content page that is displayed when users
attempt to access a private page that they do not have permissions for. This may also
occur when the user is currently logged into the incorrect Biofinity lab. The page does not
contain any panels and does not have any functionality associated with it. Consequently,
The Consent page, as seen in Figure 4.16, displays the full text of the IRB
Informed Consent Form, and it is used to obtain explicit permission from intelligent wiki
users to include their tracking data, evaluations of the system, and/or their scores for the
course (when applicable) for this thesis and other future studies. When it is first loaded,
it checks to see whether the user has already filled out the consent form. If the user has
not, then the form shown in Figure 4.17 is populated at the bottom of the page. Otherwise,
While the consent page does not contain any panels or an associated data class, its
These handle the server response when querying for the user’s consent status and
Since user consent does not (and should not) impact a user’s experience with the
The main page (shown in Figure 4.19) is the first page seen by the user when the
“Intelligent Wiki” link in the navigation sidebar is clicked, and its primary content
consists of a hard-coded welcome message to the user along with brief instructions on
how to use it. Since it consists of non-editable content, it does not have any underlying
data classes and has very few features for the user to make use of. Consequently, few
panels are attached to the page: only panels for creating pages, displaying pages recently
edited by other group members, and obtaining user consent are attached to the main page.
The main page itself does not have any tracking associated with it.
84
Similar to the Access Error Page, the Not Logged In page simply displays an error
and has no panels, data classes, or functionality associated with it. It occurs when a user
attempts to create a new wiki or publication page when not logged into Biofinity, and no
The search results page (seen in Figure 4.23) is displayed after the user performs a
keyword search via the search bar in the header while in the wiki section of the Biofinity
system. The results are displayed in a table along with details such as the page’s title, its
date of last revision, and its URL. The results are originally ordered by their term
86
indexing and search engine, and the user can click the column headers to reorder the
results. Since the SmartGWT ListGrid used to display the results can use XML data
directly, the ResultsPage class does not parse the results into an equivalent Java object
and is thus not associated with a data object. The results page also does not have any
While a tracking entry isn’t kept for viewing the search results page, the system
does track the keywords used and the results returned for the search. These entries can
contribute towards determining a page’s topics, tags, discipline, and keywords, which
4.2.2 Panels
As previously described, we encapsulate the various “recyclable” UI elements and
interactive features and functionality into panels. The panels included on a page are
7
Spärck Jones, Karen (1972). "A statistical interpretation of term specificity and its application in retrieval".
Journal of Documentation 28 (1): 11–21.
87
initialized and added to the proper HTML content element during the execution of the
page’s “createUI” methods, i.e. after the client receives primary page-specific content
from the server. Each panel then queries the wiki server separately and finishes loading
once it receives any needed data, i.e. the panels can be loaded and displayed
asynchronously.
Similar to how pages have an interface and abstract superclass, panels have their
equivalents in the Panel interface and the AbstractPanel superclass implementing it.
The Common Controls panel in Figure 4.25 provides users with basic page
management tools, including the options to edit the page, delete the currently viewed
revision, lock the page from further editing, and publish the page to the public. Clicking
the Edit button forwards the user to the page type’s corresponding edit mode and pre-
populates its fields with that of the currently revision. Clicking the Delete button marks
the current revision as deleted and cannot be undone via the wiki interface. Additionally,
the entire page is flagged as deleted if all of its revisions have been deleted. Unlike
deletion, page locking and publishing can be undone, and the corresponding buttons will
Each of the features provided by this panel are tracked by the system, and the
created tracking entry marks the action performed (one of edit, delete, lock, unlock,
publish, or unpublish), the user performing the action, when it was performed, and the
page and revision acted upon. The edit and delete actions in particular may directly
impact the page’s quality (Chapter 3.1.6) and the user’s expertise (Chapter 3.1.3), and
contribute indirectly towards determining pages relevant to the user (Chapter 3.1.1).
This panel only appears on the view modes for pages with editable contents, i.e.
The attachments panel, seen in Figure 4.26, is used to upload and associate files to
pages that the panel appears on, and to display download links to the files that have
already been attached to the page. For image attachments in particular, hovering the
mouse cursor over the hyperlink displays a thumbnail preview of the image, as seen in
When the panel is instantiated, it requests the server for attachments currently
attached to its parent page. The server then responds with the files’ names, download
URLs, and, when appropriate, thumbnail images. After receiving this response, the panel
populates its attachments list. Note that the binary data for the (original) files are not
To upload a file, the user clicks the “Add” link in the panel header. The following
pop-up appears for the user to select the target file and enter the caption text to display
with it.
To download a file, the user clicks the corresponding download hyperlink and
The system creates a tracking entry for uploads and downloads, which includes
the ID of the attachment uploaded and the page that the attachment is associated with.
Entries created from these actions are currently not used in recommendations.
This panel can be found on both the view and edit modes of pages with editable
content, i.e. wiki, data, and publication pages, as well as on user pages.
Figure 4.32: Comments panel UI, one root-level comment with two children comments
Unlike most of the other panels, the comments panel (as seen in Figure 4.29) is
located in the primary content area due to its larger size, and it enables users to carry out
comments at the root level being considered as the “beginning” of the threads. Direct
replies to existing comments add “children” comments to them. Each comment may have
any number of children comments, but only one parent and one root. This hierarchical
nature of the comments is reflected via indentation – root-level comments are leftmost,
91
and subordinate comments are indented at one level further than their immediate parent.
Particular paths through the comments tree are thus particular “threads” of conversation.
The following features are provided for the comments panel: creating a new
collapsing of threads. New comment threads, i.e. new top-level comments, can be made
by clicking the “Create New” link and filling in the dialog box (Figure 4.30) with the
topic and comment body for the thread. Clicking the “Reply” link for a particular
comment pops up a similar dialog box with the topic field pre-populated. Comment trees
and sub-trees can be selectively collapsed and re-expanded by clicking the triangle icon
the entire page. However, it should be noted that it will not refresh the panel for other
users in real-time. Instead, users may click the “Refresh” link to manually update its
contents. We opted for this approach since an automatic refresh requires periodic requests
to the server or a persistent listener for comment-related server responses. Both require
92
additional computational resources multiplied with more active wiki users, and the latter
Tracking entries are made when a user comments upon the page or participates in
existing discussions. While these currently do not impact our recommendations generated,
The comments panel can be found on both the view and edit modes of wiki, data,
The page ratings panel in Figure 4.31 appears on wiki, data, and publication pages,
and it enables users to evaluate pages on a 1-5 star scale. The user can provide a rating by
clicking the star rating to give it on the widget. This rating persists indefinitely and is
loaded each time the user views the page. Should the user choose to re-evaluate the page,
the user can give a different rating to overwrite the old one. That is, a user can only
contribute towards page’s rating once per page. By implementing the rating system in
this manner, it prevents the practice of “spamming” ratings to guide the page’s overall
rating towards a particular score. While influencing page score is still possible through
the use of additional accounts, the work involved in setting them up may discourage
The panel also displays the page’s average rating and the number of users
contributing towards that score. By publicly displaying both, users can determine the
Tracking is performed when a user rates a page, and a user’s rating can play a
large impact on the recommendations made. First, they influence the page’s relevance to
the user since they indirectly indicate the user’s inclination to the topics relevant to the
page (Chapter 3.1.1). The ratings also impact the overall perceived quality of the page
(Chapter 3.1.2).
for wiki, data, publication, and user pages. Upon initialization, the panel generates a
request to the server, which then performs an online calculation of the requested
Number of views
Number of edits
Number of comments
Number of ratings
Rating score
It has since been removed after the wiki’s prototyping stages and is disabled for
classroom deployments. Thus, no tracking is performed for this panel since these stats
cannot be viewed.
The revisions panel (Figure 4.32) appears on the view mode of pages with
editable content, such as wiki, data, and publication pages. Through this, users can view
specific non-deleted revisions of a page as well as set another revision as the main one,
i.e. the one first loaded when a specific revision for a page isn’t specified. Further, this
panel can be used to navigate to a specific revision for deletion. Selecting a revision in
the dropdown box and clicking the “View Revision” button will reload the page with the
The system creates tracking entries when the user sets the main revision for the
page and when viewing a different revision. The revision changing aspect has no impact
The Sharing Panel as seen in Figure 4.33 is used to share or send a hyperlink: 1)
to another wiki user via the intra-wiki recommendation framework, 2) to the user’s
Twitter followers via a Twitter post to the user’s account, or 3) to the user’s Facebook
friends via a post to his/her wall. Note that people following the shared link may not be
able to access the recommended page if they do not have sufficient permission to do so.
To share a URL with another wiki user, the user first clicks on the MyLab icon
(the flasks) and then enters the URL to share and the group peers to share them with in
are displayed with first name, last name, and e-mail address to help distinguish them from
one another, i.e. when two users have the same first and last names. For convenience, a
“Here” button has been included to enable the user to quickly obtain the URL of the
currently-viewed page.
96
Figure 4.35) is displayed to them in real time, provided that the receiving user is not
currently “busy” with any work. We approximate this with the following set of rules to
If the user is currently editing a page (i.e. in edit mode), do not display the alert
box
If dialog boxes are currently open (e.g. when sharing pages with other users, when
uploading an attachment, when making a comment), do not display the alert box
If the alert box has been displayed within the last ten minutes and is currently
If the above conditions are avoided and the user has recommendations that are not
algorithms in Chapter 3.1 are displayed to users via this same alert interface. In
following/dismissing aspects of this panel. While entries created from these actions
currently have no impact in the recommendations made or the modeling performed, they
The sharing panel, and consequently the sharing dialog box, is only available on
the view modes of pages with editable content, including wiki pages, data pages, and
publication pages. However, the alert box can be displayed on any page if the display
The tags panel (Figure 4.36) is used to associate a page with user-defined key
words/phrases, and these may include words relevant to the page topics and areas of
study. Each of these can be clicked to perform a search for other pages containing or
tagged with these words, enabling users to find related content. To edit these, the user
clicks the “Edit” link in the header. The list of tags then turns into a comma-separated list
Upon submission, the list is sent to the server, which determines which tags have
This panel is available on the view modes of wiki, data, and publication pages.
to Wikipedia articles that may be related to the currently-viewed page. It provides users
with easy access to additional related information, and clicking a link will open it in a
new window. This feature provides two primary benefits to users: 1) improving
99
and related topics, and 2) aid in the contribution of content to the Biofinity intelligent
wiki.
found in the page text after processing with LingPipe. For further information on how the
Clicks to follow a Wikipedia recommendation are tracked by the system, and the
tracking entry includes the keyword clicked and the page currently viewed before the
click. While the use of this hyperlink to access Wikipedia is tracked, we cannot track
further action taken by the user on Wikipedia. This action currently impacts neither
that the pages are created from. Thus, they can only originate from the four supported
Biofinity data types: classification, event, location, and occurrence data. The data panels
for each of these have their own distinct fields to reflect the data type displayed. The
Classification
o Classification ID
o Name
Taxon Name
100
Taxon Rank
Event
o Event ID
o Location ID
o Event Date
o Habitat
o Sampling Effort
o Sampling Protocol
Location
o Location ID
o Longitude
o Latitude
o Verbatim Elevation
o Continent
o Country
o State/Province
o Island
o Island Group
Occurrence
o Occurrence ID
o Event ID
101
o Basis of Record
o Sex
o Life State
o Behavior
o Reproductive Condition
o Preparations
Upon initialization, they are provided with the Biofinity Entity ID of the data,
which is used to request information to populate the fields via a Biofinity web service.
Again, the information displayed in this panel cannot be edited directly through the wiki
The viewing of data panels is not tracked, although the viewing of the associated
page is. Refer back to Chapter 4.2.1 for the impact of these tracking entries on our work.
The create panel (seen in Figure 4.38) is shown only on the main page and is used
to display links for creating new wiki and publication pages. It is not shown when the
user is not currently logged into the system. After clicking a link, the user is forwarded to
the edit mode of the respective page type. Since there is no existing information for
newly created pages, the fields in the revision forms are not pre-populated. Further, only
Creating a page of either type causes a tracking entry to be created with the ID of
The recent pages panels display the five most recently edited wiki, publication,
and data pages seen by peers in the user’s current group. These panels only appear on the
main page, beneath the welcome text in the primary content area. While the panel
pictured is specifically for recent wiki pages, all other recent pages panels display their
The recent pages panels themselves do not require any tracking entries to be made
for them. However, the viewing of the pages listed is tracked. Refer back to Chapter
4.2.1.x for additional information on the entries created and their impact.
103
The consent panel (Figure 4.40) is only used on the Main Page and is placed in
the sidebar area upon page load. When initialized, it obtains the current user’s consent
status from the wiki server. If the user has not yet filled one out, it displays the above
message and provides a link to the Consent Form Page. Otherwise, it displays a thank-
you message. Like with the consent page, no particular tracking is performed for the
consent panel.
current user. After querying for and receiving this information from the server, each peer
is displayed in “<first name> <last name>” format in a comma-separated list, and each
name is a hyperlink to that user’s corresponding user page. This panel was originally
The user info panel (Figure 4.41) is used in user pages to display the first name,
last name, and e-mail address of a particular user. Since this panel only appears on user
pages, the information displayed is that of the page’s corresponding user. As previously
mentioned, its contents cannot be directly modified through the wiki – instead, users must
edit it through Biofinity via Accounts > Manage Accounts. There are no tracking entries
The user rating panel (Figure 4.42) provides users with the opportunity to
evaluate other users on a binary scale, i.e. “like” and “dislike.” Like with page ratings,
each user can only provide one rating at most for each other user, and this rating can be
changed at any time. The overall “score” for the user is calculated by subtracting the
Tracking entries are made when a user provides a rating for another user, and this
simple statistics on a particular user’s behavior on the wiki. These were to be included on
each user page, but were removed due to accuracy concerns and concerns that displaying
These fell into three categories, including stats on the actions taken, on the user’s
recommendation activities, and on the user’s session information. These are calculated
online as a user page is loaded. In further detail, the statistics displayed include:
Action Stats
o Number of edits
o Number of uploads/downloads
o Number of comments
o Numerical rank for each of the above, relative to all other wiki users
Recommendation Stats
o Numerical rank for each of the above, relative to all other wiki users
Session Stats
o Number of logins
106
o Numerical rank for each of the above, relative to all other wiki users
Figure 4.46: The TinyMCE WYSIWYG text editor, as seen in a wiki page’s edit mode
While the third-party developed TinyMCE rich text editor (Figure 4.43) is
technically not a panel (i.e. does not extend AbstractPanel) and does not behave similarly
to one (i.e. does not request from or post information to the wiki server), it is worth
editing interface similar to that of Microsoft Word, including features such as:
Super-/sub-script
Hyperlinking
Image insertion
A key feature of the editor is that it represents its contents as plain-text HTML.
When revisions are submitted to the server, the revision contents are transmitted (with
characters converted to hex equivalents when necessary) and stored in the database in this
form.
While edits are made through this editor, tracking entries are not made for its use.
core although both exist on the same server. The wiki back end is the only component
with direct access to the wiki database, and connections to it are distributed from a
As seen in Figure4.44, there is a table for each of the objects and pages used in
the wiki. While their purposes are self-explanatory from their names, their fields and
109
relationships to the other tables may not be as straightforward. The subsections of this
data in Biofinity. Currently, the types of data generating a data page include event,
occurrence, location, and classification data. While users can create and edit wiki content
on a data page, this table only stores the link to Biofinity data. The wiki content is instead
stored in the “wikipage” table. There is a 0..1-to-1 relationship between the entries in this
table and the “page” table in that each data page has a corresponding entry in “page” but
not vice-versa. Similarly, there is a 0..1-to-1 mapping between data page entries and wiki
others are editing a single page concurrently. That is, if an entry exists for the page that
the user wishes to edit, then the user is warned of the concurrent editors before entering
the page’s edit mode. There is a 0..1-to-1 relationship between its entries and the “page”
110
table in that each entry is tied to a corresponding page ID, but not vice-versa. Its fields
keyword tags (table “wikitag”), and there is a many-to-many relationship between them.
relationship.
The table has since been revised to “Join Wiki Page to Automated Keyword” in
page revisions, and the keywords are joined to specific wiki page revisions via the
“join_wikipage_lpkeyword” table.
The table has since been revised to “Automated Keywords” in later iterations of
the wiki.
wiki, and it provides a unique ID for the page to be referenced by regardless of type. All
wiki objects in the wiki DB (except the join_wikipage_lpkeyword table) use this
page, such as the default revision to display on page load and its locked/deleted status.
112
Note that the “IsLocked” field in this table is set when a user clicks the “Lock” button in
the common page controls panel, and is not the lock induced by the “editmarker” table
previously described. Pages that are locked cannot be edited until the lock is lifted, and
deleted pages cannot be viewed or edited. The entries in this table have a 1-to-1
pages describing and organizing publications. They are different from the other page
types in that its contents are split into distinct fields rather than being contained in a
single generic field. As such, its structure is similar to that of the “wikipage” table in
Chapter 4.2.14. There is a 0..1-to-1 relationship between the entries in this table and the
“page” table.
such as the terms used and the results returned. For simplicity, the search results are left
in the XML form generated by the wiki server to be returned to the wiki client. Each
entry in the table is associated with one user and one page whereas users and pages may
attachments. Its entries are automatically generated upon image upload and have a 1-to-1
While the user rating feature was not enabled in the wiki deployment for gathering data,
it exists to enable future work in recommending users to collaborate with. The entries in
this table have a many-to-1 relationship with the users in the system – users may be
associated with making or receiving multiple ratings of other users, but each rating entry
uploaded to the wiki. To preserve space, separate revisions of a same file are not
currently supported, and uploads with the same file name on the same page will overwrite
115
the existing one. The entries in this table have a many-to-1 relationship with pages and
users.
comments/discussions occurring in the wiki. The inclusion of the “RootId” and “ParentId”
relationship between the entries in this table and pages/users. Pages and users can be
associated with more than one comment, but each comment is only associated with one
their revisions. In spite of its name, the table also stores the editable wiki information
from data pages (i.e. each data page contains a wiki page). Its fields are largely similar to
the fields of the “publicationpage” table, but it has one large generic content field instead
of multiple smaller specialized fields. The entries of this table have a 0..1-to-1
relationship with the entries in the “page” table and a 1-to-0..1 mapping with the entries
in “datapage.”
with ratings being on a scale from 1 to 5. It should be noted that using the IDs of both the
rater and the page as the primary key, users cannot “spam” ratings to heavily influence
the page’s average, assuming that a reasonable number of other users have already voted.
Instead, the user’s newest rating will overwrite the old one given. There is a many-to-1
tags are joined to specific wiki page revisions via the “join_page_tag” table. Its entries
have a many-to-many relationship with pages since each tag can be applied to multiple
including the action performed, when it was performed, the page it occurred on, and the
object acted upon. In other words, this table stores tracking information of user
behavior and contains the bulk of the data used in our analysis. There is a many-to-1
relationship between the tracking entries and wiki users, the page involved, and the
118
involved object. That is, particular users, pages, and objects may have multiple tracking
entries associated with them, but a particular tracking entry will only be associated with
one of each.
the system, whether they be from user to user via the intra-wiki “Share” button or from
the system to users. Since the recommendation entries are stored via URL as opposed to
page IDs, recommendations to content outside the wiki can also be stored. There is a
many-to-one relationship between users and URL recommendations – each user can
make and receive multiple recommendations, but each wiki URL recommendation is only
generated user page. References to users in the other tables (e.g. as AuthorId, UserId,
RateeId, etc.) use this wiki-specific ID and not the user’s corresponding Biofinity ID.
the system can consistently detect when the user logs in, the logout time may be less
straightforward to determine if the user doesn’t log out manually. For example, when
users forget to log out, the session entry could be left “open”, i.e. with no logout time, up
120
through the user’s next login. One approach taken to avoid this is to catch browser close
events via Javascript and obtaining the timestamp when this occurs. In the event that the
Javascript solution fails, the server sets the logout time to the last update time if 15
There is a many-to-one relationship between the entries in this table and wikiusers.
Chapter 5: Results
We have deployed the Biofinity Intelligent Wiki for use in the collaborative
Intelligence Applications (RAIK 390) during the Spring 2011 semester and Multiagent
Systems (MAS 475/875) during the Fall 2011 semester. By gathering and analyzing the
usage data of the students in these courses, we aim to: 1) gain a better understanding of
the relative importance of each algorithm component and 2) identify trends in student
Section 5.1 first describes the logistics of the classes and their collaborative
assignment within the wiki. From there, it summarizes the primary activities performed
that we would like to track along with the rationale for choosing them. These attributes
are the basis upon which we perform further analysis, including active vs. passive and
minimalist vs. overachiever metrics for their activity profiles. Section 5.2 then delves
into the results themselves, presenting the processed data within various contexts and
5.1 Logistics
This section describes the classes to which the Biofinity Intelligent Wiki has been
deployed, as well as the collaborative writing assignments that make use of it. It also
covers and justifies our specific points of observation and evaluation, including: the
specific student activities to track and corresponding metrics of interest, our derived
metrics of active vs. passive and minimalist vs. overachiever for their activity profiles,
recognize the need for and apply artificial intelligence to business applications. It has a
class size of 16 students, and 15 students provided consent for their usage data to be used
in this thesis.
theories and applications of multi-agent systems. It primarily consists of junior and senior
undergraduate students and graduate students, all of whom major in computer science or
spite of the differing course topics. Initially, each student individually writes a summary
A list of praises: descriptions of what the student believes are the important/useful
A list of critiques: descriptions of what the student believes are the weaknesses of
topic
After this “individual contribution” phase is completed, the students then move on
to a 4-/5-week long “collaboration phase” where they contribute towards, rate, and tag
other students’ summaries and participate in threaded discussions. However, they are not
Halfway through the collaboration phase, the instructor provides initial feedback
on the students’ collaborative efforts. Students’ collaborative activities are graded in the
following manner:
60% Wiki Editing (at least three other essays, and amount and quality)
revisions made in the wiki and extract the number of revisions and word length of
wiki collaboration that we can observe, since the discussions can give rise to new
ideas and additional wiki content. Specifically, we will track all comments made
in the wiki and count the number of comments made and their lengths (in number
Keyword tagging – Marking wiki pages with words that can be used to mark or
summarize their contents. Although we track the user performing the tagging, the
keywords used, and the pages tagged with the keywords, we only consider the
ideally after reading and evaluating its contents. Although we track the user
making each rating, the page being rated, and the rating given, we only make use
of the number of occurrences of the rating action by each user rather than the
Based on the numbers of edits, comments, keywords tagged, and ratings provided
as well as the average edit and comment lengths, we categorize the students in each class
with this “raw” attribute data. Doing so may highlight the particular attributes that are
most valuable in grouping them as well as provide some information on student behavior
Active vs. Passive – Categorizing students based on the number and type of
actions that require low cognitive cost, relative to the cognitive cost of
o Active – Greater focus on many edits and comments, more pro-active, i.e.
comment count)
Finally, we examine the usage data for patterns in students’ page sets for
their activities within the wiki. Since the data is largely numeric, we can leverage a
clustering algorithm to classify them based on the quantitative aspects of their profiles. A
few considerations must be made when selecting one appropriate for our data:
126
The number of instances for each data set is relatively few, e.g. less than 20
appropriate for us due to its ability to: 1) determine the optimal number of clusters (from
a range between a user-specified minimum and maximum) needed for the best clustering
results and 2) its ability to operate on data lacking labels and known representative
algorithm provided by the WEKA machine learning software suite to cluster the students
Parameter Value
binValue 1
cutOffFactor 0.5
debugLevel 0
debugVectorsFile weka-3-6
maxIterations 1
maxKMeans 1000
maxKMeansForChildren 1000
minNumClusters (MinK) 1
127
useKDTree false
Unfortunately, the results from the X-means clustering were unusual and
unintuitive, either due to the characteristics and attributes of our data sets or due to
The optimal solutions consistently grouped the instances into MinK (or fewer)
clusters, where MinK is the X-means parameter specifying the minimum number
of clusters expected
Cluster assignments for each instance are not consistent across many (30)
obtained from the X-means classifier. It can be summarized in the following steps:
the MinK value specified by us, it is difficult to determine whether the resulting number
of clusters is truly optimal. Fortunately, the X-means results in the WEKA package do
return two measures of the clustering effectiveness along with the clustering
A good clustering solution ideally minimizes Distortion while maximizing BIC. However,
we’ve empirically determined that both Distortion and BIC values decrease for increasing
We then defined the following weighted sum to find the optimal number of
( ) ( )
The value of k to be targeted (and the value of MinK to be used with X-means for
5.2.1.2 Put the instances into k clusters, based on the co-occurrence data
Since the clusters formed by X-means are also dependent on the particular seed
scheme that builds from the different clustering results across many seeds: iteratively
build/merge clusters based on pairs of instances that co-occur most frequently, then
second-most frequently, etc. This is based on the intuition that instances appearing
together frequently in the X-means results, i.e. co-occur in the same cluster across many
different seeds, are more likely to “truly belong” to the same cluster. Similarly, instances
that rarely co-occur in the same clusters are less likely to “belong” together.
After k is found via the process in Section 5.2.1.1, the goal is to iteratively merge
the instances into k clusters. Ideally, the algorithm ends when exactly k clusters remain,
and this value is specifically targeted to correspond with the X-means results obtained for
MinK = k.
As previously mentioned, the central concept upon which instances and clusters
are joined is the idea of grouping frequently co-occurring instances together. We define a
maximal pair for an instance to be the set of instances with which it co-occurs the most,
excluding itself and instances within its current cluster. We also define an instance to be a
strong maximal pair of another if they are mutually maximal pairs of one another. That
is, when instances are strong maximal pairs of one another, there are no other instances
with which either co-occur more often. And thus those instances shall be joined into the
same cluster.
Figure 5.1 summarizes the flow and transformation of information from the raw
Figure 5.48: Information transformation from raw usage data to final clusters
5.2.2 Definitions
This section more formally defines the concepts of maximal pairs and strong
maximal pairs and the design decisions pertaining to their use in our maximum pairs-
based clustering algorithm. Examples are also provided to aid in understanding of the
concepts.
k – the MinK value used for X-means. This value is empirically determined by:
1) Running X-means with MinK ranging from 1 to (|S| / 2 ) and recording the
( ) ( )
131
N – the number of seeds used for the X-means portion of the process, i.e. the number of
seeds used to generate co-occurrence data. While there are no specific guidelines for
C – the set of clusters currently unmerged during a particular iteration of the algorithm.
Co-occurrence( a, b ) – the number of times instance a and instance b appear in the same
cluster, when X-means is run with MinK = k across N seeds. This value has a range of 0
to N.
holds:
The above definition suffices for finding the maximal pair of singleton clusters.
Note that a given instance may have multiple maximal pairs. For example, two distinct
132
= Co-occurrence( a, c ).
We also define the maximal pairs for an entire cluster A as the most frequently co-
occurring maximal pairs across all of the cluster’s members. That is:
For example, consider the following co-occurrence table for cluster A = { a1, a2, a3 }
b c d e f
a1 5 10 2 7 8
a2 0 0 1 10 10
a3 8 8 2 3 1
Table 5.3: Co-occurrence table for cluster A = { a1, a2, a3 }
Each cell indicates the number of times (across the range of seeds used in running
X-means) for which the instances in the corresponding row and column have been
grouped in the same cluster by the X-means algorithm. As can be seen, instance a1 co-
occurs with instance b 5 times, with instance c 10 times, etc. The maximal pairs for a1,
a2, and a3 are then the instances with which they co-occurred the most, respectively:
133
For a1, the maximal pair is c since it has the largest number of co-occurrences
For a2, the maximal pairs are both e and f, since they tie for the largest number of
a3 also has two maximal pairs since b and c tie for the largest number of co-
Then to determine the maximal pairs for the entire cluster, we iterate through the
a1’s max pair of { c } is added to the set of max pairs for A since it is the first set
a2’s max pairs of { e, f } is added to the set of max pairs for A since their co-
a3’s max pairs of { b, c } are not added to the max pairs for A since their co-
occurrences with a3 do not exceed the co-occurrences of the ones added thus far
(8 < 10)
This approach of using the most frequently co-occurring maximal pairs as the
“representative” ones for the entire cluster was selected since it strikes a balance between
1. Including all maximal pairs for every cluster member as part of the cluster’s
2. Using only one member’s maximal pairs when multiple candidates exist.
While its effects are not noticeable for singleton clusters, it provides subtly different
results in the returned maximal pairs. Returning back to the co-occurrences in Table 5.1,
the maximal pair(s) for instances a1, a2, and a3 are { c }, { e, f }, and { b, c },
respectively. Under this option, the cluster’s maximal pairs would be the union of the 3
This is problematic in a couple ways. First, since clusters are merged with maximal pairs
as the basis, this enables the cluster to be merged with another on this “weaker” maximal
pair. Second, it enables larger clusters to have a larger number of instances in their
maximal pairs sets. This in turn increases their chances of merging with another cluster,
leading to a “snowball effect” where bigger clusters continue to grow while smaller
Option 2 is restrictive towards the opposite extreme and also affects singleton
clusters that have multiple maximal pairs. It can share the same weakness as (1) if the
member chosen does not contain one of the cluster’s “strongest” maximal pairs, although
this can be remedied by only selecting among members that have a “strongest” pair.
excluded itself and instances within its current cluster. It is apparent that the instance
itself should not be a maximal pair candidate, since it would always co-occur with itself
135
across all N seeds used for X-means. However, the reasoning for the cluster membership
a.cluster != b.cluster
that already belong to the same cluster. That is, since the maximal pairs concept is used to
merge distinct clusters, it would be counterintuitive to examine instances within the same
maximal pair on every iteration of the algorithm. To better elaborate on this point,
a1 a2 b1 b2 b3 c1
a1 x 10 1 0 3 2
a2 10 x 5 6 1 3
b1 1 5 x 5 10 4
b2 0 6 5 x 10 3
b3 3 1 10 10 x 8
c1 2 3 4 3 8 x
Table 5.4: Co-occurrences between all instances in clusters A = {a1, a2}, B = {b1, b2, b3}, and C = {c1}
maximal pairs for each cluster when ignoring the same-cluster condition:
Since cluster merges only occur when there are mutually maximal pairs between
them, no merges can occur in this situation. When the same-cluster condition is followed,
Each of the clusters now have maximal pairs “outside” of themselves: cluster A
has a max pair in cluster B, cluster B has a max pair in cluster C, and cluster C still has a
max pair in cluster B. Since clusters B and C mutually have maximal pairs in one another,
The following function definitions are used in the upcoming definition of strong
maximal pairs:
MaximalPairs( a ) – the procedure to obtain the set of maximal pairs for instance
a. If there are multiple, all of them are included in the returned set.
MaximalPairs( A ) – the procedure to obtain the set of maximal pairs for cluster
A. Similar to the procedure for a single instance, this one also returns multiple instances
a MaximalPairs( b )
b MaximalPairs( a )
137
Similarly, two clusters A and B can be considered to have a strong pair “joining”
( ) s.t. MaximalPairs( B )
( ) s.t. MaximalPairs( A )
As stated earlier, the algorithm uses this as the basis for merging clusters. When
two clusters have a strong pair between them, they are merged. A triplet (or larger) can be
merged when 2+ clusters have a mutual strong pair in a cluster whose maximal pairs span
Here, the maximal pair for clusters A and B is c1, and the maximal pairs for
cluster C are both a1 and b3. Thus, A-C and B-C are both strong cluster pairs.
introduction is then:
and B have a strong pair between them, according to our definition of strong maximal
pairs.
i.e. not a strong maximal pair. It should be noted that a pair that is weak during one
iteration of the algorithm may be strong in a later one, e.g., after merges occur between
to presenting the each class’s cluster membership, we also describe our initial
impressions of the results, examine each attribute’s relative “strength” for the cluster
assignments, justify any unusual traits, and discuss implications arising from them.
Delving deeper, the attribute details for each of the clusters are:
MAS Cluster 0
Student EditCnt EditLength NumCmts CmtLength NumTags NumRates Dist.
From
Centroid
0 11 134.091 6 93.167 0 9 73.197
2 6 27.667 6 70.500 5 18 38.391
4 5 50.000 6 61.000 5 7 18.270
6 7 46.857 6 43.167 12 7 32.714
8 11 37.636 7 72.714 4 17 28.766
14 9 95.667 9 75.111 14 9 31.931
Avg 8.167 65.320 6.667 69.276 6.667 11.167 37.211
StDev 2.563 41.012 1.211 16.543 5.279 4.997 -
Table 5.9: Attribute details for MAS cluster 0
MAS Cluster 1
Student EditCnt EditLength NumCmts CmtLength NumTags NumRates Dist.
From
Centroid
1 7 47.43 5 93.6 27 9 32.117
5 6 89.17 4 140 12 7 32.117
Avg 6.500 68.298 4.500 116.800 19.500 8.000 32.117
StDev 0.707 29.513 0.707 32.810 10.607 1.414 -
139
MAS Cluster 2
Student EditCnt EditLengt NumCmt CmtLengt NumTags NumRate Dist.
h s h s From
Centroid
3 3 53.667 3 76.333 6 5 32.339
7 4 141.000 2 102.500 4 4 60.791
9 0 0.000 3 94.000 14 8 88.127
10 2 124.000 3 86.333 1 5 39.587
11 3 63.000 3 87.667 16 1 27.312
13 2 61.000 7 22.429 4 4 60.040
15 2 58.000 3 61.333 2 1 32.457
16 5 187.200 1 84.000 5 5 101.536
Avg 2.625 85.983 3.125 76.824 6.500 4.125 55.273
StDev 1.506 59.869 1.727 25.108 5.503 2.295 -
Table 5.11: Attribute details for MAS cluster 2
MAS Cluster 3
Student EditCnt EditLength NumCmts CmtLength NumTags NumRates Dist.
From
Centroid
12 0 0 0 0 0 0 0
Avg - - - - - - -
StDev - - - - - - -
Table 5.12: Attribute details for MAS cluster 3
5.2.3.1.1 Observations
The most notable impression of these cluster assignments is that Clusters 1 and 3
have fewer members than clusters 0 and 2. Looking into its attributes, the sole instance in
Cluster 3 is a student that did not contribute to the wiki. It is good that the post-maximal
pairs results singled out this extreme of lacking activity! On the other hand, Cluster 1
does not seem to represent a particular extreme of (quantitative) contributions to the wiki,
Number of edits
Clusters 0 and 1 seem to jointly comprise the “high” (5+) edit counts. However,
the division between clusters 0 and 1 based on this attribute alone is not strong:
Clusters 2 and 3 seem to comprise the “low” (< 5) edit counts. Some separation
(i.e. the two clusters do not overlap within one stdev from their
Edit length
cluster assignments. The clusters have a mix of “relatively low” and “relatively high” edit
lengths. That is, edit lengths for clusters 0, 1, and 2 all overlap one another within one
stdev of their respective means. However, it may be possible for this attribute to be a
distinguishing factor between clusters 2 and 3. This is currently uncertain since cluster 3
Number of comments
Clusters 0 and 1 seem to comprise the “high” (4+) comment counts. Some
separation between clusters 0 and 1 based on this attribute seems to exist: does
not overlap with (i.e. doesn’t overlap within one stdev). However,
overlaps with .
141
Clusters 2 and 3 seem to comprise the “low” (< 4) comment counts, although
student 13 appears to be an outlier for this attribute with a comment count of 7. Some
separation between clusters 2 and 3 seems to exist: does not overlap with
Comment length
and cluster assignments. The clusters have a mix of “relatively low” and “relatively high”
comment lengths, and comment lengths for clusters 0, 1, and 2 all overlap one another
Cluster 1 has a notably higher average comment length than the other clusters.
Perhaps this can be a distinguishing factor between clusters 0 and 1? It may also be
possible for this attribute to be a distinguishing factor between clusters 2 and 3. However,
Number of tags
assignments and the number of tags provided. The students with high tag counts are
distributed across clusters 0, 1, and 2, which suggests the attribute’s lack of relevance in
the cluster assignments. Tag counts for clusters 0, 1, and 2 all overlap one another within
Number of rates
142
overlaps with (i.e. overlap within one stdev from the means) for this
attribute.
although student 9 appears to be an outlier for this attribute (8 ratings). Some separation
(i.e. the two clusters do not overlap within one stdev from their
Interestingly, the action counts seem to play a larger role in determining cluster
students:
5.2.3.1.2 Justifications
Why are clusters 1 and 3 so small?
143
For the singleton cluster 3, it was previously stated that the assignment appears to
be suitable, since the student did not participate in the wiki collaborate phase. With
values of 0 for every attribute, it is likely to be a relatively larger distance away from the
other instances.
Also previously stated, cluster 1 did not appear to consist of students who
appeared to be extreme outliers as with cluster 3. Rather, their values on the primary
categorization attributes trend towards the lower end of the range for the corresponding
the cluster may perhaps represent a niche “middle ground” between clusters 0 and 2.
counts?
One possible explanation for this emphasis on activity counts is that students may
believe that the raw action counts factor into their grade. Thus, the number of edits may
be artificially inflated via adding content in small increments and/or making multiple
“minor” edits (e.g., fixing typos and text formatting) after the “meat” of the content is
agreement to an opinion or ones providing “obvious” remarks that require little insight.
Since rates are also easy to perform, these may also be done in high quantity.
Tags may be relatively tricky to contribute towards, as good ones are relevant to
page content and/or related categories. It can be difficult to contribute additional tags
when the few obvious tags are already entered, making it harder for students to “inflate”
144
tag counts when they are not among the first collaborative contributors. As such, it would
be difficult for this attribute to be correlated in the fashion of the other activity counts.
Why would edit and comment lengths have a lesser bearing on categorization? Is
Combined with the focus on activity counts, a general lack of correlation between
edit and comment counts vs. lengths may contribute towards their apparent lack of
importance in categorizing students in this data set. This may also be a result of student
5.2.3.1.3 Implications
What are the implications of the results for the MAS class? We’ve identified the
As seen in the above results, half of the clusters are of considerably smaller size
relative to the others, and a larger sample size will aid in confirming the validity of the
clusters 1 and 3 will clarify the distinctions between clusters 0-1 and clusters 2-3,
respectively. Additional instances for clusters 0 and 2 may also tighten the
clustering.
145
emphasizes attributes that we value more, akin to the weighted sum used to emphasize
edit and comment actions. While this introduces a bias to the clustering results, such
actions over others. Different activities might be weighted differently based on the
instructor and assignment metrics, and the different weights serve to motivate students
differently in their activities. So, a more prudent approach would be to incorporate these
assignment scoring weights into the distance formula. For example, clusters based
primarily on edit and comment count, length, and quality will be of greater interest for a
The attribute range information for each of the clusters can be leveraged to guide
recommendations for a user, relative to characteristics of its cluster peers. That is, if a
student is identified to belong to a cluster that does not favor performing ratings or
could be a lower priority. It can also be used to guide “reminders” for particular actions
when a student’s performance is lacking relative to its cluster peers. Caution should be
taken to avoid “locking” students into a cluster – such recommendations may reinforce
the student behaviors that place them into the cluster to begin with.
RAIK Cluster 0
Student RaikNumEdits RaikEditLength NumCmts CmtLength NumTags NumRates Dist.
From
Centroid
0 4 408.250 6 108.333 0 11 211.377
2 3 285.000 1 113.000 0 18 93.736
3 6 256.667 2 99.500 3 7 62.113
8 5 48.200 1 52.000 0 5 153.172
11 3 57.667 0 0.000 0 16 160.856
14 3 141.000 1 81.000 0 0 59.489
Avg 4.000 199.464 1.833 75.639 0.500 9.500 123.457
Std Dev 1.265 141.835 2.137 43.227 1.225 6.834 -
RAIK Cluster 1
Student RaikNumEdits RaikEditLength NumCmts CmtLength NumTags NumRates Dist.
From
Centroid
1 5 106.400 4 33.000 34 16 0
Avg - - - - - - -
Std Dev - - - - - - -
RAIK Cluster 2
Student RaikNumEdits RaikEditLength NumCmts CmtLength NumTags NumRates Dist.
From
Centroid
4 7 83.714 4 69.250 10 6 20.501
147
5.2.3.2.1 Observations
In terms of quick, immediate impressions of the results, the presence of only three
clusters in the results is particularly noteworthy. It seems unusual for our algorithm to
end with three clusters when the target number of clusters for this data set (and
As with the MAS class, the RAIK class also has a singleton cluster. However at a
glance, it is difficult to tell whether this cluster assignment is appropriate for the instance,
Number of edits
148
Excluding the singleton cluster, there seems to be split based upon edit counts.
Cluster 0 comprises the “low” edit counts and Cluster 2 comprises the “high” edit counts,
and the separation is somewhat distinct: Clusters 0 and 2 do not overlap within one
Based on the above, Cluster 1 is aligned closest to cluster 0 for this activity.
Edit length
Excluding the singleton cluster, there seems to be split along “high” and “low”
edit lengths. The average edit length of cluster 0 is distinctly higher than that of cluster 2.
However, the clusters overlap within one standard deviation, i.e., and
Note: outliers do exist in the cluster assignments, i.e. instances with “low” edit
length in cluster 0 and “high” edit length in cluster 2. The following are the noted outliers
Cluster 1 is more closely aligned to cluster 0 than cluster 2 for this activity.
Number of comments
149
cluster assignments. Clusters 0 and 2 overlap within one standard deviation. That is, the
Comment length
There doesn’t seem to be a strong correlation between comment length and cluster
assignments. Clusters 0 and 2 overlap within one standard deviation, i.e., the intervals
and overlap.
characteristic. However, this currently cannot be confirmed due to the lack of members in
this cluster.
Number of tags
Cluster 0 comprises the “low” tag counts, and cluster 2 comprises the relatively
higher tag counts. The intervals for the two clusters do not overlap within one standard
deviation. That is, and do not overlap. However, the two clusters overlap
150
noted that cluster 2 has a relatively large standard deviation on this action and contains
For this action, cluster 1 is more-closely aligned with cluster 2. However, the
number of tags its instance performed is the highest out of all instances, suggesting that
Number of rates
and cluster assignments. The number of ratings for all three clusters overlaps within one
students:
5.2.3.2.2 Justification
Why 3 clusters?
151
Note that the check for the number of clusters remaining is on the outermost loop.
That is, our algorithm currently stops after all clusters with strong max pairs are merged
for a particular iteration of the while-loop. With this, it’s possible for the number of
clusters to go from above the target number to below during a single iteration. In this
particular case, the final iteration of our algorithm started with five clusters and
categorizing the students in this class. This is justifiable since 90% of the collaborative
contribution grade is based solely upon edits and comments, and this students aiming for
A particular feature of this data set is that it seems to be split between two
particular editing /commenting paradigms: 1) low count, high length and 2) high count,
low length. The cluster results here demonstrate that there may be a tradeoff between
number of edits/comments versus average edit/comment length – that is, due to limited
time and/or cognitive resources, students will either make few large contributions or
152
many small contributions to the wiki pages. This tradeoff is expounded upon further in
Section 5.3.
Previously, our impression was that the purpose of the singleton cluster wasn’t
readily apparent. After examining the cluster instance’s attributes relative to the others, it
appears that the cluster may represent the middle region between the relative extremes of
low counts with high lengths and high counts with low lengths. In particular, it shares the
low count and high length paradigm for edits, but high count and low length for
comments.
5.2.3.2.3 Implications
The following implications were derived from the maximal pairs algorithm results.
As with the MAS results, our procedure resulted in another set of clusters where
one is a singleton cluster. A larger sample size will also aid in: 1) confirming the validity
of the observations and justifications drawn, and 2) further distinguishing cluster 1’s
attributes for this data set. See the corresponding implication for the MAS class for
further detail.
As with the MAS class, tailoring the distance formula used for generating the
clusters may increase their value to users leveraging the cluster assignments for decision
making. It may be particularly relevant in this case since the student collaborative
153
contributions are graded in a slightly different manner between the MAS and RAIK class,
i.e. the number of tags is not part of the grade for the RAIK class. Refer back to the
membership to these clusters. See the corresponding discussion in the MAS class results
This data set highlights the issue that it is possible for our procedure to end with
fewer clusters than the target number desired. With this, the possibility exists for it to end
with much fewer clusters than the target number, e.g., ending with one or two clusters
when four or five are desired, and can arise when multiple multi-cluster maximal pairs
are available. While the probability of our algorithm resulting in much fewer clusters than
the target number is rare, we wish to re-evaluate the stopping condition and determine
Due to this inconsistency between the target number of clusters and the number of
clusters in our algorithm results, we subsequently evaluated the results arising from
stopping our algorithm at the point when exactly four clusters remain.
The four-cluster results are largely similar to these three-cluster results, with two
of the clusters formed by “reverting” one of the clusters from 5.2.3.2 to a pre-merge state.
While this “split” identifies an additional level of granularity when categorizing students,
154
it does not exhibit any strong deviations from the observations and conclusions drawn
from the three-cluster data. For further detail, please refer to Appendix B.
5.2.4 Summary
In this section, we introduced and applied our Maximal Pairs Algorithm to the
results obtained from the X-Means clustering algorithm, using the raw tracked attribute
data collected from student activity as the attributes upon which the clusters are formed.
For the MAS class, categorization appears to be primarily based upon the number
For the RAIK class, categorization appears to be primarily based on the number
As we can see, attributes pertaining to edits and comments are a common factor in
clustering the students. This is consistent with our expectations that student behavior is
centered on these attributes, since those two collaborative elements comprise 90% of the
This leads to the question: why are there differences in categorization factors
despite the collaborative wiki assignment being similarly structured and graded between
the two classes? A possible answer is that the general approach or strategy taken by the
With the RAIK students, emphasis is placed on edits which covers the number of
edits, the average edit lengths, and the tradeoff between the two. A similar but lesser
between number and length observed in the clusters formed. This is in line with the
expectation for edits to be prioritized due to their 60% weight in the grade and for
The MAS class may be focused on the counts of the different actions instead, with
students possibly believing that performing the actions more times results in a better
grade. Particularly, the primary categorization for this class is based on number of edits,
number of comments, and number of ratings, all three of which correspond to the three
action types contributing towards the collaboration grade for the assignment. With this in
mind, these actions are still in line with expectations for the assignment, even if the
One interesting thing to note is that while the average number of tags created is
relatively similar between the two classes, the average number of ratings performed for
the MAS class is significantly lower (by approximately 50%). Recall in chapter 5.1.2
that the grading for the final 10% of the assignment differs slightly between the two
classes. Specifically, the final 10% for the RAIK class is based solely on ratings provided
whereas the final 10% for the MAS class consists of ratings, tags, and views. With this
difference in criteria, the RAIK class places greater emphasis on the number of ratings
given, since the 10% is based solely on number of ratings. On the other hand, this 10% is
“spread” between ratings, tags, and views for the MAS class, and the average number of
Overall, what does all of this suggest? On one hand, factoring in the idea of
different strategies towards the assignment improves user modeling for recommendations,
since students would be more receptive towards recommendations that are in line with
their strategy. Also, this confirms to a certain extent that the instructor’s expectations, e.g.
grading requirements, for the wiki assignment will correlate to student behavior on the
wiki.
the wiki assignment, “quantity vs. quality”, and the relative “ease” of making a
Student resources for working on the assignment, e.g., time and effort, are
generally limited, and the total amount of such resources vary from student to student
based on their individual schedules. Thus the factors listed may intuitively guide how
The discussion for the first item, evaluation criteria of the wiki assignment, has
Working on the assumption that student resources for working on the wiki
assignment are limited, “quantity vs. quality” refers to the tradeoff between the quality
and the number of contributions made. Students with “more” total resources can appear
to have more contributions of better quality than those with “less” total resources. The
157
degree to which students focus on quality or quantity may be affected by the “ease of
contribution towards a page, and this may vary depending on the action taken and its
timing. For instance, it may be “easier” to make a relatively high-quality edit to a page
that is initially poorly written, and “harder” to make an impactful edit to a well-written
page. Additionally, it is “easier” to rate or write tags for a page than it is to make an edit
or a comment. Timing plays a role in that students who act earlier have access to more
requirements and evaluation criteria, student behavior can also be influenced by the
Clusters found via this approach can be leveraged to guide recommendations and
profiling of students.
found via this approach can be used to guide student profiling and recommendations.
Since the clustering is based on behavior, the clusters can be used as “behavioral
archetypes” for tailoring assistance towards groups of users. Recommendations can thus
recommendations that the user would be likely to take (reinforcing the archetype) or
158
helping the user develop new habits outside of her current behavior (breaking the
archetype).
the actions tracked within the wiki, we consider the categorization of active/passive
actions to be based upon the count and type of actions performed. That is, we will count
the number of active actions (i.e. comments and edits) and the number of passive actions
(i.e. rates and add tags), and based on how the two compare for a given student, a
categorization is made. However because the passive actions can be easily made on a
larger order of magnitude than the active ones, we weight the total active actions prior to
the comparison. We propose the following weighted sum to calculate the weighted total
collaborative actions:
( ) ( ) ( ) ( )
Note that this categorization based on action types is different from categorizing
based on overall activity level, i.e. number of actions performed. For instance, a user with
only 1 edit as her sole collaborative activity has a type distribution of 100% active,
whereas a user with 10 edits and 30 rates would have a type distribution of 50% active. In
terms of activity level, the latter user has a greater activity count, but the former has a
159
larger proportion of activities of the “active” type. However, the person with 10 edits and
Tables 5.18 and 5.19 show the derived counts for the two action types in the
RAIK Cluster 0
Student Rate Comment Add Edit Total Active Weighted Total % Active
ID Tag Actions Total Active Passive Actions
(Edits + (3 * Edits + Actions
Comments) 2* (Rate +
Comments) Add
Tag)
1 11 6 0 4 10 24 11 69%
4 7 2 3 6 8 22 10 69%
160
5 6 4 10 7 11 29 16 64%
7 17 2 0 9 11 31 17 65%
8 17 5 8 11 16 43 25 63%
9 5 1 0 5 6 17 5 77%
10 8 4 11 11 15 41 19 68%
11 12 2 5 9 11 31 17 65%
15 0 1 0 3 4 11 0 100%
Table 5.20: RAIK active/passive action categorization – Cluster 0
RAIK Cluster 1
Student Rate Comment Add Edit Total Active Weighted Total % Active
ID Tag Actions Total Active Passive Actions
(Edits + (3 * Edits + Actions
Comments) 2* (Rate +
Comments) Add
Tag)
6 29 6 16 8 14 36 45 44%
13 14 4 25 13 17 47 39 55%
14 15 3 3 5 8 21 18 54%
Table 5.21: RAIK active/passive action categorization – Cluster 1
RAIK Cluster 2
Student Rate Comment Add Edit Total Active Weighted Total % Active
ID Tag Actions Total Active Passive Actions
(Edits + (3 * Edits + Actions
Comments) 2* (Rate +
Comments) Add
Tag)
2 16 4 34 5 9 23 50 32%
3 18 1 0 3 4 11 18 38%
12 16 0 0 3 3 9 16 36%
Table 5.22: RAIK active/passive action categorization – Cluster 2
5.3.1.1 Observations
The following were observed from the RAIK data.
Clusters:
Correlations
algorithm on the % active actions value, two key correlations were observed. First, there
is a correlation of 0.531 between total active actions and total passive actions. Second,
collaborative actions and % active actions. These are relatively strong and worth
5.3.1.2 Justification
Biases towards particular actions?
It is likely that a bias towards the active actions is introduced in student behavior
since the majority of the assignment grade is based on edits and comments. Students that
have a high action count for both passive and active types may either be overachieving
Why the positive correlation between total active actions and total passive actions?
The correlation between total active actions and total passive actions (and
consequently a possible cause for the high number of students with “balanced” action
Since grades are also based on overall contribution quality, students who are
unable to make a sizeable contribution in “one shot” (i.e. late contributors) may
Lacks in edit and comment quality may also be (somewhat) compensated for with
the minor activities (i.e. rating and tagging) since they also contribute a small
There may be students who believe that action counts factor into their grade, so
the number of edits may be artificially inflated via adding content in small
increments. Since tags and rates are also easy to perform, these may also be done
in high quantity.
Why the negative correlation between the total number of unweighted collaborative
Students who are able to make sizeable, significant, and/or high quality
additional actions for a “good grade.” As such, they may not be motivated to
perform many passive actions (i.e. rates and tags). With the low number of edits
and comments made dominating their low number of total actions, their profile
for them. Since ratings and tags are easier to provide, they can be carried out in
higher quantity, thus increasing the number of total actions and decreasing the
5.3.1.3 Implications
Tradeoff between percentage active actions and collaboration initiation rate.
initiation rate (i.e. fewer total collaborative actions, as suggested in the third justification
point). Conversely, students with a lower active action percentage have a higher
collaboration initiation rate (i.e. greater total collaborative actions, as suggested in the
Cross comparison with the MAS class will be necessary to determine if this
implication is true only for the RAIK class or if it may be applicable in general.
The percent active metric (and subsequent categorization based upon it) is not
adequate on its own to profile a student, since it does not account for the absolute
164
quantity of the active/passive actions performed. However, this metric may be useful to
instructors by providing the instructor with a quick overview of how the student is
editing wiki pages and participating in discussions, a low % active actions may serve as a
“red flag” identifying students who may be struggling with making such contributions.
The instructor may also tweak the weights used in our metric according to the
For future work, the raw total passive actions and (weighted) active actions can be
The results from clustering the students based on the % Active Actions value via
MAS Cluster 0
Student Rate Comment Add Edit Total Active Weighted Total % Active
ID Tag Actions Total Active Passive Actions
(Edits + (3 * Edits + Actions
Comments) 2* (Rate +
Comments) Add
Tag)
2 9 5 27 4 9 23 36 39%
10 8 3 14 0 3 9 22 29%
13 0 0 0 0 0 0 0 0%
Table 5.24: MAS active/passive action categorization – Cluster 0
MAS Cluster 1
Student Rate Comment Add Edit Total Active Weighted Total % Active
ID Tag Actions Total Active Passive Actions
(Edits + (3 * Edits + Actions
Comments) 2* (Rate +
Comments) Add
Tag)
1 9 6 0 5 11 28 9 76%
3 18 6 5 3 9 24 23 51%
4 5 3 6 3 6 15 11 58%
5 7 6 5 4 10 26 12 68%
6 7 4 12 5 9 22 19 54%
7 7 6 12 4 10 26 19 58%
8 4 2 4 4 6 14 8 64%
9 17 7 4 5 12 31 21 60%
11 5 3 1 2 5 13 6 68%
12 1 3 16 3 6 15 17 47%
14 4 6 4 2 8 22 8 73%
15 9 8 14 8 16 40 23 63%
16 1 3 2 2 5 13 3 81%
17 5 1 5 4 5 11 10 52%
Table 5.25: RAIK active/passive action categorization – Cluster 1
5.3.2.1 Observations
The following was observed in the MAS data.
166
Clusters:
o The target MinK value for this class was determined to be two instead of
Action .
Correlations
whether the correlations discovered with the RAIK data also apply to the MAS class. As
with the other data set, there is a relatively strong correlation (0.519) between total active
actions and total passive actions in the MAS class. However, unlike the RAIK class there
is little/no correlation (0.083) between % active actions and unweighted total actions.
5.3.2.2 Justification
Biases towards particular actions?
As with the RAIK class, it is likely that a bias towards the active actions is
introduced in student behavior since the majority of the assignment grade is based on
edits and comments (60% edits, 30% comments, 10% passive actions). Students that have
a high action count for both passive and active types may either be overachieving or may
be compensating for a lack of quality in each of their (active) actions performed (see next
justification point).
167
Why the positive correlation between total active actions and total passive actions?
The correlation between total active actions and total passive actions is also
observed in the MAS class. Please refer back to 5.3.1.2 for additional details.
5.3.2.3 Implications
Usefulness and adequacy of the active action percentage metric?
As previously stated in the RAIK implications, the percent active actions metric
(and subsequent categorization based upon it) is not adequate on its own to profile a
student, since it does not account for the absolute quantity of the active/passive actions
5.3.3 Summary
In this sub-chapter, we introduced the active vs. passive approach for categorizing
student activity profiles, and it is based on the concept of “active” (e.g., edits and
comments) and “passive” (e.g., rates and tag adding) action types. This metric computes
the percentage of actions in student’s activity profiles that are of the “active” type. Note
that this is different from computing the frequency at which students perform
A positive correlation exists between the number of total active actions and total
passive actions.
Students generally have limited resources (time, attention, etc.) when doing
assignments for the class, so it is assumed that there will be a tradeoff between the
168
number of active and passive actions due to the larger resource cost to perform the active
actions. However, this is shown to be incorrect in both RAIK and MAS classes – contrary
What is the motivation for a student to behave in this manner? Reiterating the
Since grades are also based on overall contribution quality, students who are
unable to make a sizeable contribution in “one shot” (i.e., late contributors) may
Lacks in edit and comment quality may also be (somewhat) compensated for with
the minor activities (i.e., rating and tagging) since they also contribute a small
There may be students who believe that action counts factor into their grade, so
the number of edits may be artificially inflated via adding content in small
increments. Since tags and rates are also easy to perform, these may also be done
in high quantity.
Tradeoff (or lack thereof) between the total number of unweighted collaborative
In the results for the RAIK class, a tradeoff was found between the % active
actions metric and the total number of unweighted collaborative actions. At a glance, this
169
result seems contradictory to the previous finding that total active actions and total
passive actions are positively correlated – if the two are correlated and there are a large
number of both active and passive actions in the user’s activity profile, then wouldn’t the
active actions made increase the active %, resulting in a positive correlation between the
two? To reconcile this, we need to consider that although total active actions and total
passive actions are positively correlated, active actions still cannot be performed as
Students who are able to make sizeable, significant, and/or high quality
additional actions for a “good grade.” As such, they may not be motivated to
perform many passive actions (i.e. rates and tags). With the low number of edits
and comments made dominating their low number of total actions, their profile
for them. Since ratings and tags are easier to provide, they can be carried out in
higher quantity, thus increasing the number of total actions and decreasing the
Interestingly, the MAS class did not exhibit this negative correlation between %
active action and unweighted total actions. Rather, the two appear to lack any correlation
in this data set. Why might this be the case? The key may be in the slightly different
170
grading criteria in the final 10% of the collaboration grade. Recall that for the RAIK class,
the final 10% is comprised solely of the ratings action. In the MAS class, rates, tags, and
views all comprise the 10%. The addition of tags to the MAS assignment grading
criteria motivates grade-conscientious students to also add tags. In the RAIK class,
this action had no bearing on the grade and was thus ignored by students able to
and the “overachievers” in the class has benefits that could aid in recommendation and
reading” within the wiki, and alerting the instructor when students performing minimal
overachiever scale to be based on the user’s actions relative to the minimum requirements
of the collaborative writing assignment. For both classes, this minimum requirement is to
tagging) on three different wiki pages. The assignment also specifies that edits must be
made to at least three other pages. Based on these two criteria, a student may choose to
make three edits total, each on a different page, as the minimum effort.
There are a couple difficulties in basing this metric on the minimum requirements
of the assignment. First, the baseline for minimalism in this particular assignment is
relatively low. Second, the relative scale for the number of collaborative actions
171
performed is also low (on the scale of tens), due to the limited time span of the
assignment and limited student resources. With these two traits, it is difficult to determine
The process for determining minimalists and overachievers among the students in
a class is as follows:
1. Cluster the students in the class based on each of the tracked attributes. This
clustering identifies the different activity levels (e.g. high, medium, and low)
among the students for each attribute, highlighting areas where a student is
2. Map student placement within each attribute to a score between -1 (low) and +1
3. Calculate the net minimalist/overachiever score for each student by summing their
The principle behind this particular scoring scheme is for “High” placements and
A negative one indicates minimalist behavior, with a more negative score indicating
stronger minimalistic tendencies. A net score near zero is indicative of a balance between
the two.
the scoring discussed above, we also wish to consider the “quality” of the user’s
contributions. We thus manually evaluate the content of each user’s prose (i.e., edits and
comments) and include it as two additional categories contributing towards the net score:
In the following sub-sections, we apply this metric to the students in the RAIK
Student Edit Edit Edit Cmt Cmt Cmt Rates Tags Net
ID Count Length Quality Count Length Quality Count Count Score
(2) (3) (3) (2) (2) (3) (2) (2)
1 L H H H H H H L +4
2 L M L H L L H H -1
3 L H H L H M H L +1
4 L H H L H M L L -1
5 L M M H L H L L -2
6 H L H H L H H H +4
7 H L H L L H H L 0
8 H L H H L M H L +1
9 L L L L L L L L -8
10 H L M H L H L L -1
11 H L M L L L H L -3
173
12 L L L L L L H L -6
13 H L M H H M H H +4
14 L M M H L M H L -1
15 L M M L L H L L -4
Table 5.27: RAIK student activity level for each tracked attribute. The number following each attribute, i.e. (2),
indicates the number of clusters used to categorize the students. See Table 5.25 for the score values mapped to
each activity level.
Clustering the students on their net scores gives us the following clusters:
RAIK Cluster 0
Student ID Net Score Collaboration Phase Score
1 +4 100
6 +4 95
13 +4 95
Average +4 96
RAIK Cluster 1
Student ID Net Score Collaboration Phase Score
2 -1 85
3 +1 95
4 -1 100
5 -2 100
7 0 85
8 +1 *
9 -8 85
10 -1 90
11 -3 85
12 -6 50
14 -1 90
15 -4 *
Average -2.08 86.50
Table 5.28: Clusters and members for the RAIK class, based on minimalist vs. overachiever net score. Collaboration
phase score is also listed for comparison. Asterisks denote users who did not consent to the use of assignment
grade for this analysis.
5.4.1.1 Observations
The two clusters formed by the Maximal Pairs algorithm appear to consist of the
extreme overachievers for the class (Cluster 0) and the remaining students (Cluster 1).
Due to this basis for grouping the students, there is a relatively lopsided distribution
between the two, with 20% of the students placed in Cluster 0 and 80% placed in Cluster
1.
174
We are interested in whether the students’ net scores correlate to the grades they
received during the collaboration phase of the wiki assignment, and the grades for
consenting students are thus listed alongside their net scores in Table 5.27. Excluding the
non-consenting users from the calculation, the correlation coefficient between the net
score and collaboration phase score is +0.529. It should also be noted that the average
grade of the overachievers’ cluster is higher than that of the other cluster.
Additionally, we are also interested in whether the quantity vs. quality tradeoff
identified in previous sections is still present in these results. In this analysis, we identify
tradeoffs (or negative correlation) between two attributes to be a situation where one
attribute is ranked high (H) and the other is ranked low (L). Consequently, if either are
mid ranked (M), then it does not count. We also identify positive correlation between two
attributes to be a situation when the rank for the two attributes match, such as when both
The following were observed in the categorizations for edits and comments
specifically:
Edits
o A tradeoff between count and length exists for 9 out of 15 students. There
are 2 students for which count and length are positively correlated.
o For edit quality vs. count, 3 out of 15 students exhibit a tradeoff between
the two attributes. However, the two appear to be positively correlated (i.e.,
correlation (i.e. length level matches quality level) between edit quality
Comments
pattern was found: students on the extreme ends of the spectrum (e.g.,
o For comment quality vs. count, only 3 out of 15 students exhibit a tradeoff
students. There does not appear to be a distinct relation between the two
attributes.
176
5.4.1.2 Justifications
Why is there a positive correlation between Net Scores and Grades?
Since the net scores are based upon the same criteria for assignment grades, there
is a positive correlation between the two. In particular, the grading by the instructor
places particular emphasis on counts (i.e. comments and edits across 3 pages, for a
contribution length for edits and comments, and the tradeoff is more immediately
apparent for edits. The tradeoff on edits can be explained per previous discussion in sub-
chapter 5.2.3.2.2. More interestingly, the tradeoff between comment count and length
appears to be dependent on the student’s net score: the extreme minimalists and
overachievers have a positive correlation between comment count and length, whereas
the students in between the extremes exhibit the tradeoff. Intuitively, this can be justified
However, there is a lack of a tradeoff between length and content quality for both
edits and comments in general. This may be attributed to the two not being particularly
correlated, at least for this class. Intuitively, using more words is associated with more-
detailed explanations, which consequently leads to the idea that a longer contribution is
contributions can be made in relatively few words if the writer is concise, and
contribution length can be bloated while adding little value if the writer is long-winded.
This can similarly explain the lack of correlation between contribution count and content
quality.
5.4.1.3 Implications
Usefulness of Two Clusters
The Maximal Pairs algorithm identified 2 as the optimal number for categorizing
this class on the students’ net scores. Upon closer inspection, there are three possible
else
One cluster containing extreme minimalists, one cluster containing everyone else
One cluster containing overachievers (both extreme and slight), one cluster
In this particular case, the clusters formed are one with extreme overachievers and
one with the rest of the students. While this can be useful for identifying which students’
works to highlight (average grade of 96), these clusters are not as useful for identifying
178
minimalists for the instructor. Of the listed outcomes for 2 clusters, only the second
would be particularly useful for this scenario, as “slight” minimalistic tendencies might
results to 3. With it, it is possible for the clustering to group the students into extreme
overachievers, extreme minimalists, and the remaining students in between. The results
of such a clustering would be more valuable for application to the suggested scenarios in
One of the potential uses of the net score is to assist instructors in grading student
activity in the wiki. However, the current calculation of the net score may not be
The range of collaboration phase grades for the majority of students is also
relatively narrow, spanning from 85 to 100 points. Although the calculated correlation
between the net score and grade is +0.529, one can see that some of the grades received
may not align with expectations arising from net score, when using those two values as
the benchmarks for overachieving and minimalist behavior. Thus, some modifications to
the net score calculation will be needed before we can confidently use it as an indicator
for grades.
179
One possible area for modification would be in the weighting of the individual
tracked attributes. Currently, each category of tracked activity is equally weighted with
one another, contributing up to +/- 1, i.e. 1/8th, of the total net score. By modifying the
weights in this sum to reflect assignment expectations, the net score may better reflect the
Student Edit Edit Edit Cmt Cmt Cmt Rates Tags Net
ID Count Length Quality Count Length Quality Count Count Score
(1) (2) (3) (2) (2) (3) (2) (2)
1 M H H H H H H L +5
2 M L M H H H H H +4
3 M L M H H M H L +1
4 M L L L H H L L -3
5 M L M H H M H L +1
6 M H M L H H H H +4
7 M L L H L M H H 0
8 M H M L H M L L -1
9 M L H H H H H L +3
10 M L L L H H H H +1
11 M H H L H H L L +1
12 M L M L H M L H -1
13 M L L L L L L L -7
14 M L L H L L L L -5
15 M H M H H H H H +6
16 M L M L H H L L -2
17 M H H L H H L L +1
Table 5.29: MAS student activity level for each tracked attribute. The number following each attribute, i.e. (2),
indicates the number of clusters used to categorize the students. See Table 5.25 for the score values mapped to
each activity level.
Clustering the students on their net scores, we obtain the following clusters:
MAS Cluster 0
Student ID Net Score Collaboration Phase Score
1 +5 135
2 +4 100
180
3 +1 95
5 +1 100
6 +4 100
7 0 85
9 +3 100
10 +1 35
11 +1 80
15 +6 105
17 +1 85
Average 2.45 92.73
MAS Cluster 1
Student ID Net Score Collaboration Phase Score
4 -3 83
8 -1 90
12 -1 95
13 -7 0
14 -5 70
16 -2 75
Average -3.17 68.83
Table 5.30: Clusters and members for the MAS class, based on minimalist vs. overachiever net score. Collaboration
phase score is also listed for comparison.
5.4.2.1 Observations
The two clusters formed by the Maximal Pairs algorithm yields a cluster of
students whose net scores are positive (Cluster 0), and a cluster of students whose net
scores are negative (Cluster 1). This distribution of students is relatively more balanced
compared to the RAIK class, with 64.7% (11 out of 17) belonging to Cluster 0 and 35.3%
As with the RAIK class, we are interested in whether the students’ net scores
correlate to the grades they received during the collaboration phase of the wiki
assignment. The grades for the students are thus listed alongside their net scores in Table
5.30. The correlation coefficient between the net score and collaboration phase score is
+0.724, which is higher than that of the RAIK class. It should also be noted that the
average grade of the overachievers’ cluster is higher than that of the other cluster, at
92.73 vs. 68.83. This still holds when the relative outlier scores for each cluster (35 and
181
135 for Cluster 0 and 0 for Cluster 1) is excluded, averaging at 94.44 for Cluster 0 and
Additionally, we are also interested in whether the quantity vs. quality tradeoff
identified in previous sections is still present in these results. The following were
Edits
o Since the MAS class lacks distinct separation in its edit counts to
positive correlation between count and length exists for this class.
o Similarly, the relation between edit counts and quality is not distinct, due
quality.
Comments
in the RAIK class, the students on the extreme ends of the spectrum
appear to exhibit the positive correlation whereas the ones towards the
correlation status and the students’ net scores. The relation between
and length. No tradeoffs were observed between these for any of the
5.4.2.2 Justifications
Why is there a positive correlation between Net Scores and Grades?
Since the net scores are based upon the same criteria for assignment grades, there
is a positive correlation between the two. In particular, the grading by the instructor
places particular emphasis on counts (i.e. comments and edits across 3 pages, for a
As previously highlighted, there is a tradeoff between the edit count and edit
length for the RAIK class. However, this tradeoff is not observed via this this analysis for
edits. Since the Maximal Pairs algorithm identified the optimal number of clusters for
categorizing on edit count to be 1, we cannot observe the count vs. length tradeoff via the
high/mid/low categories.
183
On the other hand, a tradeoff between comment count and comment length
similar to that of the RAIK class is observed. The students’ net scores also play a role in
the MAS class: the extreme minimalists and overachievers exhibit a positive correlation
between the two attributes, and the students in between exhibit the tradeoff.
As with the RAIK class, there is a lack of a tradeoff between count and content
quality for both edits and comments in general. Instead, a positive correlation is observed
for edit and comment lengths and content quality instead. This follows the intuition that
5.4.2.3 Implications
Usefulness of Two Clusters
As with the RAIK class, the optimal number of clusters for the MAS class is 2.
However, the clusters appear to be of a different outcome (#3 of the ones listed in sub-
chapter 5.4.1.3) – that of grouping overachievers, both slight and extreme, into one
cluster and minimalists into the other. This class may similarly benefit from forcing the
Please refer back to the discussion in sub-chapter 5.4.1.3 for additional discussion on this
topic.
There is a stronger correlation between net score and grade for this class (+0.724
for MAS vs. +0.529 for RAIK). However, it may be possible to strengthen the correlation
even further. Please see the corresponding heading in section 5.4.1.3 for additional
information.
5.4.3 Summary
In this sub-chapter, we introduced the notion of minimalist vs. overachiever
activity within the wiki as well as our approach towards measuring where students fall
within the spectrum. It differs from the previous categorization analysis in sub-chapter
5.2 in that the previous analysis treats the tracked attributes as a holistic entity, whereas
examining the tracked attributes independently, we can rank each student’s activity level
on each of the individual attributes. These individual ranks are then combined into a
spectrum.
discussion points:
Count vs. Length, Count vs. Quality, and Length vs. Quality are three distinct
As seen in the results of the individual classes, it is easy to conflate and intuitively
assume particular correlations between the three attributes of count, length, and quality
185
used for edits and comments. The general intuitive belief is that the three share the
following relations:
Count and length are believed to be negatively correlated and thus have a tradeoff
and/or cognitive resources, students will either make few large (length-wise)
Count and quality are believed to be negatively correlated and thus have a
tradeoff relationship. The reasoning for this belief is akin to the one between
count and length, where one needs to sacrifice “quality” if “quantity” is desired,
and vice-versa.
Length and quality are believed to be positively correlated and thus have a
directly proportional relationship. The reasoning is that using more words allows
In the RAIK class, this tradeoff is relatively apparent with 9 out of 15 students
exhibiting a high-low tradeoff between edit count and edit length. While two instances of
low-low positive correlation were found, they belonged to students scoring as extreme
minimalists – relatively rare occurrences in both class’s data. This generally supports the
However, we are unable to confirm or debunk this finding in the MAS class, due
to all students being placed into one cluster for edit counts. Why did this unfavorable
186
result surface for this attribute? Unfortunately, a number of factors can play a role into
this happening, including but not limited to: the distribution of edit counts being too
tightly packed to form consistent clusters, the distribution of edit counts is sub-optimal
There is a possibility that the relation between these two edit attributes may also
be influenced by the student’s net score, as observed in the upcoming discussion between
Both positive and negative correlations are observed between comment counts
and comment lengths. The correlation type is dependent on the absolute value
For both the RAIK and MAS class, it was observed that students on the extreme
ends of the spectrum have positive correlation between the two attributes. That is,
true to their namesakes, overachievers are willing to spend the cognitive effort to make
multiple lengthy comments, whereas minimalists are not motivated to perform up to the
threshold where the tradeoff is visible. The remaining students in between the
Why is this observed for comments but not edits? While both comments and edits
are indisputably the activities with the highest cognitive cost relative to the other tracked
activities, we believe that between the two of them, edits generally have the higher
cognitive cost to perform since the topics for edits are limited to what is relevant and
useful to the topic. In conversation threads, students can ask questions, teach, bounce
ideas, and are generally not as limited. Consequently, it is more difficult to exhibit
187
overachieving behavior through edits. As for minimalist behavior, the RAIK class
hints at the possibility of extreme minimalists exhibiting positive correlation between the
two attributes. There are already two such instances in the data. However, we cannot
verify this in the MAS class due to the rankings for edit counts. Data from more
classes/students will be needed to verify whether this tradeoff between counts and lengths
A positive correlation is observed between length and quality for both comments
and edits.
In the MAS class, it was observed that there is a positive correlation between
content length and content quality for both edits and comments. This supports the
intuitive belief listed at the start of the discussion. But why does this not apply to the
RAIK class? As discussed in 5.4.1.2, this expectation for a positive correlation may be a
the writer is concise. Similarly, contribution length can be bloated while adding little
that the positive correlation is dependent upon the writing skill of the student or the
general writing skill of the class. That is, a skilled writer is able to “say more with
fewer words,” whereas a lesser-skilled one may need more words to convey the same
quality of ideas.
grading
188
than +0.500) between the calculated net score and the grade that the student earned for
the wiki collaboration. With such strong correlations that can be improved via weighting
net score calculation according to assignment criteria, it seems feasible to use this metric
The metric is particularly adept at identifying the quantifiable work put in by students,
There are two particular challenges to using this metric as an aid to grading. First
relative to the other students in the class. That is, particular net scores do not correspond
to specific grades, nor are they comparable to the net scores of students in other classes.
For example, it was observed that some students with relatively extreme minimalistic
tendencies still attain scores such as 70 and 85 in spite of having a relatively low net
score. While this may not make sense on an absolute scale (e.g., on the assumption that
“a net score of -8 should always correspond to a grade of 0”), the received grade can
make sense on a relative scale, since the 70-85 is towards the lower end of the grades
distributed. Thus, when using the Net Score to assist in determining a grade, the
instructor still needs to determine the relative score range for the class.
or rank the quality of a contribution’s contents, so this ranking may need to be performed
manually. The edit and comment quality are arguably the most vital parts of a student’s
collaborative contributions, and thus the net score metric is not a total substitute for the
A discussion point arising in the RAIK and MAS class data is the optimal number
of clusters for the minimalist vs. overachiever categorization. For these classes, k was
reiterate the points raised, there are three possible cluster outcomes for k = 2:
else
2. One cluster containing extreme minimalists, one cluster containing everyone else
3. One cluster containing overachievers (both extreme and slight), one cluster
For recommendation, all three outcomes when k=2 have their flaws. Clusters
1 and 2 are only able to identify and consequently generate recommendations that are
only useful to one class of users, the extreme minimalists or extreme overachievers. In
both those cases, the cluster containing everyone else spans too large a range of user
behaviors such that one recommendation to the group would not meet the needs of all of
the users within it. By forcing the optimal number of clusters to 3, it is possible for
the cluster the students into a group of extreme overachievers, a group of extreme
pages that a student edits and the pages that the same student comments upon. Examining
190
this aspect can improve wiki recommendations by gauging the breadth of collaborative
activity that a particular student will participate in for a particular page. For example, it
can help determine whether the student will also contribute towards a discussion on the
page after the system successfully guides her to make an edit on it. While ratings and tags
can also be used in this analysis, we limit the focus to comments and edits due to their
| |
| |
That is, we calculate the proportion between the set of pages that the student both
edited and commented upon and the total set of pages edited or commented.
Intuitively speaking, if students have limited resources, they would aim to reduce
the total cognitive cost of doing the assignment by minimizing the number of unique
pages edited and discussed. That is, we expect the % Overlap score to be relatively high
are as follows:
12 - 51 0 0.000 0 0
Overlap within the class as a whole. However, as exemplified in Sections 5.2 and 5.3,
student behavior and motivations are not necessarily homogeneous. There is a possibility
that the user’s favored activity type (i.e. active or passive actions) may influence this
overlap in pages edited and pages commented upon. If we separate the users based on the
On
1 31, 36, 46, 45, 47, 48 2 0.333 2 0
47, 48
4 49, 50 49, 50, 51 2 0.667 2 0
5 31, 37, 38, 38, 46, 48, 1 0.125 1 0
43 49, 50
7 45, 48 34, 45, 58 2 0.667 2 0
8 31, 38, 41, 36, 41, 43, 2 0.286 1 1
43 45, 51
9 40 34, 38, 42, 47 0 0.000 0 0
10 34, 41, 42, 34, 41, 42 3 0.750 3 0
48
11 31, 45 36, 46, 50 0 0.000 0 0
15 46 36, 46, 51 1 0.333 0 1
Average 2.778 pages 3.556 pages 1.444 0.351 1.222 0.222
Table 5.32: RAIK unique pages commented/edited and the overlap between the two, for students placed in Cluster
0
5.5.1.1 Observations
When looking at the RAIK class as a whole, the overall % Overlap average is
0.324. However, dividing the students based on their activity type clusters results in
Comparing averages between the Active and Balanced profile students, the
average number of pages edited, number of pages commented, and the overlaps (both
number and percentage) between the two clusters are relatively comparable. However,
students with passive action-biased profiles have notably lower averages on all columns
Interestingly, for 85% (17 out of 20) of the overlaps, the user edits the wiki page
5.5.1.2 Justification:
Why would students with profiles favoring “active” actions have a greater % Overlap?
Students may find it easier to make their edit and comment contributions to the
same page since the “commitment cost” (e.g. time to read and familiarize page content,
commitment cost of commenting and editing two different pages. This “two-for-one” in
providing edits and comments to the same page can drive a student’s activity profile
194
towards the “Active” end of the active-passive actions spectrum, particularly if the
Why the preference for edits over comments? Why the preference for edit-first?
The RAIK class students appear to strongly prefer editing a wiki page before
The students prefer to focus on edits first, since they make up 60% of the
Immediately after reading the wiki page to become familiar with the topic, it may
thread.
5.5.1.3 Implications:
When a student favors active actions, the student’s collaboration initiation rate on a page
comment mechanism.
mechanism for wiki page discussion, we believe that the feature encourages particular
students to collaborate more on a single page. Specifically, we posit that students in this
class that have an “active” or “balanced” action profile are more likely to participate in or
begin a discussion on the page after making an edit on it than students preferring “passive”
195
actions. This implication follows from the previous justification for why there is a
Comparison to the MAS class results will be needed to further support this.
13 - - 0 0.000 0 0
196
when looking at the MAS class as a whole. We thus similarly separate the users based on
5.5.2.1 Observations:
When examining the MAS class as a whole, the overall % Overlap average is
0.438. However, dividing the students by activity type cluster results in average %
Overlaps of:
students with a profile biased towards “active” activities also have a notably higher
Unlike the RAIK class, the MAS class has a relatively greater balance between
edit-first and comment-first overlaps: only 61% of the overlaps (19 out of 31) stem from
5.5.2.2 Justification:
Why would students with profiles favoring “active” actions have a greater % Overlap?
The justification for this can be similarly explained with the one provided in sub-
chapter 5.5.1.2. That is, students may find it easier to make their edit and comment
contributions to the same page since the “commitment” or “cognitive” cost of doing so is
lower than the cost of commenting and editing two different pages.
Why preference for edits over comments? (Why the preference for edit-first?)
As with the RAIK class, the MAS class students generally edit wiki pages before
commenting on them when both actions are performed on the same page since the
majority of overlaps are edit-first. However, the proportion of overlaps where the edit
was performed before the discussion participation is only 61%, compared to the RAIK
class’s 85%. Since the grading for this class is similar to that of the other one, the
justification for this preference in sub-chapter 5.5.1.2 may also apply here.
5.5.2.3 Implications:
When a student favors active actions, the student’s collaboration initiation rate on a page
comment mechanism.
199
Similar to the RAIK class, the data suggests that a comment mechanism enables
students preferring “active” actions to have increased participation on the wiki pages they
contribute towards. However, it isn’t as clear whether the student comments because she
has edited the page or vice-versa, due to the 61%/39% split between edit-first and
Additional verification against more classes will be needed to further verify this
implication.
5.5.3 Summary
This section examined the possible correlation between the set of pages edited and
the set of pages commented upon for users. The biggest key finding across the results is
that students with “active” (or “balanced”) activity profiles tend to have a greater %
actions have a greater % Overlap between their pages edited and commented than
students favoring “passive” actions. Sub-chapter 5.5.1.2 and 5.5.2.2 justify this as arising
from the lower cognitive cost of editing and discussing on a single page (one topic) rather
than editing and discussing two separate pages (and two separate topics). This option is
discussions for three distinct pages, as discussed in sub-chapter 5.4. This finding is
students will likely contribute to both edits and discussion on the recommended
page.
200
An interesting difference observed between the two classes is that the MAS class
has a greater number (12 vs. 3) and proportion (39% vs. 15%) of comment-first overlaps,
compared to the RAIK class. Solely examining the grading criteria for the collaboration
phase for the two classes, there doesn’t seem to be an apparent difference between them
that would suggest such an effect. However, there is a notable difference in the individual
contribution phase: students in the MAS class are required to start three discussion
threads on their own pages. It is possible that these conversation “kick-starters” are
responsible for the increased comment-first overlaps observed in the MAS class by
longer have to take on the cost of deciding upon an appropriate topic for open-
ended discussion and creating the thread, if they choose not to. This is another
more-indirect way.
reasons varying from personal affinity and familiarity over the course of their education,
to reputation with regards to intelligence and work ethic. A “clique” is a group of such
students, preferring to collaborate with others within the clique than those outside it. For
a wiki setting in particular, students with a strong preference towards their cliques will
prefer to contribute towards (e.g. editing and commenting upon) pages that the other
In this section, we wish to examine whether editing cliques exist and play a role in
the collaboration initiation rate for the two classes. While the interpersonal relations
between users are not known or feasible to deduce from the tracked attributes, we can
look for the appearance of cliques regardless of their reasons for forming. We are
particularly interested in the presence of “strong” cliques – cliques that occur frequently
across multiple wiki pages. They can be leveraged in recommendation by: 1) notifying
cliques when one of its members participates on a wiki page (e.g., reinforcing cliques), or
2) biasing recommendations towards students outside of the target user’s cliques (e.g.,
weakening cliques).
steps:
1. Determine the unique contributors for each wiki page for the class.
2. Let i = 2.
a. Count the number of wiki pages where that particular i-student grouping
occurs.
clique.
4. Increment i and repeat Step 3 until there are no i-student groupings with more
With the above process alone, strong editing cliques would be defined solely by a
scenario: Student A has only edited three unique pages, and Student B has edited twelve
unique pages. Let us suppose that the number of mutually-edited pages between them is
three. Would the A-B grouping be considered a strong editing clique even though the
number of occurrences arises from Student B editing a relatively large number of unique
pages? Compare this to a scenario where Students A and C both edited three unique
pages and the three pages edited are the same for both of them.
To account for this possibility of larger clique occurrences due to students editing
a larger number of unique pages, we thus introduce the notion of “clique strength,” which
adjusts a clique’s occurrence count relative to the average number of unique pages edited
for each wiki page. Table 5.37 lists the IDs of the wiki pages along with the consenting
49 4, 5, 11
50 2, 4, 5, 11, 12, 13
51 4, 6, 8, 12, 14, 15
Table 5.38: Distinct contributors for each wiki page in the RAIK class
Using the above, we count the number of occurrences of each clique. Tables 5.38
and 5.39 list the cliques with more than two occurrences and their clique strengths. Note
that the tables are for cliques of size 2 and 3 – there are no cliques of larger sizes with
9, 10 2 4 0.500
11, 13 3 4.5 0.667
11, 15 2 4.5 0.444
14, 15 3 4.5 0.667
Table 5.39: RAIK clique occurrences and clique strengths for groups of size 2 with 2+ occurrences
5.6.1.1 Observations
The following was observed from the RAIK class clique analysis. The largest
editing cliques found on more than one wiki page is of size 3, with the largest number of
occurrences being exactly 2 for all such cliques. Surprisingly, there are a relatively large
number of cliques occurring on at least two pages: 35 for cliques of size 2, and 12 for
cliques of size 3. In spite of the numbers, few of these cliques are “strong” (i.e., have a
clique strength larger than 0.500). For size-2 cliques, 8 out of the 35 are considered
“strong” (0.229), and none of the size-3 cliques qualify with this criteria.
5.6.1.2 Justifications
Why are there a relatively large number of cliques with 2+ occurrences?
The relatively large number of cliques may arise from the following factors:
The number of pages available for the class to edit is approximately equal to the
The minimum number of unique pages edited for the assignment is 4: 1 from a
With relatively few total pages in the pool for the entire class, the probability of
clique occurrences on at least two pages is relatively larger than if there were more pages
in the pool.
Why are there few strong editing cliques relative to the many potential cliques?
It is possible that the current threshold for “strong” cliques may be set too high
relative to the average number of unique pages edited for the class.
5.6.1.3 Implications
Clique identification may have limited usefulness within the constraints of the particular
assignment specifications.
This may be particularly true for the RAIK class, due to: 1) the small number of
required unique pages for contributions (4), and 2) the small number of total editable
pages for the class (15). That is, although cliques can be identified as “strong” due to the
relative nature of its strength calculation, they may not be significant due to the scale of
the short-term assignment. For example, although a clique strength of 0.666 is relatively
“strong,” it can be achieved by two users having two mutually edited pages when each
have edited only three unique pages. Compare this to two users with 200 mutually edited
pages when each edited 300 unique pages throughout their membership on the wiki – the
206
users in the 200/300 scenario is considered to be “more” of a clique than the ones in the
2/3 scenario.
The strength formula will need to be revised to account for absolute number of
contributors for each wiki page. Table 5.40 lists the wiki pages that the MAS class’s
We then repeat the occurrence counts and clique strength calculations with this
data. Table 5.41 lists the cliques of size 2 with two or more occurrences. As can be
inferred from the lack of additional tables, there were no cliques of sizes 3 or larger that
5.6.2.1 Observations
The following were observed in the clique analysis results for the MAS class. The
largest clique size for the MAS class is smaller than that of the RAIK class at a size of 2.
Ten different size-2 cliques occurred on at least two wiki pages. This is relatively
interesting since there are considerably fewer cliques in spite of there being more
consenting students in the MAS class data than the RAIK class data (albeit a lower
percentage). Although there are relatively fewer such cliques identified, three of the ten
5.6.2.2 Justification
Why are there a smaller number of identified cliques for the MAS class?
This may arise from the following factors of the MAS class. First, there is a
larger number of students in the MAS class than the RAIK class (29 vs. 16).
208
Consequently, there are a larger number of pages available for students to contribute
towards, since each wiki page in the pool is created by one student in the class. Since the
contribution page + 3 different pages in collaboration phase) regardless of class size, this
results in the dilution of student participation across more pages. Combined with a lower
percentage of the students in the MAS class filling out the consent form for data usage,
this creates the appearance of the wiki pages having a sparse set of contributors, and
Why are there few strong editing cliques relative to the number of potential cliques?
As mentioned in the RAIK class POJI, the current threshold for “strong” cliques
may be set too high relative to the average number of unique pages edited.
5.6.2.3 Implications
A larger class size dilutes the efforts of the student body and limits the largest possible
although MAS has approximately the same number of consenting users, there are
considerably more cliques (i.e. more unique sets of students with more 2+ occurrences)
found in the RAIK class. Consent issues also contribute towards this. See sub-chapter for
Clique identification may have limited usefulness within the constraints of the particular
assignment specifications.
As with the RAIK class, using cliques may have limited usefulness within the
constraints of the particular assignment specifications for the MAS class, due to the small
number of required unique pages for contributions (4). See sub-chapter 5.6.1.3 for
5.6.3 Summary
This section covered the concept and analysis of editing cliques among the
bias results for or against clique members, depending on instructor goals. There were two
Possible clique sizes and frequency of clique occurrences are dependent on the number
of students in the class, the number of pages in the wiki, and the minimum number of
Due to the size differences of the classes, the RAIK and MAS classes highlight
the effects of varying the number of students in the class and the number of pages in the
wiki. Sections 5.6.1.2 and 5.6.2.2 discuss these in greater detail. Table 5.42 summarizes
While the last item in the table, minimum number of unique pages to edit, is not
explicitly covered in individual class results, we deduce the effects with the assumption
that students will strive to meet the minimum specified by the instructor. When the
minimum is increased, students will edit a greater portion of the page pool for the wiki.
Consequently, this can lead to more cliques being identified, particular cliques occurring
more frequently, and/or larger cliques in general. Conversely, decreasing the minimum
may result in reduced participation on the wiki, and consequently lead to fewer identified
It should be noted that it may be possible for the effects from altering multiple
factors simultaneously to result in a net offset of zero. For example, the specifications for
this particular assignment call for an increase in the number of wiki pages with an
increase in the number of students, since each are responsible for being the primary
contributor to their own unique page. We predict that clique identification between two
classes with different user and page populations but comparable student-page ratios will
be relatively similar.
211
occurrences, relative number of unique pages edited among clique members, absolute
number of pages edited among clique members, number of users and pages in wiki.
In the opening to this section (5.6), we defined a “strong” clique as a function of:
1) The number of unique pages that the editing clique occurred on (“clique
occurrences”), and…
2) The average number of unique pages edited by the members of the editing clique
The latter condition was introduced to separate users who simply have a large
editing page set from those who deliberately seek out particular users to collaborate with.
In sections 5.6.1.3 and 5.6.2.3, we questioned whether this purely relative strength
There may be other factors that need to be considered for the strength calculation.
First, the absolute number of pages edited may need to be considered, as the 2/3 vs.
200/300 example in section 5.6.1.3 highlighted. Second, the number of users and pages in
the class/wiki may also be a factor in clique strength. Referring back to Table 5.42, there
are scenarios where it would be harder to form cliques. It would then logically follow that
cliques formed in such an environment may possibly be stronger than those formed in
more-ideal environments.
from students in two separate classes carrying out a collaborative writing assignment
212
within the Biofinity Intelligent Wiki. Table 5.43 lists the specific user actions that we
focus our analyses upon, along with the specific attributes pertaining to each.
Action Attribute(s)
Edit Number/count, word length, quality
Comment/Discussion Number/count, word length, quality
Rate Number/count
Tag Number/count
Table 5.44: User actions and associated attributes focused upon in post-hoc analyses
In our first analysis, we clustered the students in each class based on these “raw”
attributes by using our X-means based “Maximal Pairs” clustering algorithm (detailed in
5.2.1). In the following analyses, we clustered the students upon two composite metrics:
active vs. passive profile percentage (5.3) and minimalist vs. overachiever score (5.4).
The former calculates the weighted proportion of edits and comments over the total
actions performed by a user, whereas the latter calculates a user’s overall activity level
relative to peers in the class. Finally, we analyze the users with respect to overlaps in
their editing and commenting page sets (5.5) and possible editing cliques between them
(5.6).
Each analysis provided us with a series of insights regarding our metrics, user
behavior, and instructor influence within the wiki. Table 5.44 lists the highlights of each
sub-chapter, along with its analogous equivalent for virtual collaboration outside of the
5.4 The optimal number of clusters: 3 vs. k. The optimal number of clusters: 3 vs. k.
For recommendation, all three outcomes For recommendation, all three outcomes
when k=2 have their flaws. By forcing the when k=2 have their flaws. By forcing the
optimal number of clusters to 3, it is optimal number of clusters to 3, it is
possible for the cluster the students into possible for the cluster the users into a
a group of extreme overachievers, a group of extreme overachievers, a group
group of extreme minimalists, and a of extreme minimalists, and a group of
group of the remaining students in the remaining students in between.
between.
5.5 Students with “active” (or “balanced”) Users with “active” (or “balanced”)
activity profiles tend to have a greater % activity profiles tend to have a greater %
Overlap than those with “passive” activity Overlap than those with “passive” activity
profiles. The value of a successful profiles. The value of a successful
recommendation to students favoring recommendation to users favoring
“active” actions is increased, since such “active” actions is greater than those
students will likely contribute to both favoring “passive” actions, since such
edits and discussion on the users will likely contribute to both edits
recommended page. and discussion on recommended pages.
5.5 It is possible that these conversation Conversation “kick-starters” can increase
“kick-starters” are responsible for the comment-first overlaps by reducing the
increased comment-first overlaps cognitive cost of participating in a
observed in the MAS class by reducing discussion. The cost of starting a
the cost of participating in a discussion. discussion (i.e., deciding upon an
That is, potential contributors no longer appropriate topic for open-ended
have to take on the cost of deciding upon discussion) is now optional for potential
an appropriate topic for open-ended contributors.
discussion and creating the thread, if they
choose not to.
5.6 Possible clique sizes and frequency of Possible clique sizes and frequency of
clique occurrences are dependent on the clique occurrences are dependent on the
number of students in the class, the number of users on the wiki, the number
number of pages in the wiki, and the of pages in the wiki, and the minimum
minimum number of unique pages number of unique pages
edited/commented required by the edited/commented requested by the wiki
assignment. moderator.
5.6 Relevant factors in the identification of (Same as column 1)
“strong” cliques include: number of
clique occurrences, relative number of
unique pages edited among clique
members, absolute number of pages
edited among clique members, number
of users and pages in wiki.
Table 5.45: Primary findings within the various Chapter 5 analyses
215
Chapter 6: Conclusion
In this chapter, we summarize the purpose of the thesis in Chapter 6.1, briefly
cover the accomplishments and results found in Chapter 6.2, and close out the thesis by
collaboration and the benefits of improving it (Chapter 1). We focused our scope on
widespread use to this day. While multiple tools have been developed to support and
improve collaboration in such a setting (e.g., Annoki platform by Tansey and Stroulia and
the Socs application by Atzenbeck and Hicks), there were relatively few intelligent ones
beyond the SuggestBot of Cosley et al. (Chapter 2). We thus identified recommendation-
based intelligent support for improving wiki collaboration as the target niche for our
work.
The primary goal of the thesis is to investigate and provide the foundation for
contributions our work provides include: 1) wiki-based user and data models and a
implementation of an intelligent wiki that allows for the addition of social and intelligent
features, and 3) insights to user behavior, moderator influence, and model efficacy that
the design and implementation of our own intelligent wiki. Rather than modifying an
existing wiki, we implemented the majority of it from the ground up using common
technologies including Java, Javascript, HTML, Glassfish, and the Google Web Toolkit.
In addition to providing us with a great deal of flexibility for future development and
design the wiki to include social/Web 2.0 features (e.g., page tagging, page ratings, intra-
wiki and social network sharing, thread-based discussions) and intelligent features (e.g.,
page and user modeling, user tracking via an agent-based framework, recommendation
framework) from the start. For additional implementation details, please refer back to
Chapter 4.
Another major achievement is the insights discovered in the analysis of the wiki
usage data which span topics such as user behavior and contribution strategies, the degree
that moderator influence and guidelines affect users, relations between tracked attributes,
and the applicability and relevance of our new metrics. Our investigation was carried out
in the following manner: 1) deploy the Biofinity Intelligent Wiki to multiple settings, 2)
collect usage data as the various users in each setting carry out their tasks on it, and 3)
perform post-hoc analysis on the data once a setting-specific milestone is reached or the
purposes for using the wikis were fulfilled. Ultimately, the two data sets used in this
thesis originated from a classroom setting where the wiki is used for a collaborative
writing assignment. Our analyses of the usage data provided us with the following
generalized findings:
217
expectations and guidelines, the “quantity vs. quality” tradeoff, and the relative
behavior with rules and other expectations, user behavior can also be influenced
A positive correlation exists between the number of total active actions and total
Tradeoff (or lack thereof) between the total number of unweighted collaborative
actions and % active actions. Although total active actions and total passive
Count vs. Length, Count vs. Quality, and Length vs. Quality are three distinct
and edits.
The optimal number of clusters: 3 vs. k. For recommendation, all three outcomes
when k=2 have their flaws. By forcing the optimal number of clusters to 3, it is
possible for the cluster the users into a group of extreme overachievers, a group of
Users with “active” (or “balanced”) activity profiles tend to have a greater %
Overlap than those with “passive” activity profiles. The value of a successful
“passive” actions, since such users will likely contribute to both edits and
Possible clique sizes and frequency of clique occurrences are dependent on the
number of users on the wiki, the number of pages in the wiki, and the minimum
absolute number of pages edited among clique members, number of users and
pages in wiki.
219
With the work presented in this thesis – a proposed user and data model, a proposed
recommendation algorithm, and insights to user behavior and moderator influence over it
intelligent support. However, this is only a “first step”: there is still much work to be
done such as revising the model and recommendation algorithm per our findings and
deploying and evaluating the wiki under different scenarios. Chapter 6.3 delves more into
performed on a relatively small number of data sets and users. Specifically, we only had
permission to examine 32 students across two classes. It is not surprising that we are in
need of additional data sets to further support (or possibly refute) our findings from the
assignments structured and graded differently from the assignments used for the MAS
and RAIK classes, and 2) wiki usage for non-classroom purposes, i.e., a user-base
motivated by the group’s goals rather than a grade. With the first data set, we can confirm
the generality of the RAIK and MAS class implications while still constrained to a
classroom setting. With the second, we can confirm the generality of the findings to the
while this approach facilitates the discovery of insights and patterns, our findings do not
take into account the various obstacles and challenges that arise during the live use of the
user and data models. For example, the “Cold Start” problem may be of particular
weeks long, and the time allotted for the collaborative writing assignment will
Due to the above, students typically will not have an extensive activity history
data from previous classes/assignments. However, this will only serve to reduce the
amount of information that is needed from the student and will not eliminate the need for
it entirely. Additional investigation and preparation will be needed before live usage of
that may be valuable to add to the Biofinity Intelligent Wiki. Possibilities include:
extreme “active” or “passive” action profiles are identified. Users can also
References
Adomavicius G., Tuzhilin A. (2005). Toward the Next Generation of Recommender
Systems: a Survey of the State-of-the-Art and Possible Extensions. In IEEE Transactions
on Knowledge and Data Engineering, Volume 17, Number 6, pp. 734-749.
Anderson, T., Gunawardena, C., Lowe, C. (1997). Analysis Of A Global Online Debate
And The Development Of An Interaction Analysis Model For Examining Social
Construction Of Knowledge In Computer Conferencing. Journal of Educational
Computing Research, 397-431.
Atzenbeck C., Hicks D. (2008). Socs: increasing social and group awareness for Wikis by
example of Wikipedia. In Proceedings of the 4th International Symposium on Wikis, pp.
9:1-9:10.
Chen J., Geyer W., Dugan C., Muller M., Guy I. (2009). "Make New Friends, but Keep
the Old" – Recommending People on Social Networking Sites. In Proceedings of the 27th
International Conference on Human Factors in Computing Systems, pp. 201-210.
Chen J.R., Wolfe S.R., Wragg S.D. (2000). A Distributed Multi-Agent System for
Collaborative Information Management and Sharing. In Proceedings of the 9th
International Conference on Information and Knowledge Management, pp.382-388.
Chen J., Geyer W., Dugan C., Muller M., Guy I. (2009)."Make New Friends, but Keep
the Old" - Recommending People on Social Networking Sites. Proceedings of the 27th
International Conference on Human Factors in Computing Systems, (pp. 201-210).
Cosley, D., Frankowski, D., Terveen, L., Riedl, J. (2007). SuggestBot: Using Intelligent
Task Routing to Help People Find Work in Wikipedia. Proceedings of the 12th
International Conference on Intelligent User Interfaces, (pp. 32-41).
Guy I., Ronen I., Wilcox E. (2009). Do You Know? Recommending People to Invite into
Your Social Network. In Proceedings of the 13th International Conference on Intelligent
User Interfaces, pp. 77-86.
Hara N., Solomon P., Sonnenwald D.H., Kim S-L. (2003). An Emerging View of
Scientific Collaboration: Scientists' Perspectives on Collaboration and Factors that
Impact Collaboration. In Journal of the American Society for Information Science and
Technology, Volume 54, Issue 10, pp. 952-965.
Hu M., Lim E-P., Sun A., Lauw H., Vuong B-Q. (2007). Measuring article quality in
Wikipedia: models and evaluation. In Proceedings of the 16th ACM Conference on
Information and Knowledge Management, pp. 243-252.
Jermann P., Soller A., Muehlenbrock M. (2001). From Mirroring to Guiding: a Review of
State of the Art Technology for Supporting Collaborative Learning. In Proceedings of the
First European Conference on Computer-Supported Collaborative Learning, pp. 324-331.
Karoly L., Panis C. (2005). The 21st Century at Work: Forces Shaping the Future
Workforce and Workplace in the United States.
Katz J.S., Martin B.R. (1997). What is Research Collaboration? In Research Policy,
Volume 26, Issue 1, pp. 1-18.
Kester L., van Rosmalen P., Sloep P., Brouns F., Koné M., Koper R. (2007).
Matchmaking in Learning Networks: Bringing Learners Together for Knowledge Sharing.
In Interactive Learning Environments, pp. 117-126.
Khosravifar B., Bentahar J., Moazin A., Thiran P. (2010). On the Reputation of Agent-
Based Web Services. In Proceedings of the 24th AAAI Conference on Artificial
Intelligence, pp. 1352-1357.
224
Konstan J., Miller B., Maltz D., Herlocker J., Gordon L., Riedl J. (1997). GroupLens:
Applying Collaborative Filtering to Usenet News. Communications of the ACM, 40(3),
pp. 77-87.
Kraut R., Egido C., Galegher J. (1988). Patterns of Contact and Communication in
Scientific Research Collaboration. In Proceedings of the 1988 ACM Conference on
Computer-Supported Cooperative Work, pp. 1-12.
Limerick D., Cunningham B. (1993). Collaborative Individualism and the End of the
Corporate Citizen.
Liu G., Wang Y., Orgun M. (2010). Optimal Social Trust Path in Complex Social
Networks. In Proceedings of the 24th AAAI Conference on Artificial Intelligence, pp.
1391-1398.
Martin F., Donaldson J., Ashenfelter A., Torrens M., Hangarter R. (2011). The Big
Promise of Recommender Systems. In AI Magazine, Volume 32, Number 3, pp. 19-27.
Meier A., Spada H., Rummel N. (2007). A rating scheme for assessing the quality of
computer-supported collaboration processes. In International Journal of Computer-
Supported Collaborative Learning, Volume 2, Number 1, pp. 63.68.
225
O’Reilly T. (2004). What is Web 2.0: Design patterns and business models for the next
generation of software.
Paek T., Gamon M., Counts S., Chickering D.M., Dhesi A. (2010). Predicting the
Importance of Newsfeed Posts and Social Network Friends. In Proceedings of the 24th
AAAI Conference on Artificial Intelligence, pp. 1419-1424.
Sabater J., Sierra C. (2001). Regret: a Reputation Model for Gregarious Societies. In
Proceedings of the Fourth Workshop on Deception, Fraud, and Trust in Agent Societies,
pp. 61-69.
Sabater J., Sierra C. (2002). Reputation and Social Network Analysis in Multi-agent
Systems. In Proceedings of the First International Joint Conference on Autonomous
Agents and Multiagent Systems: Part 1, pp. 475-482.
Scott S., Henninger S., Jameson M.L., Moriyama E., Soh L-K., Harris S., Ocampo F.,
Simpson R. (2008). NSF Proposal for the Semantic Cyberinfrastructure for Information
Discovery.
Shepitsen A., Gemmell J., Mobasher B., Burke R. (2008). Personalized Recommendation
in Social Tagging Systems using Hierarchical Clustering. Proceedings of the 2008 ACM
Conference on Recommender Systems, pp. 259-266.
226
Song M., Lee W., Kim J. (2010). Extraction and Visualization of Implicit Social
Relations on Social Networking Sites. In Proceedings of the 24th AAAI Conference on
Artificial Intelligence, pp. 1425-1430.
Suh B., Chi E., Kittur A., Pendleton B. (2008). Lifting the Veil: Improving
Accountability and Social Transparency in Wikipedia with WikiDashboard. In
Proceeding of the 26th Annual SIGCHI Conference on Human Factors in Computing
Systems, pp. 1037-1040.
Vassileva J., McCalla G., Greer J. (2003). Multi-Agent Multi-User Modeling in I-Help.
User Modeling and User-Adapted Interaction, 13(1-2), pp. 179-210.
Vivacqua A., Moreno M., de Souza J. (2003). Profiling and Matchmaking Strategies in
Support of Opportunistic Collaboration. In Lecture Notes in Computer Science, Volume
2888, pp. 162-177.
Vivacqua A., Moreno M., de Souza J. (2006). Using Agents to Detect Opportunities for
Collaboration. In Lecture Notes in Computer Science, Volume 3865, pp. 244-253.
Wang Y., Vassileva J. (). Trust and Reputation Model in Peer-to-Peer Networks. In
Proceedings of the 3rd International Conference on Peer-to-Peer Computing, pp. 150-157.
Warkentin M., Sayeed L., Hightower R. (1997). Virtual Teams versus Face-to-Face
Teams: An Exploratory Study of a Web-based Conference System. In Decision Sciences,
Volume 28, Number 4, pp. 975-996.
∑
𝑊𝐻𝑀( ) where:
∑ ∑
j is the index of the time period evaluated. The corresponding weights and period
j Period
1 within six months from the current timestamp 0.40
2 between six to 12 months from the current timestamp 0.30
3 between 12 to 18 months from the current timestamp 0.20
4 older than 18 months 0.10
Table A.1: Time periods and corresponding weights used in Weighted Harmonic Mean calculations
( ) ∑ ∑ ( )
Where:
is the set of all pages that the user u has interacted with
The weights for each of the user actions are preliminarily assigned in Table A.2 (below),
Action
Create 2.25
Edit 2.25
Discuss 2.0
Recommend 1.5
Rate Up 1
View 1
Table A.2: Weighted counts used for calculating user interests.
229
to the moment the target number of clusters is reached. That is, the pseudocode is now:
When multiple cluster merges are available for the iteration, the one with the larger
The new memberships for the four-cluster results of the RAIK class are as follows:
Cluster 0 0, 2, 3, 8, 11, 14
Cluster 1 1
Cluster 2 4, 6, 10, 13
Cluster 3 5, 7, 9, 12
The following tables list the attributes for each cluster member:
Cluster 0
Dist.
From
Student RaikNumEdits RaikEditLength NumCmts CmtLength NumTags NumRates Centroid
0 4 408.250 6 108.333 0 11 211.377
2 3 285.000 1 113.000 0 18 93.736
3 6 256.667 2 99.500 3 7 62.113
8 5 48.200 1 52.000 0 5 153.172
11 3 57.667 0 0.000 0 16 160.856
230
Cluster 1
Dist.
From
Student RaikNumEdits RaikEditLength NumCmts CmtLength NumTags NumRates Centroid
1 5 106.400 4 33.000 34 16 0
Avg - - - - - - -
Std Dev - - - - - - -
Table B.3: Attribute details for RAIK cluster 1
Cluster 2
Dist.
From
Student RaikNumEdits RaikEditLength NumCmts CmtLength NumTags NumRates Centroid
4 7 83.714 4 69.250 10 6 13.625
6 9 54.222 2 73.000 0 17 21.351
10 9 40.667 2 54.500 5 12 35.041
13 5 115.400 3 69.667 3 15 42.187
Avg 7.500 73.501 2.750 66.604 4.500 12.500 28.051
Std Dev 1.915 33.214 0.957 8.242 4.203 4.796 -
Table B.4: Attribute details for RAIK cluster 2
Cluster 3
Dist.
From
Student RaikNumEdits RaikEditLength NumCmts CmtLength NumTags NumRates Centroid
5 8 66.25 6 86.5 16 29 13.687
7 11 76.73 5 79.6 8 17 19.305
9 11 41.73 4 70 11 8 27.571
12 13 57.62 4 114 25 14 28.711
Avg 10.750 60.580 4.750 87.525 15.000 17.000 22.318
Std Dev 2.062 14.800 0.957 18.902 7.439 8.832 -
Table B.5: Attribute details for RAIK cluster 3
231
Since the cluster assignments are fairly similar to that of the three-cluster results,
B.1 Observations
Regarding impressions from the results, the singleton cluster from the three-
cluster results persists in these four-cluster results. For additional discussion pertaining to
this cluster, refer back to the previous section. Particularly notable is the cluster “split” –
that is, the New Clusters 2 and 3 appear to be a “split” of one of the clusters from the “old”
three-cluster result. In other words, they are the two clusters that comprise Old Cluster 2.
while the intervals for each attribute of New Clusters 2 and 3 overlap with each other
within one standard deviation from their means, the new clusters seem to split some
Number of edits - Cluster 2 comprises the lower half of the range covered by the
Number of comments - Cluster 2 comprises the lower half of the range covered by
Comment length - Cluster 2 comprises the lower half of the range covered by the
Number of tags - Cluster 2 comprises the lower half of the range covered by the
Our procedure can be seen as a hierarchical clusterer, with the metric for merging
clusters being the presence of strong maximal pairs. That is, at some particular point in
the original algorithm execution, the number of clusters goes from four to three. Thus, to
go from three clusters to four, execution is “halted” when four clusters are remaining, and
one of the “old” clusters would appear to be “split” into two separate ones.
Why was the “Old” Cluster 2, i.e., Cluster 2 from the three-cluster results, “split”
Old Cluster 1 is a singleton cluster and thus is not eligible to be “split”. Also, Old
Cluster 0 had a “stronger” maximal pair, i.e. larger co-occurrence count with their
maximal pairs, between the clusters to be merged than the one between new clusters 2
and 3. It thus had higher “priority” for merging over these two.
For this particular data, the four-cluster results do not seem to significantly differ
from the three-cluster results. This appearance of having little benefit may arise from:
1. The discrepancy between the target number of clusters and the actual number of
clusters in the results being only one short, rather than being much smaller.
233
2. The relatively small standard deviations and relative lack of outliers in the “old”
3. The “new” clusters being a relatively “even” split that do not highlight/isolate
outliers.