Upload
dinhanh
View
219
Download
2
Embed Size (px)
Citation preview
1
Online Resource 1 for:
The Effect of Contextualized Conversational
Feedback in a Complex Open-Ended Learning
Environment
Published in: Educational Technology Research & Development
Corresponding Author:
James R. Segedy
Institute for Software Integrated Systems, Vanderbilt University
PMB351829
Nashville, TN 37235
Phone: +1-615-417-6551
Fax: +1-615-343-7440
Email: [email protected]
John S. Kinnebrew
Institute for Software Integrated Systems, Vanderbilt University
PMB351829
Nashville, TN 37235
Email: [email protected]
Gautam Biswas
Institute for Software Integrated Systems, Vanderbilt University
PMB351829
Nashville, TN 37235
Email: [email protected]
2
Analyzing Learning Behaviors and Strategies
Introduction
This document serves as a comprehensive description and explanation of a
data mining methodology for assessing and comparing student learning behaviors,
which includes three major components: (1) action abstraction while maintaining
context using a relevance metric; (2) generation and comparison of hidden
Markov models to analyze aggregated student learning behaviors; and (3)
combining sequential pattern mining and episode discovery/mining to identify
differences in specific learning behavior patterns. Additionally, this document
demonstrates the usefulness of these techniques by analyzing a set of data from
the Betty’s Brain computer-based learning environment (Leelawong and Biswas
2008).
The traditional approach to measuring students’ SRL thinking has been
through the use of self-report questionnaires (e.g., Pintrich, Marx, and Boyle
1993; Weinstein, Schulte, & Palmer 1987; Zimmerman and Pons 1986). The
underlying assumption in these questionnaires is that self-regulation is an aptitude
that students possess. For example, the questionnaire items might attempt to
assess students’ inclination to elaborate as they read a passage, or to determine
their approach to managing available time resources (Perry and Winne 2006;
Zimmerman 2008). This approach has been useful, as the self-report
questionnaires have been shown to be good predictors of students’ standard
achievement test scores, and they correlate well with achievement levels (Pintrich
et al. 1993; Zimmerman and Pons 1986). However, Hadwin and others (Azevedo
and Witherspoon 2009; Hadwin, Nesbit, Jamieson-Noel, Code, and Winne 2007;
Hadwin, Winne, Stockley, Nesbit, and Woszczyna 2001; Perry and Winne 2006)
have argued that while the questionnaires provide valuable information about the
learners’ self-perceptions, they fail to capture the dynamic and adaptive nature of
SRL as students are involved in learning, knowledge-building, and problem-
solving tasks.
Recently, studies have illustrated the need to supplement self-report
questionnaires with measures of SRL as dynamic processes (Hadwin et al. 2007;
Winne and Jamieson-Noel 2002). For example, Hadwin et al. (2007) performed a
study that collected activity traces of eight students using the gStudy system
(Perry and Winne 2006). The activity traces were analyzed in four different ways:
3
(i) frequency of studying events, (ii) patterns of studying activity, (iii) timing and
sequencing of events, and (iv) content analyses of students’ notes and summaries.
The results of this analysis were compared against students’ self-reports on their
self-regulated learning. One of the important findings was that many participants’
self-reports of studying tactics were not well-calibrated with studying events
traced in the gStudy system. The researchers found that the best matched item
showed only a 40% agreement, and the average agreement was 27%. The authors
concluded from this study that direct analyses of student activity traces in e-
learning environments are important for furthering our understanding of both how
students apply SRL, and how that application affects their learning performance.
Increasingly, researchers have begun to utilize trace methodologies in
order to examine the complex temporal patterns of self-regulated learning
(Aleven, McLaren, Roll and Koedinger 2006; Azevedo and Witherspoon 2009;
Azevedo, Witherspoon, Chauncey, Burkett, and Fike 2009; Biswas et al. 2010;
Hadwin et al. 2007; Jeong and Biswas 2008; Zimmerman 2008). Underlying these
approaches is a move away from assessing self-regulation as an intrinsic aptitude
and, instead, assessing it as dynamic and adaptive event occurrences. By
identifying and analyzing temporal event sequences from trace data (e.g., student
actions in a computer-based learning environment), we hope to better understand
student’s self-regulated learning strategies and develop online measurement
schemes to provide more adaptive feedback and scaffolding.
Action Abstraction with Context
Like many other computer-based learning environments, Betty’s Brain
produces extensive logs of the actions performed by students using the system. To
effectively perform sequential data mining/analysis, raw logs of learning
interactions must first be transformed into an appropriate sequence of actions by
raising the level of abstraction from raw log events to a canonical set of distinct
actions. In this step, researcher-identified categories of actions define an initial
alphabet (set of action symbols) for the sequences. This also allows us to filter out
irrelevant information (e.g., cursor position) and to combine qualitatively similar
actions (e.g., querying an agent through different interfaces or about different
concepts in a given topic). Further, this categorization focuses the data mining
4
analyses on researcher-directed distinctions and features in the student learning
activities.
In Betty’s Brain, we categorize student activities as five, qualitatively
distinct actions:
(1) READ: students access one of the pages in the resources;
(2) EDIT: students add, delete, or modify a concept or link in the map;
(3) QUER: students ask Betty a question;
(4) EXPL: students ask Betty to explain her answer to a query; and
(5) QUIZ: students tell Betty to take a quiz.
Sometimes abstracting the raw log traces through action categorization
also strips important context from the actions in the trace sequence. For example,
while differentiating the addition of every possible link in a concept map would
result in an unwieldy set of distinct actions, knowing that the link added is the
same as one described in the last read action or in a following query can be
important context information. To maintain a balance between minimizing the
number of distinct actions in the sequences and keeping relevant context
information, we employ a metric indicating the relevance of each action to recent,
previous actions. Based on this relevance metric we split each categorized action
into two distinct actions: (1) high relevance and (2) low relevance to recent
actions. High relevance actions were those whose score exceeded the mean for
that category of action in the full set of student activity traces, while low
relevance actions had scores less than the mean (Biswas et al. 2010).
For the current study, each student action was assigned a relevance score
that depended on the number of relevant previous actions within a 3-action
window. For Betty’s Brain, a prior action is considered relevant to the current
action if it is related to, or operates on, one of the same map concepts or links.
This score provides a measure of informedness for map building/refinement
actions and, similarly, a measure of diagnosticity for map monitoring activities
(Biswas et al. 2010). Overall, the relevance scoring maintains some action context
by providing a rough measure of strategy focus or consistency over a sequence of
actions. For example, if a student edits a link that is part of a causal chain used to
answer a recent query, the query action is counted in the edit’s relevance score.
The increased relevance score suggests a more informed edit because it was
related to a recent query. Similarly, if a student asks a query where the answer
5
uses a recently-edited link, the edit action is counted toward the relevance score of
the query, implying the student’s use of the query was diagnostic of the edited
link.
Modeling Learning Behaviors with HMMs
The student activities logged by a learning environment result from a
variety of internal cognitive/metacognitive states, strategies, and processes used
by the student. Analyses that do not take into account the sequential nature of
student interactions with the system (e.g. action frequency counts) capture limited,
indirect information about students’ learning strategies and cognitive states.
Employing a direct representation of these internal states and strategies with a
probabilistic automaton, such as a hidden Markov model (HMM) (Rabiner 1989),
has the potential for facilitating identification, interpretation, and comparison of
student learning behaviors (Biswas et al. 2010). In particular, HMMs can provide
a concise representation of student learning strategies and behaviors (e.g.,
strategies and their relationships, as opposed to simple action or sequence
frequencies) derived from learning activity sequences (Jeong and Biswas 2008).
HMMs include a set of states and probabilistic transitions (i.e., more likely
transitions are assigned higher probabilities) between those states.
Like a student’s mental processes, the states of an HMM are hidden, meaning that
they cannot be directly observed, but they produce observable emissions (e.g.,
actions in a learning environment). Another important aspect of an HMM is the
likelihood of starting in a given state, which can be important information for
understanding learning behaviors and strategies. Together, three sets of
probabilities form a complete model: (1) transition probabilities, which determine
the likelihood of going from one state to another at each step; (2) state output
probabilities, which define the likelihood of observing different outputs in each
state; and (3) state initial probabilities, which define the likelihood that a state will
be the starting state for an output sequence.
In recent years, a number of researchers have had success in applying
HMM generation techniques to learning interaction data. For example, Pardos
and Heffernan (2011) have successfully applied HMMs (in conjunction with a
bagged decision tree classifier) to predict correctness of student answers in
computer math learning environments. In contrast to research using HMMs for
6
action prediction, our exploratory methodology uses HMM generation to identify
learning behaviors of student groups and to compare behaviors across those
groups. Boyer et al. (2009) have used HMMs for identifying strategies and high-
level flow in student dialogues with human tutors. Cohen and Beal (2009) have
used HMMs to analyze dialogues with an intelligent tutoring system to identify
different states of engagement. Some of the more closely related work with
HMMs for identification of student learning strategies is the HMM clustering
technique employed by Shih, Koedinger, and Scheines (2010). Performance and
learning metrics are used to direct the generation of HMM collections that predict
learning outcomes. In contrast, our technique generates HMMs from a set of
student activity sequences to build a model that represents an aggregate view of
each group’s learning behaviors. The technical details of our approach to
generating HMMs from a set of student activity sequences are presented in
(Biswas et al. 2010; Jeong and Biswas 2008).
The generated HMM for a set of student activity traces provides an
overview of aggregate learning behaviors and strategies in the group of students.
We compare the HMMs between groups based on presence or absence of
identified learning behaviors, as well as differences in their proportion of expected
behavior occurrences. In many cases, the learning behavior patterns and strategies
identified by comparison of HMMs can illustrate marked differences between
groups, such as experimental versus control conditions or high versus low
performers (e.g., Biswas et al. 2010; Jeong and Biswas 2008).
Comparative Sequence Mining
While the HMM analysis allows for effective interpretation of common
learning behaviors, states, and strategies in a student group, it may not be
sufficient for a detailed understanding of differences in student learning behavior.
By aggregating various less common behaviors and specific variations of general
strategies into one state, valuable information about the order of actions within a
state can be overlooked by this level of analysis/modeling. For example, an HMM
state that includes map editing, querying, explanation, and quizzing actions (with
a high self-loop probability) would only indicate what proportion of each activity
was used, but not the order in which those actions occur. Thus, it is impossible to
determine the precise strategy (or strategies) indicated by the state. Perhaps a
7
student was using regular queries to check Betty’s knowledge during teaching and
periodically assessing her overall progress using a quiz. Another equally plausible
scenario involves the student periodically administering a quiz during teaching
and then using queries and explanations based on quiz questions to isolate
incorrect information in the map.
This issue becomes even more pronounced when trying to compare across
groups of students. When the HMMs generated for two groups contain similar
states, it is still possible that those states represent different, although often
related, learning strategies. To provide a finer-grained, complementary level of
analysis, our methodology employs a novel combination of sequence mining
techniques (Kinnebrew & Biswas 2011).
An important part of this extended analysis is the use of sequential pattern
mining (Agrawal and Srikant 1995) to identify frequent1 patterns of actions within
a group. However, this often results in a very large number of frequent patterns
that can be difficult to analyze effectively. In general, a major challenge in
effective analysis of frequent patterns is limiting the often large set of results to
interesting or important patterns (i.e., the “effectiveness” of a mining technique as
opposed to its “efficiency” in finding frequent patterns (Agrawal and Imielinski
1993)).
To find these important patterns in a comparative analysis between student
groups, our methodology combines sequential pattern mining with episode
discovery/mining to identify differentially frequent patterns. In contrast to
sequential pattern mining, episode mining (Mannila, Toivonen, and Verkamo
1997) define frequency as the number of times the pattern can be matched within
a single sequence of actions. Combining sequential pattern mining and episode
mining, our methodology focuses on learning activity patterns that are used
differentially between two groups of students.
Although generally considered distinct subfields of sequence mining, there
has been some recent work in combining techniques and concepts from sequential
pattern mining and episode mining. For example, Ding, Lo, Han, and Khoo
(2009) employ episode mining over a set of sequences, rather than a single
sequence. However, our methodology separately employs sequential pattern
1 Pattern frequency in sequential pattern mining is defined as the number of different sequences (in
this case the number of different student interaction traces) that include the pattern.
8
mining to first identify common patterns within each comparison group. Lo,
Khoo, and Liu (2008) employ multiple frequency thresholds and techniques (from
sequential pattern mining and episode mining) in a combined algorithm for the
discovery of recurrent rules in a single set of sequences.
In our analysis of differentially frequent patterns, the sequential pattern
mining frequency measure (i.e., how many students exhibit the given pattern) is
important for identifying patterns common to a group (Kinnebrew & Biswas
2011). We refer to this as the “sequence support” (s-support) of the pattern,
following the convention of Lo et al. (2008), and we call patterns meeting a given
s-support threshold s-frequent. The other frequency metric, employed for
comparing patterns across groups (Kinnebrew & Biswas 2011), is the episode
frequency (i.e., the frequency with which the pattern is repeated within an
interaction trace). For a given trace, we refer to this as the “instance support” (i-
support), following Lo et al. (2008), and we call patterns meeting a given i-
support threshold i-frequent. Further, we allow a maximum gap of one action
between consecutive actions in the patterns2 to effectively identify frequent
learning interaction patterns despite the noisy nature of learning activities
(Kinnebrew & Biswas 2011).
To compare the identified frequent patterns across groups, we calculate the
difference in group i-support frequencies of each pattern. This focuses the
comparison on patterns that are employed significantly more often by one group,
and it produces four distinct categories of frequent patterns. The two categories
where the patterns are s-frequent in only one group illustrate patterns primarily
employed by that group but not the other, and the two categories where the
patterns are common to both groups are differentiated by which group uses the
pattern more often (i.e., higher i-frequency).
Example Analysis
To demonstrate the effectiveness of these techniques, this section presents
an analysis of data from a recent classroom experimental study that tested the
effect of contextualized conversational (CC) feedback in Betty’s Brain
(Leelawong and Biswas 2008) by comparing it to a baseline feedback approach
called prompt-based action-oriented (PA) feedback. The research hypothesis was
2 A maximum gap of one means that between each consecutive pair of actions in a frequent
pattern, the mining algorithm allows up to one additional action that is not part of the given
pattern.
9
that the conversational and contextual nature of the CC feedback helps students
better understand the relevance and context of the feedback. Thus, students
receiving CC feedback would more often take actions in accordance with the
feedback, and therefore would better integrate reading, causal map editing, and
assessing activities.
To identify differences in learning behaviors between the two conditions,
we analyzed learning activity traces for students in the PA and CC groups by
employing the HMM and comparative sequence mining methods described
previously. The resulting group HMMs contain five distinct states that are
described by the state-action emission probabilities in Figure 1. The value shown
for an action in a given state is the probability of that action being produced in the
state. When an action ends in -L or -H, it indicates low or high relevance,
respectively, of that action to previous actions. Using the interpretation methods
detailed in (Biswas et al. 2010), we interpreted these states as representing:
Reading - students are primarily reading the resources.
Informed editing - students are primarily making high-relevance edits,
suggesting a more focused map-building effort.
Uninformed editing - students are primarily making low-relevance edits,
possibly indicating the use of guessing behaviors.
Uninformed editing and checking - students are performing assessment
behaviors like querying and quizzing to check the correctness of their
causal maps, but are also making a significant number of low-relevance
edits to their maps. This implies that students are not reflecting on the
results of their assessments, and they may have resorted to trial-and-error
methods to correct their maps.
Informed editing and checking - students are making high-relevance
changes to their map while also using queries and quizzes to assess the
correctness of their causal maps. This state likely corresponds to more
effective attempts at employing monitoring strategies to identify and
correct erroneous (or missing) information in the map.
The HMM results illustrate a similar set of behaviors employed by both
the PA and CC groups, although all of the uninformed editing actions in the PA
group are combined in the uninformed editing and checking state rather than
being split between it and a separate uninformed editing state in the CC group.
10
Additionally, the proportions of expected state occurrences are also relatively
similar between the two groups. However, two interesting differences in behavior
patterns are worth noting: (i) the PA group’s model has a smaller likelihood
(39%) of transitioning directly from informed editing to informed (editing and)
checking activities compared to the CC group’s model (61% transition
probability); and (ii) the CC group’s model exhibits a higher likelihood of
following these informed editing and checking activities with more reading (29%
versus 14% for the PA group’s model).
[Fig. 1 Learning Activity Emissions in HMM States]
The HMM results (Figure 2) illustrate a similar set of behaviors employed
by both the PA and CC groups, although all of the uninformed editing actions in
the PA group are combined in the uninformed editing and checking state rather
11
than being split between it and a separate uninformed editing state in the CC
group. Additionally, the proportions of expected state occurrences are also
relatively similar between the two groups. However, three interesting differences
in behavior patterns are worth noting: (1) the PA group’s HMM has a smaller
likelihood (39%) of transitioning directly from informed editing to informed
(editing and) checking activities compared to the CC group’s HMM (61%
transition probability); (2) the CC group’s HMM exhibits a higher likelihood of
following these informed editing and checking activities with more reading (29%
versus 14% for the PA group’s HMM); (3) even in the uninformed editing and
checking state, the proportion of diagnostic to non-diagnostic queries (QUER-H
to QUER-L) was higher in the CC group. Altogether, these results indicate that
the CC group (1) was more likely to intersperse diagnostic checking activities
between their reading and informed editing activities and (2) was more likely to
return to reading after informed editing and checking, which suggests that they
may have used monitoring to identify weak points for further exploration with the
resources, as well as to confirm their current understanding. Presumably, this
periodic, diagnostic checking behavior helped students in the CC group identify,
research, and correct mistakes in their concept maps in a more systematic fashion.
[Fig. 2 HMMs for the PA and CC Groups]
To provide a complementary, finer-grained level of analysis, we augment
the HMM results with a comparison of frequent learning activity patterns as
described previously. The initial comparative sequence mining analysis revealed
12
that many of the differentially frequent patterns differed only by the number of
consecutive reads or edits in the pattern. To improve this exploratory analysis, we
revised the log pre-processing to distinguish a single action from repeated actions,
which were condensed to one “action” with the “-MULT” identifier. Using the
transformed sequences, our comparative sequence mining technique allowed us to
identify additional differences in behavior between the two groups. Table 1
presents the top three differentially frequent patterns in four comparison
categories (i.e., categorized by whether the patterns were s-frequent in one or both
groups and in which group they were more i-frequent). In this analysis, we
employed an s-support threshold of 50% to analyze patterns that were evident in
the majority of either or both groups of students.
Learning Activity Pattern i-Support
(CC – PA)
s-Frequent
Group
READ-MULT EDIT-H-MULT READ 4.39 CC
READ-MULT READ-MULT EDIT-H READ 4.11 CC
READ-MULT EDIT-H READ-MULT EDIT-H
READ
3.11 CC
EDIT-H READ 3.76 Both
EDIT-H READ-MULT 2.14 Both
READ-MULT EDIT-H READ 2.08 Both
EDIT-L QUIZ-L -0.91 Both
QUIZ-L EDIT-l -1.29 Both
EDIT-L QUER-H -1.53 Both
QUER-H EXPL-H -2.29 PA
EDIT-L EDIT-L -2.35 PA
QUER-L EDIT-L -2.47 PA
[Table 1 Top Differentially Frequent Patterns between CC and PA groups]
This analysis illustrates that many of the patterns that were differentially
frequent in the CC group (i.e., patterns with a positive value in the i-support
difference column defined as CC i-support minus PA i-support for the given
pattern) were repeated read-and-edit patterns. Many other read-and-edit
sequences were frequent in both groups, but the differentially frequent sequences
for the CC group often included single reads and informed edits, while the PA
13
group generally relied on multiple reads and their differentially frequent patterns
with single edits tended to employ uninformed edits. These results suggest that the
CC group’s students may have more frequently employed a knowledge
construction strategy of reading a small portion of the resources and adding a link
to their map each time they identified a causal relationship. Conversely, the PA
group students may have engaged in less systematic knowledge construction
strategies. In particular, they may have been less focused and encountered more
difficulties during extended editing series. This difference in learning behavior
suggests that the conversational, contextualized feedback in the CC group resulted
in more effective deployment of knowledge construction strategies.
Learning Activity Pattern i-Support
(CC – PA)
s-Frequent
Group
READ-MULT EDIT-H 2.21 CC
READ-MULT EDIT-H
READ-MULT QUER-H
1.61 CC
READ-MULT EDIT-H
QUIZ-H READ-MULT
1.61 CC
QUIZ-H EDIT-H 1.33 Both
EDIT-H QUER-H EDIT-H 1.31 Both
EDIT-H QUER-H READ-MULT 1.31 Both
EDIT-L QUIZ-L -0.91 Both
QUIZ-L EDIT-l -1.29 Both
EDIT-L QUER-H -1.53 Both
QUER-H QUER-H -2.29 PA
QUER-H EXPL-H -2.29 PA
QUER-L EDIT-L -2.47 PA
[Table 2 Top Differentially Frequent Patterns including Queries or Quizzes]
The top differentially frequent patterns in the PA group primarily involve
the use of queries or quizzes in conjunction with single, uninformed edits.
Therefore, we extended our comparative sequence mining analysis by using a
regular expression constraint to focus on monitoring patterns. In particular, we
identified interesting results for patterns including Query or Quiz actions, which
are important for effective monitoring of the knowledge construction and map
14
building process. Table 2 presents the top three differentially frequent
querying/quizzing patterns in each comparative category.
This analysis illustrates that the CC group was more likely to use queries
and quizzes in between sequences of reading and single, informed edits. In
contrast, the PA group had a differential preference for using queries before and
after individual, uninformed edits. This provides further detail to the HMM
results that illustrated a difference in transitions to the “informed editing and
checking” states and from it to the “reading” state in these groups. These results
suggest that the CC group employed monitoring activities, such as queries and
quizzes, in a way that was more related to the shorter sequences of reading and
editing that they were currently involved in. Overall, the identified difference in
learning behavior indicate that the conversational, contextualized feedback in the
CC group resulted in a beneficial change in the students’ understanding and/or
employment of both knowledge construction and monitoring strategies.
References
Agrawal, R., & Imielinski, T. (1993). Mining association rules between sets of items in large
databases. ACM SIGMOD Record, (May), 207-216. New York, New York, USA: ACM Press.
Agrawal, R., & Srikant, R. (1995). Mining sequential patterns. In P.S. Yu & A.L.P. Chen (Eds.),
Proceedings of the eleventh international conference on data engineering (pp. 3-14).
Washington, DC: IEEE Computer Society.
Aleven, V., McLaren, B., Roll, I., & Koedinger, K. (2006). Toward meta-cognitive tutoring: A
model of help-seeking with a cognitive tutor. International Journal of Artificial Intelligence in
Education, 16(2), 101–128. IOS Press.
Azevedo, R. (2005). Using hypermedia as a metacognitive tool for enhancing student learning?
The role of self-regulated learning. Educational Psychologist, 40(4), 199-209.
doi:10.1207/s15326985ep4004_2
Azevedo, R., Witherspoon, A., Chauncey, A., Burkett, C., & Fike, A. (2009). MetaTutor: A
metacognitive tool for enhancing self-regulated learning. In Cognitive and metacognitive
educational systems: Papers from the AAAI Fall Symposium (pp. 14–19). Menlo Park, CA:
AAAI Press.
Biswas, G., Jeong, H., Kinnebrew, J. S., Sulcer, B., & Roscoe, R. (2010). Measuring self-regulated
learning skills through social interactions in a teachable agent environment. Research and
Practice in Technology Enhanced Learning, 5(2), 123-152.
Cohen, P., & Beal, C. (2009). Temporal data mining for educational applications. International
Journal of Software Informatics, 3(1), 31-46.
15
Ding, B., Lo, D., Han, J., & Khoo, S.C. (2009). Efficient mining of closed repetitive gapped
subsequences from a sequence database. 2009 IEEE 25th International Conference on Data
Engineering, 1024-1035. IEEE.
Hadwin, A.F., Nesbit, J.C., Jamieson-Noel, D., Code, J., & Winne, P.H. (2007). Examining trace
data to explore self-regulated learning. Metacognition and Learning, 2(2-3), 107-124.
Hadwin, A.F., Winne, P.H., Stockley, D.B., Nesbit, J.C., & Woszczyna, C. (2001). Context
moderates students’ self-reports about how they study. Journal of Educational Psychology,
93(3), 477–487.
Jeong, H., & Biswas, G. (2008). Mining student behavior models in learning-by-teaching
environments. In R.S.J.d. Baker, T. Barnes, & J.E. Beck (Eds.),
proceedings (pp. 127-136).
Montreal, Quebec, Canada.
Kinnebrew, J.S. & Biswas, G. (2011). Modeling and measuring self-regulated learning in
teachable agents environments. Journal of e-Learning and Knowledge Society, 7(2), 19-35.
Leelawong, K., & Biswas, G. (2008). Designing learning by teaching agents: The Betty’s Brain
system. International Journal of Artificial Intelligence in Education, 18(3), 181–208.
Lo, D., Khoo, S.C., & Liu, C. (2008). Efficient mining of recurrent rules from a sequence
database. Proceedings of the 13th International Conference on Database Systems for
Advanced Applications (pp. 67–83). Springer-Verlag.
Mannila, H., Toivonen, H., & Verkamo, A. I. (1997). Discovery of frequent episodes in event
sequences. Data Mining and Knowledge Discovery, 1(3), 259-289.
doi:10.1023/A:1009748302351
Pardos, Z.A., & Heffernan, N.T. (2011). Using HMMs and bagged decision trees to leverage rich
features of user and skill from an intelligent tutoring system dataset. To appear in: The
Journal of Machine Learning Research.
Perry, N. E., & Winne, P. H. (2006). Learning from learning kits: gStudy traces of students’ self-
regulated engagements with computerized content. Educational Psychology Review, 18(3),
211-228. doi:10.1007/s10648-006-9014-3
Pintrich, P.R., Marx, R.W., & Boyle, R.A. (1993). Beyond cold conceptual change: The role of
motivational beliefs and classroom contextual factors in the process of conceptual change.
Review of Educational Research, 63(2), 167-199.
Rabiner, L. R. (1989). A tutorial on hidden Markov models and selected applications in speech
recognition. Proceedings of the IEEE, 77(2), 257–286. IEEE Computer Society.
Shih, B., Koedinger, K.R., & Scheines, R. (2010). Unsupervised discovery of student learning
tactics. Proceedings of the 3rd International Conference on Educational Data Mining, 201-
210.
Weinstein, C., Schulte, A., & Palmer, D. (1987). The Learning and Study Strategies Inventory.
Clearwater, FL: H&H Publishing.
Winne, P.H. & Jamieson-Noel, D. (2002). Exploring students’ calibration of self reports about
study tactics and achievement. Contemporary Educational Psychology, 27(4), 551-572.
16
Zimmerman, B.J. (2008). Investigating self-regulation and motivation: Historical background,
methodological developments, and future prospects. American Educational Research Journal,
45(1), 166-183.
Zimmerman, B.J. & Pons, M.M. (1986). Development of a structured interview for assessing
student use of self-regulated learning strategies. American Educational Research Journal,
23(4), 614-628.