You are on page 1of 10

Addressing the Impact of learning Analytics LAK’19, March 2019, Arizona, USA

Increasing the Impact of Learning Analytics


Shane Dawson Srecko Joksimovic Oleksandra Poquet George Siemens
Centre for Change and Centre for Change and School of Computing Centre for Change and
Complexity in Learning Complexity in Learning National University of Complexity in Learning
University of South University of South Singapore University of South
Australia Australia Singapore Australia
Australia Australia sspoquet@gmail.com Australia
shane.dawson@unisa.edu.au srecko.joksimovic@unisa.edu.au george.siemens@unisa.edu.au

ABSTRACT KEYWORDS
Learning Analytics (LA) studies the learning process in order to Learning Analytics, epistemic network analysis, adoption
optimize learning opportunities for students. Although LA has
quickly risen to prominence, there remain questions regarding the ACM Reference Format:
impact LA has made to date. To evaluate the extent that LA has Dawson, S., Joksimovic, S., Poquet, S. & Siemens, G. (2019). In-
impacted our understanding of learning and produced insights that creasing the Impact of Learning Analytics. In The 9th Interna-
have been translated to mainstream practice or contributed to the- tional Learning Analytics and Knowledge Conference (LAK19),
ory, we reviewed the research published in 2011-2018 LAK con- March, 2019, Tempe, AZ, USA. ACM, New York, NY, USA. 10
ferences and Journal of Learning Analytics. The reviewed studies pages. https://doi.org/10.1145/3303772.3303784
were coded according to five dimensions: study focus, data types,
purpose, institutional setting, and scale of research and implemen- 1 INTRODUCTION
tation. The coding and subsequent epistemic network analysis indi-
It is commonly noted that learning analytics (LA) has the potential
cates that while LA research has developed in the areas of focus
to address many of the challenges confronting contemporary edu-
and sophistication of analyses, the impact on practice, theory and
cation [20]. Cited challenges include the impact of reduced govern-
frameworks have been limited. We hypothesize that this finding is
ment funding and resources on teaching quality, increased account-
due to a continuing predominance of small-scale techno-centric ex-
ability, growing student diversity, alongside demands for flexible
ploratory studies that to date have not fully accounted for the multi-
offerings and global competition [27]. LA promises to bring new
disciplinarity that comprises education. For the field to reach its
insights into the learning process to enable practices that can en-
potential in understanding and optimizing learning and learning en-
hance student success. Alongside its proclaimed potential for im-
vironments, there must be a purposeful shift to move from explor-
pact, LA is also noted to be a developing domain of research. The
atory models to more holistic and integrative systems-level re-
rhetoric of researchers in a young field, against the backdrop of
search. This necessitates greater effort applied to understanding the
broad changes in education, requires greater scrutiny.
research cycles that emerge when multiple knowledge domains co-
The first LA conference (LAK11) was held in Banff, Canada,
alesce into new fields of research.
bringing together some one hundred researchers from around the
world with an intent to explore and shape learning analytics as a
CCS CONCEPTS discipline. Almost a decade later, it is timely to reflect on the impact
• Applied computing~Education • Applied computing~Learning LA has made: can we, as frequently claimed, demonstrate that LA
management systems is addressing the often-cited challenges facing education? Is the
field of LA beginning to realize its potential?
This paper argues that to date, LA research has failed to fully
Permission to make digital or hard copies of all or part of this work deliver on its promise to impact education and educational institu-
for personal or classroom use is granted without fee provided that tions. For the field of LA to achieve its stated goals, there is a ne-
copies are not made or distributed for profit or commercial advantage cessity for research to make a purposeful and transformative shift
and that copies bear this notice and the full citation on the first page. from a predominance of exploratory studies that are often limited
Copyrights for components of this work owned by others than ACM in scale to evaluative research that explores the impact of LA ac-
must be honored. Abstracting with credit is permitted. To copy oth- tions and outcomes at scale such as institutional adoption or sys-
erwise, or republish, to post on servers or to redistribute to lists, re- temic transformations. In making this argument, we first explain
quires prior specific permission and/or a fee. Request permissions our approach to research maturity through the concept of research
from Permissions@acm.org cycles: from conceptualization and exploration to evaluation stud-
ies that provide the necessary foundation for refining, revising or
LAK19 March, 2019, Tempe, AZ, USA
generating theory. Next, we outline how access to new sets of data
© 2019 Association for Computing Machinery.
and methods influence the types of questions posed and subse-
ACM ISBN 978-1-4503-6256-6/19/03…$15.00
https://doi.org/10.1145/3303772.3303784

446
LAK’19, March 2019, Arizona, USA S. Dawson, et al .

quently the progression of LA research. Finally, we detail the spe- order to analyze these problems in the context of the changing ed-
cific methods of this study and its findings, before discussing the ucational landscape. At the same time, the combination of multiple
need for replication and evaluation studies that contribute to the re- disciplines converging on addressing such problems raises the pos-
fining or developing theory. sibility to access new data sets or apply novel combinations of pre-
viously disparate disciplinary methodologies. This process prompts
2 RESEARCH MATURITY the initiation of numerous smaller scale exploratory works as re-
searchers interrogate the affordances of alternate methods applied
LA is commonly labeled as a bricolage field, emerging from the
into new domains. An example is the application of machine learn-
convergence of multiple disciplines [11]. While this entanglement
ing models to predict student academic performance. The studies
of ontologies, methodologies and theories generates many benefits
within the exploratory phase can lead to a higher level of evaluation
it is also accompanied with many challenges. The advantages of
and replication research that enables revisions and new contribu-
multi-disciplinary research stems from the new perspective and in-
tions to theory. Figure 2 depicts the process of research maturation
sights different theoretical and computational approaches can de-
from the surrounding environmental pressures to conceptualiza-
velop. Evidence of the impact such convergences have made can
tion, exploration, evaluation and theory building. Although the
be found in the prominence of behavioral economics. This line of
model depicts a linear set of phases, we recognize that the research
research evolved from the joining of psychology and economics
is cyclical and will constantly inform all phases from conceptual-
and has profoundly impacted on research into human decision mak-
ization to exploration, evaluation and contributions to theory. The
ing [33]. Conversely, the challenges lie in the ability to establish a
progression of research maturation that has transpired in LA is
common understanding of the strengths and weaknesses of differ-
noted by Ferguson [11], and recently by Mangarsoka and Gianna-
ent methodologies or articulating and accepting various discipli-
kos [26]. The prominence of LA grew from government and fund-
nary theories and related assumptions. However, the benefits of
ing pressures alongside the increased availability of student learn-
bringing multiple perspectives, approaches and theoretical frame-
ing data. The increased adoption of technology into the learning-
works to bear on complex problems far outweighs the transitional
teaching process generated new and accessible sets of learning data.
difficulties linked to such convergences.
The availability of learning data was notably accelerated with the
In evaluating its impact, LA can learn from other established
development of open online courses such as MOOCs. Additional
multi-disciplinary fields. Similar to LA, Information Systems (IS)
challenges such as reduced resourcing and increased accountability
research involves an interplay between a wide range of factors, in-
requirements by government agencies, helped to distinguish LA
cluding people, technology, organizational structures and culture.
and raise the importance of LA through new opportunities for re-
For example, Hevner [21] proposed a conceptual research model to
search. The transitions and progression of LA is reflected in the
understand, execute, and evaluate IS research and provide a unify-
initial works being strongly positioned around defining and con-
ing framework. Hevner and colleagues [21, 22], consider three in-
ceptualizing
ter-related cycles of IS research that are applicable to the field of
LA (Figure 1). These cycles are termed as relevance, design, and
rigor. The relevance cycle represents the problem identification
phase, whereby the conceptualization of these problems is heavily
influenced by multiple environmental factors (e.g. funding, data ac-
cess, computational methods, government and organizational pol-
icy, technologies). The design cycle is based on iterations of work
that seek to build or explore in order to gather data and insights to
refine and evaluate outcomes. Linked to the outcomes of evaluation
studies, the rigor cycle relates to feedback into theory and additions
to new knowledge. The boundaries of the research cycles are not
discrete, they are blurred and integrated. Hevner’s model illustrates
there is a level of maturity of work as research transitions from the the field and associated research possibilities [9]. In this paper we
Figure 1: Design science research cycles [20]
conceptualization of problems and the initiation of exploration
studies to the replication and evaluation of outcomes. The develop-
ment of the field of LA is analogous to the model of IS research
cycles.
The application of these research cycles to LA can enable an
understanding of how the field is maturing (Figure 2, modified from
Hevner [21]). The emergent environmental factors generate novel
pressures that require new approaches and methods for scientific
inquiries – often collaborative and multi-disciplinary [11]. The
range of environmental pressures impacting LA also invoked a nec-
essary definitional and conceptualization phase. A re-framing of Figure 2: Conceptual model of research maturation phases
problems such as student retention and success were necessary in review the state of field in terms of the types and scale of research
undertaken to determine the level of maturation of the field and

447
Addressing the Impact of learning Analytics LAK’19, March 2019, Arizona, USA

overall impact to date. In this context we view maturation as studies focus groups or interviews), or viii) other data sources (e.g., litera-
building on outcomes of prior works, rather than simply the number ture reviews). Finally, we observed the research settings - whether
of studies conducted. This process mirrors health research and drug the study collected data from i) university, ii) other formal educa-
trials. Modeling and small-scale trials lead to human trials and fi- tion organizations (e.g. k-12), or iii) informal learning contexts (e.g.
nally broad population studies, applying prior works and research. workplace).
This results in ongoing evolution of research models and focal The coding was performed through several stages. Initially,
points. three coders reviewed 50 studies for training purposes and to gain
familiarity with the literature, process and coding scheme. Next,
3 METHODS each of the coders independently coded a further 40 studies (in two
iterations – 2x20), until reaching the average agreement of 90 per-
This paper aims to identify if learning analytics research is address-
cent, with average Fleiss’ Kappa [15, 16] of 0.61. The agreement
ing its goal of developing insights and understanding into the learn-
varied from 100 percent (Fleiss’ Kappa 1) for items such as “stu-
ing process and then making an impact through improving learner
dent focus” or “other educational settings” to 77 percent (Fleiss’
success at scale. If LA research is fulfilling its self-stated potential
Kappa 0.32) for “other data”. The consistency issue was due to the
of improving student success there will be strong evidence of in-
limited number of studies for a particular coded dimension, hence,
creased research activity evaluating the interventions or support
reaching high agreement was challenging. Therefore, a decision
processes that build-on established predictive models or prior ex-
was made to revisit the definition for each data-related item (di-
ploratory work. It is feasible that as these evaluation and replication
mension) before continuing the coding process. Moreover, as there
studies are undertaken there will be an increased focus on a research
were very few studies coded as “evaluate”, “understand”, and
cycle where feedback and contributions of new knowledge provide
“feedback” (Figure 3), agreement would be revisited and consid-
insight and refinements to theory and established frameworks. It is
ered for each of the studies coded with at least one of these three
at this juncture that the field can begin to demonstrate influence on
dimensions. Finally, the remaining studies were equally split be-
student success and learning. The following research question un-
tween the three coders. Overall, each coder independently reviewed
derpins this broad aim:
approximately 150 studies.
• Has the scope and focus of LA research progressed the
field in terms of understanding the learning process and 3.2 Epistemic Network Analysis
optimizing learning for student success?
Post the coding process, the relationships between the five dimen-
3.1 Literature Analysis Procedure sions described in our coding scheme (i.e., Focus, Purpose, Scale,
Data, and Settings) were explored using Epistemic Network Anal-
To address the research question 522 papers published in LAK
ysis (ENA) [30]. ENA is a graph-based method for analyzing logs
Conference proceedings and Journal for Learning Analytics (JLA) of coded data that reveals the occurrence of relationships between
were analyzed. The process included all research and practitioner
the constructs that emerge from the underlying dataset [30]. ENA
papers between 2011 and 2018 inclusive. Editorials and posters is often used to model complex and highly dynamic problem do-
were excluded from the analyses. All studies were coded according
mains [28, 29]. Specifically, ENA models an epistemic network us-
to five overarching dimensions – Focus, Purpose, Scale, Data and ing the codes assigned to each of the items in the dataset, and where
Settings (Figure 3). Specifically, the papers were noted for their
the codes represent nodes in the network. The method has been
study Focus, that is whether the study concentrated on i) using stu- used to understand collaboration patterns across various, formal
dent data to improve learning outcome, ii) improving teaching effi-
and informal, educational domains [14, 18]. In this study, ENA is
cacy, or iii) addressing policy issues related to learning analytics used to explore the relationships between the five coded dimensions
adoption. Moreover, we also coded the reviewed papers for their
(i.e., Focus, Purpose, Scale, Data, and Settings), extracted from the
Purpose. The study purpose was adopted from Dawson, et al [9] papers reviewed in the LAK conferences and JLA.
and explains whether the goal of the study reflects i) exploratory
There are two central points of interest when modeling epis-
(i.e., predict/build), ii) evaluation, iii) conceptual (i.e., understand), temic networks - namely weights of the connections between the
iv) solution validation (i.e., feedback) types of research, or v) opin-
nodes and stanzas (i.e., the data subsets where the co-occurrence of
ion papers (i.e., other). A focus of the work was to examine the
codes is established) [29]. In addition to using binary codes (the
scale of data used in the published LA studies. Each paper was
code does or does not exist), ENA allows for specifying numeric
coded for i) having more than 100 data points (i.e., N>100); ii)
values that represent the strength of a given code. Stanza further
whether data were collected at the course or project, iii) program, defines the analysis unit or the subset of data where the links be-
iv) institution, or v) multiple-institution level. The nature and type
tween the codes are expected to be established. For example, ana-
of Data collected was also coded. This included i) instrument/pro- lyzing collaboration within a discussion forum, we can model the
tocol (e.g., survey, questionnaire), ii) multimodal data sources, iii)
epistemic network with the stanza at the level of each discussion
institutional data (e.g., learning environment, LMS, or student in-
formation system, library data), iv) third-party data (e.g., Publish-
ers, gaming environments or mobile apps); v) assessment data
(grades, GPA or pre-post test scores), vi) social media and social
interaction data (e.g. twitter, facebook), vii) qualitative data (e.g.

448
LAK’19, March 2019, Arizona, USA S. Dawson, et al .

participant or perhaps at the level of discussion thread, as being


more commonly used [14, 18]. In this study, each coded paper was
treated as stanza for ENA. As the coders reviewed each paper they
assigned a binary variable 0,1 for each of the sub-dimensions (Fig-
ure 3). For example, if the study was undertaken in a university
setting it was coded as 1 for this dimension. Specifically, we ob-
served whether a specific code (e.g., “Focus.Student” or “Pur-
pose.Evaluate”) occurred at the level of a single study. The use of
studies as a unit of analysis enabled us to explore the connections
between the five dimensions for each study. ENA in this instance,
provided a means to visualize and statistically analyze the relation-
ships between the dimensions both in terms of distance (how close
they are to appearing together) and strength (how often they appear
together). Figure 4 visualizes the relationships between the coded
dimensions plotted separately for each year. The nodes in the graph
are colored with respect to the five broad dimensions noted in fig-
ure 3 (i.e. Focus, Purpose, Scale, Data, and Settings). For the pur-
pose of simplifying the visual representation, we applied a thresh-
old to filter out the least prominent nodes in the epistemic network
(Figure 4).

4 RESULTS

4.1 General Overview


The findings from the coding process demonstrate there is a relative
consistency in the quantity of papers aligned to the categories of
focus, purpose, scale and settings from LAK conference proceed-
ings (2011-2018) and JLA publications (2014-2018). As demon-
strated in Figures 2 and 3, throughout the years, the Focus dimen-
sion of the vast majority of LA studies is on the student. This find-
ing was also observed by Knobbout and van der Stappen [24] in
their systematic review of how LA measures affects learning. The
authors noted that the majority of research were related to student
performance or learning process. In the present study, there was an
observed subtle increase in LA work directed towards educators or
instructors for the 2018 published literature (LAK conference and
JLA) from prior years. The number of studies involving policy, par-
ticularly in the area of privacy and ethics has been relatively stable
from 2011-2018 (see Figure 3).
The second category is related to the purpose of the published
studies (Figure 3). In this instance we can see that the majority of
research undertaken has been exploratory. That is, the researchers
are aiming to build a tool or through analysis of particular data sets
predict a discrete student outcome (e.g. performance,
retention). This category of research can be considered as novel but
early in development. Papers related to understanding can also be
Figure 3: Percentage of papers per year for the coded dimen-
classified as early research approaches. These studies are often con-
sions. Each paper coded may include more than one
ceptual and seeking to unpack a particular theoretical model in the
dimension.
context of LA. The number of evaluation studies peaked in 2017.
These types of studies aim to build upon the foundations of earlier studies, i.e. studies that actively provide feedback into theory for
exploratory works. For instance, the prior research may have estab-
the purposes of validation, refinement or revision, have been a rel-
lished a predictive model and suggested support actions. In evalua- atively minor focus of LAK and JLA work. For example, self-reg-
tion studies this model is tested and applied in situ. The evaluation
ulated learning is an often-used theory to interpret the findings ob-
studies, defined as seeking evaluation of implemented models or served in LA. However, using SRL as an example, there are a very
tools, can be considered as more mature and provides evidence to
limited number of studies that have sought to further revise related
lead towards the feedback category of research. To date, feedback

449
Addressing the Impact of learning Analytics LAK’19, March 2019, Arizona, USA

theoretical models. An indication of maturity of research and ap- nance of studies in university settings though there was an emer-
plied methods relates to the scale of the studies in applied research gence of work in alternate education settings. Namely, this is re-
and practice [21, 34]. In the subset of exploratory studies reviewed, flective of research in K-12 education incorporating the ASSIST-
the scale of the research tended to be confined to the course ments platform. However, these studies tended to remain at the
(S.Course) or project level (S.Project). As noted in Figure 3, a mi- course and project level.
nority of studies were undertaken at the level of program or institu- For 2014, the relationships between categories can be seen to be
tion. Studies conducted at the scale of an institution would repre- relatively consistent with the prior year. Although, there was an ob-
sent data from numerous programs of study or data that has been served increase in the scale of studies undertaken (i.e. larger popu-
collected across the entire institution (e.g. Admissions). The coding lation size) as well as a growing use of data from institutional sys-
revealed that the bulk of published works have been confined to a tems such as the LMS, Student Information Systems (learning en-
course level or project. This finding is further evidence of the small vironments) and student assessment. The number of evaluation
number of institutions globally that have large scale LA instantia- studies was more prominent than prior years. This is an important
tions [8, 10]. The lack of scale also highlights the difficulty for re- focus as LA research matures beyond exploration and development
searchers to access data that lies outside a course or program of towards evaluation and impact on theory and practice.
study often due to security and privacy concerns and institutional The pattern observed in 2013, 2014 was replicated in 2015.
challenges of accessing data [8, 10]. The coding of multi-institution However, there was an increase in the number of studies involving
data are representative of studies such as ASSISTments that involve a substantial population size (N>100). This lies in part to the num-
multiple schools or education institutions. While this category ap- ber of MOOC studies undertaken. While the population size of the
pears to reflect further scale the studies involved were also at a class research increased the studies were predominantly limited to single
level – though involving multiple classes from multiple organiza- MOOC instance. Research undertaken in “other education settings”
tions. The coding in this study also highlights the broad range of (outside the university context) also gained further significance.
data sources. There is a strong emphasis on data from institutional Additionally, an increased use of third-party tools perhaps indicates
“learning environments” such as the LMS, Student Information the growth of commercial LA products. The number of Evaluation
System and Assessment. Third party tools from publishers, ITS or studies were consistent with 2014 and remained a minority of the
gaming environments are also common. total works. Conceptual papers describing a proposed framework
The final coding category relates to the setting or context or understanding a model further decreased from its peak in 2012.
wherein the study was undertaken. As detailed in Figure 2, the Uni- An increase in the number of presented papers incorporating
versity setting continues to be the major site for LA research. Since data collected from survey instruments and self-reported data was
2011 the number of studies from “other education settings” such as observed in 2016. Although these methods were employed in ear-
K-12 have been steadily increasing. This may suggest that LA is lier conferences and papers, there was a substantial increase in
broadening its reach beyond university and is gaining traction in K- 2016. This may be reflective of an interest in studies of learner dis-
12. Interestingly, the number of studies involving informal settings positions and cognitive psychology such as attention and mind
such as the workplace have decreased over time. wandering, as well as a sign of the recognition that LA studies were
lacking the ‘social’ element of the socio-technical systems where
4.1 Network Relationships learning occurs. There was also a focus on mapping qualitative data
Visual and statistical analysis of the change in the focus of reviewed and self-reported psychological surveys to clickstream data. The
learning analytics literature was undertaken using ENA. The results number of evaluation studies remained consistent.
of the ENA (Figure 4) illustrate how the reviewed studies addressed For 2017 there was a rise in the number of evaluation studies.
the aspects of Focus, Purpose, Scale, Data, and Settings over the This category is representative of more mature studies that seek to
past eight years of learning analytics research and practice. The assess and refine earlier forms of exploratory research. For exam-
group average network for 2011 indicates that the main focus of the ple, the evaluation of a predictive model and noted support inter-
presented papers related to the “other” classification of papers. That ventions through field and case studies. Although an increase in
is, the presented papers were of a conceptual nature reflecting the 2017 was observed, for 2018 the quantity of evaluation studies de-
discussions pertaining to the definition or framing of learning ana- clined. The network structure linking relationships between student
lytics in university settings. focused studies that undertake predict/ build research in university
In 2012, the focus of the research tended to be on student suc- settings identified from 2013 has remained consistent over the
cess and retention in university settings. However, the scale of the years. In order to evaluate the differences between the years with
reported studies were of a small population size indicative of a sin- respect to the first dimension of the ENA space (x-axis Figure 3),
gle project or course. The papers reported findings from explora- we performed the non-parametric Kruskal-Wallis test. The test was
tory studies or the development of predictive models or LA tools. 2(1, 7) = 15.61, p=0.03. However, the further post hoc analysis us-
Although minor, in 2012 there was an introduction of policy papers ing Dunn’s test of multiple comparisons using rank sums, showed
related to privacy and ethics as well as studies of LA in informal that only the difference in ENA space between 2011 and 2015 was
learning settings. However, 2013, saw an observed increase in the statistically significant.
use of third party and assessment data to build models that would
predict student retention or success. There remained a predomi-

450
LAK’19, March 2019, Arizona, USA S. Dawson, et al .

Figure 4: Epistemic networks between the groups of coded dimensions. Each label is constructed using a letter to represent the five
coded dimensions (Focus - F, Purpose - P, Scale - S, Data - D, and Settings – St) and the subcategory (e.g., student, learning environ-
ment). Nodes are colored according to the dimension. X-axis represents the first dimension of the epistemic space (svd1), and the y-
axis represents the second (svd2) dimension of the epistemic space.

451
Addressing the Impact of learning Analytics LAK’19, March 2019, Arizona, USA

simplistic level, this is a two-part problem requiring researchers to


5 DISCUSSION first understand the learning process before it can be optimized.
This study identifies the extent learning analytics research, as However, the complexity of this inquiry is reflected in the existing
evidenced in LAK conference proceedings and JLA articles, is ad- literature, as the majority of LA publications to date have failed to
dressing the goal of developing insights into, and an understanding reach the adequate maturity of the research rigor cycle to be able to
of, the learning process. The findings indicate that the majority of make a broader impact on the existing theories, frameworks, and
LA studies can be classified as exploratory research, focusing on models (see Figure 4). For LA to reach its potential as a research
students enrolled in independent courses or small projects. The data domain, it is necessary to untangle the complex interactions be-
types tend to be derived from institutional technologies such as the tween social, technical, and environmental factors at the systems
Learning Management System (LMS), Student Information Sys- level [10].
tems, and assessment data. Example studies include investigating A student’s learning experience rarely starts and ends in one
factors that influence student success outcomes leading to the de- course. An integrative, complex, and holistic view is needed to un-
velopment of a particular predictive model or suggested interven- derstand the dynamics outside of a specific course that influence
tions and support actions. However, as the findings demonstrate, learning performance. However, the focus, purpose, and the scale
there are relatively few studies to date that have attempted to extend of the existing research, seem to be impacted by the current suite of
exploratory works to evaluate the results or suggested actions to institutional technologies employed and the associated data sets
interrogate the proposed model and feedback to theory. The re- (Figure 4). The reviewed publications suggests that the abundance
search reviewed for this paper suggests the field has yet to under- of, and access to, technology-mediated learning data has encour-
take a substantial engagement with what Hevner [21] described as aged the deployment and advancement of computational methods.
the rigor cycle, which would indicate development of the field and Furthermore, the access to large sets of institutional learning data
growing impact on theoretical models. While the maturation of LA has also influenced the scoping of research problems and hence
has transitioned from conceptualization, it remains mired within a overall direction and progression of LA.
phase of exploratory analyzes. Technologies and teaching practices can provide a rich data
We argue that the field requires further impetus to facilitate the source at the same time a limited source of data. For instance, the
transition from technical exploratory works to the evaluation of LA reviewed papers demonstrate the heavy use of trace data collected
actions within the broader socio-technical system LA operates in. through various technologies such as Intelligent Tutoring Systems
As illustrated in Figures 1 and 2, a focus on evaluation research can (ITS) (e.g. ASSISTments) or Learning Management Systems
fuel the development of new insights into theory and frameworks (LMS). These learning technologies produce large volumes of stu-
to support the ongoing maturation of the field. These findings re- dent interaction data. While we are witnessing an ever-growing so-
flect Gasevic, et al’s [17] assertion that LA research has all too fre- phistication of analyses of these forms of trace data such as statis-
quently centered on the identification of attributes and behavioral tical network analysis, soft clustering of student learning strategies
characteristics of students that are associated with discrete learning and transitions, the data lacks the personal, social and cultural in-
outcomes. LA must extend beyond individual analysis and provide formation that also impacts on student learning. In short, the access
rigor in evaluating assumptions, interventions and actions to refine to the data obtained from these and similar systems can be auto-
established or generate new theories of learning. If the factors un- mated and scaled, making such data readily accessible for the ap-
derpinning a predictive model of student success holds true, then, plication of statistical and machine learning methods. In contrast,
neither educators nor students have sufficiently changed practice to despite the recent advances in multimodal LA [3], data concerning
improve learning outcomes. social and personal dimensions such as motivations, emotions,
Our premise for this paper is that to date, LA has not, and will health and culture are reliant on self-reports or collected from ex-
not, make a significant impact at scale while research and deploy- pensive and intrusive equipment. Even when physiological and so-
ment are undertaken at the level of an individual course instructor cio-emotional data are captured through technologies there are
or in small-scale experimental studies, in lieu of a more systems- challenges in synchronizing such discrete data sources. Although
oriented approach. Further research is required for replication and there is a lack of richness in trace data, the reviewed works demon-
evaluation studies to refine, revise or establish new theories of strate that LA research has well developed numerous predictive
learning, adoption or design. To achieve this goal a systems orien- models of student success. The capacity to predict student out-
tation is needed to identify how LA can improve student success in comes from behavioral and assessment data affords educators early
complex organizational systems [25]. and timely opportunity to intervene [17]. The ability for early inter-
vention to aid student success drives LA practice and research, re-
5.1 LA and Socio-technical Systems flecting a pursuit to understand both learning process and context
with the intent to improve student outcomes. Hence, the ability to
Similar to information systems research or health sciences, LA
generate actionable, timely formative and personalized feedback is
operates at the intersection of multiple disciplines in solving com-
a noted goal of LA research and practice. However, to achieve an
plex problems that involve the confluence of social and technical
understanding of the learning process requires more than the gen-
factors. Instead of improving the efficiency of organizational pro-
eration of interventions. The notion that such actionable intelli-
cesses or the effectiveness of health services, LA focuses on the
gence is generated through LA is central to much of the existing
specific problem of understanding and optimizing factors that lead
research. Greller and Drachsler [19], and Chatti, et al. [5], proposed
to a successful educational experience for all learners [26]. At a

452
LAK’19, March 2019, Arizona, USA S. Dawson, et al .

a set of dimensions that influences both LA uptake from a socio- fall into a hierarchical structure. For LA this may also provide a
technical and organizational perspective. Clow [6], and Campbell useful model for thinking how learning manifests within education
and Oblinger [4] were more direct to articulate the process of LA settings and how LA can be deployed for deep impact in education
as a research cycle involving discrete but interrelated activities. organizations. For LA, the research is yet to interrogate how actors
Clow’s [7] LA cycle is the most generic. The cycle commences in complex systems interact – both in terms of a) the learning pro-
with learners – as students undertake various course activities, they cess and b) the application of LA.
generate data that is then analyzed to provide the basis for some The findings in this study raise several important implications
action. Campbell and Oblinger [4] stress the importance of includ- for LA research and practice. Research does not operate in isolation
ing an additional step involving the testing of the prescribed support of other systems nor is it de-contextualized from practical applica-
actions to further revise and refine the findings tion. All research is impacted and influenced at numerous levels
The ENA results (see Figure 4) illustrate that there is predomi- such as politics, funding sources from international, national and
nance of small-scale exploratory studies and a deficit of research local sources, availability and access to data, opportunities and lim-
undertaking evaluative studies that attempt to verify findings and itations of methods and access to expertise through our own collab-
provide explicit feedback to revise, refine or develop new theories orative networks. As a practical example of the impact of external
and models (see Figure 2). This requires researchers to shift from factors on the nature of LA research, consider the influence of re-
small-scale research projects to areas that address and target sys- duced government funding to education and establishment of reten-
temic impact and outcomes. The achievement of this aim will re- tion as a performance metric in driving a research focus towards
quire a foundation of, and culture for, supporting replication stud- models of student success and retention. To date though, research
ies. Such an approach is difficult to achieve as research into social consideration for how LA manifests within a complex system has
phenomena are often contradictory [35]. Results identified in one been minimal.
study are frequently contrasted by others. Learning analytics is no Education is highly representative of a complex system residing
exception [1]. What has been demonstrated to improve or predict within multiple broader system structures. Our models of teaching
student success in one instance, classroom or institutional context, from on campus to fully online, the diversity of students, the tech-
may or may not be replicated in alternate studies. An explanation nical ecosystem required to support the myriad of academic prac-
for the lack of replicable outcomes lies in the dynamics of the sys- tices and policies, local and global economic forces and societal
tems within which LA operates. This is not to say that the research expectations all play a role in shaping how education functions, pri-
reviewed in this paper is not rigorous, but rather that researchers oritizes actions, and responds to such environmental pressures. The
frequently adopt alternate theories and models and statistical ap- inclusion of new practices or processes within such a complex sys-
proaches which in turn influence interpretations and outcomes. tem requires highly skilled leadership, substantial and persistent ef-
Hence, achieving the goal of improving learning processes and im- fort and organizational agility [31]. Of equal importance in driving
pacting outcomes can be difficult as much of the data being col- the broader impact of LA is an understanding of how these systems
lected is not only limited in scope, but also fails to account for so- dynamics influence the adoption of insights gained from more gran-
cial, cultural, and contextual variables, requiring researchers to ular LA research such as cognitive profiles of learners, predictive
adopt proxies related to grades and completion rates, rather than models of learner success, and feedback and support interventions.
more sophisticated approaches to optimize learning for all learners Over the past decade there have been a number of LA institu-
[23]. The research cycles provide a high level framework for as- tional adoption models proposed that attempt to reflect the role data
sessing LA work to date. The results highlight the influence that and data analytics plays in education systems [see for example: 2,
surrounding environmental factors and access to data plays on the 8, 13, 19]. In particular, Greller and Drachsler [19] argued that any
types of problems that are being explored and methodologies successful large scale LA instantiations must capture the interre-
adopted. Due to these constraints, LA as a field has been stifled in lated nature that operates between the identified dimensions that
terms of addressing complex social challenges and has not devel- can mediate adoption. Similarly, Ferguson, et al., [13] proposed the
oped the types of evaluative models that are required for ongoing use of the RAPID Outcomes Mapping Approach (ROMA) as a
impact and growth of the field. model to identify the factors influencing adoption and how these
factors and agents intersect within a dynamic system. Despite the
5.2 Implications number of proposed models to there remain very few large scale
Learning analytics like social science cannot be reduced to simplis- LA instantiations. LA research is clearly multi-disciplinary and de-
tic analytical problems. In recent years there has been momentum ploying a LA initiative similarly comprises numerous interactions
in developing more complex models of research and thinking that with the many sub-units that make up an education organization.
utilize simulations, alternate agents, biological networks, adjacent For example, LA adoption will involve, technical systems, aca-
possibilities and evolutionary pathways to name but a few. In es- demic systems, teaching and workload models, ethics and privacy,
sence, these approaches aim to identify how the multitude of inter- assessment policies, retention practices, communication processes
connected nodes within a system impact and influence each other. and so on. The ability to call on replicated studies that demonstrate
These relationships are not necessarily linear nor predictable. Yet, the impact of LA can only serve to aid uptake. However, without a
as Herbert Simon [32] remarked, an observed theme in complex purposeful intent to move LA works from decontextualized and ex-
systems is the concept of hierarchy. That is actors within a system ploratory settings to replicated and evaluation, there will not be a
are often comprised of inter-related sub-systems that can be seen to substantial impact on theory, adoption, nor sector wide transfor-

453
Addressing the Impact of learning Analytics LAK’19, March 2019, Arizona, USA

mation. This problem is not solely borne by LA. Social science re- [3] Blikstein, P. and Worsley, M. (2016). Multimodal learning analytics
search has long been confronted with issues of replication and val- and education data mining: Using computational technologies to measure
complex learning tasks. Journal of Learning Analytics, 32, 2, 220-238.
idation of research findings [35]. We can collectively begin to [4] Campbell, J. and Oblinger, D. (2007). Academic analytics.
tackle this challenge through deeper collaborations and investiga- EDUCAUSE Quarterly Magazine, 2007, 25th October.
tions into the types of questions posed, and data and methods ap- [5] Chatti, M. A., Dyckhoff, A. L., Schroeder, U. and Thüs, H. (2012). A
plied to LA. Incorporating for example, enhanced data sets beyond reference model for learning analytics. International Journal of
technology mediated trace data to embrace other design, cognitive Technology Enhanced Learning, 4, 5-6, 318-331.
[6] Clow, D. (2012). The learning analytics cycle: closing the loop
and socio-emotional dimensions at scale. effectively. In Proceedings of the Proceedings of the 2nd International
The findings from this study reflect the complexity in address- Conference on Learning Analytics and Knowledge (Vancouver, British
ing the optimization of learning for a field established from the con- Columbia, Canada). ACM.
vergence of multiple disciplines. However, the exploratory studies [7] Clow, D. (2013). An overview of learning analytics. Teaching in
undertaken can and do provide a substantial foundation in order to Higher Education, 18, 6, 683-695.
[8] Colvin, C., Rogers, T., Wade, A., Dawson, S., Gašević, D.,
transition to more evaluation and replication studies. Such a step Buckingham Shum, S. and Fisher, J. (2015). Student retention and
may require stimulus from existing research organizations, funding learning analytics: A snapshot of Australian practices and a framework
agencies, collaborations and the active promotion of established for advancement. Australian Office for Learning and Teaching, Sydney,
works such as the LACE initiative [12]. Collectively, the field must Australia.
pursue and engage in the complexity of these systems and in ad- [9] Dawson, S., Gašević, D., Siemens, G. and Joksimovic, S. (2014).
Current state and future trends: a citation network analysis of the learning
vancing understanding of how principles of dynamic systems influ- analytics field. In Proceedings of the Fourth International Conference on
ences adoption. Learning Analytics And Knowledge (LAK '14) (Indianapolis, USA). ACM,
New York, NY, USA, .
[10] Dawson, S., Poquet, O., Colvin, C., Rogers, T., Pardo, A. and
6 CONCLUSIONS Gasevic, D. (2018). Rethinking learning analytics adoption through
While LA research has not yet reached its potential, it is advancing complexity leadership theory. ACM, NY, USA, City.
and is on the right path to fulfil its stated promise of generating [11] Ferguson, R. (2012). Learning analytics: Drivers, developments and
sector wide transformations. The realist in this situation would flag challenges. International Journal of Technology Enhanced Learning, 4,
5/6, 304-317.
this will only come to fruition through understanding and engineer- [12] Ferguson, R. and Clow, D. (2016). Learning analytics community
ing the influential actors that reside in the education system. Tar- exchange: evidence hub. In Proceedings of the Sixth International
geting research towards LA as a complex model may help craft the Conference on Learning Analytics & Knowledge (Edinburgh, UK). ACM.
understandings required to transition from the current deep insights [13] Ferguson, R., Macfadyen, L., Clow, D., Tynan, B., Alexander, S. and
Dawson, S. (2014). Setting Learning Analytics in Context: Overcoming
about learning behavior to knowledge about the application of LA
the Barriers to Large-Scale Adoption. Journal of Learning Analytics, 1, 3,
for action and impact within the hierarchical sub-units of a complex 120-144.
education system. Now is the opportune time to achieve this goal [14] Ferreira, R., Kovanović, V., Gašević, D. and Rolim, V. (2018).
by engineering and influencing the direction of our future research Towards Combined Network and Text Analytics of Student Discourse in
to consider LA within a more holistic model. That is, to undertake Online Discussions. In Proceedings of the International Conference on
Artificial Intelligence in Education. Springer, Cham.
research that aims to understand the learning process as well as op- [15] Fleiss, J. L. and Cohen, J. (1973). The equivalence of weighted kappa
timizing that process for all students across all contexts. Research and the intraclass correlation coefficient as measures of reliability.
directions and priorities can be influenced. The strength of our work Educational and Psychological Measurement, 33, 613–619.
from individual researchers to international collaborations and [16] Freelon, D. (2013). ReCal OIR: Ordinal, interval, and ratio intercoder
reliability as a web service. International Journal of Internet Science, 8, 1,
noted research organizations (such as SoLAR and IEDM) can col-
10-16.
lectively present a powerful voice to generate the opportunities to [17] Gašević, D., Dawson, S. and Siemens, G. (2015). Let’s not forget:
have a profound influence on the types of research that are funded, Learning analytics are about learning. TechTrends, 59, 1, 64-71.
supported and ultimately undertaken in order to encourage new di- [18] Gasevic, D., Joksimovic, S., Eagan, R. and Shaffer, D. W. (2018).
rections in the field. Discussions with granting agencies and direct SENS: Network analytics to combine social and cognitive perspectives of
collaborative learning. Computers in Human Behavior.
funding of PhD and early career scholarships and research grants [19] Greller, W. and Drachsler, H. (2012). Translating learning into
can provide the impetus to expand the level of LA research into numbers: A generic framework for leanring analytics. Educational
evaluation and replication studies to fuel the knowledge base and Technology & Society, 15, 3, 42-57.
revise, refine and contribute new theory. This will require a pur- [20] Haythornthwaite, C., de Laat, M. and Dawson, S. (2013).
poseful shift to target, promote and enact such research. Introduction to the Special Issue on Learning Analytics. American
Behavioral Scientist, 57, 10, 1371-1379.
[21] Hevner, A. R. (2007). A Three Cycle View of Design Science
Research. Scandinavian Journal of Information Systems, 19, 2, 87-92.
[22] Hevner, A. R., March, S. T., Park, J. and Ram, S. (2004). Design
REFERENCES Science in Information Systems Research. Management Information
Systems Research Center, 28, 1, 75-105.
[1] Andres, J. M. L., Baker, R. S., Siemens, G., Gašević, D. and Spann, C.
[23] Joksimović, S., Poquet, O., Kovanović, V., Dowell, N., Mills, C.,
A. (2017). Replicating 21 findings on student success in online learning.
Gašević, D., Dawson, S., Graesser, A. C. and Brooks, C. (2018). How do
Technology, Instruction, Cognition, & Learning, 10, 4, 313-333. we model learning at scale? A systematic review of research on MOOCs.
[2] Bichsel, J. (2012). Research computing: The enabling role of
Review of Educational Research, 88, 1, 43-86.
information technology. EDUCAUSE Center for Applied Research.

454
LAK’19, March 2019, Arizona, USA S. Dawson, et al .

[24] Knobbout, J. and van der Stappen, E. (2018). Where Is the Learning
in Learning Analytics? A systematic literature review to identify measures
of affected learning. In Proceedings of the European Conference on
Technology Enhanced Learning. Springer, Cham.
[25] Macfadyen, L. P., Dawson, S., Pardo, A. and Gaševic, D. (2014).
Embracing big data in complex educational systems: The learning
analytics imperative and the policy challenge. Research & Practice in
Assessment, 9, 17-28.
[26] Mangaroska, K. and Giannakos, M. N. (2018). Learning analytics for
learning design: A systematic literature review of analytics-driven design
to enhance learning. IEEE Transactions on Learning Technologies.
[27] Ognjanovic, I., Gasevic, D. and Dawson, S. (2016). Using
institutional data to predict student course selections in higher education.
The Internet and Higher Education, 29, 49-62.
[28] Shaffer, D. W. (2006). Epistemic frames for epistemic games.
Computers & Education, 46, 3, 223-234.
[29] Shaffer, D. W., Collier, W. and Ruis, A. R. (2016). A tutorial on
epistemic network analysis: Analyzing the structure of connections in
cognitive, social, and interaction data. Journal of Learning Analytics, 3, 3,
9-45.
[30] Shaffer, D. W., Hatfield, D., Svarovsky, G. N., Nash, P., Nulty, A.,
Bagley, E., Frank, K., Rupp, A. A. and Mislevy, R. (2009). Epistemic
Network Analysis: A Prototype for 21st-Century Assessment of Learning.
International Journal of Learning and Media, 1, 2, 33-53.
[31] Siemens, G., Dawson, S. and Eshleman, K. (2018). Complexity: A
Leader’s Framework for Understanding and Managing Change in Higher
Education. EDUCAUSE Review, 53, 6.
[32] Simon, H. A. (1962). The Architecture of Complexity. Proceedings
of the American Philosophical Society, 106, 6, 467-482.
[33] Thaler, R. H. (2016). Behavioral economics: past, present, and future.
American Economic Review, 106, 7, 1577-1600.
[34] Torraco, R. J. (2002). Research Methods for Theory Building in
Applied Disciplines: A Comparative Analysis. Advances in Developing
Human Resources, 4, 3, 355-376.
[35] Watts, D. J. (2017). Should social science be more solution-oriented?
Nature Human Behavior, 1.

455

You might also like