Learning analytics

The educational technology and digital learning wiki
Jump to navigation Jump to search

<pageby nominor="false" comments="false"/>

Introduction

One could define learning analytics as collection of methods that allow teachers and maybe the learners to understand what is going on. I.e. all sorts of tools that allow to gain insight on participant's behaviors and productions (including discussion and reflections). The learning analytics community in their call for the 1st International Conference on learning analytics provides a more ambitious definition: “Learning analytics is the measurement, collection, analysis and reporting of data about learners and their contexts, for purposes of understanding and optimising learning and the environments in which it occurs”, i.e. also view it in a clearly transformative perspective. Educational data mining is a field that has similar interest, but that focuses more on research. In addition, favorite data gathering, aggregation/clustering, analysis and visualization techniques differ a bit.

This EduTechWiki article will particularly focus on analytics for project-oriented learning such as the Knowledge-building community model or Inquiry-based learning. We have the impression that learning analytics is mostly thought of as something that could be used to improve low-quality education, e.g. main-stream e-learning, and not as a cognitive tool designed to improve activity-based learning designs. Although we will try to cover this area as a whole, please follow-up links or the literature (both below) if you are interested in tracking reading and quizzing activities or improving your school or the educational system. An additional reason for looking at other introductions and discussions is that we have a bias against what we call "bureaucratization through data graveyarding". While we do acknowledge the usefulness of all sorts of analytics, we must make sure that it will not replace design and communication. - Daniel K. Schneider 23:00, 11 April 2012 (CEST)

See also:

The Society for Learning Analytics Research Open Learning Analytics (2011) proposal associates learning analytics with the kind of "big data" that are used in busines intelligence:

The rapid development of “big data” methods and tools coincides with new management and measurement processes in corporations. The term “business intelligence” is used to describe this intersection of data and insight. When applied to the education sector, analytics fall into two broad sectors (Table 1): learning and academic.

Learning analytics (LA) is the measurement, collection, analysis and reporting of data about learners and their contexts, for purposes of understanding and optimizing learning and the environments in which it occurs. Learning analytics are largely concerned with improving learner success.

Academic analytics is the improvement of organizational processes, workflows, resource allocation, and institutional measurement through the use of learner, academic, and institutional data. Academic analytics, akin to business analytics, are concerned with improving organizational effectiveness.
Learning and Academic Analytics. © 2011 George Siemens and Phil Long, EDUCAUSE Review, vol. 46, no. 5 (September/October 2011) . Licensed under the Commons Attribution-NonCommercial 3.0 License

In other words, learning analytics concerns people concerned by teaching and learning, e.g. learners themselves, teachers, course designers, course evaluators, etc. Learning analytics is seen “LA as a means to provide stakeholders (learners, educators, administrators, and funders) with better information and deep insight into the factors within the learning process that contribute to learner success. Analytics serve to guide decision making about educational reform and learner-level intervention for at-risk students.” (Simons et al. 2011: 5)

This definition of the society of learning analytics and the US Dept. of education report that we shall discuss below are political and like many other constructs in the education sciences "learning analytics" promises better education. We therefore conclude the introduction with the impression that learning analytics either can be seen (1) as tools that should be integrated into the learning environment and scenario with respect to specific teaching goals or (2) as a more general and bigger "framework" for doing "education intelligence". The latter also could include the former, but not necessarily so.

We therefore distinguish three kinds of learning analytics:

  • Small scale learning analytics that is just another name for "awareness tools", "learning cockpits" and so forth and that have a long tradition in educational technology
  • Mid scale analytics who's primary function is to help improve courses, detect and assist students with difficulties
  • Large scale analytics (also called academic analytics) whose function is to improve institutions or even whole educational systems.

Approaches, fields and conceptual frameworks

This chapter attempts to identify various strands of research and development and to discuss some of these in more detail.

Approaches to learning analytics and educational data mining

Some kinds of learning analytics and educational data mining have been known and used since education exists. Examples are:

Modern learning analytics (i.e. that is labelled as such) probably draws a lot of ideas from industry and its various attempts to deal with information, such as business intelligence, Quality management, Enterprise resource planning (ERP), Balanced score cards, marketing research and Web analytics. The common idea is to extract Key performance indicators from both structured and unstructured data and to improve decision making.

A related second important influence is a political will to articulate and measure quality standards in education leading to a culture of quantiative assessment (Oliver and Whelan, 2011; Hrabowski, Suess et al. 2011).

One difference between various approaches to learning analytics (in a wide sense) concerns the locus of data collection and data analysis, and on how big a scale (number of students, activities and time) it is done. In radical constructivist thought, students should be self-regulators and the role of the system is just provide enough tools for them as individuals or groups for understanding what is going on. In more teacher-oriented socio-constructivism, the teachers, in addition, need indicators in order to monitor orchestrations (activities in pedagogical scenarios). In other designs such as direct instruction, the system should provide detail performance data that are rather based on quantitative test data. Finally, powers above like schools or the educational systems would like to monitor various aspects.

Educational data mining (EDM) could be described as a research area that focuses more on data retrieval and analysis. I mainly targeted educational research using and developing all sorts of technology, e.g. information retrieval, clustering, classification, sequence mining, text mining, visualization, and social network analyis. Today, EDM also can be seen as "helper discipline" for more sophisticated learning analytics approaches.

According to Baker & Yacef (2009) “Educational Data Mining researchers study a variety of areas, including individual learning from educational software, computer supported collaborative learning, computer-adaptive testing (and testing more broadly), and the factors that are associated with student failure or non-retention in courses.”. Key areas are:

  • Improvement of student models, i.e. models that are typically used in artificial intelligence and education systems such as intelligent tutoring systems.
  • Discovering or improving models of a domain’s knowledge structure
  • Studying pedagogical support (both in learning software, and in other domains, such as collaborative learning behaviors), towards discovering which types of pedagogical support are most effective.
  • Empirical evidence to refine and extend educational theories and well-known educational phenomena, towards gaining deeper understanding of the key factors impacting learning, often with a view to design better learning system.

This overview paper also points out that relationship mining was dominant between 1995 and 2005, Prediction moved to the dominant position in 2008-2009. In addition, the authors noted an increase of data coming from the instrumentation of existing online courses as opposed to harvesting the researcher's-as-teacher's own data.

Fournier et al. (2011) quote Downs (2010) “There are different tools for measuring learning engagement, and most of them are quantificational. The obvious ones [measure] page access, time-on-task, successful submission of question results – things like that. Those are suitable for a basic level of assessment. You can tell whether students are actually doing something. That’s important in certain circumstances. But to think that constitutes analytics in any meaningful sense would be a gross oversimplification. [...] There is a whole set of approaches having to do with content analysis. The idea is to look at contributions in discussion forums, and to analyze the kind of contribution. Was it descriptive? Was it on-topic? Was it evaluative? Did it pose a question?”. For the later kind (btw. not just forums but all sorts of student productions), we do need text mining technology. In addition, Fournier point out that data mining technologies may have to be combined with semi-manual qualitative analysis, in particular social networking analysis.

The SoLAR framework

This proposal addresses the need for integrated toolsets through the development of four specific tools and resources:

  1. A Learning analytics engine, a versatile framework for collecting and processing data with various analysis modules (depending on the environment)
  2. Adaptive content engine
  3. Intervention engine: recommendations, automated support
  4. Dashboard, reporting, and visualization tools

A short discussion

The proposal starts somehow with the assumption that education continues to use rather simple tools like courseware or somewhat now popular web 2.0 tools like personal learning environments. In other words, the fog is identified as the problem and not the way education is designed. I.e. this proposal focuses on "intelligence" as opposed to "design". If you like, "learning design" could become more like a "General motors" approach (marketing oriented) than "Apple" (design-oriented). Related to that we also can identify the general assumption that "metrics" work, while in reality test-crazy systems like the US high-school education have a much lower performance than design-based systems like the Finnish one. We are not really familiar with the long history of business intelligence, management information systems but we are familiar with the most prominent critique, i.e. that such systems put a distance between decision makers and staff, kill communication lines and inhibit managers from thinking.

The project proposes a general framework based on modern service-oriented architectures. So far, all attempts to use such architectures in education did fail, probably because of the lack of substantial long-term funding, e.g. see the e-framework project. We also a wonder a bit how such funding could be found, since not even the really simple IMS Learning Design has been implemented in a usable set of tools. In addition, even more simple stuff, like simple and useful wiki tracking is not available, e.g. read wiki metrics, rubrics and collaboration tools

We would like to draw parallels with (1) the metadata community that spent a lot of time designing standards for describing documents and instead of working on how to create more expressive documents and understanding how people compose documents, (2) with business that spends energy on marketing and related business intelligence instead of designing better products, (3) with folks who believe in adaptive systems forgetting that learner control is central to deep learning and that most higher education is collaborative, (4) with the utter failure of intelligent tutoring systems trying to give control to the machine and (5) finally with the illusion of learning style. These negative remarks are not meant to say that this project should or must fail, but they are meant to state two things: The #1 issue is in education is not analytics, but designing good learning scenarios within appropriate learning environments (most are not). The #2 issue is massive long term funding. Such a system won't work before at least 50 man years over a 10 year period will be spent.

Somewhat it also is assumed that teachers don't know what is going on and that learners can't keep track of their progress or more precisely that teachers can't design scenarios that will help both teachers and students knowing what is going on. We mostly share that assumption, but would like to point out that knowledge tools do exist, e.g. knowledge forum, but these are never used. This also can be said with respect to CSCL tools that usually include scaffolding and meta-reflective tools. In other words, this proposal seems to imply that that education and students will remain simple, but "enhanced" with both teacher and student cockpits that sort of rely on either fuzzy data mined from social tools (SNSs, forums, etc.) or quantitative data from learning management systems.

Finally, this project raises deep privacy and innovation issues. Since analytics can be used for assessment, there will be attempts to create life-long scores. In addition, if students are required to play the analytics game in order to improve chances, this will be other blow to creativity. If educational systems formally adopt analytics, it opens the way for keeping education in line with "main-stream" e-learning, an approach designed for training simple subject matters (basic facts and skills) through reading, quizzing and chatting. This, because analytics will work fairly easily with simple stuff, e.g. scores, lists of buttons clicked, number of blog and forum posts, etc.

This being said, we find the SoLAR project interesting, but under the condition that it should not hamper use of creative learning designs and environments. Standardization should be enabling (e.g. like notation systems in music are) and not reduce choice.

US Dept. of education view

Bienkowski et al. (2012) in their Enhancing Teaching and Learning Through Educational Data Mining and Learning Analytics (draft) prepared for the US Department of Education firstly defines Educational data mining. It is “emerging as a research area with a suite of computational and psychological methods and research approaches for understanding how students learn. New computer-supported interactive learning methods and tools— intelligent tutoring systems, simulations, games—have opened up opportunities to collect and analyze student data, to discover patterns and trends in those data, and to make new discoveries and test hypotheses about how students learn. Data collected from online learning systems can be aggregated over large numbers of students and can contain many variables that data mining algorithms can explore for model building.” (p.9, draft).

Learning analytics is defined as data mining plus interpretation and action. “Learning analytics draws on a broader array of academic disciplines than educational data mining, incorporating concepts and techniques from information science and sociology in addition to computer science, statistics, psychology, and the learning sciences. [...] Learning analytics emphasizes measurement and data collection as activities that institutions need to undertake and understand and focuses on the analysis and reporting of the data. Unlike educational data mining, learning analytics does not generally address the development of new computational methods for data analysis but instead addresses the application of known methods and models to answer important questions that affect student learning and organizational learning systems.” (p.12, draft)

Discussion

We find it very suprising that the Bienkowski report argues that learning analytics do not address the development of new computational methods when in reality they most interesting learning analytics projects (e.g. Buckingham Shum & Crick, 2012 or Govaerts et al., 2010) exactly do that. In addition, we find that the report seems to missing one crucial issue: Enabling the learner and learner communities. 20 years of research in educational technology is missing (e.g. cognitive tools, learning with portfolioss, project-oriented learning and writing-to-learn). On the other hand, AI&Ed concepts that have been given up in the 1990's (and for good reasons) seem to re-emerge without sufficient grounding in contemporary learning theory and first principles of instructional design. Analytics that attempt to improve the learning process through awareness tools are maybe "small scale analytics" and definitely not predictive.

Below is a figure explaining how adaptive learning systems should work. Other scenarios are discussed in the report and we singled this one out, because adaptive learning systems never had any success since they do not fit into either constructivist or direct instructional design models. Note the relative absence of the learner. In the picture the student is just directed to look at contents ...

The Components and Data Flow Through a Typical Adaptive Learning System, Bienkowski et. al. 2012 (Draft version)

The OU Netherlands framework

According to the Learning Analytics framework page of the Centre for Learning Sciences and Technologies, Open Universiteit (retrieved April 13 2012), the “topic of Learning Analytics is structured in six critical dimensions, shown in the diagramme below. Each of these dimensions has instantiations which are exemplified in the inner circle items. These are only typical examples and not a complete list of possible instantiations.”

In a presentation, called Turning Learning into Numbers - A Learning Analytics Framework, Hendrik Drachsler and Wolfgang Geller presented the following global component framework of learning analytics. The framework also was presented in Hendrik Drachsler's lecture on recommender systems for learning (April 2012).

Drachsler / Geller learning analytics framework. Copyright: Drachsler & Geller, Open Universiteit. Reproduced with permission by H. Drachsler

Such a framework helps thinking about the various "components" that make up the learning analytics topic.

Tanya Elias learning analytics model

Tanya Elias (2011) in her Learning Analytics: Definitions, Processes and Potential argues that learning analytics can be matched to Baker's (2007) knowledge continuum, which somewhat could be compared to learning level models, such as the Bloom taxonomy.

  • Data - Obtain Raw Facts
  • Information - Give Meaning to Obtained Data
  • Knowledge - Analyze and Synthesize Derived Information
  • Wisdom - Use Knowledge to Establish and Achieve Goals.

In particular, a comparison of Five Setps of Analytics (Campbell and Oblinger, 2008), Web Analytics Objectives (Hendricks, Plantz and Pritchard, 2008), Collective Applications Model (Dron and Anderson, 2009) lead to seven related Processes of Learning Analytics as shown in the following table:

Comparison of Analytics frameworks and models (Tanya Elias, 2011) / Layout changed by DKS
Knowledge Continuum Five Steps of Analytics Web Analytics Objectives Collective Applications Model Processes of Learning Analytics (Elias).
Data Capture Define goals
Measure
Select
Capture
Select
Capture
Information Report Measure Aggregate Aggregate
Report
Knowledge Predict Measure Process Predict
Wisdom Act
Refine
Use
Share
Display Use
Refine
Share

According to the author (Elias:11) learning analytics consist of computer, people, theory and organizations. “Learning analytics uses of four types of technology resources to complete ongoing pattern of three-phase cycles aimed at the continual improvement of learning and teaching” (Elias, 2011:17).

Learning analytics continuous improvement cycle. © Tanya Elias reproduced with permission by T. Elias

Discussion

This model has the merit to explicitly state that learning analytics is a complex iterative process. In particular, being able to put analytics to good use is not the same as producing reports or predictions. By the way, I wouldn't associate "knowledge" exclusively with prediction in the overview table. The latter refers to just one kind of academic knowledge. For participants (learners and teachers) it is much more important to create a deep understanding what "is going on". Also interesting is the "sharing" part.

In addition, we believe that interpretation of data should be negotiated, i.e. we also would add a sharing tab between "information processing" and "knowledge application". A simple example can illustrate that. A Mediawiki analytics tool such as StatMediaWiki can show that some students actively use the discussion pages and others do not. It turns out that those who don't can be either weak students or very good ones. The latter meet face to face on a regular basis. In other words, one has to present the data to students and discuss with them.

Ferguson and Buckingham Shum's five approaches to social learning analytics

Ferguson and Buckingham Shum (2012)'s Social Learning Analytics: Five Approaches defines five dimensions of social learning for which one could create instruments:

  • social network analytics — interpersonal relationships define social platforms
  • discourse analytics — language is a primary tool for knowledge negotiation and construction
  • content analytics — user-generated content is one of the defining characteristics of Web 2.0
  • disposition analytics — intrinsic motivation to learn is a defining feature of online social media, and lies at the heart of engaged learning, and innovation
  • context analytics — mobile computing is transforming access to both people and content.

(Ferguson and Buckingham Shum, Preprint 2012), retrieved April 10 2012.)

Work on a system is in progress.

Technology for supporting Collaborative learning

Soller, Martinez, Jermann and Muehlenbrock (2004, 2005:Abstract) develop a collaboration management cycle framework that distinguishs between mirroring systems, which display basic actions to collaborators, metacognitive tools, which represent the state of interaction via a set of key indicators, and coaching systems, which offer advice based on an interpretation of those indicators.

“The framework, or collaboration management cycle is represented by a feedback loop, in which the metacognitive or behavioral change resulting from each cycle is evaluated in the cycle that follows. Such feedback loops can be organized in hierarchies to describe behavior at different levels of granularity (e.g. operations, actions, and activities). The collaboration management cycle is defined by the four following phases” (Soller et al., 2004:6):

The Collaboration Management Cycle, by A. Soller, A. Martinez, P. Jermann & M. Muehlenbrock (2004). ITS 2004 Workshop on Computational Models of Collaborative Learning
  • Phase 1: The data collection phase involves observing and recording the interaction of users. Data are logged and stored for later processing.
  • Phase 2: Higher-level variables, termed indicators are computer to represent the current state of interaction. For example, an agreement indicator might be derived by comparing the problem solving actions of two or more students, or a symmetry indicator might result from a comparison of participation indicators.
  • Phase 3: The current state of interaction can then be compared to a desired model of interaction, i.e. a set of indicator values that describe productive and unproductive interaction states. “For instance, we might want learners to be verbose (i.e. to attain a high value on a verbosity indicator), to interact frequently (i.e. maintain a high value on a reciprocity indicator), and participate equally (i.e. to minimize the value on an asymmetry indicator).” (p. 7).
  • Phase 4: Finally, remedial actions might be proposed by the system if there are discrepancies.
  • Soller et al. (2005:) add a phase 5: “After exiting Phase 4, but before re-entering Phase 1 of the following collaboration management cycle, we pass through the evaluation phase. Here, we reconsider the question, “What is the final objective?”, and assess how well we have met our goals”. In other words, the "system" is analysed as a whole.

Depending on the locus and amount amount of computer-support, the authors then identify three kinds of systems:

  • Mirroring tools automatically collect and aggregate data about the students’ interaction (phases 1 and 2 in Figure 1), and reflect this information back to the user with some kind of visualization. Locus of processing thus remains in the hand of learners or teachers.
  • Metacognitive tools display information about what the desired interaction might look like alongside a visualization of the current state of indicators (phases 1, 2 and 3 in the Figure), i.e. offer some additional insight to both learners and teachers.
  • Guiding systems perform all the phases in the collaboration management process and directly propose remedial actions to the learners. Details of the current state of interaction and the desired model may remain hidden.

Let's now examine an example: Anna De Liddo et al. (2011:4) present a research that uses a structured argumentation tool - the Cohere web application - as research vehicle. “Following the approach of structured deliberation/argument mapping, Cohere renders annotations on the web, or a discussion, as a network of rhetorical moves: users must reflect on, and make explicit, the nature of their contribution to a discussion.”. Since the tool allows participants (1) to tag a post with a rhetorical role and (2) to link posts or posts with participants, the following learning analytics can be gained per learner and/or per group (idem: 6):

  • Learners’ attention: what do learners focus on? What problems and questions they raised, what comments they made, what viewpoints they expressed etc.
  • learners’ rhetorical attitude to discourse contributions: With what and who do a learner agrees/disagrees? What ideas he supports? What data he questioned?
  • learning topics distribution: What are the hottest learning topics, by who they have been proposed and discussed?
  • learners’ social interactions: How do learners act within a discussion group? What are the relationships between learners?

From that data, more high level constructs can be derived, such as

  • analyzing learner's attention
  • learner's moves, e.g. learner's rhetorical roles, ability to compare, or ability to broker information (between others)
  • topics distribution, e.g. most central concepts in a group discussion
  • social interactions, e.g. learner's with the highest centrality

Technology overview

Since "Analytics" is a term that can include any kind of purpose as well as any kind of activity to observe and analyze, the set of technologies is rather huge and being developped in very different contexts. Recent interest in learning analytics seems to be sparked by the growing importance of on-line marketing (including search engine optimization) as well as user experience design of e-commerce sites. However, as we tried to show in our short discussion of CSCL technology, mining student "traces" has a longer tradition.

The categories we describe below overlap to a certain extent.

The technical side of analytics include:

  • Creating data for mining (optional) either automatically or manually by the users
  • Collecting data from many sources, e.g. log files or web contents
  • Aggregating, clustering, etc.
  • All sorts of relational analysis
  • Visualization of results (and raw data aggregations)

Data mining

(Romero&Ventura, 2007) according to Baker & Yacef (2009) identifies the following types of educational data mining

  1. Statistics and visualization
  2. Web mining
    1. Clustering, classification, and outlier detection
    2. Association rule mining and sequential pattern mining
    3. Text mining

Baker & Yacef (2009) then summarize a new typology defined in Baker (2010):

  1. Prediction
    • Classification
    • Regression
    • Density estimation
  2. Clustering
  3. Relationship mining
    • Association rule mining
    • Correlation mining
    • Sequential pattern mining
    • Causal data mining
  4. Distillation of data for human judgment
  5. Discovery with models

Data mining relies on several types of sources:

  • Log files (if you have access)
  • Analytics databases filled with data from client-side JavaScript code (user actions such as entering a page can be recorded and the user can be traced through cookies)
  • Web page contents
  • Data base contents
  • Productions (other than website), e.g. word processing documents
  • ...

Types of analytics that can be obtained

  • quality of text
  • richness of content
  • content (with respect to some benchmark text)
  • similarity of content (among productions)
  • etc. (this list needs to be completed)

Trace analysis

Probably very close to data mining, but a slightly distinct community. In earlier publications, in particular in the AI&Ed tradition, papers did not use the term analytics but referred to "traces" and "student modeling". E.g. see the publications from the SILEX group at LIRIS (University of Lyon), e.g. Settouti (2009). Theses systems collect data from specially crafted log files (or databases) that belong to the application.

(to be expanded...)

client-side analytics

By client-side analytics we identify tools like user-side JavaScript programs that add functionality to a web browser. E.g. SNAPP is a JavaScript booklet that allows to conduct a social network analysis of a forum page, i.e. its discussion threads.

A similar technique would be the Firefox Greasemonkey add-on. When last checked, we didn't find any SNAPP-like tools, but a few dozen scripts that will modify or enhance Google Analytics, e.g. filter out referrers from social media sites.

A different technique (which can't be called client-side analytics) concerns web widgets that owners of web pages can insert. These then can collect data and send it to a server. A typical example is the little "Share" widget to the left that uses the well known addThis service. For example, one can imagine that teachers could require students of a class to share all visited pages on a common service.

Special purpose built-in logging/analytics

Much educational software has built-in tools for collecting data. Data collected is principled, i.e. designed to be used either in education or research or both. Logging of user actions has two purposes:

  1. Research in educational technology and related fields about learning, research in related fields that use education for data (e.g. computer science).
  2. Most systems also provide data to the learner and the teacher in order to enhance learning and teaching processes.
  • Educational workflow systems such as LAMS (Dalziel, 2003) allow to monitor student progress with respect to activity completion, time spent on task, etc. LAMS also can export student productions as portfolio and then allows to look at student's productions and performance within various tools.
  • Intelligent tutoring systems, gone out of favor in the 1990's but still alive, include by define analytics. However, its purpose is to enhance the individual learner's process by various means, e.g. diagnosing and adapting.
  • As principle, most systems (initially) developed by research grants include data logging that then can be analysed by various analysis techniques. A good example is CoFFEE

Software and systems

According to Wikipedia, retrieved 14:06, 2 March 2012 (CET). “Much of the software that is currently used for learning analytics duplicates functionality of web analytics software, but applies it to learner interactions with content. Social network analysis tools are commonly used to map social connections and discussions”

None of the software listed below comes even close to the goals of SoLAR framework. We roughly can distinguish between software that gathers (and optionally displays) data in very specific applications, software that mines the web and finally general purpose analysis/visualization software.

Ali, Hatala, Gašević & Jovanović (2011), Baker & Yacef (2009) and Asadi, Jovanović, Gašević & Hatala (2011) provide overviews of various learning analytics tools. We started looking a some of these systems. Some are not available and most others only seem work within very specific environments, although they may have been designed in a more general spirit. From what we understand right now (that is little), there exist currently very few interesting tools that could be used by teachers (or even somewhat tech-savy educational technologist). In particular, ready to use tools for innovative learning designs that engage students in various forms of writing seems to be almost totally missing. E.g. most creative teachers use wikis in one or another way and indeed could use a whole lot of analytics. Read more about that in the wiki metrics, rubrics and collaboration tools.

It will take me some time to get a clearer picture - Daniel K. Schneider 19:41, 15 March 2012 (CET).

General purpose tools

A variety of tools made for other purposes is being used for educational analytics. Some of these have been developed for use in education. Below we list a few.

Web analytics tools

Most of these require you to insert a JavaScript snippet in pages. That way some web site can track users. E.g. in this wiki, we use Google analytics (where users remain anonymous)

Web analytics tools are both used for Search engine optimization, Interaction design, user experience and usability studies.

Besides Google Analytics, there are many other tools. Most are commercial, but minimal services are often free. E.g. Bing WebMaster tools, Search Metrics, SEMRush, Yahoo site explorer, Userfly, Tynt, Kissmetrics, Mixpanel.

Social network analysis

See social network analysis

Two popular ones are:

  • SNAPP, Social Networks Adapting Pedagogical Practice (Dawson, 2009), is “a software tool that allows users to visualize the network of interactions resulting from discussion forum posts and replies. The network visualisations of forum interactions provide an opportunity for teachers to rapidly identify patterns of user behaviour – at any stage of course progression. SNAPP has been developed to extract all user interactions from various commercial and open source learning management systems (LMS) such as BlackBoard (including the former WebCT), and Moodle.”, retrieved 14:06, 2 March 2012 (CET). SNAPP is implemented as JavaScript booklet that you can just drag on a brower's toolbar. It works with Firefox, Chrome and Internet explorer and for the visualization to work, you need Java installed. In Moodle (tested), this tool can visualize forum threads that are displayed hierarchically. Data also can be exported for further analysis.
  • Pajek, is a program, for Windows, for analysis and visualization of large networks.

Discussion/forum analysis tools

  • The ACODEA Framework (Automatic Classification of Online Discussions with Extracted Attributes) by Mu et al. (2012). This is not a tool, but a configuration and use scenario for the SIDE tool (Mayfield and Rosé, 2010a).

Awareness tools

So-called awareness tools provide real-time analytics. According to Dourish and Belloti (1992), awareness is an understanding of the activities of others, which provides a context for your own activity. Gutwin and Greenberg (1996:209) define eleven elements for workspace awareness, for example: presence (who is participating in the activitiy ?), location (where are they working ?), activity level (how active are they in the workspace ?), actions (what are they doing ? what are their current activites and tasks ?), and changes (what changes are they making, and where ?) Note that "where" can refer both to a real and virtual location (e.g. a living room or a forum in a learning platform).

Real-time awareness tools are popular in CSCL system, bust mostly absent from various forms of virtual learning environments, such as LMSs, content management systems and wikis

Text analytics (heavy data mining)

  • TADA-Ed (Tool for Advanced Data Analysis in Education) combines different data visualization and mining methods in order to assist educators in detecting pedagogically important patterns in students’ assignments (Merceron & Yacef, 2005). We don't know if this project is alive. The CHAI team at the University of Sydney does have other more recent educational data mining projects. - Daniel K. Schneider 19:41, 15 March 2012 (CET)
  • LightSide will be an online service for automated revision asistance and classroom support. It also is available as free software that can be download, but that must be trained by yourself.

See also:

web scraping, visualization and app building software

Various general purpose tools could be used, such as:

These tools are combinations of page scrapping (content extraction), text summarization a comparison), sorting, and visualization. Their scope and ease of use seems to differ a lot.

See also:

Data visualization software

Some online tools:

Software:

  • OpenDX (not updated since 2007)
  • NetDraw a Windows program for visualizing social network data. (Borgatti, 2002).

See also: Visualization and social network analysis

Harvesting data from badges systems

Educational badges systems that are managed through some kind of server technology can be used to harvest information about learner achievements, paths, skills, etc.

For example, “Mozilla's Open Badges project is working to solve that problem, making it easy for any organization or learning community to issue, earn and display badges across the web. The result: recognizing 21st century skills, unlocking career and educational opportunities, and helping learners everywhere level up in their life and work.” (About, retrieved 15:37, 14 March 2012 (CET).

See also: gamification

Other tracking tools

  • Most research systems built in educational technology, do have built-in tracking technology, see for example the CoFFEE system
  • e-science environments have access to a wide range of data. I don't know if any of these could be of use for learning analytics (except if students participate in an e-science project)
  • Some citizen science projects do have at least some light-weight analytics with respect to user participation
  • We don't have neither the time nor the resources to list tools developed for other domains (business analytics in particular). An interesting non-business initiative is Global pulse, a UN project. Part of Global Pulse’s approach is learning how to distill information from data so we can identify insights that can be useful in program planning and, in the long run, can contribute to shaping policy. Tools developed include for example "Twitter and perceptions of crisis-related stress" and "impact of financial crisis on primary schools".

Tracking in e-learning platforms

By default, most so-called learning management systems have built-in tracking tools. In addition, there may be extra tools, either developed for research or available as additional modules in commercial systems. Such tools may help to improve traditional e-learning designs.

LOCO Analyst

LOCO-Analyst “is implemented as an extension of Reload Content Packaging Editor, an open-source tool for creating courses compliant with the IMS Content Packaging (IMS CP) specification. By extending this tool with the functionalities of LOCO-Analyst, we have ensured that teachers effectively use the same tool for creating learning objects, receiving and viewing automatically generated feedback about their use, and modifying the learning objects accordingly.” (LOCO-Analyst, retrieved 14:06, 2 March 2012 (CET)).

As of March 2012, the systems works with iHelp Courses, an open-source standards-compliant LCMS developed at the University of Saskatchewan.

Features:

  • In LOCO-Analyst, “feedback about individual student was divided into four tab panels: Forums, Chats, Learning, and Annotations [...]. The Forums and Chats panels show student’s online interactions with other students during the learning process. The Learning panel presents information on student’s interaction with learning content. Finally, the Annotations panel provides feedback associated with the annotations (notes, comments, tags) created or used by a student during the learning process.” (Ali et al. 2001). For each of the four "areas" key data can then be consulted.
  • “The generation of feedback in LOCO-Analyst is based on analysis of the user tracking data. These analyses are based on the notion of Learning Object Context (LOC) which is about a student (or a group of students) interacting with a learning content by performing certain activity (e.g. reading, quizzing, chatting) with a particular purpose in mind. The purpose of LOC is to facilitate abstraction of relevant concepts from user-tracking data of various e-learning systems and tools.” (LOCO-Analyst, retrieved 14:06, 16 March 2012).
  • LOCO Analyst is implemented through a set of formal RDF ontologies, i.e. the LOCO (Learning Object Context Ontologies) ontological framework
  • download

Virtual location-based meta-activity analytics

Jaillet (2005) presents a model with three indicators for measuring learning in the Acolad distance teaching platform. These indices are are based on Leontiev's activity theory that distinguishes three levels: Activities are high level and they are goal and needs driven. Activities are conducted using strategies and performing actions using tools (instruments). Actions are performed using operators. In the context of education, both actions and operations can be either individual or collaborative (social).

The study measures student activity (called meta-activity) through three dimensions: attendance, availability and involvement: All indicators are measured with respect to the most active student.

  • Attendance: Number of connections (% of max.)
  • Availability: Duration of connection (% of max.)
  • Involvement: Number of actions performed (% of max., e.g. upload a document, go to virtual place, participate in group discussion)

Example:

653(61;84;42)2629480 means that a student connected 61% of 653 (with respect to the student that has the highest connection), during 84% of 480 hours and that he completed 42% operations out of 2692 (max.).

It is not very clear from the paper whether the data leading to "meta-activity" indicators are supposed to measure actions or operations. The study also includes an other interesting idea.

Since Acolad uses a 2D spatial metaphor, i.e. 3D drawings, for various learning activities, one can guess from the student's location what they are working on. There is evidence from research in CSCW that people do tend to relocated in virtual environment. In other words, learning analytics should provide better data from environments that associate learning activities with a space.

Blackboard Analytics for learn

  • The Blackboard Analytics for Learn™ module is advertized as “easy, self-service access to data that can help give you an enterprise-level perspective. Gain insight into user activity, course design and student performance data across departments and colleges — enabling you to improve your use of the Blackboard Learn™ platform in support of teaching, learning, and student success.” (retrieved April 5, 2012).

According to the same source, one could get answers to questions like:

  • How are my students performing on learning standards over time?
  • How can I easily find students who aren’t engaged in their online courses?
  • Which tools are being used in courses the most? The least?
  • How many logins, page views, and other metrics have occurred over time?
  • Who are the most active instructors? Who are the least active?
  • What strategies to improve the quality of course design and instruction result in better student performance? Instructor training courses? Course reviews?

Moodle analytics with MocLog

Quote: Teachers, students, study program managers and administrators need feedback about the status of the activities in online-courses. MOCLog will realize a monitoring system, that helps to analyze logfiles of the Moodle LMS more effectively and efficiently thus contributing to improve the quality of teaching and learning.

List of other tools to explore

  • Purdue Course Signals combines predictive modeling with data-mining from Blackboard Vista. This software has been bought by Sunguard (now Ellucian), a supplier of campus software. “Course Signals’ advisor view helps your institution’s professional and faculty advisors have a better window into how a student is performing in their current courses, making it easy for them to identify and reach out to students who are struggling in multiple courses, have risk factors such as low overall GPAs, or are first-generation status or transfer students. A real-time view of this data means advisors can act more immediately to help students before they abandon a program of study, or worse, leave school discouraged and unprepared for their futures” (Redefining “early” to support student success, retrieved April 10 2012). See also its PowerCAMPUS Analytics and Datatel Colleague
  • WCET Predictive Analytics Reporting (PAR) Framework. Aim: to conduct large-scale analyses of federated data sets within postsecondary institutions to better inform loss prevention and identify drivers related to student progression and completion. This project (as of April 2012) seems to be in planning stage.
  • Teacher ADVisor (TADV): uses LCMS tracking data to extract student, group, or class models (Kobsa, Dimitrova, & Boyle, 2005).
  • Student Inspector: analyzes and visualizes the log data of student activities in online learning environments in order to assist educators to learn more about their students (Zinn & Scheuer, 2007).
  • CourseViz: extracts information from WebCT.
  • Learner Interaction Monitoring System (LIMS): (Macfadyen & Sorenson, 2010)
  • ASTRO and BBStats, some kind of Blackboard plugins (?) that provide a dashboard.

Classroom tools

Tools like Metrix (working with any mobile device) allow teachers to define a skill set and then enter student data that allow to "track, analyze and differentiate" students.

General purpose educational tracking tools

Some institutions did or will implement tracking tools that pull in data from several sources. Some tools are designed for both teachers and learnings, others are in the realm of academic analytics (as opposed to learning analytics).

UWS Tracking and Improvement System for Learning and Teaching (TILT)

The University of Western Sydney implemented a system with the following uses in mind (Scott, 2010).

  1. To prove quality – e.g. data can be drawn from different quantitative and qualitative data bases and, after triangulation, can be used to prove the quality of what UWS is doing. Data analysis can use absolute criteria and/or relative (benchmarked) standards.
  2. To improve quality – e.g. data can be used to identify key improvement priorities at the university, college, division, course and unit levels.

This framework uses about ten data sources, mostly surveys (student satisfaction, after graduation) and peformance reports. Interestingly, it includes time series, e.g. student satisfaction during, at exit and after obtaining a degree.

Course quality assurance at QUT

Towers (2010) describes Queensland University of Technology (QUT) annual Courses Performance Report as a combination of Individual Course Reports (ICRs), the Consolidated Courses Performance Report (CCPR), Underperforming Courses Status Update and the Strategic Faculty Courses Update (SFCU). “Individual Course Reports (ICRs) is to prompt an annual health check of each course’s performance, drawing upon course viability, quality of learning environment, and learning outcomes data. In total, data on 16 indicators are included in the ICRs.” (Towers, 2010: 119). The paper include an example of an 8-page report in the appendix.

Tools for power teaching

Courses that aim at deep learning, e.g. applicable knowledge and/or higher order knowledge are usually taught in smaller classes most often uses software that is distinct from the ones used in so-called e-learning, e.g. wikis, learning e-portfolios, content management systems, blogs, webtops and combinations of these. In addition, some systems developed in educational technology research are being used. The latter often implement analytics for research, but often there is a teacher interface using the same data. Most creative power teaching, however, uses what we would call "street technology" and for these, analytics tool rarely exist. A good example would be the Enquiry Blog Builder we shortly describe below. Another example is StatMediaWiki, a tool that we use to learn about student contributions in the french version of this wiki. Most analytics tools we found so far, require some learner participation like tagging or filling in shorter or longer questionnaires.

Since project-oriented learning and related constructivist designs do rarely include "pages to view" and quizzes, simple quantitative data do not provide good enough analytics, although some quantitative data such as "number of edits", "number of connections", etc. are useful. Some teachers, in particular in the engineering sciences also use project management software that by definition include analytics. However these are less suited for writing, collaboration and creating communities of learning than typical out-of-the-box multi-purpose portalware.

The main issue of what we would call "small scale analytics" is not how to measure learning, but how to improve learning scenarios through tools that help both the learner and the teacher understand "what is going on" and therefore improve reflection.

Analytics for e-portfolios

Oliver and Whelan (2010) point out that “Many agree that current available data indicators - such as grades and completion rates - do not provide useable evidence in terms of learning achievements and employability (Goldstein, 2005; Norris, Baer et al. 2008)” and they argue that “qualitative evidence of learning is traditionally housed in student portfolio systems”. Such systems then should include mechanisms for self-management of learning that could feed into institutional learning analytics. At Curtin, to encourage students to reflect on and assess their own achievement of learning, the iPortfolio incorporates a self-rating tool based on Curtin's graduate attributes, “enabling the owner to collate evidence and reflections and assign themselves an overall star-rating based on Dreyfus and Dreyfus' Five-Stage Model of Adult Skill Acquisition (Dreyfus, 2004). The dynamic and aggregated results are available to the user: [as shown in the figure], the student can see a radar graph showing their self-rating in comparison with the aggregated ratings of their invited assessors (these could include peers, mentors, industry contacts, and so on).” (Oliver & Wheelan:2010).

This example as well as De Liddo et al. (2011) tagging system discussed above demonstrates a very simple principle that one could implement in any other environment: (Also) ask the learner.

Student Activity Monitor and Cam dashboard for PLEs

The SAM (Student Activity Monitor) was developed in the context of the EU ROLE project. “To increase self-reflection and awareness among students and for teachers, we developed a tool that allows analysis of student activities with different visualizations”. Although SAM was developed in a project that focused on developing a kit for constructing personal learning environments it can be used in any other context where the learning process is largely driven by rather autonomous learning activities. SAM is implemented as Flex software application and it is documented in the sourceforge ROLE project wiki as Widget Student Activity Meter. It relies on a widget that users must use: the Widget CAM Widget that enables/disables user monitoring. Enabling the Monitoring provides learners with benefits regarding recommendation and self-reflection mechanisms.

Its four purposes are (Govaerts et al., 2010):

  • Self-monitoring for learners
  • Awareness for teachers
  • Time tracking
  • Learning resource recommendation

The Cam dashboard is a an application that enables the visualization of usage data as a basis to detect changes in usage patterns. The purpose is to detect variations in the use of PLEs based on changes in usage patterns with widgets and services.

Links:

Download:

  • So far we didn't find any .... - Daniel K. Schneider 19:34, 11 April 2012 (CEST) (but maybe we will have to spend some more time searching or download the software kit).

The EnuiryBlogger

Enquiry Blog Builder provides a series of plugins for the very popular Wordpress bloggingware (Ferguson, R., Buckingham Shum, S. and Deakin Crick, R., 2011). As of April 2012, the list includes:

  • MoodView - This displays a line graph plotting the mood of the student as their enquiry progresses. The widget displays the past few moods and allows a new one to be selected at any time. Changing moods (a hard coded drop-down list from 'going great' to 'it's a disaster') creates a new blog entry with an optional reason for the mood change. The graph is created using the included Flot JavaScript library.
  • EnquirySpiral - This widget provides a graphical display of the number of posts made in the first nine categories. A spiral of blobs appears over an image with each blob representing a category. The blobs are small and red when no posts have been made. They change to yellow for one or two posts, and green for three or more. In this way it is easy for the student to see how they are progressing, assuming the nine categories are well chosen.
  • EnquirySpider - This widget works in the same way as the EnquirySpiral, except that the blobs are arranged in a star shape. They are associated with seven categories. (from nine to sixteen so they don't conflict with the EnquirySpiral). The categories are intended to match with the Effective Lifelong Learning Inventory diagram.

(Enquiry Blog Builder, retrieved April 10 2012).

The kit also includes BlogBuilder allowing batch creation of blogs in one go with provided teacher and student names. Teachers who login will then see the a dashboard showing the progress of the students assigned to them. In other words, the teacher will get monitoring widgets that correspond to the three widgets above. These will show the student widgets.

Learner disposition

Buckingham Shum and Deakin Crick (2012) describe a setup that collects learning dispositions through a questionnaire. Results then are shown to the students and teachers. “The inventory is a self-report web questionnaire comprising 72 items in the schools version and 75 in the adult version. It measures what learners say about themselves in a particular domain, at a particular point in time” (Preprint: page 4).

The seven measured dimensions are:

  • Critical curiosity
  • Meaning Making
  • Dependence and Fragility
  • Creativity
  • Learning Relationships
  • Strategic Awareness

Students are shown a spider diagram of these dimensions. Since ELLI can be applied repetitively, a diagram also can display change, i.e. (hopefully) increase. Furthermore data can be aggregated across groups of learners in order to provide collective profiles for all dimensions or details about a specific dimension.

Other writing tools and cognitive tools with built-in Analytics

There are many writing-to-learn environments, other cognitive tools, knowledge building environments, etc. allowing to co-create knowledge and that do include some information about "what is going on". Some are developed for education, others for various forms of knowledge communities, such as communities of practice

Examples:

  • Cohere from Open University is a visual tool to create, connect and share Ideas. E.g. de Liddo et al. (2011) analyse quality of contribution to discourse as an indicator for meaningful learning.

Discussion

Small vs. big analytics

Small analytics (individual learners, teachers and the class) are not new, although we hope that more tools that will help us understanding what is going on will be available in the future. Such tools are mostly lacking for environments that engage students in rich activities and productions (as opposed to so-called e-learning that uses learning management systems). In project-oriented learning designs, we also should think about building tools into the scenario that let the students add data, e.g. "what did I contribute" and "why".

Big time analytics should carefully think about strategy, i.e. ways of thinking about improving education. In particular we believe it important that one should not automatically focus on alignment as in a balanced scorecard method and rather adopt a "learning organization" perspective, e.g. adopt some kind of expansive learning philosophy. Below is a picture from a talk by Stefano Mastrogiacomo that illustrates various way of thinking.

Catalogue of strategies (© S. Mastrogiacomo)

.

Of course, discussion about the aims of learning analytics already takes place, in particular ethical issues (refs?). However, we suggest to focus on the idea that Change should be related to learner and teacher empowerment as opposed to strengthening the already sometimes counterproductive education bureaucracy.

A list of questions

Items to expand (also should merge some stuff above)

  • Learning analytics is an old topic in educational technology. Some CSCL research already was mentioned. Much older artificial intelligence and education always had some kind of analytics for both the learner and the teacher built-in. Adaptive systems have been produced by the dozen and failed to interest teachers. An educationally more light-weight variant of tutoring systems are so-called recommended systems. These may become more popular, although attention must be payed of not atttempting to lock students into a box. E.g. one could image adding 20% of "noise" (i.e. recommendation that are rather far as opposed to close). That's how genetic algorithms work.
  • New in learning analytics are the ability to track students across environments (e.g. all social services) and the interest for using analytics as evaluation tool az institution level and above. This leads to privacy issues but not just. Analytics can have intrinsic adverse effects. See the next item.
  • Analytics in other fields often had an adverse effect. In public policy, monitoring and reporting tool had the effect to blow up procedures, kill initiative, remove communication, etc. Analytics can (but must not) lead to absolute disasters both in terms of both cost and effectiveness. Reliance on key indicators and reporting can bring out the dark side of management: “toxic leadership, managerial worst practices, project failures and suffering at work.” (Stefano Mastrogiacomo) and increase the power of bureaucracy at the expense of individual creativity and communication. In education that could mean: Teachers will have to use mind crippling e-learning platforms (else the organization can't collect data), participants will learn how to produce data for the system (as opposed to produce meaning) and participants will talk less to each other (since time has to be spend on collecting and archiving data).
  • Lack of tools. Use of data mining and visualization tools need month of training (at least)
  • Why don't teachers ask more information from the learners (self-reporting)? Many tools allow students to tell what they did and in addition to reflect upon, and teachers can require it. For example, LAMS activities can be configured with a (mandatory) notebook to be filled in at the end. Wiki's such as this one include personal home pages where students can add links to their contributions. Actually, tools we found for project-oriented designs such as inquiry learning or portfolio construction do this.
  • Are there analytics that can distinguish between development of routine expertise, as opposed to adaptive expertise (new schemas that allow to respond to new situtations) ? (Alexander 2003, cited by Werner et al. 2013)).

Links

Literature

Overviews and bibliographies

Journals

  • JEDM - Journal of Educational Data Mining

Organizations and communities

Organizations and conferences

Web sites

Communities

Training

Free online courses

Talks

Tools

Also see links in this text (above)

Ideas

Bibliography

We split it into sections with the idea that this article may become separated some day. It's getting longish ...

Technical

Analytics and data mining

Acknowledgement: This bibliography was started from:

Entries

  • Ali, L., Hatala, M. Gašević, D., Jovanović, J., "A Qualitative Evaluation of Evolution of a Learning Analytics Tool," Computers & Education, Vol. 58, No. 1, 2012, pp. 470-489 Abstract/HTML/ (Science direct)
  • Arnold, K. E. (2010). Signals: Applying Academic Analytics, EDUCAUSE Quarterly 33(1). HTML
  • Atkisson, Michael and David Wiley. 2011. Learning analytics as interpretive practice: applying Westerman to educational intervention. In Proceedings of the 1st International Conference on Learning Analytics and Knowledge (LAK '11). ACM, New York, NY, USA, 117-121. DOI=10.1145/2090116.2090133
  • Anjewierden, Anjo and Lilia Efimova. Understanding weblog communities through digital traces: a framework, a tool and an example. In “International Workshop on Community Informatics (COMINF 2006)”, Robert Meersman, Zahir Tari, and Pilar Herrero (eds.), pp. 279–289, LNCS 4277, Springer, Montpellier, 2006 (November).
  • Anjewierden, Anjo; Rogier Brussee and Lilia Efimova. Shared conceptualisations in weblogs, in ”BlogTalks 2.0: The European Conference on Weblogs”, Thomas N. Burg (ed.), Danube, University of Krems, 110–138, 2005.
  • Aviv, R., Erlich, Z., Ravid, G., & Geva, A. (2003). Network analysis of knowledge construction in asynchronous learning networks. Journal of Asynchronous Learning Networks, 7(3), 1-­‐23.
  • Baepler, P. M., Cynthia James. (2010). Academic Analytics and Data Mining in Higher Education. International Journal for the Scholarship of Teaching and Learning, 4(2).
  • Baker, R.S.J.d. (2010) Data Mining for Education. In McGaw, B., Peterson, P., Baker, E. (Eds.) International Encyclopedia of Education (3rd edition), vol. 7, pp. 112-118. Oxford, UK: Elsevier. PDF Reprint
  • Baker, R.S.J.d. (2010) Mining Data for Student Models. In Nkmabou, R., Mizoguchi, R., & Bourdeau, J. (Eds.) Advances in Intelligent Tutoring Systems, pp. 323-338. Secaucus, NJ: Springer. PDF Reprint
  • Baker, R.S.J.d., Yacef, K. (2009) The State of Educational Data Mining in 2009: A Review and Future Visions. Journal of Educational Data Mining, 1 (1), 3-17. PDF
  • Baker, R.S.J.d., de Carvalho, A. M. J. A. (2008) Labeling Student Behavior Faster and More Precisely with Text Replays. Proceedings of the 1st International Conference on Educational Data Mining, 38-47.
  • Blikstein, Paulo (2011). Using learning analytics to assess students' behavior in open-ended programming tasks. In Proceedings of the 1st International Conference on Learning Analytics and Knowledge (LAK '11). ACM, New York, NY, USA, 110-116. DOI=10.1145/2090116.2090132
  • Borgatti, S.P., 2002. NetDraw Software for Network Visualization. Analytic Technologies: Lexington, KY
  • Buckingham Shum, S. and Deakin Crick, R. (2012). Learning Dispositions and Transferable Competencies: Pedagogy, Modelling and Learning Analytics. Proc. 2nd International Conference on Learning Analytics & Knowledge. (29 Apr-2 May, 2012, Vancouver, BC). ACM Press: New York. Eprint: http://oro.open.ac.uk/32823
  • Campbell, John P.; Peter B. DeBlois, and Diana G. Oblinger (2007). Academic Analytics: A New Tool for a New Era, EDUCAUSE Review, vol. 42, no. 4 (July/August 2007), pp. 40–57, http://www.educause.edu/library/erm0742.
  • Champalle, O; K. Sehaba, A. Mille, Y. Prié (2011) A framework for observation and analysis of learners’ behavior in a full-scope simulator of a nuclear power plant - Approach based on modelled traces. ICALT 2011 - The 11th IEEE International Conference on Advanced Learning Technologies, Athens, Georgia, USA. pp. 30-31. ISBN 978-1-61284-209-7. ISSN 2161-3761. PDF
  • Coates, H. (2010). New directions in quality management. In C. S. Nair, L. Webster & P. Mertova (Eds), Leadership and management of quality in higher education. Oxford, UK: Chandos Publishing: 169-186.
  • Dawson, S. (2008). A study of the relationship between student social networks and sense of community, Educational Technology and Society 11(3), pp. 224-38. PDF
  • Dawson, S. M., Erica. (2008). Investigating the application of IT generated data as an indicator of learning and teaching performance. Wollongong: Australian Learning and Teaching Council.
  • Dawson, S., Heathcote, L. and Poole, G. (2010), “Harnessing ICT Potential: The Adoption and Analysis of ICT Systems for Enhancing the Student Learning Experience,” The International Journal of Educational Management, 24(2), pp. 116-129.
  • Dawson, S., Heathcote, L. and Poole, G. (2010). Harnessing ICT potential: The adoption and analysis of ICT systems for enhancing the student learning experience, International Journal of Educational Management 24(2) pp. 116-128.
  • Dawson, S., McWilliam, E., & Tan, J. P.-L. (2008). Teaching Smarter: How mining ICT data can inform and improve learning and teaching practice. Paper presented at the ASCILITE 2008. PDF
  • Dawson, Shane. 2010. 'Seeing' the Learning Community: An Exploration of the Development of a Resource for Monitoring Online Student Networking. British Journal of Educational Technology 41(5): 736-752.
  • Dawson,. (2010) Seeing the learning community. An exploration of the development of a resource for monitoring online student networking. British Journal of Educational Technology, 41(5), 736-752. doi:10.1111/j.1467-8535.2009.00970.x.
  • De Laat, M., Lally, V., Lipponen, L., & Simons, R.J. (2006). Analysing student engagement with learning and tutoring activities in networked learning communities: a multi-method approach. International Journal of Web Based Communities, 2(4), 394-412.
  • De Laat, M.; V. Lally, L. Lipponen, R. Simons (2007). Investigating patterns of interaction in networked learning and computer-supported collaborative learning: a role for social network analysis, International Journal of Computer-Supported Collaborative Learning, 2 (1), pp. 87–103
  • De Liddo, A., Buckingham Shum, S., Quinto, I., Bachler, M. and Cannavacciuolo, L. (2011). Discourse-Centric Learning Analytics. Proc. 1st International Conference on Learning Analytics & Knowledge. Feb. 27-Mar 1, 2011, Banff. ACM Press: New York. Eprint: http://oro.open.ac.uk/25829
  • Djouad, Tarek; A. Mille, M. Benmohammed. (2011). SBT-IM: Système à base de traces-Indicateurs d'interactions Moodle. EIAH 2011, Mons. PDF
  • Djouad, Tarek; L. S Settouti, Y. Prié, C. Reffay, A. Mille. (2010). Un Système à Base de Traces pour la modélisation et l’élaboration d’indicateurs d’activités éducatives individuelles et collectives. Mise à l’épreuve sur Moodle.. TSI 29(6):721-741. PDF
  • Downs, Stephen (2010). Collaboration, Analytics, and the LMS: A Conversation with Stephen Downes, Campus Technology, October 14, 2010, retrieved 18:37, 19 March 2012 (CET).
  • Drachsler,H.,et al.(2010). Issues and Considerations regarding Sharable Data Sets for Recommender Systems in Technology Enhanced Learning .1st Workshop on Recommender Systems in Technology Enhanced Learning(RecSysTEL@EC-TEL2010)September,28,2010, Barcelona, Spain.
  • Dron, J. and Anderson, T. (2009). On the design of collective applications, Proceedings of the 2009 International Conference on Computational Science and Engineering, Volume 04, pp. 368-374.
  • Eckerson, W. W. (2006). Performance dashboards: Measuring, monitoring, and managing your business. Hoboken, New Jersey: John Wiley & Sons.
  • Elias, Tanya (2011). Learning Analytics: Definitions, Processes and Potential.Athabasca University (?) PDF
  • Ferguson, R., Buckingham Shum, S. and Deakin Crick, R. (2011). EnquiryBlogger – Using widgets to support awareness and reflection in a PLE setting. Awareness and Reflection in Personal Learning Environments: Workshop at the Personal Learning Environments Conference 2011, 11th July 2011, University of Southampton, UK PDF
  • Ferguson, R. and Buckingham Shum, S. (2011). Learning Analytics to Identify Exploratory Dialogue within Synchronous Text Chat. Proc. 1st International Conference on Learning Analytics & Knowledge. Feb. 27-Mar 1, 2011, Banff. ACM Press: New York. Eprint: http://oro.open.ac.uk/28955
  • Ferguson, Rebecca (2012). The State Of Learning Analytics in 2012: A Review and Future Challenges. Technical Report KMI-12-01, Knowledge Media Institute, The Open University, UK. Abstract/PDF.
  • Few, S. (2007). Dashboard Confusion Revisited. Visual Business Intelligence Newsletter. January 2007. PDF
  • Fournier, Helene; Kop, Rita; Sitlia, Hanan (2011). The Value of Learning Analytics to Networked Learning on a Personal Learning Environment, 1st International Conference Learning Analytics and Knowledge. PDF (from nrc-cnrc.gc.ca)
  • Greller, Wolfgang and Hendrik Drachsler (in press, 2013?), Translating Learning into Numbers: Generic Framework for Learning Analytics, Educational Technology & Society, PDF Draft
  • Goldstein, P. J. (2005) Academic Analytics: Uses of Management Information and Technology in Higher Education PDF
  • Goldstein, Philip J. and Richard N. Katz (2005). Academic Analytics: The Uses of Management Information and Technology in Higher Education, Educause Center for Applied Research (ECAR). Table of Contents - PDF full report
  • Govaerts, Sten; Katrien Verbert, Joris Klerkx, Erik Duval (2010). "Visualizing Activities for Self-reflection and Awareness", The 9th International Conference on Web-based Learning, ICWL 2010, Springer, Lecture Notes on Computer Science. Abstract/PDF
  • Hadwin, A. F., Nesbit, J. C., Code, J., Jamieson-Noel, D. L., & Winne, P. H. (2007). Examining trace data to explore self-regulated learning. Metacognition and Learning, 2, 107-124
  • Haiming, L. Macintyre, R. and Ferguson, R. (2012). Exploring Qualitative Analytics for E-Mentoring Relationships Building in an Online Social Learning Environment. Proc. 2nd International Conference on Learning Analytics & Knowledge, (29 Apr-2 May, Vancouver, BC). ACM Press: New York. PDF
  • Hanani et al.(2001)Information Filtering: Overview of Issues, Research and Systems", User Modeling and User-Adapted Interaction (11)
  • Haythornthwaite, C. (2008). Learning relations and networks in web-based communities. International Journal of Web Based Communities, 4(2), 140-158.
  • Hendricks, M., Plantz, M.C., & Pritchard, K.J. (2008). Measuring outcomes of United Wayfunded programs: Expectations and reality. In J.G. Carman & K.A. Fredricks (Eds.), Nonprofits and evaluation. New Directions for Evaluation, 119, pp. 13-35.
  • Hijon R. and Carlos, R. (2006). E-learning platforms analysis and development of students tracking functionality, Proceedings of the 18th World Conference on Educational Multimedia,Hypermedia & Telecomunications pp. 2823-2828.
  • Hübscher, R. & Puntambekar, S. (2008). Integrating knowledge gained from data mining with pedagogical knowledge. In Proceedings of the 1st International Conference on Educational Data Mining (EDM2008), 97–106. (PDF)
  • Hübscher, R., Puntambekar, S., & Nye, A. H. (2007). Domain specific interactive data mining. In Proceedings of Workshop on Data Mining for User Modeling, 11th International Conference on User Modeling, Corfu, Greece, 81–90. (PDF)
  • Johnson, L., Smith, R., Willis, H., Levine, A. & Haywood, K. (2011). The 2011 Horizon Report. Austin, Texas, The New Media Consortium. http://wp.nmc.org/horizon2011/
  • Koedinger, K.R., Baker, R.S.J.d., Cunningham, K., Skogsholm, A., Leber, B., Stamper, J. (2010) A Data Repository for the EDM community: The PSLC DataShop. In Romero, C., Ventura, S., Pechenizkiy, M., Baker, R.S.J.d. (Eds.) Handbook of Educational Data Mining. Boca Raton, FL: CRC Press, pp. 43-56.
  • Leah P. Macfadyen and Shane Dawson, “Mining LMS Data to Develop an ‘Early Warning System’ for Educators: A Proof of Concept,” Computers & Education, vol. 54, no. 2 (2010), pp. 588–599.
  • Libby V. Morris, Catherine Finnegan, and Sz-Shyan Wu, “Tracking Student Behavior, Persistence, and Achievement in Online Courses,” The Internet and Higher Education, vol. 8, no. 3 (2005), pp. 221–231.
  • Macfadyen, L. P. and Sorenson. P. (2010) “Using LiMS (the Learner Interaction Monitoring System) to track online learner engagement and evaluate course design.” In Proceedings of the 3rd international conference on educational data mining (pp. 301–302), Pittsburgh, USA.
  • Macfayden, L. P., & Dawson, S. (2010). Mining LMS data to develop an “early warning” system for educators: a proof of concept. Computers & Education, 54(2), 588–599.
    • This is an often cited text defining how EDM could be directly useful to educators.
  • Macfadyen, L. P., & Dawson, S. (2012). Numbers Are Not Enough. Why e-Learning Analytics Failed to Inform an Institutional Strategic Plan. Educational Technology & Society,15(3), 149–163. PDF.
  • Mayfield, E., & Rosé, C. (2010a). An interactive tool for supporting error analysis for text mining. Paper presented at the Proceedings of the NAACL HLT 2010 Demonstration Session (pp. 25 – 28), Los Angeles, California
  • Mazza R., and Dimitrova, V. (2004). Visualising student tracking data to support instructors in web-based distance education, 13th International World Wide Web conference, New York, NY, USA: ACM Press, pp. 154-161. PDF, retrieved 19:41, 15 March 2012 (CET).
  • Mazza, R. & V. Dimitrova. (2003) "CourseVis: Externalising Student Information to Facilitate Instructors in Distance Learning" in Proceedings of 11th International Conference on Artificial Intelligence in Education (AIED03), F. Verdejo and U. Hoppe (Eds), Sydney: IOS Press.
  • Mazza, R. B., Luca. (2007). Monitoring an Online Course with the GISMO Tool: A Case Study. Journal of Interactive Learning Research, 18(2), 15.
  • Mazza, R., & Dimitrova, V. (2007). CourseVis: a graphical student monitoring tool for supporting instructors in webbased distance courses. International Journal of Human-Computer Studies, 65(2), 125-139.
  • Merceron, A. and K. Yacef (2005c). Clustering students to help evaluate learning, volume 171 of Technology Enhanced Learning, pages 31-42. Springer,
  • Merceron, A. & K. Yacef (2003). "A Web-based Tutoring Tool with Mining Facilities to Improve Learning and Teaching" in Proceedings of 11th International Conference on Artificial Intelligence in Education., F. Verdejo and U. Hoppe (Eds), pp 201-208, Sydney: IOS Press.
  • Merceron, A. and K. Yacef (2005b). Educational data mining: a case study. In C. K. Looi, G. McCalla, B. Bredeweg, and J. Breuker, editors, Proceedings of the 12th Conference on Artificial Intelligence in Education, pages 467-474, Amsterdam, The Netherlands, 2005. IOS Press.
  • Merceron, Agathe and Kalina Yacef (2005). TADA-Ed for Educational Data Mining, Interactive Multimedia Journal of Computer-Enhanced Learning (IMEJ), http://www.imej.wfu.edu/articles/2005/1/03/, retrieved 19:41, 15 March 2012 (CET).
  • Najjar. J; M. Wolpers, and E. Duval. Contextualized attention metadata. D-Lib Magazine, 13(9/10), Sept. 2007.
  • Norris, Donald; Linda Baer, Joan Leonard, Louis Pugliese, and Paul Lefrere (2008). Action Analytics: Measuring and Improving Performance That Matters in Higher Education, EDUCAUSE Review, vol. 43, no. 1 (January/February 2008): 42–67 html
  • Oliver, Beverly and Barbara Whelan (2011). Designing an e-portfolio for assurance of learning focusing on adoptability and learning analytics, Australasian Journal of Educational Technology, 27(6), 1026-1041. HTML
  • Paulsen, Michael B. and Kenneth A. Feldman, Exploring the dimensions of the scholarship of teaching and learning: Analytics for an emerging literature, New directions for Institutional Research 129, 21-36, DOI: 10.1002/ir.169. Not learning analytics, but interesting from a technical point of view.
  • Rath, A. S., D. Devaurs, and S. Lindstaedt. UICO: an ontology-based user interaction context model for automatic task detection on the computer desktop. In Proceedings of the 1st Workshop on Context, Information and Ontologies, CIAO ’09, pages 8:1—-8:10, New York, NY, USA, 2009. ACM.
  • Romero, C. , Ventura, S.N., & Garcia, E. (2008). Data mining in course management systems: Moodle case study and tutorial, Computers & Education, 51(1), 368-384.
  • Romero, C. Ventura, S. (in press) Educational Data Mining: A Review of the State-of-the-Art. IEEE Transaction on Systems, Man, and Cybernetics, Part C: Applications and Reviews. PDF
  • Romero, C., & Ventura, S. (2007). Educational data mining: A survey from 1995 to 2005. Expert Systems with Applications, 33(1), 135-146.
  • Scott, G. (2010). TILT: UWS Tracking and Improvement System for Learning and Teaching, University of Western Sydney. PDF, retrieved 12:45, 10 April 2012 (CEST).
  • Siemens George; Dragan Gasevic, Caroline Haythornthwaite, Shane Dawson, Simon Buckingham Shum, Rebecca Ferguson, Erik Duval, Katrien Verbert and Ryan S. J. d. Baker (2011). Open Learning Analytics: an integrated & modularized platform Proposal to design, implement and evaluate an open platform to integrate heterogeneous learning analytics techniques, Society for Learning Analytics Research. PDF, retrieved 22:02, 1 March 2012 (CET).
  • Siemens, G. (2007) “Connectivism: Creating a Learning Ecology in Distributed Environment,” in Didactics of Microlearning: Concepts, discourses, and examples, in T. Hug, (ed.),Waxmann Verlag, New York, pp. 53-68
  • Siemens, G. and Long, P. (2011). Penetrating the Fog: Analytics in learning and education. EDUCAUSE Review, vol. 46, no. 4 (July/August 2011)
  • Sylvan, Elisabeth Amy . (2007). The Sharing of Wonderful Ideas: Influence and Interaction in Online Communities of Creators, PhD Thesis, Massachusetts Institute Of Technology. PDF. (Good reading if you are interested in analyzing online communities of creators (OCOC) with Network Analysis.
  • Ternier, S.; K. Verbert, G. Parra, B. Vandeputte, J. Klerkx, E. Duval, V. Ordonez, and X. Ochoa. The Ariadne Infrastructure for Managing and Storing Metadata. IEEE Internet Computing, 13(4):18–25, July 2009.
  • Werner,Linda Charlie McDowell, Jill Denner (2013). A First Step in Learning Analytics: Pre-processing Low-Level Alice Logging Data of Middle School Students, JEDM - Journal of Educational Data Mining, Vol 5, No 2, HTML/PDF
  • Wolpers. M; J. Najjar, K. Verbert, and E. Duval. Tracking actual usage: the attention metadata approach. Educational Technology and Society, 10(3):106–121, 2007.
  • Woolley, A.W., Chabris, C., Pentland, A, Hashmi, N. and Malone, T. W. (2010). Evidence for a Collective Intelligence Factor in the Performance of Human Groups. Science Magazine, Vol. 330, pp. 686-688. DOI 10.1126/science.1193147
  • Zaiane, O.R. "Web Usage Mining for a Better Web-Based Learning Environment" in Proceedings of Conference on Advanced Technology for Education (CATE'01), pp 60-64, Banff, Alberta (2001).
  • Zhang, H. and Almeroth, K. (2010). Moodog: Tracking Student Activity in Online Course Management Systems. Journal of Interactive Learning Research, 21(3), 407-429. Chesapeake, VA: AACE. Retrieved October 5, 2010 from http://www.editlib.org.aupac.lib.athabascau.ca/p/32307.
  • Zhang, H., Almeroth, K., Knight, A., Bulger, M., and Mayer, R. (2007). Moodog: Tracking Students' Online Learning Activities in C. Montgomerie and J. Seale (Eds.), Proceedings of World Conference on Educational Multimedia, Hypermedia and Telecommunications, Chesapeake, VA: AACE, 2007, pp. 4415􀂱4422. DOI 10.1.1.82.951


Planned:

CSCL, AD&Ed and learning sciences literature

  • Constantino-Gonzalez, M.A., Suthers, D.D. and Escamilla de los Santos, J.G. (2002). Coaching web-based collaborative learning based on problem solution differences and participation. International Journal of Artificial Intelligence in Education, 13, 263-299.
  • Dönmez, P., Rosé, C. P., Stegmann, K., Weinberger, A., & Fischer, F. (2005). Supporting CSCL with Automatic Corpus Analysis Technology. In T. Koschmann & D. Suthers & T. W. Chan (Eds.), Proceedings of the International Conference on Computer Support for Collaborative Learning (pp. 125-134). Taipeh, Taiwan.
  • Fischer, F., Bruhn, J., Gräsel, C. & Mandl, H. (2002). Fostering collaborative knowledge construction with visualization tools. Learning and Instruction, 12, 213-232.
  • Hutchins (1995). How a cockpit remembers its speeds. Cognitive Science, 19, 265-288.
  • Jermann, P., Soller, A., & Muehlenbrock, M. (2001). From Mirroring to Guiding: A Review of State of the Art Technology for Supporting Collaborative Learning. Proceedings of the First European Conference on Computer-Supported Collaborative Learning, Maastricht, The Netherlands, 324-331.
  • Khayat, N. Mock, M., Kindermann, J. (2011). Towards automatic behavior analysis of learners in a technology-enhanced learning environment. CSEDU 2011. May 6-8, 2011. Noordwijkerhout, the Netherlands.
  • Kolodner, J., & Guzdial, M. (1996). Effects with and of CSCL: Tracking learning in a new paradigm. In T. Koschmann (Ed.) CSCL: Theory and Practice of an Emerging Paradigm (pp. 307-320). Mahwah NJ: Lawrence Erlbaum Associates.
  • Kop, Rita (2011). The Challenges to Connectivist Learning on Open Online Networks: Learning Experiences during a Massive Open Online Course. International Review of Research in Open and Distance Learning 12 (3). PDF (case study using learning analytics, see Kop et al. 2011)
  • Kop, R., Fournier, H., & Sitlia, H. (2011). The value of learning analytics to networked learning on a personal learning environment. First International Conference on Learning Analytics and Knowledge 2011, Banff, Alberta, Canada, February 27–March 1, 2011.
  • Merceron, A. and K. Yacef. Interestingness measures for association rules in educational data. In Proceedings of Educational Data Mining Conference, pages 57-66, 2008.
  • Mühlenbrock, M., & Hoppe, U. (1999). Computer supported interaction analysis of group problem solving. In C. Hoadley & J. Roschelle (Eds.) Proceedings of the Conference on Computer Supported Collaborative Learning (CSCL-99) (pp. 398-405). Mahwah, NJ: Erlbaum.
  • Perera, D.; J. Kay, K. Yacef, and I. Koprinska. Mining learners' traces from an online collaboration tool. In Proceedings of Educational Data Mining workshop, pages 60-69, 2007. HTML
  • Schmitz, H.C., Scheffel, M., Friedrich, M., Jahn, M., Niemann, K., Wolpers, M.: Camera for PLE. In: EC-TEL. Volume 5794 of Lecture Notes in Computer Science., Springer (2009) 507–520
  • Settouti, L.S.; N. Guin, V. Luengo, A. Mille. (2010). A Trace-Based Learner Modelling Framework for Technology-Enhanced Learning Systems. 1Oth IEEE International Conference on Advanced Learning Technologies, Sousse, Tunisia. pp. 73-77. ISBN 978-1-4244-7144-7.
  • Settouti, L. S; Y. Prié, J.C. Marty, A. Mille. (2009). A Trace-Based System for Technology-Enhanced Learning Systems Personalisation. 9th IEEE International Conference on Advanced Learning Technologies, Riga, Latvia.
  • Soller, A., & Lesgold, A. (2003). A computational approach to analyzing online knowledge sharing interaction. Proceedings of Artificial Intelligence in Education 2003, Sydney, Australia, 253-260.
  • Soller et al. (eds.) (2004). The 2nd International Workshop on Designing Computational Models of Collaborative Learning Interaction PDF
  • Suthers, D. D., Ravi, V., Medina, R., Joseph, S., & Dwyer, N. (2008). Beyond threaded discussion: representational guidance in asynchronous collaborative learning environments. Computers & Education, 50, 1103-­‐1127.
  • Stegmann, K., & Fischer, F. (2011). Quantifying Qualities in Collaborative Knowledge Construction: The Analysis of Online Discussions. Analyzing Interactions in CSCL, 247–268. Springer.
  • Southavilay. V, K. Yacef, and R. A. Calvo. Process mining to support students’ collaborative writing (best student paper award). In Educational Data Mining conference proceedings, pages 257-266, 2010
  • Southavilay. V, K. Yacef, and R. A. Calvo. Analysis of collaborative writing processes using hidden markov models and semantic heuristics. In Submitted to Workshop on Semantic Aspects in Data Mining (SADM) at ICDM2010, 2010.
  • Wang Yi-Chia, Cui, Y., Arguello, J., Stegmann, K., Weinberger, A., Fischer, F., & Rosé, C. P. (2008). Analyzing collaborative learning processes automatically: Exploiting the advances of computational linguistics in computer-supported collaborative learning. International Journal of Computer-Supported Collaborative Learning, 3(3), 237–271. doi: 10.1007/s11412-007-9034-0.
  • Weinbrenner, S., Engler, J., Wichmann, A., & Hoppe, U. (2010). Monitoring and analysing students' systematic behaviour - the SCY pedagogical agent framework. Paper presented at the ECTEL conference. Barcelona.

Acknowledgements and copyright modification

Most of this article is licensed under the Creative Commons Attribution-NonCommercial 3.0 License and you also must cite cite Penetrating the Fog: Analytics in Learning and Education and Elias, T. (2011) Learning Analytics: Definitions, Processes, Potential if you reuse parts from these sources, e.g. the figures.

Some figures are not licensed for reuse and are reproduced with permission. You can't reproduce these.