Thanks to Jim and several others who have recently brought my attention to IBM's rather grandiose claim that, in a few years, classrooms will learn us. The kinds of technology described in this article are not really very new. They have been just around the corner since the 60s and have been around in quantity since the early 90s when adaptive hypermedia (AH) and intelligent tutoring systems (ITS) rose to prominence, spawning a great many systems, and copious research reported on in hundreds of conferences, books and journal articles. A fair bit of my early work in the late 90s was on applying such things to an open corpus, which is the kind of thing that has blossomed (albeit indirectly) into the recently popular learning analytics movement. Learning analytics systems are essentially very similar to AH systems but mostly leave the adaptation stage of the process up to the learner and/or teacher and tend to focus more on presenting information about the learning process in a useful way than on acting on the results. I've maintained more than a passing interest in this area but I remain a little on the edge of the field because my ambitions for such tools have never been to direct the learning process. For me, this has always been about helping people to help one another to learn, not to tell them or advise them on how to learn, because people are, at least till now, the best teachers and an often-wasted resource. This seemed intuitively obvious to me from the start and, as a design pattern, it has served me well. Of late, I have begun to understand better why it works, hence this post.
The general principle behind any adaptive system for learning is that there are learners, some kind of content, and some means of adapting the content to the learners. This implies some kind of learner model and a means of mapping that to the content, although I believe (some disagree) that the learner model can be disembodied in constituent pieces and can even happily exist outside the systems we build, in the heads of learners. Learning analytics systems are generally all about the learner model and not much else, while adaptive systems also need a content model and a means of bringing the two together.
Beyond some dedicated closed-corpus systems, there are some big obstacles to building effective adaptive systems for learning, or that support the learning process by tracking what we are doing. It's not that these are bad ideas in principle - far from it. The problem is more to do with how they are automated and what they automate. Automation is a great idea when it works. If the tasks are very well defined and can be converted into algorithms that won't need to be changed too much over time, then it can save a lot of effort and let us do things we could not do before, with greater efficiency. If we automate the wrong things, use the wrong data, or get the automation a little wrong, we create at least as many problems as we solve. Learning management systems are a simple case in point: they automated abstracted versions of existing teaching practice, thus making it more likely that existing practices would be continued in an online setting, even though they had in many cases emerged for pragmatic rather than pedagogic reasons that made little sense in an online environment. In fact, the very process of abstraction made this more likely to happen. Worse, we make it very much harder to back out when we automate, because we tend to harden a system, making it less flexible and less resilient. We set in stone what used to be flexible and open. It's worse still if we centralize that, because then whole systems depend on what we have set in stone and you cannot implement big changes in any area without scrapping the whole thing. If the way we teach is wrong then it is crazy to try to automate it. Again, learning management systems show this in spades, as do many of the more popular xMOOC systems. They automate at least some of the wrong things (e.g. courses, grading, etc). So we had better be mighty sure about what we are automating and why we are doing it. And this is where things begin to look a bit worrying for IBM's 'vision'. At the heart of it is the assumption that classrooms, courses, grades and other paraphenalia of educational systems are all good ideas that are worth preserving. The problem here is that these evolved in an ecosystem that made them a sensible set of technologies at the time but that have very little to do with best practice or research into learning. This is not about learning - it is about propping up a poorly adapted system.
If we ignore the surrounding systems and start with a clean slate, then this should be a set of problems about learning. The first problem for learning analytics is to identify what are we should be analyzing, the second is to understand what the data mean and how to process them, the third to decide what to do about that. Our knowledge on all three stages is intermediate at best. There are issues concerning what to capture, what we can dicover about learners through the information we capture, and how we should use that knowledge to help them learn better. Central to all of this is what we actually know about education and what we have discovered works best - not just statistically or anecdotally, but for any and all individuals. Unfortunately, in education, the empirical knowledge we have to base this on is very weak indeed.
So far, the best we can come up with that is fairly generalizable (my favourite example being spaced learning) is typically only relevant to small and trivial learning tasks like memorization or simple skill acquisition. We're pretty good at figuring out how to teach simple things well, and ITS and AH systems have done a pretty fair job under such circumstances, where goals (seldom learning goals - more often proxies like marks on tests or retention rates) are very clear and/or learning outcomes very simple. As soon as we aim for more complex learning tasks, the vast majority of studies of education are either specific, qualitative and anecdotal, or broad and statistical, or (more often than should be the case) both. Neither is of much value when trying to create an algorithmic teacher, which is the explicit goal of AH and ITS, and is implied in the teaching/learning support systems provided by learning analytics.
There are many patterns that we do know a lot about, though they don't help much here. We know, for example, that one-to-one mastery teaching on average works really brilliantly - Bloom's 2-sigma challenge still stands, about 30 years after it was first made. One-to-one teaching is not a process that can be replicated algorithmically: it is simply a configuration of people that allows the participants to adapt, interact and exchange or co-develop knowledge with each other more effectively than configurations where there is less direct contact between people. It lets learners express confusion or enthusiasm as directly as possible, and for the teacher to provide tailored responses, giving full and undistracted attention. It allows teachers to directly care both for the subject and for the student, and to express that caring effectively. It allows targeted teaching to occur, however that teaching might be enacted. It is great for motivation because it ticks all the boxes on what makes us self-motivated. But it is not a process and tells us nothing at all about how best to teach nor how best to learn in any way that can be automated, save that people can, on the whole, be pretty good at both, at least on average.
We also know that social constructivist models can, on average, be effective, for probably related reasons. it can also be a complete disaster. But fans of such approaches wilfully ignore the rather obvious fact that lots of people often learn very well indeed without them - the throwaway 'on average' covers a massive range of differences between real people, teachers and learners, and between the same people at different times in different contexts. This shouldn't come as a surprise because a lot of teaching leads to some learning and most teaching is neither one-to-one nor inspired by social constructivist thinking. Personally, I have learned phenomenal amounts, been inspired and discovered many things through pretty dreadful teaching technologies and processes, including books and lectures and even examined quizzes. Why does it work? Partly because how we are taught is not the same thing at all as how we learn. How you and I learn from the same book is probably completely different in myriad ways. Partly it is because it ain't what you do to teach but how you do it that makes the biggest difference. We do not yet have an effective algorithmic way of making or even identifying creative and meaningful decisions about what will help people to learn best - it is something that people and only people do well. Teachers can follow an identical course design with identical subject matter and turn it into a pile of junk or a work of art, depending on how they do it, how enthusiastic they are about it, how much eye contact they make, how they phrase it, how they pace it, their intonation, whether they turn to the wall, whether they remembered to shave, whether they stammer etc, etc, etc, and the same differentiators may work sometimes and not work others, may work for some people sometimes and not others. Sometimes, even awful teaching can lead to great learning, if the learners are interested and learn despite rather than because of the teacher, taking things into their own hands because the teaching is so awful. Teaching and learning, beyond simple memory and training tasks, are arts and not sciences. True, some techniques appear to work more often than not (but not always), but there is always a lot of mysterious stuff that is not replicable from one context to the next, save in general patterns and paradigms that are mostly not easily reduced to algorithms. It is over-ambitious to think that we can automate in software something we do not understand well enough to turn into an algorithm. Sure, we learn tricks and techniques, just like any artist, and it is possible to learn to be a good teacher just as it is possible to learn to be a good sculptor, painter or designer. We can learn much of what doesn't work, and methods for dealing with tricky situations, and even a few rules of thumb to help us to do it better and processes for learning from our mistakes. But, when it comes down to basics, it is a creative process that can be done well, badly or with inspiration, whether we follow rules of thumb or not, and it takes very little training to become proficient. Some of the best teachers I've ever known have used the worst techniques. I quite like the emphasis that Alexandra Cristea and others have put on designing good authoring environments for adaptive systems because they then become creative tools rather than ends in themselves, but a good authoring tool has, to date, proved elusive and far too few people are working on this problem.
The proponents of learning analytics reckon they have an answer to this problem, by simply providing more information, better aggregated and more easily analyzed. It is still a creative and responsive teacher doing the teaching and/or a learner doing learning, so none of the craft or art is lost, but now they have more information, more complete, more timely, better presented, to help them with the task so that they can do it better. The trouble is that, if the information is about the wrong things, it will be worse than useless. We have very little idea what works in education from a process point of view so we do not know what to collect or how to represent it, unless all we are doing is relying on proxies that are based on an underlying model that we know with absolute certainty is at least partly incorrect or, at best, is massively incomplete. Unless we can get a clearer idea of how education works, we are inevitably going to be making a system that we know to be flawed to be more efficient than it was. Unfortunately, it is not entirely clear where the flaws lie especially as what may be a flaw for one may not be for another, and a flaw in one context may be a positive benefit in another. When performing analytics or building adaptive systems of any kind, we focus on proxies like grades, attention, time-on-task, and so on - things that we unthinkingly value in the broken system and that mean different things to different people in different contexts. Peter Drucker made an important observation about this kind of thing:
'Nothing is less productive than to make more efficient what should not be done at all'.
A lot of systems of this nature improve the efficiency of bad ideas. Maybe they valorize behaviourist learning models and/or mediaeval or industrial forms of teaching. Maybe they increase the focus on grading. Maybe they rely on task-focused criteria that ignore deeper connective discoveries. Maybe they contain an implied knowledge model that is based on experts' views of a subject area, which does not normally equate to the best way to come by that knowledge. Maybe they assume that time on task matters or, just as bad, that less time spent learning means the system is working better (both and neither are true). Maybe they track progress through a system that, at its most basic level, is anti-educational. I have seen all these flaws and then some. The vast majority of tools are doing education-process analytics, not learning analytics. Even those systems that use a more open form of analytics which makes fewer assumptions about what should be measured, using data mining techniques to uncover hidden patterns, typically have risky systemic effects: they afford plentiful opportunities for filter bubbles, path dependencies, Matthew Effects and harmful feedback loops, for example. But there is a more fundamental difficulty for these systems. Whenever you make a model it is, of necessity, a simplification, and the rules for simplification make a difference. Models are innately biased, but we need them, so the models have to be good. If we don't know what it is that works in the first place then we cannot have any idea whether the patterns we pick out and use to help people guide their learning journeys are a cause, an effect or a by-product of something else entirely. If we lack an explicit and accurate or useful model in the first place, we could just again be making something more efficient that should never be done at all. This is not to suggest that we should abandon the effort, because it might be a step to finding a better model, but it does suggest we should treat all findings gathered this way with extreme scepticism and care, as steps towards a model rather than an end in themselves.
In conclusion, from a computing perspective, we don't really know much about what to measure, we don't have great grounds for deciding how to process what we have measured, and we don't know much at all about how to respond to what we have processed. Real teachers and learners know this kind of thing and can make sense of the complexity because we don't just rely on algorithms to think. Well, OK, that's not necessarily entirely true, but the algorithms are likely at a neural network level as well as an abstract level and are probably combinatorially complex in ways we are not likely to understand for quite a while yet. It's thus a little early to be predicting a new generation of education. But it's a fascinating area to research that is full of opportunities to improve things, albeit with one important proviso: we should not be entrusting a significant amount of our learning to such systems just yet, at least not on a massive scale. If we do use them, it should be piecemeal and we should try diverse systems rather than centralizing or standardizing in ways that the likes of Knewton are trying to do. It's bit like putting a computer in charge of decisions whether or not to launch nuclear missiles. If the computer were amazingly smart, reliable and bug-free, in a way that no existing computer even approaches, it might make sense. If not, if we do not understand all the processes and ramifications of decisions that have to be made along the way, including ways to avoid mistakes, accidents and errors, it might be better to wait. If we cannot wait, then using a lot of different systems and judging their different outputs carefully might be a decent compromise. Either way, adaptive teaching and learning systems are undoubtedly a great idea, but they are, have long been, and should remain on the fringes until we have a much clearer idea of what they are supposed to be doing.
The Landing is a social site for Athabasca University staff, students and invited guests. It is a space where they can share, communicate and connect with anyone or everyone.
Unless you are logged in, you will only be able to see the fraction of posts on the site that have been made public. Right now you are not logged in.
If you have an Athabasca University login ID, use your standard username and password to access this site.
We welcome comments on public posts from members of the public. Please note, however, that all comments made on public posts must be moderated by their owners before they become visible on the site. The owner of the post (and no one else) has to do that.
If you want the full range of features and you have a login ID, log in using the links at the top of the page or at https://landing.athabascau.ca/login (logins are secure and encrypted)
Posts made here are the responsibility of their owners and may not reflect the views of Athabasca University.
Comments
Hi Jon
Good piece and sums up the possible objections to adaptive learning. However, in the interests of not throwing the bath and the baby out with the bathwater, there are several important reasons for seeing adpative learning as a fruitful path in online learning.
First, I'd like to say that the same arguments can be used against traditional teaching. Where's the evidence that slabs of lectures and occassional essays actually achieve what people assume they achieve? In truth, the majority of those who teach in HE have no underlying learning theory, even rudimentary good practice. It's a crapshoot. In fact, bothe good tehory and proven practice are often studiously ignored.
More importantly, adaptive systems can objectify good theory and practice to good effect. We know that this is true in one effective learning task - search. Google uses an ensemble of algorithms that are constantly tweaked and produce stunning results. Teachers and learners use it every working day. Cookies on websites often do the same thing and most software is algorithmic in the sense that conditional code is used to execute the presentation of content based on the users request/input. Adaptive learning simply takes us further along this obviously fruitful path from simple conditionals to rule-based systems and ensembles of algorithms. This is not an easy path, as basic learning theory is hard to identify (the mistake here, I think is to follow the social constructivist path). However, there is clearly diagnostic data, from user inpute, both individual and aggregated, that is useful here.
I have seen this used to great effect in games and simulations, where the assessment of competence, through the actions of the learner is proven by the very fact that the learner performs, or faisl to perform, the task. Flight simulatirs, for example, have been doing tis for decades.
Adaptive teachniques use algorithms that make a real attempt to capture god practice, teachers, in my experience, simply wing this. Adaptive learning will take time to develop, but at least we have a path out of the messy practice that currently exists. Algorithms have made a significant contribution to increased productivity in almost every area of human endeavour, to deny its relevance in learning is to give learning a special status it doesn't deserve. Sure, it will take time, but so did most other software tasks that we now take for granted.
Teachers love to think of themselves as unique, irreplacable and practising some sort of black art. History shows us that this is just another problem to be solved.
Some good stuff on this on www.cogbooks.com
- Donald Clark
Thanks Donald - yes, I entirely agree that it's worth pursuing, to the extent that a good chunk of my research career has been dedicated to the pursuit. But we understand far too little about input, processing and output to do more than the crudest of things with it yet, so I want to throw a bucket of cold water on the idea that the world is about to change for the better as a result.
I disagree with you about teachers: teachers are irreplaceable and they are and should be doing art more than science. But they don't have to be professionals, they don't have to be people that label themselves as teachers, and they certainly don't have to be in classrooms. Nor need they be virtually, temporally, physically or socially co-present with the learner. Authors, directors, editors, designers, other learners, forum participants, janitors, architects and a host of others that provide or support the content and environment that we learn from are teachers too, as much or more than the people calling themselves teachers in the classrooms. Adaptive systems don't work in a vacuum, they work with the stuff teachers have made, whether they think of themselves as teachers or not. The better AH systems adapt to things that change rather than a fixed set of content in a closed system, but the open-corpus problem remains a huge one and we are not very close to figuring out more than the crudest of solutions to that one.
Following the social constructivist path will indeed not work for most kinds of adaptive systems for the reasons I mention. It does work quite often, though the fact that it is 100% dependent on things that cannot be codified and relies completely on how well people teach and learn (not really addressed by most constructivist models) means it is equally likely not to work: it depends entirely on the people and context. Constructivism provides a set of loosely related theories that include hardly any methods that can be turned into algorithms that machines can use. It's a way of seeing the world, not a way of teaching. A big mistake to think of it as occupying an epistemologically similar position to cognitivist and behaviourist methods. Indeed, those methods remain central in all constructivist teaching, they just get distributed a bit more among the participants. I still like it though because it reminds us that it is not all about method but at least as much about people, learning and building knowledge together.
I totally agree about Google which is perhaps the most successful learning technology in the world today, along with Wikipedia and an assortment of ever-changing social media systems. These and others (including the adaptive systems I've built and indeed this very site), work because they make it easier for people to share, connect and communicate in more interesting ways, to find people and the stuff people make more easily and efficiently. They are not completely neutral to method and pedagogy, but they are not far off. They create conditions that help people to get together and share what they know. This is much like social constructivism, not a set of methods but a set of conditions. Google's PageRank is a great idea to help people help one another - crowd wisdom enacted elegantly and effectively. However, I have extremely mixed feelings about Google's relatively recent efforts to personalize the search results that it has been pushing more and more over the past few years. The filter bubbles that result are in many ways very dangerous indeed. Increasingly, I see what I want and expect to see, with an ever-decreasing diversity and ever-decreasing chance for serendipity and mind-jarring discoveries. This might work for some people undertaking some kinds of learning, but it's a rabbit hole factory of gargantuan proportions that makes expansive learning less likely. It's exactly the kind of danger that I am trying to highlight here, where a poorly thought-through model of what people need to see becomes ubiquitous and, almost invisibly, changes how we learn without any clear idea of how best to do that.