Young people, common sense and a new government

I spent a humbling morning yesterday at Young Futures and learned a lot.

It’s an inspiring organisation which helps young people who have experienced seriously tough childhoods progress from care into a decent future.

Without diminishing it, what struck me most was that Young Futures’ work was hardly difficult to describe. They simply cared for vulnerable, often damaged people in a thoughtful, structured, respectful, individual way — from ensuring arriving teenagers can personalise their own bedrooms, to fighting an unusual student’s cause in the face of inflexible organisations and crude government targets.

It was shocking to me that Young Futures seem to be outliers in an otherwise bleak and under-resourced system. It seems that “education” is now so narrowly viewed that it does not cover nurturing the social and emotional skills and resilience that many of us take for granted. Even ignoring the moral imperative here, this costs our society enormous amounts of money over the long-term — from healthcare to incarceration.

As we contemplate five years of government by an administration not notable for its compassion or acknowledgement of complexity, I only hope a little of the spirit I experienced yesterday can infuse our politics.

Virgin Unite / B Team report on hiring for the future

I kicked off and scaffolded this substantial report for Virgin Unite and the B Team earlier in the year and I’m delighted that Holly Branson launched it on the Virgin website today.

It can be found at

Let me know what you think!

Why I don’t like the idea of “Netflix for Education”

So many companies seem to be selling themselves as the “Netflix for Education” — perhaps most notably, Pearson in a recent article. It’s an idea I find depressing.

It’s easy to see why, on first glance, this is such an appealing elevator pitch.

For the learner and teacher, education as:

  • personalised;
  • drawing on a huge range of materials which are often of very high production values;
  • allowing you to go at your own pace;
  • hugely enjoyable.

For the investor, a business which is:

  • clearly scaleable across the world;
  • reusing content with light localisation;
  • using AI to reinforce its network effect;
  • deploying a subscription business model with recurring revenues.

What’s not to like?


Unlike Netflix, the complicated business of teaching and learning is:

  • collaborative (between teachers and students, students and students, teachers and teachers, and more, in multiple complex ways…) — whether this is just learning algebra in a classroom or practising key twenty-first century skills such as working in teams, empathy or critical thinking;
  • deeply contextual: teachers judge what’s best to do next in a given moment using a wide range of very human inputs, many of which computers won’t be able accurately to assess any time soon. It’s not just about what has been chosen in the past —it’s about things like has it been hot today? or did that child eat breakfast this morning? or Jill is upset after that row in the playground;
  • active not passive — to coin an old phrase comparing the internet and television, “lean forward” not “lean back”;
  • sometimes inevitable, grinding hard work (however hard we try, learning Mandarin characters and sounds is just difficult and dull);
  • often profoundly un-digital, involving practical messy work with the hands, or outdoor investigations, or physical activity like sport, drama or music.

I guess I feel the whole idea diminishes what I feel education is really about. And, incidentally, the scope of how Pearson and others might deliver impact, revenue and profit. A shame.

Photo credit: “South Carolina First Steps 4K students recite the alphabet in class at the Child Development Center at Shaw Air Force Base, S.C., May 31, 2018. The children learned through play, structured lessons and social interactions in preparation for further schooling. (U.S. Air Force photo by Airman 1st Class Kathryn R.C. Reaves)”

Treat with care

The widespread use of medical language to describe education innovation needs challenging

Perhaps it’s because I have needed to consult doctors and nurses recently. Perhaps it’s because there is one over-simplification about education that annoys me more than most others – the meme which unfavourably compares the relative use of technology in operating theatres and classrooms now and a hundred years ago. Perhaps it’s because I’m spending a lot of time working with the medically trained founders of adaptive learning company Area9 at the moment. But the use of medical language in the debates around the future of education seems to be growing, and doesn’t seem to be questioned very much. Whilst the analogy can sometimes be helpful, it can also be lazy, misleading and counter-productive – both as we think about teaching and learning, and about medicine.

One area where the medical comparisons are useful is when talking about the value of professional training. Most researchers agree with the not-exactly-controversial idea that teachers are the single most important ingredient in student success. It’s hard to argue differently about doctors and nurses in the treatment of patients. Yet in the medical world, continuing professional development is not only regarded as common sense, but in many countries it is a legal requirement. In formal education, not so much. A failure to provide resources, incentives and career pathways which recognise teachers as reflective professionals is, in my view, a glaring failure of many education systems around the world.

It also means we aren’t widely gathering evidence and reflecting on it to make things better. As Michael Feldstein recently wrote:

“…effective ed tech cannot evolve without a trained profession of self-consciously empirical educators any more than effective medication could have evolved without a profession of self-consciously empirical physicians”

Ah, “Researchers”. “Student success”. “Empirical”. Michael and I are of course adopting the language of “learning science” – the fraught and freighted sphere of establishing “what works” in education. And this is where the problems start.

Being empirical is relatively easy in medicine. People can be scientifically tested for the presence of ailments, diseases, pathogens and “abnormalities” within reasonable, globally agreed levels of probability. To caricature, you’ve got tuberculosis or you haven’t. Once the drugs have been administered, doctors and their managers can agree on whether or not the illness has gone.

Classrooms aren’t operating theatres. As I have written before, education is a world of values, subjectivity, and Politics (big “P” deliberate). Whilst most of the world can probably agree on whether or not three individual children from England, China and India can solve quadratic equations effectively, views on the “right” history of Hong Kong under the British Empire might be highly divergent. It’s very difficult to establish a definition of “success” here without acknowledging the (inevitable and necessary) value judgements which are involved. Which means a universal definition of “empirical” is often impossible. We all need to be honest about this, and devise intelligent ways to deal with it.

There’s also a more detailed debate about medical research methodologies being applied in education, thoughtfully covered by the UK National Foundation for Educational Research here, so I won’t go into detail. Whilst a randomised controlled trial (RCT) is the gold standard in the world of medicine, it seems it isn’t necessarily always the right way forward in education.

More interesting, perhaps, is a linked broader issue. RCTs aim to find drugs and processes which make people better – medicine, at least in conventional Western circles, is mostly a business of repair. In classrooms or educational systems, we probably shouldn’t be “fixing” people. We should be nurturing them to fulfil their unique potential and giving them tools for life (note the Western liberal assumptions here, naturally).

The description of some educational software as “interventions” is almost militaristic – implying a dramatic break with a present which is unsatisfactory. When governments and thought leaders start talking about education as “broken”, there are worrying implications of retrospective conservatism, rather than creative, hopeful imagination.  When I’ve worked on successful educational resources, the process has been one of co-creation and discussion – something done with teachers, institutions, researchers and learners, rather than to them. Medical metaphors take us down a troubling road here, and it could be argued that the top-down “intervention” approach to improving education has seen some notable failures (inBloom’s collapse or Zuckerberg’s doomed initiative in Newark being good examples).

Then again, many major healthcare systems are now adopting the idea that theirs is not just a work of treating problems as they arrive. Living well, for a long period of time, can be seen as a partnership between medical researchers and the evidence they provide, trained reflective professionals acting as advisors and deploying the researched resources at their disposal, and patients taking responsibility for their own choices. A reflective, comparative discussion between health and education – about what they do and how they describe it – seems ever more fruitful.

Image: Engraving by Daniel Chodowiecki from Franz Heinrich Ziegenhagen, Lehre vom richtigen Verhältnis zu den Schöpfungswerken und die durch öffentliche Einführung desselben allein zu bewürkende algemeine Menschenbeglükkung (1792). Wikimedia Commons/Public Domain.


Meta: early indicator of Zuckerberg’s impact on education

Last week the Chan Zuckerberg Initiative (CZI) bought a company called Meta. This small acquisition is potentially seismic in its implications for the worlds of science and education.

CZI is the philanthropic vehicle for Priscilla Chan and her husband Mark Zuckerberg (I put the couple in that order because they do). After the Initiative was announced with some fanfare, there have only been a few announcements – ambitious, striking, but mostly short on detail. Chan Zuckerberg Science was announced with the intention of “helping cure, prevent or manage all diseases in our children’s lifetime.” The highly regarded Jim Shelton was appointed head of Chan Zuckerberg Education. They announced their first lead investment in edtech and employability company Andela, alongside a group of other well-known investors in the space. So far, so expected.

The Meta acquisition is different. In the company’s own words, Meta is “a tool that helps researchers understand what is happening globally in science and shows them where science is headed.” Facilitated by the move towards Open Access in scientific publishing, the venture-backed company was selling Artificial Intelligence-driven insights about the state and potential trajectory of research to publishers, pharma companies, and others.

Now, Meta’s toolset is going to be totally free, and the company has a clear mission to open up its technology to humans and machines, and collaborate with the research community to make itself better. The for-profit drive of the company has been up-ended, towards one of impact only. The venture investors are going to take their financial exit and leave (presumably having made profits they feel happy with).

This is unprecedented as far as I’m aware. I’ve never known a philanthropic organisation acquire a venture-backed startup and mould it to its own ends. Invest, make grants, lobby – all familiar stuff from the activities of the Bill and Melinda Gates Foundation and others. Acquire – no.

The effects of this are, in my view, many. There are relatively parochial issues about what happens to the world of Scientific, Technical and Medical publishing, which I spent several years analysing with my team when at Holtzbrinck. Market leader Elsevier’s apparent strategy of building competitive advantage through amassing proprietary data and the resulting algorithms looks challenged. Clarivate Analytics (previously part of Thomson Reuters) will be worried about the long term viability of its Web of Science product.

More interesting are the broader consequences and issues. I should say first that I feel that CZI is doing something really interesting and worthwhile here, and genuinely breaking new ground. But there are wrinkles.

  1. CZI just changed a commercial market. The value chain and power relationships amongst players in scientific publishing are altered substantially – not least as this is probably just the beginning of what CZI might do, given their financial firepower. Investors and shareholders will be worried and may hold off making spending decisions. Who knows what will be changed next?
  2. There is a risk that CZI actually reduce choice and innovation in the available algorithms and tools which analyse the corpus of scientific knowledge. It’s very difficult to compete with free. I suspect they have thought of this and may well want to put money into a range of approaches, but it’s worth pointing up.
  3. Venture capitalists may now have a new exit route to add to trade sale (i.e. sale to another company) or IPO: sell to CZI if your idea is clever enough and aligns with their theses. I wonder how soon I will see this in a fundraising document.
  4. There’s a clear issue of democratic accountability here. Without malice or deliberate intention on anyone’s part, we have ended up in a situation where two people can spend huge amounts of money according to their definition of “good”, and consequently affect a major component of our society. I suspect CZI are thoughtful and well aware of this, and consulted with the scientific community before making their acquisition. Nevertheless, this is another slightly scary illustration of how polarised our world is becoming.

So what about education? All of these issues are directly transferable. I would now be very surprised if Jim Shelton and his team don’t make some major moves which re-balance power in education ecosystems, particularly in the USA but likely more broadly. Investors and existing players need to take note of whatever signals we get from San Francisco.

Yet in my view CZI need to ensure they don’t transfer their model wholesale. Science has a clear, globally accepted value system based on “standing on the shoulders of giants” – rigorous process, peer review, publication. Teaching and learning are different. As I (and many others) have written before, establishing “what works” in education is a highly subjective, value-laden exercise. CZI will need to show how they are fostering innovation and impact in education with a diverse range of cultures and contexts – or unapologetically and publicly adopt a clear set of values, and argue passionately for them. They need to do these whilst avoiding being cast as latter-day cultural imperialists and facing a backlash. This is a tall order in a “post-truth” world where Zuckerberg has already struggled with issues of editorial responsibility in his day job at Facebook. The team have awe-inspiring ambition, power, and potential, which they have only just started to deploy. Things just got interesting.

When I was asked to write about good uses of technology in education…

…I passionately felt that there was a more important question to answer. Post-Brexit and post-Trump, a thousand words came out in two hours. Quite understandably, the publication that asked me to pitch an article to them rejected what I sent. So here it is.

After two decades spent at the intersection of education, digital technology and commerce, I have never felt such mixed emotions. I am disappointed at how little has changed; accepting of how naïve I was when I started; and never more determined to engage with the complex, essential and exciting challenge of what I feel needs to be done.

We need to be frank. Despite widespread excitement and rhetoric, digital learning has so far shown few clear gains to societies around the world. School systems have not been transformed, curricula remain broadly unchanged, and in a number of countries (including the UK, my home) digital initiatives have withered. The evidence for “what works” is disputed and thin. Children are leaving educational systems ill-equipped for the new realities of globalisation, and many adults are unable to re-tool themselves as the world changes around them.

For sure, there are some contexts in which the use of technology has gained adoption. These are frequently those in which there are clear crises of value and low-cost gaps which need to be filled. In US Higher Education, online learning is slowly but firmly moving into the mainstream, not least because of major demographic changes in the student body, and the widespread questioning of the cost/benefit of a traditional four-year residential degree. Alongside universities offering online degrees we are seeing the rise of the coding bootcamp – both virtual and physical – which foregrounds practical digital skills. In language learning, “good enough” solutions have emerged – whilst a free app with no human interaction is highly unlikely to make you fluent in English, it may well give you enough to improve your job prospects. Digital test preparation tools and services for high-stakes public exams are mushrooming, with particularly substantial investment going to providers in India and China. Especially in Africa, interesting but controversial experiments are under way to raise education levels with low-cost private schools facilitated by technology.

Yet all of these so far feel like tinkering rather than the unleashing of dramatic potential. In developed economies, digital pedagogy is often explored by teachers in spite of the systems in which they are operating. Throughout the world, programmes to introduce technology in schools come and go, frequently driven by political whim rather than evidence and best practice, often losing funding before there is a chance to embed new ways of working. Teachers and systems resist change, for entirely understandable reasons – they don’t want to risk their children’s futures or their careers (both frequently based on unchanged traditional public examination results), and they have seen so many initiatives come and go that cynicism has set in. Not everyone is instantly comfortable with new technology, and the support and infrastructure they need to engage with it is rarely resourced properly. Educators are rarely given the space to develop as reflective professionals. There is a lack of patient, incremental capital which allows companies to grow at the pace of discursive educational change, rather than at the pace of Uber.

Technology still holds immense potential for learning. It holds out the opportunities of greater personalisation, more exciting resources, wider choices, and better evidence with consequent improved practice. More importantly perhaps, we need to reframe the issue:  learning requires technology as our world becomes deeply digital. Our working tools are increasingly electronic. All of us need to understand the underpinnings (and underbelly) of the technology-driven societies we call our own, in the same way as most of us have long agreed that children need to know the basics of science. We have to know what we are dealing with, and know how to question it.

So – how do we make progress?

The problem here is agreeing what we want progress to be.

Education is frequently presented as an amoral scientific endeavour, particularly by technology companies who have solutions to sell. Learning algorithms will find the best pathways for our children. Teachers with more data will make better decisions. Virtual reality will bring History alive in unprecedented ways. Mobile devices will let us learn anything, anywhere.

Yet education is fundamentally, unavoidably, always about values. Algorithms are programmed by humans who decide on their priorities. Data is useless without filtering and interrogation. At its best, the study of History is learning a process of weighing evidence to find a provisional, consciously incomplete idea of the past – at its worst, a political endeavour to reinforce power. Content and data on mobile devices is mediated by powerful companies with agendas which may not always match our own.

In these days of division, fear and uncertainty, somehow we need to agree on an agenda for education in our various countries which serves to cement what holds us together rather than grow fissures into chasms. Technology can be a vital part of this. We have an extraordinary new ability to talk to and see people in the next community, or furthest country, at tiny cost. We may be able to develop new ways to assess and understand people’s abilities and potential, and help them develop the attitudes and skills which give them the best chance in life. It may even be possible more quickly to find them jobs that they can do wherever they find themselves, or wherever they want to move to, giving them the self-esteem and confidence which comes from a stable, sufficient income. We can bring unprecedented deep and diverse experiences – of people and environments – to the learning process.

Those of us who have positions of possibility, knowledge and power within the world of education face a new imperative to articulate how technology is no more and no less than a tool in the service of what we want our children – and ourselves, and our fellow citizens – to experience. We need to talk about aims and values first and evaluate our educational tools, systems, practices and people (including ourselves) against them. We need to acknowledge the inevitable moral angle of our work and argue, in the spirit of honest dialogue, for what we believe in.

As ever, references are available if you contact me via Twitter @nkind88

Why we need to predict the futures, plural


It’s been a bad year for those paid to predict. This week has shown that perhaps more than any other. The pollsters and pundits have failed again to call the way in which people will vote, meaning for the second electoral shock of 2016 – after Brexit, Trump. As someone who has just stopped being paid to think about the future for a world-leading media group, I’m intrigued. As someone who is hopefully about to start helping at least one significant organisation to develop its future strategy, I’m daunted. And as someone who hopes regularly to deconstruct ideas about the future of education for investors and decision makers, I am unsurprised. (As a citizen of the world, both events sadden me greatly, but that’s in passing).

The educational technology writer Audrey Watters regularly points out how we need to recognise that some venture capital firms, corporations and others create stories about the future to further their own agendas, and often dress them up as unassailable fact. She is essential reading for anyone making thoughtful strategy in education, and her latest, characteristically acerbic piece was published last week and persuasively skewered a number of well-known and influential organisations offering such predictions.

Yet as I read Audrey’s post I felt there was more to say, with more general application. There are times when we have to predict the future, and sometimes that has to be public. If you have the responsibility of allocating resources – for example, making decisions about where to invest money or cut jobs – I would argue you have a duty to do so having thought carefully about what might happen in the next months or years. If you have public shareholders and/or accountability, you may have to justify those choices openly.

The problems come when story becomes presented as fact, and complexity is ignored for the sake of the story. For example, personalised learning technologies hold great promise theoretically, but are substantially unproven and likely to be very context-specific. They are not “a magic pill”, but that sounds a lot better and more confident to many people.

The way through, I think, is to be humble. One way to do this is to hold multiple futures as hypotheses simultaneously and to acknowledge that none of them are likely to be entirely accurate, but all may have elements of truth in them. Those hypothetical futures should be rooted in as much evidence about the trends leading to them as you can practically obtain, and perhaps by examining the motives and likely actions of key players as events unfold. This often involves looking for data which isn’t easily available. In other words, tell multiple stories and critically compare and contrast them.

If you can explore the implications of a number of possible futures for your decision, it will be better – and the next decisions as the world changes (unexpectedly) should be better too. Additionally – and importantly – you may be able to explore the unintended consequences of your possible paths of action. Sometimes the best intentions lead to the worst of outcomes, as many politicians have found in 2016.

As Audrey points out, the future is not inevitable. We make decisions which make things happen. If we acknowledge that predictions about the future are not only stories but also more or less likely possibilities, we are better placed to evaluate the probability and desirability of those possibilities. Then, we can debate and aim for the right ones. Given the times we are in, this seems more necessary than ever.

Image credit: (CC BY SA 2.0)

As ever, links and references available on request.

Education and Artificial Intelligence – critical optimism

We shouldn’t believe the hype or succumb to dystopian panic – it is vital to be simultaneously positive and reflective

Artificial Intelligence (AI) seems to be at peak media frenzy right now, with its effects on education frequently in the mix. My own reading in the last few days has included the first published output from Stanford University’s One Hundred Year Study on Artificial Intelligence, news of the American technology giants working together to address societal concerns around AI, and a review of Yuval Noah Harari’s new book Homo Deus, which follows on from his highly influential Sapiens to examine the future of mankind in the light of new technologies.

The attention is not new, even if the intensity and pitch seem higher. Education has long been touted as a sector where advanced machines can make a real difference – and as Audrey Watters regularly points out, this history pre-dates digital computers. “Personalisation at scale” is the holy grail of education technology, seemingly promising better learning with fewer resources. Education can be portrayed as an engineering problem, with the notion of “learning science” becoming prevalent and indeed key to some major corporations’ strategies and marketing literature.

The market is full of marketing, of course. Personally, I’ve become convinced of the promise of better analytics – providing aggregated and digestibly presented data to inform the decisions of students, teachers, parents and administrators. After all, the markbook (aka gradebook in the USA) has long been a key tool of the teacher in its paper and pencil form. However, I am instinctively sceptical the moment analytics become “predictive” or software starts to recommend or even enforce paths of action to/on people who are not empowered to understand, question or circumvent them. I am particularly dubious if a commercial vendor does not wish to explain the theory behind a software algorithm, and/or claims wide, generic application, and/or does not wish for their claims of learning improvement to be subjected to rigorous independent evaluation.

My knee-jerk reaction is clearly questionable. There will never be any clear boundaries between analytics, recommendations, or artificial intelligence – the moment you ask a question of data and choose how to present its results, you have prioritised that line of enquiry over many others. Some vendors have persuaded me – often via an enjoyable debate – that their algorithms and user experiences work effectively, but only for specific aspects of learning (for example, in delivering and reinforcing factual knowledge, but not in teaching critical thinking). There is thoughtful work being done on student retention, and interesting research taking place on evaluating adaptive technologies.

My concerns remain, however, based on the following:

  • Context is crucial in effective learning – and there are so many different contexts. There are plenty of commentators who talk as if a tool or technology is context-agnostic – a “magic pill”. Yet I have had murmurs of recognition pretty much every time I discuss what I call the “carrot and piece of string” issue: a truly great teacher can probably deliver a great lesson on any subject with a carrot and a piece of string, but a poor teacher will struggle even when equipped with every fancy tool available. This is why thoughtful implementations of technology in education focus on the contexts in which effective learning happens with the provided tool, and almost invariably include training (perhaps for both teachers and learners), evaluation and reflection as part of the rollout.
  • Computers can never capture all of the data which go into learning. In the past I have caricatured this as “the breakfast problem”: there was some research some years ago in the UK which (perhaps unsurprisingly) indicated that kids who didn’t have breakfast learnt worse during the day. I haven’t yet come across a system which asks kids about their diet, but even if one did capture this relatively simple data point, it would be impossible to gather the totality of the experience the child has brought to school that day and therefore to make the absolutely best recommendations for what and how (s)he should be learning. A teacher can’t capture all the data either, but they have many more things to work with, most importantly human empathy.
  • Misleading claims of scientific amorality. All algorithms are written by humans; and education is a world particularly freighted with values. Vladimir Putin may view – and teach – the recent history of Syria rather differently to me. The notion of “learning science” can be used to imply fixed, unarguable goals as much as it can be used to imply a critical, evidence-based approach to acknowledging the complexities of understanding and delivering “good” education.
  • Crass comparisons between education and other sectors, sometimes by people who should know better. Education is not retail or entertainment. To take one example, learning recommendations are sometimes compared to Amazon’s “you bought this, so you might want these things other people like you bought” functionality. Retail is binary, unitary and effectively irreversible. You bought something or you didn’t. (Yes, you can return an item, but you still bought it). Learning is none of these – you can partially understand a concept, you may need to learn other things before you can master it, and you can forget it too (just ask me to try and explain calculus).

Discussions of the role of AI in learning have been given fresh energy, complexity and relevance by the dizzying advance of technologies which can be described with the general term “AI”, and by the extreme, often dystopian, predictions for the future which have been inspired by this rise (cue Martin Ford and Harari). Deep learning, NLP, expression recognition et. al. may, it seems, lead to a future where there are many fewer jobs, irrevocably deep societal divides, or even computers which are more intelligent than humans and present us with an existential threat.

Written (and it seems extensively debated) by a panel of genuine experts, the Stanford report is a useful counterweight to the prophets of doom. According to the report,

“Contrary to the more fantastic predictions for AI in the popular press, the Study Panel found no cause for concern that AI is an imminent threat to humankind. No machines with self-sustaining long-term goals and intent have been developed, nor are they likely to be developed in the near future. Instead, increasingly useful applications of AI, with potentially profound positive impacts on our society and economy are likely to emerge between now and 2030, the period this report considers.”

Frankenstein, mercifully, seems some way off. But the very next sentence in the text shows its nuanced approach, another impact on education, and how we all now have to step up:

“At the same time, many of these developments will spur disruptions in how human labor is augmented or replaced by AI, creating new challenges for the economy and society more broadly.”

Note that there is a deliberate choice not to say that there will be less jobs – there is a clear acknowledgement of the uncertainty around this matter elsewhere – but it is clear that jobs will be different. Here and implicitly threaded through much of the document is a highlighting of choice – the decisions that we have to make, collectively and individually, as these technologies become more prevalent and more powerful.

What does this mean for those of us that work in creating new educational products, services and companies, for those of us that research, teach and/or manage in educational contexts, for learners, or for the interested citizen? I’d like to see the following emerging:

  • Transparency from education software developers about the inputs and outputs used and produced by their code. A comparison here would be the approach taken by Android and iOS to contact information, location data, or Facebook posting – the user has to authorise apps to access external data and functionality, and therefore engages with what the software is doing behind the scenes.
  • Digital education products and services which actively include and react to input from their users as their algorithms produce recommendations – empowering teachers, learners, administrators, parents, and more to influence the real-world results (and potentially even improve the software). The Stanford report is worth quoting again here:

“Design strategies that enhance the ability of humans to understand AI systems and decisions (such as explicitly explaining those decisions), and to participate in their use, may help build trust and prevent drastic failures. Likewise, developers should help manage people’s expectations, which will affect their happiness and satisfaction with AI applications. Frustration in carrying out functions promised by a system diminishes people’s trust and reduces their willingness to use the system in the future.”

  • A marketplace for personalised learning algorithms based on an open operating system for education, so that software can compete on the basis of its learning outcomes in particular contexts rather than other barriers to market entry. I’ve written at length about this in my last blog post.
  • Investment and growth in the ecosystem of companies and other organisations aiming to up-skill, re-skill and motivate the widest possible range of people as we face the “disruption” in the job market due to AI anticipated by so many forecasters. (This is a key area of investigation for me right now as I explore my next paid job, so please get in touch if you are working on something interesting!)
  • A sophisticated ongoing debate around the deployment of AI in education, particularly amongst those developing new tools. Conference organisers please have a think!

In short, I believe in a “critical optimism” approach. AI in education is critical – it is essential for our future. We also need to be critical – we must carefully examine and challenge the actions of ourselves and others, including major organisations. And we need to be optimistic as this exciting set of technologies reveals its possibilities and pitfalls.

Note: The Stanford report is here. Audrey Watters is here. However, nearly every sentence in this piece could have had a reference to evidence or research, so I’ve left all other references out to make for clean reading. Please reach out to me via Twitter @nkind88 or in the comments, and I’ll post links.

Image credit: yumikrum on Flickr (CC-BY 2.0)

Open Educational Resources (OER) and the challenge of choice

Today I read in EdSurge, an educational technology newsletter and site, that “Menlo Park, Calif.-based nonprofit Open Up Resources launched today with $10 million in foundation funding“. The company is riding the Open Educational Resources wave which seems to be particularly high at the moment. Other buzz has included the release of a recent Babson report highlighting that “most higher education faculty are unaware of Open Educational Resources”, the US government’s #GoOpen campaign, Amazon Inspire, and more.

Broadly speaking, Open Educational Resources (usually shortened to OER) are educational materials which are free at the point of use, usually only in digital form. An often-used more restricted definition only includes resources using components (e.g. text, media) which can all be freely remixed and reused by anyone to suit individual learning contexts.

I can’t count the number of times I have sat in conferences or bars or meeting rooms and debated whether or not OER will be (or should be) the death of the publisher. Generally I feel that this is both a dull conversation (publishers can evolve) and missing the more important issues here (see below).

Consciously or unconsciously, OER development is a decision about costs and where they land. There is always a cost in developing educational resources – even a simple worksheet uploaded by a teacher onto a sharing website has a cost in terms of time which could have been deployed in another way (for example on planning a new lesson, or childcare, or going to the pub). Foundations in the USA such as the William and Flora Hewlett Foundation deploy substantial funding towards the development of both materials and OER organisations. Some of these organisations (e.g. Open Up, from a scan of their website) seek sustainable business models through the sale of add-ons such as teacher training or printed versions.

There used to be a common knee-jerk reaction from publishers that OER could never be of decent quality. Ignoring the evidence of great materials out there, this is a priori not true – OER development is just deploying many of the skills and tools of quality publishing (pedagogy, design, selection, aggregation, proof reading) with another business model. Some will be excellent, some good, some bad.

Sustainability of OER materials is a more serious matter, but equally a red herring. Many will have a very long half-life (given my degree in English Literature, I think about commentaries on Shakespeare from the eighteenth century which are read today. Digitize them, and they don’t need updating). But time-sensitive materials such as those covering US politics, or the state of Britain in Europe (sad face) need almost daily review, and can quickly become irrelevant or positively misleading if left alone. However, there are plenty of ways to ensure that this management takes place.

The real – and complex, no-easy-answers – issue around OER is that of choice and innovation. A market for textbooks served by for-profit publishers and software developers is a government policy decision, not an immutable monolith – and in many countries, such markets don’t exist. Whilst in the UK funding is given to schools to spend on the resources they feel they need for their particular context, in some Indian states (example) one textbook per subject is developed by the government and provided to schools as a fait accompli. Sometimes there is a hybrid model with a restricted list of approved publications, as with the Brazilian government’s PNLD programme or adoption states in the USA.

Governments are making their own judgements here about what’s appropriate. On the one hand, developing a single textbook has the merits of economies of production scale and of control (this last aspect being particularly attractive in authoritarian regimes). On the other, a completely free commercial market with funding devolved to schools means that more money is likely spent on developing materials than the government might have centrally (because some producers will spend on unsuccessful materials and lose money), markets can often foster innovation through competition, and schools have much greater choice. However, funding goes to for-profits, which some find uncomfortable.

OER are equally a double-edged sword. If they are funded by a foundation or uploaded gratis by teachers, that’s effectively new money coming into the system funding fresh, hopefully innovative and likely more flexible materials – at first sight, not much to dislike here. But there is a potentially restrictive effect on choice. For-profit organisations obviously won’t compete with free, and schools will usually choose to redirect restricted budgets away from educational resources if quality gratis alternatives are available. There is an extreme potential resulting scenario in which foundations become the main funders of educational resources used in schools. No educational materials are free of political or moral judgements. So, if this situation looks likely, it will become essential somehow to ensure that there is a really good choice of OER so that diverse viewpoints are available. Any suggestions?

Side note: there’s a point about remixing here which is often missed. It would be quite possible for governments to mandate schools only to buy materials which can be remixed and publishers would have to follow suit. The challenge is that many educational resources are better with copyrighted content which simply can’t be licensed to allow people to re-mix them (for example, clips from Hollywood films, or images of major artworks). 

Image by