A letter to the Sunday Times on education reform

I was a co-signatory to a letter published yesterday in the UK Sunday Times, co-ordinated by Robert Halfon MP, Chair of the House of Commons Education Select Committee, David Davis MP, Professor Rose Luckin, and Priya Lakhani MBE. It calls for essential change in the education system and a potential silver lining in the COVID cloud.

Robert’s tweet, with images of the full letter, is here.

The Sunday Times page is here, but it’s behind a paywall.

Young people, common sense and a new government

I spent a humbling morning yesterday at Young Futures and learned a lot.

It’s an inspiring organisation which helps young people who have experienced seriously tough childhoods progress from care into a decent future.

Without diminishing it, what struck me most was that Young Futures’ work was hardly difficult to describe. They simply cared for vulnerable, often damaged people in a thoughtful, structured, respectful, individual way — from ensuring arriving teenagers can personalise their own bedrooms, to fighting an unusual student’s cause in the face of inflexible organisations and crude government targets.

It was shocking to me that Young Futures seem to be outliers in an otherwise bleak and under-resourced system. It seems that “education” is now so narrowly viewed that it does not cover nurturing the social and emotional skills and resilience that many of us take for granted. Even ignoring the moral imperative here, this costs our society enormous amounts of money over the long-term — from healthcare to incarceration.

As we contemplate five years of government by an administration not notable for its compassion or acknowledgement of complexity, I only hope a little of the spirit I experienced yesterday can infuse our politics.

Virgin Unite / B Team report on hiring for the future

I kicked off and scaffolded this substantial report for Virgin Unite and the B Team earlier in the year and I’m delighted that Holly Branson launched it on the Virgin website today.

It can be found at https://www.virgin.com/richard/holly-sam-0/guide-hiring-future

Let me know what you think!

Why I don’t like the idea of “Netflix for Education”

So many companies seem to be selling themselves as the “Netflix for Education” — perhaps most notably, Pearson in a recent article. It’s an idea I find depressing.

It’s easy to see why, on first glance, this is such an appealing elevator pitch.

For the learner and teacher, education as:

  • personalised;
  • drawing on a huge range of materials which are often of very high production values;
  • allowing you to go at your own pace;
  • hugely enjoyable.

For the investor, a business which is:

  • clearly scaleable across the world;
  • reusing content with light localisation;
  • using AI to reinforce its network effect;
  • deploying a subscription business model with recurring revenues.

What’s not to like?

Well…

Unlike Netflix, the complicated business of teaching and learning is:

  • collaborative (between teachers and students, students and students, teachers and teachers, and more, in multiple complex ways…) — whether this is just learning algebra in a classroom or practising key twenty-first century skills such as working in teams, empathy or critical thinking;
  • deeply contextual: teachers judge what’s best to do next in a given moment using a wide range of very human inputs, many of which computers won’t be able accurately to assess any time soon. It’s not just about what has been chosen in the past —it’s about things like has it been hot today? or did that child eat breakfast this morning? or Jill is upset after that row in the playground;
  • active not passive — to coin an old phrase comparing the internet and television, “lean forward” not “lean back”;
  • sometimes inevitable, grinding hard work (however hard we try, learning Mandarin characters and sounds is just difficult and dull);
  • often profoundly un-digital, involving practical messy work with the hands, or outdoor investigations, or physical activity like sport, drama or music.

I guess I feel the whole idea diminishes what I feel education is really about. And, incidentally, the scope of how Pearson and others might deliver impact, revenue and profit. A shame.

Photo credit: “South Carolina First Steps 4K students recite the alphabet in class at the Child Development Center at Shaw Air Force Base, S.C., May 31, 2018. The children learned through play, structured lessons and social interactions in preparation for further schooling. (U.S. Air Force photo by Airman 1st Class Kathryn R.C. Reaves)”

Treat with care

Perhaps it’s because I have needed to consult doctors and nurses recently. Perhaps it’s because there is one over-simplification about education that annoys me more than most others – the meme which unfavourably compares the relative use of technology in operating theatres and classrooms now and a hundred years ago. Perhaps it’s because I’m spending a lot of time working with the medically trained founders of adaptive learning company Area9 at the moment. But the use of medical language in the debates around the future of education seems to be growing, and doesn’t seem to be questioned very much. Whilst the analogy can sometimes be helpful, it can also be lazy, misleading and counter-productive – both as we think about teaching and learning, and about medicine.

One area where the medical comparisons are useful is when talking about the value of professional training. Most researchers agree with the not-exactly-controversial idea that teachers are the single most important ingredient in student success. It’s hard to argue differently about doctors and nurses in the treatment of patients. Yet in the medical world, continuing professional development is not only regarded as common sense, but in many countries it is a legal requirement. In formal education, not so much. A failure to provide resources, incentives and career pathways which recognise teachers as reflective professionals is, in my view, a glaring failure of many education systems around the world.

It also means we aren’t widely gathering evidence and reflecting on it to make things better. As Michael Feldstein recently wrote:

“…effective ed tech cannot evolve without a trained profession of self-consciously empirical educators any more than effective medication could have evolved without a profession of self-consciously empirical physicians”

Ah, “Researchers”. “Student success”. “Empirical”. Michael and I are of course adopting the language of “learning science” – the fraught and freighted sphere of establishing “what works” in education. And this is where the problems start.

Being empirical is relatively easy in medicine. People can be scientifically tested for the presence of ailments, diseases, pathogens and “abnormalities” within reasonable, globally agreed levels of probability. To caricature, you’ve got tuberculosis or you haven’t. Once the drugs have been administered, doctors and their managers can agree on whether or not the illness has gone.

Classrooms aren’t operating theatres. As I have written before, education is a world of values, subjectivity, and Politics (big “P” deliberate). Whilst most of the world can probably agree on whether or not three individual children from England, China and India can solve quadratic equations effectively, views on the “right” history of Hong Kong under the British Empire might be highly divergent. It’s very difficult to establish a definition of “success” here without acknowledging the (inevitable and necessary) value judgements which are involved. Which means a universal definition of “empirical” is often impossible. We all need to be honest about this, and devise intelligent ways to deal with it.

There’s also a more detailed debate about medical research methodologies being applied in education, thoughtfully covered by the UK National Foundation for Educational Research here, so I won’t go into detail. Whilst a randomised controlled trial (RCT) is the gold standard in the world of medicine, it seems it isn’t necessarily always the right way forward in education.

More interesting, perhaps, is a linked broader issue. RCTs aim to find drugs and processes which make people better – medicine, at least in conventional Western circles, is mostly a business of repair. In classrooms or educational systems, we probably shouldn’t be “fixing” people. We should be nurturing them to fulfil their unique potential and giving them tools for life (note the Western liberal assumptions here, naturally).

The description of some educational software as “interventions” is almost militaristic – implying a dramatic break with a present which is unsatisfactory. When governments and thought leaders start talking about education as “broken”, there are worrying implications of retrospective conservatism, rather than creative, hopeful imagination.  When I’ve worked on successful educational resources, the process has been one of co-creation and discussion – something done with teachers, institutions, researchers and learners, rather than to them. Medical metaphors take us down a troubling road here, and it could be argued that the top-down “intervention” approach to improving education has seen some notable failures (inBloom’s collapse or Zuckerberg’s doomed initiative in Newark being good examples).

Then again, many major healthcare systems are now adopting the idea that theirs is not just a work of treating problems as they arrive. Living well, for a long period of time, can be seen as a partnership between medical researchers and the evidence they provide, trained reflective professionals acting as advisors and deploying the researched resources at their disposal, and patients taking responsibility for their own choices. A reflective, comparative discussion between health and education – about what they do and how they describe it – seems ever more fruitful.

Image: Engraving by Daniel Chodowiecki from Franz Heinrich Ziegenhagen, Lehre vom richtigen Verhältnis zu den Schöpfungswerken und die durch öffentliche Einführung desselben allein zu bewürkende algemeine Menschenbeglükkung (1792). Wikimedia Commons/Public Domain.

 

Meta: early indicator of Zuckerberg’s impact on education

Last week the Chan Zuckerberg Initiative (CZI) bought a company called Meta. This small acquisition is potentially seismic in its implications for the worlds of science and education.

CZI is the philanthropic vehicle for Priscilla Chan and her husband Mark Zuckerberg (I put the couple in that order because they do). After the Initiative was announced with some fanfare, there have only been a few announcements – ambitious, striking, but mostly short on detail. Chan Zuckerberg Science was announced with the intention of “helping cure, prevent or manage all diseases in our children’s lifetime.” The highly regarded Jim Shelton was appointed head of Chan Zuckerberg Education. They announced their first lead investment in edtech and employability company Andela, alongside a group of other well-known investors in the space. So far, so expected.

The Meta acquisition is different. In the company’s own words, Meta is “a tool that helps researchers understand what is happening globally in science and shows them where science is headed.” Facilitated by the move towards Open Access in scientific publishing, the venture-backed company was selling Artificial Intelligence-driven insights about the state and potential trajectory of research to publishers, pharma companies, and others.

Now, Meta’s toolset is going to be totally free, and the company has a clear mission to open up its technology to humans and machines, and collaborate with the research community to make itself better. The for-profit drive of the company has been up-ended, towards one of impact only. The venture investors are going to take their financial exit and leave (presumably having made profits they feel happy with).

This is unprecedented as far as I’m aware. I’ve never known a philanthropic organisation acquire a venture-backed startup and mould it to its own ends. Invest, make grants, lobby – all familiar stuff from the activities of the Bill and Melinda Gates Foundation and others. Acquire – no.

The effects of this are, in my view, many. There are relatively parochial issues about what happens to the world of Scientific, Technical and Medical publishing, which I spent several years analysing with my team when at Holtzbrinck. Market leader Elsevier’s apparent strategy of building competitive advantage through amassing proprietary data and the resulting algorithms looks challenged. Clarivate Analytics (previously part of Thomson Reuters) will be worried about the long term viability of its Web of Science product.

More interesting are the broader consequences and issues. I should say first that I feel that CZI is doing something really interesting and worthwhile here, and genuinely breaking new ground. But there are wrinkles.

  1. CZI just changed a commercial market. The value chain and power relationships amongst players in scientific publishing are altered substantially – not least as this is probably just the beginning of what CZI might do, given their financial firepower. Investors and shareholders will be worried and may hold off making spending decisions. Who knows what will be changed next?
  2. There is a risk that CZI actually reduce choice and innovation in the available algorithms and tools which analyse the corpus of scientific knowledge. It’s very difficult to compete with free. I suspect they have thought of this and may well want to put money into a range of approaches, but it’s worth pointing up.
  3. Venture capitalists may now have a new exit route to add to trade sale (i.e. sale to another company) or IPO: sell to CZI if your idea is clever enough and aligns with their theses. I wonder how soon I will see this in a fundraising document.
  4. There’s a clear issue of democratic accountability here. Without malice or deliberate intention on anyone’s part, we have ended up in a situation where two people can spend huge amounts of money according to their definition of “good”, and consequently affect a major component of our society. I suspect CZI are thoughtful and well aware of this, and consulted with the scientific community before making their acquisition. Nevertheless, this is another slightly scary illustration of how polarised our world is becoming.

So what about education? All of these issues are directly transferable. I would now be very surprised if Jim Shelton and his team don’t make some major moves which re-balance power in education ecosystems, particularly in the USA but likely more broadly. Investors and existing players need to take note of whatever signals we get from San Francisco.

Yet in my view CZI need to ensure they don’t transfer their model wholesale. Science has a clear, globally accepted value system based on “standing on the shoulders of giants” – rigorous process, peer review, publication. Teaching and learning are different. As I (and many others) have written before, establishing “what works” in education is a highly subjective, value-laden exercise. CZI will need to show how they are fostering innovation and impact in education with a diverse range of cultures and contexts – or unapologetically and publicly adopt a clear set of values, and argue passionately for them. They need to do these whilst avoiding being cast as latter-day cultural imperialists and facing a backlash. This is a tall order in a “post-truth” world where Zuckerberg has already struggled with issues of editorial responsibility in his day job at Facebook. The team have awe-inspiring ambition, power, and potential, which they have only just started to deploy. Things just got interesting.

http://meta.com/#letter

When I was asked to write about good uses of technology in education…

…I passionately felt that there was a more important question to answer. Post-Brexit and post-Trump, a thousand words came out in two hours. Quite understandably, the publication that asked me to pitch an article to them rejected what I sent. So here it is.

After two decades spent at the intersection of education, digital technology and commerce, I have never felt such mixed emotions. I am disappointed at how little has changed; accepting of how naïve I was when I started; and never more determined to engage with the complex, essential and exciting challenge of what I feel needs to be done.

We need to be frank. Despite widespread excitement and rhetoric, digital learning has so far shown few clear gains to societies around the world. School systems have not been transformed, curricula remain broadly unchanged, and in a number of countries (including the UK, my home) digital initiatives have withered. The evidence for “what works” is disputed and thin. Children are leaving educational systems ill-equipped for the new realities of globalisation, and many adults are unable to re-tool themselves as the world changes around them.

For sure, there are some contexts in which the use of technology has gained adoption. These are frequently those in which there are clear crises of value and low-cost gaps which need to be filled. In US Higher Education, online learning is slowly but firmly moving into the mainstream, not least because of major demographic changes in the student body, and the widespread questioning of the cost/benefit of a traditional four-year residential degree. Alongside universities offering online degrees we are seeing the rise of the coding bootcamp – both virtual and physical – which foregrounds practical digital skills. In language learning, “good enough” solutions have emerged – whilst a free app with no human interaction is highly unlikely to make you fluent in English, it may well give you enough to improve your job prospects. Digital test preparation tools and services for high-stakes public exams are mushrooming, with particularly substantial investment going to providers in India and China. Especially in Africa, interesting but controversial experiments are under way to raise education levels with low-cost private schools facilitated by technology.

Yet all of these so far feel like tinkering rather than the unleashing of dramatic potential. In developed economies, digital pedagogy is often explored by teachers in spite of the systems in which they are operating. Throughout the world, programmes to introduce technology in schools come and go, frequently driven by political whim rather than evidence and best practice, often losing funding before there is a chance to embed new ways of working. Teachers and systems resist change, for entirely understandable reasons – they don’t want to risk their children’s futures or their careers (both frequently based on unchanged traditional public examination results), and they have seen so many initiatives come and go that cynicism has set in. Not everyone is instantly comfortable with new technology, and the support and infrastructure they need to engage with it is rarely resourced properly. Educators are rarely given the space to develop as reflective professionals. There is a lack of patient, incremental capital which allows companies to grow at the pace of discursive educational change, rather than at the pace of Uber.

Technology still holds immense potential for learning. It holds out the opportunities of greater personalisation, more exciting resources, wider choices, and better evidence with consequent improved practice. More importantly perhaps, we need to reframe the issue:  learning requires technology as our world becomes deeply digital. Our working tools are increasingly electronic. All of us need to understand the underpinnings (and underbelly) of the technology-driven societies we call our own, in the same way as most of us have long agreed that children need to know the basics of science. We have to know what we are dealing with, and know how to question it.

So – how do we make progress?

The problem here is agreeing what we want progress to be.

Education is frequently presented as an amoral scientific endeavour, particularly by technology companies who have solutions to sell. Learning algorithms will find the best pathways for our children. Teachers with more data will make better decisions. Virtual reality will bring History alive in unprecedented ways. Mobile devices will let us learn anything, anywhere.

Yet education is fundamentally, unavoidably, always about values. Algorithms are programmed by humans who decide on their priorities. Data is useless without filtering and interrogation. At its best, the study of History is learning a process of weighing evidence to find a provisional, consciously incomplete idea of the past – at its worst, a political endeavour to reinforce power. Content and data on mobile devices is mediated by powerful companies with agendas which may not always match our own.

In these days of division, fear and uncertainty, somehow we need to agree on an agenda for education in our various countries which serves to cement what holds us together rather than grow fissures into chasms. Technology can be a vital part of this. We have an extraordinary new ability to talk to and see people in the next community, or furthest country, at tiny cost. We may be able to develop new ways to assess and understand people’s abilities and potential, and help them develop the attitudes and skills which give them the best chance in life. It may even be possible more quickly to find them jobs that they can do wherever they find themselves, or wherever they want to move to, giving them the self-esteem and confidence which comes from a stable, sufficient income. We can bring unprecedented deep and diverse experiences – of people and environments – to the learning process.

Those of us who have positions of possibility, knowledge and power within the world of education face a new imperative to articulate how technology is no more and no less than a tool in the service of what we want our children – and ourselves, and our fellow citizens – to experience. We need to talk about aims and values first and evaluate our educational tools, systems, practices and people (including ourselves) against them. We need to acknowledge the inevitable moral angle of our work and argue, in the spirit of honest dialogue, for what we believe in.

As ever, references are available if you contact me via Twitter @nkind88

Why we need to predict the futures, plural

22606265795_74526e9e87_k

It’s been a bad year for those paid to predict. This week has shown that perhaps more than any other. The pollsters and pundits have failed again to call the way in which people will vote, meaning for the second electoral shock of 2016 – after Brexit, Trump. As someone who has just stopped being paid to think about the future for a world-leading media group, I’m intrigued. As someone who is hopefully about to start helping at least one significant organisation to develop its future strategy, I’m daunted. And as someone who hopes regularly to deconstruct ideas about the future of education for investors and decision makers, I am unsurprised. (As a citizen of the world, both events sadden me greatly, but that’s in passing).

The educational technology writer Audrey Watters regularly points out how we need to recognise that some venture capital firms, corporations and others create stories about the future to further their own agendas, and often dress them up as unassailable fact. She is essential reading for anyone making thoughtful strategy in education, and her latest, characteristically acerbic piece was published last week and persuasively skewered a number of well-known and influential organisations offering such predictions.

Yet as I read Audrey’s post I felt there was more to say, with more general application. There are times when we have to predict the future, and sometimes that has to be public. If you have the responsibility of allocating resources – for example, making decisions about where to invest money or cut jobs – I would argue you have a duty to do so having thought carefully about what might happen in the next months or years. If you have public shareholders and/or accountability, you may have to justify those choices openly.

The problems come when story becomes presented as fact, and complexity is ignored for the sake of the story. For example, personalised learning technologies hold great promise theoretically, but are substantially unproven and likely to be very context-specific. They are not “a magic pill”, but that sounds a lot better and more confident to many people.

The way through, I think, is to be humble. One way to do this is to hold multiple futures as hypotheses simultaneously and to acknowledge that none of them are likely to be entirely accurate, but all may have elements of truth in them. Those hypothetical futures should be rooted in as much evidence about the trends leading to them as you can practically obtain, and perhaps by examining the motives and likely actions of key players as events unfold. This often involves looking for data which isn’t easily available. In other words, tell multiple stories and critically compare and contrast them.

If you can explore the implications of a number of possible futures for your decision, it will be better – and the next decisions as the world changes (unexpectedly) should be better too. Additionally – and importantly – you may be able to explore the unintended consequences of your possible paths of action. Sometimes the best intentions lead to the worst of outcomes, as many politicians have found in 2016.

As Audrey points out, the future is not inevitable. We make decisions which make things happen. If we acknowledge that predictions about the future are not only stories but also more or less likely possibilities, we are better placed to evaluate the probability and desirability of those possibilities. Then, we can debate and aim for the right ones. Given the times we are in, this seems more necessary than ever.

Image credit: https://www.flickr.com/photos/39908901@N06/22606265795 (CC BY SA 2.0)

As ever, links and references available on request.