AI and intelligence based learning

, , 1 Comment


In The Robots Are Coming Andrés Oppenheimer writes that no one is safe including a chapter titled “They’re Coming for Lawyers!” So what are the implications of AI and the study of Law?

Well AI is here; we write with Grammarly or Hemmingway; Google answers all our questions and legal firms use chatbots such as ‘Billy Bot’ claiming to do the work of a barristers’ clerk. (1) In trying to find a definition of artificial intelligence I found it helpful to realise that as a concept it is interdisciplinary; made up of cognitive psychology, philosophy, computer science, math, linguistics as well as the notion of ‘intelligence’. Some could argue that the definition of intelligence that surrounds AI is only based on content or knowledge rather than what makes up the rich complexity of human intelligence; which includes our physicality and emotional reality and is far more than just content.

Indeed across the education sector, at all levels, there have been debates about moving away from knowledge-based curriculum towards a more intelligence-based one. If this is the case that would mean that any discipline, including Law, should focus more on non-cognitive skills such as creativity, critical thinking, communication and decision making as well as digital and data literacy. All skills, that we are increasingly told by employers, are needed in the workplace. This could also include mental resilience because real breakthroughs in any field only come about after multiple failed attempts, so the concept of ‘failure’ itself should be reframed and seen as part of an intelligence-based learning approach: 

Teaching law students how to respond productively in the face of failure will help increase resilience and efficacy in the legal profession’ (2)

Central to this is the idea of learning to learn, the ability to be flexible rather than tied down to hard rules. In a world of fake news or deep fake videos on social media sites for example; individuals need to be able to develop their self-efficacy and critical skills so that they are not easily sucked in. This leads to the issue of ethical behaviours as well, particularly in view of events like the Cambridge Analytica data misuse. It was Cambridge Analytica’s whistleblower Christopher Wylie who later spoke about how the unchecked use of data reduces human beings to a commodity: 

‘There are very few examples in human history of industries where people themselves become products and those are scary industries – slavery and the sex trade. And now, we have social media’ (Campaign 2018)

Since then in the business world, different approaches are developing. For example, Swedish fashion company H&M argues for the importance of responsibility in using AI:

“AI is a powerful tool and allows for many new possibilities. But that we can do something doesn’t mean that we should do it.”Linda Leopold, Head of AI Policy (3)

 One of H&M’s key methods is an Ethical AI Debate Club where employees meet to debate ethical dilemmas that potentially could arise in their industry. Indeed in a Jisc 2019 presentation on Why-ethical-debate-is-crucial-in-the-classroom (4) Dr. Miranda Mowbray uses the example of Cambridge Analytica’s Christopher Wylie, who at the age of 24 was a research director, to argue that young professionals are increasingly under great pressure when faced with complex decisions and that it is crucial that universities support students with building these ethical decision-making skills; as educational institutions are safe and trusted spaces to do so.

However, AI itself could be an ally to a more intelligence-based learning approach. It could support the challenge of assessing away from the knowledge curriculum or help in assessing those students who might not perform well in standard assessments. ‘Through data collection approaches that work on modelling techniques, students can benefit from a richer fair assessment system that can evaluate students across a longer period of time and from an evidence-based, value added perspective (Luckin 2017a). 

So if the likes of AI legal chatbots are increasingly going to cover the ‘content’  stuff such as answering basic questions from clients or overturning parking tickets could that mean that lawyers are freed up to deal with the more complex and subtle matters of Law? This means that the focus when then be for education to really work on developing those so-called ‘soft’ skills that are probably now not so much an extracurricular bonus but essential: 

“Lawyering requires human-human interaction, creativity, language processing at the highest level, deep understanding of how society works, and a sort of experience that can (currently) only be gained by humans.”(6)


 (2) Framing Failure in the Legal Classroom: Techniques for Encouraging Growth and Resilience. Kaci Bishop January2019

(3 )–head-of-ai-policy-.html




Print Friendly, PDF & Email

One Response

Leave a Reply

Your email address will not be published. Required fields are marked *