Trends and issues of AI-enabled legal and compliance services

As AI continues to transform many industries[1], including the legal service industry, many experts are unanimous in predicting exponential growth in AI as a paramount technology to bring new tools and features to improve legal services and access to justice. Already, many aspects of the estimated $786B[2] market for legal services are being digitised, automated and AI-enabled whether discovery in litigation (e.g. RelativityAI), divorce (e.g. HelloDivorce), dispute resolution (e.g. DoNotPay) or contract management (e.g. IronClad).

As with many disruptive technologies, there are many experts who believe that AI will significantly disrupt (rather than extend) the legal market:

“AI will impact the availability of legal sector jobs, the business models of many law firms, and how in-house counsel leverage technology. According to Deloitte, about 100,000 legal sector jobs are likely to be automated in the next twenty years. Deloitte claims 39% of legal jobs can be automated; McKinsey estimates that 23% of a lawyer’s job could be automated. Some estimates suggest that adopting all legal technology (including AI) already available now would reduce lawyers’ hours by 13%”[3]

The real impact will be more nuanced over the long-term as whilst AI will eliminate certain tasks and some legal jobs, it will also augment and extend the way legal services are provided and consumed. In doing so, it will drive new ways of working and operating for both established and new entrant firms who will need to invest in new capabilities and skills to support the opening up new markets, new business models and new service innovations. In the past few decades, we have seen the impact of emerging and disruptive technologies on established players across many sectors, including banking (e.g. FinTechs), media and entertainment (e.g. music, movies, gambling), publishing (e.g. news), travel (e.g. Airbnb) and transportation (e.g. Uber). It is very likely traditional legal providers will be faced with the same disruptive challenges from AI and AI-enabled innovations bundling automation, analytics, and cloud with new business models including subscription, transaction or freemium.

Although AI and AI-enabled solutions present tremendous opportunities to support, disrupt or extend traditional legal services, they also present extremely difficult ethical questions for society, policy-makers and legal bodies (e.g. Law Society) to decide.

This is the focus of this article which sets out a summary of these issues, and is structured into two parts:

  1. Current and future use cases and trends of AI in legal and compliance services;
  2. Key issues for stakeholders including legal practitioners, society, organisations, AI vendors, and policy-makers.

A few notes:

  • This article is not designed to be exhaustive, comprehensive or academically detailed review and analysis of the existing AI and legal services literature. It is a blog post first and foremost (albeit a detailed one) on a topic of personal and professional interest to me, and should be read within this context;
  • Sources are referenced within the footnotes and acknowledged where possible, with any errors or omissions are my own.
  • Practical solutions and future research areas of focus is lightly touched on in the conclusion, however is not a focus for this article.

Part 1 – Current and future use cases of AI in legal and compliance services

Historically, AI in legal services has focused on automating tasks via software to achieve the same outcome as if a law practitioner had done the work. However, increasing innovation in AI and experimentation within the legal and broader ecosystem have allowed solutions to accelerate beyond this historical perspective.

The graphic below provides a helpful segmentation of four main use cases of how AI tools are being used in legal services[4]:

A wider view of use cases, which links to existing legal and business processes, is provided below:

  • e-discovery;
  • document and contract management
  • expertise automation;
  • legal research and insight
  • contract management
  • predictive analytics
  • dispute resolution
  • practice automation
  • transactions and deals
  • access to justice

Further context on a selection of these uses is summarised below (note, there is overlap between many of these areas):

  • E-Discovery – Over the past few years, the market for e-discovery services has accelerated beyond the historical litigation use case and into other enterprise processes and requirements (e.g. AML remediation, compliance, cybersecurity, document management). This has allowed for the development of more powerful and integrated business solutions enabled by the convergence of technologies including cloud, AI, automation, data and analytics. Players in the legal e-discovery space include Relativity, DISCO, and Everlaw.
  • Document and contract management The rapid adoption of cloud technologies have accelerated the ability of organisations across all sectors to invest in solutions to better solve, integrate and automate business processes challenges, such as document and contract lifecycle management. For contracts, they need to be initiated (e.g. templates, precedents), shared, stored, monitored (e.g. renewals) or searched and tracked for legal, regulatory or dispute reasons (e.g. AI legaltech start-ups like Kira, LawGeex, and eBrevia). In terms of drafting and collaboration, the power of Microsoft Word, Power Automate and G-Suite solutions has expanded along with a significant number of  AI-powered tools or sites (e.g. LegalZoom) that help lawyers (and businesses or consumers) to find, draft and share the right documents whether for commercial needs, transactions or litigation. New ‘alternative legal service’ entrants have combined these sorts of powerful solutions (and others in this list) with lower-cost labour models (with non-legal talent and/or lower-cost legal talent) to provide a more integrated offering for Fortune500 legal, risk and compliance teams (e.g. Ontra, Axiom, UnitedLex, Elevate, Integreon);
  • Expertise Automation –In the access to justice context, there are AI-powered services that automate contentious or bureaucratic situations for individuals such as utility bill disputes, small claims, immigration filing, or fighting traffic tickets (e.g. DoNotPay). Other examples include workflow automation software to enable consumers to draft a will (for a fixed fee or subscription) or chatbots in businesses to give employees access to answers to common questions in a specific area, such as employment law. It is forseeable that extending this at scale in a B2C context (using AI-voice assistants Siri or Alexa) with a trusted brand (e.g. Amazon Legal perhaps?) – and bundled into your Prime subscription alongside music, videos and same-day delivery – will be as easy as checking the weather or ordering an Uber.
  • Legal Research – New technologies (e.g. AI, automation, analytics, e-commerce) and business models (e.g. SaaS) have enabled the democratisation of legal knowledge beyond the historic use cases (e.g. find me an IT contract precedent or Canadian case law on limitation of liability). New solutions make it easy for clients and consumers (as well as lawyers) to find answers or solutions to legal or business challenges without interacting with a lawyer. In more recent times, legal publishing companies (e.g. LexisNexis, PLC, Westlaw) have leveraged legal sector relationships and huge databases of information including laws and regulations in multiple jurisdictions to build different AI-enabled solutions and business models for clients (or lawyers). These offerings promise fast, accurate (and therefore cost-effective) research with a variety of analytical and predictive capabilities. In the IP context, intellectual property lawyers can use AI-based software from companies like TrademarkNow and Anaqua to perform IP research, brand protection and risk assessment;
  • Legal and predictive analytics – This area aims to generate insights from unstructured, fragmented and other types of data sets to improve future decision-making.  A key use case are the tools that will analyse all the decisions in a domain (e.g. software patent litigation cases), input the specific issues in a case including factors (e.g. region, judge, parties etc) and provide a prediction of likely outcomes. This may significantly impact how the insurance and medical industry operate in terms of risk, pricing, and business models. For example, Intraspexion leverages deep learning to predict and warn users of their litigation risks, and predictive analytical company CourtQuant has partnered with two litigation financing companies to help evaluate litigation funding opportunities using AI. Another kind of analytics will review a given piece of legal research or legal submission to a court and help judges (or barristers) identify missing precedents In addition, there is a growing group of AI providers that provide what are essentially do-it-yourself tool kits to law firms and corporations to create their own analytics programs customized to their specific needs;
  • Transactions and deals – Although no two deals are the same, similar deals do require similar processes of pricing, project management, document due diligence and contract management. However, for various reasons, many firms will start each transaction with a blank sheet of paper (or sale and purchase agreement) or a sparsely populated one. However, AI-enabled document and contract automation solutions – and other M&A/transaction tools – are providing efficiencies during each stage of the process. In more advanced cases, data room vendors in partnership with law firms or end clients are using AI to analyse large amounts of data created by lawyers from previous deals. This data set is capable of acting as an enormous data bank for future deals where the AI has the ability to learn from these data sets in order to then:
    • Make clause recommendations to lawyers based on previous drafting and best practice.
    • Identify “market” standards for contentious clauses.
    • Spot patterns and make deal predictions.
    • Benchmark clauses and documents against given criteria.
    • Support pricing decisions based on key variables
  • Access to justice – Despite more lawyers in the market than ever before, the law has arguably never been more inaccessible. From a small consumer perspective, there are thousands of easy-to-use and free or low cost apps or online services which solve many simple or challenging aspects of life, whether buying properties, consulting with a doctor, making payments, finding on-demand transport, or booking household services. However, escalating costs and increasing complexity (both in terms of the law itself and the institutions that apply and enforce it) mean that justice is often out of reach for many, especially the most vulnerable members of society. With the accelerating convergence of various technologies and business models, it is starting to play a role in opening up the (i) provision of legal services to a greater segment of the population and (ii) replacing or augmenting the role of legal experts. From providing quick on-demand access to a lawyer via VC, accelerating time to key evidence, to bringing the courtroom to even the most remote corners of the world and digitizing many court processes, AI, augmented intelligence, and automation is dramatically improving the accessibility and affordability of legal representation. Examples include:
    • VC tools e.g. Zoom, FaceTime
    • Document and knowledge automation e.g. LegalZoom
    • ADR to ODR (online dispute resolution) e.g. eBay, Alibaba
    • Speed to evidence – Cloud-based, AI-powered technology e.g. DISCO

2. Key issues for the future of AI-power legal and compliance services  

There are many significant issues and challenges for the legal sector when adopting AI and AI-powered solutions. Whilst every use case of AI-deployment is unique, there are some overarching issues to be explored by key stakeholders including the legal profession, regulators, society, programmers, vendors and government.  

A sample of key questions include the following:

  • Will AI in the future make lawyers obsolete?
  • How does AI impact the duty of competence and related professional responsibilities?
  • How do lawyers, users and clients and stakeholders navigate the ‘black box’ challenge?
  • Do the users (e.g. lawyers, legal operations, individuals) and clients trust the data and the insights the systems generate?
  • How will liability be managed and apportioned in a balanced, fair and equitable way?
  • How do organisations identify, procure, implement and govern the ‘right’ AI-solution for their organisation?
  • Are individuals, lawyers or clients prepared to let data drive decision outcomes?
  • What is the role of ethics in developing AI systems?

Other important questions include:

  • How do AI users (e.g. lawyers), clients or regulators ‘audit’ an AI system?
  • How can AI systems be safeguarded from cybercriminals?
  • To what extent do AI-legal services need to be regulated and consumers be protected?
  • Have leaders in businesses identified the talent/skills needed to realise the business benefits (and manage risks) from AI?
  • To what extent is client consent to use data an issue in the development and scaling of AI systems?
  • Are lawyers, law students, or legal service professionals receiving relevant training to prepare for how they need to approach the use of AI in their jobs?
  • Are senior management and employees open to working with or alongside AI systems in their decisions and decision-making?

Below we further explore a selection of the above questions:

  • Obsolescence – When technology performs better than humans at certain tasks, job losses for those tasks are inevitable. However, the dynamic role of a lawyer — one that involves strategy, negotiation, empathy, creativity, judgement, and persuasion — can’t be replaced by one or several AI programs. As such, the impact of AI on lawyers in the profession may not be as dire as some like to predict. In his book Online Courts and the Future of Justice, author Richard Susskind discusses the ‘AI fallacy’ which is the mistaken impression that machines mimic the way humans work. For example, many current AI systems review data using machine learning, or algorithms, rather than cognitive processes. AI is adept at processing data, but it can’t think abstractly or apply common sense as humans can. Thus, AI in the legal sector enhances the work of lawyers, but it can’t replace them (see chart below[5]).
  • Professional Responsibility – Lawyers in all jurisdictions have specific professional responsibilities to consider and uphold in the delivery of legal and client services. Sample questions include:
    • Can a lawyer discharge professional duties of competence if they do not understand how the technology works?
    • Is a legal chatbot practicing law?
    • How does a lawyer provide adequate supervision where the lawyer does not understand how the work is being done or even ‘who’ is doing it?
    • How will a lawyer explain decisions made if they do not even know how those decisions were derived?

To better understand these complex questions, the below summaries some of the key professional duties and how they are being navigated by various jurisdictions:

Duty of Competence: The principal ethical obligation of lawyers when they are developing or assisting clients is the duty of competence. Over the past decade, many jurisdictions are specifically requiring lawyers to understand how (and why) new technologies such as AI, impact that duty (and related duties). This includes the requirement for lawyers to develop and maintain competence in ‘relevant technologies’. In 2012, in the US the American Bar Association (the “ABA”) explicitly included the obligation of “technological competence” as falling within the general duty of competence which exists within Rule 1.1 of its Model Rules of Professional Conduct (“Model Rules”)[6]. To date, 38 states have adopted some version of this revised comment to Rule 1.1. In Australia, most state solicitor and barrister regulators have incorporated this principle into their rules. In the future, jurisdictions may consider it unethical for lawyers or legal service professionals to avoid technologies that could benefit one’s clients. A key challenge is that there is no easy way to provide objective and independent analysis of the efficacy of any given AI solution, so that neither lawyers nor clients can easily determine which of several products or services actually achieve either the results they promise. In the long-term, it will very likely be one of the tasks of the future lawyer to assist clients in making those determinations and in selecting the most appropriate solution for a given problem. At a minimum, lawyers will need to be able to identify and access the expertise to make those judgments if they do not have it themselves.

Duty to Supervise – This supervisory duty assumes that lawyers are competent to select and oversee team members and the proper use of third parties (e.g. law firms) in the delivery of legal services[7]. However, the types of third parties used has expanded in recent times due to liberalisation of legal practice in some markets (e.g. UK due to the ABS laws allowing non-lawyers to operate legal services businesses). For example, alternative service providers, legal process outsourcers, tech vendors, and AI vendors have historically been outside of the remit of the solicitor or lawyer regulators (this is changing in various jurisdictions as discussed in below sections). By extension, to what extent is this more than just a matter of the duty to supervise what goes on with third parties, but how those third-parties provide services especially if technologies and tools are used? In such a case, potential liability issues arise if client outcomes are not successful: did the lawyer appropriately select the vendor, and did the lawyers properly manage the use of the solution?

The Duty to Communicate – In the US, lawyers also have an explicit duty to communicate to material matters to clients in connection with the lawyers’ services. This duty is set out in ABA Model Rue 1.4 and other jurisdictions have adopted similar rules[8]. Thus, not only must lawyers be competent in the use of AI, but they will need to understand its use sufficiently to explain to clients the question of the selection, use, and supervision of AI tools.

Black Box Challenge  

  • Transparency – A basic principle of justice is transparency – the requirement to explain and justify the reasons for a decision. As AI algorithms grow more advanced and rely on increasing volumes of structured and unstructured data sets, it becomes more difficult to make sense of their inner workings or how outcomes have been derived. For example, Michael Kearns and Aaron Roth report in Ethical Algorithm Design Should Guide Technology Regulation[9]:

“Nearly every week, a new report of algorithmic misbehaviour emerges. Recent examples include an algorithm for targeting medical interventions that systematically led to inferior outcomes for black patients, a resume-screening tool that explicitly discounted resumes containing the word “women” (as in “women’s chess club captain”), and a set of supposedly anonymized MRI scans that could be reverse-engineered to match to patient faces and names”.

Part of the problem is that many of these types of AI systems are ‘self-organising’ so they are inherently without external supervision or guidance. The ‘secrecy’ of AI vendors – especially those in a B2B and legal services context – regarding the inner workings of the AI algorithms and data sets doesn’t make the transparency and trust issue difficult for customers, regulators and other stakeholders. For lawyers, to what extent must they know the inner workings of that black box to ensure that she meets her ethical duties of competence and diligence? Without addressing this, these problems will likely continue as the legal sector increases its reliance on technology increases and injustices, in all likelihood, continue to arise. Over time, many organisations will need to have a robust and integrated AI business strategy designed at the board and management level to guide the wider organisation on these AI issues across areas including governance, policy, risk, HR and more. For example, during procurement of AI solutions, buyers, stakeholders and users (e.g. lawyers) must consider broader AI policies and mitigate these risk factors during vendor evaluation and procurement.

  • Algorithms – There are many concerns that AI algorithms are inherently limited in their accuracy, reliability and impartiality[10]. These limitations may be the direct result of biased data, but they may also stem from how the algorithms are created. For example, how software engineers choose a set of variables to include in an algorithm, deciding how to use variables, whether to maximize profit margins or maximize loan repayments, can lead to a biased algorithm. Programmers may also struggle to understand how an AI algorithm generates its outputs—the algorithm may be unpredictable, thus validating “correctness” or accuracy of those outputs when piloting a new AI system. This brings up the challenge of auditing algorithms:

“More systematic, ongoing, and legal ways of auditing algorithms are needed. . . . It should be based on what we have come to call ethical algorithm design, which begins with a precise understanding of what kinds of behaviours we want algorithms to avoid (so that we know what to audit for), and proceeds to design and deploy algorithms that avoid those behaviours (so that auditing does not simply become a game of whack-a-mole).”[11]

In terms of AI applications, most AI algorithms within legal services are currently able to perform only a very specific set of tasks based on data patterns and definitive answers. Conversely, it performs poorly when applied to the abstract or open-ended situations requiring judgment, such as the situations that lawyers often operate in[12]. In these circumstances, human expertise and intelligence are still critical to the development of AI solutions. Many are not sophisticated enough to understand and adapt to nuances, and to respond to expectations and layered meaning, and comprehend the practicalities of human experience. Thus, AI still a long way from the ‘obsolescence’ issue for lawyers raised above, and further research is necessary on programmers’ and product managers’ decision-making processes and methodologies when ideating, designing, coding, testing and training an AI algorithm[13]:

  • Data – Large volumes of data is a critical part of AI algorithm development as training material and input material. However, data sets may be of poor quality for a variety of reasons. For example, the data an AI system is ‘trained’ on may well include systemic ‘human’ bias, such as recruiters’ gender or racial discrimination of job candidates. In terms of data quality in law firms, most are slow at adopting new technologies and tend to be “document rich, and data poor” due, in large part, to legacy on-premise systems (or hybrid cloud) which do not integrate with each other. As more firms and enterprises transition to the cloud, this will accelerate the automation of business processes (e.g. contract management) with more advanced data and analytics capabilities to enable and facilitate AI system adoption (in theory, however there are many constraints within traditional law firm business and operating models which makes the adoption of AI-enabled solutions at scale unlikely). However, 3rd party vendors within the legal sector including e-discovery, data rooms, and legal process outsourcers – or new tech-powered entrants from outside of the legal sector – do not have such constraints and are able to innovate more effectively using AI, cloud, automation and analytics in these contexts (however other constants exist such as client consent and security). In the court context, public data such as judicial decisions and opinions are either not available or so varied in format as to be difficult to use effectively[14]. Beyond data quality issues, significant data privacy, client confidentiality and cybersecurity concerns exist which raises the need to define and implement standards (including safeguards) to build confidence in the use of algorithmic systems – and especially in legal contexts. As AI becomes more pervasive within law firms, legal departments, legal vendors (including managed services) and new entrants outside of legal, a foundation with strong guidelines for ethical use, transparency, privacy, cross-department sharing and more becomes even crucial[15].
  • Implementation – Within the legal sector, law firms and legal departments are laggards when it comes to adopting new technologies, transforming operations, and implementing change. With business models based on hours billed (e.g. law firms), this may not incentivize the efficiency improvements that AI systems can provide.  In addition:

“Effective deployment of AI requires a clearly defined use case and work process, strong technical expertise, extensive personnel and algorithm training, well-executed change management processes, an appetite for change and a willingness to work with the new technologies. Potential AI users should recognize that effectively deploying the technology may be harder than they would expect. Indeed, the greatest challenge may be simply getting potential users to understand and to trust the technology, not necessarily deploying it[16].

However, enterprises (e.g. Fortune500), start-ups, alternative service providers (e.g. UnitedLex) and new entrants from outside of legal do not suffer from these constraints, and are likely to be more successful – from a business model and innovation perspective – in adopting new AI-enabled solutions for use with clients (although AI-enabled providers must work to overcome client concerns as discussed above).   

  • Liability – There are a number of issues to consider on the topic of liability. Key questions are set out below:
    • Who is responsible when things do go wrong? Although AI might be more efficient than a human lawyer at performing these tasks, if the AI system misses clauses, mis-references definitions, or provides incorrect outcome/price predictions caused by AI software, all parties risk claims depending on how the parties apportioned liability. The role of contract and insurance is key, however this assumes that law firms have the contractual means of passing liability (in terms of professional duties) onto third parties. In addition, when determining relative liability between the provider of the defective solution and the lawyer, should a court consider the steps the lawyer took to determine whether the solution was the appropriate one for use in the particular client’s matter?
    • Should AI developers be liable for damage caused by their product? In most other fields, product liability is an established principle. But if the product is performing in ways no-one could have predicted, is it still reasonable to assign blame to the developer? AI systems also often interact with other systems so assigning liability becomes difficult. AI solutions are also fundamentally reliant on the data they were trained on, so liability may exists with the data sources.  Equally, there are risks of AI systems that are vulnerable to hacking.
    • To what extent are, or will, lawyers be liable when and how they use, or fail to use, AI solutions to address client needs? One example explained above is whether a lawyer or law firm will be liable for malpractice if the judge in a matter accesses software that identifies guiding principles or precedents that the lawyer failed to find or use. It does not seem to be a stretch to believe that liability should attach if the consequence of the lawyer’s failure to use that kind of tool is a bad outcome for the client and the client suffers injury as a result.
  • Regulatory Issues – As discussed above, addressing the significant issues of bias and transparency in AI tools, and, in addition, advertising standards, will grow in importance as the use of AI itself grows. Whilst the wider landscape for regulating AI is fragmented across industry and political spheres, there are signs the UK, EU and US are starting to align.[17] Within the legal services sector, some jurisdictions (e.g. England, Wales, Australia and certain Canadian provinces) are in the process of adopting and implementing a broader regulatory framework. This approach enables the legal regulators to oversee all providers of legal services, not just traditional law firms and/or lawyers. However, in the interim the implications of this regulatory imbalance will become more pronounced as alternative legal service providers play an increasing role in providing clients with legal services, often without any direct involvement of lawyers. In the long run, a broader regulatory approach is going to be critically important in establishing appropriate standards for all providers of AI-based legal services.
  • Ethics – The ethics of AI and data uses remains a high concern and key topic for debate in terms of the moral implications or unintended consequences that result from the coming together of technology and humans. Even proponents of AI, such as Elon Musk’s OpenAI group, recognise the need to police AI that could be used for ‘nefarious’ means. A sample of current ethical challenges in this area include:
    • Big data, cloud and autonomous systems provoke questions around security, privacy, identify, and fundamental rights and freedoms;
    • AI and social media challenge us to define how we connect with each other, source news, facts and information, and understand truth in the world;
    • Global data centres, data sources and intelligent systems means there is limited control of the data outside our borders (although regimes including GDPR is addressing this);
    • Is society content with AI that kills? Military applications including lethal autonomous weapons are already here;
    • Facial recognition, sentiment analysis, and data mining algorithms could be used to discriminate against disfavoured groups, or invade people’s privacy, or enable oppressive regimes to more effectively target political dissidents;
    • It may be necessary to develop AI systems that disobey human orders, subject to some higher-order principles of safety and protection of life;

Over the years, the private and public sectors have attempted to provide various frameworks and standards to ensure ethical AI development. For example, the Aletheia Framework[18] (developed by Rolls-Royce in an open partnership with industry) is a recent, practical one-page toolkit that guides developers, executives and boards both prior to deploying an AI, and during its use. It asks system designs and relevant AI business managers to consider 32 facets of social impact, governance and trust and transparency and to provide evidence which can then be used to engage with approvers, stakeholders or auditors. A new module added in December 2021 is a tried and tested way to identify and help mitigate the risk of bias in training data and AIs. This complements the existing five-step continuous automated checking process, which, if comprehensively applied, tracks the decisions the AI is making to detect bias in service or malfunction and allow human intervention to control and correct it.

Within the practice of law, while AI offers cutting-edge advantages and benefits, it also raises complicated questions for lawyers around professional ethics. Lawyers must be aware of the ethical issues involved in using (and not using) AI, and they must have an awareness of how AI may be flawed or biased. In 2016, The House of Commons Science and Technology Committee (UK Parliament) recognised the issue:

“While it is too soon to set down sector-wide regulations for this nascent field, it is vital that careful scrutiny of the ethical, legal and societal dimensions of artificially intelligent systems begins now”.

In a 2016 article in the Georgetown Journal of Legal Ethics, the authors Remus and Levy were concerned that:

“…the core values of legal professionalism meant that it might not always be desirable, even if feasible, to replace humans with computers because of the different way they perform the task. This assertion raises questions about what the core values of the legal profession are and what they should or could be in the future. What is the core value of a solicitor beyond reserved activities? And should we define the limit of what being a solicitor or lawyer is?[19]

These are all extremely nuanced, complex and dynamic issues for lawyers, society, developers and regulators at large. How the law itself may need to change to deal with these issues will be a hot topic of debate in the coming years.

Conclusion

Over the next few years there can be little doubt that AI will begin to have a noticeable impact on the legal profession and consumers of legal services. Law firms, in-house legal departments and alternative legal services firms and vendors – plus new entrants outside of legal perhaps unencumbered by the constraints of established legal sector firms – have opportunities to explore and challenges to address, but it is clear that there will be significant change ahead. What is required of a future ‘lawyer’ (this term may mean something different in the future) or legal graduate today – let alone in 2025 or 2030 versus new lawyers of a few decades ago, will likely be transformed in many ways. There are also many difficult ethical questions for society to decide, for which the legal practice regulators (e.g. Law Society in England and Wales) may be in a unique position to grasp the opportunity of ‘innovating the profession’ and lead the debate. On the other hand, as the businesses of the future become more AI-enabled at their core (e.g. Netflix, Facebook, Google, Amazon etc), the risk that many legal services become commoditised or a ‘feature set’ within a broader business or service model is a real possibility in the near future.

At the same time, AI itself poses significant legal and ethical questions across all sorts of sectors and priority global challenges, from health, to climate change, to war, to cybersecurity. Further analysis on the legal and ethical implications of AI for society, legal practitioners, organisations, AI vendors, and policy-makers, plus what practical solutions can be employed to navigate the safe and ethical deployment of AI in the legal and other sectors, will be critical.


[1] AI could contribute up to $15.7 trillion1 to the global economy in 2030, more than the current output of China and India combined. Of this, $6.6 trillion is likely to come from increased productivity and $9.1 trillion is likely to come from consumption side effects.

[2] https://www.statista.com/statistics/605125/size-of-the-global-legal-services-market/

[3] https://jolt.law.harvard.edu/digest/a-primer-on-using-artificial-intelligence-in-the-legal-profession

[4] https://www.morganlewis.com/-/media/files/publication/presentation/webinar/2020/session-11_the-ethics-of-artificial-intelligence-for-the-legal-profession_18june20.pdf

[5] https://kirasystems.com/learn/can-ai-be-problematic-in-legal-sector/

[6] https://www.americanbar.org/groups/professional_responsibility/publications/professional_lawyer/27/1/the-future-law-firms-and-lawyers-the-age-artificial-intelligence

[7] Australian Solicitors Conduct Rules 2012, Rule 37 Supervision of Legal Services.

[8] https://lawcat.berkeley.edu/record/1164159?ln=en

[9] https://www.brookings.edu/research/ethical-algorithm-design-should-guide-technology-regulation/

[10] https://hbr.org/2019/05/addressing-the-biases-plaguing-algorithms

[11] https://www.brookings.edu/research/ethical-algorithm-design-should-guide-technology-regulation/

[12] https://hbr.org/2019/05/addressing-the-biases-plaguing-algorithms

[13] https://bostonreview.net/articles/annette-zimmermann-algorithmic-political/

[14] https://www.law.com/legaltechnews/2019/10/29/uninformed-or-underwhelming-most-lawyers-arent-seeing-ais-value/

[15] https://www.crowell.com/NewsEvents/Publications/Articles/A-Tangled-Web-How-the-Internet-of-Things-and-AI-Expose-Companies-to-Increased-Tort-Privacy-and-Cybersecurity-Litigation

[16] https://www.lexisnexis.co.uk/pdf/lawyers-and-robots.pdf

[17] https://www.brookings.edu/blog/techtank/2022/02/01/the-eu-and-u-s-are-starting-to-align-on-ai-regulation/

[18] https://www.rolls-royce.com/sustainability/ethics-and-compliance/the-aletheia-framework.aspx

[19]https://go.gale.com/ps/i.do?id=GALE%7CA514460996&sid=googleScholar&v=2.1&it=r&linkaccess=abs&issn=10415548&p=AONE&sw=w&userGroupName=anon%7E138c97cd

Advertisement

Top ten traps which cause GC’s to fail around technology, automation and change

I recently came across a brilliant guide from Plexus, a legal software and services provider based in Australia. Although most GCs will rate rate ‘investing in technology and automation’ as their top priority, many initiatives will never get past the starting line.

There are many reasons for this and everyone’s context is different. It is certainly not for lack of intent, ambition, need, or interest.

According to Plexus, the biggest challenge functional leaders have is when they are required to rapidly work and operate in a cross-functional way. These skills are a new ‘core’ capability for a legal function or executive:

If you ask a GC how they select and sign up a new law firm to spend $100,000 - the answer is often as simple as a few emails and signed engagement letter. Ask them how they will spend half that amount on technology, and they will scratch their head… even though - because of the ‘sunk cost’ nature of professional services spend’ it is far more likely that it will not generate value

Plexus.co

I have summarised the top ten traps below:

You let risk aversion get the better of you.

You approach the project like a contract negotiation.

You frame the business case around ‘what legal needs’

You get hung up in ‘integrations’

You get too many people involved.

You try to go too big too early.

You delegate responsibility to others.

You allow other functions to drive the agenda.

You fall for the ‘tomorrow fallacy’.

You work with vendors with the wrong orientation.

I highly recommend reading the full version here which contains more detail on the above traps plus extremely practical ‘do this’ and ‘do not do this’ strategies and tactics.

Gartner’s Top 5 Legal Technology Trends Changing In-House Legal Departments

This week I came across an article from Gartner (Feb 21) covering future trends within in-house legal departments. I think the list still holds true, although it will be interesting to check when the 2022 version comes out soon.

Obviously most lawyers have long been resistant to and risk-averse about new technologies and automation, but the effects of the pandemic forced many to shift gears in 2020 and pursue — or at least actively consider — more extensive automation of certain legal activities, especially those around major corporate transactions. According to Gartner, the challenge now is deciding which technologies to embrace to drive real business outcomes.

“The new pressures brought about by the coronavirus pandemic certainly have acted as a catalyst for this shift,” says Zack Hutto, Director, Advisory, Gartner. “Legal and compliance teams have rarely been frontrunners to modernize, digitalize and automate. The pandemic has flattened staffing budgets and increased legal workloads; technology is the most obvious solution for many legal departments.”

Here are the Top 5 trends:

Trend No. 1: By 2025, legal departments will increase their spend on legal technology threefold

Trend No. 2: By 2024, legal departments will replace 20% of generalist lawyers with non-lawyer staff

Trend No. 3: By 2024, legal departments will have automated 50% of legal work related to major corporate transactions

Trend No. 4: By 2025, corporate legal departments will capture only 30% of the potential benefit of their contract life cycle management investments

Trend No. 5: By 2025, at least 25% of spending on corporate legal applications will go to nonspecialist technology providers

You can read the full article here

BBC’s Reith Lectures – Living with AI

Last Wednesday BBC R4 hosted the first of 4 weekly lectures hosted by Professor Stuart Russell, a world-renowned AI expert at UCLA. The talks (followed by Q&A) examine the impact of AI on our lives and discuss how we can retain power over machines more powerful than ourselves.

I think this area (e.g. AI commercialisation, AI governance, AI safety, AI ethics, AI regulation etc) is going to be one of the hot topics of the next decade alongside trends including climate change, fintech (crypto), AR/VR, quantum computing etc. Accordingly I couldn’t wait to hear Professor Russell speak.

The event blurb states the following:

The lectures will examine what Russell will argue is the most profound change in human history as the world becomes increasingly reliant on super-powerful AI. Examining the impact of AI on jobs, military conflict and human behaviour, Russell will argue that our current approach to AI is wrong and that if we continue down this path, we will have less and less control over AI at the same time as it has an increasing impact on our lives. How can we ensure machines do the right thing? The lectures will suggest a way forward based on a new model for AI, one based on machines that learn about and defer to human preferences.

As I write, I have heard 2 talks both of which have been absolutely fascinating (and quite honestly, scary. Especially regarding military applications of AI which is already here). I didn’t take notes however the BBC interviewed Professor Russell ahead of the talks. I have provided a summary of the Q&A below which is well worth a read:

How have you shaped the lectures?

The first drafts that I sent them were much too pointy-headed, much too focused on the intellectual roots of AI and the various definitions of rationality and how they emerged over history and things like that.

So I readjusted – and we have one lecture that introduces AI and the future prospects both good and bad.

And then, we talk about weapons and we talk about jobs.

And then, the fourth one will be: “OK, here’s how we avoid losing control over AI systems in the future.”

Do you have a formula, a definition, for what artificial intelligence is?

Yes, it’s machines that perceive and act and hopefully choose actions that will achieve their objectives.

All these other things that you read about, like deep learning and so on, they’re all just special cases of that.

But could a dishwasher not fit into that definition?

It’s a continuum.

Thermostats perceive and act and, in a sense, they have one little rule that says: “If the temperature is below this, turn on the heat.

“If the temperature is above this, turn off the heat.”

So that’s a trivial program and it’s a program that was completely written by a person, so there was no learning involved.

All the way up the other end – you have the self-driving cars, where the decision-making is much more complicated, where a lot of learning was involved in achieving that quality of decision-making.

But there’s no hard-and-fast line.

We can’t say anything below this doesn’t count as AI and anything above this does count.

And is it fair to say there have been great advances in the past decade in particular?

In object recognition, for example, which was one of the things we’ve been trying to do since the 1960s, we’ve gone from completely pathetic to superhuman, according to some measures.

And in machine translation, again we’ve gone from completely pathetic to really pretty good.

So what is the destination for AI?

If you look at what the founders of the field said their goal was, general-purpose AI, which means not a program that’s really good at playing Go or a program that’s really good at machine translation but something that can do pretty much anything a human could do and probably a lot more besides because machines have huge bandwidth and memory advantages over humans.

Just say we need a new school.

The robots would show up.

The robot trucks, the construction robots, the construction management software would know how to build it, knows how to get permits, knows how to talk to the school district and the principal to figure out the right design for the school and so on so forth – and a week later, you have a school.

And where are we in terms of that journey?

I’d say we’re a fair bit of the way.

Clearly, there are some major breakthroughs that still have to happen.

And I think the biggest one is around complex decision-making.

So if you think about the example of building a school – how do we start from the goal that we want a school, and then all the conversations happen, and then all the construction happens, how do humans do that?

Well, humans have an ability to think at multiple scales of abstraction.

So we might say: “OK, well the first thing we need to figure out is where we’re going to put it. And how big should it be?”

We don’t start thinking about should I move my left finger first or my right foot first, we focus on the high-level decisions that need to be made.

You’ve painted a picture showing AI has made quite a lot of progress – but not as much as it thinks. Are we at a point, though, of extreme danger?

I think so, yes.

There are two arguments as to why we should pay attention.

One is that even though our algorithms right now are nowhere close to general human capabilities, when you have billions of them running they can still have a very big effect on the world.

The other reason to worry is that it’s entirely plausible – and most experts think very likely – that we will have general-purpose AI within either our lifetimes or in the lifetimes of our children.

I think if general-purpose AI is created in the current context of superpower rivalry – you know, whoever rules AI rules the world, that kind of mentality – then I think the outcomes could be the worst possible.

Your second lecture is about military use of AI and the dangers there. Why does that deserve a whole lecture?

Because I think it’s really important and really urgent.

And the reason it’s urgent is because the weapons that we have been talking about for the last six years or seven years are now starting to be manufactured and sold.

So in 2017, for example, we produced a movie called Slaughterbots about a small quadcopter about 3in [8cm] in diameter that carries an explosive charge and can kill people by getting close enough to them to blow up.

We showed this first at diplomatic meetings in Geneva and I remember the Russian ambassador basically sneering and sniffing and saying: “Well, you know, this is just science fiction, we don’t have to worry about these things for 25 or 30 years.”

I explained what my robotics colleagues had said, which is that no, they could put a weapon like this together in a few months with a few graduate students.

And in the following month, so three weeks later, the Turkish manufacturer STM [Savunma Teknolojileri Mühendislik ve Ticaret AŞ] actually announced the Kargu drone, which is basically a slightly larger version of the Slaughterbot.

What are you hoping for in terms of the reaction to these lectures – that people will come away scared, inspired, determined to see a path forward with this technology?

All of the above – I think a little bit of fear is appropriate, not fear when you get up tomorrow morning and think my laptop is going to murder me or something, but thinking about the future – I would say the same kind of fear we have about the climate or, rather, we should have about the climate.

I think some people just say: “Well, it looks like a nice day today,” and they don’t think about the longer timescale or the broader picture.

And I think a little bit of fear is necessary, because that’s what makes you act now rather than acting when it’s too late, which is, in fact, what we have done with the climate.

The Reith Lectures will be on BBC Radio 4, BBC World Service and BBC Sounds.

How to Fail Fast and Pivot: Lessons from the Legal Ops Front Lines

This week I attended a virtual Summit hosted by CLOC (Corporate Legal Operations Consortium). One interesting session covered lessons learned from developing and implementing legaltech and operation changes within legal and compliance teams of large corporates and SMEs.

Panelists including range of lawyers, project managers and legalops experts from Netflix, Salesforce and GE and covered topics including:

*How to manage change and being comfortable with being uncomfortable

*Avoiding big bang deployments which are so risky now vs POC/MVP and more agile approaches to change

*Learning how to take some budget off the total and use it for experimenting and be prepared to fail.

The below is a blurb introducing the session:

“…On the path to success, failure is not only an option, it’s inevitable. Mistakes, missteps, and misunderstandings are opportunities to acquire new skills and knowledge that can contribute to your professional growth. The growing and evolving legal operations profession is filled with opportunities to evolve beyond errors. 

In this honest and impactful session, legal operations professionals will share a key moment of failure and how they learned and grew from it. After hearing from the panelists on their vulnerable moments of growth, we will spend time as a group sharing our own stories and offering our peers perspectives and possible solutions for overcoming some of their failures…”

Below I captured a few nuggets of gold from the panellists:

  • Give a purpose to failure; this helps to gain buy-in from users and clarify the bigger picture
  • Allow the community to own the new way to work rather than push to them
  • Leadership (e.g. town hall) to set the tone
  • Wider business context, and show why the change is important
  • Be transparent – there will be failure. Expect it. Tolerate it!
  • If leading the change, need to get to a stage of comfortable with being uncomfortable…but not too uncomfortable. Have to be mindful of current state of culture, empathise with users
  • Need to balance focus on the big picture using storytelling, sales skills etc as can’t control every details of the change
  • Experimental in communication, design thinking, courageous leadership, state of culture a huge consideration on how to balance approach
  • Big bang projects are so risky now vs POC/MVP and more agile approaches

Ultimate Resource List: Optimising & Transforming Legal Services

Here is a list of great resources I have started to compile recently. I’ll continue to add here over time and hopefully build up a pretty good catalogue for those people interested in optimising and transforming legal services, whether legal and compliance teams and departments, law firms, or other service providers.

As the list get bigger I’ll start to add headings to make it easier to follow. If you come across any interesting resources, frameworks, guides, research etc, be sure to email me at andrewessa@gmail.com

What is Legal Operations? A CLOC Guide

InCloudCounsel Guide for Successful B2B Vendor Outsourcing

Top Priorities for Legal and Compliance in 2021

KPMG 2021 Global Legal Department Benchmarking Survey

CLOC Templates and Resources

Gartner Guide for New General Counsels: 8 Step Action Plan

Gartner Guide For Selecting and Implementing LegalTech

25 Legal Tech Stats for 2020/21

This week I came across a blog post from ImpactMyBiz which compiled a list of great statistics, use cases and market data pertaining to the current state of technology in the legal sector.

In sum, there’s a lot of good progress but the sector is still subject to a lot of hype and extremely slow adoption when compared to other sectors. This is moreso in the B2B space with B2C innovation moving at a faster rate of adoption in improvement over time.

Perhaps the continued challenges presented by COVID around the world, increasing regulatory complexity, competitive pressures from alternative legal service providers (ALSP) and new entrants, remote working, client cost pressures, access to justice, and other key drivers will continue to move the needle forward.

25 legal tech stats to shed light on where where the industry is heading for in the new decade:

1.  In 2018, legal tech investments broke the $1 billion mark. That figure was topped in 2019, with $1.23 billion in funding by the end of the third quarter alone.

2. With the help of AI, a contract can be reviewed in less than an hour, saving 20-90% of the time needed to perform this work manually without sacrificing accuracy.

3. AI legal technology offerings for businesses increased nearly two-thirds in 2020 compared to 2019.

4. JP Morgan launched their in-house program, COIN, which extracts 150 attributes from 12,000 commercial credit agreements and contracts in a few seconds. This is equivalent to 360,000 hours of legal work by lawyers and loan officers per year.

5. Cloud usage among firms is 58%, with smaller firms and solos leading the way.

6. Security measures are lacking, with no more than 35% of firms using precautionary cybersecurity measures to protect their businesses. A staggering 7% of firms have no security measures at all.

7. Despite some reservations, lawyers continue to use popular consumer cloud services like Google Apps, iCloud and Evernote at higher rates than dedicated legal cloud services. Clio and NetDocuments ranked the highest among the legal cloud services.

8. The percentage of the ABA 2019 Legal Technology Survey participants answering “Yes” to the basic question of whether they had used web-based software services or solutions grew slightly, from 55% to 58%. 31% said “No”, a small decrease. 

9. When asked what prevented their law firms from adopting the cloud, 50% cited confidentiality/security concerns, 36% cited the loss of control and 19% cited the cost of switching.

10. 26% of respondents in a 2019 survey report that their law firms have experienced some sort of security breach

11. In 2018, just 25% of law firms reported having an incident response plan. In 2019, this figure had risen to 31%, and we expect the same for 2020.

12. Interest in cloud services from law firms is high, but expectations of adoption among them remain low, with just 8% of firms indicating they will replace existing legacy software with cloud tools.

13. Only one-third of lawyers (34%) believe their organizations are very prepared to keep up with technology changes in the legal market.

14. Firms described as “technology leading” fared better, with 50% prepared to meet digital technology demands in the industry.

15. 49% of law firms report that they are effectively using technology today, and 47% say they can improve technology adoption and plan to do so.

16. Over half (53%) of lawyers in the US and Europe say their organizations will increase technology investment over the next three years.

17. While over half of lawyers expect to see transformational change in their firms from technology like AI, big data and analytics, fewer than one quarter say they understand them.

18. The biggest trends cited by lawyers that are driving legal tech adoption are “Coping with increased volume and complexity of information” and “Emphasis on improved productivity and efficiency.”

19. It is estimated that 23% of work done by lawyers can be automated by existing technology.

20. 27% of the senior executives at firms believe that using digital transformation is not a choice, but a matter of survival.

21. The top challenges for corporate legal departments today include reducing and controlling outside legal costs; improving case and contract management; and automating routine tasks and leveraging technology in work processes.

22. 60% of lawyers believe their legal firm is ready to adopt new technology for routine tasks.

23. According to research conducted by Gartner, only 19% of law firms’ in-house teams are ready to move forward with enterprise-level digital strategies.

24. A recent study uncovered that 70% of consumers would rather use an automated online system or “lawbot” to handle their legal affairs instead of a human lawyer because of three important factors—cost, speed, and ease of use.

25. 70% of businesses indicated that “using tech to simplify workflow and manual processes” to cut costs was a top priority going forward.

“New Law” Opportunities for Law Firms

I recently came across a presentation I gave in April 2015 to senior partners at Eversheds LLP in London. At the time, Eversheds were proactive in starting to diversify their professional services offerings away from traditional legal and transactional work into ‘alternative’ services areas, such as business improvement consulting for in-house legal teams, and flexible resourcing solutions.

At the time, it was unusual for a major corporate firm to be experimenting into different areas.

The question for the presentation was as follows:

Downward cost pressure, deregulation and new technology are transforming the legal industry, as ‘New Law’ providers compete with traditional law firms.  What are the opportunities for large law firms in this evolving marketplace? 

I focused on 2 main themes of (a) Changing the mind-set and (b) Managing innovation.

Since then, in six years a lot of innovation has been introduced into the legal sector. However, it has been a fairly low-bar for many years with the legal sector ‘glacial’ when it comes to change and technology.

Certainly the ‘legaltech’ and/or ‘lawtech’ markets have received significant injections of VC to build next generation B2C and B2B solutions. Most large firms are now experimenting with different AI and automation solutions, running incubators, offering flexible resourcing arrangements, investing in start-ups, and so on.

To better support Fortune500 General Counsels with their efficiency challenges, the Big4 are building services and capability at scale, as are legal process outsourcers and ALSP’s.

Many of these ideas were referenced in the presentation.

However, the critical question is has anything really changed in how legal services are delivered, bought and sold? How much of this is ‘innovation theatre’ and nibbling around the edges versus real change?

For example:

  • Does the partner in the Freshfields office in HK work any differently then they did as a trainee 20 years ago?
  • Are the skills and requirements of a newly qualified lawyer any different?
  • Does the single lawyer law office in Bristol run their practice any differently?
  • Does the COO of a regional law firm run the business any differently?
  • Do consumers who need a family lawyer do this any differently?
  • Does the barrister or judge involved in a trial do this any differently?

The short answer I think is not a great deal of change across the industry as a whole. However there has been a tonne of experimentation and innovation in some fragmented areas, especially in B2C (e.g. DoNotPay). COVID-19 has certainly accelerated this, and that can only be a good thing.

I think what we are seeing is a marathon, not a sprint. In fact, it is more like the start of a triathlon where there’s a washing-machine effect as participants fight their way forward before a steadier state emerges.

We see this with most new technologies, where things often take much longer to truly disrupt. In retail and e-Commerce, it is only recently that the Internet is causing significant challenges for traditional players, almost 20 years after the Dot.Com crash in 2001.

One thing is for sure – the next 10 to 15 years in the legal sector will be fascinating.

Pandemic Pivots by Small Businesses

The COVID-19 crisis caused many businesses to make crunch decisions such as rapidly pivot offerings or building out new products/services. Often we hear stories of how big companies (e.g. Uber pivoting away from ride-sharing to food delivery) have done this (or not as the case may be), but rarely do we hear of pivots by small or local businesses.

In the course of research for my REIGNITE! 2020 Report which analysed strategic responses of 439 international organisations (large and small) around the world between March-June 2020, I came across many inspirational stories of incredible small business pivots.

In a recent speech to the NED Forum (slides here), I described the story of one particular business who had managed to turn crisis into opportunity.

To tell the story of a brilliant pandemic pivot by a small business, I’ve pasted the excerpt from the talk below:

Let me tell you a quick story about ABC Learning Company, based here in Gsy. Obviously that is not their real name but I came across them in some research I did during Q2 and lockdown. 

In the research which later became the REIGNITE 2020 Report – which I’ll introduce shortly – there was so much devastation across sectors including travel, hospitality, retail, construction, manufacturing, and so on. 

In fact 50% of the 439 leaders surveyed were in total despair, in terms of closures, restructuring, uncertainty and so on. 

However…there was a glimmer of hope!

About 10% of businesses were doing extraordinary things. They were using the crisis as an opportunity to reset, rethink, and reinvent. They were pivoting, quickly using technology to launch new offerings, testing new business models, and at the same time becoming more efficient, productive and reducing costs.

In terms of ABC Learning, it was a typical lifestyle business providing high school tutors, owned by one person with 5 tutors on the payroll. No online presence, web-site or anything. Business stopped overnight with lockdown, but by rethinking things quickly and using simple online and digital tools – google spreadsheets for CRM and bookings, zoom for delivery of live sessions, stripe for online or over the phone payments, the owner was not only able to quickly survive but doubled revenue during lockdown, hired 10 more tutors on contracts, and created a scalable solution which allowed for recorded training on-demand on popular topics. So better CX, more revenue and profits.

So what is interesting here is the combination of human psychology and business strategy during a crisis: so how did the leader reinvent whilst everyone was retreating, what can we learn, and how can we emulate this for our own contexts

This is what underpins today’s talk and certainly the REIGNITE 2020 Report which I’ll introduce shortly.


OneTrust: How A Privacy-Law-Compliance Tech Start-Up Became America’s Fastest Growing Company

Today I came across an incredible story of OneTrust, a privacy-law-compliance start-up based in Atlanta.

OneTrust landed at No. 1 on this year’s Inc. 5000, with more than $70 million in 2019 revenue and a staggering 48,337.2 percent three-year growth rate. It is among the global leaders in privacy-law-compliance technology with a suite of digital tools that gives companies a clearer view of all the user data they accumulate.

This enables them to comply with privacy laws, like the European Union’s General Data Protection Regulation (GDPR) which gives consumers greater control of how com­panies can use their data.

Whilst Enterprise B2B SaaS and analytics isn’t the most sexiest space, in many cases, firms that play there can be the fastest-growing, most scaleable and profitable (and do good things at the same time). 

In an age of Big Tech monopolies, increasingly intelligent AI and API-powered platform business models, growing regulatory oversight and appetite, and increasing consumer-awareness, OneTrust and others are clearly riding a tidal wave. 

Read the full story here 

 

 

6 Ways To Make Digital Investments More Successful

Recently I posted here about how organisations can go back to basics and understand what digital really means. In the context of today’s rapid acceleration of digital and IT investments to support remote or new ways of working – from cloud to SaaS tools to desktop VC solutions – this is critical to understand.

Another key fact to consider is that some of the most successful companies ever were started during or just after times of crisis (e.g. GE, GM, IBM, Disney, Facebook).

For leaders who can seize the ‘re-set’ opportunity this crisis provides – and start to engage with more long-term, future-focused, and exploratory strategic planning with digital at the core – this presents a potentially game-changing moment.

This presents a critical question: how should firm’s approach and organise to make digital or innovation investments and transformations successful?

Whilst there is no playbook, below I pull together a number of perspectives from some of the world’s leading management thinkers and practitioners on strategy, digital, innovation and change.

The Challenge

Digital transformation is extremely complex and requires new ways of approaching strategy. Starting big, spending a lot, and assuming you have all the information is likely to produce a full-on attack from corporate antibodies—everything from risk aversion and resentment of your project to simple resistance to change.

  1. Start Small, Think Big

Professor Rita McGrath calls this ongoing learning approach to strategy: discovery-driven planning (DDP). It was developed in the 1990s as a product innovation methodology, and it was later incorporated into the popular “lean start-up” tool kit for launching businesses in an environment of high uncertainty. At its center is a low-cost process for quickly testing assumptions about what works, obtaining new information, and minimizing risks. According to Rita:

By starting small, spending a little on an ongoing portfolio of experiments, and learning a lot, you can win early supporters and early adopters. By then moving quickly and demonstrating clear impact on financial performance indicators, you can build a case for and learn your way into a digital strategy. You can also use your digitization projects to begin an organizational transformation. As people become more comfortable with the horizontal communications and activities that digital technologies enable, they will also embrace new ways of working.

2. Soft and Hard Facts About Change

Managing change is tough, but part of the problem is that there is little agreement on what factors most influence transformation initiatives. Ask five executives to name the one factor critical for the success of these programs, and you’ll probably get five different answers.

In recent years, many change management gurus have focused on soft issues, such as culture, leadership, and motivation. Such elements are important for success, but managing these aspects alone isn’t sufficient to implement transformation projects.

According to consultants from BCG in an Harvard Business Review article entitled The Hard Side Of Change Management:

What’s missing, we believe, is a focus on the not-so-fashionable aspects of change management: the hard factors. These factors bear three distinct characteristics. First, companies are able to measure them in direct or indirect ways. Second, companies can easily communicate their importance, both within and outside organizations. Third, and perhaps most important, businesses are capable of influencing those elements quickly. Some of the hard factors that affect a transformation initiative are the time necessary to complete it, the number of people required to execute it, and the financial results that intended actions are expected to achieve. Our research shows that change projects fail to get off the ground when companies neglect the hard factors. That doesn’t mean that executives can ignore the soft elements; that would be a grave mistake. However, if companies don’t pay attention to the hard issues first, transformation programs will break down before the soft elements come into play.

3. Breaking Down the Barriers

According to a 2019 article from the partners from Innosight, a critical reason for businesses failing to get the impact they want is because they’ve failed to address a huge underlying obstacle: the day-to-day routines and rituals that stifle innovation.

Shifting+the+Culture+Iceberg

Innosight Partner Scott Anthony talks further about this below:

4. A Systematic Approach

A study by McKinsey here of leaders post-transformation has shown there are 21 best practices for organisation’s to implement to improve the chances of success.

These characteristics fall into five categories: leadership, capability building, empowering workers, upgrading tools, and communication. Specifically:

  • having the right, digital-savvy leaders in place
  • building capabilities for the workforce of the future
  • empowering people to work in new ways
  • giving day-to-day tools a digital upgrade
  • communicating frequently via traditional and digital methods

One interesting best practice was that firm’s who deploy multiple forms of technologies, tools and methods tended to have a great success rate with transformation (see below).

This might seem counterintuitive, given that a broader suite of technologies could result in more complex execution of transformation initiatives and, therefore, more opportunities to fail. But the organizations with successful transformations are likelier than others to use more sophisticated technologies, such as artificial intelligence, the Internet of Things, and advanced neural machine-learning techniques.

4. Execute AND Innovate

For any followers of the work of the late Professor Clayton Christensen on Disruptive Innovation (view his HBR collection of popular articles here), this is a fundamental challenge for almost every established firm which often becomes a matter of survival during industry, business model, technology or other shifts.

According to Alex Osterwalder:

This continues to be one of the biggest challenges we see companies face: to create two parallel cultures of world-class execution and world class innovation that collaborate harmoniously.

How Airbnb Cut 25% of Its Workforce

Every day we have been witnessing examples of great leadership (or not so great). On its face, the approach to layoff 25% of Airbnb’s workforce – which until the crisis was on track for a bumper IPO – seemed like another one of those great examples.

However, the coverage has been both supportive and negative. Without going into the detail of it, my initial thoughts are that if there was ever a ‘classy’ way to do this, this was it. 

You can read the full statement from CEO Brian Cesky here and judge for yourself. 

 

What Digital Really Means

“Everyone wants to go digital. The first step is truly understanding what that means” – McKinsey

I was talking to a COO of an off-shore investment bank yesterday and he mentioned something which gave me the impression that his bank did not understand what ‘digital’ really meant. According to McKinsey:

For some executives, it’s about technology. For others, digital is a new way of engaging with customers. And for others still, it represents an entirely new way of doing business. None of these definitions is necessarily incorrect. But such diverse perspectives often trip up leadership teams because they reflect a lack of alignment and common vision about where the business needs to go. This often results in piecemeal initiatives or misguided efforts that lead to missed opportunities, sluggish performance, or false starts.

As COVID-19 continues to rapidly accelerates the shift to building more digital capabilities within organisations, it is a critical time to take a step back and reevaluate existing efforts in light of the new challenges ahead. This means properly understanding what digital means, assessment of existing efforts, aligning to future strategy, and identifying what capabilities are needed across leadership, culture, and execution.

Whilst extremely hard, now is the best time to refocus efforts toward accelerating digitisation as the case for such change is for some a matter of survival. Think about how many food and other retailers are rapidly shifting to e-commerce models requiring new skills, software, tools and mindsets.

You can read more on this from McKinsey here

 

Reimagine The Future Online Conference

This Reimagine The Future virtual conference starts today and I have signed up for it.

It is being run by Thinkers50 and Outthinker and features 24 top management experts doing 24 sessions in 24 hours including Renee Mauborgne (INSEAD and Blue Ocean Strategy), Scott Anthony (Innosight), Daniel Pink, and Hal Gregerson. Recordings of every session will be available on-demand so there’s no need to be live.

All profits are going to a range of charities involved in COVID-19 relief. You can access tickets here.

Turbocharging Legal Industry Transformation

“COVID-19 will produce a thinning of the herd and a reimagined legal industry” – Mark Cohen

Last Thursday I watched a great online webinar run by LegalGeek entitled ‘An Uncertain Decade’. Legal sector experts Mark Cohen and Richard Susskind ran the sessions. Whilst I have read various thought leadership from each expert, it was the first time I had seen either speak. Not surprisingly, they both were very impressive in both domain expertise and thoughtfulness around their points of view.

Here are some key takeaways (along with my thoughts):

  • Disruption: COVID-19 will dramatically turbocharge legal industry transformation which has been slowly accelerating since the 2008 Financial Crisis. This may not be that surprising to many outsiders, however many lawyers – including those in Generation X – still tend to be conservative when thinking about competition, new technologies, business models, and structural market changes. Transformation and disruptive models and services will continue to come from outside traditional law firms. Whilst Disruptive Innovation theories of Professor Clayton Christensen were not referenced, his work on how established firms often are disrupted by low-end entrants who move up-market over time, will provide insight as to why and how this is happening within the legal sector (click here for his articles from Harvard Business Review).
  • Innovation: Enterprise clients and consumers will continue to drive the shift away from bundled services toward a more productised, customer-centric mode of consumption at scale which leverage new technologies, business models, and regulatory changes. This has already been happening to some extent ‘around the edges’, and facilitated by ABS models in the UK (whereby retailers (e.g. Co-Op), real estate agents, insurers and other firms can compete head-on with traditional legal practices with their own legal services ventures). DoNotPay in the US was cited as a recent example of a start-up which has over 100 use cases of dispute resolution services (e.g. fight a parking ticket). A new and current use case in the US for them has been to make it easy, cheap and convenient to file unemployment and other worker claims.
  • Unbundling: The impending depression driven by COVID-19 and resultant cost pressures for clients will accelerate the shifting of lower value, high-volume work to more flexible, alternative providers (e.g. LPOs, ABS Licensed Firms, Big4, Axiom, UnitedLex etc) and digital platforms (e.g. DoNotPay). This will continue to enable these players to move further up-scale into higher-value, more complex work and jobs. This is how the Indian IT outsourcing firms managed to make significant in-roads against Accenture, IBM and others in the 2000s, and how Toyota managed to become a US car manufacturing leader with its low-cost model US market strategy. Over the coming years, the legal industry will continue to rapidly fragment beyond traditional structural boundaries to incorporate a much bigger share for alternative providers (which will grow rapidly at the expense of incumbents), but significant new markets will be created especially for those consumers (i.e. non-consumers) who historically have never able to access low-cost, convenient legal services;
  • Business Models: A next generation of technology-enabled service providers (e.g. FisherBroyles) will gain rapid scale over the next decade in the same way as FinTechs have within Retail Banking. Continued experimentation by established law firms (e.g. non-legal services diversification, in-sourcing IT, new product development etc) and further consolidation within and amongst traditional law firms and alternative services providers, vendors and legaltechs unable to re-capitalise or scale. Large traditional law firms with the foresight and capital to invest over the coming years will likely continue to struggle to properly allocate resources and organise these innovative models efficiently and effectively within the established firm.
  • Online Dispute Resolution: COVID-19 has provided an MVP to as legal systems and courts globally have had to re-think how to deliver this. According to Mark Cohen:

The inaccessibility, cost, formality, abstruse rules, and protracted processes of courts in their present guise is misaligned with life in the digital age. The urgency of modernization is unprecedented. Courts around the world have ground to a halt when demand for accessible, efficient, and widespread administration of justice is desperately needed.

  • Education: Law Schools have not really changed their content, formats or approach to skills in over 40 years. Combined with EdTech disruption, providers will be under significant pressure to change in line with industry and client demands. Traditional JD/LLB’s offered by mid-market schools in the short-term will see massive disruption and closures, whilst the degree as it stand may become a more commoditised requirement, augmented by other specialty courses run by others or industry. Clearly, now is the time for online law ‘degree’ or course models, assuming the solicitor/lawyer regulatory boards provide approval (if they haven’t already)
  • Training: Traditional insistence of a junior being trained up or supervised by a senior lawyer/partner will be turned on its head. Assuming a longer-term shift to more remote working for a large number of the workforce and demand for more multi-skilled lawyers (e.g. project management), training for juniors (and all staff) will need to be redesigned.

In a recent Forbes article here, Mark Cohen concludes the following:

The old guard will cling to the hope these are temporary changes. They will point to the recession precipitated by the 2008 global economic crisis and suggest the current one will take a similar course. This time  is different. Technology and new delivery models are far more advanced than they were in 2008. Consumers have a different mindset and a greater urgency to solve a growing list of complex challenges. The potential of technology and its ability to support new models, processes, and paradigms is already on display. The genie is out of law’s bottle, and it will not return.

Opportunities for Law (and other) Firms

Interestingly, there wasn’t too much discussion on this during the webinar. For me, Disruptive Innovation theory might provide some guidance for any progressive law firms who wish to take on the inevitable structural and business model disruptions described above. I’ll save this analysis for another post soon.

 

%d bloggers like this: