The adoption of artificial intelligence (AI), in combination with other disruptive technologies, such as big data and cloud computing, offers huge potential for the justice system; however the introduction of AI in the judicial system also highlights a number of interesting issues and challenges. These include machine bias (i.e. humans are more likely to defer to a decision made by a machine); a lack of transparency (due to the ‘black box’ nature of much AI); inherent biases within data sets; and a worry that we are seeing the creeping erosion of human rights, and a lack of accountability, with privately developed code becoming equivalent to binding law and regulations.
With this in mind, the Council of Europe has adopted its first European Ethical Charter on the use of AI in judicial systems.
The Charter was developed by The European Commission for the Efficiency of Justice (CEPEJ).
It provides a framework of principles to aid policy makers, legislators and justice professionals when looking at the implementation and operation of AI in national judicial processes.
The CEPEJ takes the view that whilst the application of AI in the field of justice will help to improve efficiency and quality, it must be implemented in a responsible manner. In particular, the implementation of AI must comply with the fundamental rights guaranteed in the European Convention on Human Rights (ECHR) and the Council of Europe Convention on the Protection of Personal Data (the Conventions). Accordingly, AI must operate “in the service of the general interest” and its use must respect individual rights.
To achieve this, the Charter sets out five core principles to be respected in the field of AI and justice:
Speaking recently, Stéphane Leyenberger, Secretary to the CEPEJ, stated that the next step is for the CEPEJ to engage with private companies about a process whereby AI developers and technology companies could self-certify compliance of their products and services with the Charter’s principles.
The principles in the Charter will assist those involved in the development and testing of AI in judicial systems, notably AI applications processing judicial decisions and data (machine learning or any other methods deriving from data science). Principle 1, in particular, will be relevant with a strong preference to be given to ethical-by-design approaches, meaning that straight from the design and learning phases, rules prohibiting direct or indirect violations of the fundamental values protected by the Conventions are fully integrated in the AI itself. Appendix IV of the Charter sets out a checklist for evaluating proposed processing methods to ensure compatibility with the Charter.
 The Law Society, Technology and Law Policy Commission evidence session
New legislation introduced to extend digital connectivity, regulate direct marketing and protect con...Read more
The UK Court of Appeal has recently provided some limited guidance on the fine line between using a ...Read more
Given that most high-profile competition law actions tend to involve the decisions of large-scale re...Read more
Tailored insights delivered to your inbox