Note: This article is just one of 60+ sections from our full report titled: The 2024 Legal AI Retrospective - Key Lessons from the Past Year. Please download the full report to check any citations.
Supporting Multi-Dimensional Explainability
Explanation: Reliability of AI outputs is a function of the different dimensions the AI uses to source its output. Like how a legal case needs to be looked at from criminal law, civil law and constitutional law.
Challenge:
Need for comprehensive trustworthiness explanations
Research Direction:
Develop integrated trustworthiness metrics
Challenge:
Barriers to interdisciplinary collaboration
Research Direction:
Foster collaboration through shared resources and training

Alex Denne, Head of Growth (Open Source Law) at Ƶ, is a legal tech leader and serial founder with over a decade of experience driving innovation and making legal services more accessible. Since joining in 2021, he has scaled the platform from 200 to over 120,000 users, combining deep contract law expertise with a data-driven, open-source approach. He is passionate about democratizing legal knowledge through AI, backed by strong academic credentials and experience leading major product and innovation initiatives.
Alex Denne, Head of Growth (Open Source Law) at Ƶ, is a legal tech leader and serial founder with over a decade of experience driving innovation and making legal services more accessible. Since joining in 2021, he has scaled the platform from 200 to over 120,000 users, combining deep contract law expertise with a data-driven, open-source approach. He is passionate about democratizing legal knowledge through AI, backed by strong academic credentials and experience leading major product and innovation initiatives.