Launched in 2024, the Allens AI Australian Law Benchmark assesses the ability of LLMs to provide legal guidance effectively.
Which LLMs were tested?
We tested the following publicly available LLMs from the following developers:
|
Methodology
|
LLMs not tested
Notably, the 2025 edition excludes various legal-industry specific tools such as Harvey, Thomson Reuters's CoCounsel and LexisNexis's Lexis+ AI.
While Allens continues to benchmark and leverage several of these (and other) products, none are included in this report to avoid drawing broad performance comparisons between general purpose products and the products that have been optimised for specific use-cases or user-interactions, as well as (increasingly) 'multi-model' products that use different AI models for different tasks. Our use of 'single prompt' questioning (for example) to ensure consistency between LLMs may not best demonstrate the capabilities of products that offer multiple structured pathways for answering the same question.