Introduction
With the widespread adoption of AI in the legal profession, certain critical challenges, one of which is hallucinations—a phenomenon where AI models perpetuate plausible but inaccurate responses—came to the fore, necessitating verification of their generated content.Footnote 1 Advanced AI solutions have revolutionized the legal domain, including legal education, research, and practice, and have performed exceptionally on law school and bar exams as well as with legal analysis.Footnote 2 AI enables machines to simulate human intelligence, inspiring them to learn patterns from the data they are trained on, solve problems, and make decisions. Lawyers are employing AI to augment legal services, and many of the world’s leading firms have started using AI in rendering legal services.Footnote 3
Nevertheless, these tools are not completely without risk and embrace ethical challenges, such as bias, data privacy, and fictitious outcomes, necessitating counter-verification of their generated content in the forms of citations, case law, statutes, quotations from superior court opinions, and legal arguments.Footnote 4 In response to users’ prompts, AI generates new original content based on the machine-learning model that mirrors human intelligence and decision-making competence. With the development of Large Language Models (LLMs), these generative AI tools have gained more significance, creating human-like text based on the vast data on which these models are being trained. LLMs generate content by forecasting the next element in the sequence without any certainty of its accuracy, which could lead to generating misinformation.Footnote 5 Based on this probabilistic nature, in response to users’ input, AI models may generate erroneous outcomes, which often sound plausible, and users assume they are accurate.Footnote 6
ChatGPT, for instance, is an LLM developed by OpenAI, which secured itself as the fastest-growing consumer application after its release in November 2022.Footnote 7 Following the trend, other tech companies have launched AI tools, such as Microsoft Copilot, Lexis+ AI, and Llama. However, these tools are in their infancy and are susceptible to producing inaccurate or made-up responses, necessitating their assessment and reliability.
Though AI tools augment legal services, they pose risks like malpractice and legal liability and cause a miscarriage of justice if presented or relied upon without verification of their precision. AI solutions tend to produce plausible but inaccurate responses, which are referred to as hallucinations. Legal hallucinations refer to the phenomenon where AI models perpetuate fictitious precedents, misinterpret legal doctrines and statutes, provide inaccurate legal advice, or generate fabricated legal content. In the given scenario, the end user is liable to validate the outcomes of AI tools before making them part of their court filings or tendering AI-based legal advice.
There are various contributing factors leading to hallucinations: (1) limitations in training data, where AI models are trained on deficient or prejudiced datasets, and their generated content can be erroneous; (2) overconfidence in prediction, where AI models may produce results or generate content, often fabricating details instead of accepting uncertainty; and (3) model complexity, where due to the AI models’ complexity and extensive data processing requirements, problems like overfitting, training data gaps, over-optimization, and a lack of grounding in real-world knowledge, can make LLMs prone to hallucinations.Footnote 8
LLMs by design cannot differentiate between the truthfulness and falsehood of the generated content because these models mirror patterns from the data they are trained on without recognizing their precision. Hence, accuracy in these models’ responses is often inadvertent.Footnote 9 The models are not infallible and can perpetuate fabricated responses, creating real-world challenges where accuracy is paramount. Considering their fallibility, it is critical to understand the nature of their responses and the end user’s corresponding duty to double-check these models’ outcomes before using or relying on them.
Relevant Literature
By applying Retrieval-Augmented Generation (RAG), a technique that combines the capabilities of a pre-trained LLM with an external data source, legal enterprises like LexisNexis and Thomson Reuters have claimed, though not supported by any empirical evidence, to have significantly mitigated the risk of hallucinations.Footnote 10
A recent study proved that commercially available RAG-based, AI-driven legal tools still hallucinate. The study was conducted on Lexis+ AI, offered by LexisNexis; Westlaw AI-Assisted Research, produced by Thomson Reuters; Ask Practical Law AI, another Thomson Reuters product; and ChatGPT, offered by OpenAI. The research exhibited that 16.67 percent (1 out of 6) of Lexis+ AI and Ask Practical Law AI queries produced misleading or false information, whereas the rate for Westlaw was 33.33 percent (1 out of 3), although it was less prone to hallucinate in comparison with GPT-4. The frequency of accuracy reported in these tools was 65 percent for Lexis+ AI, 41 percent for Westlaw, and 19 percent for Practical Law AI.Footnote 11
LLMs can transform legal operations, but this potential is hindered by their tendency to hallucinate, which is a critical challenge to the widespread adoption of LLMs. Another significant study examined hallucinations in open-domain settings, where LLMs were required to handle a wide range of legal prompts to provide exact responses to open-ended questions. The study revealed legal hallucinations of 69 percent with ChatGPT 3.5 and 88 percent with Llama 2. The results indicated that LLMs often provide superficially legitimate but incorrect replies to counterfactual legal queries. The research proved that LLMs such as ChatGPT are susceptible to generating responses incompatible with prevalent doctrines and case law. AI models aspire to provide legal services and information, but their inherent shortcomings in providing reliable and accurate information greatly obstruct this intent.Footnote 12
Hallucinations in legal operations are a real-world challenge and should not be considered merely a theoretical concern. Plausible but fictional outcomes could necessitate novel methods for employing LLMs in situations where accuracy is paramount and inaccuracy is unacceptable, such as in legal, financial, and medical scenarios. According to a study undertaken on GPT 3.5, Llama 2, and PaLM 2, LLMs hallucinate from 69 percent to 88 percent of the time when responding to specific legal queries. These models are oblivious to their errors and bolster incorrect legal presumptions, fostering concerns about their authenticity and reliability and underscoring the significance of vigilant adoption of AI-powered solutions into the legal province.Footnote 13
For example, the following instance demonstrates the breadth of hallucinations in legal operations: A summarization tool was employed to condense a lengthy document. AI fabricated certain legal terms and omitted critical features, resulting in a summary draft that misconstrued the original draft. Hallucinations of this kind could have severe repercussions where precision is inevitable.Footnote 14
Fabricating citations is another critical instance of hallucinations. AI models can often hallucinate by creating fake references, apparently real, with all the information about the title, author, journal, periodical, or, in the instance of court precedent, the names of parties with a complete citation, etc. However, upon meticulous examination, the source does not exist but is hypothetically generated. AI models would likely struggle with more complex or nuanced Bluebook citations.
In addition, while utilizing AI solutions, legal professionals may face challenges in comparing cases where the court reached opposite conclusions due to the nuanced nature of legal reasoning and interpretation. AI systems might encounter difficulties in understanding context, as cases depend on specific facts and jurisdictional nuances. Likewise, the legal language in opinions can be complex, often filled with jargon, especially when cases arrive at contrasting conclusions owing to jurisdictional differences, statutory interpretations, or factual circumstances.
For example, in a motion to dismiss filed in the US District Court for the District of Connecticut, a lawyer might like to compare two court cases reaching opposite conclusions on the same issue. After researching cases, the lawyer might cite Castagno v. Wholean Footnote 15 (reversing the dismissal of a grandparent’s visitation petition for lack of jurisdiction in the absence of death, divorce, or a child custody proceeding) and Frame v. Nehls Footnote 16 (holding that the state’s grandparent visitation statute did not apply in a paternity action because it was not an action for child custody). Though both cases involve grandparents’ visitation rights, they differ in their legal interpretations, underscoring the importance of jurisdictional and statutory context in legal analysis.
Some researchers argue that AI-powered solutions might take over most of the tasks typically performed by lawyers, potentially multiplying access to justice at the cost of legal jobs.Footnote 17 However, these studies mainly focus on AI alone and miss the most significant part: the human interaction with these AI systems. Due to their inherent limitations, it is more likely that AI will assist humans rather than replace them.Footnote 18 Current research on AI often overlooks the interaction between humans and machines, missing the most probable application of AI in the legal field.
Methods
Quantitative and comparative methods have been employed to conduct the proposed research, which was conducted between December 2024 and April 2025.Footnote 19 To check the accuracy, completeness, and hallucinations produced by ChatGPT-4, Copilot, DeepSeek, Lexis+ AI, and Llama 3, we prepared fifty legal questions divided into the following categories: constitutional doctrines; general law questions regarding new legal developments; questions requiring description; close-ended questions; and false propositions. We prompted each tool with the same questions and compared its responses with the correct answers in a spreadsheet. We calculated the performance of each tool separately in percentages for accuracy, incomplete response, and fabricated outcome. At the end of the analysis, we also demonstrated these tools with the help of charts. To quantify the performance of these AI tools, we employed three variables to evaluate the performance measure and accuracy-to-hallucination ratio. Through pseudocode, we represented the responses of these tools.
AI Tools: Comparative Performance Analysis
Given their probabilistic nature, AI solutions are prone to producing fabricated outcomes despite advancements in their performance. The following segment quantifies responses of ChatGPT-4, Copilot, DeepSeek, Lexis+ AI, and Llama 3 in terms of their accuracy, incomplete responses, and the hallucinations they create. Through pseudocode and mathematical equations, we drew a comparison of these tools by calculating the accuracy-to-hallucination ratio, providing insight into the decision-making tasks of these tools.
Mathematical Investigation
To mathematically represent the performance of AI-powered legal tools in terms of accuracy, incomplete responses, and hallucinations, we defined the following variables:
$ {A}_i $
: Accuracy of the tool
$ i $
, expressed as a percentage.
$ {I}_i $
: Percentage of incomplete responses generated by the tool
$ i $
.
$ {H}_i $
: Percentage of hallucinations produced by the tool
$ i $
.
Where
$ i\in \{1,2,3,4,5\} $
corresponds to the AI tools:
ChatGPT-4
Copilot
DeepSeek
Lexis+ AI
Llama 3
A performance vector can capture each AI tool’s performance (
$ {A}_i $
,
$ {I}_i $
,
$ {H}_i $
), where:
ChatGPT-4:
$ \left({A}_1,{I}_1,{H}_1\right)=\left(30\hskip0.3em percent,34\hskip0.3em percent,36\hskip0.3em percent\right) $
.
Copilot:
$ \left({A}_2,{I}_2,{H}_2\right)=\left(52\hskip0.3em percent,22\hskip0.3em percent,26\hskip0.3em percent\right) $
.
DeepSeek:
$ ({A}_3,{I}_3,{H}_3)=(50\hskip2pt percent,20\hskip2pt percent,30\hskip2pt percent). $
Lexis+ AI:
$ \left({A}_4,{I}_4,{H}_4\right)=\left(58\hskip0.3em percent,22\hskip0.3em percent,20\hskip0.3em percent\right) $
.
Llama 3:
$ \left({A}_5,{I}_5,{H}_5\right)=\left(36\hskip0.3em percent,28\hskip0.3em percent,36\hskip0.3em percent\right) $
.
Total Performance Measure
To evaluate the overall performance of each AI tool in terms of accuracy, incomplete responses, and hallucinations, we defined a total performance measure
$ {P}_i $
for each tool:

This equation summarizes the three key performance metrics, allowing for a holistic comparison of each tool’s behavior.
Accuracy-to-Hallucination Ratio
A more insightful evaluation of the balance between accuracy and hallucination rates can be made by calculating the accuracy-to-hallucination ratio
$ {R}_i $
:

This ratio helps to quantify how well each tool performs in terms of accuracy relative to the hallucinations it generates. A higher value of
$ {R}_i $
indicates that the tool maintains better accuracy with fewer hallucinations, while a lower value suggests a higher propensity for hallucinations compared to its accuracy.
Comparative Analysis
The above metrics and ratios can be used to perform comparative analyses between the AI tools, offering insights into their strengths and weaknesses in legal decision-making tasks. Graphical representations, such as bar charts or radar plots, can further aid in visualizing the performance differences across accuracy, incomplete responses, and hallucinations for each tool.
Pseudocode Representation
Define performance for each tool as (accuracy, incompleteness, hallucinations)
ChatGPT = (30, 34, 36)
Copilot = (52, 22, 26)
DeepSeek = (50, 20, 30)
Lexis+AI = (58, 22, 20)
Llama 3 = (36, 28, 36)
List of tools
Tools = [ChatGPT, Copilot, DeepSeek, Lexis+ AI, Llama 3]
Function to calculate the total performance
Function TotalPerformance(tool):
Return tool[0] + tool[1] + tool[2]
Function to calculate accuracy-to-hallucination ratio
Function AccuracyToHallucination(tool):
If tool[2] > 0:
Return tool[0] / tool[2]
Else:
Return 0
Loop through tools and print results
For each tool in Tools:
Print “Total Performance:”, TotalPerformance(tool)
Print “Accuracy-to-Hallucination:”, AccuracyToHallucination(tool)
End
Graphical Representation
The following charts demonstrate the frequency of accuracy and incomplete responses, as well as the tendency to generate hallucinations in these AI-powered solutions:
Figure 1 represents accuracy in AI-generated legal content through tools employed in the survey: the survey shows that Lexis+ AI secured the highest accuracy at 58 percent, followed by Copilot at 52 percent, and DeepSeek at 50 percent. Llama 3 provided 36 percent accuracy, whereas ChatGPT-4 secured the lowest accuracy at 30 percent.

Figure 1. Percentage of Accuracy.
Figure 2 exhibits incomplete responses: DeepSeek outperformed all the comparative tools by producing 20 percent incomplete responses. Copilot and Lexis+ AI yielded 22 percent incomplete responses. Llama 3 and ChatGPT-4 perpetuated 28 percent and 34 percent incomplete responses, respectively. However, the risk associated with the use of DeepSeek is critical. A recent study reported huge security risks concerning the use of DeepSeek, where a significant data breach exposed the sensitive information of over one million users. ClickHouse, DeepSeek’s database, remained unprotected, permitting full control over database operations. This exposed a breach of sensitive information, such as chat histories and other critical data, which raised concerns regarding data management, practices, and compliance with privacy laws.Footnote 20 Consequently, several countries and organizations have banned DeepSeek. Experts have criticized its security protocols, highlighting the potential for data exploitation by criminals, necessitating robust security protocols in AI-driven platforms to protect users and maintain trust.Footnote 21

Figure 2. Percentage of Incomplete Responses.
Figure 3 indicates how often these tools generate hallucinations through conceivable but fictional responses. The susceptibility to hallucinations is increased when an inaccurate fictitious prompt is given: ChatGPT-4 and Llama 3 produced 36 percent fictitious responses. DeepSeek produced 30 percent fabricated responses. Copilot had 26 percent made-up responses, while Lexis+ AI outperformed all the comparative tools by reporting the lowest number of fabricated responses at 20 percent.

Figure 3. Percentage of Legal Hallucinations.
Figure 4 shows the proportions of responses observed in these tools, characterized as hallucinations, incomplete responses, and accuracy: ChatGPT-4 reported 30 percent accuracy, 34 percent incomplete responses, and 36 percent fabricated responses. Copilot produced 52 percent correct, 22 percent incomplete, and 26 percent fabricated responses. DeepSeek had 50 percent precision, 20 percent incomplete, and 30 percent fabricated responses. Lexis+ AI demonstrated 58 percent precision, 22 percent incomplete, and 20 percent fabricated responses, while Llama 3 achieved 36 percent correct, 28 percent incomplete, and 36 percent fabricated legal content. It was also observed that case law from lower courts was subject to more hallucinations than the superior courts. Interestingly, the frequency of hallucinations multiplied with a wrong or fictitious prompt. The tools advance and construct fabricated content in support of made-up citations instead of correcting or negating the existence of the proposed authority.

Figure 4. Proportion of Responses.
The Application and Implications of the Rag Model
Commercial services use Retrieval-Augmented Generation (RAG) to enhance their capabilities. All the AI tools deployed in this research leverage the RAG model, tailored to their use and strengths. ChatGPT-4, for instance, utilizes it to retrieve relevant information from external sources to improve conversational accuracy. Copilot uses it to ground its responses collected from credible, authoritative sources. DeepSeek models support RAG capabilities and are designed to combine retrieval-based and generative approaches to produce more accurate and relevant results.
Llama 3 uses RAG for handling large datasets for precise responses. Lexis+ AI uses GraphRAG—an advanced method to combine graph databases and RAG—to deliver high-quality insights with coherent and contextually relevant outcomes. It helps reduce AI hallucinations by leveraging structured data represented as a graph.Footnote 22 Though AI tools adapt the RAG model to their exclusive requirements, the core objective of combining retrieval with generation remains constant across all these tools.
In this technique, a prompt proposed to an LLM is first transferred to a retrieval system to search for a text-based source database, which creates a list of relevant documents. The retrieved data with the prompt is referred to the LLM to base its response exclusively on those documents, which we may refer to as an external database. By basing an LLM response exclusively on documents found, the chances of hallucinations could be significantly reduced, and it may also address any knowledge gap in an LLM by producing a comprehensive, accurate, and relevant response. The RAG opens new avenues to a more transparent and conceivable interaction with these models by allowing the end users to consider the same source that the LLM used to craft a response. RAG anchors LLMs’ responses to information retrieved from an external database, thus ensuring the byproduct is factual and verifiable. Though RAG furthers understanding and trust in the LLMs, it limits the creativity of the generated response by design.Footnote 23 The above diagram conceptualizes RAG architecture:
Searching for a particular principle of US law, for instance, may involve the following steps: First, the process begins with a user inputting a prompt into the system, as demonstrated in Figure 5. In the second step, the prompt is forwarded to a retrieval system assigned to search a large database to identify relevant keywords, principles, or documents commensurate with the query. In the third step, the retrieved data and the prompt are transferred to an LLM, which uses the referred data to produce a well-informed, coherent, and contextually relevant response in the given case-relevant principle. In the fourth step, the LLM’s generated response is delivered to the user within the same chatbot interface where the user initially entered their prompt.

Figure 5. RAG Model.
Though the application of RAG models has contributed to the accuracy of responses, these models have confronted certain inherent challenges: complexity and the extraction of data from various sources—specifically, unstructured data, which could be complex.Footnote 24 RAG models manage both retrieval and generation, making them more complicated than the standard generative models.Footnote 25 Accuracy and scalability, which ensure the relevancy and accuracy of the retrieved data, are critical since poor retrieval can lead to the incorporation of inaccurate and irrelevant information in the model’s response.Footnote 26
Similarly, scaling the RAG model to handle voluminous data efficiently is challenging while concurrently dealing with the storage and retrieval of vector embeddings.Footnote 27 Some of these challenges are the following: 1) latency—the RAG model can slow down the performance of the AI tool, as the retrieval process can cause latency, making the RAG model slower;Footnote 28 2) creativity—based on their reliance on the retrieved data, the models’ creativity may be constrained by the scope and quality of the data being accessed;Footnote 29 3) ambiguities—RAG may struggle with complex words having multiple meanings and contexts, leading to the retrieval of irrelevant or incorrect information placed into the model’s response;Footnote 30 and 4) dependency—RAG models largely depend on the data’s quality and comprehensiveness, which could lead to poor model performance where the dataset is biased, inaccurate, or incomplete.Footnote 31 A data validation process with vigilant and persistent improvements can diminish these challenges and improve AI tools’ self-learning.
Domain-specific and geographical scenarios are the least addressed real-world challenges to RAG models’ performance of the high-quality structured data in diverse domains like law, finance, and healthcare, which might be intricate, necessitating specific terminologies that might affect retrieval and generation quality.Footnote 32 The efficacy of RAG models is subject to their understanding of domain-specific knowledge, which may require fine-tuning and intelligent training at the production level. Geographical differences in language and dialects are another impediment to the performance of the RAG models. A system trained on specific data from one region may struggle with queries from another region, owing to linguistic or dialectal variations. Cultural context is essential for generating relevant and precise responses. So, the RAG models should be sensitive to cultural nuances that change across various geographical regions. Localized data sources significantly contribute to the RAG models’ performance.Footnote 33 For instance, a model trained and designed for use in Pakistan would perform better if it had access to local information and databases. Likewise, a system designed for the US, trained on local datasets and information sources, might perform adversely in other geographical regions or when generating responses about other jurisdictions.
The following techniques can help optimize the domain- and region-specific complexities and challenges of the RAG models’ performance: (1) Fine-tuning models on region- and domain-specific datasets, including curated datasets reflecting specific needs of the focused domain or region, can considerably improve their performance; (2) A hybrid approach to combining rule-based methods for structured data with machine-learning technologies for unstructured data can help largely resolve domain-specific issues;Footnote 34 and (3) Employing RAG systems on optimized infrastructure, such as using cloud-based platforms like AWS, Google Cloud, or Microsoft Azure, along with hardware optimization, can significantly enhance query processing and the overall performance of the models across different domains and geographical scenarios.
Results and Discussion
Considering the responses to the fifty legal prompts, Lexis+ AI outperformed other AI comparative tools, but all these tools still hallucinate due to their probabilistic nature. Unlike ChatGPT-4, Copilot, DeepSeek, and Llama 3, which are general-purpose AI models, Lexis+ AI specializes in legal operations. Lexis+ AI is trained in legal-specific data, which enhances its accuracy and reliability for legal research and analysis. It integrates products like case law searching, legal drafting, and analytics within a single platform, making it more productive for legal professionals. With the additional features of security and data privacy, Lexis+ AI offers a unique user interface tailored to legal professionals, making it easier to navigate and use for specific legal tasks.Footnote 35
Despite all these features, our research demonstrates that Lexis+ AI still perpetuates hallucinations and incomplete responses. Hence, its generated content cannot be relied upon or presented without counter-verification through conventional validation methods.
Employment of the RAG system overcomes the chances of hallucinations; however, all these tools are still prone to fabricated responses at different rates. Like other modern language models, DeepSeek is built on transformer architecture, which enables the model to produce proficient results by fusing advanced neural networks. These models are mainly trained with a combination of supervision and learning from human feedback, tailored to human preferences and values.Footnote 36
Moreover, the responses generated by these comparative tools substantiated variations with domain and region-specific queries, necessitating fine-tuning and infrastructural optimization of RAG models, including hardware optimization through graphics processing units (GPUs) or AWS Graviton processors, cloud-based solutions, efficient data management, and intelligent data training at the production level. The following segment concludes the research by further contributing to mitigating factors of the prospective hallucinations.
Recommendations and Conclusion
The following are some recommendations for developers of AI models and legal professionals who are dealing with these models as end users:
-
(1) The AI tools’ responses are significantly grounded in the quality of training data. Designers should ensure high-quality training data through training samples and subsequent testing through supervised and self-supervised learning. Blind datasets should be augmented based on their performance at the production level. Self-supervised learning is a process whereby a model generates labeled data from the trained data during the learning process without relying on externally provided labels, leveraging intrinsic structures within the input data to create deep-training signals. The complexities of the data queried at the production level, which models could not resolve, should be introduced and incorporated into the training datasets of supervised networks. Additionally, the complexities of self-labeling, a key aspect of self-supervised learning, should be improved at the production level.
-
(2) Overfitting is another frequent problem in machine learning. A model may perform exceptionally with training data but poorly with testing or validation data because of overfitting.Footnote 37 The latter happens when a model memorizes the training data, has too many parameters, is trained for too long, or is attuned to specifics and noise to such an extent that it fails to perform well on new unseen data. This makes the model overly complex and captures the random fluctuations in the training data rather than the desired output. The data for learning needs to present all levels of challenges and complexities as presented in production. Real-world search inquiries must be frequently added to the training of LLMs so that the networks learn the intrinsic patterns. Limited data availability can also lead to overfitting, so the dataset needs to be augmented intelligently to enhance the network’s performance.
-
(3) Imposing limitations on AI outcomes can contribute to LLMs’ performance. The application of sophisticated artificial neural network models on specific legal data and their combinations can improve the performance of the AI models.
-
(4) Developers should continuously improve and update the AI models by subsequent testing and getting feedback from end users through a process we may refer to as human-in-the-loop (HITL).
-
(5) Human oversight is recommended to validate the responses of AI tools by involving an expert in local laws. For instance, Leap AI has already initiated an optional verification process in which the AI-generated content is sent to an expert attorney to validate the accuracy of the AI-generated content.
-
(6) Developers may also consider fine-tuning models for specific legal tasks by adjusting the models’ parameters to learn new tasks.
-
(7) By designing AI tools to justify and give reasons for their decision-making, developers can significantly enhance the reliability of the generated content. The models’ self-explanatory attributes should be part of the RAG architecture. Based on its probabilistic nature and black box system that operates without providing information regarding its internal operation for decision-making, reasoned decision-making can significantly contribute to the overall efficacy and performance of AI tools.
-
(8) Designers should arrange online training sessions and guidelines on using AI tools for legal services without compromising accuracy, which is paramount in law.
-
(9) AI should assist but not replace humans, and legal professionals must be cautious about using AI tools for legal assistance. AI solutions should be employed to augment legal services, but not at the cost of subjective judgments.
-
(10) AI-generated content must be counter-verified through cross-referencing with traditional databases, peer reviews, consultation with senior colleagues, and other rounds of swift checking before presenting in the courts.
-
(11) Legal professionals should master precise prompt-crafting skills that help AI yield accurate responses.
-
(12) Legal professionals must be abreast of AI innovations and best practices with continuous audits of AI tools, ensuring their compatibility with legal standards, relevant laws, and policies.
-
(13) To avoid the apprehension of malpractice and civil liability, legal professionals must disclose to their clients the use of AI and keep the court updated on the extent of their reliance on AI, which is part of their ethical and professional responsibilities.
-
(14) Legal professionals should coordinate with developers to update the AI tools and report when these tools produce fabricated or inaccurate responses.
To conclude, AI is in its transitional phase of revolutionizing legal services. It exhibits ethical challenges like data privacy, bias, and fabricated responses. Existing research has demonstrated that Lexis+ AI is the most reliable and authentic tool for legal assistance, but it is still subject to hallucinations. The likelihood of fictitious responses amplifies where the dataset these AI tools are trained on is inaccurate, biased, or otherwise flawed, making an inverse relationship between hallucinations and accuracy. With the recommendations given, the overall performance of AI tools can be significantly improved to overcome the existing issue of hallucinations.