Skip to content

2.8 Inherent Technical Limitations of AI

The Halos and Shadows of Intelligence: Understanding AI’s Inherent Technical Limitations

Section titled “The Halos and Shadows of Intelligence: Understanding AI’s Inherent Technical Limitations”

Artificial Intelligence (AI), particularly modern AI systems represented by Machine Learning (ML) and Deep Learning (DL), often appears like a prodigiously talented newcomer. Its remarkable ability to enhance efficiency and handle complex tasks frequently cloaks it in a dazzling aura of “intelligence.” However, law, as a discipline rooted in logic, evidence, fairness, and justice, with an utmost demand for rigor, requires its practitioners to possess sharp insight and prudence when embracing the disruptive changes brought by AI. We must profoundly recognize that current AI technology is far from the mythical “omniscient and omnipotent” being; it harbors a series of deep-seated, arguably inherent technical limitations.

These limitations are not mere programming bugs or temporary technological bottlenecks. They are deeply rooted in AI’s core principles (like the nature of statistical learning), the methods of acquiring and processing training data, and the design philosophies of current mainstream algorithms. In the high-risk, high-stakes legal field, ignoring or downplaying these limitations can trigger severe, even catastrophic consequences – ranging from incorrect legal information retrieval and drafting invalid contract clauses, to discriminatory case risk assessments, flawed judicial decision support, threats to system security, and potential erosion of fundamental rule-of-law principles.

This section aims to dissect the major challenges and inherent shortcomings facing current AI technology, like sketching the shadows alongside the bright star, with a particular focus on their specific implications and warnings for legal practice.

I. Data Dependency: The Foundation and Achilles’ Heel of Intelligence

Section titled “I. Data Dependency: The Foundation and Achilles’ Heel of Intelligence”

The “intelligence” of modern AI (especially ML and DL models) largely stems not from genuine understanding, reasoning, or consciousness, but from “digesting” massive amounts of data and learning complex statistical patterns and correlations within it. This extreme reliance on data is both the source of its powerful capabilities—being “well-read”—and constitutes its fundamental vulnerability—a form of “determinism by origin.”

The “No Bricks Without Straw” Dilemma: A Dual Test of Quantity and Quality

Section titled “The “No Bricks Without Straw” Dilemma: A Dual Test of Quantity and Quality”

The performance of AI models exhibits an almost iron-clad positive correlation with the quantity and quality of their training data. The industry adage—“Garbage in, garbage out” (GIGO)—aptly captures this harsh reality.

  • Insufficient Quantity: If the amount of data used to train a model is inadequate, it cannot learn sufficiently rich and robust patterns. Its capabilities will be severely limited, like a student who has only read a few books trying to tackle complex, diverse problems.

  • Poor Quality: If the training data quality is compromised—containing significant noise (errors, irrelevant information), incorrect labels (e.g., labeling a valid contract as invalid), missing crucial information (e.g., case records lacking key factual descriptions), or being poorly formatted and inconsistent—the model is likely to “learn badly,” internalizing erroneous or distorted patterns and producing misleading outputs.

  • The Harsh Reality in Legal Practice: Acquiring large-scale, high-quality, and professionally, accurately labeled datasets in the legal domain is often an extremely arduous and costly task. This is not only because legal data frequently involves highly sensitive privacy and confidentiality issues (client information, case details, trade secrets) but also due to:

    • High Labeling Costs: Accurate annotation of legal documents (e.g., identifying clause types, assessing risk levels, marking argumentation relations) requires substantial time investment from lawyers or legal experts with deep domain knowledge, resulting in high labor costs.
    • Labeling Consistency Challenges: Many legal concepts (“reasonableness,” “material impact,” “good faith”) inherently involve interpretation and subjective judgment. Annotations from different experts may vary, affecting data quality.
    • Data Standardization Difficulties: Legal texts like judgments, contracts, and regulations vary widely in format, terminology, and argumentation style, making standardization for machine processing a challenge in itself.
    • Difficulty Capturing Tacit Knowledge: Much valuable legal knowledge and experience reside in lawyers’ discussions, memos, informal communications, and personal experience—this unstructured, tacit knowledge is hard to capture directly for model training.

    Therefore, the reliability and performance ceiling of many seemingly promising legal AI applications (like accurate case prediction or perfect automated contract generation) are often constrained by data availability bottlenecks. Blindly trusting models trained on limited or low-quality data for critical legal analysis, risk assessment, or decision support is akin to building professional judgment on quicksand—it is extremely dangerous.

III. The “Black Box” Dilemma: Elusive Decision Logic and Lack of Explainability

Section titled “III. The “Black Box” Dilemma: Elusive Decision Logic and Lack of Explainability”

Modern AI, especially advanced models based on Deep Neural Networks (DNNs), often possess incredibly complex internal structures, potentially involving millions, billions, or even trillions of interconnected parameters. This extreme complexity leads to a significant problem: even when a model provides seemingly highly accurate predictions or decisions, we often cannot clearly and intuitively understand the specific reasons, internal logic, or key evidence it relied upon to reach that judgment. This phenomenon is aptly termed the “Black Box Problem.”

The Predicament of “Knowing It Works, But Not Knowing Why”

Section titled “The Predicament of “Knowing It Works, But Not Knowing Why””

Imagine a sophisticated AI legal analysis tool (perhaps based on an LLM) reads a complex contract and outputs: “Clause X of this contract poses a significant legal risk.” Or an AI image recognition system analyzes a surveillance screenshot and flags “suspected possession of contraband.” When we probe further—“Why did you reach this conclusion?” “What specific facts or logic led to this judgment?”—these “black box” models often fail to provide a satisfactory explanation that aligns with human cognitive habits.

We might not know precisely:

  • Which key features in the input information (contract text, image pixels) did the model focus on?
  • How were these features combined, weighed, and transformed internally by the model?
  • What reasoning path or decision rules did the model follow internally?
  • How was the model’s Confidence Score calculated, and is it reliable?

We can only observe the final input and output; the intricate information processing in between remains opaque, like a black box whose inner workings are hidden from view.

Law’s “Rigid Demand” for Explainability

Section titled “Law’s “Rigid Demand” for Explainability”

In many technical application areas, “black box” models might be acceptable based on their superior performance (e.g., you might not care how a recommendation system works if it accurately suggests movies you like). However, in law—a field demanding logical rigor, evidentiary sufficiency, procedural fairness, and reasoned justification—a lack of explainability is often an intolerable, even fatal flaw:

  • Accountability and Liability Attribution: If an AI-based system provides faulty legal advice, makes improper auxiliary decisions (e.g., wrongly marking key evidence as irrelevant), or its actions cause harm, but we cannot understand its decision process, then:

    • Tracing the cause of error becomes difficult: Was it bad data? A flawed algorithm? Insufficient training?
    • Identifying the liable party becomes challenging: Was it the AI developer? The deployer? The data provider? Or the user who relied on the AI’s decision? This renders effective Accountability mechanisms illusory, contradicting the principles of the rule of law.
  • Due Process and Right to Appeal: A cornerstone of legal decision-making is the requirement for Reasoned Decision-Making and procedural transparency. Whether it’s an administrative agency’s decision or a court’s judgment, parties have the Right to Reasons, which is fundamental to effectively exercising their right to appeal or review. If a decision significantly affecting a party’s rights and obligations relies heavily on an unexplainable “black box” AI model (even for assistance), it could violate the party’s Right to Information and Due Process.

  • Reliability Assessment and Trust Building: Lawyers, judges, regulators, and the public will struggle to fully trust a system that cannot explain its workings. If we don’t understand how AI reaches conclusions, it’s hard to:

    • Assess the reliability of its conclusions: We can’t determine if they are based on correct legal principles and factual evidence.
    • Detect potential flaws: Hidden biases, unusual sensitivities to certain inputs, logical loopholes, or knowledge gaps within the model become difficult to uncover.
    • Build genuine trust: Especially when dealing with complex, contentious cases or those involving significant interests, trust is foundational for human-machine collaboration.
  • Debugging, Correction, and Continuous Improvement: When a model errs or performs poorly, if the cause cannot be understood (e.g., which part of the reasoning went wrong? What incorrect information influenced it?), then effective debugging, targeted correction, and continuous model improvement become extremely difficult. This hinders the enhancement of AI system reliability.

  • Meeting Regulatory Compliance Requirements: As AI applications become widespread, more laws and regulations are focusing on the transparency and explainability of algorithmic decision-making. For example:

    • The EU’s General Data Protection Regulation (GDPR) grants data subjects the right, under certain conditions, to obtain “meaningful information about the logic involved” in automated decisions, including profiling.
    • Highly regulated sectors like finance and healthcare also impose explicit requirements on model risk management, validation, and transparency. Legal AI applications must also meet these growing compliance demands for explainability.

Exploring the “Black Box”: Efforts and Limits of Explainable AI (XAI)

Section titled “Exploring the “Black Box”: Efforts and Limits of Explainable AI (XAI)”

To address the severe challenges posed by the “black box” problem, the field of Explainable AI (XAI) emerged. XAI aims to develop techniques and methods to explain the logic and basis behind AI models’ predictions or decisions, enabling humans to understand, trust, and effectively manage these intelligent systems.

Current mainstream XAI techniques can be broadly categorized into two types:

  • Intrinsic Interpretability: This approach involves selecting or designing models whose structures are relatively simple and whose working mechanisms are inherently understandable by humans. Examples include:

    • Linear Models (Linear Regression, Logistic Regression): Their decision boundaries are linear, and weights directly reflect feature importance.
    • Decision Trees: (Especially shallow ones) Their decision process can be visually represented as a series of “If-Then” rules.
    • Rule Lists: Composed directly of an ordered set of rules.
    • Generalized Additive Models (GAMs): Represent predictions as sums of non-linear transformations of individual features. The advantage is that explanations come directly from the model itself, usually being reliable and Faithful. The downside is that such models typically have limited expressive power and often struggle to match the performance (e.g., prediction accuracy) of complex “black box” models (like deep neural networks) when dealing with high-dimensional data exhibiting complex non-linear relationships. There’s a Trade-off between performance and interpretability.
  • Post-hoc Explanation: This approach accepts the use of complex “black box” models (due to their typically better performance) and attempts to provide some form of explanation “after the fact”—once the model has made a prediction or decision. There are many such methods, common ones include:

    • Feature Importance Analysis: Aims to assess which input features had the most influence on a specific prediction outcome.
      • Global Importance: Analyzes which features are most important for the model’s overall performance (e.g., Permutation Importance).
      • Local Importance: Explains, for a single prediction, which input features were key drivers of that result. Representative methods are LIME (Local Interpretable Model-agnostic Explanations) (approximates the black box model’s decision boundary near a specific sample using a simple, interpretable local surrogate model) and SHAP (SHapley Additive exPlanations) (based on Shapley values from game theory, assigns a contribution value to each feature for a prediction).
    • Model Distillation/Surrogate Models: Attempts to train a simpler, interpretable model (like a decision tree or linear model) to mimic the input-output behavior of the complex “black box” model. Understanding the simpler surrogate model indirectly helps infer the logic of the black box.
    • Rule Extraction: Attempts to extract a set of “If-Then” rules from a trained “black box” model (especially neural networks) that approximate its behavior.
    • Example-based Explanation: Provides explanations by identifying training data samples most similar to the current prediction case (Prototypes), or by finding critical samples that, if slightly changed, would alter the prediction outcome (Counterfactual Explanations). E.g., “The system flagged this contract as high-risk because it is very similar to previous contracts A and B in clauses X and Y, both of which led to litigation.”

IV. “Hallucinations”: AI Confidently Talking Nonsense

Section titled “IV. “Hallucinations”: AI Confidently Talking Nonsense”

Beneath the dazzling glow of Generative AI (GenAI), particularly technologies represented by Large Language Models (LLMs), while they exhibit astonishing abilities in text creation, conversation, knowledge Q&A, and code writing, they also expose an extremely worrying phenomenon, particularly fatal in the legal domain: they sometimes confidently generate content that appears grammatically fluent, logically coherent, and professionally toned, but actually contains false information, severely contradicts objective facts, or is entirely fabricated. This phenomenon of AI “confidently talking nonsense” is vividly termed “Hallucination” by researchers and developers.

How Do “Hallucinations” Occur? Why Does AI “Lie”?

Section titled “How Do “Hallucinations” Occur? Why Does AI “Lie”?”

Understanding the root cause of “hallucinations” requires a deep recognition of how LLMs fundamentally work. They do not possess a “brain” storing factual knowledge like humans, nor do they have genuine logical reasoning or fact-checking capabilities. Instead, their output is generated based on extremely complex Statistical Patterns learned from massive amounts of training text data (from the internet, books, etc.).

When given an input prompt, an LLM’s core task is to predict the next most likely word (or Token, the basic text unit), append this predicted word to the existing text, and then continue predicting the next word based on the new sequence… This process repeats, ultimately “weaving” a text sequence that statistically looks the most natural, coherent, and consistent with the language patterns learned during training.

Their primary objective is “Fluency and Coherence,” not necessarily “Factual Accuracy.”

“Hallucinations” are more likely to occur under several conditions:

  1. Knowledge Boundaries & Outdated Information (Knowledge Cut-off): An LLM’s “knowledge” comes from its training data. If a user asks about something outside the scope of its training data (e.g., highly specialized, niche, or very recent information), or if its training data has a specific cut-off date, rendering its internal knowledge outdated, the model might “not know” the answer. However, to fulfill the generation task and maintain conversational flow, it might resort to “guessing,” “associating,” or even “fabricating” based on existing, vaguely related patterns, rather than honestly stating “I don’t know.”

    • Example: Asking a model trained on data up to 2022 about the specifics of a law amended in 2024 will likely yield an inaccurate answer, but it might invent plausible-sounding legal text to respond.
  2. Ambiguous or Leading Prompts: If the user’s prompt is unclear, ambiguous, or carries strong suggestions or presuppositions, it might lead the model to “associate” and generate along incorrect lines. The model might “play along” even if the premise is false.

    • Example: Asking, “Please detail the main contents of the ‘Lunar Resources Development and Management Act’?” Even though such a law doesn’t exist (in most jurisdictions), a less cautious LLM might obligingly invent several official-sounding clauses.
  3. Challenges with Complex Reasoning or Fact-Checking: Tasks requiring multi-step logical reasoning, cross-verification of multiple sources, or judgment amidst conflicting information are often difficult for LLMs relying solely on statistical patterns. Their “reasoning” is more like an “associative chain” based on pattern matching, prone to logical leaps, factual confusion, or self-contradiction within the chain.

    • Example: Asking an LLM to analyze a complex contract dispute involving cross-interpretation of multiple clauses and a complex evidence chain might yield a fluently written analysis, but closer inspection could reveal skipped crucial steps or misinterpretations of clause meanings.
  4. “Creative Leaps” When Aiming for Creativity/Diversity (Creativity vs. Factuality Trade-off): In scenarios designed for creative content generation (like writing poems, stories) or providing diverse answers, model parameters like “Temperature” might be set higher. This encourages “thinking outside the box,” which, while boosting creativity, can also increase the risk of deviating from facts and producing “hallucinations.”

Section titled “High-Risk Manifestations of “Hallucinations” in Legal Scenarios: From Misleading to Disastrous”

For the legal industry, where rigor and accuracy are paramount, AI “hallucinations” are not harmless technical glitches but potentially fatal flaws leading from severe misinformation to professional catastrophe. Their specific manifestations in legal contexts are particularly alarming:

  • Fabricating Case Law:

    • Harm: This is one of the most dangerous and notorious forms of “hallucination,” directly challenging legal professionalism. LLMs might generate completely non-existent case names, fictional courts, judges, decision dates, and even convincingly “cite” non-existent paragraphs or reasoning from fabricated judgments.
    • Deceptiveness: These fictional cases often look incredibly realistic, with proper formatting and potentially citing real legal terms or principles, easily fooling unwary lawyers.
    • Real-World Warning: There have been multiple real, shocking cases internationally (especially in the US) where lawyers faced severe court sanctions (fines, public reprimands, potential disciplinary action) for citing fake cases invented by AI tools like ChatGPT in court filings (e.g., the widely reported Mata v. Avianca case and subsequent similar incidents). This serves as the highest level of warning to the global legal community!
  • Misrepresenting Statutes/Regulations:

    • Citing a non-existent law section number.
    • Incorrectly stating the content, applicability conditions, or exceptions of a real legal provision.
    • Confusing provisions from different laws (or even different jurisdictions).
    • Citing repealed or amended outdated regulations.
  • Inventing Scholarly Sources:

    • Generating a fake paper title, author name, journal name, or publication year and citing or summarizing it convincingly.
    • Incorrectly attributing a viewpoint or argument to a well-known legal scholar or judge.
  • Providing Information Inconsistent with Case Facts:

    • When generating case summaries, background info, evidence timelines, etc., inaccurately describing key factual details like dates, locations, involved parties’ relationships, contract amounts, or performance status.
    • Confusing facts between different cases.
  • Producing Illogical or Irrelevant Arguments:

    • Generated legal analysis or arguments, though fluent and perhaps eloquent, might contain logical leaps, contradictions, or circular reasoning upon careful review.
    • Arguments presented might lack substantive connection to the point they claim to support, appearing relevant but being logically unsound.
Section titled “Countering “Hallucinations”: The Mandatory Verification Baseline for Legal Professionals”

Given the prevalence of “hallucinations” and their extreme danger in legal contexts, legal professionals using any generative AI tool (especially LLMs) must establish and strictly adhere to a comprehensive risk control and verification process:

  • Human Verification is the “Iron Law,” No Exceptions!:

    • Every piece of AI-generated content involving substantive facts, data, legal authorities (cases, statutes, doctrines), or source citations must undergo 100% independent, diligent, cross-verification.
    • Never treat AI as a reliable source of knowledge or a “fact engine.” View it as an assistant that might be eloquent but occasionally speaks falsehoods.
    • Verification must use authoritative, reliable sources: Official legal databases (like Westlaw, LexisNexis, official government legislation sites), official court websites, reputable case law databases, original academic literature, verified factual records, etc. Never use one AI to verify another AI’s output!
  • Leverage Retrieval-Augmented Generation (RAG):

    • Principle: RAG is a promising technique to partially mitigate “hallucinations.” The idea is: before letting the LLM generate an answer, first have it Retrieve relevant snippets of information from a trusted, up-to-date, domain-specific external knowledge base (e.g., a firm’s internal case database, a professional legal statute library, selected law review articles). Then, instruct the LLM to Generate its answer based on this retrieved factual information, rather than relying solely on its internal, potentially outdated or inaccurate “memory.”
    • Effect: This can help “anchor” the model’s response to reliable facts, reducing the likelihood of fabrication and improving relevance and accuracy to some extent.
    • Limitation: RAG is not a silver bullet. The retrieval process itself might be imperfect (failing to find the most relevant info, or retrieving inaccurate info); the LLM might still misinterpret, over-extrapolate, or introduce new errors when summarizing and generating based on retrieved info. Therefore, even with RAG, human verification remains the essential final safeguard.
  • Optimize Prompt Engineering:

    • Design more precise, specific, and constraining prompts to better guide the model towards facts and limit its “creative freedom.”
    • Explicitly instruct the model to state “I don’t know” when uncertain, rather than forcing an answer.
    • Request the model to cite its sources (though the cited sources themselves might be hallucinated, this provides a starting point for verification).
    • Avoid asking leading questions or questions based on false premises.
  • Utilize Model Fine-tuning:

    • If resources permit (sufficient high-quality data and computation), Fine-tuning a general foundation LLM on rigorously fact-checked, high-quality internal legal data (like a firm’s case precedents, standard contract templates, expert Q&A pairs) can be beneficial.
    • This helps the model better adapt to the specific terminology, norms, writing style, and knowledge base of a particular legal domain, potentially reducing domain-specific “hallucinations” (as it’s more familiar with the “correct way” to talk about things in that domain).
    • Limitation: Fine-tuning cannot completely eliminate hallucinations, especially regarding new knowledge outside the training data or tasks requiring complex reasoning.
  • Clear Risk Disclosure and User Education:

    • AI tool providers have a responsibility to clearly and prominently inform users (via product interfaces or documentation) about the risk of hallucinations and provide guidance on proper usage and necessary verification.
    • Users (lawyers, paralegals, etc.) themselves need adequate training and education to deeply understand LLM limitations, master effective verification methods, and cultivate healthy skepticism.
  • Maintain Perpetual Critical Thinking:

    • Always approach AI-generated content with vigilance and skepticism, no matter how fluent, confident, or professional it appears.
    • Apply your own legal expertise, experience, and logical judgment to scrutinize, question, and evaluate the AI’s output. Ask yourself: “Does this sound reasonable?” “Does this align with legal principles I know?” “Have I heard of this case/statute before?”
    • Treat AI as a (sometimes unreliable) research assistant, not a legal expert itself.

Summary: At the current state of AI technology, “hallucination” is an inherent characteristic of generative AI (especially LLMs) that is difficult to completely eradicate. Legal professionals must treat it as a normal, manageable risk. By establishing strict verification processes, leveraging auxiliary techniques (like RAG), optimizing usage patterns (like prompt engineering), and most importantly—always maintaining critical awareness and professional judgment, we can harness AI’s power to enhance efficiency while safeguarding the lifelines of legal work: truth, accuracy, and responsibility.

V. Robustness Concerns and the Threat of Adversarial Attacks

Section titled “V. Robustness Concerns and the Threat of Adversarial Attacks”

The Robustness of an AI model, colloquially its “resistance to interference,” refers to its ability to maintain performance (e.g., prediction accuracy, decision stability) when faced with minor perturbations or noise in input data, or new data slightly different from the training distribution. Worryingly, many modern AI models, particularly complex deep learning ones, exhibit considerable fragility in this regard.

Extreme Sensitivity to “Minor Perturbations”: The “Butterfly Effect” in AI

Section titled “Extreme Sensitivity to “Minor Perturbations”: The “Butterfly Effect” in AI”

Research has long revealed a disturbing phenomenon: sometimes, making incredibly small, almost imperceptible changes to a model’s input data (e.g., slightly altering the color values of a few pixels in an image, replacing a synonym or adding a few seemingly meaningless characters in a text) can cause the model to make a completely wrong prediction, often while still maintaining high confidence.

  • Classic Examples:
    • Images: Adding a carefully crafted layer of tiny, human-imperceptible noise (“adversarial perturbation”) to an image correctly identified by a model as a “panda” can cause the model to misclassify it as a “gibbon,” “starfish,” or even a “jet plane” with high confidence.
    • Text: Inserting a few special, seemingly irrelevant words or characters into text originally judged harmless by a model can trick it into classifying it as harmful; slightly rephrasing a sentence can cause a sentiment analysis model’s judgment to flip 180 degrees.

This phenomenon starkly reveals that the decision boundaries learned by AI models (especially deep ones) can be very fragile and counter-intuitive. They might be overly reliant on certain unstable feature patterns that humans struggle to comprehend.

Adversarial Attacks: Exploiting Vulnerability to Deceive AI like a “Trojan Horse”

Section titled “Adversarial Attacks: Exploiting Vulnerability to Deceive AI like a “Trojan Horse””

By exploiting these inherent robustness vulnerabilities, attackers can deliberately craft malicious inputs designed to deceive, mislead, or manipulate AI systems. These are known as Adversarial Attacks. These attacks don’t exploit traditional software code vulnerabilities but cleverly leverage the model’s own characteristics and learning mechanisms. They act like tailored “Trojan horses” for AI, capable of silently infiltrating and disrupting normal system functions.

Common types of adversarial attacks include:

  • Evasion Attacks:

    • Timing: Occur after model deployment, the most common type.
    • Method: Without altering the model itself, the attacker crafts malicious input samples (e.g., images/text/speech with adversarial perturbations, or physical objects like stickers with special patterns) that cause the model to produce the attacker’s desired incorrect output during inference (prediction).
    • Examples:
      • Designing special glasses or stickers that cause facial recognition systems to fail or misidentify someone.
      • Placing tiny disruptive patterns on traffic signs to make autonomous vehicle recognition systems misinterpret them.
      • Making minor, non-functional changes to malware code to evade AI-based virus detection engines.
      • Inputting “adversarial audio commands” into voice assistants that are hard for humans to hear but recognized by the machine.
  • Poisoning Attacks:

    • Timing: Occur during the model’s training phase.
    • Method: The attacker manages to inject a small number of carefully crafted malicious samples (“poison data”) into the model’s training dataset.
    • Goals:
      • Availability Attack: Aim to degrade the model’s overall performance, making it perform worse on all or specific tasks.
      • Backdoor Attacks: More insidious and dangerous. Aim to implant a “backdoor” into the model—it behaves normally on regular inputs, but when presented with an input containing a specific trigger (e.g., a small mark on an image, a special word in text), it is forced to output a predefined, usually incorrect or malicious result desired by the attacker.
    • Challenge: Poisoning attacks are harder to execute (requiring influence over training data) but can have more persistent and stealthy effects if successful.
  • Model Stealing / Model Extraction Attacks:

    • Method: Attackers typically query the model’s API interface repeatedly (sending numerous crafted inputs and observing outputs) to try to infer, replicate, or reverse-engineer the target model’s internal parameters, architecture, or decision logic.
    • Goals: Stealing commercially valuable model intellectual property; preparing for more effective evasion attacks or analyzing adversarial defenses.
  • New Attacks Targeting Large Language Models (LLMs): With the rise of LLMs, attacks targeting their unique mechanisms have emerged:

    • Prompt Injection: Attackers cleverly embed hidden, malicious instructions within the user prompt provided to the model, attempting to override or bypass the developer’s original system instructions or safety guardrails, tricking the model into performing unintended or harmful actions.
      • Example: User A asks a legal AI to summarize a confidential document. Attacker B, in a subsequent conversation with the same AI, uses prompt injection to trick the AI into revealing the content User A summarized; or tricking a legal AI into generating phishing emails, malicious code, defamatory statements, etc., that it’s supposed to refuse.
    • Jailbreaking: Similar to prompt injection but focuses more on using cleverly designed conversations, role-playing scenarios, or complex prompts exploiting model logic loopholes to induce an LLM to violate its own safety, ethical, or content restriction policies. This makes it answer normally refused questions (like asking for illegal information) or generate inappropriate, biased, or harmful content. It’s like finding a way to “jailbreak” the model’s built-in rules.

VI. Other Deep-seated Limitations: Bottlenecks in Understanding, Generalization, and Learning

Section titled “VI. Other Deep-seated Limitations: Bottlenecks in Understanding, Generalization, and Learning”

Beyond the major limitations discussed above, current AI technologies (especially deep learning models) face other fundamental, deep-seated restrictions that also profoundly impact the depth and breadth of their application in complex, nuanced legal scenarios.

Lack of True Understanding and Common Sense

Section titled “Lack of True Understanding and Common Sense”

Despite the ability of state-of-the-art AI models (especially LLMs) to generate remarkably fluent natural text, engage in seemingly logical conversations, and even pass the Turing Test, their “intelligence” remains fundamentally based on learning and mimicking statistical correlations from massive data, rather than possessing human-like deep understanding of how the world works, grasp of causation, flexible use of abstract concepts, or the vast, implicit, self-evident knowledge base of common sense.

  • Correlation vs. Causation: A model might learn from vast text that “demand letters” and “increased litigation risk” frequently co-occur, but it doesn’t truly “understand” that the former can cause the latter. It only knows they are statistically correlated.

  • Lack of Deep Semantic Understanding: A model might know “contract” and “agreement” are often used interchangeably, but may not grasp their subtle differences or varying legal effects in specific legal contexts.

  • Common Sense Knowledge Gap: AI models often lack or possess very fragilely the vast background knowledge humans take for granted in daily life—about the physical world (objects fall, water is wet), social rules (queuing, keeping promises), interpersonal dynamics (intentions, emotions, trust), etc.

    • Fun Example: You could ask an LLM “Why shouldn’t I wipe a power outlet with a wet towel?” It might give a plausible-sounding answer based on correlations between “electricity,” “water,” and “danger” in its data. But it doesn’t truly understand the physics of “water conductivity,” “short circuits,” or “electric shock.” If asked a question requiring more common sense reasoning, like “I put a red ball in a blue box, closed the box and shook it. What color is the ball now?”, even advanced LLMs might answer incorrectly or nonsensically, lacking an inherent understanding of object permanence or color property invariance.
  • Legal Significance: Legal practice is not just rule application; it’s replete with interpreting intent, judging reasonableness, considering fairness, balancing ambiguous concepts (like “good faith,” “material mistake,” “public interest”), and understanding complex social contexts. AI’s deficiency in these areas requiring deep understanding, causal reasoning, and common sense judgment severely limits its ability to independently handle novel, complex, atypical legal problems or those requiring value trade-offs. They might perform adequately at “applying” rules but falter in scenarios demanding true “legal wisdom.”

Limited Generalization to Out-of-Distribution Data

Section titled “Limited Generalization to Out-of-Distribution Data”

AI models typically perform well within the distribution scope represented by their training data. However, when they encounter new situations significantly different from the training distribution, very rare events (the “Long-tail Distribution” problem), or need to transfer knowledge to entirely new domains, their performance often degrades sharply. Models might have “Overfitted” to the patterns they’ve seen, struggling to Generalize effectively to unseen contexts.

  • Legal Significance: Legal practice is precisely full of unique, non-standard “edge cases.” Every case has its unique factual background and points of contention. Applying legal principles often requires flexible, creative interpretation and adaptation to specific circumstances. Over-reliance on AI models trained on relatively standardized, common case data may render them ineffective in handling these difficult, rare, or novel cases that fall outside their “experience range” and require special treatment. They might offer incorrect, oversimplified, or completely inapplicable analyses and recommendations.

Static Knowledge & Difficulty with Continual Learning: Struggling to Keep Pace with Change

Section titled “Static Knowledge & Difficulty with Continual Learning: Struggling to Keep Pace with Change”

The training of most large AI models (especially foundation models underlying many applications) is an extremely time-consuming and resource-intensive process, typically completed using datasets collected before a specific point in time. This means the “knowledge” embedded within the model is static and becomes outdated over time.

  • Dynamic Nature of the Legal World: Law itself is a constantly evolving system. New laws and regulations are continuously enacted and amended, significant guiding precedents emerge, judicial interpretations are updated, and even societal understanding and values regarding certain legal issues shift.

  • AI Learning Challenges: Enabling these massive AI models to continuously and efficiently learn new knowledge like humans, without interfering with or forgetting previously learned knowledge (i.e., avoiding “Catastrophic Forgetting”), remains a major, unsolved technical challenge. While models can be updated through periodic complete retraining (extremely costly) or by employing techniques like Incremental Learning or Continual Learning, these approaches often have limited efficiency and struggle to guarantee seamless integration of old and new knowledge.

  • Legal Significance: For legal applications, the Timeliness and Accuracy of knowledge are critical. An AI legal tool based on outdated laws, overruled precedents, or old judicial interpretations is not just useless but potentially severely misleading and legally risky. Ensuring legal AI systems promptly and accurately reflect the latest legal developments is a prerequisite for their reliable application, yet remains a persistent challenge requiring ongoing investment and solutions within current technical frameworks. Users must remain vigilant about the “knowledge cut-off date” of the AI tools they use.

Conclusion: Embrace Change, But Proceed with Caution; Human-Machine Collaboration is the Path Forward

Section titled “Conclusion: Embrace Change, But Proceed with Caution; Human-Machine Collaboration is the Path Forward”

Clearly recognizing the inherent limitations of current AI technology—from data dependency and bias risks, to the opacity of the “black box,” the potential harm of “hallucinations,” and challenges in robustness, understanding, generalization, and knowledge updating—is not intended to negate the immense potential value and transformative power AI holds for the legal field. Rather, it advocates for a more safe, effective, and Responsible AI approach to introducing and applying this powerful enabling technology.

For every legal professional in this intelligent era, this implies a profound shift in mindset and behavior:

  • Uphold Perpetual Critical Thinking: Never treat AI output as absolute truth or the final answer, no matter how authoritative or convenient it appears. Apply your own professional knowledge, practice experience, and independent judgment for rigorous, independent verification, review, and critical assessment. View AI as a powerful (but sometimes error-prone, even nonsensical) assistant or tool, not an “expert” or “decision-maker” to be blindly relied upon.
  • Clarify the Boundaries of Human-Machine Collaboration: Be acutely aware of which legal tasks are suitable for AI assistance given current technology and limitations (e.g., initial information retrieval, preliminary screening/review of large document volumes, formatting standardized documents, initial responses to simple repetitive queries), and which core, high-risk tasks requiring deep understanding, creative thinking, value judgment, and interpersonal communication (e.g., complex legal strategy formulation, critical evidence admissibility decisions, building client trust, courtroom advocacy, final legal decision-making) must remain human-led.
  • Strengthen Risk Management Awareness and Processes: Before introducing any AI tool, conduct thorough, comprehensive risk assessments. Systematically consider data security and privacy, algorithmic bias and fairness safeguards, model robustness and security testing, intellectual property ownership, regulatory compliance requirements, potential ethical conflicts, etc., and develop corresponding risk mitigation plans and usage protocols.
  • Insist on Human Oversight and Final Responsibility: In all human-machine collaboration models, AI should always be positioned as an auxiliary tool, not the decision-maker. The ultimate judgment, decision-making authority, and the resulting legal and professional responsibility must, and can only, be borne by qualified and experienced human professionals. Establishing clear and effective Human-in-the-Loop or Human-on-the-Loop mechanisms for supervision, review, and intervention is crucial.
  • Actively Participate in Governance and Norm-Setting: Legal professionals are not just users of AI technology; leveraging their professional expertise and understanding of the rule of law, they should be active participants and contributors to discussions on AI ethics, governance, and legal regulation. Pay attention to legal challenges arising from AI development (e.g., evidence rules, liability determination, IP, regulatory frameworks), participate in setting industry standards, best practices, and codes of conduct, promoting technology development in a direction that benefits the rule of law, fosters social justice, and safeguards fundamental human rights in a healthy, sustainable manner.

Only by deeply understanding both the shining “halos” and the lurking “shadows” of artificial intelligence can we, in the coming new era of legal services profoundly shaped by AI, fully harness the power granted by technology to enhance the efficiency, accessibility, and quality of legal services, while simultaneously steadfastly upholding the rigor, prudence, fairness, and responsibility demanded by the legal profession. Ultimately, this allows for the harmonious coexistence and mutual reinforcement of technological progress and the spirit of the rule of law.