Dispelling the Mirage of Comprehension: MIT Reveals the Fallacy of AI’s Formal Specifications

Dispelling the Mirage of Comprehension: MIT Reveals the Fallacy of AI’s Formal Specifications

MIT Discredits AI’s Formal Specifications

MIT Lincoln Laboratory discovered that while formal specifications provide mathematical accuracy, they don’t guarantee human comprehensibility. Participants faced challenges in confirming AI behaviors based on these specifications, exposing a gap between theoretical claims and practical interpretation. The study emphasizes the necessity for realistic evaluations of AI’s interpretability.

Despite high hopes, formal specifications don’t seem to help autonomous systems communicate effectively with humans, according to recent research.

As AI and autonomous systems gain prevalence in everyday life, novel approaches are being developed to verify their expected behavior. One such method, formal specifications, translates mathematical formulas into human-readable expressions, with claims that this technique aids in explaining AI decisions.

However, MIT Lincoln Laboratory researched to assess the interpretability of formal specifications. Contrary to claims, their findings revealed that humans struggle to interpret these specifications. In a study involving participants assessing an AI agent’s plan in a virtual game using formal specifications, correctness was achieved less than half the time.

Study: Humans struggle with formal specifications, touted as a way to make AI decisions understandable. Credit: Bryan Mastergeorge

MIT Discredits AI’s Formal Specifications: machine behavior

Hosea Siu, a researcher from the AI Technology Group at the laboratory, conveyed the implications of their findings, stating, “The results are disappointing for researchers asserting that formal methods provide interpretability to systems. While this might hold in certain confined and theoretical contexts, it falls short of practical system validation.” The group’s paper was recently accepted for presentation at the 2023 International Conference on Intelligent Robots and Systems.

Interpretability is pivotal in fostering trust in machine behavior, ensuring users can discern the need for adjustments or trust the fairness of decisions made by robots or AI in real-world scenarios. It grants users—beyond developers—an understanding of a technology’s capabilities, yet achieving interpretability remains a longstanding challenge in AI and autonomy. The inherent opacity of the machine learning process often results in an inability to explain decision-making.

Hosea Siu emphasizes that claims of interpretability in machine learning systems need the same scrutiny as claims of accuracy, highlighting the need for more transparency in justifying interpretability.

Logical expression and mathematical proofs

The researchers investigated whether formal specifications improved the interpretability of a system’s behavior, focusing on users’ ability to verify if the system aligned with user goals. Formal specifications, primarily used for logical expression and mathematical proofs within a system, have been repurposed to aid human comprehension, bridging the gap between a model’s logic and human understanding.

Siu emphasizes the misconception between the precision of formal specifications and their interpretability to humans. The study revealed a lack of assessments regarding people’s actual understanding of these outputs.

During the experiment, participants, both experts and non-experts in formal methods, were tasked with validating a simple set of behaviors in a robot playing a game of capture the Flag. The goal was to determine if the robot consistently adhered to specified rules, ensuring victory.

The participants received the formal specifications in three formats: as a raw logical formula, translated into more natural language expressions, and presented as a decision tree—a format often perceived as human-interpretable in AI.

MIT Discredits AI’s Formal Specifications: The study

Surprisingly, the validation performance across all presentation types was notably poor, hovering around 45 percent accuracy, indicating a struggle regardless of the information format.

The study uncovered that individuals trained in formal specifications only marginally outperformed novices. Curiously, experts exhibited more confidence in their responses, regardless of accuracy. Participants overly trusted the given specifications, overlooking rules that could lead to game losses, raising concerns about system validation by potentially ignoring failure modes.

However, the researchers don’t advocate abandoning formal specifications as a means to elucidate system behaviors. Instead, they emphasize the necessity for improved design in presenting these specifications and their integration into human workflows.

Siu acknowledges that individuals versed in formal methods aren’t specifically trained for the type of specifications examined in the experiment. Moreover, anticipating all potential outcomes from a set of rules is challenging. Yet, the rule sets provided to participants were relatively concise, akin to a brief paragraph of text—considerably shorter than those encountered in actual systems.

The endeavor

Team to explore interpretability claims within formal logic, sidestepping direct links to real-world robot validation.

Siu and team aim to improve human-robot interactions, especially in military settings, as part of their larger project. The project aims to help operators train robots directly, similar to how humans teach each other. Such an approach could bolster operator confidence in the robots and the robots’ adaptability.

The team believes their research will improve how autonomy is integrated into human decision-making.

Siu emphasizes the necessity for human evaluations of various autonomous systems and AI concepts before making sweeping claims about their effectiveness in human contexts. This cautious approach advocates for scrutinizing these technologies’ utility in practical human applications.


Read the original article on scitechdaily.

Read more: Journal of Philosophical Logic.

Share this post