y0news
← Feed
Back to feed
🧠 AI NeutralImportance 6/10

Empirical Characterization of Rationale Stability Under Controlled Perturbations for Explainable Pattern Recognition

arXiv – CS AI|Abu Noman Md Sakib, Zhensen Wang, Merjulah Roby, Zijie Zhang|
🤖AI Summary

Researchers propose a new metric to assess consistency of AI model explanations across similar inputs, implementing it on BERT models for sentiment analysis. The framework uses cosine similarity of SHAP values to detect inconsistent reasoning patterns and biased feature reliance, providing more robust evaluation of model behavior.

Key Takeaways
  • Novel metric developed to quantify consistency of AI model explanations across samples with same class labels.
  • Framework implemented using BERT, RoBERTa, and DistilBERT models on sentiment analysis datasets with SHAP feature importance.
  • Method uses cosine similarity to detect biased reliance on features and inconsistent reasoning patterns.
  • Approach enables more robust verification of rationale stability in explainable AI systems.
  • Research code publicly available for reproducibility and further development.
Read Original →via arXiv – CS AI
Act on this with AI
Stay ahead of the market.
Connect your wallet to an AI agent. It reads balances, proposes swaps and bridges across 15 chains — you keep full control of your keys.
Connect Wallet to AI →How it works
Related Articles