23 articles tagged with #interpretable-ai. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBullisharXiv – CS AI · 2d ago7/10
🧠Researchers demonstrate that robots equipped with minimal embodied sensorimotor capabilities learn numerical concepts significantly faster than vision-only systems, achieving 96.8% counting accuracy with 10% of training data. The embodied neural network spontaneously develops biologically plausible number representations matching human cognitive development, suggesting embodiment acts as a structural learning prior rather than merely an information source.
AIBullisharXiv – CS AI · Mar 177/10
🧠Researchers developed new methods for extracting symbolic formulas from Kolmogorov-Arnold Networks (KANs), addressing a key bottleneck in making AI models more interpretable. The proposed Greedy in-context Symbolic Regression (GSR) and Gated Matching Pursuit (GMP) methods achieved up to 99.8% reduction in test error while improving robustness.
AIBullisharXiv – CS AI · Mar 117/10
🧠Researchers introduce Logos, a compact AI model that combines multi-step logical reasoning with chemical consistency for molecular design. The model achieves strong performance in structural accuracy and chemical validity while using fewer parameters than larger language models, and provides transparent reasoning that can be inspected by humans.
AIBullisharXiv – CS AI · Mar 46/103
🧠Researchers developed an interpretable AI framework for detecting structural heart disease from electrocardiograms, achieving better performance than existing deep-learning methods while providing clinical transparency. The model demonstrated improvements of nearly 1% across key metrics using the EchoNext benchmark of over 80,000 ECG-ECHO pairs.
AIBullisharXiv – CS AI · Mar 46/102
🧠Researchers introduce CoBELa, a new AI framework for interpretable image generation that uses concept bottlenecks on energy landscapes to enable transparent, controllable synthesis without requiring decoder retraining. The system achieves strong performance on benchmark datasets while allowing users to compositionally manipulate concepts through energy function combinations.
AIBullisharXiv – CS AI · 3d ago6/10
🧠Researchers propose a neuro-symbolic deep reinforcement learning approach that integrates logical rules and symbolic knowledge to improve sample efficiency and generalization in RL systems. The method transfers partial policies from simple tasks to complex ones, reducing training data requirements and improving performance in sparse-reward environments compared to existing baselines.
AIBullisharXiv – CS AI · Apr 66/10
🧠Researchers have developed HIL-CBM, a new hierarchical interpretable AI model that enhances explainability by mimicking human cognitive processes across multiple semantic levels. The model outperforms existing Concept Bottleneck Models in classification accuracy while providing more interpretable explanations without requiring manual concept annotations.
AIBullisharXiv – CS AI · Mar 266/10
🧠Researchers introduced ES-LLMs, a new AI tutoring architecture that separates decision-making from language generation to create more reliable and interpretable educational AI systems. The system outperformed traditional monolithic LLMs in human evaluations (91.7% preference) while reducing costs by 54% and achieving 100% adherence to pedagogical constraints.
AIBullisharXiv – CS AI · Mar 266/10
🧠Researchers introduce Learning to Guide (LTG), a new AI framework where machines provide interpretable guidance to human decision-makers rather than making automated decisions. The SLOG approach transforms vision-language models into guidance generators using human feedback, showing promise in medical diagnosis applications.
AIBullisharXiv – CS AI · Mar 176/10
🧠Researchers propose a new framework that uses LLMs as code generators rather than per-instance evaluators for high-stakes decision-making, creating interpretable and reproducible AI systems. The approach generates executable decision logic once instead of querying LLMs for each prediction, demonstrated through venture capital founder screening with competitive performance while maintaining full transparency.
🧠 GPT-4
AIBullisharXiv – CS AI · Mar 96/10
🧠Researchers developed an interpretable AI framework for fetal ultrasound image classification that incorporates medical concepts and clinical knowledge. The system uses graph convolutional networks to establish relationships between key medical concepts, providing explanations that align with clinicians' cognitive processes rather than just pixel-level analysis.
AIBullisharXiv – CS AI · Mar 37/107
🧠Researchers developed InstructX2X, a new AI model for generating counterfactual medical images that provides interpretable explanations and prevents unintended modifications. The model achieves state-of-the-art performance in creating high-quality chest X-ray images with visual guidance maps for medical applications.
AIBullisharXiv – CS AI · Mar 36/106
🧠Researchers introduce GlassMol, a new interpretable AI model for molecular property prediction that addresses the black-box problem in drug discovery. The model uses Concept Bottleneck Models with automated concept curation and LLM-guided selection, achieving performance that matches or exceeds traditional black-box models across thirteen benchmarks.
AIBullisharXiv – CS AI · Mar 36/107
🧠Researchers have developed QIME, a new framework for creating interpretable medical text embeddings that uses ontology-grounded questions to represent biomedical text. Unlike black-box AI models, QIME provides clinically meaningful explanations while achieving performance close to traditional dense embeddings in medical text analysis tasks.
AIBullisharXiv – CS AI · Mar 27/1015
🧠Researchers have developed DeBiasLens, a new framework that uses sparse autoencoders to identify and deactivate social bias neurons in Vision-Language models without degrading their performance. The model-agnostic approach addresses concerns about unintended social bias in VLMs by making the debiasing process interpretable and targeting internal model dynamics rather than surface-level fixes.
AIBullisharXiv – CS AI · Mar 26/1017
🧠Researchers introduce VISTA, a framework for vessel trajectory imputation that uses knowledge-driven LLM reasoning to repair incomplete maritime tracking data. The system provides 'repair provenance' - documented reasoning behind data repairs - achieving 5-91% accuracy improvements over existing methods while reducing inference time by 51-93%.
AIBullisharXiv – CS AI · Feb 276/103
🧠Researchers developed DisQ-HNet, a new AI framework that synthesizes tau-PET brain scans from MRI data to detect Alzheimer's disease pathology. The method uses advanced neural network architectures to generate cost-effective alternatives to expensive PET imaging while maintaining diagnostic accuracy.
AINeutralarXiv – CS AI · 2d ago5/10
🧠Researchers propose Enhanced-FQL(λ), a fuzzy reinforcement learning framework that combines fuzzified eligibility traces and segmented experience replay to improve interpretability and efficiency in continuous control tasks. The method demonstrates competitive performance with neural network approaches while maintaining computational simplicity through interpretable fuzzy rule bases rather than complex black-box architectures.
$FET
AINeutralarXiv – CS AI · Mar 174/10
🧠Researchers developed a symbolic machine learning approach for predicting failures in chemical processes, specifically testing on ethylene oxidation. The method outperformed traditional AI models while maintaining interpretability through rule-based systems, addressing safety concerns in chemical industries where black-box AI models are unsuitable.
AINeutralarXiv – CS AI · Mar 95/10
🧠Researchers introduce BM25-V, a new image retrieval method that combines sparse visual-word activations from Vision Transformers with BM25 scoring for efficient and interpretable image search. The approach achieves 99.3%+ recall across seven benchmarks while offering explainable results and serving as an efficient first-stage retriever for dense reranking systems.
AINeutralarXiv – CS AI · Mar 44/104
🧠Researchers have developed TVF (Time-Varying Filtering), a lightweight 1 million parameter speech enhancement model that combines digital signal processing with deep learning for real-time speech denoising. The model uses a neural network to predict coefficients for a 35-band IIR filter cascade, offering interpretable processing while adapting dynamically to changing noise conditions.
AINeutralarXiv – CS AI · Mar 25/108
🧠Researchers introduce Hierarchical Concept Embedding Models (HiCEMs), a new approach to make deep neural networks more interpretable by modeling relationships between concepts in hierarchical structures. The method includes Concept Splitting to automatically discover fine-grained sub-concepts without additional annotations, reducing the burden of manual labeling while improving model accuracy and interpretability.
AINeutralarXiv – CS AI · Feb 274/105
🧠Researchers propose Knob, a new framework that applies control theory principles to neural networks by mapping gating dynamics to mechanical systems. The approach enables real-time human adjustment of AI model behavior through intuitive physical parameters like damping and frequency, offering both static and continuous processing modes.