y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto
🤖All25,985🧠AI11,519⛓️Crypto9,585💎DeFi974🤖AI × Crypto505📰General3,402
🧠

AI

11,520 AI articles curated from 50+ sources with AI-powered sentiment analysis, importance scoring, and key takeaways.

11520 articles
AINeutralarXiv – CS AI · Mar 167/10
🧠

HCP-DCNet: A Hierarchical Causal Primitive Dynamic Composition Network for Self-Improving Causal Understanding

Researchers introduce HCP-DCNet, a new AI framework that combines physical dynamics with symbolic causal reasoning to enable AI systems to understand cause-and-effect relationships. The system uses hierarchical causal primitives and can self-improve through interventions, potentially addressing current limitations in AI's ability to handle distribution shifts and counterfactual reasoning.

AIBullisharXiv – CS AI · Mar 167/10
🧠

Darwin Godel Machine: Open-Ended Evolution of Self-Improving Agents

Researchers introduce the Darwin Gödel Machine (DGM), a self-improving AI system that can iteratively modify its own code and validate changes through benchmarks. The system demonstrated significant performance improvements, increasing coding capabilities from 20.0% to 50.0% on SWE-bench and from 14.2% to 30.7% on Polyglot benchmarks.

AINeutralarXiv – CS AI · Mar 167/10
🧠

On Deepfake Voice Detection -- It's All in the Presentation

Researchers have identified why current deepfake voice detection systems fail in real-world applications, finding that existing datasets don't account for how audio changes when transmitted through communication channels. A new framework improved detection accuracy by 39-57% and emphasizes that better datasets matter more than larger AI models for effective deepfake detection.

AIBullisharXiv – CS AI · Mar 167/10
🧠

Efficient Reasoning with Balanced Thinking

Researchers propose ReBalance, a training-free framework that optimizes Large Reasoning Models by addressing overthinking and underthinking issues through confidence-based guidance. The solution dynamically adjusts reasoning trajectories without requiring model retraining, showing improved accuracy across multiple AI benchmarks.

AIBullisharXiv – CS AI · Mar 167/10
🧠

Pyramid MoA: A Probabilistic Framework for Cost-Optimized Anytime Inference

Researchers have developed Pyramid MoA, a new framework that optimizes large language model inference costs by using a hierarchical router system that escalates queries to more expensive models only when necessary. The system achieves up to 62.7% cost savings while maintaining Oracle-level accuracy on various benchmarks including coding and mathematical reasoning tasks.

🧠 Llama
AIBullisharXiv – CS AI · Mar 167/10
🧠

Active Causal Structure Learning with Latent Variables: Towards Learning to Detour in Autonomous Robots

Researchers propose Active Causal Structure Learning with Latent Variables (ACSLWL) as a necessary component for building AGI agents and robots. The paper demonstrates how this approach enables simulated robots to learn complex detour behaviors when encountering unexpected obstacles, allowing them to adapt to new environments by constructing internal causal models.

AIBullisharXiv – CS AI · Mar 167/10
🧠

Learnable Koopman-Enhanced Transformer-Based Time Series Forecasting with Spectral Control

Researchers propose a new family of learnable Koopman operators that combine linear dynamical systems theory with deep learning for time series forecasting. The approach integrates with existing transformer architectures like Patchtst and Autoformer, offering improved stability and interpretability in predictive models.

AIBearisharXiv – CS AI · Mar 167/10
🧠

MalURLBench: A Benchmark Evaluating Agents' Vulnerabilities When Processing Web URLs

Researchers have released MalURLBench, the first benchmark to evaluate how LLM-based web agents handle malicious URLs, revealing significant vulnerabilities across 12 popular models. The study found that existing AI agents struggle to detect disguised malicious URLs and proposed URLGuard as a defensive solution.

AIBullisharXiv – CS AI · Mar 167/10
🧠

Learnability and Privacy Vulnerability are Entangled in a Few Critical Weights

Researchers discovered that privacy vulnerabilities in neural networks exist in only a small fraction of weights, but these same weights are critical for model performance. They developed a new approach that preserves privacy by rewinding and fine-tuning only these critical weights instead of retraining entire networks, maintaining utility while defending against membership inference attacks.

AIBullisharXiv – CS AI · Mar 167/10
🧠

Human-AI Governance (HAIG): A Trust-Utility Approach

Researchers introduce the Human-AI Governance (HAIG) framework that treats AI systems as collaborative partners rather than mere tools, proposing a trust-utility approach to governance across three dimensions: Decision Authority, Process Autonomy, and Accountability Configuration. The framework aims to enable adaptive regulatory design for evolving AI capabilities, particularly as foundation models and multi-agent systems demonstrate increasing autonomy.

AIBearisharXiv – CS AI · Mar 167/10
🧠

Purify Once, Edit Freely: Breaking Image Protections under Model Mismatch

Researchers have identified a critical vulnerability in image protection systems that use adversarial perturbations to prevent unauthorized AI editing. Two new purification methods can effectively remove these protections, creating a 'purify-once, edit-freely' attack where images become vulnerable to unlimited manipulation.

AIBullisharXiv – CS AI · Mar 167/10
🧠

ARL-Tangram: Unleash the Resource Efficiency in Agentic Reinforcement Learning

Researchers introduced ARL-Tangram, a resource management system that optimizes cloud resource allocation for agentic reinforcement learning tasks involving large language models. The system achieves up to 4.3x faster action completion times and 71.2% resource savings through action-level orchestration, and has been deployed for training MiMo series models.

AIBullisharXiv – CS AI · Mar 167/10
🧠

LightMoE: Reducing Mixture-of-Experts Redundancy through Expert Replacing

Researchers introduce LightMoE, a new framework that compresses Mixture-of-Experts language models by replacing redundant expert modules with parameter-efficient alternatives. The method achieves 30-50% compression rates while maintaining or improving performance, addressing the substantial memory demands that limit MoE model deployment.

AIBullisharXiv – CS AI · Mar 167/10
🧠

Disentangling Recall and Reasoning in Transformer Models through Layer-wise Attention and Activation Analysis

Researchers used mechanistic interpretability techniques to demonstrate that transformer language models have distinct but interacting neural circuits for recall (retrieving memorized facts) and reasoning (multi-step inference). Through controlled experiments on Qwen and LLaMA models, they showed that disabling specific circuits can selectively impair one ability while leaving the other intact.

AIBearisharXiv – CS AI · Mar 167/10
🧠

Altered Thoughts, Altered Actions: Probing Chain-of-Thought Vulnerabilities in VLA Robotic Manipulation

Research reveals critical vulnerabilities in Vision-Language-Action robotic models that use chain-of-thought reasoning, where corrupting object names in internal reasoning traces can reduce task success rates by up to 45%. The study shows these AI systems are vulnerable to attacks on their internal reasoning processes, even when primary inputs remain untouched.

AIBullisharXiv – CS AI · Mar 167/10
🧠

Cost-Efficient Multimodal LLM Inference via Cross-Tier GPU Heterogeneity

Researchers developed HeteroServe, a system that optimizes multimodal large language model inference by partitioning vision encoding and language generation across different GPU tiers. The approach reduces data transfer requirements and achieves 31-40% cost savings while improving throughput by up to 54% compared to existing systems.

AIBearisharXiv – CS AI · Mar 167/10
🧠

Experimental evidence of progressive ChatGPT models self-convergence

Research reveals that recent ChatGPT models show declining ability to generate diverse text outputs, a phenomenon called 'model self-convergence.' This degradation is attributed to training on increasing amounts of synthetic data as AI-generated content proliferates across the internet.

🧠 ChatGPT
AIBullisharXiv – CS AI · Mar 167/10
🧠

Development of Ontological Knowledge Bases by Leveraging Large Language Models

Researchers have developed a new methodology that leverages Large Language Models to automate the creation of Ontological Knowledge Bases, addressing traditional challenges of manual development. The approach demonstrates significant improvements in scalability, consistency, and efficiency through automated knowledge acquisition and continuous refinement cycles.

AINeutralarXiv – CS AI · Mar 167/10
🧠

The Economics of AI Supply Chain Regulation

A game-theoretic study analyzes how regulatory policies affect AI supply chains where foundation model providers serve downstream firms. The research finds that price competition policies work best with high compute costs, while quality competition policies always improve consumer surplus, offering guidance for effective AI market regulation.

AIBullisharXiv – CS AI · Mar 167/10
🧠

SRAM-Based Compute-in-Memory Accelerator for Linear-decay Spiking Neural Networks

Researchers developed an SRAM-based compute-in-memory accelerator for spiking neural networks that uses linear decay approximation instead of exponential decay, achieving 1.1x to 16.7x reduction in energy consumption. The innovation addresses the bottleneck of neuron state updates in neuromorphic computing by performing in-place decay directly within memory arrays.

AIBullisharXiv – CS AI · Mar 167/10
🧠

Spend Less, Reason Better: Budget-Aware Value Tree Search for LLM Agents

Researchers propose Budget-Aware Value Tree (BAVT), a training-free framework that improves LLM agent efficiency by intelligently managing computational resources during multi-hop reasoning tasks. The system outperforms traditional approaches while using 4x fewer resources, demonstrating that smart budget management beats brute-force compute scaling.

AINeutralarXiv – CS AI · Mar 167/10
🧠

Aligning Large Language Model Agents with Rational and Moral Preferences: A Supervised Fine-Tuning Approach

Researchers developed a supervised fine-tuning approach to align large language model agents with specific economic preferences, addressing systematic deviations from rational behavior in strategic environments. The study demonstrates how LLM agents can be trained to follow either self-interested or morally-guided strategies, producing distinct outcomes in economic games and pricing scenarios.

AINeutralarXiv – CS AI · Mar 167/10
🧠

Epistemic diversity across language models mitigates knowledge collapse

Research published on arXiv demonstrates that training diverse AI model ecosystems can prevent knowledge collapse, where AI systems degrade when trained on their own outputs. The study shows that optimal diversity levels increase with training iterations, and larger, more homogeneous systems are more susceptible to collapse.

AINeutralarXiv – CS AI · Mar 167/10
🧠

Embedded Quantum Machine Learning in Embedded Systems: Feasibility, Hybrid Architectures, and Quantum Co-Processors

Research paper explores embedded quantum machine learning (EQML) feasibility for edge devices like IoT nodes and drones by 2026. The study identifies hybrid workflows and embedded quantum co-processors as the most viable implementation pathways, while highlighting major barriers including latency, data encoding overhead, and energy constraints.

AIBullisharXiv – CS AI · Mar 167/10
🧠

Revisiting Model Stitching In the Foundation Model Era

Researchers introduce improved methods for stitching Vision Foundation Models (VFMs) like CLIP and DINOv2, enabling integration of different models' strengths. The study proposes VFM Stitch Tree (VST) technique that allows controllable accuracy-latency trade-offs for multimodal applications.

← PrevPage 45 of 461Next →
Filters
Sentiment
Importance
Sort
Stay Updated
Everything combined