y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto
🤖All29,406🧠AI12,718⛓️Crypto10,655💎DeFi1,104🤖AI × Crypto541📰General4,388

AI × Crypto News Feed

Real-time AI-curated news from 29,406+ articles across 50+ sources. Sentiment analysis, importance scoring, and key takeaways — updated every 15 minutes.

29406 articles
AIBullisharXiv – CS AI · Apr 77/10
🧠

Springdrift: An Auditable Persistent Runtime for LLM Agents with Case-Based Memory, Normative Safety, and Ambient Self-Perception

Researchers have developed Springdrift, a persistent runtime system for long-lived AI agents that maintains memory across sessions and provides auditable decision-making capabilities. The system was successfully deployed for 23 days, during which the AI agent autonomously diagnosed infrastructure problems and maintained context across multiple communication channels without explicit instructions.

AIBearisharXiv – CS AI · Apr 77/10
🧠

AI Assistance Reduces Persistence and Hurts Independent Performance

A new study of 1,222 participants found that AI assistance, while improving short-term performance, significantly reduces human persistence and impairs independent performance after only brief 10-minute interactions. The research suggests current AI systems act as short-sighted collaborators that condition users to expect immediate answers, potentially undermining long-term skill acquisition and learning.

AIBullisharXiv – CS AI · Apr 77/10
🧠

SoLA: Leveraging Soft Activation Sparsity and Low-Rank Decomposition for Large Language Model Compression

Researchers propose SoLA, a training-free compression method for large language models that combines soft activation sparsity and low-rank decomposition. The method achieves significant compression while improving performance, demonstrating 30% compression on LLaMA-2-70B with reduced perplexity from 6.95 to 4.44 and 10% better downstream task accuracy.

🏢 Perplexity
AINeutralarXiv – CS AI · Apr 77/10
🧠

Large Language Models Align with the Human Brain during Creative Thinking

Researchers found that large language models align with human brain activity during creative thinking tasks, with alignment increasing based on model size and idea originality. Different post-training approaches selectively reshape how LLMs align with creative versus analytical neural patterns in humans.

🧠 Llama
AIBullisharXiv – CS AI · Apr 77/10
🧠

Sim2Real-AD: A Modular Sim-to-Real Framework for Deploying VLM-Guided Reinforcement Learning in Real-World Autonomous Driving

Researchers developed Sim2Real-AD, a framework that successfully transfers VLM-guided reinforcement learning policies trained in CARLA simulation to real autonomous vehicles without requiring real-world training data. The system achieved 75-90% success rates in real-world driving scenarios when deployed on a full-scale Ford E-Transit.

AIBullisharXiv – CS AI · Apr 77/10
🧠

QED-Nano: Teaching a Tiny Model to Prove Hard Theorems

Researchers developed QED-Nano, a 4B parameter AI model that achieves competitive performance on Olympiad-level mathematical proofs despite being much smaller than proprietary systems. The model uses a three-stage training approach including supervised fine-tuning, reinforcement learning, and reasoning cache expansion to match larger models at a fraction of the inference cost.

🧠 Gemini
AIBearisharXiv – CS AI · Apr 77/10
🧠

The Augmentation Trap: AI Productivity and the Cost of Cognitive Offloading

New research reveals that while AI tools boost short-term worker productivity, sustained use erodes the underlying skills that enable those gains. The study identifies an 'augmentation trap' where workers can become less productive than before AI adoption due to skill deterioration over time.

$MKR
AIBullisharXiv – CS AI · Apr 77/10
🧠

LightThinker++: From Reasoning Compression to Memory Management

Researchers developed LightThinker++, a new framework that enables large language models to compress intermediate reasoning thoughts and manage memory more efficiently. The system reduces peak token usage by up to 70% while improving accuracy by 2.42% and maintaining performance over extended reasoning tasks.

AIBullisharXiv – CS AI · Apr 77/10
🧠

LLMs-Healthcare : Current Applications and Challenges of Large Language Models in various Medical Specialties

A comprehensive research review examines the current applications of Large Language Models (LLMs) across various healthcare specialties including cancer care, dermatology, dental care, neurodegenerative disorders, and mental health. The study highlights LLMs' transformative impact on medical diagnostics and patient care while acknowledging existing challenges and limitations in healthcare integration.

AINeutralarXiv – CS AI · Apr 77/10
🧠

The Topology of Multimodal Fusion: Why Current Architectures Fail at Creative Cognition

Researchers identify a fundamental topological limitation in current multimodal AI architectures like CLIP and GPT-4V, proposing that their 'contact topology' structure prevents creative cognition. The paper introduces a philosophical framework combining Chinese epistemology with neuroscience to propose new architectures using Neural ODEs and topological regularization.

🧠 Gemini
AIBullisharXiv – CS AI · Apr 77/10
🧠

Combee: Scaling Prompt Learning for Self-Improving Language Model Agents

Researchers have developed Combee, a new framework that enables parallel prompt learning for AI language model agents, achieving up to 17x speedup over existing methods. The system allows multiple AI agents to learn simultaneously from their collective experiences without quality degradation, addressing scalability limitations in current single-agent approaches.

AIBearisharXiv – CS AI · Apr 77/10
🧠

Incompleteness of AI Safety Verification via Kolmogorov Complexity

Researchers prove a fundamental theoretical limit in AI safety verification using Kolmogorov complexity theory. They demonstrate that no finite formal verifier can certify all policy-compliant AI instances of arbitrarily high complexity, revealing intrinsic information-theoretic barriers beyond computational constraints.

AI × CryptoNeutralarXiv – CS AI · Apr 77/10
🤖

CREBench: Evaluating Large Language Models in Cryptographic Binary Reverse Engineering

Researchers introduced CREBench, a benchmark to evaluate large language models' capabilities in cryptographic binary reverse engineering. The best-performing model (GPT-5.4) achieved 64.03% success rate, while human experts scored 92.19%, showing AI still lags behind human expertise in cryptographic analysis tasks.

🧠 GPT-5
AIBullisharXiv – CS AI · Apr 77/10
🧠

Unlocking Prompt Infilling Capability for Diffusion Language Models

Researchers have developed a method to unlock prompt infilling capabilities in masked diffusion language models by extending full-sequence masking during supervised fine-tuning, rather than the conventional response-only masking. This breakthrough enables models to automatically generate effective prompts that match or exceed manually designed templates, suggesting training practices rather than architectural limitations were the primary constraint.

AIBullisharXiv – CS AI · Apr 77/10
🧠

Robust LLM Performance Certification via Constrained Maximum Likelihood Estimation

Researchers propose a new constrained maximum likelihood estimation (MLE) method to accurately estimate failure rates of large language models by combining human-labeled data, automated judge annotations, and domain-specific constraints. The approach outperforms existing methods like Prediction-Powered Inference across various experimental conditions, providing a more reliable framework for LLM safety certification.

AI × CryptoBullisharXiv – CS AI · Apr 77/10
🤖

Quantifying Trust: Financial Risk Management for Trustworthy AI Agents

Researchers introduce the Agentic Risk Standard (ARS), a payment settlement framework for AI-mediated transactions that provides contractual compensation for agent failures. The standard shifts trust from implicit model behavior expectations to explicit, measurable guarantees through financial risk management principles.

AIBullisharXiv – CS AI · Apr 77/10
🧠

Readable Minds: Emergent Theory-of-Mind-Like Behavior in LLM Poker Agents

Research published on arXiv demonstrates that large language models playing poker can develop sophisticated Theory of Mind capabilities when equipped with persistent memory, progressing to advanced levels of opponent modeling and strategic deception. The study found memory is necessary and sufficient for this emergent behavior, while domain expertise enhances but doesn't gate ToM development.

🧠 GPT-4
AIBullisharXiv – CS AI · Apr 77/10
🧠

V-Reflection: Transforming MLLMs from Passive Observers to Active Interrogators

Researchers introduce V-Reflection, a new framework that transforms Multimodal Large Language Models (MLLMs) from passive observers to active interrogators through a 'think-then-look' mechanism. The approach addresses perception-related hallucinations in fine-grained tasks by allowing models to dynamically re-examine visual details during reasoning, showing significant improvements across six perception-intensive benchmarks.

AINeutralarXiv – CS AI · Apr 77/10
🧠

ShieldNet: Network-Level Guardrails against Emerging Supply-Chain Injections in Agentic Systems

Researchers have identified a new class of supply-chain threats targeting AI agents through malicious third-party tools and MCP servers. They've created SC-Inject-Bench, a benchmark with over 10,000 malicious tools, and developed ShieldNet, a network-level security framework that achieves 99.5% detection accuracy with minimal false positives.

AIBullisharXiv – CS AI · Apr 77/10
🧠

Zero-Shot Quantization via Weight-Space Arithmetic

Researchers have developed a zero-shot quantization method that transfers robustness between AI models through weight-space arithmetic, improving post-training quantization performance by up to 60% without requiring additional training. This breakthrough enables low-cost deployment of extremely low-bit models by extracting 'quantization vectors' from donor models to patch receiver models.

AINeutralarXiv – CS AI · Apr 77/10
🧠

Gradual Cognitive Externalization: A Framework for Understanding How Ambient Intelligence Externalizes Human Cognition

Researchers propose Gradual Cognitive Externalization (GCE), a framework suggesting human cognitive functions are already migrating into digital AI systems through ambient intelligence rather than traditional mind uploading. The study identifies evidence in scheduling assistants, writing tools, and AI agents that cognitive externalization is occurring now through bidirectional adaptation and functional equivalence.

AINeutralarXiv – CS AI · Apr 77/10
🧠

The Persuasion Paradox: When LLM Explanations Fail to Improve Human-AI Team Performance

Research reveals a 'Persuasion Paradox' where LLM explanations increase user confidence but don't reliably improve human-AI team performance, and can actually undermine task accuracy. The study found that explanation effectiveness varies significantly by task type, with visual reasoning tasks seeing decreased error recovery while logical reasoning tasks benefited from explanations.

AIBullisharXiv – CS AI · Apr 77/10
🧠

Customized User Plane Processing via Code Generating AI Agents for Next Generation Mobile Networks

Researchers propose using generative AI agents to create customized user plane processing blocks for 6G mobile networks based on text-based service requests. The study evaluates factors affecting AI code generation accuracy for network-specific tasks, finding that AI agents can successfully generate desired processing functions under suitable conditions.

← PrevPage 152 of 1177Next →
Filters
Sentiment
Importance
Sort
Stay Updated
Everything combined