y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto
🤖All31,146🧠AI13,276⛓️Crypto11,247💎DeFi1,160🤖AI × Crypto566📰General4,897
🧠

AI

13,277 AI articles curated from 50+ sources with AI-powered sentiment analysis, importance scoring, and key takeaways.

13277 articles
AIBullisharXiv – CS AI · Mar 26/109
🧠

Preference Packing: Efficient Preference Optimization for Large Language Models

Researchers propose 'preference packing,' a new optimization technique for training large language models that reduces training time by at least 37% through more efficient handling of duplicate input prompts. The method optimizes attention operations and KV cache memory usage in preference-based training methods like Direct Preference Optimization.

AINeutralarXiv – CS AI · Mar 26/1014
🧠

Jailbreak Foundry: From Papers to Runnable Attacks for Reproducible Benchmarking

Researchers introduce Jailbreak Foundry (JBF), a system that automatically converts AI jailbreak research papers into executable code modules for standardized testing. The system successfully reproduced 30 attacks with high accuracy and reduces implementation code by nearly half while enabling consistent evaluation across multiple AI models.

AIBullisharXiv – CS AI · Mar 26/1010
🧠

SHINE: Sequential Hierarchical Integration Network for EEG and MEG

Researchers developed SHINE, a Sequential Hierarchical Integration Network for analyzing brain signals (EEG/MEG) to detect speech from neural activity. The system achieved high F1-macro scores of 0.9155-0.9184 in the LibriBrain Competition 2025 by reconstructing speech-silence patterns from magnetoencephalography signals.

AIBullisharXiv – CS AI · Mar 27/1016
🧠

MINT: Multimodal Imaging-to-Speech Knowledge Transfer for Early Alzheimer's Screening

Researchers developed MINT, a framework that transfers knowledge from MRI brain scans to speech analysis for early Alzheimer's detection. The system achieves comparable performance to speech-only methods while being grounded in neuroimaging biomarkers, enabling population-scale screening without requiring expensive MRI scans at inference.

AINeutralarXiv – CS AI · Mar 27/1010
🧠

Ask don't tell: Reducing sycophancy in large language models

Research identifies sycophancy as a key alignment failure in large language models, where AI systems favor user-affirming responses over critical engagement. The study demonstrates that converting user statements into questions before answering significantly reduces sycophantic behavior, offering a practical mitigation strategy for AI developers and users.

AIBullisharXiv – CS AI · Mar 27/1015
🧠

PointCoT: A Multi-modal Benchmark for Explicit 3D Geometric Reasoning

Researchers introduce PointCoT, a new AI framework that enables multimodal large language models to perform explicit geometric reasoning on 3D point cloud data using Chain-of-Thought methodology. The framework addresses current limitations where AI models suffer from geometric hallucinations by implementing a 'Look, Think, then Answer' paradigm with 86k instruction-tuning samples.

AINeutralarXiv – CS AI · Mar 27/1011
🧠

FaultXformer: A Transformer-Encoder Based Fault Classification and Location Identification model in PMU-Integrated Active Electrical Distribution System

Researchers developed FaultXformer, a Transformer-based AI model that achieves 98.76% accuracy in fault classification and 98.92% accuracy in fault location identification in electrical distribution systems using PMU data. The dual-stage architecture significantly outperforms traditional deep learning methods like CNN, RNN, and LSTM, particularly in systems with distributed energy resources integration.

AIBullisharXiv – CS AI · Mar 26/1015
🧠

Green or Fast? Learning to Balance Cold Starts and Idle Carbon in Serverless Computing

Researchers developed LACE-RL, a deep reinforcement learning framework that optimizes serverless computing by balancing cold-start latency and carbon emissions. The system dynamically adjusts keep-alive durations based on real-time carbon intensity and workload patterns, achieving 51.69% fewer cold starts and 77.08% lower idle carbon emissions compared to static policies.

AIBullisharXiv – CS AI · Mar 27/1012
🧠

The Geometry of Transfer: Unlocking Medical Vision Manifolds for Training-Free Model Ranking

Researchers developed a new framework for selecting optimal medical AI foundation models without costly fine-tuning, achieving 31% better performance than existing methods. The topology-driven approach evaluates manifold tractability rather than statistical overlap to better assess model transferability for medical image segmentation tasks.

AINeutralarXiv – CS AI · Mar 26/1012
🧠

Ref-Adv: Exploring MLLM Visual Reasoning in Referring Expression Tasks

Researchers introduce Ref-Adv, a new benchmark for testing multimodal large language models' visual reasoning capabilities in referring expression tasks. The benchmark reveals that current MLLMs, despite performing well on standard datasets like RefCOCO, rely heavily on shortcuts and show significant gaps in genuine visual reasoning and grounding abilities.

AIBullisharXiv – CS AI · Mar 26/1021
🧠

Multi-View Encoders for Performance Prediction in LLM-Based Agentic Workflows

Researchers developed Agentic Predictor, a lightweight AI system that uses multi-view encoding to optimize LLM-based agent workflows without expensive trial-and-error evaluations. The system incorporates code architecture, textual prompts, and interaction graphs to predict task success rates and select optimal configurations across different domains.

AINeutralarXiv – CS AI · Mar 27/1017
🧠

Exploring Robust Intrusion Detection: A Benchmark Study of Feature Transferability in IoT Botnet Attack Detection

Researchers conducted a benchmark study on IoT botnet intrusion detection systems, finding that models trained on one network domain suffer significant performance degradation when applied to different environments. The study evaluated three feature sets across four IoT datasets and provided guidelines for improving cross-domain robustness through better feature engineering and algorithm selection.

AIBullisharXiv – CS AI · Mar 27/1011
🧠

Foundation World Models for Agents that Learn, Verify, and Adapt Reliably Beyond Static Environments

Researchers propose a new framework for foundation world models that enables autonomous agents to learn, verify, and adapt reliably in dynamic environments. The approach combines reinforcement learning with formal verification and adaptive abstraction to create agents that can synthesize verifiable programs and maintain correctness while adapting to novel conditions.

AINeutralarXiv – CS AI · Mar 27/1010
🧠

From Static Benchmarks to Dynamic Protocol: Agent-Centric Text Anomaly Detection for Evaluating LLM Reasoning

Researchers propose a dynamic agent-centric benchmarking system for evaluating large language models that replaces static datasets with autonomous agents that generate, validate, and solve problems iteratively. The protocol uses teacher, orchestrator, and student agents to create progressively challenging text anomaly detection tasks that expose reasoning errors missed by conventional benchmarks.

AINeutralarXiv – CS AI · Mar 27/1015
🧠

SLA-Aware Distributed LLM Inference Across Device-RAN-Cloud

Researchers tested distributed AI inference across device, edge, and cloud tiers in a 5G network, finding that sub-second AI response times required for embodied AI are challenging to achieve. On-device execution took multiple seconds, while RAN-edge deployment with quantized models could meet 0.5-second deadlines, and cloud deployment achieved 100% success for 1-second deadlines.

$NEAR
AIBullisharXiv – CS AI · Mar 27/1016
🧠

TradeFM: A Generative Foundation Model for Trade-flow and Market Microstructure

Researchers introduced TradeFM, a 524M-parameter generative AI model that learns from billions of trade events across 9,000+ equities to understand market microstructure. The model can generate synthetic market data and generalizes across different markets without asset-specific calibration, potentially enabling new applications in trading and market simulation.

$COMP
AIBullisharXiv – CS AI · Mar 26/1010
🧠

SAGE-LLM: Towards Safe and Generalizable LLM Controller with Fuzzy-CBF Verification and Graph-Structured Knowledge Retrieval for UAV Decision

Researchers propose SAGE-LLM, a novel framework that combines Large Language Models with Control Barrier Functions for safe UAV autonomous decision-making. The system addresses LLM safety limitations through formal verification mechanisms and graph-based knowledge retrieval, demonstrating improved safety and generalization in drone control scenarios.

AIBullisharXiv – CS AI · Mar 27/1010
🧠

UPath: Universal Planner Across Topological Heterogeneity For Grid-Based Pathfinding

Researchers developed UPath, a universal AI-powered pathfinding algorithm that improves A* search performance by up to 2.2x across diverse grid environments. The deep learning model generalizes across different map types without retraining, achieving near-optimal solutions within 3% of optimal cost on unseen tasks.

AIBullisharXiv – CS AI · Mar 26/1012
🧠

TRIZ-RAGNER: A Retrieval-Augmented Large Language Model for TRIZ-Aware Named Entity Recognition in Patent-Based Contradiction Mining

Researchers developed TRIZ-RAGNER, a retrieval-augmented large language model framework that improves patent analysis and systematic innovation by extracting technical contradictions from patent documents. The system achieved 84.2% F1-score accuracy, outperforming existing methods by 7.3 percentage points through better integration of domain-specific knowledge.

AINeutralarXiv – CS AI · Mar 27/1021
🧠

The Compulsory Imaginary: AGI and Corporate Authority

A research paper analyzes how leading AGI companies OpenAI and Anthropic use similar rhetorical strategies to construct sociotechnical imaginaries that position themselves as indispensable to AI's future development. The study identifies four shared rhetorical operations that help these firms project corporate authority over technological futures despite their different public approaches.

AIBullisharXiv – CS AI · Mar 26/1016
🧠

FlexGuard: Continuous Risk Scoring for Strictness-Adaptive LLM Content Moderation

Researchers introduce FlexGuard, a new AI content moderation system that provides continuous risk scoring instead of binary decisions, allowing platforms to adapt moderation strictness as needed. The system addresses limitations of existing guardrail models that break down when content moderation requirements change across platforms or over time.

AINeutralarXiv – CS AI · Mar 26/1012
🧠

DLEBench: Evaluating Small-scale Object Editing Ability for Instruction-based Image Editing Model

Researchers introduce DLEBench, the first benchmark specifically designed to evaluate instruction-based image editing models' ability to edit small-scale objects that occupy only 1%-10% of image area. Testing on 10 models revealed significant performance gaps in small object editing, highlighting a critical limitation in current AI image editing capabilities.

← PrevPage 249 of 532Next →
Filters
Sentiment
Importance
Sort
Stay Updated
Everything combined