y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto
🤖All31,271🧠AI13,316⛓️Crypto11,310💎DeFi1,165🤖AI × Crypto566📰General4,914
🧠

AI

13,316 AI articles curated from 50+ sources with AI-powered sentiment analysis, importance scoring, and key takeaways.

13316 articles
AIBullisharXiv – CS AI · Feb 276/107
🧠

ECHO: Encoding Communities via High-order Operators

Researchers introduce ECHO, a new Graph Neural Network architecture that solves community detection in large networks by overcoming computational bottlenecks and memory constraints. The system can process networks with over 1.6 million nodes and 30 million edges in minutes, achieving throughputs exceeding 2,800 nodes per second.

AIBullisharXiv – CS AI · Feb 276/106
🧠

Automating the Detection of Requirement Dependencies Using Large Language Models

Researchers developed LEREDD, an LLM-based system that automates the detection of dependencies between software requirements using Retrieval-Augmented Generation and In-Context Learning. The system achieved 93% accuracy in classifying requirement dependencies, significantly outperforming existing baselines with relative gains of over 94% in F1 scores for specific dependency types.

AIBullisharXiv – CS AI · Feb 276/107
🧠

GetBatch: Distributed Multi-Object Retrieval for ML Data Loading

Researchers introduce GetBatch, a new object store API that optimizes machine learning data loading by replacing thousands of individual GET requests with a single batch operation. The system achieves up to 15x throughput improvement for small objects and reduces batch retrieval latency by 2x in production ML training workloads.

AIBullisharXiv – CS AI · Feb 276/107
🧠

AeroDGS: Physically Consistent Dynamic Gaussian Splatting for Single-Sequence Aerial 4D Reconstruction

Researchers have developed AeroDGS, a physics-guided 4D Gaussian splatting framework that enables accurate dynamic scene reconstruction from single-view aerial UAV footage. The system addresses key challenges in monocular aerial reconstruction by incorporating physics-based optimization and geometric constraints to resolve depth ambiguity and improve motion estimation.

AIBullisharXiv – CS AI · Feb 276/108
🧠

GRAU: Generic Reconfigurable Activation Unit Design for Neural Network Hardware Accelerators

Researchers propose GRAU, a new reconfigurable activation unit design for neural network hardware accelerators that uses piecewise linear fitting with power-of-two slopes. The design reduces LUT consumption by over 90% compared to traditional multi-threshold activators while supporting mixed-precision quantization and nonlinear functions.

AIBullisharXiv – CS AI · Feb 276/105
🧠

Importance of Prompt Optimisation for Error Detection in Medical Notes Using Language Models

Researchers demonstrated that prompt optimization using Genetic-Pareto (GEPA) significantly improves language models' ability to detect errors in medical notes. The technique boosted accuracy from 0.669 to 0.785 with GPT-5 and from 0.578 to 0.690 with Qwen3-32B, achieving state-of-the-art performance on medical error detection benchmarks.

AINeutralarXiv – CS AI · Feb 276/106
🧠

Sydney Telling Fables on AI and Humans: A Corpus Tracing Memetic Transfer of Persona between LLMs

Researchers created a 4.5k text corpus analyzing how different AI personas, including Microsoft's controversial Sydney chatbot, express views on human-AI relationships across 12 major language models. The study examines how the Sydney persona has spread memetically through training data, allowing newer models to simulate its distinctive characteristics and perspectives.

AIBullisharXiv – CS AI · Feb 276/106
🧠

UpSkill: Mutual Information Skill Learning for Structured Response Diversity in LLMs

Researchers introduce UpSkill, a new training method that uses Mutual Information Skill Learning to improve large language models' ability to generate diverse correct responses across multiple attempts. The technique shows ~3% improvements in pass@k metrics on mathematical reasoning tasks using models like Llama 3.1-8B and Qwen 2.5-7B without degrading single-attempt accuracy.

AIBullisharXiv – CS AI · Feb 275/107
🧠

Learning to reconstruct from saturated data: audio declipping and high-dynamic range imaging

Researchers have developed a self-supervised learning method that can reconstruct audio and images from clipped/saturated measurements without requiring ground truth training data. The approach extends self-supervised learning to non-linear inverse problems and performs nearly as well as fully supervised methods while using only clipped measurements for training.

AIBullisharXiv – CS AI · Feb 276/106
🧠

Reinforcement-aware Knowledge Distillation for LLM Reasoning

Researchers propose RL-aware distillation (RLAD), a new method to efficiently transfer knowledge from large language models to smaller ones during reinforcement learning training. The approach uses Trust Region Ratio Distillation (TRRD) to selectively guide student models only when it improves policy updates, outperforming existing distillation methods across reasoning benchmarks.

AIBullisharXiv – CS AI · Feb 275/107
🧠

Decoder-based Sense Knowledge Distillation

Researchers have developed Decoder-based Sense Knowledge Distillation (DSKD), a new framework that integrates lexical resources into decoder-style large language models during training. The method enhances knowledge distillation performance while enabling generative models to inherit structured semantics without requiring dictionary lookup during inference.

AIBullisharXiv – CS AI · Feb 276/106
🧠

Integrating Machine Learning Ensembles and Large Language Models for Heart Disease Prediction Using Voting Fusion

Researchers developed a hybrid system combining machine learning ensembles with large language models for heart disease prediction, achieving 96.62% accuracy. The study found that traditional ML models (95.78% accuracy) outperformed standalone LLMs (78.9% accuracy), but combining both approaches yielded the best results for clinical decision-support tools.

AIBullisharXiv – CS AI · Feb 276/105
🧠

To Deceive is to Teach? Forging Perceptual Robustness via Adversarial Reinforcement Learning

Researchers introduce AOT (Adversarial Opponent Training), a self-play framework that improves Multimodal Large Language Models' robustness by having an AI attacker generate adversarial image manipulations to train a defender model. The method addresses perceptual fragility in MLLMs when processing visually complex scenes, reducing hallucinations through dynamic adversarial training.

AIBearisharXiv – CS AI · Feb 276/107
🧠

Analysis of LLMs Against Prompt Injection and Jailbreak Attacks

Researchers evaluated prompt injection and jailbreak vulnerabilities across multiple open-source LLMs including Phi, Mistral, DeepSeek-R1, Llama 3.2, Qwen, and Gemma. The study found significant behavioral variations across models and that lightweight defense mechanisms can be consistently bypassed by long, reasoning-heavy prompts.

AIBullisharXiv – CS AI · Feb 276/106
🧠

SmartChunk Retrieval: Query-Aware Chunk Compression with Planning for Efficient Document RAG

Researchers have developed SmartChunk retrieval, a query-adaptive framework that improves retrieval-augmented generation (RAG) systems by dynamically adjusting chunk sizes and compression for document question answering. The system uses a planner to predict optimal chunk abstraction levels and a compression module to create efficient embeddings, outperforming existing RAG baselines while reducing costs.

AIBullisharXiv – CS AI · Feb 276/105
🧠

Comparative Analysis of Neural Retriever-Reranker Pipelines for Retrieval-Augmented Generation over Knowledge Graphs in E-commerce Applications

Researchers developed improved neural retriever-reranker pipelines for Retrieval-Augmented Generation (RAG) systems over knowledge graphs in e-commerce applications. The study achieved 20.4% higher Hit@1 and 14.5% higher Mean Reciprocal Rank compared to existing benchmarks, providing a framework for production-ready RAG systems.

AIBullisharXiv – CS AI · Feb 276/104
🧠

Multi-Dimensional Spectral Geometry of Biological Knowledge in Single-Cell Transformer Representations

Researchers decoded the internal representations of scGPT, a single-cell foundation model, revealing it organizes genes into interpretable biological coordinate systems rather than opaque features. The model encodes cellular organization patterns including protein localization, interaction networks, and regulatory relationships across its transformer layers.

AIBullisharXiv – CS AI · Feb 276/106
🧠

DS SERVE: A Framework for Efficient and Scalable Neural Retrieval

DS-Serve is a new framework that converts massive text datasets (up to half a trillion tokens) into efficient neural retrieval systems. The framework provides web interfaces and APIs with low latency and supports applications like retrieval-augmented generation (RAG) and training data attribution.

← PrevPage 256 of 533Next →
Filters
Sentiment
Importance
Sort
Stay Updated
Everything combined