y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto
🤖All29,625🧠AI12,738⛓️Crypto10,731💎DeFi1,113🤖AI × Crypto546📰General4,497
🧠

AI

12,738 AI articles curated from 50+ sources with AI-powered sentiment analysis, importance scoring, and key takeaways.

12738 articles
AINeutralarXiv – CS AI · Apr 66/10
🧠

Xpertbench: Expert Level Tasks with Rubrics-Based Evaluation

Researchers introduce XpertBench, a new benchmark for evaluating Large Language Models on expert-level professional tasks across domains like finance, healthcare, and legal services. Even top-performing LLMs achieve only ~66% success rates, revealing a significant 'expert-gap' in current AI systems' ability to handle complex professional work.

AIBullisharXiv – CS AI · Apr 66/10
🧠

Unified Thinker: A General Reasoning Modular Core for Image Generation

Researchers introduce Unified Thinker, a new AI architecture that improves image generation by separating reasoning from visual generation. The modular system addresses the gap between closed-source models like Nano Banana and open-source alternatives by enabling better instruction following through executable reasoning and reinforcement learning.

AIBullisharXiv – CS AI · Apr 66/10
🧠

Attribution Gradients: Incrementally Unfolding Citations for Critical Examination of Attributed AI Answers

Researchers have developed "attribution gradients," a new technique to improve AI answer engines by making citations more informative and easier to evaluate. The method consolidates evidence amounts, supporting/contradictory excerpts, and contextual explanations in one place, while also allowing users to explore second-degree citations without leaving the interface.

AIBullisharXiv – CS AI · Apr 66/10
🧠

SmartCLIP: Modular Vision-language Alignment with Identification Guarantees

Researchers introduce SmartCLIP, a new AI model that improves upon CLIP by addressing information misalignment issues between images and text through modular vision-language alignment. The approach enables better disentanglement of visual representations while preserving cross-modal semantic information, demonstrating superior performance across various tasks.

AIBullisharXiv – CS AI · Apr 66/10
🧠

The More, the Merrier: Contrastive Fusion for Higher-Order Multimodal Alignment

Researchers introduce Contrastive Fusion (ConFu), a new multimodal machine learning framework that aligns individual modalities and their fused combinations in a unified representation space. The approach captures higher-order dependencies between multiple modalities while maintaining strong pairwise relationships, demonstrating competitive performance on retrieval and classification tasks.

AINeutralarXiv – CS AI · Apr 66/10
🧠

Human Psychometric Questionnaires Mischaracterize LLM Psychology: Evidence from Generation Behavior

Research reveals that standard human psychological questionnaires fail to accurately assess the true psychological characteristics of large language models (LLMs). The study of eight open-source LLMs found significant differences between self-reported questionnaire responses and actual generation behavior, suggesting questionnaires capture desired behavior rather than authentic psychological traits.

AIBearisharXiv – CS AI · Apr 66/10
🧠

What Is The Political Content in LLMs' Pre- and Post-Training Data?

Research reveals that large language models exhibit political biases stemming from systematically left-leaning training data, with pre-training datasets containing more politically engaged content than post-training data. The study finds strong correlations between political stances in training data and model behavior, with biases persisting across all training stages.

AIBullisharXiv – CS AI · Apr 66/10
🧠

ForgeryGPT: A Multimodal LLM for Interpretable Image Forgery Detection and Localization

Researchers have developed ForgeryGPT, a new multimodal AI framework that can detect, localize, and explain image forgeries through natural language interaction. The system combines advanced computer vision techniques with large language models to provide interpretable analysis of tampered images, addressing limitations in current forgery detection methods.

🧠 GPT-4
AINeutralarXiv – CS AI · Apr 66/10
🧠

StructEval: Benchmarking LLMs' Capabilities to Generate Structural Outputs

Researchers introduce StructEval, a comprehensive benchmark for evaluating Large Language Models' ability to generate structured outputs across 18 formats including JSON, HTML, and React. Even state-of-the-art models like o1-mini only achieve 75.58% average scores, with open-source models performing approximately 10 points lower.

AIBullisharXiv – CS AI · Apr 66/10
🧠

Gradient Boosting within a Single Attention Layer

Researchers introduce gradient-boosted attention, a new method that improves transformer performance by applying gradient boosting principles within a single attention layer. The technique uses a second attention pass to correct errors from the first pass, achieving lower perplexity (67.9 vs 72.2) on WikiText-103 compared to standard attention mechanisms.

🏢 Perplexity
AIBullisharXiv – CS AI · Apr 66/10
🧠

InCoder-32B-Thinking: Industrial Code World Model for Thinking

Researchers introduce InCoder-32B-Thinking, an AI model trained with Error-driven Chain-of-Thought (ECoT) framework and Industrial Code World Model (ICWM) for industrial software development. The model generates reasoning traces for hardware-constrained programming and achieves top-tier performance on 23 benchmarks, scoring 81.3% on LiveCodeBench v5 and 84.0% on CAD-Coder.

AIBullisharXiv – CS AI · Apr 66/10
🧠

Valence-Arousal Subspace in LLMs: Circular Emotion Geometry and Multi-Behavioral Control

Researchers developed a method to identify valence-arousal subspaces in large language models, enabling controlled emotional steering of AI outputs. The technique demonstrates cross-architecture effectiveness on multiple models and reveals that emotional control can bidirectionally influence AI behaviors like refusal and sycophancy.

🧠 Llama
AIBearisharXiv – CS AI · Apr 66/10
🧠

Can VLMs Truly Forget? Benchmarking Training-Free Visual Concept Unlearning

Researchers introduce VLM-UnBench, the first benchmark for evaluating training-free visual concept unlearning in Vision Language Models. The study reveals that realistic prompts fail to genuinely remove sensitive or copyrighted visual concepts, with meaningful suppression only occurring under oracle conditions that explicitly disclose target concepts.

AIBullisharXiv – CS AI · Apr 66/10
🧠

R2-Write: Reflection and Revision for Open-Ended Writing with Deep Reasoning

Researchers introduce R2-Write, a new AI framework that improves large language models' performance on open-ended writing tasks by incorporating explicit reflection and revision patterns. The study reveals that existing reasoning models show limited gains in creative writing compared to mathematical tasks, prompting the development of an automated system with writer-judge interactions and process reward mechanisms.

AIBearisharXiv – CS AI · Apr 66/10
🧠

From Abstract to Contextual: What LLMs Still Cannot Do in Mathematics

A new study reveals that large language models, despite excelling at benchmark math problems, struggle significantly with contextual mathematical reasoning where problems are embedded in real-world scenarios. The research shows performance drops of 13-34 points for open-source models and 13-20 points for proprietary models when abstract math problems are presented in contextual settings.

AINeutralarXiv – CS AI · Apr 66/10
🧠

Split and Conquer Partial Deepfake Speech

Researchers developed a new AI framework for detecting partial deepfake speech by splitting the problem into boundary detection and segment classification stages. The method achieves state-of-the-art performance on benchmark datasets, significantly improving detection and localization of manipulated audio regions within otherwise authentic speech.

AIBearisharXiv – CS AI · Apr 66/10
🧠

LogicPoison: Logical Attacks on Graph Retrieval-Augmented Generation

Researchers have discovered LogicPoison, a new attack method that exploits vulnerabilities in Graph-based Retrieval-Augmented Generation (GraphRAG) systems by corrupting logical connections in knowledge graphs without altering text semantics. The attack successfully bypasses GraphRAG's existing defenses by targeting the topological integrity of underlying graphs, significantly degrading AI system performance.

AIBullisharXiv – CS AI · Apr 66/10
🧠

Rubrics to Tokens: Bridging Response-level Rubrics and Token-level Rewards in Instruction Following Tasks

Researchers propose Rubrics to Tokens (RTT), a novel reinforcement learning framework that improves Large Language Model alignment by bridging response-level and token-level rewards. The method addresses reward sparsity and ambiguity issues in instruction-following tasks through fine-grained credit assignment and demonstrates superior performance across different models.

AIBullisharXiv – CS AI · Apr 66/10
🧠

QAPruner: Quantization-Aware Vision Token Pruning for Multimodal Large Language Models

Researchers developed QAPruner, a new framework that simultaneously optimizes vision token pruning and post-training quantization for Multimodal Large Language Models (MLLMs). The method addresses the problem where traditional token pruning can discard important activation outliers needed for quantization stability, achieving 2.24% accuracy improvement over baselines while retaining only 12.5% of visual tokens.

AIBullisharXiv – CS AI · Apr 66/10
🧠

NavCrafter: Exploring 3D Scenes from a Single Image

NavCrafter is a new AI framework that creates flexible 3D scenes from a single image by generating novel-view video sequences with controllable camera movement. The system uses video diffusion models and enhanced 3D Gaussian Splatting to achieve superior 3D reconstruction and novel-view synthesis under large viewpoint changes.

AINeutralarXiv – CS AI · Apr 66/10
🧠

Random Is Hard to Beat: Active Selection in online DPO with Modern LLMs

Research from arXiv shows that Active Preference Learning (APL) provides minimal improvements over random sampling in training modern LLMs through Direct Preference Optimization. The study found that random sampling performs nearly as well as sophisticated active selection methods while being computationally cheaper and avoiding capability degradation.

AIBullisharXiv – CS AI · Apr 66/10
🧠

Efficient3D: A Unified Framework for Adaptive and Debiased Token Reduction in 3D MLLMs

Researchers have developed Efficient3D, a framework that accelerates 3D Multimodal Large Language Models (MLLMs) while maintaining accuracy through adaptive token pruning. The system uses a Debiased Visual Token Importance Estimator and Adaptive Token Rebalancing to reduce computational overhead without sacrificing performance, showing +2.57% CIDEr improvement on benchmarks.

AINeutralarXiv – CS AI · Apr 66/10
🧠

DocShield: Towards AI Document Safety via Evidence-Grounded Agentic Reasoning

Researchers introduce DocShield, a new AI framework that uses evidence-based reasoning to detect text-based image forgeries in documents. The system combines visual and logical analysis to identify, locate, and explain document manipulations, showing significant improvements over existing detection methods.

🧠 GPT-4
← PrevPage 169 of 510Next →
Filters
Sentiment
Importance
Sort
Stay Updated
Everything combined