y0news
#deep-learning16 articles
16 articles
AIBullisharXiv – CS AI Β· 4h ago7
🧠

Reasoning-Driven Multimodal LLM for Domain Generalization

Researchers developed RD-MLDG, a new framework that uses multimodal large language models with reasoning chains to improve domain generalization in deep learning. The approach addresses challenges in cross-domain visual recognition by leveraging reasoning capabilities rather than just visual feature invariance, achieving state-of-the-art performance on standard benchmarks.

AINeutralarXiv – CS AI Β· 4h ago2
🧠

DLEBench: Evaluating Small-scale Object Editing Ability for Instruction-based Image Editing Model

Researchers introduce DLEBench, the first benchmark specifically designed to evaluate instruction-based image editing models' ability to edit small-scale objects that occupy only 1%-10% of image area. Testing on 10 models revealed significant performance gaps in small object editing, highlighting a critical limitation in current AI image editing capabilities.

AI Γ— CryptoBullisharXiv – CS AI Β· 4h ago8
πŸ€–

Blockchain-Enabled Routing for Zero-Trust Low-Altitude Intelligent Networks

Researchers propose a blockchain-enabled zero-trust architecture for secure routing in low-altitude intelligent networks using unmanned aerial vehicles. The framework combines blockchain technology with AI-based routing algorithms to improve security and performance in UAV networks.

AIBullisharXiv – CS AI Β· 4h ago3
🧠

UPath: Universal Planner Across Topological Heterogeneity For Grid-Based Pathfinding

Researchers developed UPath, a universal AI-powered pathfinding algorithm that improves A* search performance by up to 2.2x across diverse grid environments. The deep learning model generalizes across different map types without retraining, achieving near-optimal solutions within 3% of optimal cost on unseen tasks.

AINeutralarXiv – CS AI Β· 4h ago2
🧠

FaultXformer: A Transformer-Encoder Based Fault Classification and Location Identification model in PMU-Integrated Active Electrical Distribution System

Researchers developed FaultXformer, a Transformer-based AI model that achieves 98.76% accuracy in fault classification and 98.92% accuracy in fault location identification in electrical distribution systems using PMU data. The dual-stage architecture significantly outperforms traditional deep learning methods like CNN, RNN, and LSTM, particularly in systems with distributed energy resources integration.

AIBullisharXiv – CS AI Β· 4h ago4
🧠

CUDA Agent: Large-Scale Agentic RL for High-Performance CUDA Kernel Generation

Researchers developed CUDA Agent, a reinforcement learning system that significantly outperforms existing methods for GPU kernel optimization, achieving 100% faster performance than torch.compile on benchmark tests. The system uses large-scale agentic RL with automated verification and profiling to improve CUDA kernel generation, addressing a critical bottleneck in deep learning performance.

AIBullisharXiv – CS AI Β· 4h ago3
🧠

Carr\'e du champ flow matching: better quality-generalisation tradeoff in generative models

Researchers introduce CarrΓ©e du champ flow matching (CDC-FM), a new generative AI model that improves the quality-generalization tradeoff by using geometry-aware noise instead of standard uniform noise. The method shows significant improvements in data-scarce scenarios and non-uniformly sampled datasets, particularly relevant for AI applications in scientific domains.

AINeutralarXiv – CS AI Β· 4h ago0
🧠

Hierarchical Concept-based Interpretable Models

Researchers introduce Hierarchical Concept Embedding Models (HiCEMs), a new approach to make deep neural networks more interpretable by modeling relationships between concepts in hierarchical structures. The method includes Concept Splitting to automatically discover fine-grained sub-concepts without additional annotations, reducing the burden of manual labeling while improving model accuracy and interpretability.

AINeutralarXiv – CS AI Β· 4h ago0
🧠

NuBench: An Open Benchmark for Deep Learning-Based Event Reconstruction in Neutrino Telescopes

NuBench is a new open benchmark for deep learning-based event reconstruction in neutrino telescopes, comprising seven large-scale simulated datasets with nearly 130 million neutrino interactions. The benchmark enables comparison of machine learning reconstruction methods across different detector geometries and evaluates four algorithms including ParticleNeT and DynEdge on core reconstruction tasks.

AINeutralarXiv – CS AI Β· 4h ago0
🧠

General vs Domain-Specific CNNs: Understanding Pretraining Effects on Brain MRI Tumor Classification

Research comparing CNN architectures for brain tumor classification found that general-purpose models like ConvNeXt-Tiny (93% accuracy) outperformed domain-specific medical pre-trained models like RadImageNet DenseNet121 (68% accuracy). The study suggests that contemporary general-purpose CNNs with diverse pre-training may be more effective for medical imaging tasks in data-scarce scenarios.

AINeutralarXiv – CS AI Β· 4h ago0
🧠

Micro-expression Recognition Based on Dual-branch Feature Extraction and Fusion

Researchers developed a dual-branch neural network for micro-expression recognition that combines residual and Inception networks with parallel attention mechanisms. The method achieved 74.67% accuracy on the CASME II dataset, significantly outperforming existing approaches like LBP-TOP by over 11%.

AINeutralarXiv – CS AI Β· 4h ago0
🧠

Intrinsic Lorentz Neural Network

Researchers propose the Intrinsic Lorentz Neural Network (ILNN), a fully intrinsic hyperbolic architecture that performs all computations within the Lorentz model for better handling of hierarchical data structures. The network introduces novel components including point-to-hyperplane layers and GyroLBN batch normalization, achieving state-of-the-art performance on CIFAR and genomic benchmarks while outperforming Euclidean baselines.

AINeutralarXiv – CS AI Β· 4h ago0
🧠

Bridging the Performance Gap Between Target-Free and Target-Based Reinforcement Learning

Researchers introduce iterated Shared Q-Learning (iS-QL), a new reinforcement learning method that bridges target-free and target-based approaches by using only the last linear layer as a target network while sharing other parameters. The technique achieves comparable performance to traditional target-based methods while maintaining the memory efficiency of target-free approaches.

AINeutralarXiv – CS AI Β· 4h ago0
🧠

Heterogeneous Multi-Agent Reinforcement Learning with Attention for Cooperative and Scalable Feature Transformation

Researchers propose a new multi-agent reinforcement learning framework that uses three cooperative agents with attention mechanisms to automate feature transformation for machine learning models. The approach addresses key limitations in existing automated feature engineering methods, including dynamic feature expansion instability and insufficient agent cooperation.