y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#deep-learning News & Analysis

257 articles tagged with #deep-learning. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

257 articles
AIBullisharXiv โ€“ CS AI ยท Mar 37/106
๐Ÿง 

General Proximal Flow Networks

Researchers introduce General Proximal Flow Networks (GPFNs), a generalization of Bayesian Flow Networks that allows for arbitrary divergence functions instead of fixed Kullback-Leibler divergence. The framework enables iterative generative modeling with improved generation quality when divergence functions are adapted to underlying data geometry.

$LINK
AIBullisharXiv โ€“ CS AI ยท Mar 36/109
๐Ÿง 

Data-Free PINNs for Compressible Flows: Mitigating Spectral Bias and Gradient Pathologies via Mach-Guided Scaling and Hybrid Convolutions

Researchers developed a data-free Physics-Informed Neural Network (PINN) that can solve compressible flows around circular cylinders at extreme speeds up to Mach 15. The system uses hybrid convolutions and Mach-guided scaling to overcome traditional limitations and successfully captures shock waves without requiring training data.

AIBullisharXiv โ€“ CS AI ยท Mar 36/108
๐Ÿง 

A Deep Learning Framework for Heat Demand Forecasting using Time-Frequency Representations of Decomposed Features

Researchers developed a deep learning framework using Continuous Wavelet Transform and CNNs for heat demand forecasting in district heating systems. The model achieved 36-43% reduction in forecasting errors compared to existing methods, reaching up to 95% accuracy in predicting day-ahead heat demand across multiple European cities.

AIBullisharXiv โ€“ CS AI ยท Mar 36/105
๐Ÿง 

Dataset Color Quantization: A Training-Oriented Framework for Dataset-Level Compression

Researchers propose Dataset Color Quantization (DCQ), a new framework that compresses visual datasets by reducing color-space redundancy while preserving information crucial for AI model training. The method achieves significant storage reduction across major datasets including CIFAR-10, CIFAR-100, Tiny-ImageNet, and ImageNet-1K while maintaining training performance.

AIBullisharXiv โ€“ CS AI ยท Mar 36/108
๐Ÿง 

GRAD-Former: Gated Robust Attention-based Differential Transformer for Change Detection

Researchers introduce GRAD-Former, a novel AI framework for detecting changes in satellite imagery that outperforms existing methods while using fewer computational resources. The system uses gated attention mechanisms and differential transformers to more efficiently identify semantic differences in very high-resolution satellite images.

AIBullisharXiv โ€“ CS AI ยท Mar 36/108
๐Ÿง 

AlignVAR: Towards Globally Consistent Visual Autoregression for Image Super-Resolution

Researchers introduced AlignVAR, a new visual autoregressive framework for image super-resolution that delivers 10x faster inference with 50% fewer parameters than leading diffusion-based approaches. The system addresses key challenges in image reconstruction through improved spatial consistency and hierarchical constraints, establishing a more efficient paradigm for high-quality image enhancement.

AINeutralarXiv โ€“ CS AI ยท Mar 36/104
๐Ÿง 

Near--Real-Time Conflict-Related Fire Detection in Sudan Using Unsupervised Deep Learning

Researchers developed a lightweight AI model using unsupervised deep learning to detect conflict-related fires in Sudan within 24-30 hours using commercially available satellite imagery. The Variational Auto-Encoder (VAE) approach outperformed traditional methods in identifying burn signatures from 4-band Planet Labs satellite data at 3-meter resolution.

$CRV$NEAR
AIBullisharXiv โ€“ CS AI ยท Mar 37/105
๐Ÿง 

CHLU: The Causal Hamiltonian Learning Unit as a Symplectic Primitive for Deep Learning

Researchers propose the Causal Hamiltonian Learning Unit (CHLU), a physics-based deep learning primitive that addresses stability issues in temporal dynamics models. The CHLU uses symplectic integration and Hamiltonian structure to maintain infinite-horizon stability while preserving information, potentially solving the memory-stability trade-off in neural networks.

AIBullisharXiv โ€“ CS AI ยท Mar 36/103
๐Ÿง 

Explanation-Guided Adversarial Training for Robust and Interpretable Models

Researchers propose Explanation-Guided Adversarial Training (EGAT), a framework that combines adversarial training with explainable AI to create more robust and interpretable deep neural networks. The method achieves 37% improvement in adversarial accuracy while producing semantically meaningful explanations with only 16% increase in training time.

AIBullisharXiv โ€“ CS AI ยท Mar 36/103
๐Ÿง 

Detection-Gated Glottal Segmentation with Zero-Shot Cross-Dataset Transfer and Clinical Feature Extraction

Researchers developed a detection-gated AI pipeline combining YOLOv8 and U-Net for accurate glottal segmentation in medical videoendoscopy. The system achieved state-of-the-art performance with zero-shot transfer capabilities across different clinical datasets, enabling real-time extraction of vocal function biomarkers at 35 frames per second.

AIBullisharXiv โ€“ CS AI ยท Mar 36/103
๐Ÿง 

A Graph Meta-Network for Learning on Kolmogorov-Arnold Networks

Researchers developed WS-KAN, the first weight-space architecture designed specifically for Kolmogorov-Arnold Networks (KANs), which learns directly from neural network parameters. The study shows KANs share permutation symmetries with MLPs and introduces a graph representation to better understand their computation structure.

AIBullisharXiv โ€“ CS AI ยท Mar 36/104
๐Ÿง 

Scalable Multi-Task Learning through Spiking Neural Networks with Adaptive Task-Switching Policy for Intelligent Autonomous Agents

Researchers have developed SwitchMT, a novel methodology using Spiking Neural Networks with adaptive task-switching for multi-task learning in autonomous agents. The approach addresses task interference issues and demonstrates competitive performance in multiple Atari games while maintaining low power consumption and network complexity.

AINeutralarXiv โ€“ CS AI ยท Mar 35/103
๐Ÿง 

General Protein Pretraining or Domain-Specific Designs? Benchmarking Protein Modeling on Realistic Applications

Researchers introduce Protap, a comprehensive benchmark comparing protein modeling approaches across realistic applications. The study finds that large-scale pretrained models often underperform supervised encoders on small datasets, while structural information and domain-specific biological knowledge can enhance specialized protein tasks.

AINeutralarXiv โ€“ CS AI ยท Mar 36/104
๐Ÿง 

Vision-DeepResearch Benchmark: Rethinking Visual and Textual Search for Multimodal Large Language Models

Researchers introduce Vision-DeepResearch Benchmark (VDR-Bench) with 2,000 VQA instances to better evaluate multimodal AI systems' visual and textual search capabilities. The benchmark addresses limitations in existing evaluations where answers could be inferred without proper visual search, and proposes a multi-round cropped-search workflow to improve model performance.

$NEAR
AIBullisharXiv โ€“ CS AI ยท Mar 36/104
๐Ÿง 

AIRMap: AI-Generated Radio Maps for Wireless Digital Twins

Researchers developed AIRMap, a deep-learning framework that generates radio maps for wireless network simulation over 100x faster than traditional ray tracing methods. The AI model achieves under 4 dB RMSE accuracy in 4 ms per inference and significantly outperforms traditional simulators when calibrated with field measurements.

$NEAR
AINeutralarXiv โ€“ CS AI ยท Mar 35/103
๐Ÿง 

FIRE: Frobenius-Isometry Reinitialization for Balancing the Stability-Plasticity Tradeoff

Researchers propose FIRE, a new reinitialization method for deep neural networks that balances stability and plasticity when learning from nonstationary data. The method uses mathematical optimization to maintain prior knowledge while adapting to new tasks, showing superior performance across visual learning, language modeling, and reinforcement learning domains.

AINeutralarXiv โ€“ CS AI ยท Mar 35/104
๐Ÿง 

Spurious Correlation-Aware Embedding Regularization for Worst-Group Robustness

Researchers propose SCER (Spurious Correlation-Aware Embedding Regularization), a new deep learning approach that improves AI model robustness by regularizing feature representations to suppress spurious correlations. The method demonstrates superior performance in worst-group accuracy across vision and language tasks compared to existing state-of-the-art approaches.

AIBullisharXiv โ€“ CS AI ยท Mar 37/106
๐Ÿง 

Expert Divergence Learning for MoE-based Language Models

Researchers introduce Expert Divergence Learning, a new pre-training strategy for Mixture-of-Experts language models that prevents expert homogenization by encouraging functional specialization. The method uses domain labels to maximize routing distribution differences between data domains, achieving better performance on 15 billion parameter models with minimal computational overhead.

AIBullisharXiv โ€“ CS AI ยท Mar 26/1018
๐Ÿง 

Reasoning-Driven Multimodal LLM for Domain Generalization

Researchers developed RD-MLDG, a new framework that uses multimodal large language models with reasoning chains to improve domain generalization in deep learning. The approach addresses challenges in cross-domain visual recognition by leveraging reasoning capabilities rather than just visual feature invariance, achieving state-of-the-art performance on standard benchmarks.

AINeutralarXiv โ€“ CS AI ยท Mar 26/1012
๐Ÿง 

DLEBench: Evaluating Small-scale Object Editing Ability for Instruction-based Image Editing Model

Researchers introduce DLEBench, the first benchmark specifically designed to evaluate instruction-based image editing models' ability to edit small-scale objects that occupy only 1%-10% of image area. Testing on 10 models revealed significant performance gaps in small object editing, highlighting a critical limitation in current AI image editing capabilities.

AIBullisharXiv โ€“ CS AI ยท Mar 27/1013
๐Ÿง 

CUDA Agent: Large-Scale Agentic RL for High-Performance CUDA Kernel Generation

Researchers developed CUDA Agent, a reinforcement learning system that significantly outperforms existing methods for GPU kernel optimization, achieving 100% faster performance than torch.compile on benchmark tests. The system uses large-scale agentic RL with automated verification and profiling to improve CUDA kernel generation, addressing a critical bottleneck in deep learning performance.

AIBullisharXiv โ€“ CS AI ยท Mar 27/1010
๐Ÿง 

UPath: Universal Planner Across Topological Heterogeneity For Grid-Based Pathfinding

Researchers developed UPath, a universal AI-powered pathfinding algorithm that improves A* search performance by up to 2.2x across diverse grid environments. The deep learning model generalizes across different map types without retraining, achieving near-optimal solutions within 3% of optimal cost on unseen tasks.