y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#deep-learning News & Analysis

257 articles tagged with #deep-learning. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

257 articles
AI × CryptoBullisharXiv – CS AI · Mar 26/1027
🤖

Blockchain-Enabled Routing for Zero-Trust Low-Altitude Intelligent Networks

Researchers propose a blockchain-enabled zero-trust architecture for secure routing in low-altitude intelligent networks using unmanned aerial vehicles. The framework combines blockchain technology with AI-based routing algorithms to improve security and performance in UAV networks.

AIBullisharXiv – CS AI · Mar 27/1010
🧠

UPath: Universal Planner Across Topological Heterogeneity For Grid-Based Pathfinding

Researchers developed UPath, a universal AI-powered pathfinding algorithm that improves A* search performance by up to 2.2x across diverse grid environments. The deep learning model generalizes across different map types without retraining, achieving near-optimal solutions within 3% of optimal cost on unseen tasks.

AIBullisharXiv – CS AI · Feb 275/106
🧠

Quality-Aware Robust Multi-View Clustering for Heterogeneous Observation Noise

Researchers propose QARMVC, a new AI framework for multi-view clustering that addresses heterogeneous noise in real-world data. The system uses quality scores to identify contamination levels and employs hierarchical learning to improve clustering performance, showing superior results across benchmark datasets.

AIBullisharXiv – CS AI · Feb 276/105
🧠

From Open Vocabulary to Open World: Teaching Vision Language Models to Detect Novel Objects

Researchers have developed a framework that enables open vocabulary object detection models to operate in real-world settings by identifying and learning previously unseen objects. The method introduces techniques called Open World Embedding Learning (OWEL) and Multi-Scale Contrastive Anchor Learning (MSCAL) to detect unknown objects and reduce misclassification errors.

$NEAR
AIBullisharXiv – CS AI · Feb 276/104
🧠

HARU-Net: Hybrid Attention Residual U-Net for Edge-Preserving Denoising in Cone-Beam Computed Tomography

Researchers developed HARU-Net, a novel AI architecture for denoising cone-beam computed tomography (CBCT) medical images that outperforms existing state-of-the-art methods while using less computational resources. The system addresses critical noise issues in low-dose dental and maxillofacial imaging by combining hybrid attention mechanisms with residual U-Net architecture.

AIBullisharXiv – CS AI · Feb 275/107
🧠

RepSPD: Enhancing SPD Manifold Representation in EEGs via Dynamic Graphs

Researchers have developed RepSPD, a novel geometric deep learning model that enhances EEG brain activity decoding using symmetric positive definite manifolds and dynamic graphs. The framework introduces cross-attention mechanisms on Riemannian manifolds and bidirectional alignment strategies to improve brain signal representation and analysis.

AINeutralLast Week in AI · Dec 96/10
🧠

LWiAI Podcast #227 - Jeremie is back! DeepSeek 3.2, TPUs, Nested Learning

DeepSeek releases version 3.2 AI model claiming improved speed, cost-efficiency and performance. NVIDIA partners are reportedly shifting toward Google's TPU ecosystem, while new research explores nested learning in deep learning architectures.

LWiAI Podcast #227 - Jeremie is back! DeepSeek 3.2, TPUs, Nested Learning
🏢 Nvidia
AINeutralOpenAI News · Dec 146/104
🧠

Weak-to-strong generalization

Researchers present a new approach to AI alignment called weak-to-strong generalization, exploring whether deep learning's generalization properties can be used to control powerful AI models using weaker supervisory systems. The work addresses the superalignment problem of maintaining control over increasingly capable AI systems.

AINeutralLil'Log (Lilian Weng) · Jan 276/10
🧠

The Transformer Family Version 2.0

This article presents an updated and expanded version of a comprehensive guide to Transformer architecture improvements, building upon a 2020 post. The new version is twice the length and includes recent developments in Transformer models, providing detailed technical notations and covering both encoder-decoder and simplified architectures like BERT and GPT.

🏢 OpenAI
AINeutralOpenAI News · Jun 95/108
🧠

Techniques for training large neural networks

Large neural networks are driving recent AI advances but present significant training challenges that require coordinated GPU clusters for synchronized calculations. The technical complexity of orchestrating distributed computing resources remains a key engineering obstacle in scaling AI systems.

AINeutralLil'Log (Lilian Weng) · Sep 246/10
🧠

How to Train Really Large Models on Many GPUs?

This article reviews training parallelism paradigms and memory optimization techniques for training very large neural networks across multiple GPUs. It covers architectural designs and methods to overcome GPU memory limitations and extended training times for deep learning models.

🏢 OpenAI
AINeutralLil'Log (Lilian Weng) · Jul 116/10
🧠

What are Diffusion Models?

Diffusion models are a new type of generative AI model that can learn complex data distributions and generate high-quality images competitive with state-of-the-art GANs. The article covers recent developments including classifier-free guidance, GLIDE, unCLIP, Imagen, latent diffusion models, and consistency models.

AIBullishHugging Face Blog · Sep 106/105
🧠

Block Sparse Matrices for Smaller and Faster Language Models

The article discusses block sparse matrices as a technique to create smaller and faster language models. This approach could significantly reduce computational requirements and memory usage in AI systems while maintaining performance.

AINeutralOpenAI News · Jan 306/105
🧠

OpenAI standardizes on PyTorch

OpenAI has announced it is standardizing its deep learning framework on PyTorch, consolidating its AI development infrastructure. This decision represents a significant technical choice for one of the leading AI companies and could influence broader industry adoption patterns.

AIBullishLil'Log (Lilian Weng) · Jan 316/10
🧠

Generalized Language Models

This article discusses the evolution of generalized language models including BERT, GPT, and other major pre-trained models that achieved state-of-the-art results on various NLP tasks. The piece covers the breakthrough progress in 2018 with large-scale unsupervised pre-training approaches that don't require labeled data, similar to how ImageNet helped computer vision.

🏢 OpenAI
AIBullishOpenAI News · Nov 86/106
🧠

Spinning Up in Deep RL

OpenAI has released Spinning Up in Deep RL, a comprehensive educational resource designed to help anyone learn deep reinforcement learning. The resource includes clear code examples, educational exercises, documentation, and tutorials for practitioners.

AIBullishOpenAI News · Jul 96/108
🧠

Glow: Better reversible generative models

Researchers introduce Glow, a reversible generative AI model that uses invertible 1x1 convolutions to generate high-resolution images with efficient sampling capabilities. The model simplifies previous architectures while enabling feature discovery for data attribute manipulation, with code and visualization tools being made publicly available.

AINeutralLil'Log (Lilian Weng) · Sep 286/10
🧠

Anatomize Deep Learning with Information Theory

Professor Naftali Tishby applied information theory to analyze deep neural network training, proposing the Information Bottleneck method as a new learning bound for DNNs. His research identified two distinct phases in DNN training: first representing input data to minimize generalization error, then compressing representations by forgetting irrelevant details.

AIBullishOpenAI News · Oct 115/104
🧠

Transfer from simulation to real world through learning deep inverse dynamics model

The article discusses research on transferring AI models from simulation environments to real-world applications through deep inverse dynamics modeling. This approach aims to bridge the sim-to-real gap in robotics and AI systems by learning how to map actions to outcomes in physical environments.

← PrevPage 7 of 11Next →