y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#self-supervised-learning News & Analysis

18 articles tagged with #self-supervised-learning. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

18 articles
AIBullisharXiv โ€“ CS AI ยท 2d ago7/10
๐Ÿง 

Instructions are all you need: Self-supervised Reinforcement Learning for Instruction Following

Researchers propose a label-free self-supervised reinforcement learning framework that enables language models to follow complex multi-constraint instructions without external supervision. The approach derives reward signals directly from instructions and uses constraint decomposition strategies to address sparse reward challenges, demonstrating strong performance across both in-domain and out-of-domain instruction-following tasks.

AIBullisharXiv โ€“ CS AI ยท 3d ago7/10
๐Ÿง 

TimeRewarder: Learning Dense Reward from Passive Videos via Frame-wise Temporal Distance

TimeRewarder is a new machine learning method that learns dense reward signals from passive videos to improve reinforcement learning in robotics. By modeling temporal distances between video frames, the approach achieves 90% success rates on Meta-World tasks using significantly fewer environment interactions than prior methods, while also leveraging human videos for scalable reward learning.

AINeutralarXiv โ€“ CS AI ยท 3d ago6/10
๐Ÿง 

A Minimal Model of Representation Collapse: Frustration, Stop-Gradient, and Dynamics

Researchers present a minimal mathematical model demonstrating how representation collapse occurs in self-supervised learning when frustrated (misclassified) samples exist, and show that stop-gradient techniques prevent this failure mode. The work provides closed-form analysis of gradient-flow dynamics and fixed points, offering theoretical insights into why modern embedding-based learning systems sometimes lose discriminative power.

AINeutralarXiv โ€“ CS AI ยท Apr 106/10
๐Ÿง 

A Lightweight Library for Energy-Based Joint-Embedding Predictive Architectures

Facebook Research releases EB-JEPA, an open-source library for learning representations through Joint-Embedding Predictive Architectures that predict in representation space rather than pixel space. The framework demonstrates strong performance across image classification (91% on CIFAR-10), video prediction, and action-conditioned world models, making self-supervised learning more accessible for research and practical applications.

AIBullisharXiv โ€“ CS AI ยท Mar 176/10
๐Ÿง 

Pixel-level Scene Understanding in One Token: Visual States Need What-is-Where Composition

Researchers propose CroBo, a new visual state representation learning framework that helps robotic agents better understand dynamic environments by encoding both semantic identities and spatial locations of scene elements. The framework uses a global-to-local reconstruction method that compresses observations into compact tokens, achieving state-of-the-art performance on robot policy learning benchmarks.

AIBullisharXiv โ€“ CS AI ยท Mar 37/107
๐Ÿง 

MetaMind: General and Cognitive World Models in Multi-Agent Systems by Meta-Theory of Mind

Meta researchers introduced MetaMind, a cognitive world model for multi-agent systems that enables agents to understand and predict other agents' behaviors without centralized supervision or communication. The system uses a meta-theory of mind framework allowing agents to reason about goals and beliefs of others through self-reflective learning and analogical reasoning.

AIBullisharXiv โ€“ CS AI ยท Mar 36/106
๐Ÿง 

DINOv3 Meets YOLO26 for Weed Detection in Vegetable Crops

Researchers developed a foundational crop-weed detection model combining DINOv3 vision transformer with YOLO26 architecture, achieving significant improvements in precision agriculture applications. The model showed up to 14% better performance on cross-domain datasets while maintaining real-time processing at 28.5 fps despite increased computational requirements.

AINeutralarXiv โ€“ CS AI ยท Mar 36/105
๐Ÿง 

A SUPERB-Style Benchmark of Self-Supervised Speech Models for Audio Deepfake Detection

Researchers introduced Spoof-SUPERB, a new benchmark for evaluating self-supervised learning models' ability to detect audio deepfakes. The study tested 20 SSL models and found that large-scale discriminative models like XLS-R and WavLM Large consistently outperformed others, especially under acoustic degradations.

AIBullisharXiv โ€“ CS AI ยท Mar 36/103
๐Ÿง 

Latent Diffusion Model without Variational Autoencoder

Researchers introduce SVG, a new latent diffusion model that eliminates the need for variational autoencoders by using self-supervised representations. The approach leverages frozen DINO features to create semantically structured latent spaces, enabling faster training, fewer sampling steps, and better generative quality while maintaining semantic capabilities.

AIBullisharXiv โ€“ CS AI ยท Feb 275/107
๐Ÿง 

Learning to reconstruct from saturated data: audio declipping and high-dynamic range imaging

Researchers have developed a self-supervised learning method that can reconstruct audio and images from clipped/saturated measurements without requiring ground truth training data. The approach extends self-supervised learning to non-linear inverse problems and performs nearly as well as fully supervised methods while using only clipped measurements for training.

AIBullisharXiv โ€“ CS AI ยท Mar 175/10
๐Ÿง 

Human-like Object Grouping in Self-supervised Vision Transformers

Researchers developed a behavioral benchmark showing that self-supervised vision transformers, particularly those trained with DINO objectives, align closely with human object perception and segmentation behavior. The study found that models with stronger object-centric representations better predict human visual judgments, with Gram matrix structure playing a key role in perceptual alignment.

AINeutralarXiv โ€“ CS AI ยท Mar 54/10
๐Ÿง 

Directional Neural Collapse Explains Few-Shot Transfer in Self-Supervised Learning

Researchers propose directional CDNV (decision-axis variance) as a key geometric quantity explaining why self-supervised learning representations transfer well with few labels. The study shows that small variability along class-separating directions enables strong few-shot transfer and low interference across multiple tasks.

AINeutralarXiv โ€“ CS AI ยท Mar 35/107
๐Ÿง 

SIGMAS: Second-Order Interaction-based Grouping for Overlapping Multi-Agent Swarms

Researchers introduce SIGMAS, a self-supervised AI framework for identifying group structures in multi-agent swarms like drone fleets without ground-truth supervision. The system uses second-order interactions to infer latent group memberships from agent trajectories, demonstrating robust performance across diverse synthetic swarm scenarios.

AINeutralarXiv โ€“ CS AI ยท Feb 274/105
๐Ÿง 

FM-RME: Foundation Model Empowered Radio Map Estimation

Researchers introduce FM-RME, a foundation model for radio map estimation that combines geometry-aware feature extraction with attention-based neural networks. The model uses self-supervised pre-training to enable zero-shot generalization across spatial, temporal, and spectral domains without scenario-specific retraining.

AINeutralarXiv โ€“ CS AI ยท Feb 274/103
๐Ÿง 

DyGnROLE: Modeling Asymmetry in Dynamic Graphs with Node-Role-Oriented Latent Encoding

Researchers introduce DyGnROLE, a new AI architecture that better models directed dynamic graphs by treating source and destination nodes differently. The system uses role-specific embeddings and a self-supervised learning approach called Temporal Contrastive Link Prediction to achieve superior performance on future edge classification tasks.

$LINK
AINeutralarXiv โ€“ CS AI ยท Feb 274/107
๐Ÿง 

A Semi-Supervised Learning Method for the Identification of Bad Exposures in Large Imaging Surveys

Researchers developed a semi-supervised machine learning pipeline using vision transformers and k-Nearest Neighbor classifiers to automatically detect poor-quality exposures in astronomical imaging surveys. The method was successfully applied to the DECam Legacy Survey, identifying 780 problematic exposures that were verified through visual inspection.

AINeutralarXiv โ€“ CS AI ยท Feb 274/105
๐Ÿง 

TokEye: Fast Signal Extraction for Fluctuating Time Series via Offline Self-Supervised Learning From Fusion Diagnostics to Bioacoustics

Researchers developed TokEye, a self-supervised AI framework that can extract coherent signals from noisy time-series data in 0.5 seconds, initially designed for fusion reactor diagnostics. The system demonstrates applications beyond fusion research, including bioacoustics, suggesting broader potential for real-time signal processing across industries.