y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#few-shot-learning News & Analysis

25 articles tagged with #few-shot-learning. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

25 articles
AIBullisharXiv – CS AI · Mar 117/10
🧠

BiCLIP: Domain Canonicalization via Structured Geometric Transformation

Researchers introduce BiCLIP, a new framework that improves vision-language models' ability to adapt to specialized domains through geometric transformations. The approach achieves state-of-the-art results across 11 benchmarks while maintaining simplicity and low computational requirements.

AIBullisharXiv – CS AI · Mar 56/10
🧠

TATRA: Training-Free Instance-Adaptive Prompting Through Rephrasing and Aggregation

Researchers introduce TATRA, a training-free prompting method for Large Language Models that creates instance-specific few-shot prompts without requiring labeled training data. The method achieves state-of-the-art performance on mathematical reasoning benchmarks like GSM8K and DeepMath, matching or outperforming existing prompt optimization methods that rely on expensive training processes.

AIBullisharXiv – CS AI · Mar 56/10
🧠

Relational In-Context Learning via Synthetic Pre-training with Structural Prior

Researchers introduce RDB-PFN, the first relational foundation model for databases trained entirely on synthetic data to overcome privacy and scarcity issues with real relational databases. The model uses a Relational Prior Generator to create over 2 million synthetic tasks and demonstrates strong few-shot performance on 19 real-world relational prediction tasks through in-context learning.

AIBullisharXiv – CS AI · Mar 57/10
🧠

SPRINT: Semi-supervised Prototypical Representation for Few-Shot Class-Incremental Tabular Learning

Researchers introduce SPRINT, the first Few-Shot Class-Incremental Learning (FSCIL) framework designed specifically for tabular data domains like cybersecurity and healthcare. The system achieves 77.37% accuracy in 5-shot learning scenarios, outperforming existing methods by 4.45% through novel semi-supervised techniques that leverage unlabeled data and confidence-based pseudo-labeling.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Boosting In-Context Learning in LLMs Through the Lens of Classical Supervised Learning

Researchers propose Supervised Calibration (SC), a new framework to improve In-Context Learning performance in Large Language Models by addressing systematic biases through optimal affine transformations in logit space. The method achieves state-of-the-art results across multiple LLMs including Mistral-7B, Llama-2-7B, and Qwen2-7B in few-shot learning scenarios.

🧠 Llama
AIBullishOpenAI News · Nov 77/107
🧠

Learning concepts with energy functions

Researchers developed an energy-based AI model that can learn spatial concepts like 'near' and 'above' from just five demonstrations using 2D point sets. The model demonstrates cross-domain transfer capabilities, applying concepts learned in 2D particle environments to solve 3D physics-based robotics tasks.

$NEAR
AINeutralarXiv – CS AI · 3d ago6/10
🧠

Legal2LogicICL: Improving Generalization in Transforming Legal Cases to Logical Formulas via Diverse Few-Shot Learning

Researchers introduce Legal2LogicICL, an LLM-based framework that improves the conversion of natural-language legal cases into logical formulas through retrieval-augmented few-shot learning. The method addresses data scarcity in legal AI systems and introduces a new annotated dataset (Legal2Proleg) to advance interpretable legal reasoning without requiring model fine-tuning.

AINeutralarXiv – CS AI · 3d ago6/10
🧠

A Survey of Inductive Reasoning for Large Language Models

Researchers present the first comprehensive survey of inductive reasoning in large language models, categorizing improvement methods into post-training, test-time scaling, and data augmentation approaches. The survey establishes unified benchmarks and evaluation metrics for assessing how LLMs perform particular-to-general reasoning tasks that better align with human cognition.

AINeutralarXiv – CS AI · 4d ago6/10
🧠

GNN-as-Judge: Unleashing the Power of LLMs for Graph Learning with GNN Feedback

Researchers propose GNN-as-Judge, a framework combining Large Language Models with Graph Neural Networks to improve learning on text-attributed graphs in low-resource settings. The approach uses collaborative pseudo-labeling and weakly-supervised fine-tuning to generate reliable labels while reducing noise, demonstrating significant performance gains when labeled data is scarce.

AIBullisharXiv – CS AI · Mar 266/10
🧠

Uni-DAD: Unified Distillation and Adaptation of Diffusion Models for Few-step Few-shot Image Generation

Researchers introduce Uni-DAD, a unified approach that combines diffusion model distillation and adaptation into a single pipeline for efficient few-shot image generation. The method achieves comparable quality to state-of-the-art methods while requiring less than 4 sampling steps, addressing the computational cost issues of traditional diffusion models.

AIBullisharXiv – CS AI · Mar 45/103
🧠

Quantum-Inspired Fine-Tuning for Few-Shot AIGC Detection via Phase-Structured Reparameterization

Researchers propose Q-LoRA, a quantum-enhanced fine-tuning method that integrates quantum neural networks into LoRA adapters for improved AI-generated content detection. The study also introduces H-LoRA, a classical variant using Hilbert transforms that achieves similar 5%+ accuracy improvements over standard LoRA at lower computational cost.

AIBullisharXiv – CS AI · Mar 36/103
🧠

Meta-Adaptive Prompt Distillation for Few-Shot Visual Question Answering

Researchers developed a meta-learning approach for Large Multimodal Models (LMMs) that uses distilled soft prompts to improve few-shot visual question answering performance. The method outperformed traditional in-context learning by 21.2% and parameter-efficient finetuning by 7.7% on VQA tasks.

AIBullishGoogle Research Blog · Sep 236/105
🧠

Time series foundation models can be few-shot learners

The article discusses advancements in time series foundation models and their capability for few-shot learning in generative AI applications. These models can learn patterns from limited data samples, potentially improving forecasting and prediction tasks across various domains.

AIBullishHugging Face Blog · Sep 266/107
🧠

SetFit: Efficient Few-Shot Learning Without Prompts

SetFit is a new machine learning framework that enables efficient few-shot learning without requiring prompts. This approach could significantly reduce the computational resources and data requirements for training AI models in various applications.

AIBullishLil'Log (Lilian Weng) · Jun 236/10
🧠

Meta Reinforcement Learning

Meta reinforcement learning enables AI agents to rapidly adapt to new tasks by learning from a distribution of training tasks. The approach allows agents to develop new RL algorithms through internal activity dynamics, focusing on fast and efficient problem-solving for unseen scenarios.

AINeutralarXiv – CS AI · Mar 54/10
🧠

Directional Neural Collapse Explains Few-Shot Transfer in Self-Supervised Learning

Researchers propose directional CDNV (decision-axis variance) as a key geometric quantity explaining why self-supervised learning representations transfer well with few labels. The study shows that small variability along class-separating directions enables strong few-shot transfer and low interference across multiple tasks.

AINeutralarXiv – CS AI · Mar 44/102
🧠

Boosting Meta-Learning for Few-Shot Text Classification via Label-guided Distance Scaling

Researchers propose a Label-guided Distance Scaling (LDS) strategy to improve few-shot text classification by leveraging label semantics during both training and testing phases. The method addresses misclassification issues when randomly selected labeled samples don't provide effective supervision signals, demonstrating significant performance improvements over state-of-the-art models.

AINeutralarXiv – CS AI · Mar 34/103
🧠

Exploiting Low-Dimensional Manifold of Features for Few-Shot Whole Slide Image Classification

Researchers propose a Manifold Residual (MR) block to address overfitting in few-shot Whole Slide Image classification by preserving the low-dimensional manifold geometry of pathology foundation model features. The geometry-aware approach achieves state-of-the-art results with fewer parameters by using a fixed random matrix as geometric anchor and a trainable low-rank residual pathway.

AINeutralarXiv – CS AI · Mar 34/104
🧠

MAGIC: Few-Shot Mask-Guided Anomaly Inpainting with Prompt Perturbation, Spatially Adaptive Guidance, and Context Awareness

MAGIC is a new AI framework for few-shot anomaly detection in industrial quality control that uses mask-guided inpainting to generate high-fidelity synthetic anomalies. The system introduces three key innovations: Gaussian prompt perturbation, spatially adaptive guidance, and context-aware mask alignment to improve anomaly generation while preserving normal regions.

AINeutralGoogle Research Blog · Oct 204/108
🧠

Teaching Gemini to spot exploding stars with just a few examples

Google's Gemini AI is being trained to identify exploding stars (supernovas) using few-shot learning techniques. This demonstrates AI's capability to recognize rare astronomical phenomena with minimal training examples.

AIBullisharXiv – CS AI · Mar 34/103
🧠

MAML-KT: Addressing Cold Start Problem in Knowledge Tracing for New Students via Few-Shot Model-Agnostic Meta Learning

Researchers introduce MAML-KT, a meta-learning approach that addresses the cold start problem in knowledge tracing systems when predicting performance of new students with limited interaction data. The model uses few-shot learning to rapidly adapt to unseen students, achieving higher early accuracy than existing knowledge tracing models across multiple datasets.

AINeutralHugging Face Blog · Dec 63/107
🧠

SetFitABSA: Few-Shot Aspect Based Sentiment Analysis using SetFit

The article appears to discuss SetFitABSA, a methodology for performing aspect-based sentiment analysis using SetFit with minimal training examples. However, the article body is empty, making it impossible to provide meaningful analysis of the content or implications.

AINeutralOpenAI News · May 281/103
🧠

Language models are few-shot learners

The article title references few-shot learning capabilities in language models, but no article body content was provided for analysis. Without the actual article content, a comprehensive analysis cannot be performed.