y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#optimization News & Analysis

282 articles tagged with #optimization. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

282 articles
AIBullisharXiv โ€“ CS AI ยท Mar 36/104
๐Ÿง 

Contribution-aware Token Compression for Efficient Video Understanding via Reinforcement Learning

Researchers developed CaCoVID, a reinforcement learning-based algorithm that compresses video tokens for large language models by selecting tokens based on their actual contribution to correct predictions rather than attention scores. The method uses combinatorial policy optimization to reduce computational overhead while maintaining video understanding performance.

AINeutralarXiv โ€“ CS AI ยท Mar 35/103
๐Ÿง 

FIRE: Frobenius-Isometry Reinitialization for Balancing the Stability-Plasticity Tradeoff

Researchers propose FIRE, a new reinitialization method for deep neural networks that balances stability and plasticity when learning from nonstationary data. The method uses mathematical optimization to maintain prior knowledge while adapting to new tasks, showing superior performance across visual learning, language modeling, and reinforcement learning domains.

AIBullisharXiv โ€“ CS AI ยท Mar 36/102
๐Ÿง 

SWE-MiniSandbox: Container-Free Reinforcement Learning for Building Software Engineering Agents

Researchers introduced SWE-MiniSandbox, a container-free method for training software engineering AI agents using reinforcement learning that reduces disk usage to 5% and environment setup time to 25% of traditional container-based approaches. The system uses kernel-level isolation and lightweight pre-caching instead of bulky container images while maintaining comparable performance.

AIBullisharXiv โ€“ CS AI ยท Mar 26/1021
๐Ÿง 

Multi-View Encoders for Performance Prediction in LLM-Based Agentic Workflows

Researchers developed Agentic Predictor, a lightweight AI system that uses multi-view encoding to optimize LLM-based agent workflows without expensive trial-and-error evaluations. The system incorporates code architecture, textual prompts, and interaction graphs to predict task success rates and select optimal configurations across different domains.

AIBullisharXiv โ€“ CS AI ยท Mar 27/1019
๐Ÿง 

Smoothing DiLoCo with Primal Averaging for Faster Training of LLMs

Researchers propose Generalized Primal Averaging (GPA), a new optimization method that improves training speed for large language models by 8-10% over standard AdamW while using less memory. GPA unifies and enhances existing averaging-based optimizers like DiLoCo by enabling smooth iterate averaging at every step without complex two-loop structures.

AIBullisharXiv โ€“ CS AI ยท Mar 27/1019
๐Ÿง 

Thompson Sampling via Fine-Tuning of LLMs

Researchers developed ToSFiT (Thompson Sampling via Fine-Tuning), a new Bayesian optimization method that uses fine-tuned large language models to improve search efficiency in complex discrete spaces. The approach eliminates computational bottlenecks by directly parameterizing reward probabilities and demonstrates superior performance across diverse applications including protein search and quantum circuit design.

AIBullisharXiv โ€“ CS AI ยท Mar 26/1018
๐Ÿง 

Taming Momentum: Rethinking Optimizer States Through Low-Rank Approximation

Researchers introduce LoRA-Pre, a memory-efficient optimizer that reduces memory overhead in training large language models by using low-rank approximation of momentum states. The method achieves superior performance on Llama models from 60M to 1B parameters while using only 1/8 the rank of baseline methods.

AIBullisharXiv โ€“ CS AI ยท Mar 26/1016
๐Ÿง 

Does Your Reasoning Model Implicitly Know When to Stop Thinking?

Researchers introduce SAGE (Self-Aware Guided Efficient Reasoning), a novel sampling paradigm that improves AI reasoning efficiency by helping large reasoning models know when to stop thinking. The approach addresses the problem of redundant, lengthy reasoning chains that don't improve accuracy while reducing computational costs and response times.

AIBullisharXiv โ€“ CS AI ยท Mar 26/1014
๐Ÿง 

Trust Region Masking for Long-Horizon LLM Reinforcement Learning

Researchers propose Trust Region Masking (TRM) to address off-policy mismatch problems in Large Language Model reinforcement learning pipelines. The method provides the first non-vacuous monotonic improvement guarantees for long-horizon LLM-RL tasks by masking entire sequences that violate trust region constraints.

AIBullisharXiv โ€“ CS AI ยท Mar 26/1013
๐Ÿง 

RF-Agent: Automated Reward Function Design via Language Agent Tree Search

Researchers introduce RF-Agent, a framework that uses Large Language Models as agents to automatically design reward functions for control tasks through Monte Carlo Tree Search. The method improves upon existing approaches by better utilizing historical feedback and enhancing search efficiency across 17 diverse low-level control tasks.

AIBullisharXiv โ€“ CS AI ยท Mar 27/1011
๐Ÿง 

Learning Flexible Job Shop Scheduling under Limited Buffers and Material Kitting Constraints

Researchers developed a deep reinforcement learning approach using heterogeneous graph networks to solve Flexible Job Shop Scheduling Problems with limited buffers and material kitting constraints. The method outperforms traditional heuristics by improving buffer utilization and decision quality through better modeling of complex dependencies in production scheduling.

AIBullisharXiv โ€“ CS AI ยท Mar 26/1010
๐Ÿง 

Long Range Frequency Tuning for QML

Researchers have developed a new quantum machine learning optimization technique using ternary encodings that significantly improves frequency tuning efficiency. The method achieves 22.8% better performance than existing approaches while requiring exponentially fewer encoding gates than traditional fixed-frequency methods.

AIBullisharXiv โ€“ CS AI ยท Mar 27/1011
๐Ÿง 

KEEP: A KV-Cache-Centric Memory Management System for Efficient Embodied Planning

Researchers from PKU-SEC-Lab have developed KEEP, a new memory management system that significantly improves the efficiency of AI-powered embodied planning by optimizing KV cache usage. The system achieves 2.68x speedup compared to text-based memory methods while maintaining accuracy, addressing a key bottleneck in memory-augmented Large Language Models for complex planning tasks.

AIBullisharXiv โ€“ CS AI ยท Mar 27/1016
๐Ÿง 

SMAC: Score-Matched Actor-Critics for Robust Offline-to-Online Transfer

Researchers developed Score Matched Actor-Critic (SMAC), a new offline reinforcement learning method that enables smooth transition to online RL algorithms without performance drops. SMAC achieved successful transfer in all 6 D4RL tasks tested and reduced regret by 34-58% in 4 of 6 environments compared to best baselines.

AIBullisharXiv โ€“ CS AI ยท Mar 27/1010
๐Ÿง 

UPath: Universal Planner Across Topological Heterogeneity For Grid-Based Pathfinding

Researchers developed UPath, a universal AI-powered pathfinding algorithm that improves A* search performance by up to 2.2x across diverse grid environments. The deep learning model generalizes across different map types without retraining, achieving near-optimal solutions within 3% of optimal cost on unseen tasks.

AIBullisharXiv โ€“ CS AI ยท Mar 27/1012
๐Ÿง 

FedNSAM:Consistency of Local and Global Flatness for Federated Learning

Researchers propose FedNSAM, a new federated learning algorithm that improves global model performance by addressing the inconsistency between local and global flatness in distributed training environments. The algorithm uses global Nesterov momentum to harmonize local and global optimization, showing superior performance compared to existing FedSAM approaches.

AIBullisharXiv โ€“ CS AI ยท Mar 27/1012
๐Ÿง 

Rudder: Steering Prefetching in Distributed GNN Training using LLM Agents

Researchers introduced Rudder, a software module that uses Large Language Models (LLMs) to optimize data prefetching in distributed Graph Neural Network training. The system shows up to 91% performance improvement over baseline training and 82% over static prefetching by autonomously adapting to dynamic conditions.

AINeutralarXiv โ€“ CS AI ยท Mar 27/1015
๐Ÿง 

What Makes a Reward Model a Good Teacher? An Optimization Perspective

Research reveals that reward model accuracy alone doesn't determine effectiveness in RLHF systems. The study proves that low reward variance can create flat optimization landscapes, making even perfectly accurate reward models inefficient teachers that underperform less accurate models with higher variance.

AIBullisharXiv โ€“ CS AI ยท Feb 276/107
๐Ÿง 

Duel-Evolve: Reward-Free Test-Time Scaling via LLM Self-Preferences

Researchers introduce Duel-Evolve, a new optimization algorithm that improves LLM performance at test time without requiring external rewards or labels. The method uses self-generated pairwise comparisons and achieved 20 percentage points higher accuracy on MathBench and 12 percentage points improvement on LiveCodeBench.

AINeutralarXiv โ€“ CS AI ยท Feb 275/105
๐Ÿง 

Scaling Laws for Precision in High-Dimensional Linear Regression

Researchers developed theoretical scaling laws for low-precision AI model training, analyzing how quantization affects model performance in high-dimensional linear regression. The study reveals that multiplicative and additive quantization schemes have distinct effects on effective model size, with multiplicative maintaining full precision while additive reduces it.

AIBullisharXiv โ€“ CS AI ยท Feb 276/106
๐Ÿง 

Q$^2$: Quantization-Aware Gradient Balancing and Attention Alignment for Low-Bit Quantization

Researchers propose Qยฒ, a new framework that addresses gradient imbalance issues in quantization-aware training for complex visual tasks like object detection and image segmentation. The method achieves significant performance improvements (+2.5% mAP for object detection, +3.7% mDICE for segmentation) while introducing no inference-time overhead.

$ADA
AIBearisharXiv โ€“ CS AI ยท Feb 276/106
๐Ÿง 

ConstraintBench: Benchmarking LLM Constraint Reasoning on Direct Optimization

Researchers introduced ConstraintBench, a new benchmark testing whether large language models can directly solve constrained optimization problems without external solvers. The study found that even the best frontier models only achieve 65% constraint satisfaction, with feasibility being a bigger challenge than optimality.

AIBullisharXiv โ€“ CS AI ยท Feb 276/107
๐Ÿง 

A Minimum Variance Path Principle for Accurate and Stable Score-Based Density Ratio Estimation

Researchers propose the Minimum Variance Path (MVP) Principle to improve score-based machine learning methods by addressing the path variance problem that makes theoretically path-independent methods practically path-dependent. The approach uses a closed-form variance expression and Kumaraswamy Mixture Model to learn data-adaptive, low-variance paths, achieving new state-of-the-art results on benchmarks.

AIBullisharXiv โ€“ CS AI ยท Feb 276/106
๐Ÿง 

Exploratory Memory-Augmented LLM Agent via Hybrid On- and Off-Policy Optimization

Researchers propose EMPOยฒ, a new hybrid reinforcement learning framework that improves exploration capabilities for large language model agents by combining memory augmentation with on- and off-policy optimization. The framework achieves significant performance improvements of 128.6% on ScienceWorld and 11.3% on WebShop compared to existing methods, while demonstrating superior adaptability to new tasks without requiring parameter updates.

AIBullisharXiv โ€“ CS AI ยท Feb 276/106
๐Ÿง 

Large Language Model Compression with Global Rank and Sparsity Optimization

Researchers propose a novel two-stage compression method for Large Language Models that uses global rank and sparsity optimization to significantly reduce model size. The approach combines low-rank and sparse matrix decomposition with probabilistic global allocation to automatically detect redundancy across different layers and manage component interactions.