y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#optimization News & Analysis

282 articles tagged with #optimization. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

282 articles
AIBullisharXiv – CS AI · Mar 45/103
🧠

Quantum-Inspired Fine-Tuning for Few-Shot AIGC Detection via Phase-Structured Reparameterization

Researchers propose Q-LoRA, a quantum-enhanced fine-tuning method that integrates quantum neural networks into LoRA adapters for improved AI-generated content detection. The study also introduces H-LoRA, a classical variant using Hilbert transforms that achieves similar 5%+ accuracy improvements over standard LoRA at lower computational cost.

AIBullisharXiv – CS AI · Mar 45/102
🧠

Enhancing Physics-Informed Neural Networks with Domain-aware Fourier Features: Towards Improved Performance and Interpretable Results

Researchers have developed Domain-aware Fourier Features (DaFFs) to enhance Physics-Informed Neural Networks (PINNs), achieving orders-of-magnitude lower errors and faster convergence. The approach incorporates domain-specific characteristics like geometry and boundary conditions while eliminating the need for explicit boundary condition loss terms, making PINNs more accurate, efficient, and interpretable.

AINeutralarXiv – CS AI · Mar 45/103
🧠

Why Adam Can Beat SGD: Second-Moment Normalization Yields Sharper Tails

Research paper establishes the first theoretical separation between Adam and SGD optimization algorithms, proving Adam achieves better high-probability convergence guarantees. The study provides mathematical backing for Adam's superior empirical performance through second-moment normalization analysis.

AINeutralarXiv – CS AI · Mar 37/108
🧠

DIVA-GRPO: Enhancing Multimodal Reasoning through Difficulty-Adaptive Variant Advantage

Researchers have developed DIVA-GRPO, a new reinforcement learning method that improves multimodal large language model reasoning by adaptively adjusting problem difficulty distributions. The approach addresses key limitations in existing group relative policy optimization methods, showing superior performance across six reasoning benchmarks.

AIBullisharXiv – CS AI · Mar 36/109
🧠

Information-Theoretic Framework for Self-Adapting Model Predictive Controllers

Researchers introduced Entanglement Learning (EL), an information-theoretic framework that enhances Model Predictive Control (MPC) for autonomous systems like UAVs. The framework uses an Information Digital Twin to monitor information flow and enable real-time adaptive optimization, improving MPC reliability beyond traditional error-based feedback systems.

AIBullisharXiv – CS AI · Mar 37/107
🧠

Learning Structured Reasoning via Tractable Trajectory Control

Researchers propose Ctrl-R, a new framework that improves large language models' reasoning abilities by systematically discovering and reinforcing diverse reasoning patterns through structured trajectory control. The method enables better exploration of complex reasoning behaviors and shows consistent improvements across mathematical reasoning tasks in both language and vision-language models.

AIBullisharXiv – CS AI · Mar 37/106
🧠

Expert Divergence Learning for MoE-based Language Models

Researchers introduce Expert Divergence Learning, a new pre-training strategy for Mixture-of-Experts language models that prevents expert homogenization by encouraging functional specialization. The method uses domain labels to maximize routing distribution differences between data domains, achieving better performance on 15 billion parameter models with minimal computational overhead.

AINeutralarXiv – CS AI · Mar 37/109
🧠

Universal NP-Hardness of Clustering under General Utilities

Researchers prove that clustering problems in machine learning are universally NP-hard, providing theoretical explanation for why clustering algorithms often produce unstable results. The study demonstrates that major clustering methods like k-means and spectral clustering inherit fundamental computational intractability, explaining common failure modes like local optima.

AIBullisharXiv – CS AI · Mar 37/107
🧠

MuonRec: Shifting the Optimizer Paradigm Beyond Adam in Scalable Generative Recommendation

Researchers introduce MuonRec, a new optimization framework for recommendation systems that significantly outperforms the widely-used Adam/AdamW optimizers. The framework reduces training steps by 32.4% on average while improving ranking quality by 12.6% in NDCG@10 metrics across traditional and generative recommenders.

AIBullisharXiv – CS AI · Mar 36/109
🧠

QANTIS: A Hardware-Validated Quantum Platform for POMDP Planning and Multi-Target Data Association

QANTIS is a hardware-validated quantum computing platform that demonstrates quadratic improvements in autonomous navigation planning problems and multi-target data association tasks. The research shows successful implementation on IBM quantum hardware, achieving 5.1x amplification of rare observation probabilities while maintaining Bayesian posterior accuracy.

AINeutralarXiv – CS AI · Mar 37/107
🧠

EraseAnything++: Enabling Concept Erasure in Rectified Flow Transformers Leveraging Multi-Object Optimization

Researchers introduced EraseAnything++, a new framework for removing unwanted concepts from advanced AI image and video generation models like Stable Diffusion v3 and Flux. The method uses multi-objective optimization to balance concept removal while preserving overall generative quality, showing superior performance compared to existing approaches.

AI × CryptoBullisharXiv – CS AI · Mar 37/1010
🤖

Communication-Efficient Quantum Federated Learning over Large-Scale Wireless Networks

Researchers present a novel quantum federated learning framework for large-scale wireless networks that combines quantum computing with privacy-preserving federated learning. The study introduces a sum-rate maximization approach using quantum approximate optimization algorithm (QAOA) that achieves over 100% improvement in performance compared to conventional methods.

AIBullisharXiv – CS AI · Mar 36/109
🧠

Provable and Practical In-Context Policy Optimization for Self-Improvement

Researchers introduce In-Context Policy Optimization (ICPO), a new method that allows AI models to improve their responses during inference through multi-round self-reflection without parameter updates. The practical ME-ICPO algorithm demonstrates competitive performance on mathematical reasoning tasks while maintaining affordable inference costs.

AIBullisharXiv – CS AI · Mar 37/107
🧠

LFPO: Likelihood-Free Policy Optimization for Masked Diffusion Models

Researchers propose Likelihood-Free Policy Optimization (LFPO), a new framework for improving Diffusion Large Language Models by bypassing likelihood computation issues that plague existing methods. LFPO uses geometric velocity rectification to optimize denoising logits directly, achieving better performance on code and reasoning tasks while reducing inference time by 20%.

AIBullisharXiv – CS AI · Mar 36/108
🧠

FAST-DIPS: Adjoint-Free Analytic Steps and Hard-Constrained Likelihood Correction for Diffusion-Prior Inverse Problems

Researchers propose FAST-DIPS, a new training-free diffusion prior method for solving inverse problems that achieves up to 19.5x speedup while maintaining competitive image quality metrics. The method replaces computationally expensive inner optimization loops with closed-form projections and analytic step sizes, significantly reducing the number of required denoiser evaluations.

AIBullisharXiv – CS AI · Mar 36/109
🧠

Surgical Post-Training: Cutting Errors, Keeping Knowledge

Researchers introduce Surgical Post-Training (SPoT), a new method to improve Large Language Model reasoning while preventing catastrophic forgetting. SPoT achieved 6.2% accuracy improvement on Qwen3-8B using only 4k data pairs and 28 minutes of training, offering a more efficient alternative to traditional post-training approaches.

AINeutralarXiv – CS AI · Mar 36/104
🧠

AMemGym: Interactive Memory Benchmarking for Assistants in Long-Horizon Conversations

Researchers introduce AMemGym, an interactive benchmarking environment for evaluating and optimizing memory management in long-horizon conversations with AI assistants. The framework addresses limitations in current memory evaluation methods by enabling on-policy testing with LLM-simulated users and revealing performance gaps in existing memory systems like RAG and long-context LLMs.

AIBullisharXiv – CS AI · Mar 36/104
🧠

OrbitFlow: SLO-Aware Long-Context LLM Serving with Fine-Grained KV Cache Reconfiguration

OrbitFlow is a new KV cache management system for long-context LLM serving that uses adaptive memory allocation and fine-grained optimization to improve performance. The system achieves up to 66% better SLO attainment and 3.3x higher throughput by dynamically managing GPU memory usage during token generation.

AIBullisharXiv – CS AI · Mar 36/103
🧠

MeanCache: From Instantaneous to Average Velocity for Accelerating Flow Matching Inference

MeanCache introduces a training-free caching framework that accelerates Flow Matching inference by using average velocities instead of instantaneous ones. The framework achieves 3.59X to 4.56X acceleration on major AI models like FLUX.1, Qwen-Image, and HunyuanVideo while maintaining superior generation quality compared to existing caching methods.

AIBullisharXiv – CS AI · Mar 36/104
🧠

FMIP: Joint Continuous-Integer Flow For Mixed-Integer Linear Programming

Researchers have developed FMIP, a new generative AI framework that models both integer and continuous variables simultaneously to solve Mixed-Integer Linear Programming problems more efficiently. The approach reduces the primal gap by 41.34% on average compared to existing baselines and is compatible with various downstream solvers.

AIBullisharXiv – CS AI · Mar 36/103
🧠

Training Large Language Models To Reason In Parallel With Global Forking Tokens

Researchers developed Set Supervised Fine-Tuning (SSFT) and Global Forking Policy Optimization (GFPO) methods to improve large language model reasoning by enabling parallel processing through 'global forking tokens.' The techniques preserve diverse reasoning modes and demonstrate superior performance on math and code generation benchmarks compared to traditional fine-tuning approaches.

AIBullisharXiv – CS AI · Mar 36/104
🧠

Contribution-aware Token Compression for Efficient Video Understanding via Reinforcement Learning

Researchers developed CaCoVID, a reinforcement learning-based algorithm that compresses video tokens for large language models by selecting tokens based on their actual contribution to correct predictions rather than attention scores. The method uses combinatorial policy optimization to reduce computational overhead while maintaining video understanding performance.

← PrevPage 6 of 12Next →