282 articles tagged with #optimization. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBullisharXiv β CS AI Β· 2d ago7/10
π§ Researchers propose Proximal Supervised Fine-Tuning (PSFT), a new method that applies trust-region constraints from reinforcement learning to improve how foundation models adapt to new tasks. The technique maintains model capabilities while fine-tuning, outperforming standard supervised fine-tuning on out-of-domain generalization tasks.
AIBullisharXiv β CS AI Β· 2d ago7/10
π§ Researchers introduce PnP-CM, a new method that reformulates consistency models as proximal operators within plug-and-play frameworks for solving inverse problems. The approach achieves high-quality image reconstructions with minimal neural function evaluations (4 NFEs), demonstrating practical efficiency gains over existing consistency model solvers and marking the first application of CMs to MRI data.
AIBullisharXiv β CS AI Β· 2d ago7/10
π§ A comprehensive tutorial examines how deep learning complements operations research and optimization for sequential decision-making under uncertainty. The framework positions AI not as a replacement for traditional optimization but as an enhancement, with applications across supply chains, healthcare, energy, and autonomous systems.
AIBullisharXiv β CS AI Β· 3d ago7/10
π§ AlphaLab is an autonomous research system using frontier LLMs to automate experimental cycles across computational domains. Without human intervention, it explores datasets, validates frameworks, and runs large-scale experiments while accumulating domain knowledgeβachieving 4.4x speedups in CUDA optimization, 22% lower validation loss in LLM pretraining, and 23-25% improvements in traffic forecasting.
π§ GPT-5π§ Claudeπ§ Opus
AIBullisharXiv β CS AI Β· 3d ago7/10
π§ Researchers introduce CSAttention, a training-free sparse attention method that accelerates LLM inference by 4.6x for long-context applications. The technique optimizes the offline-prefill/online-decode workflow by precomputing query-centric lookup tables, enabling faster token generation without sacrificing accuracy even at 95% sparsity levels.
AIBullisharXiv β CS AI Β· 3d ago7/10
π§ Researchers propose Advantage-Guided Diffusion (AGD-MBRL), a novel approach that improves model-based reinforcement learning by using advantage estimates to guide diffusion models during trajectory generation. The method addresses the short-horizon myopia problem in existing diffusion-based world models and demonstrates 2x performance improvements over current baselines on MuJoCo control tasks.
AIBullisharXiv β CS AI Β· Apr 77/10
π§ Researchers propose a new method for aligning AI language models with human preferences that addresses stability issues in existing approaches. The technique uses relative density ratio optimization to achieve both statistical consistency and training stability, showing effectiveness with Qwen 2.5 and Llama 3 models.
π§ Llama
AIBullisharXiv β CS AI Β· Apr 77/10
π§ Researchers propose SLaB, a novel framework for compressing large language models by decomposing weight matrices into sparse, low-rank, and binary components. The method achieves significant improvements over existing compression techniques, reducing perplexity by up to 36% at 50% compression rates without requiring model retraining.
π’ Perplexityπ§ Llama
AIBullisharXiv β CS AI Β· Apr 67/10
π§ Research shows that large language models significantly outperform traditional AI planning algorithms on complex block-moving problems, tracking theoretical optimality limits with near-perfect precision. The study suggests LLMs may use algorithmic simulation and geometric memory to bypass exponential combinatorial complexity in planning tasks.
AIBullisharXiv β CS AI Β· Apr 67/10
π§ Researchers introduce Textual Equilibrium Propagation (TEP), a new method to optimize large language model compound AI systems that addresses performance degradation in deep, multi-module workflows. TEP uses local learning principles to avoid exploding and vanishing gradient problems that plague existing global feedback methods like TextGrad.
AIBullisharXiv β CS AI Β· Apr 67/10
π§ Researchers have developed ClinicalReTrial, a multi-agent AI system that can redesign clinical trial protocols to improve success rates. The system demonstrated an 83.3% improvement rate in trial protocols with a mean 5.7% increase in success probability at minimal cost of $0.12 per trial.
AIBullisharXiv β CS AI Β· Mar 277/10
π§ Researchers propose GlowQ, a new quantization technique for large language models that reduces memory overhead and latency while maintaining accuracy. The method uses group-shared low-rank approximation to optimize deployment of quantized LLMs, showing significant performance improvements over existing approaches.
π’ Perplexity
AIBullisharXiv β CS AI Β· Mar 267/10
π§ Researchers introduce Hybrid Distillation Policy Optimization (HDPO), a new method that improves large language model training for mathematical reasoning by addressing 'cliff prompts' where standard reinforcement learning fails. The technique uses privileged self-distillation to provide learning signals for previously unsolvable problems, showing measurable improvements in coverage metrics while maintaining accuracy.
AINeutralarXiv β CS AI Β· Mar 267/10
π§ Researchers propose a theory of LLM information susceptibility that identifies fundamental limits to how large language models can improve optimization in AI agent systems. The study shows that nested, co-scaling architectures may be necessary for open-ended AI self-improvement, providing predictive constraints for AI system design.
AIBullisharXiv β CS AI Β· Mar 267/10
π§ Researchers have developed QUARK, a quantization-enabled FPGA acceleration framework that significantly improves Transformer model performance by optimizing nonlinear operations through circuit sharing. The system achieves up to 1.96x speedup over GPU implementations while reducing hardware overhead by more than 50% compared to existing approaches.
AIBullisharXiv β CS AI Β· Mar 267/10
π§ Researchers have developed DVM, a real-time compiler for dynamic AI models that uses bytecode virtual machine technology to significantly speed up compilation times. The system achieves up to 11.77x better operator/model efficiency and up to 5 orders of magnitude faster compilation compared to existing solutions like TorchInductor and PyTorch.
AIBullisharXiv β CS AI Β· Mar 177/10
π§ Researchers introduce Mask Fine-Tuning (MFT), a novel approach that improves large language model performance by applying binary masks to optimized models without updating weights. The method achieves consistent performance gains across different domains and model architectures, with average improvements of 2.70/4.15 in IFEval benchmarks for LLaMA models.
AIBullisharXiv β CS AI Β· Mar 177/10
π§ Researchers propose Simple Energy Adaptation (SEA), a new algorithm for aligning large language models with human feedback at inference time. SEA uses gradient-based sampling in continuous latent space rather than searching discrete response spaces, achieving up to 77.51% improvement on AdvBench and 16.36% on MATH benchmarks.
AIBullisharXiv β CS AI Β· Mar 177/10
π§ Researchers introduce EcoAlign, a new framework for aligning Large Vision-Language Models that treats alignment as an economic optimization problem. The method balances safety, utility, and computational costs while preventing harmful reasoning disguised with benign justifications, showing superior performance across multiple models and datasets.
AIBullisharXiv β CS AI Β· Mar 177/10
π§ Researchers introduce POLCA (Prioritized Optimization with Local Contextual Aggregation), a new framework that uses large language models as optimizers for complex systems like AI agents and code generation. The method addresses stochastic optimization challenges through priority queuing and meta-learning, demonstrating superior performance across multiple benchmarks including agent optimization and CUDA kernel generation.
AIBullisharXiv β CS AI Β· Mar 177/10
π§ PrototypeNAS is a new zero-shot neural architecture search method that rapidly designs and optimizes deep neural networks for microcontroller units without requiring extensive training. The system uses a three-step approach combining structural optimization, ensemble zero-shot proxies, and Hypervolume subset selection to identify efficient models within minutes that can run on resource-constrained edge devices.
AINeutralarXiv β CS AI Β· Mar 177/10
π§ Researchers developed Prefix-Shared KV Cache (PSKV), a new technique that accelerates jailbreak attacks on Large Language Models by 40% while reducing memory usage by 50%. The method optimizes the red-teaming process by sharing cached prefixes across multiple attack attempts, enabling more efficient parallel inference without compromising attack success rates.
AIBullisharXiv β CS AI Β· Mar 177/10
π§ Justitia is a new scheduling system for task-parallel LLM agents that optimizes GPU server performance through selective resource allocation based on completion order prediction. The system uses memory-centric cost quantification and virtual-time fair queuing to achieve both efficiency and fairness in LLM serving environments.
π’ Meta
AIBullisharXiv β CS AI Β· Mar 177/10
π§ ADV-0 is a new closed-loop adversarial training framework for autonomous driving that uses min-max optimization to improve robustness against rare but safety-critical scenarios. The system treats the interaction between driving policy and adversarial agents as a zero-sum game, converging to Nash Equilibrium while maximizing real-world performance bounds.
AIBullisharXiv β CS AI Β· Mar 167/10
π§ Researchers introduce a novel optimization framework that integrates the Minimum Description Length (MDL) principle directly into deep neural network training dynamics. The method uses geometrically-grounded cognitive manifolds with coupled Ricci flow to create autonomous model simplification while maintaining data fidelity, with theoretical guarantees for convergence and practical O(N log N) complexity.