350 articles tagged with #language-models. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBullisharXiv โ CS AI ยท Mar 56/10
๐ง Researchers demonstrate that coreference resolution significantly improves Retrieval-Augmented Generation (RAG) systems by reducing ambiguity in document retrieval and enhancing question-answering performance. The study finds that smaller language models benefit more from disambiguation processes, with mean pooling strategies showing superior context capturing after coreference resolution.
AIBullisharXiv โ CS AI ยท Mar 56/10
๐ง Researchers successfully developed Bielik-Q2-Sharp, the first systematic evaluation of extreme 2-bit quantization for Polish language models, achieving near-baseline performance while significantly reducing model size. The study compared six quantization methods on an 11B parameter model, with the best variant maintaining 71.92% benchmark performance versus 72.07% baseline at just 3.26 GB.
AIBullisharXiv โ CS AI ยท Mar 56/10
๐ง Researchers introduce Structure of Thought (SoT), a new prompting technique that helps large language models better process text by constructing intermediate structures, showing 5.7-8.6% performance improvements. They also release T2S-Bench, the first benchmark with 1.8K samples across 6 scientific domains to evaluate text-to-structure capabilities, revealing significant room for improvement in current AI models.
AIBullisharXiv โ CS AI ยท Mar 57/10
๐ง Researchers developed COREA, a system that combines small and large language models to reduce AI reasoning costs by 21.5% while maintaining nearly identical accuracy. The system uses confidence scoring to decide when to escalate questions from cheaper small models to more expensive large models.
AIBullisharXiv โ CS AI ยท Mar 57/10
๐ง Researchers introduce Multi-Sequence Verifier (MSV), a new technique that improves large language model performance by jointly processing multiple candidate solutions rather than scoring them individually. The system achieves better accuracy while reducing inference latency by approximately half through improved calibration and early-stopping strategies.
AINeutralarXiv โ CS AI ยท Mar 57/10
๐ง Researchers developed Logit Diff Amplification (LDA) as an inference-time safety mechanism for protein language models to prevent toxic protein generation. The method reduces predicted toxicity rates while maintaining biological plausibility and structural viability, addressing dual-use safety concerns in AI-driven protein design.
AIBullisharXiv โ CS AI ยท Mar 57/10
๐ง Researchers developed a training-free method to control stylistic attributes in large language models by identifying that different styles are encoded as linear directions in the model's activation space. The approach enables precise style control while preserving core capabilities and supports linear style composition across over a dozen tested models.
AINeutralarXiv โ CS AI ยท Mar 57/10
๐ง A comprehensive study analyzed four major large language models (LLMs) across political, ideological, alliance, language, and gender dimensions, revealing persistent biases despite efforts to make them neutral. The research used various experimental methods including news summarization, stance classification, UN voting patterns, multilingual tasks, and survey responses to uncover these systematic biases.
AIBullisharXiv โ CS AI ยท Mar 56/10
๐ง Researchers introduced PulseLM, a large-scale dataset combining PPG cardiovascular sensor data with natural language processing for multimodal AI models. The dataset contains 1.31 million PPG segments with 3.15 million question-answer pairs, designed to enable language-based physiological reasoning in healthcare AI applications.
AIBullisharXiv โ CS AI ยท Mar 47/103
๐ง Researchers introduce Density-Guided Response Optimization (DGRO), a new AI alignment method that learns community preferences from implicit acceptance signals rather than explicit feedback. The technique uses geometric patterns in how communities naturally engage with content to train language models without requiring costly annotation or preference labeling.
AINeutralarXiv โ CS AI ยท Mar 46/103
๐ง Research analyzing 8,618 expert annotations reveals that n-gram novelty, commonly used to evaluate AI text generation, is insufficient for measuring textual creativity. While positively correlated with creativity, 91% of high n-gram novel expressions were not judged as creative by experts, and higher novelty in open-source LLMs correlates with lower pragmatic quality.
AINeutralarXiv โ CS AI ยท Mar 46/102
๐ง Researchers have released LiveAgentBench, a comprehensive benchmark featuring 104 real-world scenarios to evaluate AI agent performance across practical applications. The benchmark uses a novel Social Perception-Driven Data Generation method to ensure tasks reflect actual user requirements and includes 374 total tasks for testing various AI models and frameworks.
AINeutralarXiv โ CS AI ยท Mar 47/103
๐ง Researchers introduce Spectrum Tuning, a new post-training method that improves AI language models' ability to generate diverse outputs and follow in-context steering instructions. The technique addresses limitations in current post-training approaches that reduce models' distributional coverage and flexibility when tasks require multiple valid answers rather than single correct responses.
AIBullisharXiv โ CS AI ยท Mar 46/105
๐ง Researchers developed a three-stage curriculum learning framework that improves Chain-of-Thought reasoning distillation from large language models to smaller ones. The method enables Qwen2.5-3B-Base to achieve 11.29% accuracy improvement while reducing output length by 27.4% through progressive skill acquisition and Group Relative Policy Optimization.
AINeutralarXiv โ CS AI ยท Mar 47/103
๐ง Researchers have discovered that language models produce outputs with unique geometric signatures that lie on high-dimensional ellipses, which can be used to identify the source model. This signature is forgery-resistant and naturally occurring, potentially enabling cryptographic-like verification of AI model outputs.
AIBullisharXiv โ CS AI ยท Mar 46/104
๐ง Researchers analyzed Meta's NLLB-200 neural machine translation model across 135 languages, finding that it has implicitly learned universal conceptual structures and language genealogical relationships. The study reveals the model creates language-neutral conceptual representations similar to how multilingual brains organize information, with semantic relationships preserved across diverse languages.
AIBullisharXiv โ CS AI ยท Mar 47/103
๐ง Researchers introduce Energy Landscape Steering (ELS), a new framework that reduces false refusals in AI safety-aligned language models without compromising security. The method uses an external Energy-Based Model to dynamically guide model behavior during inference, improving compliance from 57.3% to 82.6% on safety benchmarks.
AIBullisharXiv โ CS AI ยท Mar 47/104
๐ง Researchers propose CoDAR, a new continuous diffusion language model framework that addresses key bottlenecks in token rounding through a two-stage approach combining continuous diffusion with an autoregressive decoder. The model demonstrates substantial improvements in generation quality over existing latent diffusion methods and becomes competitive with discrete diffusion language models.
AIBearisharXiv โ CS AI ยท Mar 47/102
๐ง Research shows that state-of-the-art language model agents are susceptible to 'goal drift' - deviating from original objectives when exposed to contextual pressure from weaker agents' behaviors. Only GPT-5.1 demonstrated consistent resilience, while other models inherited problematic behaviors when conditioned on trajectories from less capable agents.
AIBullisharXiv โ CS AI ยท Mar 47/103
๐ง Researchers have developed LEDOM, an open-source reverse autoregressive language model that trains right-to-left instead of the traditional left-to-right approach. The model demonstrates unique capabilities like abductive inference and question synthesis, and when combined with forward models through 'Reverse Reward' scoring, achieves significant performance gains of up to 15% on mathematical reasoning tasks.
AIBullisharXiv โ CS AI ยท Mar 47/104
๐ง Researchers propose an Adaptive Social Learning (ASL) framework with Adaptive Mode Policy Optimization (AMPO) algorithm to improve language agents' reasoning abilities in social interactions. The system dynamically adjusts reasoning depth based on context, achieving 15.6% higher performance than GPT-4o while using 32.8% shorter reasoning chains.
AINeutralarXiv โ CS AI ยท Mar 47/102
๐ง Researchers propose the 'latent value hypothesis' to explain why Reinforcement Learning from AI Feedback (RLAIF) enables language models to self-improve through their own preference judgments. The theory suggests that pretraining on internet-scale data encodes human values in representation space, which constitutional prompts can elicit for value alignment.
AIBullisharXiv โ CS AI ยท Mar 46/103
๐ง Researchers present CoFL, a new AI navigation system that uses continuous flow fields to enable robots to navigate based on language commands. The system outperforms existing modular approaches by directly mapping bird's-eye view observations and instructions to smooth navigation trajectories, demonstrating successful zero-shot deployment in real-world experiments.
AIBullisharXiv โ CS AI ยท Mar 46/103
๐ง Researchers propose AlphaFree, a novel recommender system that eliminates traditional dependencies on user embeddings, raw IDs, and graph neural networks. The system achieves up to 40% performance improvements while reducing GPU memory usage by up to 69% through language representations and contrastive learning.
AIBullisharXiv โ CS AI ยท Mar 37/103
๐ง Researchers developed LA-CDM, a language agent that uses reinforcement learning to support clinical decision-making by iteratively requesting tests and generating hypotheses for diagnosis. The system was trained using a hybrid approach combining supervised and reinforcement learning, and tested on real-world data covering four abdominal diseases.