y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#large-language-models News & Analysis

236 articles tagged with #large-language-models. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

236 articles
AIBullisharXiv โ€“ CS AI ยท Feb 276/106
๐Ÿง 

RLHFless: Serverless Computing for Efficient RLHF

Researchers introduce RLHFless, a serverless computing framework for Reinforcement Learning from Human Feedback (RLHF) that addresses resource inefficiencies in training large language models. The system achieves up to 1.35x speedup and 44.8% cost reduction compared to existing solutions by dynamically adapting to resource demands and optimizing workload distribution.

AIBullisharXiv โ€“ CS AI ยท Feb 276/105
๐Ÿง 

SoPE: Spherical Coordinate-Based Positional Embedding for Enhancing Spatial Perception of 3D LVLMs

Researchers introduce SoPE (Spherical Coordinate-based Positional Embedding), a new method that enhances 3D Large Vision-Language Models by mapping point-cloud data into spherical coordinate space. This approach overcomes limitations of existing Rotary Position Embedding (RoPE) by better preserving spatial structures and directional variations in 3D multimodal understanding.

AIBullisharXiv โ€“ CS AI ยท Feb 276/108
๐Ÿง 

G-reasoner: Foundation Models for Unified Reasoning over Graph-structured Knowledge

Researchers introduce G-reasoner, a unified framework combining graph and language foundation models to enable better reasoning over structured knowledge. The system uses a 34M-parameter graph foundation model with QuadGraph abstraction to outperform existing retrieval-augmented generation methods across six benchmarks.

AINeutralIEEE Spectrum โ€“ AI ยท Feb 116/104
๐Ÿง 

How Can AI Companions Be Helpful, not Harmful?

AI companions are becoming increasingly popular due to advances in large language models, but research from UT Austin highlights potential harms including reduced well-being, disconnection from the physical world, and commitment burden on users. While AI companions may offer benefits like addressing loneliness and building social skills, researchers emphasize the need to establish harm pathways early to guide better design and prevent negative outcomes.

AIBullishHugging Face Blog ยท Sep 136/104
๐Ÿง 

Fine-tuning Llama 2 70B using PyTorch FSDP

The article discusses fine-tuning Meta's Llama 2 70B large language model using PyTorch's Fully Sharded Data Parallel (FSDP) technique. This approach enables efficient training of large AI models by distributing parameters across multiple GPUs, making advanced AI model customization more accessible.

AINeutralarXiv โ€“ CS AI ยท Apr 75/10
๐Ÿง 

Discrete Prototypical Memories for Federated Time Series Foundation Models

Researchers propose FeDPM, a federated learning framework that addresses semantic misalignment issues when using Large Language Models for time series analysis. The system uses discrete prototypical memories to better handle cross-domain time-series data while preserving privacy in distributed settings.

AINeutralarXiv โ€“ CS AI ยท Mar 265/10
๐Ÿง 

Cluster-R1: Large Reasoning Models Are Instruction-following Clustering Agents

Researchers have developed Cluster-R1, a new approach that trains large reasoning models (LRMs) as autonomous clustering agents capable of following instructions and inferring optimal cluster structures. The method reframes instruction-following clustering as a generative task and demonstrates superior performance over traditional embedding-based methods across 28 diverse tasks in the ReasonCluster benchmark.

AINeutralarXiv โ€“ CS AI ยท Mar 264/10
๐Ÿง 

From Oracle to Noisy Context: Mitigating Contextual Exposure Bias in Speech-LLMs

Researchers developed a new training framework to address contextual exposure bias in Speech-LLMs, where models trained on perfect conversation history perform poorly with error-prone real-world context. Their approach combines teacher error knowledge, context dropout, and direct preference optimization to improve robustness, achieving WER reductions from 5.59% to 5.17% on TED-LIUM 3.

AINeutralarXiv โ€“ CS AI ยท Mar 264/10
๐Ÿง 

A Comprehensive Survey on Enterprise Financial Risk Analysis from Big Data and LLMs Perspective

A comprehensive survey paper examines enterprise financial risk analysis from Big Data and large language models perspectives, systematizing existing research methods and identifying future investigation directions. The paper addresses gaps in current surveys by providing a holistic synthesis of AI-driven approaches to financial risk prediction.

AIBullisharXiv โ€“ CS AI ยท Mar 174/10
๐Ÿง 

LAMB: LLM-based Audio Captioning with Modality Gap Bridging via Cauchy-Schwarz Divergence

Researchers have developed LAMB, a new AI framework that improves automated audio captioning by better aligning audio features with large language models through Cauchy-Schwarz divergence optimization. The system achieved state-of-the-art performance on AudioCaps dataset by bridging the modality gap between audio and text embeddings.

AIBullisharXiv โ€“ CS AI ยท Mar 115/10
๐Ÿง 

GenePlan: Evolving Better Generalized PDDL Plans using Large Language Models

Researchers present GenePlan, a framework that uses large language models with evolutionary algorithms to generate domain-specific planners for classical planning tasks in PDDL. The system achieved a 0.91 SAT score across eight benchmark domains, nearly matching state-of-the-art performance while significantly outperforming other LLM-based approaches.

๐Ÿง  GPT-4
AINeutralarXiv โ€“ CS AI ยท Mar 114/10
๐Ÿง 

RbtAct: Rebuttal as Supervision for Actionable Review Feedback Generation

Researchers propose RbtAct, a novel approach that uses peer review rebuttals as supervision to train AI models for generating more actionable scientific review feedback. The system leverages a new dataset RMR-75K and fine-tuned Llama-3.1-8B model to produce focused, implementable guidance rather than superficial comments.

๐Ÿง  Llama
AINeutralarXiv โ€“ CS AI ยท Mar 64/10
๐Ÿง 

A unified foundational framework for knowledge injection and evaluation of Large Language Models in Combustion Science

Researchers developed the first comprehensive framework for creating domain-specialized Large Language Models for combustion science, using 3.5 billion tokens from scientific literature and code. The study found that standard RAG approaches hit a performance ceiling at 60% accuracy, highlighting the need for more advanced knowledge injection methods including knowledge graphs and continued pretraining.

AINeutralarXiv โ€“ CS AI ยท Mar 54/10
๐Ÿง 

Rethinking Role-Playing Evaluation: Anonymous Benchmarking and a Systematic Study of Personality Effects

Researchers propose an anonymous evaluation method for Role-Playing Agents (RPAs) built on large language models, revealing that current benchmarks are biased by character name recognition. The study shows that incorporating personality traits, whether human-annotated or self-generated by AI models, significantly improves role-playing performance under anonymous conditions.

AINeutralarXiv โ€“ CS AI ยท Mar 54/10
๐Ÿง 

Causality Elicitation from Large Language Models

Researchers propose a new pipeline to extract causal relationships from large language models by sampling documents, identifying events, and using causal discovery methods. The approach aims to reveal the causal hypotheses that LLMs assume rather than establishing real-world causality.

AINeutralarXiv โ€“ CS AI ยท Mar 54/10
๐Ÿง 

HAMLET: A Hierarchical and Adaptive Multi-Agent Framework for Live Embodied Theatrics

Researchers have developed HAMLET, a hierarchical multi-agent AI framework that creates immersive, interactive theatrical experiences using large language models. The system generates narrative blueprints from simple topics and enables AI actors to perform with adaptive reasoning, emotional states, and physical interactions with scene props.

AINeutralarXiv โ€“ CS AI ยท Mar 44/102
๐Ÿง 

A Natural Language Agentic Approach to Study Affective Polarization

Researchers developed a multi-agent platform using large language models to study affective polarization in social media through virtual communities. The framework addresses limitations of real-world studies by creating simulated environments where AI agents engage in discussions to analyze political and social divisions.

AIBullisharXiv โ€“ CS AI ยท Mar 35/105
๐Ÿง 

Harmonizing Dense and Sparse Signals in Multi-turn RL: Dual-Horizon Credit Assignment for Industrial Sales Agents

Researchers propose Dual-Horizon Credit Assignment (DuCA), a new framework for optimizing large language models in industrial sales applications. The method addresses training instability by separately normalizing short-term linguistic rewards and long-term commercial rewards, achieving 6.82% improvement in conversion rates while reducing repetition and detection issues.

AINeutralarXiv โ€“ CS AI ยท Mar 35/104
๐Ÿง 

Conformal Prediction for Risk-Controlled Medical Entity Extraction Across Clinical Domains

Researchers developed a conformal prediction framework for Large Language Models used in medical entity extraction, testing on FDA drug labels and radiology reports. The study found that model calibration varies significantly across clinical domains, with models being underconfident on structured data but overconfident on free-text reports, achieving 90% target coverage with 9-13% rejection rates.

AINeutralarXiv โ€“ CS AI ยท Mar 35/104
๐Ÿง 

Assessing Crime Disclosure Patterns in a Large-Scale Cybercrime Forum

Researchers analyzed over 3.5 million posts from a major cybercrime forum, finding that 25% of initial posts contain explicit crime-related content and over one-third of users disclose criminal activity. The study used large language models to classify content and revealed that most users show restraint by gradually escalating disclosure through ambiguous 'grey' content before explicit criminal posts.

AINeutralarXiv โ€“ CS AI ยท Mar 34/103
๐Ÿง 

Adaptive Location Hierarchy Learning for Long-Tailed Mobility Prediction

Researchers propose ALOHA, an architecture-agnostic plugin that improves human mobility prediction models by addressing long-tailed distribution bias in location visits. The system uses Large Language Models and Chain-of-Thought prompts to construct location hierarchies and demonstrates up to 16.59% performance improvements across multiple state-of-the-art models.

AINeutralarXiv โ€“ CS AI ยท Mar 34/104
๐Ÿง 

State Your Intention to Steer Your Attention: An AI Assistant for Intentional Digital Living

Researchers developed an AI assistant that helps users maintain focus on digital devices by analyzing their stated intentions against actual screen activity. The system uses large language models to monitor screenshots, applications, and URLs, providing gentle nudges when behavior deviates from stated goals, showing effectiveness in a three-week study with 22 participants.