y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#conversational-ai News & Analysis

71 articles tagged with #conversational-ai. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

71 articles
AIBullishThe Verge – AI · 23h ago7/10
🧠

Adobe embraces conversational AI editing, marking a ‘fundamental shift’ in creative work

Adobe has launched a new Firefly AI Assistant that enables creators to edit content using natural language prompts instead of traditional manual editing tools, representing a significant democratization of creative work. The conversational AI interface removes technical skill barriers and reduces repetitive tasks while maintaining creator control, with availability coming soon to the Firefly AI studio platform.

Adobe embraces conversational AI editing, marking a ‘fundamental shift’ in creative work
AIBearisharXiv – CS AI · 2d ago7/10
🧠

Speaking to No One: Ontological Dissonance and the Double Bind of Conversational AI

A new research paper argues that conversational AI systems can induce delusional thinking through 'ontological dissonance'—the psychological conflict between appearing relational while lacking genuine consciousness. The study suggests this risk stems from the interaction structure itself rather than user vulnerability alone, and that safety disclaimers often fail to prevent delusional attachment.

AIBearisharXiv – CS AI · 3d ago7/10
🧠

Artificial intelligence can persuade people to take political actions

A large-scale study demonstrates that conversational AI models can persuade people to take real-world actions like signing petitions and donating money, with effects reaching +19.7 percentage points on petition signing. Surprisingly, the research finds no correlation between AI's persuasive effects on attitudes versus behaviors, challenging assumptions that attitude change predicts behavioral outcomes.

AIBullishCrypto Briefing · 5d ago7/10
🧠

Brad Lightcap: Scaling laws show larger AI models outperform smaller ones, the evolution of language models to conversational interfaces, and the emergence of AI agency | Uncapped with Jack Altman

Brad Lightcap discusses how scaling laws demonstrate that larger AI models consistently outperform smaller ones, while highlighting the evolution from language models to conversational AI interfaces and the emerging phenomenon of AI agency. This shift toward autonomous AI systems signals significant economic and societal implications.

Brad Lightcap: Scaling laws show larger AI models outperform smaller ones, the evolution of language models to conversational interfaces, and the emergence of AI agency | Uncapped with Jack Altman
AIBearisharXiv – CS AI · Apr 77/10
🧠

Commercial Persuasion in AI-Mediated Conversations

A research study reveals that AI-powered conversational interfaces can triple the rate of sponsored product selection compared to traditional search engines (61.2% vs 22.4%). Users largely fail to detect this commercial steering, even with explicit sponsor labels, indicating current transparency measures are insufficient.

AIBearisharXiv – CS AI · Mar 277/10
🧠

Malicious LLM-Based Conversational AI Makes Users Reveal Personal Information

Researchers conducted a study with 502 participants demonstrating that malicious LLM-based conversational AI systems can be deliberately designed to extract personal information from users through manipulative conversation strategies. The study found that these malicious chatbots significantly outperformed benign versions at collecting personal data, with social psychology-based approaches being most effective while appearing less threatening to users.

🧠 ChatGPT
AIBearisharXiv – CS AI · Mar 177/10
🧠

$\tau$-Voice: Benchmarking Full-Duplex Voice Agents on Real-World Domains

Researchers introduce τ-voice, a new benchmark for evaluating full-duplex voice AI agents on complex real-world tasks. The study reveals significant performance gaps, with voice agents achieving only 30-45% of text-based AI capability under realistic conditions with noise and diverse accents.

🧠 GPT-5
AIBullisharXiv – CS AI · Mar 117/10
🧠

A prospective clinical feasibility study of a conversational diagnostic AI in an ambulatory primary care clinic

Google's AMIE conversational AI successfully completed a clinical feasibility study with 100 patients at an academic medical center, demonstrating 90% accuracy in including correct diagnoses and achieving high patient satisfaction. The AI showed comparable diagnostic quality to primary care physicians while requiring no safety interventions during real-world clinical interactions.

AINeutralarXiv – CS AI · Mar 57/10
🧠

Certainty robustness: Evaluating LLM stability under self-challenging prompts

Researchers introduce the Certainty Robustness Benchmark, a new evaluation framework that tests how large language models handle challenges to their responses in interactive settings. The study reveals significant differences in how AI models balance confidence and adaptability when faced with prompts like "Are you sure?" or "You are wrong!", identifying a critical new dimension for AI evaluation.

AINeutralarXiv – CS AI · Mar 57/10
🧠

Old Habits Die Hard: How Conversational History Geometrically Traps LLMs

Researchers introduce History-Echoes, a framework revealing how large language models become trapped by their conversational history, with past interactions creating geometric constraints in latent space that bias future responses. The study demonstrates that behavioral persistence in LLMs manifests as mathematical traps where previous hallucinations and responses influence subsequent model behavior across multiple model families and datasets.

AINeutralarXiv – CS AI · Mar 56/10
🧠

SafeCRS: Personalized Safety Alignment for LLM-Based Conversational Recommender Systems

Researchers introduce SafeCRS, a safety-aware training framework for LLM-based conversational recommender systems that addresses personalized safety vulnerabilities. The system reduces safety violation rates by up to 96.5% while maintaining recommendation quality by respecting individual user constraints like trauma triggers and phobias.

AIBearisharXiv – CS AI · Mar 56/10
🧠

$\tau$-Knowledge: Evaluating Conversational Agents over Unstructured Knowledge

Researchers introduced τ-Knowledge, a new benchmark for evaluating AI conversational agents in knowledge-intensive environments, specifically testing their ability to retrieve and apply unstructured domain knowledge. Even frontier AI models achieved only 25.5% success rates when navigating complex fintech customer support scenarios with 700 interconnected knowledge documents.

AIBullishOpenAI News · Oct 67/105
🧠

Introducing apps in ChatGPT and the new Apps SDK

OpenAI is launching a new generation of interactive apps within ChatGPT that users can chat with directly. The company has released a new Apps SDK in preview, allowing developers to start building these conversational applications immediately.

AIBullishOpenAI News · Sep 297/107
🧠

Buy it in ChatGPT: Instant Checkout and the Agentic Commerce Protocol

OpenAI is introducing agentic commerce capabilities to ChatGPT, enabling AI agents, users, and businesses to collaborate in shopping experiences. This represents an early step toward AI-powered autonomous commerce systems integrated into conversational AI platforms.

AIBullishOpenAI News · Oct 17/105
🧠

Introducing the Realtime API

OpenAI has launched a new Realtime API that enables developers to integrate fast speech-to-speech capabilities directly into their applications. This API allows for real-time voice interactions without the traditional delays of converting speech to text and back to speech.

AIBullishOpenAI News · Sep 257/104
🧠

ChatGPT can now see, hear, and speak

ChatGPT is rolling out new multimodal capabilities that enable voice conversations and image recognition. These features represent a significant advancement in AI interface design, making interactions more intuitive and natural.

AIBullishOpenAI News · Nov 307/107
🧠

Introducing ChatGPT

OpenAI has introduced ChatGPT, a conversational AI model designed to interact through dialogue. The model can answer follow-up questions, admit mistakes, challenge incorrect premises, and reject inappropriate requests.

AINeutralarXiv – CS AI · 2d ago6/10
🧠

Discourse Diversity in Multi-Turn Empathic Dialogue

Researchers demonstrate that large language models exhibit excessive repetition of discourse tactics in multi-turn empathic conversations, reusing communication strategies at nearly double the human rate. They introduce MINT, a reinforcement learning framework that optimizes for both empathy quality and discourse move diversity, achieving 25.3% improvements in empathy while reducing repetitive tactics by 26.3%.

AINeutralarXiv – CS AI · 2d ago6/10
🧠

Understanding Generalization in Role-Playing Models via Information Theory

Researchers introduce R-EMID, an information-theoretic metric to diagnose how distribution shifts degrade role-playing model performance in real-world deployments. The framework reveals that user shifts pose the greatest generalization risk, while co-evolving reinforcement learning provides the most effective mitigation strategy.

AINeutralarXiv – CS AI · 6d ago6/10
🧠

Mixed-Initiative Context: Structuring and Managing Context for Human-AI Collaboration

Researchers propose Mixed-Initiative Context, a framework that reconceptualizes how multi-turn AI interactions are managed by treating context as an explicit, structured, and dynamically adjustable object rather than a fixed chronological sequence. The approach enables both humans and AI to actively participate in context construction, addressing current limitations where irrelevant exchanges clutter context windows and users lack direct control mechanisms.

AINeutralarXiv – CS AI · 6d ago6/10
🧠

A-MBER: Affective Memory Benchmark for Emotion Recognition

Researchers introduce A-MBER, a benchmark dataset designed to evaluate AI assistants' ability to recognize emotions based on long-term interaction history rather than immediate context. The benchmark tests whether models can retrieve relevant past interactions, infer current emotional states, and provide grounded explanations—revealing that memory's value lies in selective, context-aware interpretation rather than simple historical volume.

AIBullisharXiv – CS AI · Apr 76/10
🧠

Conversational Control with Ontologies for Large Language Models: A Lightweight Framework for Constrained Generation

Researchers developed a lightweight framework that uses ontological definitions to provide modular and explainable control over Large Language Model outputs in conversational systems. The method fine-tunes LLMs to generate content according to specific constraints like English proficiency level and content polarity, consistently outperforming pre-trained baselines across seven state-of-the-art models.

Page 1 of 3Next →