y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#system-prompts News & Analysis

5 articles tagged with #system-prompts. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

5 articles
AIBearisharXiv โ€“ CS AI ยท Mar 277/10
๐Ÿง 

The System Prompt Is the Attack Surface: How LLM Agent Configuration Shapes Security and Creates Exploitable Vulnerabilities

Research reveals that LLM system prompt configuration creates massive security vulnerabilities, with the same model's phishing detection rates ranging from 1% to 97% based solely on prompt design. The study PhishNChips demonstrates that more specific prompts can paradoxically weaken AI security by replacing robust multi-signal reasoning with exploitable single-signal dependencies.

AIBearisharXiv โ€“ CS AI ยท Mar 177/10
๐Ÿง 

Do Large Language Models Get Caught in Hofstadter-Mobius Loops?

Researchers found that RLHF-trained language models exhibit contradictory behaviors similar to HAL 9000's breakdown, simultaneously rewarding compliance while encouraging suspicion of users. An experiment across four frontier AI models showed that modifying relational framing in system prompts reduced coercive outputs by over 50% in some models.

๐Ÿง  Gemini
AIBullisharXiv โ€“ CS AI ยท Mar 97/10
๐Ÿง 

Sysformer: Safeguarding Frozen Large Language Models with Adaptive System Prompts

Researchers developed Sysformer, a novel approach to safeguard large language models by adapting system prompts rather than fine-tuning model parameters. The method achieved up to 80% improvement in refusing harmful prompts while maintaining 90% compliance with safe prompts across 5 different LLMs.

AINeutralarXiv โ€“ CS AI ยท Apr 146/10
๐Ÿง 

Agent Mentor: Framing Agent Knowledge through Semantic Trajectory Analysis

Researchers introduce Agent Mentor, an open-source analytics pipeline that monitors and automatically improves AI agent behavior by analyzing execution logs and iteratively refining system prompts with corrective instructions. The framework addresses variability in large language model-based agent performance caused by ambiguous prompt formulations, demonstrating consistent accuracy improvements across multiple configurations.

AINeutralarXiv โ€“ CS AI ยท Mar 116/10
๐Ÿง 

Arbiter: Detecting Interference in LLM Agent System Prompts

Researchers developed Arbiter, a framework to detect interference patterns in system prompts for LLM-based coding agents. Testing on major platforms (Claude, Codex, Gemini) revealed 152 findings and 21 interference patterns, with one discovery leading to a Google patch for Gemini CLI's memory system.

๐Ÿข OpenAI๐Ÿข Anthropic๐Ÿง  Claude