y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#human-ai-alignment News & Analysis

6 articles tagged with #human-ai-alignment. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

6 articles
AINeutralarXiv โ€“ CS AI ยท Mar 56/10
๐Ÿง 

Automated Concept Discovery for LLM-as-a-Judge Preference Analysis

Researchers developed automated methods to discover biases in Large Language Models when used as judges, analyzing over 27,000 paired responses. The study found LLMs exhibit systematic biases including preference for refusing sensitive requests more than humans, favoring concrete and empathetic responses, and showing bias against certain legal guidance.

AIBullisharXiv โ€“ CS AI ยท Feb 276/107
๐Ÿง 

Is This Just Fantasy? Language Model Representations Reflect Human Judgments of Event Plausibility

Researchers have identified 'modal difference vectors' in language models that can distinguish between possible, impossible, and nonsensical statements, revealing better modal categorization abilities than previously thought. The study shows these vectors emerge consistently as models become more capable and can even predict human judgment patterns about event plausibility.

AINeutralarXiv โ€“ CS AI ยท Apr 64/10
๐Ÿง 

Social Meaning in Large Language Models: Structure, Magnitude, and Pragmatic Prompting

Research reveals that large language models can reproduce the qualitative structure of human social reasoning but struggle with quantitative magnitude calibration. Pragmatic prompting strategies that consider speaker knowledge and motives can improve this calibration, though fine-grained accuracy remains partially unresolved.

AIBullisharXiv โ€“ CS AI ยท Mar 175/10
๐Ÿง 

Human-like Object Grouping in Self-supervised Vision Transformers

Researchers developed a behavioral benchmark showing that self-supervised vision transformers, particularly those trained with DINO objectives, align closely with human object perception and segmentation behavior. The study found that models with stronger object-centric representations better predict human visual judgments, with Gram matrix structure playing a key role in perceptual alignment.

AINeutralarXiv โ€“ CS AI ยท Mar 54/10
๐Ÿง 

MuSaG: A Multimodal German Sarcasm Dataset with Full-Modal Annotations

Researchers have released MuSaG, the first German multimodal sarcasm detection dataset featuring 33 minutes of annotated television content with text, audio, and video data. The study reveals a significant gap between human sarcasm detection (which relies heavily on audio cues) and current AI models (which perform best with text).

AINeutralarXiv โ€“ CS AI ยท Mar 34/105
๐Ÿง 

Emerging Human-like Strategies for Semantic Memory Foraging in Large Language Models

Researchers analyzed how Large Language Models access semantic memory using the Semantic Fluency Task, finding that LLMs exhibit similar memory foraging patterns to humans. The study reveals convergent and divergent search strategies in LLMs that mirror human cognitive behavior, potentially enabling better human-AI alignment or productive cognitive disalignment.