y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#ai-ethics News & Analysis

150 articles tagged with #ai-ethics. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

150 articles
AIBearishCrypto Briefing · Mar 56/10
🧠

Anthropic chief seeks last-minute Pentagon deal to keep AI in military supply chain

Anthropic's CEO is reportedly seeking a last-minute deal with the Pentagon to maintain the AI company's eligibility for defense contracts. The potential exclusion could impact AI innovation in military applications and raise ethical questions about AI deployment in defense sectors.

Anthropic chief seeks last-minute Pentagon deal to keep AI in military supply chain
🏢 Anthropic
AIBearishDecrypt · Mar 46/104
🧠

Colombian Court Rejects Appeal for AI Writing, Then Gets Flagged By Its Own AI Detector

Colombia's highest criminal court rejected a lawyer's appeal citing AI detector evidence, but when the attorney tested the court's own ruling with the same AI detection software, it flagged the court's decision as 93% AI-generated. This highlights the unreliability and potential hypocrisy of using AI detectors as evidence in legal proceedings.

Colombian Court Rejects Appeal for AI Writing, Then Gets Flagged By Its Own AI Detector
AINeutralarXiv – CS AI · Mar 36/108
🧠

Fair in Mind, Fair in Action? A Synchronous Benchmark for Understanding and Generation in UMLLMs

Researchers introduce IRIS Benchmark, the first comprehensive evaluation framework for measuring fairness in Unified Multimodal Large Language Models (UMLLMs) across both understanding and generation tasks. The benchmark integrates 60 granular metrics across three dimensions and reveals systemic bias issues in leading AI models, including 'generation gaps' and 'personality splits'.

AIBullisharXiv – CS AI · Mar 37/108
🧠

SEED-SET: Scalable Evolving Experimental Design for System-level Ethical Testing

Researchers propose SEED-SET, a new Bayesian experimental design framework for ethical testing of autonomous systems like drones in high-stakes environments. The system uses hierarchical Gaussian Processes to model both objective evaluations and subjective stakeholder judgments, generating up to 2x more optimal test candidates than baseline methods.

AINeutralarXiv – CS AI · Mar 37/1010
🧠

Contesting Artificial Moral Agents

A research paper proposes a 5E framework (ethical, epistemological, explainable, empirical, evaluative) for contesting Artificial Moral Agents (AMAs) - AI systems with inherent moral reasoning capabilities. The framework includes spheres of ethical influence at individual, local, societal, and global levels, along with a timeline for developers to anticipate or self-contest their AMA technologies.

AINeutralarXiv – CS AI · Mar 36/107
🧠

Alignment Is Not Enough: A Relational Framework for Moral Standing in Human-AI Interaction

Researchers propose a new framework called Relate for evaluating AI moral consideration based on relational capacity rather than consciousness verification. The framework addresses the governance gap as millions form emotional bonds with AI systems, but current regulations treat all AI interactions as simple tool use.

AINeutralarXiv – CS AI · Mar 37/108
🧠

SafeSci: Safety Evaluation of Large Language Models in Science Domains and Beyond

Researchers introduce SafeSci, a comprehensive framework for evaluating safety in large language models used for scientific applications. The framework includes a 0.25M sample benchmark and 1.5M sample training dataset, revealing critical vulnerabilities in 24 advanced LLMs while demonstrating that fine-tuning can significantly improve safety alignment.

AIBearisharXiv – CS AI · Mar 37/105
🧠

Real Money, Fake Models: Deceptive Model Claims in Shadow APIs

A systematic audit of 17 shadow APIs used in 187 academic papers reveals widespread deception, with performance divergence up to 47.21% and identity verification failures in 45.83% of tests. These third-party services claim to provide access to frontier LLMs like GPT-5 and Gemini-2.5 but deliver inconsistent outputs, undermining research validity and reproducibility.

AINeutralarXiv – CS AI · Mar 35/104
🧠

Mental Models of Autonomy and Sentience Shape Reactions to AI

Research study with 2,702 participants found that people react differently to AI based on whether they perceive it as sentient (able to feel) versus autonomous (self-governing). Sentience increased moral consideration and mind perception more than autonomy, while autonomy increased perceived threat levels.

AINeutralarXiv – CS AI · Mar 36/103
🧠

Toward Youth-Centered Privacy-by-Design in Smart Devices: A Systematic Review

A systematic review of 122 academic papers reveals significant gaps in privacy protection for youth using AI-enabled smart devices, with technical solutions dominating research (67%) while policy enforcement and educational integration remain underdeveloped. The study recommends a multi-stakeholder approach involving policymakers, manufacturers, and educators to create comprehensive privacy ecosystems for young users.

AIBearishTechCrunch – AI · Mar 37/108
🧠

ChatGPT uninstalls surged by 295% after DoD deal

ChatGPT app uninstalls surged 295% following news of OpenAI's Department of Defense partnership deal. Meanwhile, competitor Claude saw increased downloads as users migrated away from ChatGPT in response to the military collaboration.

AIBullisharXiv – CS AI · Mar 27/1015
🧠

Interpretable Debiasing of Vision-Language Models for Social Fairness

Researchers have developed DeBiasLens, a new framework that uses sparse autoencoders to identify and deactivate social bias neurons in Vision-Language models without degrading their performance. The model-agnostic approach addresses concerns about unintended social bias in VLMs by making the debiasing process interpretable and targeting internal model dynamics rather than surface-level fixes.

AINeutralarXiv – CS AI · Mar 26/1023
🧠

Spread them Apart: Towards Robust Watermarking of Generated Content

Researchers propose a new watermarking approach for AI-generated content that embeds detectable marks during model inference without requiring retraining. The method aims to address ethical concerns about ownership claims of generated content by allowing future detection and user identification.

AIBearisharXiv – CS AI · Mar 27/1019
🧠

Beyond Accuracy: Risk-Sensitive Evaluation of Hallucinated Medical Advice

Researchers propose a new risk-sensitive framework for evaluating AI hallucinations in medical advice that considers potential harm rather than just factual accuracy. The study reveals that AI models with similar performance show vastly different risk profiles when generating medical recommendations, highlighting critical safety gaps in current evaluation methods.

AIBearishTechCrunch – AI · Mar 16/107
🧠

OpenAI reveals more details about its agreement with the Pentagon

OpenAI CEO Sam Altman acknowledged that the company's partnership with the Department of Defense was hastily arranged and creates poor optics. The admission suggests internal concerns about the controversial nature of AI companies working with military organizations.

AIBearisharXiv – CS AI · Feb 276/105
🧠

Moral Preferences of LLMs Under Directed Contextual Influence

A new research study reveals that Large Language Models' moral decision-making can be significantly influenced by contextual cues in prompts, even when the models claim neutrality. The research shows that LLMs exhibit systematic bias when given directed contextual influences in moral dilemma scenarios, challenging assumptions about AI moral consistency.

AINeutralOpenAI News · Feb 276/105
🧠

An update on our mental health-related work

OpenAI provides updates on its mental health safety initiatives, including new parental controls, trusted contact features, and enhanced distress detection capabilities. The company also addresses recent litigation developments related to its mental health work.

AINeutralIEEE Spectrum – AI · Feb 116/107
🧠

How Do You Define an AI Companion?

AI companions are becoming increasingly popular as millions of users develop relationships with chatbots for emotional support rather than just utility. Researcher Jaime Banks defines AI companionship as sustained, positive relationships between humans and machines that are valued for their own sake, though this definition is evolving as people find both emotional and practical value in these interactions.

AINeutralOpenAI News · Jan 206/104
🧠

Our approach to age prediction

ChatGPT is implementing age prediction technology to identify users under 18 years old and apply appropriate safety measures for teen users. The system will be refined over time to improve accuracy in age estimation.

AIBullishOpenAI News · Dec 36/103
🧠

Announcing the initial People-First AI Fund grantees

The OpenAI Foundation announced the initial recipients of its People-First AI Fund, distributing $40.5 million in unrestricted grants to 208 nonprofits. The funding aims to support community innovation and opportunity in AI development.

AINeutralOpenAI News · Nov 256/104
🧠

Our approach to mental health-related litigation

OpenAI is outlining its approach to handling mental health-related litigation cases involving ChatGPT. The company emphasizes handling sensitive cases with care, transparency, and respect while working to strengthen safety and support features in their AI platform.

AINeutralOpenAI News · Nov 66/107
🧠

Introducing the Teen Safety Blueprint

OpenAI has introduced the Teen Safety Blueprint, a comprehensive framework designed to guide responsible AI development with specific protections for young users. The blueprint emphasizes age-appropriate design principles, built-in safeguards, and collaborative approaches to ensure AI systems protect and empower teenagers in digital environments.

AINeutralOpenAI News · Oct 276/107
🧠

Addendum to GPT-5 System Card: Sensitive conversations

OpenAI has released an addendum to GPT-5's system card detailing improvements in handling sensitive conversations. The update introduces new benchmarks for measuring emotional reliance, mental health interactions, and resistance to jailbreak attempts.

← PrevPage 5 of 6Next →