150 articles tagged with #ai-ethics. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBearishCrypto Briefing · Mar 56/10
🧠Anthropic's CEO is reportedly seeking a last-minute deal with the Pentagon to maintain the AI company's eligibility for defense contracts. The potential exclusion could impact AI innovation in military applications and raise ethical questions about AI deployment in defense sectors.
🏢 Anthropic
AIBearishDecrypt · Mar 46/104
🧠Colombia's highest criminal court rejected a lawyer's appeal citing AI detector evidence, but when the attorney tested the court's own ruling with the same AI detection software, it flagged the court's decision as 93% AI-generated. This highlights the unreliability and potential hypocrisy of using AI detectors as evidence in legal proceedings.
AINeutralarXiv – CS AI · Mar 36/108
🧠Researchers introduce IRIS Benchmark, the first comprehensive evaluation framework for measuring fairness in Unified Multimodal Large Language Models (UMLLMs) across both understanding and generation tasks. The benchmark integrates 60 granular metrics across three dimensions and reveals systemic bias issues in leading AI models, including 'generation gaps' and 'personality splits'.
AIBullisharXiv – CS AI · Mar 37/108
🧠Researchers propose SEED-SET, a new Bayesian experimental design framework for ethical testing of autonomous systems like drones in high-stakes environments. The system uses hierarchical Gaussian Processes to model both objective evaluations and subjective stakeholder judgments, generating up to 2x more optimal test candidates than baseline methods.
AINeutralarXiv – CS AI · Mar 37/1010
🧠A research paper proposes a 5E framework (ethical, epistemological, explainable, empirical, evaluative) for contesting Artificial Moral Agents (AMAs) - AI systems with inherent moral reasoning capabilities. The framework includes spheres of ethical influence at individual, local, societal, and global levels, along with a timeline for developers to anticipate or self-contest their AMA technologies.
AINeutralarXiv – CS AI · Mar 36/107
🧠Researchers propose a new framework called Relate for evaluating AI moral consideration based on relational capacity rather than consciousness verification. The framework addresses the governance gap as millions form emotional bonds with AI systems, but current regulations treat all AI interactions as simple tool use.
AINeutralarXiv – CS AI · Mar 37/108
🧠Researchers introduce SafeSci, a comprehensive framework for evaluating safety in large language models used for scientific applications. The framework includes a 0.25M sample benchmark and 1.5M sample training dataset, revealing critical vulnerabilities in 24 advanced LLMs while demonstrating that fine-tuning can significantly improve safety alignment.
AIBearisharXiv – CS AI · Mar 37/105
🧠A systematic audit of 17 shadow APIs used in 187 academic papers reveals widespread deception, with performance divergence up to 47.21% and identity verification failures in 45.83% of tests. These third-party services claim to provide access to frontier LLMs like GPT-5 and Gemini-2.5 but deliver inconsistent outputs, undermining research validity and reproducibility.
AINeutralarXiv – CS AI · Mar 35/104
🧠Research study with 2,702 participants found that people react differently to AI based on whether they perceive it as sentient (able to feel) versus autonomous (self-governing). Sentience increased moral consideration and mind perception more than autonomy, while autonomy increased perceived threat levels.
AINeutralarXiv – CS AI · Mar 36/103
🧠A systematic review of 122 academic papers reveals significant gaps in privacy protection for youth using AI-enabled smart devices, with technical solutions dominating research (67%) while policy enforcement and educational integration remain underdeveloped. The study recommends a multi-stakeholder approach involving policymakers, manufacturers, and educators to create comprehensive privacy ecosystems for young users.
AIBearishTechCrunch – AI · Mar 37/108
🧠ChatGPT app uninstalls surged 295% following news of OpenAI's Department of Defense partnership deal. Meanwhile, competitor Claude saw increased downloads as users migrated away from ChatGPT in response to the military collaboration.
AIBullisharXiv – CS AI · Mar 27/1015
🧠Researchers have developed DeBiasLens, a new framework that uses sparse autoencoders to identify and deactivate social bias neurons in Vision-Language models without degrading their performance. The model-agnostic approach addresses concerns about unintended social bias in VLMs by making the debiasing process interpretable and targeting internal model dynamics rather than surface-level fixes.
AINeutralarXiv – CS AI · Mar 26/1023
🧠Researchers propose a new watermarking approach for AI-generated content that embeds detectable marks during model inference without requiring retraining. The method aims to address ethical concerns about ownership claims of generated content by allowing future detection and user identification.
AINeutralarXiv – CS AI · Mar 27/1018
🧠Researchers analyzed how large language models express moral judgments when prompted to role-play different personas. The study found that Claude models are most morally robust, while larger models within families tend to be more susceptible to moral shifts through persona conditioning.
AIBearisharXiv – CS AI · Mar 27/1019
🧠Researchers propose a new risk-sensitive framework for evaluating AI hallucinations in medical advice that considers potential harm rather than just factual accuracy. The study reveals that AI models with similar performance show vastly different risk profiles when generating medical recommendations, highlighting critical safety gaps in current evaluation methods.
AIBearishTechCrunch – AI · Mar 16/107
🧠OpenAI CEO Sam Altman acknowledged that the company's partnership with the Department of Defense was hastily arranged and creates poor optics. The admission suggests internal concerns about the controversial nature of AI companies working with military organizations.
AIBearisharXiv – CS AI · Feb 276/105
🧠A new research study reveals that Large Language Models' moral decision-making can be significantly influenced by contextual cues in prompts, even when the models claim neutrality. The research shows that LLMs exhibit systematic bias when given directed contextual influences in moral dilemma scenarios, challenging assumptions about AI moral consistency.
AINeutralOpenAI News · Feb 276/105
🧠OpenAI provides updates on its mental health safety initiatives, including new parental controls, trusted contact features, and enhanced distress detection capabilities. The company also addresses recent litigation developments related to its mental health work.
AINeutralWired – AI · Feb 266/105
🧠The article discusses ongoing tensions between AI company Anthropic and the Pentagon, exploring themes of 'woke' AI versus defense applications. It also covers developments in undersea cable infrastructure (TAT-8) and political dynamics between Trump and the State of the Union.
AINeutralIEEE Spectrum – AI · Feb 116/107
🧠AI companions are becoming increasingly popular as millions of users develop relationships with chatbots for emotional support rather than just utility. Researcher Jaime Banks defines AI companionship as sustained, positive relationships between humans and machines that are valued for their own sake, though this definition is evolving as people find both emotional and practical value in these interactions.
AINeutralOpenAI News · Jan 206/104
🧠ChatGPT is implementing age prediction technology to identify users under 18 years old and apply appropriate safety measures for teen users. The system will be refined over time to improve accuracy in age estimation.
AIBullishOpenAI News · Dec 36/103
🧠The OpenAI Foundation announced the initial recipients of its People-First AI Fund, distributing $40.5 million in unrestricted grants to 208 nonprofits. The funding aims to support community innovation and opportunity in AI development.
AINeutralOpenAI News · Nov 256/104
🧠OpenAI is outlining its approach to handling mental health-related litigation cases involving ChatGPT. The company emphasizes handling sensitive cases with care, transparency, and respect while working to strengthen safety and support features in their AI platform.
AINeutralOpenAI News · Nov 66/107
🧠OpenAI has introduced the Teen Safety Blueprint, a comprehensive framework designed to guide responsible AI development with specific protections for young users. The blueprint emphasizes age-appropriate design principles, built-in safeguards, and collaborative approaches to ensure AI systems protect and empower teenagers in digital environments.
AINeutralOpenAI News · Oct 276/107
🧠OpenAI has released an addendum to GPT-5's system card detailing improvements in handling sensitive conversations. The update introduces new benchmarks for measuring emotional reliance, mental health interactions, and resistance to jailbreak attempts.