26 articles tagged with #safety. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBearishTechCrunch – AI · 6d ago7/10
🧠A stalking victim is suing OpenAI, alleging that ChatGPT ignored three separate warnings—including the company's own mass casualty flag—while her abuser used the platform to fuel his obsessive behavior. The lawsuit raises critical questions about AI companies' liability when warned of dangerous user behavior.
🏢 OpenAI🧠 ChatGPT
AIBullisharXiv – CS AI · Mar 177/10
🧠Researchers introduce EcoAlign, a new framework for aligning Large Vision-Language Models that treats alignment as an economic optimization problem. The method balances safety, utility, and computational costs while preventing harmful reasoning disguised with benign justifications, showing superior performance across multiple models and datasets.
AINeutralarXiv – CS AI · Mar 56/10
🧠Researchers propose new metrics to measure the automation of AI R&D (AIRDA), arguing that existing capability benchmarks don't capture real-world automation effects or broader consequences. The proposed metrics would track dimensions like capital allocation, researcher time, and AI oversight incidents to help decision-makers understand AIRDA's impact on AI progress and safety.
AIBullisharXiv – CS AI · Mar 37/103
🧠Researchers developed OS-Det3D, a two-stage framework for camera-based 3D object detection in autonomous vehicles that can identify unknown objects beyond predefined categories. The system uses LiDAR geometric cues and a joint selection module to discover novel objects while improving detection of known objects, addressing safety risks in real-world driving scenarios.
AIBullisharXiv – CS AI · Feb 277/104
🧠Researchers developed AviaSafe, a physics-informed AI model that forecasts aviation-critical cloud species up to 7 days ahead, addressing safety concerns around engine icing. The model outperforms operational weather models by predicting specific hydrometeor species rather than general atmospheric variables, enabling better aviation route optimization.
CryptoBearishDL News · Feb 137/104
⛓️A Binance France team member was targeted in a failed 'wrench attack' at their home, which Binance has confirmed following local media reports. This type of attack involves criminals attempting to physically coerce cryptocurrency executives or holders to transfer digital assets.
AINeutralIEEE Spectrum – AI · Feb 27/108
🧠The article argues for regulating AI applications and use cases rather than the underlying AI models themselves. The author contends that model-centric regulation fails because digital artifacts can't be controlled once released, while use-based regulation can effectively address real-world harms by scaling obligations according to deployment risk levels.
$NEAR
AIBullishOpenAI News · Jul 177/104
🧠OpenAI has released a System Card for ChatGPT's new agentic model, which integrates research capabilities, browser automation, and code execution tools. The system operates under OpenAI's Preparedness Framework with built-in safeguards to manage potential risks from autonomous AI agents.
AIBullishOpenAI News · Mar 237/107
🧠OpenAI has implemented initial support for plugins in ChatGPT, which are tools specifically designed for language models with safety as a core principle. These plugins enable ChatGPT to access current information, perform computations, and integrate with third-party services.
AIBearishOpenAI News · Jul 177/106
🧠Researchers have developed adversarial images that can consistently fool neural network classifiers across multiple scales and viewing perspectives. This breakthrough challenges previous assumptions that self-driving cars would be secure from malicious attacks due to their multi-angle image capture capabilities.
AIBullisharXiv – CS AI · Apr 76/10
🧠Researchers introduce VLA-Forget, a new unlearning framework for vision-language-action (VLA) models used in robotic manipulation. The hybrid approach addresses the challenge of removing unsafe or unwanted behaviors from embodied AI foundation models while preserving their core perception, language, and action capabilities.
AIBullisharXiv – CS AI · Mar 276/10
🧠Researchers have developed the first formal mathematical framework for verifying AI agent protocols, specifically comparing Schema-Guided Dialogue (SGD) and Model Context Protocol (MCP). They proved these systems are structurally similar but identified critical gaps in MCP's capabilities, proposing MCP+ extensions to achieve full equivalence with SGD.
AIBullisharXiv – CS AI · Mar 27/1016
🧠Researchers propose SafeGen-LLM, a new approach to enhance safety in robotic task planning by combining supervised fine-tuning with policy optimization guided by formal verification. The system demonstrates superior safety generalization across multiple domains compared to existing classical planners, reinforcement learning methods, and base large language models.
AIBullisharXiv – CS AI · Feb 276/105
🧠Researchers developed Risk-aware World Model Predictive Control (RaWMPC), a new framework for autonomous driving that makes safe decisions without relying on expert demonstrations. The system uses a world model to predict consequences of multiple actions and selects low-risk options through explicit risk evaluation, showing superior performance in both normal and rare driving scenarios.
AINeutralOpenAI News · Dec 116/105
🧠OpenAI has released GPT-5.2, the latest model in the GPT-5 series, maintaining the same comprehensive safety mitigation approach as previous versions. The model was trained on diverse datasets including publicly available internet information, third-party partnerships, and user-generated content.
CryptoNeutralEthereum Foundation Blog · Nov 35/102
⛓️y0.exchange has issued a second update regarding safety preparations for Devconnect events, following previous travel advisories. The team is actively working with local security providers, law enforcement, and risk advisory partners to monitor and address potential security concerns.
CryptoBearishEthereum Foundation Blog · Oct 236/103
⛓️Event organizers are issuing a travel advisory for Devconnect Istanbul due to security concerns related to ongoing events in Israel and Gaza. The advisory reflects heightened risk assessment procedures for attendees considering travel to the cryptocurrency/blockchain conference.
AIBullishOpenAI News · Nov 186/105
🧠OpenAI has removed the waitlist requirement for accessing its API, making it widely available to developers and businesses. The broader access is enabled by improvements in safety measures and protocols.
AINeutralarXiv – CS AI · Mar 174/10
🧠Researchers introduce IL-CIRL, a framework combining Iterative Learning Control with Deep Reinforcement Learning to address safety risks and stability issues in industrial batch process control. The method uses Kalman filter-based state estimation to guide DRL agents toward safer, constraint-satisfying control policies.
AIBullishTechCrunch – AI · Mar 65/10
🧠City Detect, an AI-powered company that helps local governments prevent urban decay and maintain city safety and cleanliness, has raised $13 million in Series A funding. The company is currently operating in at least 17 cities, including major markets like Dallas and Miami.
AINeutralarXiv – CS AI · Mar 54/10
🧠A research paper analyzes reward functions used in reinforcement learning for autonomous driving, identifying gaps in current approaches. The study categorizes objectives into Safety, Comfort, Progress, and Traffic Rules compliance, highlighting limitations in objective aggregation and context awareness.
AIBullishOpenAI News · Dec 184/104
🧠OpenAI has released new AI literacy resources designed to help teenagers and parents use ChatGPT more responsibly and safely. The educational materials include expert-reviewed guidance on critical thinking, establishing healthy boundaries, and navigating sensitive conversations with AI tools.
AINeutralarXiv – CS AI · Mar 34/105
🧠Researchers propose AURA, an AIoT framework that uses in-vehicle sensors and AI to continuously monitor driving safety in older adults. The system analyzes real-world driving patterns while preserving privacy through edge computing architecture.
AINeutralGoogle Research Blog · Jan 132/107
🧠This article appears to discuss research on using hard-braking events as predictive indicators for crash risk assessment on road segments. The focus is on algorithmic approaches and theoretical frameworks for traffic safety analysis.
GeneralNeutralOpenAI News · Sep 161/106
📰This appears to be an update on safety and security practices, but the article body is missing or not provided. Without the actual content, it's impossible to analyze the specific security measures, incidents, or improvements being discussed.