y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#safety News & Analysis

26 articles tagged with #safety. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

26 articles
AIBullisharXiv – CS AI · Mar 177/10
🧠

EcoAlign: An Economically Rational Framework for Efficient LVLM Alignment

Researchers introduce EcoAlign, a new framework for aligning Large Vision-Language Models that treats alignment as an economic optimization problem. The method balances safety, utility, and computational costs while preventing harmful reasoning disguised with benign justifications, showing superior performance across multiple models and datasets.

AINeutralarXiv – CS AI · Mar 56/10
🧠

Measuring AI R&D Automation

Researchers propose new metrics to measure the automation of AI R&D (AIRDA), arguing that existing capability benchmarks don't capture real-world automation effects or broader consequences. The proposed metrics would track dimensions like capital allocation, researcher time, and AI oversight incidents to help decision-makers understand AIRDA's impact on AI progress and safety.

AIBullisharXiv – CS AI · Mar 37/103
🧠

Towards Camera Open-set 3D Object Detection for Autonomous Driving Scenarios

Researchers developed OS-Det3D, a two-stage framework for camera-based 3D object detection in autonomous vehicles that can identify unknown objects beyond predefined categories. The system uses LiDAR geometric cues and a joint selection module to discover novel objects while improving detection of known objects, addressing safety risks in real-world driving scenarios.

AIBullisharXiv – CS AI · Feb 277/104
🧠

AviaSafe: A Physics-Informed Data-Driven Model for Aviation Safety-Critical Cloud Forecasts

Researchers developed AviaSafe, a physics-informed AI model that forecasts aviation-critical cloud species up to 7 days ahead, addressing safety concerns around engine icing. The model outperforms operational weather models by predicting specific hydrometeor species rather than general atmospheric variables, enabling better aviation route optimization.

CryptoBearishDL News · Feb 137/104
⛓️

Binance France boss targeted in failed home invasion wrench attack

A Binance France team member was targeted in a failed 'wrench attack' at their home, which Binance has confirmed following local media reports. This type of attack involves criminals attempting to physically coerce cryptocurrency executives or holders to transfer digital assets.

AINeutralIEEE Spectrum – AI · Feb 27/108
🧠

Don’t Regulate AI Models. Regulate AI Use

The article argues for regulating AI applications and use cases rather than the underlying AI models themselves. The author contends that model-centric regulation fails because digital artifacts can't be controlled once released, while use-based regulation can effectively address real-world harms by scaling obligations according to deployment risk levels.

$NEAR
AIBullishOpenAI News · Jul 177/104
🧠

ChatGPT agent System Card

OpenAI has released a System Card for ChatGPT's new agentic model, which integrates research capabilities, browser automation, and code execution tools. The system operates under OpenAI's Preparedness Framework with built-in safeguards to manage potential risks from autonomous AI agents.

AIBullishOpenAI News · Mar 237/107
🧠

ChatGPT plugins

OpenAI has implemented initial support for plugins in ChatGPT, which are tools specifically designed for language models with safety as a core principle. These plugins enable ChatGPT to access current information, perform computations, and integrate with third-party services.

AIBearishOpenAI News · Jul 177/106
🧠

Robust adversarial inputs

Researchers have developed adversarial images that can consistently fool neural network classifiers across multiple scales and viewing perspectives. This breakthrough challenges previous assumptions that self-driving cars would be secure from malicious attacks due to their multi-angle image capture capabilities.

AIBullisharXiv – CS AI · Apr 76/10
🧠

VLA-Forget: Vision-Language-Action Unlearning for Embodied Foundation Models

Researchers introduce VLA-Forget, a new unlearning framework for vision-language-action (VLA) models used in robotic manipulation. The hybrid approach addresses the challenge of removing unsafe or unwanted behaviors from embodied AI foundation models while preserving their core perception, language, and action capabilities.

AIBullisharXiv – CS AI · Mar 276/10
🧠

Formal Semantics for Agentic Tool Protocols: A Process Calculus Approach

Researchers have developed the first formal mathematical framework for verifying AI agent protocols, specifically comparing Schema-Guided Dialogue (SGD) and Model Context Protocol (MCP). They proved these systems are structurally similar but identified critical gaps in MCP's capabilities, proposing MCP+ extensions to achieve full equivalence with SGD.

AIBullisharXiv – CS AI · Mar 27/1016
🧠

SafeGen-LLM: Enhancing Safety Generalization in Task Planning for Robotic Systems

Researchers propose SafeGen-LLM, a new approach to enhance safety in robotic task planning by combining supervised fine-tuning with policy optimization guided by formal verification. The system demonstrates superior safety generalization across multiple domains compared to existing classical planners, reinforcement learning methods, and base large language models.

AIBullisharXiv – CS AI · Feb 276/105
🧠

Risk-Aware World Model Predictive Control for Generalizable End-to-End Autonomous Driving

Researchers developed Risk-aware World Model Predictive Control (RaWMPC), a new framework for autonomous driving that makes safe decisions without relying on expert demonstrations. The system uses a world model to predict consequences of multiple actions and selects low-risk options through explicit risk evaluation, showing superior performance in both normal and rare driving scenarios.

AINeutralOpenAI News · Dec 116/105
🧠

Update to GPT-5 System Card: GPT-5.2

OpenAI has released GPT-5.2, the latest model in the GPT-5 series, maintaining the same comprehensive safety mitigation approach as previous versions. The model was trained on diverse datasets including publicly available internet information, third-party partnerships, and user-generated content.

CryptoNeutralEthereum Foundation Blog · Nov 35/102
⛓️

Update 2 - Preparing for Devconnect Events

y0.exchange has issued a second update regarding safety preparations for Devconnect events, following previous travel advisories. The team is actively working with local security providers, law enforcement, and risk advisory partners to monitor and address potential security concerns.

Update 2 - Preparing for Devconnect Events
CryptoBearishEthereum Foundation Blog · Oct 236/103
⛓️

Update - Advisory on recent events and potential travel considerations

Event organizers are issuing a travel advisory for Devconnect Istanbul due to security concerns related to ongoing events in Israel and Gaza. The advisory reflects heightened risk assessment procedures for attendees considering travel to the cryptocurrency/blockchain conference.

Update - Advisory on recent events and potential travel considerations
AIBullishOpenAI News · Nov 186/105
🧠

OpenAI’s API now available with no waitlist

OpenAI has removed the waitlist requirement for accessing its API, making it widely available to developers and businesses. The broader access is enabled by improvements in safety measures and protocols.

AINeutralarXiv – CS AI · Mar 174/10
🧠

Iterative Learning Control-Informed Reinforcement Learning for Batch Process Control

Researchers introduce IL-CIRL, a framework combining Iterative Learning Control with Deep Reinforcement Learning to address safety risks and stability issues in industrial batch process control. The method uses Kalman filter-based state estimation to guide DRL agents toward safer, constraint-satisfying control policies.

AIBullishOpenAI News · Dec 184/104
🧠

AI literacy resources for teens and parents

OpenAI has released new AI literacy resources designed to help teenagers and parents use ChatGPT more responsibly and safely. The educational materials include expert-reviewed guidance on critical thinking, establishing healthy boundaries, and navigating sensitive conversations with AI tools.

AINeutralGoogle Research Blog · Jan 132/107
🧠

Hard-braking events as indicators of road segment crash risk

This article appears to discuss research on using hard-braking events as predictive indicators for crash risk assessment on road segments. The focus is on algorithmic approaches and theoretical frameworks for traffic safety analysis.

GeneralNeutralOpenAI News · Sep 161/106
📰

An update on our safety & security practices

This appears to be an update on safety and security practices, but the article body is missing or not provided. Without the actual content, it's impossible to analyze the specific security measures, incidents, or improvements being discussed.

Page 1 of 2Next →