y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#ai-standards News & Analysis

6 articles tagged with #ai-standards. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

6 articles
AINeutralarXiv – CS AI · Apr 77/10
🧠

Is your AI Model Accurate Enough? The Difficult Choices Behind Rigorous AI Development and the EU AI Act

A research paper challenges the common view of AI accuracy as purely technical, arguing it involves context-dependent normative decisions that determine error priorities and risk distribution. The study analyzes the EU AI Act's "appropriate accuracy" requirements and identifies four critical choices in performance evaluation that embed assumptions about acceptable trade-offs.

AIBullishOpenAI News · Dec 97/106
🧠

OpenAI co-founds Agentic AI Foundation, donates AGENTS.md

OpenAI co-founded the Agentic AI Foundation under the Linux Foundation and donated AGENTS.md to promote open, interoperable standards for safe agentic AI development. This initiative aims to establish industry-wide standards for AI agent safety and interoperability.

AINeutralOpenAI News · Aug 127/106
🧠

OpenAI’s letter to Governor Newsom on harmonized regulation

OpenAI has sent a letter to California Governor Gavin Newsom advocating for harmonized AI regulation between state and national levels. The company is pushing for California to lead in creating AI regulatory standards that align with emerging US and global frameworks.

AINeutralOpenAI News · Feb 27/106
🧠

Response to NIST Executive Order on AI

NIST has issued a request for information regarding its assignments under sections 4.1, 4.5, and 11 of the Executive Order on Artificial Intelligence. This represents a formal step in implementing federal AI regulatory framework and standards development.

AINeutralarXiv – CS AI · Apr 146/10
🧠

ATANT v1.1: Positioning Continuity Evaluation Against Memory, Long-Context, and Agentic-Memory Benchmarks

ATANT v1.1 is a companion paper clarifying how existing memory and context evaluation benchmarks (LOCOMO, LongMemEval, BEAM, MemoryBench, and others) fail to measure 'continuity' as defined in the original v1.0 framework. The analysis reveals that existing benchmarks cover a median of only 1 out of 7 required continuity properties, and the authors demonstrate a significant measurement gap through comparative scoring: their system achieves 96% on ATANT but only 8.8% on LOCOMO, proving these benchmarks evaluate different capabilities.

AINeutralOpenAI News · Mar 256/10
🧠

Inside our approach to the Model Spec

OpenAI has released its Model Spec, a public framework that outlines how AI models should behave by balancing safety considerations, user freedom, and accountability. The specification serves as a governance tool for managing AI system behavior as these technologies continue to advance.

🏢 OpenAI