7 articles tagged with #certification. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBullisharXiv โ CS AI ยท Apr 77/10
๐ง Researchers propose a new constrained maximum likelihood estimation (MLE) method to accurately estimate failure rates of large language models by combining human-labeled data, automated judge annotations, and domain-specific constraints. The approach outperforms existing methods like Prediction-Powered Inference across various experimental conditions, providing a more reliable framework for LLM safety certification.
AINeutralarXiv โ CS AI ยท Mar 46/105
๐ง Researchers propose Human-Certified Module Repositories (HCMRs) as a new framework to ensure trustworthy software development in the AI era. The system combines human oversight with automated analysis to certify and curate reusable code modules, addressing growing security concerns as AI increasingly generates and assembles software components.
AIBullisharXiv โ CS AI ยท Mar 47/103
๐ง Researchers propose a dual Randomized Smoothing framework that overcomes limitations of standard neural network robustness certification by using input-dependent noise variances instead of global ones. The method achieves strong performance at both small and large radii with gains of 15-20% on CIFAR-10 and 8-17% on ImageNet, while adding only 60% computational overhead.
AIBullisharXiv โ CS AI ยท Feb 277/105
๐ง Researchers introduce Certified Circuits, a framework that provides provable stability guarantees for neural network circuit discovery. The method wraps existing algorithms with randomized data subsampling to ensure circuit components remain consistent across dataset variations, achieving 91% higher accuracy while using 45% fewer neurons.
AIBullisharXiv โ CS AI ยท Apr 76/10
๐ง Researchers propose a compliance-by-construction architecture that integrates Generative AI with structured formal argument representations to ensure accountability in high-stakes decision systems. The approach uses typed Argument Graphs, retrieval-augmented generation, validation constraints, and provenance ledgers to prevent AI hallucinations while maintaining traceability for regulatory compliance.
AIBullisharXiv โ CS AI ยท Mar 176/10
๐ง Researchers developed SimCert, a probabilistic certification framework that verifies behavioral similarity between compressed neural networks and their original versions. The framework addresses critical safety challenges in deploying compressed DNNs on resource-constrained systems by providing quantitative safety guarantees with adjustable confidence levels.
AIBullishOpenAI News ยท Dec 96/106
๐ง OpenAI has launched its first certification programs and AI Foundations courses designed to help individuals develop practical AI skills. These educational offerings aim to enhance career prospects and prepare workers for an AI-driven future workplace.