y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#ai-security News & Analysis

216 articles tagged with #ai-security. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

216 articles
AINeutralOpenAI News · Feb 147/106
🧠

Disrupting malicious uses of AI by state-affiliated threat actors

AI company terminated accounts linked to state-affiliated threat actors attempting to use AI models for malicious cybersecurity purposes. Investigation revealed that the AI models provided only limited incremental capabilities for such malicious activities.

AIBullishOpenAI News · Jul 217/105
🧠

Moving AI governance forward

OpenAI and other leading AI laboratories are strengthening AI governance through voluntary commitments focused on safety, security, and trustworthiness. This represents a proactive industry approach to self-regulation in AI development.

AIBearishOpenAI News · Jul 177/106
🧠

Robust adversarial inputs

Researchers have developed adversarial images that can consistently fool neural network classifiers across multiple scales and viewing perspectives. This breakthrough challenges previous assumptions that self-driving cars would be secure from malicious attacks due to their multi-angle image capture capabilities.

AIBullishFortune Crypto · 1d ago6/10
🧠

Exclusive: Artemis raises $70M to help fight AI-powered attacks with AI

Artemis has secured $70 million in funding to develop AI-powered defense systems against increasingly sophisticated AI-driven cyberattacks. The funding reflects growing market demand for advanced security solutions as AI-enabled threats become faster and more cost-effective to deploy.

Exclusive: Artemis raises $70M to help fight AI-powered attacks with AI
AINeutralarXiv – CS AI · 1d ago6/10
🧠

SeedPrints: Fingerprints Can Even Tell Which Seed Your Large Language Model Was Trained From

Researchers have developed SeedPrints, a novel fingerprinting method that identifies Large Language Models based on their random initialization seed rather than post-training characteristics. This approach enables model attribution and provenance verification from inception through full pretraining, addressing limitations of existing methods that only work reliably after fine-tuning.

AI × CryptoNeutralCrypto Briefing · 2d ago6/10
🤖

Coinbase, Binance seek Anthropic Mythos access as crypto firms brace for AI security threats

Major cryptocurrency exchanges Coinbase and Binance are seeking access to Anthropic's Mythos AI model as the crypto industry increasingly recognizes AI as both a critical security asset and potential existential threat. This development underscores how crypto firms are strategically positioning themselves to leverage advanced AI capabilities for defending against emerging security vulnerabilities.

Coinbase, Binance seek Anthropic Mythos access as crypto firms brace for AI security threats
🏢 Anthropic
AINeutralarXiv – CS AI · 2d ago6/10
🧠

Machine Learning-Based Detection of MCP Attacks

Researchers developed machine learning models to detect malicious Model Context Protocol (MCP) attacks, achieving up to 100% F1-score on binary classification and 90.56% on multiclass detection tasks. The study addresses a critical security gap in MCP technology, which extends LLM capabilities but introduces new attack surfaces, and includes a middleware solution for real-world deployment.

AIBullisharXiv – CS AI · 2d ago6/10
🧠

QShield: Securing Neural Networks Against Adversarial Attacks using Quantum Circuits

Researchers introduce QShield, a hybrid quantum-classical neural network architecture that combines traditional CNNs with quantum processing modules to defend deep learning models against adversarial attacks. Testing on MNIST, OrganAMNIST, and CIFAR-10 datasets shows the hybrid approach maintains accuracy while substantially reducing attack success rates and increasing computational costs for adversaries.

AINeutralarXiv – CS AI · 2d ago6/10
🧠

Neuro-Symbolic Strong-AI Robots with Closed Knowledge Assumption: Learning and Deductions

This academic paper proposes a neuro-symbolic approach for AGI robots combining neural networks with formal logic reasoning using Belnap's 4-valued logic system. The framework enables robots to handle unknown information, inconsistencies, and paradoxes while maintaining controlled security through axiom-based logic inference.

AIBullishBlockonomi · 3d ago6/10
🧠

Wedbush Sees Buying Opportunity in Tech Downturn — 5 Cybersecurity Stocks to Watch

Wedbush Securities views the recent technology sector decline as an excessive market overreaction, identifying CrowdStrike, Palo Alto Networks, and three additional cybersecurity firms as attractive buying opportunities. The recommendation reflects analyst confidence that AI-driven security demand will sustain growth despite near-term volatility.

AINeutralarXiv – CS AI · 3d ago6/10
🧠

Leave My Images Alone: Preventing Multi-Modal Large Language Models from Analyzing Images via Visual Prompt Injection

Researchers introduce ImageProtector, a user-side defense mechanism that embeds imperceptible perturbations into images to prevent multi-modal large language models from analyzing them. When adversaries attempt to extract sensitive information from protected images, MLLMs are induced to refuse analysis, though potential countermeasures exist that may partially mitigate the technique's effectiveness.

AI × CryptoBearishThe Register – AI · 4d ago5/10
🤖

Anthropic's mysterious Mythos AI threatens to upend the infosec world

The article title references Anthropic's alleged 'Mythos AI' as a potential security threat to the information security industry, though no article body was provided to verify claims or assess actual impact.

🏢 Anthropic
AI × CryptoBullishCrypto Briefing · 5d ago7/10
🤖

Gavriel Cohen: AI native service companies can achieve software-like margins, the rise of AI agents in marketing, and security risks of complex architectures | MLST

Gavriel Cohen discusses how AI-native service companies can achieve software-like profit margins through minimal, secure tool design, exemplified by NanoClaw's success. The article explores the emerging role of AI agents in marketing while highlighting security vulnerabilities inherent in complex AI architectures.

Gavriel Cohen: AI native service companies can achieve software-like margins, the rise of AI agents in marketing, and security risks of complex architectures | MLST
AIBearishCrypto Briefing · 5d ago6/10
🧠

Ranjan Roy: AI marketing hype often overshadows substance, concerns about AI exploiting software vulnerabilities, and the significance of scaling laws in model performance | Big Technology

Ranjan Roy highlights how AI marketing hype often obscures substantive security concerns, particularly regarding AI systems exploiting software vulnerabilities. The analysis emphasizes the importance of scaling laws in model performance and urges critical evaluation of AI breakthroughs beyond promotional claims.

Ranjan Roy: AI marketing hype often overshadows substance, concerns about AI exploiting software vulnerabilities, and the significance of scaling laws in model performance | Big Technology
AINeutralWired – AI · 6d ago6/10
🧠

Anthropic’s Mythos Will Force a Cybersecurity Reckoning—Just Not the One You Think

Anthropic's new Mythos AI model is raising cybersecurity concerns as experts warn it could be weaponized by hackers, though the real issue lies in developers' historical neglect of security practices. The model's capabilities are forcing the industry to confront long-standing vulnerabilities in software development that predate advanced AI systems.

Anthropic’s Mythos Will Force a Cybersecurity Reckoning—Just Not the One You Think
🏢 Anthropic
AI × CryptoBullishBlockonomi · 6d ago6/10
🤖

CrowdStrike (CRWD) Stock Rebounds After Anthropic Partnership Erases AI Disruption Fears

CrowdStrike (CRWD) stock rebounded after announcing a strategic partnership with Anthropic's Project Glass Wing initiative, alleviating investor concerns about AI-driven disruption to its cybersecurity business. The partnership signals the company's adaptation to the evolving AI landscape and positions it alongside a major AI research organization.

🏢 Anthropic
AINeutralarXiv – CS AI · 6d ago6/10
🧠

Fighting AI with AI: AI-Agent Augmented DNS Blocking of LLM Services during Student Evaluations

Researchers introduce AI-Sinkhole, an AI-agent augmented DNS-blocking framework that dynamically detects and temporarily blocks LLM chatbot services during proctored exams to prevent academic integrity violations. The system uses quantized LLMs for semantic classification and Pi-Hole for network-wide DNS blocking, achieving robust cross-lingual detection with F1-scores exceeding 0.83.

AINeutralarXiv – CS AI · 6d ago6/10
🧠

SkillSieve: A Hierarchical Triage Framework for Detecting Malicious AI Agent Skills

Researchers introduced SkillSieve, a three-layer detection framework that identifies malicious AI agent skills in OpenClaw's ClawHub marketplace, where 13-26% of over 13,000 skills contain security vulnerabilities. The system combines regex/AST scanning, LLM-based analysis with parallel sub-tasks, and multi-LLM voting to achieve 0.800 F1 score at $0.006 per skill, significantly outperforming existing detection methods.

AINeutralarXiv – CS AI · 6d ago6/10
🧠

Towards Privacy-Preserving Large Language Model: Text-free Inference Through Alignment and Adaptation

Researchers introduce Privacy-Preserving Fine-Tuning (PPFT), a novel training approach that enables LLM services to process user queries without receiving raw text, addressing privacy vulnerabilities in current deployments. The method uses client-side encoders and noise-injected embeddings to maintain competitive model performance while eliminating exposure of sensitive personal, medical, or legal information.

AINeutralarXiv – CS AI · Apr 66/10
🧠

Split and Conquer Partial Deepfake Speech

Researchers developed a new AI framework for detecting partial deepfake speech by splitting the problem into boundary detection and segment classification stages. The method achieves state-of-the-art performance on benchmark datasets, significantly improving detection and localization of manipulated audio regions within otherwise authentic speech.

AIBearisharXiv – CS AI · Apr 66/10
🧠

LogicPoison: Logical Attacks on Graph Retrieval-Augmented Generation

Researchers have discovered LogicPoison, a new attack method that exploits vulnerabilities in Graph-based Retrieval-Augmented Generation (GraphRAG) systems by corrupting logical connections in knowledge graphs without altering text semantics. The attack successfully bypasses GraphRAG's existing defenses by targeting the topological integrity of underlying graphs, significantly degrading AI system performance.

← PrevPage 6 of 9Next →