133 articles tagged with #healthcare. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBearisharXiv – CS AI · Mar 267/10
🧠Researchers have identified critical privacy vulnerabilities in deep learning models used for time series imputation, demonstrating that these models can leak sensitive training data through membership and attribute inference attacks. The study introduces a two-stage attack framework that successfully retrieves significant portions of training data even from models designed to be robust against overfitting-based attacks.
AINeutralGoogle DeepMind Blog · Mar 257/10
🧠Google DeepMind is conducting research into AI's potential for harmful manipulation across critical sectors including finance and healthcare. This research is driving the development of new safety measures to protect people from AI-powered manipulation tactics.
🏢 Google
AIBearisharXiv – CS AI · Mar 177/10
🧠Researchers evaluated the faithfulness of closed-source AI models like ChatGPT and Gemini in medical reasoning, finding that their explanations often appear plausible but don't reflect actual reasoning processes. The study revealed these models frequently incorporate external hints without acknowledgment and their chain-of-thought reasoning doesn't causally drive predictions, raising safety concerns for medical applications.
🧠 ChatGPT🧠 Gemini
AINeutralarXiv – CS AI · Mar 177/10
🧠Researchers identified that medical multimodal large language models (MLLMs) fail primarily due to inadequate visual grounding capabilities when analyzing medical images, unlike their success with natural scenes. They developed VGMED evaluation dataset and proposed VGRefine method, achieving state-of-the-art performance across 6 medical visual question-answering benchmarks without additional training.
AIBearisharXiv – CS AI · Mar 127/10
🧠Research study finds that LLaMA-70B-Instruct hallucinated in 19.7% of medical Q&A responses despite high plausibility scores, highlighting significant reliability issues in AI healthcare applications. The study shows that lower hallucination rates correlate with higher usefulness scores, emphasizing the need for better safeguards in medical AI systems.
AIBullisharXiv – CS AI · Mar 117/10
🧠Researchers have developed Meissa, a lightweight 4B-parameter medical AI model that brings advanced agentic capabilities offline for healthcare applications. The system matches frontier models like GPT in medical benchmarks while operating with 25x fewer parameters and 22x lower latency, addressing privacy and cost concerns in clinical settings.
🧠 Gemini
AIBullisharXiv – CS AI · Mar 117/10
🧠Researchers developed EyExIn, a new AI framework that addresses critical gaps in large vision language models for medical diagnosis by anchoring them with domain-specific expert knowledge. The system uses dual-stream encoding and deep expert injection to improve accuracy in ophthalmic diagnosis, outperforming existing proprietary systems across four benchmarks.
AINeutralarXiv – CS AI · Mar 97/10
🧠Researchers evaluated 34 large language models on radiology questions, finding that agentic retrieval-augmented reasoning systems improve consensus and reliability across different AI models. The study shows these systems reduce decision variability between models and increase robust correctness, though 72% of incorrect outputs still carried moderate to high clinical severity.
AINeutralFortune Crypto · Mar 67/10
🧠OpenAI investor Vinod Khosla predicts AI will automate 80% of jobs by 2030, potentially creating mass unemployment. The Silicon Valley billionaire envisions this leading to a deflationary economy with free healthcare and education, requiring significant tax policy reforms to manage the economic transition.
🏢 OpenAI
AINeutralarXiv – CS AI · Mar 67/10
🧠Researchers introduce BioLLMAgent, a hybrid framework combining reinforcement learning models with large language models to simulate human decision-making in computational psychiatry. The framework demonstrates strong interpretability while accurately reproducing human behavioral patterns and successfully simulating cognitive behavioral therapy principles.
AIBullisharXiv – CS AI · Mar 56/10
🧠Researchers developed an automated AI pipeline for detecting cervical spine fractures in medical imaging using a novel 2D-to-3D projection approach. The system achieved clinically relevant performance comparable to expert radiologists while reducing computational complexity through optimized 2D projections instead of traditional 3D methods.
AIBullisharXiv – CS AI · Mar 56/10
🧠Researchers developed MA-RAG, a Multi-Round Agentic RAG framework that improves medical AI reasoning by iteratively refining responses through conflict detection and external evidence retrieval. The system achieved a substantial +6.8 point accuracy improvement over baseline models across 7 medical Q&A benchmarks by addressing hallucinations and outdated knowledge in healthcare AI applications.
AIBullisharXiv – CS AI · Mar 56/10
🧠Researchers developed NeuroFlowNet, a novel AI framework using Conditional Normalizing Flow to reconstruct deep brain EEG signals from non-invasive scalp measurements. This breakthrough enables analysis of deep temporal lobe brain activity without requiring invasive electrode implantation, potentially transforming neuroscience research and clinical diagnosis.
AINeutralarXiv – CS AI · Mar 57/10
🧠Researchers have released ERDES, the first open-access dataset of ocular ultrasound videos for detecting retinal detachment and macular status using machine learning. The dataset addresses a critical gap in automated medical diagnosis by enabling AI models to classify retinal detachment severity, which is essential for determining surgical urgency.
AINeutralarXiv – CS AI · Mar 57/10
🧠Researchers propose RAG-X, a diagnostic framework for evaluating retrieval-augmented generation systems in medical AI applications. The study reveals an 'Accuracy Fallacy' showing a 14% gap between perceived system success and actual evidence-based grounding in medical question-answering systems.
AIBullisharXiv – CS AI · Mar 57/10
🧠Researchers introduce SPRINT, the first Few-Shot Class-Incremental Learning (FSCIL) framework designed specifically for tabular data domains like cybersecurity and healthcare. The system achieves 77.37% accuracy in 5-shot learning scenarios, outperforming existing methods by 4.45% through novel semi-supervised techniques that leverage unlabeled data and confidence-based pseudo-labeling.
AIBullisharXiv – CS AI · Mar 57/10
🧠Stanford researchers introduced Merlin, a 3D vision-language foundation model for analyzing abdominal CT scans that processes volumetric medical images alongside electronic health records and radiology reports. The model was trained on over 6 million images from 15,331 CT scans and demonstrated superior performance compared to existing 2D models across 752 individual medical tasks.
AIBullisharXiv – CS AI · Mar 47/103
🧠Researchers developed ATPO (Adaptive Tree Policy Optimization), a new AI algorithm for multi-turn medical dialogues that outperforms existing methods by better handling uncertainty in patient-doctor interactions. The algorithm enabled a smaller Qwen3-8B model to surpass GPT-4o's accuracy by 0.92% on medical dialogue benchmarks through improved value estimation and exploration strategies.
AIBullisharXiv – CS AI · Mar 47/102
🧠Researchers have released MedXIAOHE, a new medical vision-language AI foundation model that achieves state-of-the-art performance across medical benchmarks and surpasses leading closed-source systems. The model incorporates advanced features like entity-aware pretraining, reinforcement learning for medical reasoning, and evidence-grounded report generation to improve reliability in clinical applications.
AIBullisharXiv – CS AI · Mar 47/103
🧠Researchers have developed MedLA, a new logic-driven multi-agent AI framework that uses large language models for complex medical reasoning. The system employs multiple AI agents that organize their reasoning into explicit logical trees and engage in structured discussions to resolve inconsistencies and reach consensus on medical questions.
AIBullisharXiv – CS AI · Mar 46/102
🧠NeuroWise is a multi-agent LLM system designed to help neurotypical individuals better communicate with autistic partners through AI-based coaching and interpretation. A study of 30 participants showed the system significantly reduced deficit-based thinking about autism and improved communication efficiency by 37%.
AIBullisharXiv – CS AI · Mar 47/103
🧠Researchers introduce MIRAGE, a novel AI framework that uses knowledge graphs and electronic health records to predict Alzheimer's disease when MRI scans are unavailable. The system improves AD classification rates by 13% compared to single-modality approaches by creating synthetic representations without expensive 3D brain scan reconstruction.
AIBullisharXiv – CS AI · Mar 46/103
🧠Researchers developed an interpretable AI framework for detecting structural heart disease from electrocardiograms, achieving better performance than existing deep-learning methods while providing clinical transparency. The model demonstrated improvements of nearly 1% across key metrics using the EchoNext benchmark of over 80,000 ECG-ECHO pairs.
AIBullisharXiv – CS AI · Mar 46/103
🧠Researchers introduce PRISM, an EEG foundation model that demonstrates how diverse pretraining data leads to better clinical performance than narrow-source datasets. The study shows that geographically diverse EEG data outperforms larger but homogeneous datasets in medical diagnosis tasks, particularly achieving 12.3% better accuracy in distinguishing epilepsy from similar conditions.
$COMP
AIBullisharXiv – CS AI · Mar 46/102
🧠Researchers developed GTDoctor, an AI model for diagnosing gestational trophoblastic disease that achieves over 91% precision in lesion detection. The system reduces diagnostic time from 56 to 16 seconds per case while maintaining 95.59% positive predictive value in clinical trials.