49 articles tagged with #human-ai-collaboration. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBullisharXiv – CS AI · Mar 116/10
🧠Researchers introduce Social-R1, a reinforcement learning framework that enhances social reasoning in large language models by training on adversarial examples. The approach enables a 4B parameter model to outperform larger models across eight benchmarks by supervising the entire reasoning process rather than just outcomes.
AIBullisharXiv – CS AI · Mar 96/10
🧠Researchers developed 'Companion,' an AI system that combines drawing robots with Large Language Models to create a collaborative artistic partner. The system engages in real-time bidirectional interaction through speech and sketching, with art experts validating its ability to produce works with distinct aesthetic identity and exhibition merit.
AINeutralarXiv – CS AI · Mar 96/10
🧠A research study involving 737 participants found that human guidance is crucial in 'vibe coding' - using natural language to generate code through AI. The study shows hybrid systems perform best when humans provide high-level instructions while AI handles evaluation, with AI-only instruction leading to performance collapse.
AINeutralarXiv – CS AI · Mar 36/1011
🧠Researchers introduce LifeEval, a new multimodal benchmark designed to evaluate how well AI assistants can help humans in real-time daily life tasks from a first-person perspective. The benchmark reveals significant challenges for current AI models in providing timely and adaptive assistance in dynamic environments.
AIBullisharXiv – CS AI · Mar 36/104
🧠Researchers developed CLEO, an AI system that enables real-time collaborative context awareness between humans and AI agents by interpreting concurrent user actions on shared artifacts. A study with professional designers identified key interaction patterns and decision factors for when to delegate work to AI versus collaborate directly.
AIBullisharXiv – CS AI · Mar 36/104
🧠Researchers introduce AIssistant, an open-source framework that combines human expertise with AI agents to streamline scientific review and perspective paper creation in data science. The system uses 15 specialized LLM-driven agents across two workflows and demonstrates 65.7% time savings while maintaining research quality through strategic human oversight.
AIBullisharXiv – CS AI · Mar 36/103
🧠Research shows that predictive AI deployment during medical training significantly improves diagnostic accuracy for novices, with the greatest benefits occurring when AI is used in both training and practice phases. The study found that AI integration not only enhances individual performance but also affects error diversity across groups, impacting collective decision-making quality.
AIBullisharXiv – CS AI · Mar 26/1010
🧠Researchers introduce CowPilot, a framework that combines autonomous AI agents with human collaboration for web navigation tasks. The system achieved 95% success rate while requiring humans to perform only 15.2% of total steps, demonstrating effective human-AI cooperation for complex web tasks.
AIBullisharXiv – CS AI · Mar 27/1015
🧠Researchers developed MACD, a Multi-Agent Clinical Diagnosis framework that enables large language models to self-learn clinical knowledge and improve medical diagnosis accuracy. The system achieved up to 22.3% improvement over clinical guidelines and 16% improvement over physician-only diagnosis when tested on 4,390 real-world patient cases.
AIBullisharXiv – CS AI · Feb 276/107
🧠Researchers introduce AHCE (Active Human-Augmented Challenge Engagement), a framework that enables AI agents to collaborate with human experts more effectively through learned policies. The system achieved 32% improvement on normal difficulty tasks and 70% on difficult tasks in Minecraft experiments by treating humans as interactive reasoning tools rather than simple help sources.
AINeutralarXiv – CS AI · Feb 276/106
🧠Researchers published a case study demonstrating successful human-AI collaboration in mathematical research, extending Hermite quadrature rule results beyond manual capabilities. The study reveals AI's strengths in algebraic manipulation and proof exploration, while highlighting the critical need for human verification and domain expertise in every step of the research process.
AINeutralOpenAI News · Nov 215/102
🧠The article discusses advancements in red teaming methodologies that combine human expertise with artificial intelligence capabilities. This represents a significant development in cybersecurity practices and AI safety testing approaches.
AIBullishOpenAI News · Aug 156/106
🧠OpenAI is implementing GPT-4 for content policy development and moderation decisions to improve consistency and efficiency. This approach reduces human moderator involvement while enabling faster policy refinement through improved feedback loops.
AIBullishOpenAI News · Jun 136/105
🧠Researchers developed AI models that can identify and describe flaws in text summaries, helping human evaluators detect problems more effectively. Larger AI models showed better self-critique capabilities than summary-writing abilities, suggesting potential for AI-assisted supervision of AI systems.
AINeutralarXiv – CS AI · Apr 75/10
🧠Researchers have developed BLK-Assist, a modular framework that enables artists to fine-tune AI diffusion models using their own artwork while maintaining privacy and stylistic control. The system includes three components for concept generation, transparency-preserving assets, and high-resolution outputs, demonstrating a consent-based approach to human-AI collaboration in creative work.
AINeutralWired – AI · Mar 265/10
🧠Independent tech reporters are increasingly integrating AI agents throughout their entire reporting workflow, from research to writing to editing. This trend raises questions about the evolving role and value proposition of human journalists in an AI-augmented media landscape.
AINeutralarXiv – CS AI · Mar 95/10
🧠Researchers analyzed how the GPT-J-6B language model internally represents and reasons about trust by comparing its embeddings to established human trust models. The study found that the AI's trust representation most closely aligns with the Castelfranchi socio-cognitive model, suggesting LLMs encode social concepts in meaningful ways.
AINeutralarXiv – CS AI · Mar 64/10
🧠This academic research paper examines the challenges of human-AI teaming as AI systems become more autonomous and agentic. The study proposes extending Team Situation Awareness theory to address structural uncertainties that arise when AI systems can take open-ended actions and evolve their objectives over time.
AINeutralarXiv – CS AI · Mar 54/10
🧠A research study examined how generative AI models perform in business decision-making contexts, particularly their ability to detect ambiguity and resist sycophantic behavior. The study found that while AI excels at identifying contradictions and contextual ambiguities, it struggles with linguistic nuances and requires human oversight to function as a reliable strategic partner.
AINeutralarXiv – CS AI · Mar 54/10
🧠Researchers propose IntPro, a new AI proxy agent that improves intent understanding by learning from individual user patterns through retrieval-conditioned inference. The system uses historical intent data and specialized training methods to better interpret user intentions in context-aware scenarios.
AINeutralarXiv – CS AI · Mar 54/10
🧠Researchers developed a web tool that uses natural language as the primary interface for LLM-assisted educational game design, allowing instructors to collaborate with AI to create games with specific learning outcomes. The tool maps pedagogy to gameplay through four linked components while maintaining human agency in critical design decisions.
AIBullisharXiv – CS AI · Mar 35/1011
🧠ViviDoc is a new human-agent collaborative system that generates interactive educational documents using a multi-agent pipeline and Document Specification framework. The system allows educators to review and refine AI-generated content plans before code production, significantly outperforming naive AI generation methods.
$RNDR
AIBullishOpenAI News · Oct 14/106
🧠Altera is leveraging OpenAI's GPT-4o to develop new collaborative frameworks between AI agents and humans. The initiative represents an advancement in human-AI partnership models using the latest GPT technology.
AINeutralarXiv – CS AI · Mar 34/104
🧠Researchers present a multi-agent Large Language Model framework for interactive AI planning systems that provides context-dependent explanations to human planners. The system aims to facilitate collaborative decision-making between humans and AI rather than replacing human planners entirely.