88 articles tagged with #natural-language-processing. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBullisharXiv – CS AI · Feb 276/106
🧠Researchers developed a three-stage framework using Small Language Models (SLMs) to automatically translate natural language queries into Kusto Query Language (KQL) for cybersecurity operations. The approach achieves high accuracy (98.7% syntax, 90.6% semantic) while reducing costs by up to 10x compared to GPT-4, potentially solving bottlenecks in Security Operations Centers.
AIBullisharXiv – CS AI · Feb 276/105
🧠Researchers introduce MoDora, an AI-powered system that uses tree-based analysis to understand and answer questions about semi-structured documents containing mixed data elements like tables, charts, and text. The system addresses challenges in processing fragmented OCR data and hierarchical document structures, achieving 5.97%-61.07% accuracy improvements over existing baselines.
AIBullisharXiv – CS AI · Feb 276/106
🧠Researchers developed MALLET, a multi-agent AI system that reduces emotional intensity in news content by up to 19.3% while preserving semantic meaning. The system uses four specialized agents to analyze, adjust, and personalize content presentation modes for calmer decision-making without restricting access to original information.
$NEAR
AINeutralApple Machine Learning · Feb 256/103
🧠Research identifies a significant performance gap between speech-adapted Large Language Models and their text-based counterparts on language understanding tasks. Current approaches to bridge this gap rely on expensive large-scale speech synthesis methods, highlighting a key challenge in extending LLM capabilities to audio inputs.
AIBullishHugging Face Blog · Feb 276/105
🧠HuggingFace has partnered with the Indian Institute of Science (IISc) to enhance AI model development for India's diverse linguistic landscape. This collaboration aims to improve natural language processing capabilities across multiple Indian languages, potentially expanding AI accessibility in the region.
AIBullishOpenAI News · Mar 156/106
🧠OpenAI has released new versions of GPT-3 and Codex with enhanced capabilities that allow users to edit and insert content into existing text, rather than only completing text. This represents a significant advancement in AI text editing functionality beyond traditional text generation.
AINeutralarXiv – CS AI · Apr 64/10
🧠Researchers developed EWAD and CPDP techniques for improving multi-teacher knowledge distillation in low-resource abstractive summarization tasks. The study across Bangla and cross-lingual datasets shows logit-level knowledge distillation provides most reliable gains, while complex distillation improves short summaries but degrades longer outputs.
AINeutralarXiv – CS AI · Mar 274/10
🧠An academic research paper provides a comprehensive historical review of chatbot technology evolution from 1906 statistical models through early systems like ELIZA to modern AI conversational agents like ChatGPT and Google Bard. The study traces key milestones and paradigm shifts that shaped conversational AI development over decades.
🧠 ChatGPT
AIBullisharXiv – CS AI · Mar 175/10
🧠Researchers introduce IDALC, a semi-supervised framework for voice-controlled dialog systems that improves intent detection and reduces manual annotation costs. The system achieves 5-10% higher accuracy and 4-8% better macro-F1 scores while requiring annotation of only 6-10% of unlabeled data.
AINeutralarXiv – CS AI · Mar 164/10
🧠Researchers developed an automated query expansion framework using multiple large language models that constructs domain-specific examples without manual intervention. The system uses a two-LLM ensemble approach where different models generate expansions that are then refined by a third LLM, showing significant improvements over traditional methods across multiple datasets.
AINeutralarXiv – CS AI · Mar 125/10
🧠Researchers introduced the Contextual Emotional Inference (CEI) Benchmark, a dataset of 300 human-validated scenarios designed to evaluate how well large language models understand pragmatic reasoning in complex communication. The benchmark tests LLMs' ability to interpret ambiguous utterances across five pragmatic subtypes including sarcasm, mixed signals, and passive aggression in various social contexts.
AINeutralarXiv – CS AI · Mar 124/10
🧠Researchers developed an automated framework to evaluate Large Language Models' effectiveness in translating Mandarin Chinese to English, comparing GPT-4, GPT-4o, and DeepSeek against Google Translate. While LLMs performed well on news translation, they showed varying results with literary texts, with DeepSeek excelling at cultural subtleties and GPT-4o/DeepSeek better at semantic conservation.
🏢 Meta🧠 GPT-4
AINeutralarXiv – CS AI · Mar 114/10
🧠Researchers introduce VoxEmo, a comprehensive benchmark for evaluating Speech Large Language Models on emotion recognition tasks across 35 emotion corpora and 15 languages. The benchmark addresses evaluation challenges in open text generation and introduces novel protocols that better align with human subjective emotion perception.
AINeutralarXiv – CS AI · Mar 114/10
🧠Researchers propose RbtAct, a novel approach that uses peer review rebuttals as supervision to train AI models for generating more actionable scientific review feedback. The system leverages a new dataset RMR-75K and fine-tuned Llama-3.1-8B model to produce focused, implementable guidance rather than superficial comments.
🧠 Llama
AINeutralarXiv – CS AI · Mar 54/10
🧠Researchers developed a web tool that uses natural language as the primary interface for LLM-assisted educational game design, allowing instructors to collaborate with AI to create games with specific learning outcomes. The tool maps pedagogy to gameplay through four linked components while maintaining human agency in critical design decisions.
AINeutralarXiv – CS AI · Mar 54/10
🧠Researchers have created CzechTopic, a new benchmark dataset for evaluating AI models' ability to identify specific topics within historical Czech documents. The study compared various large language models and BERT-based models, finding significant performance variations with the strongest models approaching human-level accuracy in topic detection.
AINeutralarXiv – CS AI · Mar 54/10
🧠TopicENA is a new framework that combines BERTopic with Epistemic Network Analysis to automatically analyze concept relationships in large text datasets without manual coding. The research demonstrates that automated topic modeling can replace expert manual coding while maintaining analytical quality, making network analysis scalable for large corpora.
AINeutralarXiv – CS AI · Mar 44/102
🧠Researchers developed new prompting-based approaches using multimodal large language models to generate real-time video commentary that considers both content relevance and timing. The study introduces dynamic interval-based decoding that adjusts prediction timing based on utterance duration, showing improved alignment with human commentary patterns without requiring model fine-tuning.
AIBullisharXiv – CS AI · Mar 44/103
🧠Researchers propose ASEGR, a novel AI framework that enhances product recommendation systems by extracting sensory attributes from user reviews using large language models. The system uses a two-stage pipeline where an LLM extracts structured sensory data which is then distilled into compact embeddings for sequential recommendation models.
AINeutralarXiv – CS AI · Mar 44/103
🧠Researchers developed a novel approach using instruction-tuned Large Language Models to improve argumentative component detection in text analysis. The method reframes the task as language generation rather than traditional sequence labeling, achieving superior performance on standard benchmarks compared to existing state-of-the-art systems.
AINeutralarXiv – CS AI · Mar 34/104
🧠Researchers have created CrimeNER, a specialized dataset of over 1,500 annotated crime-related documents for training named-entity recognition AI models. The study addresses the lack of quality training data in the crime domain by developing a database from terrorist attack reports and DOJ press notes, defining 22 types of crime-related entities.
AINeutralarXiv – CS AI · Mar 25/104
🧠A study evaluated large language models (Claude, Gemini, ChatGPT) translating Ancient Greek texts, finding high performance on previously translated works (95.2/100) but declining quality on untranslated technical texts (79.9/100). Terminology rarity was identified as a strong predictor of translation failure, with rare terms causing catastrophic performance drops.
AINeutralarXiv – CS AI · Mar 25/109
🧠Researchers explore using large language models (LLMs) as mediators rather than just moderators in online conflicts, developing a framework that combines judgment evaluation and empathetic intervention. Their study using Reddit data shows API-based models outperform open-source alternatives in de-escalating flame wars and fostering constructive dialogue.
AINeutralarXiv – CS AI · Mar 25/105
🧠Researchers developed LEC-KG, a new framework that combines Large Language Models with Knowledge Graph Embeddings to better extract and structure information from unstructured text. The system was tested on Chinese Sustainable Development Goal reports and showed significant improvements over traditional LLM approaches, particularly for identifying rare relationships in domain-specific content.
AIBullisharXiv – CS AI · Feb 274/106
🧠Researchers introduce Alignment-Aware Masked Learning (AML), a new training strategy for Referring Image Segmentation that improves pixel-level vision-language alignment. The approach achieves state-of-the-art performance on RefCOCO datasets by filtering poorly aligned regions and focusing on reliable visual-language cues.