y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#natural-language-processing News & Analysis

88 articles tagged with #natural-language-processing. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

88 articles
AIBullisharXiv – CS AI · Feb 276/106
🧠

Towards Small Language Models for Security Query Generation in SOC Workflows

Researchers developed a three-stage framework using Small Language Models (SLMs) to automatically translate natural language queries into Kusto Query Language (KQL) for cybersecurity operations. The approach achieves high accuracy (98.7% syntax, 90.6% semantic) while reducing costs by up to 10x compared to GPT-4, potentially solving bottlenecks in Security Operations Centers.

AIBullisharXiv – CS AI · Feb 276/105
🧠

MoDora: Tree-Based Semi-Structured Document Analysis System

Researchers introduce MoDora, an AI-powered system that uses tree-based analysis to understand and answer questions about semi-structured documents containing mixed data elements like tables, charts, and text. The system addresses challenges in processing fragmented OCR data and hierarchical document structures, achieving 5.97%-61.07% accuracy improvements over existing baselines.

AIBullisharXiv – CS AI · Feb 276/106
🧠

Multi-Agent Large Language Model Based Emotional Detoxification Through Personalized Intensity Control for Consumer Protection

Researchers developed MALLET, a multi-agent AI system that reduces emotional intensity in news content by up to 19.3% while preserving semantic meaning. The system uses four specialized agents to analyze, adjust, and personalize content presentation modes for calmer decision-making without restricting access to original information.

$NEAR
AINeutralApple Machine Learning · Feb 256/103
🧠

Closing the Gap Between Text and Speech Understanding in LLMs

Research identifies a significant performance gap between speech-adapted Large Language Models and their text-based counterparts on language understanding tasks. Current approaches to bridge this gap rely on expensive large-scale speech synthesis methods, highlighting a key challenge in extending LLM capabilities to audio inputs.

AIBullishHugging Face Blog · Feb 276/105
🧠

HuggingFace, IISc partner to supercharge model building on India's diverse languages

HuggingFace has partnered with the Indian Institute of Science (IISc) to enhance AI model development for India's diverse linguistic landscape. This collaboration aims to improve natural language processing capabilities across multiple Indian languages, potentially expanding AI accessibility in the region.

AIBullishOpenAI News · Mar 156/106
🧠

New GPT-3 capabilities: Edit & insert

OpenAI has released new versions of GPT-3 and Codex with enhanced capabilities that allow users to edit and insert content into existing text, rather than only completing text. This represents a significant advancement in AI text editing functionality beyond traditional text generation.

AINeutralarXiv – CS AI · Apr 64/10
🧠

Reliability Gated Multi-Teacher Distillation for Low Resource Abstractive Summarization

Researchers developed EWAD and CPDP techniques for improving multi-teacher knowledge distillation in low-resource abstractive summarization tasks. The study across Bangla and cross-lingual datasets shows logit-level knowledge distillation provides most reliable gains, while complex distillation improves short summaries but degrades longer outputs.

AINeutralarXiv – CS AI · Mar 164/10
🧠

Automatic In-Domain Exemplar Construction and LLM-Based Refinement of Multi-LLM Expansions for Query Expansion

Researchers developed an automated query expansion framework using multiple large language models that constructs domain-specific examples without manual intervention. The system uses a two-LLM ensemble approach where different models generate expansions that are then refined by a third LLM, showing significant improvements over traditional methods across multiple datasets.

AINeutralarXiv – CS AI · Mar 125/10
🧠

CEI: A Benchmark for Evaluating Pragmatic Reasoning in Language Models

Researchers introduced the Contextual Emotional Inference (CEI) Benchmark, a dataset of 300 human-validated scenarios designed to evaluate how well large language models understand pragmatic reasoning in complex communication. The benchmark tests LLMs' ability to interpret ambiguous utterances across five pragmatic subtypes including sarcasm, mixed signals, and passive aggression in various social contexts.

AINeutralarXiv – CS AI · Mar 124/10
🧠

Automated evaluation of LLMs for effective machine translation of Mandarin Chinese to English

Researchers developed an automated framework to evaluate Large Language Models' effectiveness in translating Mandarin Chinese to English, comparing GPT-4, GPT-4o, and DeepSeek against Google Translate. While LLMs performed well on news translation, they showed varying results with literary texts, with DeepSeek excelling at cultural subtleties and GPT-4o/DeepSeek better at semantic conservation.

🏢 Meta🧠 GPT-4
AINeutralarXiv – CS AI · Mar 114/10
🧠

VoxEmo: Benchmarking Speech Emotion Recognition with Speech LLMs

Researchers introduce VoxEmo, a comprehensive benchmark for evaluating Speech Large Language Models on emotion recognition tasks across 35 emotion corpora and 15 languages. The benchmark addresses evaluation challenges in open text generation and introduces novel protocols that better align with human subjective emotion perception.

AINeutralarXiv – CS AI · Mar 114/10
🧠

RbtAct: Rebuttal as Supervision for Actionable Review Feedback Generation

Researchers propose RbtAct, a novel approach that uses peer review rebuttals as supervision to train AI models for generating more actionable scientific review feedback. The system leverages a new dataset RMR-75K and fine-tuned Llama-3.1-8B model to produce focused, implementable guidance rather than superficial comments.

🧠 Llama
AINeutralarXiv – CS AI · Mar 54/10
🧠

Bridging Pedagogy and Play: Introducing a Language Mapping Interface for Human-AI Co-Creation in Educational Game Design

Researchers developed a web tool that uses natural language as the primary interface for LLM-assisted educational game design, allowing instructors to collaborate with AI to create games with specific learning outcomes. The tool maps pedagogy to gameplay through four linked components while maintaining human agency in critical design decisions.

AINeutralarXiv – CS AI · Mar 54/10
🧠

CzechTopic: A Benchmark for Zero-Shot Topic Localization in Historical Czech Documents

Researchers have created CzechTopic, a new benchmark dataset for evaluating AI models' ability to identify specific topics within historical Czech documents. The study compared various large language models and BERT-based models, finding significant performance variations with the strongest models approaching human-level accuracy in topic detection.

AINeutralarXiv – CS AI · Mar 54/10
🧠

TopicENA: Enabling Epistemic Network Analysis at Scale through Automated Topic-Based Coding

TopicENA is a new framework that combines BERTopic with Epistemic Network Analysis to automatically analyze concept relationships in large text datasets without manual coding. The research demonstrates that automated topic modeling can replace expert manual coding while maintaining analytical quality, making network analysis scalable for large corpora.

AINeutralarXiv – CS AI · Mar 44/102
🧠

Real-Time Generation of Game Video Commentary with Multimodal LLMs: Pause-Aware Decoding Approaches

Researchers developed new prompting-based approaches using multimodal large language models to generate real-time video commentary that considers both content relevance and timing. The study introduces dynamic interval-based decoding that adjusts prediction timing based on utterance duration, showing improved alignment with human commentary patterns without requiring model fine-tuning.

AIBullisharXiv – CS AI · Mar 44/103
🧠

Sensory-Aware Sequential Recommendation via Review-Distilled Representations

Researchers propose ASEGR, a novel AI framework that enhances product recommendation systems by extracting sensory attributes from user reviews using large language models. The system uses a two-stage pipeline where an LLM extracts structured sensory data which is then distilled into compact embeddings for sequential recommendation models.

AINeutralarXiv – CS AI · Mar 44/103
🧠

Compact Prompting in Instruction-tuned LLMs for Joint Argumentative Component Detection

Researchers developed a novel approach using instruction-tuned Large Language Models to improve argumentative component detection in text analysis. The method reframes the task as language generation rather than traditional sequence labeling, achieving superior performance on standard benchmarks compared to existing state-of-the-art systems.

AINeutralarXiv – CS AI · Mar 25/104
🧠

Terminology Rarity Predicts Catastrophic Failure in LLM Translation of Low-Resource Ancient Languages: Evidence from Ancient Greek

A study evaluated large language models (Claude, Gemini, ChatGPT) translating Ancient Greek texts, finding high performance on previously translated works (95.2/100) but declining quality on untranslated technical texts (79.9/100). Terminology rarity was identified as a strong predictor of translation failure, with rare terms causing catastrophic performance drops.

AINeutralarXiv – CS AI · Mar 25/109
🧠

From Moderation to Mediation: Can LLMs Serve as Mediators in Online Flame Wars?

Researchers explore using large language models (LLMs) as mediators rather than just moderators in online conflicts, developing a framework that combines judgment evaluation and empathetic intervention. Their study using Reddit data shows API-based models outperform open-source alternatives in de-escalating flame wars and fostering constructive dialogue.

AINeutralarXiv – CS AI · Mar 25/105
🧠

LEC-KG: An LLM-Embedding Collaborative Framework for Domain-Specific Knowledge Graph Construction -- A Case Study on SDGs

Researchers developed LEC-KG, a new framework that combines Large Language Models with Knowledge Graph Embeddings to better extract and structure information from unstructured text. The system was tested on Chinese Sustainable Development Goal reports and showed significant improvements over traditional LLM approaches, particularly for identifying rare relationships in domain-specific content.

AIBullisharXiv – CS AI · Feb 274/106
🧠

AMLRIS: Alignment-aware Masked Learning for Referring Image Segmentation

Researchers introduce Alignment-Aware Masked Learning (AML), a new training strategy for Referring Image Segmentation that improves pixel-level vision-language alignment. The approach achieves state-of-the-art performance on RefCOCO datasets by filtering poorly aligned regions and focusing on reliable visual-language cues.

← PrevPage 3 of 4Next →