y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#semantic-analysis News & Analysis

15 articles tagged with #semantic-analysis. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

15 articles
AINeutralarXiv โ€“ CS AI ยท Mar 177/10
๐Ÿง 

Distributional Semantics Tracing: A Framework for Explaining Hallucinations in Large Language Models

Researchers introduce Distributional Semantics Tracing (DST), a new framework for explaining hallucinations in large language models by tracking how semantic representations drift across neural network layers. The method reveals that hallucinations occur when models are pulled toward contextually inconsistent concepts based on training correlations rather than actual prompt context.

AINeutralarXiv โ€“ CS AI ยท Mar 57/10
๐Ÿง 

Escaping the BLEU Trap: A Signal-Grounded Framework with Decoupled Semantic Guidance for EEG-to-Text Decoding

Researchers propose SemKey, a novel framework that addresses key limitations in EEG-to-text decoding by preventing hallucinations and improving semantic fidelity through decoupled guidance objectives. The system redesigns neural encoder-LLM interaction and introduces new evaluation metrics beyond BLEU scores to achieve state-of-the-art performance in brain-computer interfaces.

AIBullisharXiv โ€“ CS AI ยท Mar 46/103
๐Ÿง 

LLM-MLFFN: Multi-Level Autonomous Driving Behavior Feature Fusion via Large Language Model

Researchers developed LLM-MLFFN, a new framework combining large language models with multi-level feature fusion to classify autonomous vehicle driving behaviors. The system achieves over 94% accuracy on the Waymo dataset by integrating numerical driving data with semantic features extracted through LLMs.

AINeutralarXiv โ€“ CS AI ยท 1d ago6/10
๐Ÿง 

LLM as Attention-Informed NTM and Topic Modeling as long-input Generation: Interpretability and long-Context Capability

Researchers propose a novel framework treating Large Language Models as attention-informed Neural Topic Models, enabling interpretable topic extraction from documents. The approach combines white-box interpretability analysis with black-box long-context LLM capabilities, demonstrating competitive performance on topic modeling tasks while maintaining semantic clarity.

AINeutralarXiv โ€“ CS AI ยท 2d ago6/10
๐Ÿง 

Agent Mentor: Framing Agent Knowledge through Semantic Trajectory Analysis

Researchers introduce Agent Mentor, an open-source analytics pipeline that monitors and automatically improves AI agent behavior by analyzing execution logs and iteratively refining system prompts with corrective instructions. The framework addresses variability in large language model-based agent performance caused by ambiguous prompt formulations, demonstrating consistent accuracy improvements across multiple configurations.

AINeutralarXiv โ€“ CS AI ยท 2d ago6/10
๐Ÿง 

Beyond Statistical Co-occurrence: Unlocking Intrinsic Semantics for Tabular Data Clustering

Researchers introduce TagCC, a novel deep clustering framework that combines Large Language Models with contrastive learning to enhance tabular data analysis by incorporating semantic knowledge from feature names and values. The approach bridges the gap between statistical co-occurrence patterns and intrinsic semantic understanding, demonstrating significant performance improvements over existing methods in finance and healthcare applications.

AINeutralarXiv โ€“ CS AI ยท Mar 66/10
๐Ÿง 

Simulating Meaning, Nevermore! Introducing ICR: A Semiotic-Hermeneutic Metric for Evaluating Meaning in LLM Text Summaries

Researchers introduce ICR (Inductive Conceptual Rating), a new qualitative metric for evaluating meaning in large language model text summaries that goes beyond simple word similarity. The study found that while LLMs achieve high linguistic similarity to human outputs, they significantly underperform in semantic accuracy and capturing contextual meanings.

AINeutralarXiv โ€“ CS AI ยท Mar 45/103
๐Ÿง 

FinTexTS: Financial Text-Paired Time-Series Dataset via Semantic-Based and Multi-Level Pairing

Researchers have developed FinTexTS, a new large-scale dataset that pairs financial news with stock price data using semantic matching and multi-level categorization. The framework uses embedding-based matching and LLMs to classify news into four levels (macro, sector, related company, and target company) for improved stock price forecasting accuracy.

AIBullisharXiv โ€“ CS AI ยท Mar 26/1014
๐Ÿง 

Latent Self-Consistency for Reliable Majority-Set Selection in Short- and Long-Answer Reasoning

Researchers introduce Latent Self-Consistency (LSC), a new method for improving Large Language Model output reliability across both short and long-form reasoning tasks. LSC uses learnable token embeddings to select semantically consistent responses with only 0.9% computational overhead, outperforming existing consistency methods like Self-Consistency and Universal Self-Consistency.

AINeutralarXiv โ€“ CS AI ยท Feb 275/106
๐Ÿง 

Enriching Taxonomies Using Large Language Models

Researchers have developed Taxoria, a new taxonomy enrichment pipeline that uses Large Language Models to enhance existing taxonomies by proposing, validating, and integrating new nodes. The system addresses limitations in current taxonomies such as limited coverage and outdated information while including hallucination mitigation and provenance tracking.

AIBullisharXiv โ€“ CS AI ยท Feb 276/106
๐Ÿง 

Temporal Sparse Autoencoders: Leveraging the Sequential Nature of Language for Interpretability

Researchers introduce Temporal Sparse Autoencoders (T-SAEs), a new method that improves AI model interpretability by incorporating temporal structure of language through contrastive loss. The technique enables better separation of semantic from syntactic features and recovers smoother, more coherent semantic concepts without sacrificing reconstruction quality.

AINeutralarXiv โ€“ CS AI ยท Mar 124/10
๐Ÿง 

Automated evaluation of LLMs for effective machine translation of Mandarin Chinese to English

Researchers developed an automated framework to evaluate Large Language Models' effectiveness in translating Mandarin Chinese to English, comparing GPT-4, GPT-4o, and DeepSeek against Google Translate. While LLMs performed well on news translation, they showed varying results with literary texts, with DeepSeek excelling at cultural subtleties and GPT-4o/DeepSeek better at semantic conservation.

๐Ÿข Meta๐Ÿง  GPT-4
AINeutralarXiv โ€“ CS AI ยท Feb 274/104
๐Ÿง 

What Makes an Ideal Quote? Recommending "Unexpected yet Rational" Quotations via Novelty

Researchers developed NovelQR, an AI framework for recommending quotations that are 'unexpected yet rational' by prioritizing novelty over surface-level topical relevance. The system uses a generative label agent to interpret deep meanings and a novelty estimator to rerank candidates, showing superior performance in human evaluations across bilingual datasets.

AINeutralarXiv โ€“ CS AI ยท Mar 34/107
๐Ÿง 

A Case Study on Concept Induction for Neuron-Level Interpretability in CNN

Researchers successfully applied a Concept Induction framework for neural network interpretability to the SUN2012 dataset, demonstrating the method's broader applicability beyond the original ADE20K dataset. The study assigns interpretable semantic labels to hidden neurons in CNNs and validates them through statistical testing and web-sourced images.

AINeutralarXiv โ€“ CS AI ยท Mar 24/107
๐Ÿง 

Language Models as Messengers: Enhancing Message Passing in Heterophilic Graph Learning

Researchers propose LEMP4HG, a new language model-enhanced approach for improving graph neural networks on heterophilic graphs where connected nodes have different characteristics. The method leverages language models to better understand semantic relationships between text-attributed nodes, outperforming existing methods while maintaining efficiency through selective message enhancement.