15 articles tagged with #semantic-analysis. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AINeutralarXiv โ CS AI ยท Mar 177/10
๐ง Researchers introduce Distributional Semantics Tracing (DST), a new framework for explaining hallucinations in large language models by tracking how semantic representations drift across neural network layers. The method reveals that hallucinations occur when models are pulled toward contextually inconsistent concepts based on training correlations rather than actual prompt context.
AINeutralarXiv โ CS AI ยท Mar 57/10
๐ง Researchers propose SemKey, a novel framework that addresses key limitations in EEG-to-text decoding by preventing hallucinations and improving semantic fidelity through decoupled guidance objectives. The system redesigns neural encoder-LLM interaction and introduces new evaluation metrics beyond BLEU scores to achieve state-of-the-art performance in brain-computer interfaces.
AIBullisharXiv โ CS AI ยท Mar 46/103
๐ง Researchers developed LLM-MLFFN, a new framework combining large language models with multi-level feature fusion to classify autonomous vehicle driving behaviors. The system achieves over 94% accuracy on the Waymo dataset by integrating numerical driving data with semantic features extracted through LLMs.
AINeutralarXiv โ CS AI ยท 1d ago6/10
๐ง Researchers propose a novel framework treating Large Language Models as attention-informed Neural Topic Models, enabling interpretable topic extraction from documents. The approach combines white-box interpretability analysis with black-box long-context LLM capabilities, demonstrating competitive performance on topic modeling tasks while maintaining semantic clarity.
AINeutralarXiv โ CS AI ยท 2d ago6/10
๐ง Researchers introduce Agent Mentor, an open-source analytics pipeline that monitors and automatically improves AI agent behavior by analyzing execution logs and iteratively refining system prompts with corrective instructions. The framework addresses variability in large language model-based agent performance caused by ambiguous prompt formulations, demonstrating consistent accuracy improvements across multiple configurations.
AINeutralarXiv โ CS AI ยท 2d ago6/10
๐ง Researchers introduce TagCC, a novel deep clustering framework that combines Large Language Models with contrastive learning to enhance tabular data analysis by incorporating semantic knowledge from feature names and values. The approach bridges the gap between statistical co-occurrence patterns and intrinsic semantic understanding, demonstrating significant performance improvements over existing methods in finance and healthcare applications.
AINeutralarXiv โ CS AI ยท Mar 66/10
๐ง Researchers introduce ICR (Inductive Conceptual Rating), a new qualitative metric for evaluating meaning in large language model text summaries that goes beyond simple word similarity. The study found that while LLMs achieve high linguistic similarity to human outputs, they significantly underperform in semantic accuracy and capturing contextual meanings.
AINeutralarXiv โ CS AI ยท Mar 45/103
๐ง Researchers have developed FinTexTS, a new large-scale dataset that pairs financial news with stock price data using semantic matching and multi-level categorization. The framework uses embedding-based matching and LLMs to classify news into four levels (macro, sector, related company, and target company) for improved stock price forecasting accuracy.
AIBullisharXiv โ CS AI ยท Mar 26/1014
๐ง Researchers introduce Latent Self-Consistency (LSC), a new method for improving Large Language Model output reliability across both short and long-form reasoning tasks. LSC uses learnable token embeddings to select semantically consistent responses with only 0.9% computational overhead, outperforming existing consistency methods like Self-Consistency and Universal Self-Consistency.
AINeutralarXiv โ CS AI ยท Feb 275/106
๐ง Researchers have developed Taxoria, a new taxonomy enrichment pipeline that uses Large Language Models to enhance existing taxonomies by proposing, validating, and integrating new nodes. The system addresses limitations in current taxonomies such as limited coverage and outdated information while including hallucination mitigation and provenance tracking.
AIBullisharXiv โ CS AI ยท Feb 276/106
๐ง Researchers introduce Temporal Sparse Autoencoders (T-SAEs), a new method that improves AI model interpretability by incorporating temporal structure of language through contrastive loss. The technique enables better separation of semantic from syntactic features and recovers smoother, more coherent semantic concepts without sacrificing reconstruction quality.
AINeutralarXiv โ CS AI ยท Mar 124/10
๐ง Researchers developed an automated framework to evaluate Large Language Models' effectiveness in translating Mandarin Chinese to English, comparing GPT-4, GPT-4o, and DeepSeek against Google Translate. While LLMs performed well on news translation, they showed varying results with literary texts, with DeepSeek excelling at cultural subtleties and GPT-4o/DeepSeek better at semantic conservation.
๐ข Meta๐ง GPT-4
AINeutralarXiv โ CS AI ยท Feb 274/104
๐ง Researchers developed NovelQR, an AI framework for recommending quotations that are 'unexpected yet rational' by prioritizing novelty over surface-level topical relevance. The system uses a generative label agent to interpret deep meanings and a novelty estimator to rerank candidates, showing superior performance in human evaluations across bilingual datasets.
AINeutralarXiv โ CS AI ยท Mar 34/107
๐ง Researchers successfully applied a Concept Induction framework for neural network interpretability to the SUN2012 dataset, demonstrating the method's broader applicability beyond the original ADE20K dataset. The study assigns interpretable semantic labels to hidden neurons in CNNs and validates them through statistical testing and web-sourced images.
AINeutralarXiv โ CS AI ยท Mar 24/107
๐ง Researchers propose LEMP4HG, a new language model-enhanced approach for improving graph neural networks on heterophilic graphs where connected nodes have different characteristics. The method leverages language models to better understand semantic relationships between text-attributed nodes, outperforming existing methods while maintaining efficiency through selective message enhancement.