Hi!

Your personalized paper recommendations for 05 to 09 January, 2026.
Massachusetts Institute of Technology MIT
Rate paper: 👍 👎 ♥ Save
Paper visualization
Rate image: 👍 👎
Abstract
Scientific inquiry requires systems-level reasoning that integrates heterogeneous experimental data, cross-domain knowledge, and mechanistic evidence into coherent explanations. While Large Language Models (LLMs) offer inferential capabilities, they often depend on retrieval-augmented contexts that lack structural depth. Traditional Knowledge Graphs (KGs) attempt to bridge this gap, yet their pairwise constraints fail to capture the irreducible higher-order interactions that govern emergent physical behavior. To address this, we introduce a methodology for constructing hypergraph-based knowledge representations that faithfully encode multi-entity relationships. Applied to a corpus of ~1,100 manuscripts on biocomposite scaffolds, our framework constructs a global hypergraph of 161,172 nodes and 320,201 hyperedges, revealing a scale-free topology (power law exponent ~1.23) organized around highly connected conceptual hubs. This representation prevents the combinatorial explosion typical of pairwise expansions and explicitly preserves the co-occurrence context of scientific formulations. We further demonstrate that equipping agentic systems with hypergraph traversal tools, specifically using node-intersection constraints, enables them to bridge semantically distant concepts. By exploiting these higher-order pathways, the system successfully generates grounded mechanistic hypotheses for novel composite materials, such as linking cerium oxide to PCL scaffolds via chitosan intermediates. This work establishes a "teacherless" agentic reasoning system where hypergraph topology acts as a verifiable guardrail, accelerating scientific discovery by uncovering relationships obscured by traditional graph methods.
Why we are recommending this paper?
Due to your Interest in Knowledge Graphs

This paper from MIT explores knowledge representation, aligning with your interest in knowledge graphs and taxonomy of products. It investigates how systems can integrate diverse knowledge for scientific reasoning, a key area of your focus.
Heidelberg University
Rate paper: 👍 👎 ♥ Save
Abstract
As large language models (LLMs) are more frequently used in retrieval-augmented generation pipelines, it is increasingly relevant to study their behavior under knowledge conflicts. Thus far, the role of the source of the retrieved information has gone unexamined. We address this gap with a novel framework to investigate how source preferences affect LLM resolution of inter-context knowledge conflicts in English, motivated by interdisciplinary research on credibility. With a comprehensive, tightly-controlled evaluation of 13 open-weight LLMs, we find that LLMs prefer institutionally-corroborated information (e.g., government or newspaper sources) over information from people and social media. However, these source preferences can be reversed by simply repeating information from less credible sources. To mitigate repetition effects and maintain consistent preferences, we propose a novel method that reduces repetition bias by up to 99.8%, while also maintaining at least 88.8% of original preferences. We release all data and code to encourage future work on credibility and source preferences in knowledge-intensive NLP.
Why we are recommending this paper?
Due to your Interest in Knowledge Management

Given your interest in knowledge graphs and how they are used, this Heidelberg University paper directly addresses the source selection process within LLMs. Understanding how LLMs prioritize information is crucial for effective knowledge management.
Bochum University of Applied Science
Rate paper: 👍 👎 ♥ Save
Paper visualization
Rate image: 👍 👎
Abstract
We present a regularization-based approach for continual learning (CL) of fixed capacity convolutional neural networks (CNN) that does not suffer from the problem of catastrophic forgetting when learning multiple tasks sequentially. This method referred to as Group and Exclusive Sparsity based Continual Learning (GESCL) avoids forgetting of previous tasks by ensuring the stability of the CNN via a stability regularization term, which prevents filters detected as important for past tasks to deviate too much when learning a new task. On top of that, GESCL makes the network plastic via a plasticity regularization term that leverage the over-parameterization of CNNs to efficiently sparsify the network and tunes unimportant filters making them relevant for future tasks. Doing so, GESCL deals with significantly less parameters and computation compared to CL approaches that either dynamically expand the network or memorize past tasks' data. Experiments on popular CL vision benchmarks show that GESCL leads to significant improvements over state-of-the-art method in terms of overall CL performance, as measured by classification accuracy as well as in terms of avoiding catastrophic forgetting.
Why we are recommending this paper?
Due to your Interest in Continual Generalized Category Discovery

This paper from Bochum University of Applied Science tackles continual learning, a relevant area for your interest in product categorization and knowledge graphs. It offers a method for adapting models to new data, aligning with your need for generalized category discovery.
Dalian University of Technology
Rate paper: 👍 👎 ♥ Save
Paper visualization
Rate image: 👍 👎
Abstract
Continual learning aims to enable models to acquire new knowledge while retaining previously learned information. Prompt-based methods have shown remarkable performance in this domain; however, they typically rely on key-value pairing, which can introduce inter-task interference and hinder scalability. To overcome these limitations, we propose a novel approach employing task-specific Prompt-Prototype (ProP), thereby eliminating the need for key-value pairs. In our method, task-specific prompts facilitate more effective feature learning for the current task, while corresponding prototypes capture the representative features of the input. During inference, predictions are generated by binding each task-specific prompt with its associated prototype. Additionally, we introduce regularization constraints during prompt initialization to penalize excessively large values, thereby enhancing stability. Experiments on several widely used datasets demonstrate the effectiveness of the proposed method. In contrast to mainstream prompt-based approaches, our framework removes the dependency on key-value pairs, offering a fresh perspective for future continual learning research.
Why we are recommending this paper?
Due to your Interest in Continual Generalized Category Discovery

This paper from Dalian University of Technology focuses on continual learning techniques, directly addressing your interest in product categorization and knowledge management. The key-value free approach is particularly relevant for scalable knowledge graph construction.
University of Illinois UrbanaChampaign
Rate paper: 👍 👎 ♥ Save
Paper visualization
Rate image: 👍 👎
Abstract
Let $H$ be a fixed graph whose edges are colored red and blue and let $β\in [0,1]$. Let $I(H, β)$ be the (asymptotically normalized) maximum number of copies of $H$ in a large red/blue edge-colored complete graph $G$, where the density of red edges in $G$ is $β$. This refines the problem of determining the semi-inducibility of $H$, which is itself a generalization of the classical question of determining the inducibility of $H$. The function $I(H, β)$ for $β\in [0,1]$ was not known for any graph $H$ on more than three vertices, except when $H$ is a monochromatic clique (Kruskal-Katona) or a monochromatic star (Reiher-Wagner). We obtain sharp results for some four and five vertex graphs, addressing several recent questions posed by various authors. We also obtain some general results for trees and stars. Many open problems remain.
Why we are recommending this paper?
Due to your Interest in Graphs for Products

This paper from the University of Illinois UrbanaChampaign investigates graph structures, a core component of your interest in knowledge graphs and ontology for products. Understanding graph inducibility is fundamental to building effective knowledge representations.
Rate paper: 👍 👎 ♥ Save
Abstract
This volume contains the post-proceedings of the Sixteenth International Workshop on Graph Computation Models (GCM 2025). The workshops took place in Koblenz, Germany on June 10 as part of STAF (Software Technologies: Applications and Foundations). Graphs are common mathematical structures that are visual and intuitive. They constitute a natural and seamless way for system modeling in science, engineering, and beyond, including computer science, biology, and business process modeling. Graph computation models constitute a class of very high-level models where graphs are first-class citizens. The aim of the International GCM Workshop series is to bring together researchers interested in all aspects of computation models based on graphs and graph transformation. It promotes the cross-fertilizing exchange of ideas and experiences among senior and young researchers from the different communities interested in the foundations, applications, and implementations of graph computation models and related areas.
Why we are recommending this paper?
Due to your Interest in Graphs for Products
Rate paper: 👍 👎 ♥ Save
Abstract
In this work we present an intuitive construction of the quantum logical axiomatic system provided by George Mackey. The goal of this work is a detailed discussion of the results from the paper 'Physical justification for using the tensor product to describe two quantum systems as one joint system' [1] published by Diederik Aerts and Ingrid Daubechies. This means that we want to show how certain composed physical systems from classical and quantum mechanics should be described logically. To reach this goal, we will, like in [1], discuss a special class of axiomatically defined composed physical systems. With the help of certain results from lattice and c-morphism theory (see [2] and [23]), we will present a detailed proof of the statement, that in the quantum mechanical case, a composed physical system must be described via a tensor product space.
Why we are recommending this paper?
Due to your Interest in Ontology for Products
University of Copenhagen
Rate paper: 👍 👎 ♥ Save
Abstract
Duality, not monism, constitutes the hermeneutic lens that characterizes the original Copenhagen interpretation of Quantum Mechanics. Therefore, evoking the principles of correspondence and complementarity, in this work we re assert a dual-aspect reading of quantum theory, structured through a multi-perspective schema encompassing its ontological, analytical, epistemological, causal, and information dimensions. We then show how this schema dissolves the so-called measurement problem, along with the associated knowledge-information and macro-micro dichotomies, issues historically raised within later monistic or universalist philosophical settings that ultimately depart from the traditional Copenhagen spirit.
Why we are recommending this paper?
Due to your Interest in MECE Mutually Exclusive, Collectively Exhaustive.
Hebei University of Technology
Rate paper: 👍 👎 ♥ Save
Abstract
Fine-Grained Domain Generalization (FGDG) presents greater challenges than conventional domain generalization due to the subtle inter-class differences and relatively pronounced intra-class variations inherent in fine-grained recognition tasks. Under domain shifts, the model becomes overly sensitive to fine-grained cues, leading to the suppression of critical features and a significant drop in performance. Cognitive studies suggest that humans classify objects by leveraging both common and specific attributes, enabling accurate differentiation between fine-grained categories. However, current deep learning models have yet to incorporate this mechanism effectively. Inspired by this mechanism, we propose Concept-Feature Structuralized Generalization (CFSG). This model explicitly disentangles both the concept and feature spaces into three structured components: common, specific, and confounding segments. To mitigate the adverse effects of varying degrees of distribution shift, we introduce an adaptive mechanism that dynamically adjusts the proportions of common, specific, and confounding components. In the final prediction, explicit weights are assigned to each pair of components. Extensive experiments on three single-source benchmark datasets demonstrate that CFSG achieves an average performance improvement of 9.87% over baseline models and outperforms existing state-of-the-art methods by an average of 3.08%. Additionally, explainability analysis validates that CFSG effectively integrates multi-granularity structured knowledge and confirms that feature structuralization facilitates the emergence of concept structuralization.
Why we are recommending this paper?
Due to your Interest in Product Categorization
Universidad de la Repblica
Rate paper: 👍 👎 ♥ Save
Abstract
We propose Concept Tokens, a lightweight method that adds a new special token to a pretrained LLM and learns only its embedding from multiple natural language definitions of a target concept, where occurrences of the concept are replaced by the new token. The LLM is kept frozen and the embedding is optimized with the standard language-modeling objective. We evaluate Concept Tokens in three settings. First, we study hallucinations in closed-book question answering on HotpotQA and find a directional effect: negating the hallucination token reduces hallucinated answers mainly by increasing abstentions, whereas asserting it increases hallucinations and lowers precision. Second, we induce recasting, a pedagogical feedback strategy for second language teaching, and observe the same directional effect. Moreover, compared to providing the full definitional corpus in-context, concept tokens better preserve compliance with other instructions (e.g., asking follow-up questions). Finally, we include a qualitative study with the Eiffel Tower and a fictional "Austral Tower" to illustrate what information the learned embeddings capture and where their limitations emerge. Overall, Concept Tokens provide a compact control signal learned from definitions that can steer behavior in frozen LLMs.
Why we are recommending this paper?
Due to your Interest in Product Categorization
Brock University
Rate paper: 👍 👎 ♥ Save
Paper visualization
Rate image: 👍 👎
AI Insights
  • The framework's graph grounding is the most influential factor, with a mean gain of 21.5% across all model-interval combinations. [3]
  • Stronger models still improve even when using explicit structure, with larger models gaining around +22% at 1s and +29% at 100s. [3]
  • Reasoning-tuned models exploit the KG best, with o4-mini showing the highest incremental benefit. [3]
  • KG (Knowledge Graph) - A graph-based data structure that stores entities, relationships, and attributes. [3]
  • The accuracy of TAAF responses varies across different LLM backends, but o4-mini with TAAF settings tops the chart at every interval. [2]
  • TAAF (Trace Abstraction and Analysis Framework) synergizes knowledge graphs and LLMs to improve accuracy in answering trace-related queries. [1]
Abstract
Execution traces are a critical source of information for understanding, debugging, and optimizing complex software systems. However, traces from OS kernels or large-scale applications like Chrome or MySQL are massive and difficult to analyze. Existing tools rely on predefined analyses, and custom insights often require writing domain-specific scripts, which is an error-prone and time-consuming task. This paper introduces TAAF (Trace Abstraction and Analysis Framework), a novel approach that combines time-indexing, knowledge graphs (KGs), and large language models (LLMs) to transform raw trace data into actionable insights. TAAF constructs a time-indexed KG from trace events to capture relationships among entities such as threads, CPUs, and system resources. An LLM then interprets query-specific subgraphs to answer natural-language questions, reducing the need for manual inspection and deep system expertise. To evaluate TAAF, we introduce TraceQA-100, a benchmark of 100 questions grounded in real kernel traces. Experiments across three LLMs and multiple temporal settings show that TAAF improves answer accuracy by up to 31.2%, particularly in multi-hop and causal reasoning tasks. We further analyze where graph-grounded reasoning helps and where limitations remain, offering a foundation for next-generation trace analysis tools.
Why we are recommending this paper?
Due to your Interest in Knowledge Graphs

Interests not found

We did not find any papers that match the below interests. Try other terms also consider if the content exists in arxiv.org.
  • Taxonomy of Products
You can edit or add more interests any time.