arXiv:2604.01235v1 Announce Type: new Abstract: Structured LLM routing is often treated as a prompt-engineering problem. We argue that it is, more fundamentally, a systems-level burden-allocation problem. As large language models (LLMs) become core control components in agentic AI systems, reliable structured routing must balance correctness, latency, and implementation cost under real deployment constraints. We show […]
Domain-constrained knowledge representation: A modal framework
arXiv:2604.01770v1 Announce Type: new Abstract: Knowledge graphs store large numbers of relations efficiently, but they remain weak at representing a quieter difficulty: the meaning of a concept often shifts with the domain in which it is used. A triple such as Apple, instance-of, Company may be acceptable in one setting while being misleading or unusable […]
Lifting Unlabeled Internet-level Data for 3D Scene Understanding
arXiv:2604.01907v1 Announce Type: cross Abstract: Annotated 3D scene data is scarce and expensive to acquire, while abundant unlabeled videos are readily available on the internet. In this paper, we demonstrate that carefully designed data engines can leverage web-curated, unlabeled videos to automatically generate training data, to facilitate end-to-end models in 3D scene understanding alongside human-annotated […]
Bayesian Elicitation with LLMs: Model Size Helps, Extra “Reasoning” Doesn’t Always
arXiv:2604.01896v1 Announce Type: new Abstract: Large language models (LLMs) have been proposed as alternatives to human experts for estimating unknown quantities with associated uncertainty, a process known as Bayesian elicitation. We test this by asking eleven LLMs to estimate population statistics, such as health prevalence rates, personality trait distributions, and labor market figures, and to […]
MOON3.0: Reasoning-aware Multimodal Representation Learning for E-commerce Product Understanding
arXiv:2604.00513v2 Announce Type: replace-cross Abstract: With the rapid growth of e-commerce, exploring general representations rather than task-specific ones has attracted increasing attention. Although recent multimodal large language models (MLLMs) have driven significant progress in product understanding, they are typically employed as feature extractors that implicitly encode product information into global embeddings, thereby limiting their ability […]
How and why does deep ensemble coupled with transfer learning increase performance in bipolar disorder and schizophrenia classification?
arXiv:2604.02002v1 Announce Type: new Abstract: Transfer learning (TL) and deep ensemble learning (DE) have recently been shown to outperform simple machine learning in classifying psychiatric disorders. However, there is still a lack of understanding as to why that is. This paper aims to understand how and why DE and TL reduce the variability of single-subject […]
Combating Data Laundering in LLM Training
arXiv:2604.01904v1 Announce Type: cross Abstract: Data rights owners can detect unauthorized data use in large language model (LLM) training by querying with proprietary samples. Often, superior performance (e.g., higher confidence or lower loss) on a sample relative to the untrained data implies it was part of the training corpus, as LLMs tend to perform better […]
ByteRover: Agent-Native Memory Through LLM-Curated Hierarchical Context
arXiv:2604.01599v1 Announce Type: new Abstract: Memory-Augmented Generation (MAG) extends large language models with external memory to support long-context reasoning, but existing approaches universally treat memory as an external service that agents call into, delegating storage to separate pipelines of chunking, embedding, and graph extraction. This architectural separation means the system that stores knowledge does not […]
Hybrid Energy-Based Models for Physical AI: Provably Stable Identification of Port-Hamiltonian Dynamics
arXiv:2604.00277v2 Announce Type: replace-cross Abstract: Energy-based models (EBMs) implement inference as gradient descent on a learned Lyapunov function, yielding interpretable, structure-preserving alternatives to black-box neural ODEs and aligning naturally with physical AI. Yet their use in system identification remains limited, and existing architectures lack formal stability guarantees that globally preclude unstable modes. We address this […]
GraphWalk: Enabling Reasoning in Large Language Models through Tool-Based Graph Navigation
arXiv:2604.01610v1 Announce Type: new Abstract: The use of knowledge graphs for grounding agents in real-world Q&A applications has become increasingly common. Answering complex queries often requires multi-hop reasoning and the ability to navigate vast relational structures. Standard approaches rely on prompting techniques that steer large language models to reason over raw graph context, or retrieval-augmented […]
Robust Graph Representation Learning via Adaptive Spectral Contrast
arXiv:2604.01878v1 Announce Type: cross Abstract: Spectral graph contrastive learning has emerged as a unified paradigm for handling both homophilic and heterophilic graphs by leveraging high-frequency components. However, we identify a fundamental spectral dilemma: while high-frequency signals are indispensable for encoding heterophily, our theoretical analysis proves they exhibit significantly higher variance under spectrally concentrated perturbations. We […]
ThinknCheck: Grounded Claim Verification with Compact, Reasoning-Driven, and Interpretable Models
arXiv:2604.01652v1 Announce Type: new Abstract: We present ThinknCheck, a 1B-parameter verifier for grounded claim verification that first produces a short, structured rationale and then a binary verdict. We construct LLMAggreFact-Think, a 24.1k reasoning-augmented training set derived from LLMAggreFact, and fine-tune a 4-bit Gemma3 model to follow this format. On LLMAggreFact, ThinknCheck attains 78.1 balanced accuracy […]