Abstract
We develop foundations for oriented category theory, an extension of
$(\infty,\infty)$-category theory obtained by systematic usage of the Gray
tensor product, in order to study lax phenomena in higher category theory. As
categorical dimension increases, classical category-theoretic concepts
generally prove too rigid or fully break down and must be replaced by oriented
versions, which allow more flexible notions of naturality and coherence.
Oriented category theory provides a framework to address these issues. The main
objects of study are oriented, and their conjugate antioriented, categories,
which are deformations of $(\infty,\infty)$-categories where the various
compositions only commute up to a coherent (anti)oriented interchange law. We
give a geometric description of (anti)oriented categories as sheaves on a
deformation of the simplex category $\Delta$ in which the linear graphs are
weighted by (anti)oriented cubes.
To demonstrate the utility of our theory, we show that the categorical
analogues of fundamental constructions in homotopy theory, such as cylinder and
path objects, join and slice, and suspension and loops, are not functors of
$(\infty, \infty)$-categories, but only of (anti)oriented categories,
generalizing work of Ara, Guetta, and Maltsiniotis in the strict setting. As a
main result we construct an embedding of the theory of
$(\infty,\infty)$-categories into the theory of (anti)oriented categories and
characterize the image, which we call (anti)oriented spaces.
We provide an algebraic description of (anti)oriented spaces as
(anti)oriented categories satisfying a strict (anti)oriented interchange law
and a geometric description as sheaves on suitable categories of (anti)oriented
polytopes, generalizing Grothendieck's philosophy of test categories to higher
categorical dimension and refining Campion's work on lax cubes and suitable
sites.
Squirrel Ai Learning
Abstract
Large Reasoning Models (LRMs) often suffer from the ``over-thinking''
problem, generating unnecessarily long reasoning on simple tasks. Some
strategies have been proposed to mitigate this issue, such as length penalties
or routing mechanisms, but they are typically heuristic and task-specific,
lacking a general framework for adaptive reasoning. In this paper, we present
ARM2, a unified model that adaptively balances reasoning performance and
efficiency across multiple formats through a reinforcement learning framework
augmented with length-aware optimization. Beyond conventional natural language
inference, ARM2 integrates vision understanding, extending its applicability to
multimodal. Moreover, ARM2 integrates executable code into reasoning, enabling
substantial reductions in token cost while preserving task performance compared
to long CoT. Experiments demonstrate that ARM2 achieves performance on par with
traditional reasoning models trained with GRPO, while reducing token usage by
over 70% on average. We further conduct extensive analyses to validate the
effectiveness of ARM2 and the soundness of its design.
AI Insights - ARM2 uses a lightweight RL policy to choose the best reasoning format—CoT, code, or visual inference—based on input complexity.
- It beats state‑of‑the‑art on CSQA, GSM8K, GEO3K while cutting token usage by over 70 %.
- The paper surveys multimodal reasoning, placing ARM2 among vision‑language and code‑augmented models.
- Authors admit very complex reasoning still strains ARM2, calling for more robust, adaptable architectures.
- ARM2 uses LoRA for efficient tuning and VERL to ground text in images.
- Recommended reading: “Multimodal Reasoning: A Survey” plus VERL, LoRA, CSQA, GSM8K, GEO3K papers.
- The adaptive policy adds computational overhead, a trade‑off the authors suggest can be cut with smarter RL.