pith. machine review for the scientific record. sign in

arxiv: 1105.2024 · v2 · submitted 2011-05-10 · ✦ hep-th · hep-ph· math-ph· math.MP

Recognition: unknown

Symbols of One-Loop Integrals From Mixed Tate Motives

Authors on Pith no claims yet
classification ✦ hep-th hep-phmath-phmath.MP
keywords dimensionsintegralintegralsmixedmotivesone-loopsymboltate
0
0 comments X
read the original abstract

We use a result on mixed Tate motives due to Goncharov (arXiv:alg-geom/9601021) to show that the symbol of an arbitrary one-loop 2m-gon integral in 2m dimensions may be read off directly from its Feynman parameterization. The algorithm proceeds via recursion in m seeded by the well-known box integrals in four dimensions. As a simple application of this method we write down the symbol of a three-mass hexagon integral in six dimensions.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 3 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. COMPASS: A Unified Decision-Intelligence System for Navigating Performance Trade-off in HPC

    cs.PF 2026-04 conditional novelty 6.0

    COMPASS formalizes HPC configuration questions as ML tasks on traces, quantifies recommendation trustworthiness, and delivers 65.93% lower average job turnaround time plus 80.93% lower node usage versus prior methods ...

  2. Reliability of Large Language Models for Design Synthesis: An Empirical Study of Variance, Prompt Sensitivity, and Method Scaffolding

    cs.SE 2026-04 unverdicted novelty 5.0

    Preference-based prompting raises LLM adherence to object-oriented design principles in UML generation but leaves substantial output variance and model-specific differences intact.

  3. Locate, Steer, and Improve: A Practical Survey of Actionable Mechanistic Interpretability in Large Language Models

    cs.CL 2026-01 unverdicted novelty 5.0

    The survey organizes mechanistic interpretability techniques into a Locate-Steer-Improve framework to enable actionable improvements in LLM alignment, capability, and efficiency.