pith. machine review for the scientific record. sign in

arxiv: 1607.05447 · v2 · submitted 2016-07-19 · 💻 cs.CV · math.OC

Recognition: unknown

On Differentiating Parameterized Argmin and Argmax Problems with Application to Bi-level Optimization

Authors on Pith no claims yet
classification 💻 cs.CV math.OC
keywords optimizationproblemargmaxargminbi-levelproblemssomedifferentiating
0
0 comments X
read the original abstract

Some recent works in machine learning and computer vision involve the solution of a bi-level optimization problem. Here the solution of a parameterized lower-level problem binds variables that appear in the objective of an upper-level problem. The lower-level problem typically appears as an argmin or argmax optimization problem. Many techniques have been proposed to solve bi-level optimization problems, including gradient descent, which is popular with current end-to-end learning approaches. In this technical report we collect some results on differentiating argmin and argmax optimization problems with and without constraints and provide some insightful motivating examples.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 4 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Constraint-Aware Flow Matching: Decision Aligned End-to-End Training for Constrained Sampling

    cs.LG 2026-05 unverdicted novelty 7.0

    Constraint-Aware Flow Matching integrates constraint projections into the flow matching training objective to align model dynamics with constrained sampling and reduce distributional shift.

  2. Decision-Focused Learning via Tangent-Space Projection of Prediction Error

    cs.LG 2026-05 unverdicted novelty 7.0

    Regret gradients in DFL are the tangent-space projection of prediction error scaled by curvature, enabling efficient direct computation without differentiating through solvers.

  3. Representation-Guided Parameter-Efficient LLM Unlearning

    cs.CL 2026-04 unverdicted novelty 6.0

    REGLU guides LoRA-based unlearning via representation subspaces and orthogonal regularization to outperform prior methods on forget-retain trade-off in LLM benchmarks.

  4. InfoGeo: Information-Theoretic Object-Centric Learning for Cross-View Generalizable UAV Geo-Localization

    cs.CV 2026-05 unverdicted novelty 5.0

    InfoGeo reformulates cross-view geo-localization as an information bottleneck that aligns object-centric structural relations across views while minimizing view-specific noise.