Recognition: unknown
Online Convex Optimization with Stochastic Constraints
read the original abstract
This paper considers online convex optimization (OCO) with stochastic constraints, which generalizes Zinkevich's OCO over a known simple fixed set by introducing multiple stochastic functional constraints that are i.i.d. generated at each round and are disclosed to the decision maker only after the decision is made. This formulation arises naturally when decisions are restricted by stochastic environments or deterministic environments with noisy observations. It also includes many important problems as special cases, such as OCO with long term constraints, stochastic constrained convex optimization, and deterministic constrained convex optimization. To solve this problem, this paper proposes a new algorithm that achieves $O(\sqrt{T})$ expected regret and constraint violations and $O(\sqrt{T}\log(T))$ high probability regret and constraint violations. Experiments on a real-world data center scheduling problem further verify the performance of the new algorithm.
This paper has not been read by Pith yet.
Forward citations
Cited by 2 Pith papers
-
Prox-PEP: A Proximal Partial Exact Penalty Algorithm for Weakly Convex Stochastic Nonlinear Programming
Prox-PEP achieves O(T^{-1/4}) expected oracle complexity for epsilon-KKT stationarity in weakly convex stochastic nonlinear programming via proximal steps, exact penalty, and designed quadratic subproblems.
-
A Quadratic-Approximation-Based Stochastic Approximation Method for Weakly Convex Stochastic Programming
PMQSopt achieves expected O(T^{-1/4}) convergence rates for three epsilon-KKT metrics after T iterations in weakly convex stochastic programming under weak convexity and strict feasibility assumptions.
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.