Abstract 1 Introduction 2 Revisiting SEVER 3 Optimal Rates for Robust SCO under Weak Distributional Assumptions 4 Projected Gradient Descent with Robust Gradient Estimator 5 Conclusion and Future Work References Appendix A Proof of Theorem 11 Appendix B Analysis of Convolutional Smoothing for Nonsmooth but Lipschitz Population Risk Appendix C Analysis of Algorithm 3 Appendix D Projected Biased Gradient Descent Appendix E Discussions on the Assumptions Appendix F An exponential time algorithm that achieves the minimax-optimal excess risk bound without logd factor Appendix G Dealing with Unknown 𝝈

Optimal Rates for Robust Stochastic Convex Optimization

Changyu Gao University of Wisconsin-Madison, Madison, WI, USA Andrew Lowy University of Wisconsin-Madison, Madison, WI, USA Xingyu Zhou Wayne State University, Detroit, MI, USA Stephen J. Wright University of Wisconsin-Madison, Madison, WI, USA
Abstract

Machine learning algorithms in high-dimensional settings are highly susceptible to the influence of even a small fraction of structured outliers, making robust optimization techniques essential. In particular, within the ϵ-contamination model, where an adversary can inspect and replace up to an ϵ-fraction of the samples, a fundamental open problem is determining the optimal rates for robust stochastic convex optimization (SCO) under such contamination. We develop novel algorithms that achieve minimax-optimal excess risk (up to logarithmic factors) under the ϵ-contamination model. Our approach improves over existing algorithms, which are not only suboptimal but also require stringent assumptions, including Lipschitz continuity and smoothness of individual sample functions. By contrast, our optimal algorithms do not require these stringent assumptions, assuming only population-level smoothness of the loss. Moreover, our algorithms can be adapted to handle the case in which the covariance parameter is unknown, and can be extended to nonsmooth population risks via convolutional smoothing. We complement our algorithmic developments with a tight information-theoretic lower bound for robust SCO.

Keywords and phrases:
Adversarial Robustness, Machine Learning, Optimization Algorithms, Robust Optimization, Stochastic Convex Optimization
Copyright and License:
[Uncaptioned image] © Changyu Gao, Andrew Lowy, Xingyu Zhou, and Stephen J. Wright; licensed under Creative Commons License CC-BY 4.0
2012 ACM Subject Classification:
Theory of computation Mathematical optimization
Related Version:
Full Version: https://arxiv.org/abs/2412.11003
Acknowledgements:
Changyu would like to thank Shuyao Li for helpful discussions.
Funding:
Research of CG, AL, and SW was supported by NSF Awards 2023239 and 2224213 and AFOSR Award FA9550-21-1-0084. XZ is supported in part by NSF CNS-2153220 and CNS-2312835.
Editors:
Mark Bun

1 Introduction

Machine learning models are increasingly deployed in security-critical applications, yet they remain vulnerable to data manipulation. A particular threat is data poisoning, where adversaries deliberately insert malicious points into training data to degrade model performance [1]. Even in non-adversarial settings, naturally occurring outliers can significantly impact learning algorithms, especially in high-dimensional settings. These challenges motivate our study of optimization algorithms for training machine learning models in the presence of outliers, both natural and adversarial.

Motivation for our work traces to Tukey’s pioneering research on robust estimation [17]. Recent breakthroughs have produced efficient algorithms for high-dimensional robust estimation under the ϵ-contamination model, where an adversary can arbitrarily replace up to an ϵ-fraction of the samples. Notable advances include polynomial-time algorithms for robust mean estimation in high dimensions [5, 3]. See [6] for a comprehensive survey of recent developments in high-dimensional robust estimation.

These developments in robust estimation naturally lead to a fundamental question: Can we solve stochastic optimization problems, under the ϵ-contamination model? Stochastic optimization is used in machine learning to find the parameter that minimizes the population risk using training samples. We focus specifically on robust stochastic optimization with convex objective functions whose gradients exhibit bounded covariance, a standard assumption in robust mean estimation [7]. While our goal aligns with the classical use of stochastic convex optimization in minimizing population risk, the presence of adversarial contamination introduces significant new challenges.

Prior research in robust optimization has concentrated primarily on narrow domains. One line of work focuses on robust linear regression [13, 8, 2]. While [12, 16] have explored general problems, they focus on robust regression. To our best knowledge, SEVER [4] is the only work that considers general stochastic optimization problems. However, this approach has several limitations that restrict the applicability of SEVER. First, it focuses only on achieving dimension-independent error due to corruption, with only a suboptimal sample complexity. Second, the results for SEVER depend on several stringent assumptions, including Lipschitzness and smoothness conditions on individual sample functions. Because of these limitations, optimal excess risk bounds for robust stochastic convex optimization, and under what conditions they can be achieved, remain unknown.

In this work, we develop efficient algorithms for robust stochastic convex optimization that achieve optimal excess risk bounds (up to logarithmic factors) under the ϵ-contamination model. Notably, Algorithm 1 assumes only the smoothness of the population risk. Moreover, we prove a matching lower bound to show the minimax-optimality of our algorithms.

Due to space limitations, we omit some details and proofs. These appear in the full version, available at https://arxiv.org/abs/2412.11003.

1.1 Problem Setup and Motivation

Notation.

For a vector vd, v denotes the 2 norm of v. For a matrix Ad×d, A denotes the spectral norm of A. For symmetric matrices A and B, we write AB if BA is positive semidefinite (PSD). We use O~ and Ω~ to hide logarithmic factors in our bounds.

Let 𝒲d be a closed convex set. Consider a distribution p over functions f:𝒲. Stochastic optimization aims to find a parameter vector w𝒲 minimizing the population risk f¯(w):=𝔼fp[f(w)]. For example, function f can take the form of a loss function fx(w) dependent on the data point x, and the data distribution on x induces the function distribution p. In robust stochastic optimization, some data samples may be corrupted. Following [4], we adopt the strong ϵ-contamination model, which allows the adversary to replace up to ϵ fraction of samples.

Definition 1 (ϵ-contamination model).

Given ϵ>0 and a distribution p over functions f:𝒲, data is generated as follows: first, n clean samples f1,,fn are drawn from p. An adversary is then permitted to examine the samples and replace up to ϵn of them with arbitrary samples. The algorithm is subsequently provided with this modified set of functions, which we refer to as ϵ-corrupted samples (with respect to p).

This model is strictly stronger than the Huber contamination model [11], in which the samples are drawn from a mixture of the clean and adversarial distributions of the form p=(1ϵ)p+ϵq, where p is the clean distribution and q is the adversarial distribution.

Our objective is to develop an efficient algorithm that minimizes the population risk f¯(w), even when the data is ϵ-corrupted. The following is assumed throughout the paper.

Assumption 2.
  1. 1.

    𝒲d is a compact convex set with diameter D, that is, supw,w𝒲wwD.

  2. 2.

    f is differentiable almost surely. The population risk f¯(w) is convex.

  3. 3.

    The regularity condition holds 𝔼fp[f(w)]=f¯(w).111This technical assumption allows us to exchange the expectation and the gradient. See discussions in Section E.2

We also assume in most results that the gradients of the functions have bounded covariance as in [4], which is a typical assumption used in robust mean estimation.

Assumption 3.

There is σ>0 such that for all w𝒲 and all unit vectors v, we have 𝐄fp[(v(f(w)f¯(w)))2]σ2.

An equivalent form of this assumption is that for every w𝒲, the covariance matrix of the gradients, defined by Σw:=𝔼fp[(f(w)f¯(w))(f(w)f¯(w))T] satisfies Σwσ2I. (See Appendix E for a proof.)

We will additionally assume that the population risk f¯(w) satisfies certain properties, or that certain properties are satisfied almost surely for functions f from distribution p, as needed.

To our best knowledge, SEVER [4] is the only work that studies robust stochastic optimization for general convex losses. While SEVER focuses on finding approximate critical points, our work focuses on minimizing the population risk f¯(w), and we measure the performance of our algorithm in terms of the excess risk f¯(w^)minwf¯(w), where w^ is the output of the algorithm.

We remark that SEVER also derives excess risk bounds. To contrast with SEVER, we decompose the excess risk of a stochastic optimization algorithm as follows222 We omit the term due to optimization error that depends on the number of iterations of the algorithm, since it will be dominated by the other terms when we run the optimization algorithm for a sufficient number of iterations. :

Excess risk=Error due to corruption+Statistical error,

where “error due to corruption” refers to the error due to the presence of corruption in the data, while “statistical error” denotes the error that accrues even when there is no corruption. SEVER [4] focuses only on the error due to corruption. The statistical error term is implicit in their requirement on the sample complexity n, that is,

Excess risk=Error due to corruption,if n[sample complexity].

Specifically, they design a polynomial-time algorithm that achieves O(Dσϵ) error due to corruption term for n=Ω~(dL2ϵσ2+dL4σ4), provided that ff¯ is L-Lipschitz and β-smooth almost surely for fp, and that f is smooth almost surely. (Their analysis has an incorrect sample complexity result, which we fix in the appendix of the full version of this paper.) This sample complexity can be huge (even infinite), as some functions in the distribution may have a very large (possibly unbounded) Lipschitz constant. Moreover, SEVER implicitly requires f to be smooth almost surely.

Consider functions of the form fx(w)=12xw2 for w such that wD, where xP for a probability distribution P with bounded mean and variance but with unbounded values, e.g. the normal distribution. We have fx(w)=xw. Since x is unbounded, the worst-case Lipschitz parameter and smoothness of f are both infinite. However, the population risk f¯(w)=12w2𝐄[x] is smooth and Lipschitz. This example demonstrates that the assumptions in SEVER that assume properties uniformly for individual functions fp can be too stringent. In this paper, we aim to answer the following question:

Can we design computationally efficient algorithms that achieve the optimal excess risk for robust SCO, under much milder conditions?

We give positive answers to this question and summarize our contributions below.

1.2 Our Contributions

  1. 1.

    Optimal Rates for Robust SCO (Section 3): We develop algorithms that achieve the following minimax-optimal (up to logarithmic factors) excess risk:

    f¯(w^T)minw𝒲f¯(w)=O~(D(σϵ+σdlog(1/τ)n)).

    Compared with SEVER, we achieve the same error due to corruption O(Dσϵ) provided n=Ω~(d/ϵ), a significant improvement in sample complexity.333 We remark that in excess risk bounds, O~ always hides logarithmic factors only in the statistical error term, and the robust term is always O(Dσϵ).

  2. 2.

    Much Weaker Assumptions for Robust SCO: Algorithm 1 achieves the optimal rates while only assuming the smoothness of the population risk, which is significantly weaker than the assumptions used in SEVER. By contrast, SEVER requires ff¯ to have bounded worst-case Lipschitz and smoothness parameter, and that individual functions f are smooth almost surely.

  3. 3.

    Handling unknown σ and extensions to nonsmooth case: Simple adaptations allow our algorithm to handle the case in which the covariance parameter σ is unknown. We also extend our algorithm to nonsmooth population risks using convolutional smoothing. The resulting algorithm achieves the minimax-optimal excess risk.

  4. 4.

    A Matching Lower Bound for Robust SCO: We show a matching lower bound, demonstrating that our excess risk bound is minimax-optimal (up to logarithmic factors). Consequently, our sample complexity for achieving the error due to corruption O(Dσϵ) is also minimax-optimal.

  5. 5.

    A Straightforward Algorithm for Robust SCO (Section 4): Algorithm 3 is an elementary algorithm that achieves the same optimal excess risk, with more stringent assumptions compared to Algorithm 1. Our approach builds on the “many-good-sets” assumption, which SEVER briefly introduced without providing a concrete analysis.

Our results might be surprising, as net-based approaches (e.g., uniform convergence) typically suffers from suboptimal error. Our results, however, imply that the net-based approach can indeed achieve the optimal excess risk under the ϵ-contamination model. We discuss this further in Section E.3. A high-level summary of our results appears in Table 1.

Table 1: Comparison of assumptions, rates, and sample complexity of SEVER and our two algorithms. The parameters β, L, etc are all assumed to be finite. All algorithms assume Assumption 2, and bounded covariance of the gradients, that is, the covariance matrix Σw satisfies Σwσ2I for all w. Optimality is up to logarithmic factors. For the case when f¯ is nonsmooth but Lipschitz (see Section 3.2), the excess risk is optimal (up to logarithmic factors) under the noncentral moment assumption.
Algorithm Assumptions Excess Risk Sample Complexity
SEVER [4]
1. ff¯ is L-Lipschitz a.s.
2. f is β-smooth a.s.
suboptimal Ω~(dL2ϵσ2+dL4σ4)
Algorithm 1 f¯ is β¯-smooth or L¯-Lipschitz. optimal Ω~(d/ϵ)
Algorithm 3
1.  ff¯ is L-Lipschitz a.s. and β-smooth a.s.
2. f¯ is β¯-smooth or L¯-Lipschitz.
optimal Ω~(d/ϵ)

2 Revisiting SEVER

In this section, we revisit SEVER [4] to motivate our work. Below we fix the corruption parameter ϵ and the covariance boundedness parameter σ>0. Given ϵ-corrupted function samples f1,,fn, we say a subset of functions is “good” with respect to w if their sample mean and covariance at w are close to those of the true distribution, as defined below.

Definition 4 (“Good” set).

We say a set Sgood[n] with |Sgood|(1ϵ)n is “good” w.r.t. w if the functions {fi}iSgood satisfy the following,

1|Sgood|iSgood(fi(w)f¯(w))(fi(w)f¯(w))T O(σ2), (1)
1|Sgood|iSgood(fi(w)f¯(w)) O(σϵ).

A “good” set w.r.t. w allows us to robustly estimate the gradient at w. SEVER requires the existence of a set that is uniformly good for all w, which we refer to as the “uniform-good-set” assumption.

Assumption 5 (“Uniform good set”, [4, Assumption B.1]).

There exists a set Sgood[n] with |Sgood|(1ϵ)n such that Sgood is “good” w.r.t. w, for all w𝒲.

SEVER operates through an iterative filtering framework built around a black-box learner. Its core algorithm consists of three main steps: (1) The black-box learner processes the current set of functions to find approximate critical points. (2) A filtering mechanism identifies and removes outlier functions. (3) The algorithm updates its working set with the remaining functions. This process repeats until convergence. Crucially, SEVER’s theoretical guarantees rely on its “uniform-good-set” assumption. Without this assumption (as opposed to “many-good-sets” assumption introduced later), the set of “good” functions can change at each iteration, potentially preventing the iterative filtering process from converging.

We argue that the “uniform-good-set” assumption can be too strong. Recall that SEVER requires a sample complexity of n=Ω~(dL2ϵσ2+dL4σ4). When n=Ω~(d/ϵ), the “uniform-good-set” assumption can no longer be guaranteed to hold. In contrast, the “many-good-sets” assumption introduced below is weaker, and aligns with the general framework of robustly estimating gradients in each iteration.

SEVER also assumes the existence of a black box approximate learner.

Definition 6 (γ-approximate learner).

A learning algorithm is called γ-approximate if, for any functions f1,,fm:𝒲, each bounded below on a closed domain , the output w of is a γ-approximate critical point of f^(x):=1mi=1mfi(x), that is, there exists δ>0 such that for all unit vectors v where w+δv𝒲, we have that vf^(w)γ.

 Remark 7.

We remark that the existence of a γ-approximate learner implies that the learner can find a γ-approximate critical point of any function f by choosing f1==fm=f. To our best knowledge, any polynomial-time algorithm that finds approximate critical points requires smoothness of the objective. Therefore, SEVER does not apply to problems where some functions in the distribution are nonsmooth. For example, consider a distribution p consisted of two functions with equal probability, h+g and hg, where h is smooth but g is nonsmooth. The population risk is smooth, but the individual functions are not.

In the appendix of [4], the authors consider the “many-good-sets” assumption, an alternative weaker assumption that allows the good set to depend on the point w.

Assumption 8 (“Many good sets”, [4, Assumption D.1]).

For every w, there exists a set Sgood(w)[n] with |Sgood(w)|(1ϵ)n such that Sgood(w) is “good” with respect to w.

We remark that the “many-good-sets” assumption allows us to do robust gradient estimation in each iteration. The SEVER paper mentions (without going into detail) that under the “many-good-sets” assumption, projected gradient descent can be used to find a O(σϵ)-approximate critical point. It is unclear that under what conditions “many-good-sets” assumption can be satisfied, and no excess risk bound or sample complexity is provided.

In this paper, we utilize a further relaxed assumption stated below, which only requires the existence of good sets at points in a fine net of the domain. For these purposes, we define a ξ-net of 𝒲 (for some small ξ>0) to be a set 𝒞 such that for any w𝒲, there exists w𝒞 with wwξ.

Assumption 9 (“Dense good sets”).

For a given ξ>0, there exists a ξ-net 𝒞 of the domain 𝒲 such that for every w𝒞, there exists a set Sgood(w)[n] with |Sgood(w)|(1ϵ)n such that Sgood(w) is “good” with respect to w.

When the “dense good sets” assumption holds, we can approximate the gradient at any point in the domain 𝒲 by robustly estimating the gradient at the nearest point in the net. The approximation error will be small, provided that the population risk is smooth and the net is fine enough. (The parameter ξ will depend on σ, as we will see in Algorithm 1.) This relaxed assumption allows us to circumvent the technical difficulties of dealing with infinitely many w, thus removing the requirements of uniform Lipschitzness and smoothness of ff¯ for all f that are used in SEVER. As a consequence, we are able to achieve the same corruption error as SEVER with a significantly reduced sample complexity. The next section presents our algorithm that achieves this result.

3 Optimal Rates for Robust SCO under Weak Distributional Assumptions

We now present a net-based algorithm that achieves the minimax-optimal excess risk under the weak assumption that the population risk f¯ is smooth.

Assumption 10.

Given the distribution p over functions f:𝒲 with f¯=𝐄fp[f], we have that f¯ is β¯-smooth.

Here, the β-smoothness requirement only applies to the population risk. Each individual function f can have different smoothness parameters.

We outline our algorithm below, see Algorithm 1. The algorithm is based on projected gradient descent with a robust estimator. Here, we treat the robust gradient estimator RobustEstimator as a black box, which can be any deterministic stability-based algorithm. For completeness, we provide an instantiation of the robust gradient estimator due to [7], outlined in Algorithm 2, which at a high level iteratively filters out points that are “far” from the sample mean in a large variance direction. Algorithm 2 runs in polynomial time.

The key innovation lies in its gradient estimation strategy.

Rather than computing gradients at arbitrary points, it makes use a dense net of the domain 𝒲, estimating the gradient at the current iterate w by the gradient at the nearest point w in the net to w. The smoothness of the population risk ensures that this approximation remains accurate. As mentioned at the end of Section 2, this strategy helps us avoid the technical challenges of handling infinitely many w with a net argument, thereby achieving optimal rates under significantly weaker distributional assumptions compared to SEVER.

Algorithm 1 Net-based Projected Gradient Descent with Robust Gradient Estimator.
Algorithm 2 An Instantiation of the Robust Gradient Estimator: Iterative Filtering [7].

Efficient Implementation.

For implementation efficiency, we propose a grid-based net construction. Let ξ=σϵ/β¯. We use grid points spaced ξ/d apart in each dimension, i.e.,

{ξdz=(ξdz1,ξdz2,,ξdzd):z=(z1,z2,,zd)d,ξdz2D}

to construct a ξ-net.444Technically, we can choose a grid spaced ξ/4d apart in each dimension, and add additional points to cover the boundary of the feasible set. This would reduce the size of grid points by almost a factor of 2d. Given a point w, we can find a net point within ξ distance in O(d) time through: (1) Scaling: Divide w by ξ/d. (2) Rounding: Convert to the nearest integral vector in d. (3) Rescaling: Multiply by ξ/d.

This construction yields a net of size |𝒞|=O(Dd/ξ)d, which is larger than the optimal covering number O((D/ξ)d). While this introduces an extra logd factor in the excess risk bound (due to union bound over net points), it offers two significant practical advantages: (1) Implicit net: No need to explicitly construct and store the net. (2) Efficient computation: O(d) time for finding the nearest net point. An exponential-time algorithm that achieves the excess risk without the logd factor is described in Appendix F.

Polynomial runtime.

The robust gradient estimator in Algorithm 2 runs in polynomial time, when used with the grid-based construction above. As can be seen in Appendix A, the required number of iterations is also polynomial in parameters. Therefore, the algorithm runs in polynomial time overall.

Convergence of Algorithm 1 is described in the following result.

Theorem 11.

Suppose that Assumption 2, Assumption 3, and Assumption 10 hold. There are choices of stepsizes {ηt}t=1T and T such that, with probability at least 1τ, we have

f¯(w^T)minw𝒲f¯(w)=O~(σDϵ+σDdlog(1/τ)n).

As a consequence, the algorithm achieves excess risk of O(Dσϵ) with high probability whenever n=Ω~(d/ϵ).

 Remark 12.

Theorem 11 is minimax-optimal (up to logarithmic factors). Our sample complexity n=Ω~(d/ϵ), significant improves over the sample complexity of SEVER, which is n=Ω~(dL2ϵσ2+dL4σ4).

The following matching lower bound can be established, showing the minimax-optimality (up to logarithmic factors) of Algorithm 1. A proof of the lower bound, drawing in part on an adaptation of [15], is provided in the appendix of the full version of the paper.

Theorem 13.

For d140 and n62500, there exist a closed bounded set 𝒲d with diameter at most D and a distribution p over functions f:𝒲 that satisfy the following: Let f¯=𝐄fp[f]. We have that for every w𝒲 and unit vector v that 𝐄fp[(v(f(w)f¯(w)))2]σ2. Both f (almost surely) and f¯ are convex, Lipschitz and smooth. The output w^ of any algorithm with access to an ϵ-corrupted set of functions f1,,fn sampled from p satisfies the following with probability at least 1/2:

f¯(w^)minw𝒲f¯(w)=Ω(Dσϵ+Dσdn). (2)

3.1 Proof Sketch of Theorem 11 

We defer the full proof to Appendix A and sketch the proof below. In each iteration of Algorithm 1, we estimate the gradient at the current iterate w by applying the robust gradient estimator to its nearest point in the net w. We can decompose the error as follows:

g~(w)f¯(w)g~(w)f¯(w)+f¯(w)f¯(w),

where the first term measures the bias of the robust gradient estimator, and the second term is due to the approximation error due to the net.

We will show that there exist good sets for all net (grid) points (cf. Assumption 9) with high probability, so that we can robustly estimate gradients for all points in the net. This gives a bound for the first term in the equation above, whereas the second term can be bounded using smoothness of the population risk f¯.

Once we establish the gradient estimation bias in each iteration (with high probability), we use the projected biased gradient descent analysis framework (Section 4) to establish an upper bound on the excess risk.

3.2 Handling Nonsmooth but Lipschitz Population Risks

We now consider a setting in which Assumption 2 and Assumption 3 hold, but f¯ is nonsmooth in the sense that Assumption 10 is not satisfied. That is, there is no β¯< such that f¯ is β¯-smooth. We assume instead that f¯ is L¯-Lipschitz for some finite L¯. In this setting, we can use convolutional smoothing and run Algorithm 1 on the smoothed objective. Our algorithm works as follows:

  1. 1.

    For every index i[n], we independently sample a perturbation ui𝒰s, where 𝒰s is the uniform distribution over the d-dimensional L2-norm ball of radius s centered at the origin. We replace samples {fi}i=1n by the smoothed samples {fi(+ui)}i=1n.

  2. 2.

    We run Algorithm 1 on the smoothed samples with β replaced by L¯ds and σ replaced by σ2+4L¯2.

The modified algorithm has the following convergence guarantees.

Proposition 14.

Suppose that Assumption 2 and Assumption 3 hold and that f¯ is L¯-Lipschitz. There are choices of algorithmic parameters such that, with probability at least 1τ, the output w^T of the modified algorithm satisfies the following excess risk bound:

f¯(w^T)minw𝒲f¯(w)=O~((σ+L¯)Dϵ+(σ+L¯)Ddlog(1/τ)n).
 Remark 15.

Compared to the smooth case, the excess risk bound has an extra L¯ term. Using this result, we can show that under the alternative noncentral moment assumption, that is, instead of Assumption 3, assume that for every w𝒲 and unit vector v, 𝐄fp[(vf(w))2]G2, our modified algorithm (with σ and L¯ both replaced by G) achieves the following excess risk bound.

Theorem 16.

Suppose that Assumption 2 holds, and that for every w𝒲 and unit vector v, we have 𝐄fp[(vf(w))2]G2 for some G. There are choices of algorithmic parameters such that, with probability at least 1τ, the output w^T of the modified algorithm (with σ and L¯ both replaced by G) satisfies the following excess risk bound:

f¯(w^T)minw𝒲f¯(w)=O~(GDϵ+GDdlog(1/τ)n).

See Appendix B for proofs of both results above.

We can show that the excess risk bound in Theorem 16 is minimax-optimal (up to logarithmic factors) under the noncentral moment assumption, as a matching lower bound can be established.

Theorem 17.

For d140 and n62500, there exist a closed bounded set 𝒲d with diameter at most D, and a distribution p over functions f:𝒲 that satisfy the following: Let f¯=𝐄fp[f]. We have that for every w𝒲 and unit vector v that 𝐄fp[(vf(w))2]G2. Both f (almost surely) and f¯ are convex, Lipschitz and smooth. The output w^ of any algorithm with access to an ϵ-corrupted set of functions f1,,fn sampled from p satisfies the following with probability at least 1/2,

f¯(w^)f¯=Ω(DGϵ+DGdn). (3)

The proof is essentially the same as that of Theorem 13 (available in the full version of the paper), since the same hard instances that are used to establish Theorem 13 can be reused to establish Theorem 17.

3.3 Handling Unknown Covariance Parameter σ

In Algorithm 1, σ primarily affects the fineness of the net through ξ=σϵ/β¯. When σ is unknown, we can adapt our algorithm with a preprocessing step to estimate σ: (1) Run iterative filtering (Algorithm 2) to obtain a lower bound σ^ of σ, and (2) Run Algorithm 1 with the modified fineness parameter ξ=σ^ϵ/β¯. This adaptation preserves the optimal excess risk guarantees for the known σ case. Full details are provided in Appendix G.

4 Projected Gradient Descent with Robust Gradient Estimator

Algorithm 1 uses a net-based approach to estimate gradients robustly. A more naïve approach is to directly estimate gradients at arbitrary points using a robust gradient estimator. We will show that the simple projected gradient descent algorithm can achieve the same optimal rate as Algorithm 1 under stronger assumptions. Even so, our new assumptions are still slightly weaker than those used in SEVER [4]. Concretely, following assumptions on the distribution over functions are assumed.

Assumption 18.

Let p be a distribution over functions f:𝒲 with f¯=𝐄fp[f] so that:

  1. 1.

    ff¯ is L-Lipschitz and β-smooth almost surely, where555 Without loss of generality, see Appendix E. Lσ.

  2. 2.

    f¯ is β¯-smooth or L¯-Lipschitz.

We use bars on the constants β¯ and L¯ to emphasize that they reflect properties of f¯.

Algorithm 3 follows the “many-good-sets” assumption. We are able to robustly estimate the gradient of the population risk f¯ at any point w with high probability, at the cost of requiring additional almost-sure assumptions on ff¯ compared to Algorithm 1.

Algorithm 3 Projected Gradient Descent with Robust Gradient Estimator.

Algorithm 3 achieves the same optimal excess risk bounds as in Theorem 11.

Theorem 19.

Suppose that Assumption 2, Assumption 3, and Assumption 18 hold. There are choices of stepsizes {ηt}t=1T and T such that, with probability at least 1τ, we have

f¯(w^T)minw𝒲f¯(w)=O~(σDϵ+σDdlog(1/τ)n).

As a consequence, the algorithm achieves excess risk of O(Dσϵ) with high probability whenever n=Ω~(d/ϵ). The expected excess risk is bounded by O~(σDϵ+σDd/n).

The proof is based on the net argument, similar to that of Lemma C.5 in [14]. The high level idea is as follows: For simplicity, we say w is “good” if there exists a good set of functions at w. We will show that with high probability, there exists a good set for all w (cf. Assumption 8), so that we can robustly estimate the gradient at all w. To show this, we employ a net argument, based on the claim that if w is “good”, then all points in a small neighborhood of w are also “good.” By the union bound, with high probability, all points in the net are “good”. Then it follows that all w are “good”. The full proof can be found in Appendix C.

5 Conclusion and Future Work

In this work, we have advanced robust stochastic convex optimization under the ϵ-contamination model. While the prior state of the art SEVER [4] focused on finding approximate critical points under stringent assumptions, we have developed algorithms that directly tackle population risk minimization, obtaining the optimal excess risk under more practical assumptions. Our first algorithm (Algorithm 1) achieves the minimax-optimal excess risk by leveraging our relaxed “dense-good-sets” assumption and estimating gradients only at points in a net of the domain, relaxing the stringent distributional conditions as required in SEVER. Our second algorithm (Algorithm 3) provides a simple projected gradient descent approach that achieves the same optimal excess risk, making use of the “many-good-sets” assumption briefly noted in [4]. Both of our algorithms significantly reduce sample complexity compared to SEVER.

For future work, it would be interesting to explore following directions: (1) Our excess risk bound is tight up to logarithmic factors. Can we improve the bound to remove the logarithmic factors? (2) Our lower bound is with constant probability. Is it possible to derive a lower bound that the includes log(1/τ) term with probability τ? (3) Robustness has been shown to be closely related to differential privacy [10]. Can we design optimization algorithms that are both robust and differentially private?

References

  • [1] Battista Biggio, Blaine Nelson, and Pavel Laskov. Poisoning attacks against support vector machines. arXiv preprint arXiv:1206.6389, 2012.
  • [2] Yeshwanth Cherapanamjeri, Efe Aras, Nilesh Tripuraneni, Michael I Jordan, Nicolas Flammarion, and Peter L Bartlett. Optimal robust linear regression in nearly linear time. arXiv preprint arXiv:2007.08137, 2020. arXiv:2007.08137.
  • [3] Ilias Diakonikolas, Gautam Kamath, Daniel Kane, Jerry Li, Ankur Moitra, and Alistair Stewart. Robust estimators in high-dimensions without the computational intractability. SIAM Journal on Computing, 48(2):742–864, 2019. doi:10.1137/17M1126680.
  • [4] Ilias Diakonikolas, Gautam Kamath, Daniel Kane, Jerry Li, Jacob Steinhardt, and Alistair Stewart. SEVER: A robust meta-algorithm for stochastic optimization. In International Conference on Machine Learning, pages 1596–1606. PMLR, 2019. URL: http://proceedings.mlr.press/v97/diakonikolas19a.html.
  • [5] Ilias Diakonikolas, Gautam Kamath, Daniel M Kane, Jerry Li, Ankur Moitra, and Alistair Stewart. Being robust (in high dimensions) can be practical. In International Conference on Machine Learning, pages 999–1008. PMLR, 2017. URL: http://proceedings.mlr.press/v70/diakonikolas17a.html.
  • [6] Ilias Diakonikolas and Daniel M Kane. Recent advances in algorithmic high-dimensional robust statistics. arXiv preprint arXiv:1911.05911, 2019. arXiv:1911.05911.
  • [7] Ilias Diakonikolas, Daniel M Kane, and Ankit Pensia. Outlier robust mean estimation with subgaussian rates via stability. Advances in Neural Information Processing Systems, 33:1830–1840, 2020.
  • [8] Ilias Diakonikolas, Weihao Kong, and Alistair Stewart. Efficient algorithms and lower bounds for robust linear regression. In Proceedings of the Thirtieth Annual ACM-SIAM Symposium on Discrete Algorithms, pages 2745–2754. SIAM, 2019. doi:10.1137/1.9781611975482.170.
  • [9] Vitaly Feldman. Generalization of erm in stochastic convex optimization: The dimension strikes back. Advances in Neural Information Processing Systems, 29, 2016.
  • [10] Samuel B Hopkins, Gautam Kamath, Mahbod Majid, and Shyam Narayanan. Robustness implies privacy in statistical estimation. In Proceedings of the 55th Annual ACM Symposium on Theory of Computing, pages 497–506, 2023. doi:10.1145/3564246.3585115.
  • [11] Peter J Huber. Robust estimation of a location parameter. In Breakthroughs in statistics: Methodology and distribution, pages 492–518. Springer, 1992.
  • [12] Arun Jambulapati, Jerry Li, Tselil Schramm, and Kevin Tian. Robust regression revisited: Acceleration and improved estimation rates. Advances in Neural Information Processing Systems, 34:4475–4488, 2021. URL: https://proceedings.neurips.cc/paper/2021/hash/23b023b22d0bf47626029d5961328028-Abstract.html.
  • [13] Adam Klivans, Pravesh K Kothari, and Raghu Meka. Efficient algorithms for outlier-robust regression. In Conference On Learning Theory, pages 1420–1430. PMLR, 2018. URL: http://proceedings.mlr.press/v75/klivans18a.html.
  • [14] Shuyao Li, Yu Cheng, Ilias Diakonikolas, Jelena Diakonikolas, Rong Ge, and Stephen Wright. Robust second-order nonconvex optimization and its application to low rank matrix sensing. Advances in Neural Information Processing Systems, 36, 2024.
  • [15] Andrew Lowy and Meisam Razaviyayn. Private stochastic optimization with large worst-case lipschitz parameter: Optimal rates for (non-smooth) convex losses and extension to non-convex losses. In International Conference on Algorithmic Learning Theory, pages 986–1054. PMLR, 2023. URL: https://proceedings.mlr.press/v201/lowy23a.html.
  • [16] Adarsh Prasad, Arun Sai Suggala, Sivaraman Balakrishnan, and Pradeep Ravikumar. Robust estimation via robust gradient estimation. Journal of the Royal Statistical Society Series B: Statistical Methodology, 82(3):601–627, 2020.
  • [17] John Wilder Tukey. A survey of sampling from contaminated distributions. Contributions to probability and statistics, pages 448–485, 1960.
  • [18] Farzad Yousefian, Angelia Nedić, and Uday V Shanbhag. On stochastic gradient and subgradient methods with adaptive steplength sequences. Automatica, 48(1):56–67, 2012. doi:10.1016/J.AUTOMATICA.2011.09.043.

Appendix A Proof of Theorem 11

We start with the robust estimation result from [7], then proceed with the proof of Theorem 11.

Lemma 20 ([7, Proposition 1.5]).

Let S be an ϵ-corrupted set of n samples from a distribution in d with mean μ and covariance Σ such that Σσ2I. Let ϵ=Θ(log(1/τ)/n+ϵ)c be given, for a constant c>0. Then any stability-based algorithm (e.g. Algorithm 2) on input S and ϵ, efficiently computes μ^ such that with probability at least 1τ, we have

μ^μ=O(σδ(τ)),whereδ(τ)=ϵ+d/n+log(1/τ)/n. (4)

Proof of Theorem 11.

1. Bound the bias of the gradient estimator at w. For given w, let w=argminz𝒞zw. Applying Lemma 20 to samples f1(w),f2(w),,fn(w), we have that with probability at least 1τ, the robust gradient estimator g~(w) satisfies

g~(w)f¯(w)=σO~(ϵ+d/n+log(1/τ)/n).

We have wwσϵ/β¯ by definition of the net. By β¯-smoothness of the population risk f¯, we have

f¯(w)f¯(w)β¯wwσϵ. (5)

Combining the two bounds, we have

g~(w)f¯(w)=σO~(ϵ+d/n+log(1/τ)/n). (6)

2. Apply the union bound over all points in the net 𝒞. By union bound, setting τ=τ/|𝒞|, we have that with probability at least 1τ, (6) simultaneously holds for all w𝒞. Recall |𝒞|=O(Dd/ξ)d. We have log|𝒞|=O~(d). It follows that, with probability at least 1τ, simultaneously for all w𝒲, let w=argminz𝒞zw, we have

g~(w)f¯(w)=σO~(ϵ+d/n+dlog(1/τ)/n). (7)

Therefore, with probability at least 1τ, the bias of the gradient estimator at w is bounded by the above expression, simultaneously for all w𝒲.

3. Apply the projected biased gradient descent analysis. By Lemma 27, choosing a constant step size η=1/β¯, the excess risk of the algorithm is bounded by

f¯(w^T)minw𝒲f¯(w)=O~(β¯D2T+D(σϵ+σdlog(1/τ)n)). (8)

Choosing T=Ω~(β¯Dσϵ+σdlog(1/τ)n) gives the optimal rate.

Appendix B Analysis of Convolutional Smoothing for Nonsmooth but Lipschitz Population Risk

Before proving Proposition 14, we need some properties of the convolutional smoothing.

Lemma 21 ([18]).

Suppose {f(w)} is convex and L-Lipschitz over 𝒲+B2(0,s), where B2(0,s) is the d-dimensional L2 ball of radius s centered at the origin. For w𝒲, the convolutional smoother with radius s, f~s(w):=𝔼u𝒰s[f(w+u)], where 𝒰s is the uniform distribution over B2(0,s), has the following properties:

  1. 1.

    f(w)f~s(w)f(w)+Ls;

  2. 2.

    f~s(w) is convex and L-Lipschitz;

  3. 3.

    f~s(w) is Lds-smooth.

Proof of Proposition 14.

Let f¯s(w)=𝔼u𝒰s[f¯(w+u)] be the smoothed population risk.

  1. 1.

    By properties of convolutional smoothing (part 3 of Lemma 21), we know that since f¯ is L¯-Lipschitz, f¯s is (L¯d/s)-smooth.

    As f1,f2,,fn are ϵ-corrupted samples from p, we know f1(+u1),f2(+u2),,fn(+un) are ϵ-corrupted samples from the product distribution of p and 𝒰s. Below, we show that in expectation, perturbed gradient (clean) samples f(w+u) is equal to the smoothed gradient f¯s(w).

    By the law of total expectation, we have

    𝔼fp,u𝒰s[f(w+u)]=𝔼u𝒰s[𝔼fp[f(w+u)]],

    and using the regularity condition to interchange the gradient with the expectation, we obtain

    𝔼fp,u𝒰s[f(w+u)]=𝔼u𝒰s[f¯(w+u)].

    Below we drop the distributions and write 𝔼f, 𝔼u for simplicity. Since f¯ is L¯-Lipschitz, we can exchange the order of expectation and gradient, that is,

    𝔼u[f¯(w+u)]=𝔼u[f¯(w+u)]=f¯s(w), (9)

    which shows that 𝔼f,u[f(w+u)]=f¯s(w).

  2. 2.

    Next, we bound the covariance of the perturbed gradient Covf,u(f(w+u)).

    Using law of total covariance, that is, Cov(X,Y)=𝔼(Cov(X,YZ))+Cov(𝔼(XZ),𝔼(YZ)), conditioned on u, we can write the covariance of the perturbed gradient as

    Covf,u(f(w+u)) =𝔼u[Covf(f(w+u))]+Covu[𝔼f[f(w+u)]] (10)
    =𝔼u[Covf(f(w+u))]Term 1+Covu[f¯(w+u)]Term 2,

    In the first term, for all u, the covariance inside the expectation is bounded by σ2I by assumption. Therefore, we have (Term 1)𝔼u[σ2I]=σ2I. We bound the second term using the boundedness of the gradient. Consider the following fact: for any random vector u, we have 𝔼[uu]C2I if uC almost surely, and consequently, Cov(u)4C2I. It follows that, (Term 2)4L¯2I. Therefore, the covariance parameter increases from σ2 to σ2+4L¯2 due to smoothing.

    We have verified the conditions to apply the original algorithm, with σ replaced by σ2+4L¯2=O(σ+L¯) and β¯=L¯d/s.

  3. 3.

    By applying Theorem 11 to the smoothed function f¯s, we know there are choices of {ηt}t=1T and T such that, with probability at least 1τ, the output w^T satisfies,

    f¯s(w^T)minw𝒲f¯s(w)=O~((σ+L¯)Dϵ+(σ+L¯)Ddlog(1/τ)n).

    By properties of convolutional smoothing, we have

    f¯(w^T)f¯s(w^T)andminw𝒲f¯s(w)minw𝒲f¯(w)+L¯s.

    It follows that, choosing s=O~(D(σ/L¯+1)(ϵ+dlog(1/τ)/n)), we have

    f¯(w^T)minw𝒲f¯(w) f¯s(w^T)minw𝒲f¯s(w)+L¯s (11)
    =O~((σ+L¯)Dϵ+(σ+L¯)Ddlog(1/τ)n).

As a corollary, the result under the noncentral moment condition (Theorem 16) follows using the lemma below, which the relates noncentral second moment to the mean.

Lemma 22.

For a d-dimensional random vector u that satisfies 𝔼[uu]G2Id, we have that 𝔼[u]G.

Proof.

Let v be any fixed unit vector in d. By Jensen’s inequality, we have

(v𝔼[u])2=(𝔼[vu])2𝔼[(vu)2]=v𝔼[uu]vG2 (12)

Since this inequality holds for any unit vector v, we can choose v=𝔼[u]𝔼[u] when 𝔼[u]0, which gives 𝔼[u]2G2, and thus 𝔼[u]G. If 𝔼[u]=0, inequality 𝔼[u]G holds trivially.

Proof of Theorem 16.

For any random vector u, we have for any unit vector v,

𝔼[(v(u𝔼[u]))2]=𝔼[(vu)2]v𝔼[u]2.

Substituting u=𝔼f[f(w)], we have 𝔼[u]=f¯(w). So for every w, we have 𝐄[(v(f(w)f¯(w)))2]𝐄[(vf(w))2]G2 for any unit vector v. We know that 𝐄[f(w)f(w)]G2Id is equivalent to 𝐄[(vf(w))2]G2 holding for every unit vector v. By Lemma 22, we have f¯(w)G for every w. Therefore, applying Proposition 14 with σ=G and L¯=G, we obtain the desired result.

Appendix C Analysis of Algorithm 3

Before proving Theorem 19, we need some results from robust estimation literature.

C.1 Results from Robust Mean Estimation

Recall Definition 4. The “good” set property is a special case of stability, defined as follows:

Definition 23 (Stability [3]).

Fix 0<ϵ<1/2 and δϵ. A finite set Sd is (ϵ,δ)-stable with respect to mean μd and σ2 if for every SS with |S|(1ϵ)|S|, the following conditions hold: (i) μSμσδ, and (ii) Σ¯Sσ2Iσ2δ2/ϵ, where μS=(1/|S|)xSx and ΣS=(1/|S|)xS(xμ)(xμ).

The following due to [7] establishes stability of samples from distributions with bounded covariance.

Lemma 24 ([7]).

Fix any 0<τ<1. Let S be a multiset of n i.i.d. samples from a distribution on d with mean μ and covariance Σ such that Σσ2I. Let ϵ=Θ(log(1/τ)/n+ϵ)c, for a sufficiently small constant c>0. Then, with probability at least 1τ, there exists a subset SS such that |S|(1ϵ)n and S is (2ϵ,δ)-stable with respect to μ and σ2, where δ=δ(τ) depends on τ as δ(τ)=O((dlogd)/n+ϵ+log(1/τ)/n).

With the stability condition, we can robustly estimate the mean of a distribution with bounded covariance.

Lemma 25 (Robust Mean Estimation Under Stability [3]).

Let Td be an ϵ-corrupted version of a set S with the following stability properties: S contains a subset SS such that |S|(1ϵ)|S| and S is (Cϵ,δ) stable with respect to μd and σ2, for a sufficiently large constant C>0. Then there is a polynomial-time algorithm (e.g. Algorithm 2), that on input ϵ,T, computes μ^ such that μ^μ=O(σδ).

C.2 Proof of Theorem 19

As long as the stability condition holds, we can use deterministic stability-based algorithms (e.g. Algorithm 2) to robustly estimate the mean. Using union bound over the net, it suffices to argue that at a given point w, given the existence of a stable subset of the form {fi(w)}i, where denotes the index set of the stable subset at w, such subset is also stable within a small neighborhood of w, that is, {fi(w)}i is stable for all w in a small neighborhood of w. We have the following stability result, which corresponds to “many-good-sets” Assumption 8.

Lemma 26.

Under Assumption 3 and Assumption 18, let f1,,fn denote an ϵ-corrupted set of functions sampled from p. Let ϵ=Θ(log(1/τ)/n+ϵ)c be given, for a constant c>0. With probability at least 1τ, for all w𝒲, there exists index set [n] (here depends on the choice of w) such that ||(1ϵ)n and {fi(w)}i is (2ϵ,δ(τ))-stable with respect to f¯(w) and σ2, where τ=τ/exp(O~(d)) and δ(τ)=O~(ϵ+dlog(1/τ)/n).

Proof.

We use a net argument to show that the stability condition holds for all w, following similar proof techniques used in [14]. For fixed w, by Lemma 24, with probability at least 1τ, there exists a subset [n] such that ||(1ϵ)n and {fi(w)}i is (2ϵ,δ)-stable where δ=δ(τ), with respect to f¯(w) and σ2, that is

1||ifi(w)f¯(w)σδ, (13a)
1||i(fi(w)f¯(w))(fi(w)f¯(w))σ2Iσ2δ2/ϵ. (13b)

By β-smoothness of fif¯, we have

1||ifi(w)f¯(w) 1||i(fi(w)fi(w))+fi(w)f¯(w) (14)
βww+σδ.

Therefore, (13a) holds (up to a constant factor) for all w such that wwσδ/β.

Next, Equation 13b is equivalent to the following: for any unit vector v, we have

|1||i(v(fi(w)f¯(w)))2σ2|σ2δ2/ϵ.

By L-Lipschitzness and β-smoothness of fif¯, for any unit vector v, we have

|{v(fi(w)f¯(w))}2{v(fi(w)f¯(w))}2| (15)
={v(fi(w)f¯(w))+v(fi(w)f¯(w))}
{v(fi(w)f¯(w))v(fi(w)f¯(w))}2Lβww|,

It follows that (13b) holds (up to a constant factor) for w such that wwσ2δ2/(ϵLβ).

Let ξ=min(σδ/β,σ2δ2/(ϵLβ)). Then, for all w such that wwξ, {fi(w)}i is (2ϵ,2δ)-stable with respect to f¯(w) and σ2. It suffices to choose a ξ-net 𝒞 of 𝒲, where the optimal size of the net is |𝒞|=O((D/ξ)d), and choose τ=τ/|𝒞|. By union bound, with probability at least 1|𝒞|τ, the stable subset exists for all w𝒞 simultaneously. Since we have argued that for fixed w, the same stable subset applies for all w within distance ξ from w, the subset stability holds simultaneously for all w with probability at least 1τ, as claimed.

Proof of Theorem 19.

Combining Lemma 26 and Lemma 25, with probability at least 1τ, in each iteration, we can estimate the gradient up to a bias as follows:

g~(wt)f¯(wt)=O(σδ(τ))=O~(σϵ+σdlog(1/τ)/n).

Note that the probability is simultaneously for all w, so it does not matter how many iterations we run. Conditioned on the gradient estimation bias bound being held, the excess risk bound then follows by applying Lemma 27 for smoothness loss, or Lemma 28 for Lipschitz loss with corresponding choices of stepsizes and large enough T.

Appendix D Projected Biased Gradient Descent

In this section, we analyze the convergence of the projected gradient descent algorithm with a biased gradient estimator. We assume the loss function is convex throughout this section.

Both Algorithm 1 and Algorithm 3 can be viewed as instances of the following algorithm.

Algorithm 4 Projected Gradient Descent with Biased Gradient Estimator.

Here, Π𝒲() denotes the projection operator onto the feasible set 𝒲, that is,

Π𝒲(y)=argminw𝒲wy2.

The projection operation ensures that the iterates wt remain within the feasible set 𝒲 throughout the optimization process. The projection step is crucial when the optimization problem is constrained, as it guarantees that the updates do not violate the constraints defined by 𝒲.

Let us assume that the biased gradient estimator g~t has a bias bounded by B, that is, g~tF(wt)B for all t, and the diameter of the feasible set 𝒲 is bounded by D, that is, wwD for all w,w𝒲. We have the following convergence results for the algorithm.

Lemma 27.

Suppose F is convex and β-smooth. Running Algorithm 4 with constant step size η=1β, we have

F(1Tt=1Twt)minw𝒲F(w)βD22T+BD. (16)

Alternatively, we can consider the case where the loss function F(w) is convex and L-Lipschitz. The following lemma holds.

Lemma 28.

Suppose F is convex and L-Lipschitz. Running Algorithm 4 with constant step size η=1β, we have

F(1Tt=1Twt)minw𝒲F(w)DLT+(1T+1)BD. (17)

The proof of the two convergence results above can be found in the full version.

Appendix E Discussions on the Assumptions

E.1 On the bounded covariance assumption

Without loss of generality, we can assume σL. The reason is as follows: By Lipschitzness, we have f(w)f¯(w)L almost surely. By Cauchy-Schwarz Inequality, we have 𝐄fp[(v(f(w)f¯(w)))2]L2 holds for any unit vector v. On the other hand, L can be as larger as dσ (e.g. consider standard multivariate normal).

The condition 𝐄[(v(f(w)f¯(w)))2]σ2 for every unit vector v is equivalent to requiring that the covariance matrix Σw of the gradients f(w) satisfies Σwσ2I.

Proposition 29.

Let Σw denote the covariance matrix of the gradients f(w). For given w, the following two assumptions are equivalent:

  1. 1.

    For every unit vector v, we have 𝐄fp[(v(f(w)f¯(w)))2]σ2.

  2. 2.

    The covariance matrix satisfies Σw σ2I.

Furthermore, since Σw is positive semidefinite, by definition of the spectral norm, the latter assumption can be equivalently written as Σwσ2.

Proof.

By definition,

Σw=𝐄fp[(f(w)f¯(w))(f(w)f¯(w))]. (18)

We have

𝐄fp[(v(f(w)f¯(w)))2] =𝐄fp[v(f(w)f¯(w))(f(w)f¯(w))v] (19)
=v𝐄fp[(f(w)f¯(w))(f(w)f¯(w))]v
=vΣwv.

Therefore, the two assumptions are equivalent.

E.2 On the regularity condition

For technical reasons, we need to assume regularity conditions such that we can exchange the gradient and expectation, that is, for any w,

𝔼fp[f(w)]=f¯(w). (20)

A necessary condition666Technically, we need to be more precise about the distribution p over functions. In this paper, we follow the same convention as used by [4]. To be more concrete, we can just think of f parameterized by some random variable X and the distribution p is induced by the distribution of X. See the example. due to dominated convergence theorem for the regularity condition is that there exists some functional (mapping of functions) g(f) such that 𝔼fp[g(f)]<, and for all w, f(w)g(f) almost surely.

Consider the following example, where f takes the form fX(w)=12(Xw)2 where X is a random vector in d with distribution P. The distribution P of X induces the distribution p of functions f. Let P be such that 𝔼X[XXT2]M for some M>0, but X has unbounded support (e.g., X𝒩(0,Id) is multivariate Gaussian). Note that fX(w)=XXw. So we have fX(w)XX2w. In this case, we can take g(X)=DXX so that fX(w)g(X) almost surely. We have that 𝔼X[g(X)]XX2wMC<. In this case, we can exchange the order of expectation and gradient.

E.3 Comparing bounded covariance assumption with bounded variance assumption

Net-based approaches (e.g. uniform convergence) often suffer from suboptimal error [9]. However, Algorithm 1 indeed achieves the minimax-optimal rate. We believe the reason is due to the bounded covariance assumption Σσ2I. Below, we provide a discussion on the bounded covariance assumption and compare it with the bounded variance assumption.

The bounded covariance assumption Σσ2I is different from the bounded variance assumption 𝐄f(w)f¯(w)2Φ2 as commonly used in optimization literature without corruption. Using the property tr(AB)=tr(BA), this is equivalent to tr(Σ)Φ2.

We comment that neither assumption implies the other. For isotropic Gaussian distribution, where the covariance matrix is Σ=σ2I, we have tr(Σ)=dσ2. On the other hand, consider the distribution where the variance is concentrated in one direction, i.e., Σ=Φ2vv for some unit vector v. We have tr(Σ)=Φ2 and Σ=Φ2. In general, we only know that Σtr(Σ)dΣ.

Recall Lemma 24. The complete version of the lemma is as follows:

Lemma 30 ([7]).

Fix any 0<τ<1. Let S be a multiset of n i.i.d. samples from a distribution on d with mean μ and covariance Σ. Let ϵ=Θ(log(1/τ)/n+ϵ)c, for a sufficiently small constant c>0. Then, with probability at least 1τ, there exists a subset SS such that |S|(1ϵ)n and S is (2ϵ,δ(τ))-stable with respect to μ and Σ, where δ(τ)=O((r(Σ)logr(Σ))/n+ϵ+log(1/τ)/n). Here we use r(M) to denote the stable rank (or intrinsic dimension) of a positive semidefinite matrix M, i.e., r(M):=tr(M)/M.

Following identical proof steps (recall proofs for our algorithms), we can express our excess risk bound in terms of the covariance matrix Σ as follows:

DO~(Σϵ+tr(Σ)/n+dΣlog(1/τ)/n). (21)

In our paper, we consider the bounded covariance assumption Σσ2I, which is a standard assumption in robust optimization literature. Otherwise, we cannot control the error term Σϵ due to corruption. In the worse case (e.g. isotropic Gaussian), we have Σ=σ2 and tr(Σ)=dσ2, so the bound reduces to

σO(ϵ+d/n+dlog(1/τ)/n). (22)

We see that the second term already contains the dependence on d. Therefore, the d factor in the last term due to our net-based approach in conjunction with the use of the union bound over the net points, does not affect the rate.

Appendix F An exponential time algorithm that achieves the minimax-optimal excess risk bound without logd factor

Both of our two algorithms achieve the minimax-optimal excess risk bound up to logarithmic factors. In this section, we show that the minimax-optimal excess risk bound can be achieved without the logd factor, but at the cost of exponential time complexity. Based on Lemma 24, we can remove the logd factor when estimating the gradients, by using the following framework, as shown in [7].

  1. 1.

    Set k=ϵn. Randomly partition n samples S into k buckets of size n/k (remove the last bucket if n is not divisible by k).

  2. 2.

    Compute the empirical mean within each bucket and denote the means as z1,,zk.

  3. 3.

    Run stability-based robust mean estimation over the set {z1,,zk}.

Here, the first two steps serve as preprocessing before feeding the data into the robust mean estimation algorithm. We now restate the robust estimation result without logd factor below.

Lemma 31.

Let S be an ϵ-corrupted set of n samples from a distribution in d with mean μ and covariance Σσ2I. Let ϵ=Θ(log(1/τ)/n+ϵ)c be given, for a constant c>0. Then any stability-based algorithm on input S and ϵ, efficiently computes μ^ such that with probability at least 1τ, we have μ^μ=σO(ϵ+d/n+log(1/τ)/n).

We recall that our efficient implementation using grid points cost a logd factor due to the suboptimal net size. Using a net with a size matching the covering number O((D/ξ)d) will remove the logd factor, but at the cost of exponential time complexity for constructing the net and finding a point within O(ξ) distance for a given point.

Following the same proof steps, as in Appendix A, we can derive the excess risk bound without the logd factor, at the cost of exponential time complexity.

Appendix G Dealing with Unknown 𝝈

We adapt Algorithm 1 to work without knowing σ by first getting a lower bound on σ using the filtering algorithm (Algorithm 2) and then using this lower bound to set the fineness parameter ξ of the net in Algorithm 1.

The modified algorithm is as follows: (1) Estimate σ: Choose a point w and run Algorithm 2 with input S={fi(w)}i=1n to obtain a lower bound σ^. (2) Then, we run Algorithm 1 with ξ=σ^δ/β¯.

In Algorithm 1, we use σ only to determine the fineness of the net via ξ=σϵ/β¯. A smaller ξ results in a finer net and consequently reduces the error when evaluating gradients at the net point w instead of w, that is, (5) still holds with a smaller ξ. Since the excess risk depends on ξ only through logarithmic terms, the same analysis (see Appendix A) holds with a smaller ξ. It then suffices to choose ξ=σ^δ/β¯, where σ^ is a lower bound on σ. We also need to choose T=Ω~(βDσ^ϵ+σ^dlog(1/τ)n) where we use σ^ in place of σ. When using smoothing to handle nonsmooth losses, we can choose β similarly by replacing σ with σ^.

Recall that Algorithm 2 works even when σ is unknown. Moreover, the output h satisfies Σ(h)σ2(1+O(δ2/ϵ)) (see [7]). It follows that we can use Σ(h) to obtain a lower bound on σ. Using Lemma 24, at any fixed w, we can run Algorithm 2 with input S={fi(w)}i=1n to obtain a lower bound σ^ on σ. We have that (plugging in δ(τ) in Lemma 24), with probability at least 1τ,

Σ(h)σ2(1+O(δ2/ϵ)), (23)

where δ=O~(ϵ+d/n+log(1/τ)/n). Therefore, σ^:=Σ(h)/1+O(δ2/ϵ) is a lower bound on σ with probability at least 1τ.