7
Derivative-Free Optimization of High-Dimensional Non-Convex Functions by Sequential Random Embeddings * Hong Qian and Yi-Qi Hu and Yang Yu National Key Laboratory for Novel Software Technology, Nanjing University, Nanjing 210023, China {qianh,huyq,yuy}@lamda.nju.edu.cn Abstract Derivative-free optimization methods are suitable for sophisticated optimization problems, while are hard to scale to high dimensionality (e.g., larger than 1,000). Previously, the random embedding technique has been shown successful for solving high-dimensional problems with low effective di- mensions. However, it is unrealistic to assume a low effective dimension in many applications. This paper turns to study high-dimensional problems with low optimal ε-effective dimensions, which al- low all dimensions to be effective but many of them only have a small bounded effect. We character- ize the properties of random embedding for this kind of problems, and propose the sequential ran- dom embeddings (SRE) to reduce the embedding gap while running optimization algorithms in the low-dimensional spaces. We apply SRE to several state-of-the-art derivative-free optimization meth- ods, and conduct experiments on synthetic func- tions as well as non-convex classification tasks with up to 100,000 variables. Experiment results verify the effectiveness of SRE. 1 Introduction Solving sophisticated optimization problems plays an impor- tant role in artificial intelligence. Let f : R D R be a func- tion of which we assume that a global minimizer x * always exists. An optimization problem can be formally written as x * = argmin xR D f (x). We assume that the optimization problems discussed in this paper are deterministic, i.e., every call of f (x) returns the same value for the same x. In this paper, we focus on derivative-free optimization methods, which regard f as a black-box function that can only be evaluated point-wisely, i.e., they perform optimiza- tion based on the function values f (x) for the sampled solu- tions and other information like gradient is not used. Because * This research was supported by the NSFC (61375061, 61223003), Foundation for the Author of National Excellent Doc- toral Dissertation of China (201451), and 2015 Microsoft Research Asia Collaborative Research Program. these methods do not rely on derivatives, they are suitable for optimization problems that are, e.g., with many local optima, non-differentiable, and discontinuous, which are often en- countered in a wide range of applications. The performance of a derivative-free optimization algorithm can be evaluated by the simple regret [Bubeck et al., 2009], i.e., given n func- tion evaluations, for minimization, r n = f (x(n)) - min xR D f (x), where x(n) R D is the solution with the lowest function value found by the algorithm when the budget of n function evaluations is used up. The simple regret measures the differ- ence between the best function value found by the algorithm and the minimum of f . Many derivative-free optimization methods have been de- signed under various principles. They can be roughly cate- gorized into three kinds: model-based methods, determinis- tic Lipschitz optimization methods and meta-heuristic search. Model-based methods, such as Bayesian optimization meth- ods [Brochu et al., 2010; Snoek et al., 2012; Kawaguchi et al., 2015] and classification-based methods [Yu et al., 2016], learn a model from the solutions and the model is then applied to guide sampling of solutions for the next round. Determinis- tic Lipschitz optimization methods need Lipschitz continuity assumption on f , such as [Jones et al., 1993; Pint´ er, 1996; Bubeck et al., 2011; Munos, 2014]. Meta-heuristic search is designed with inspired heuristics, such as evolutionary strate- gies [Hansen et al., 2003]. Problem. Almost all derivative-free methods are effective and efficient in low-dimensional problems (usually less than 100 dimensions), but are hard to scale to high dimensionality (say, larger than 1,000 dimensions). This is mainly due to ei- ther the low convergence rate in high-dimensional space, thus unbearably many iterations are required; or the per-iteration computational cost is very high in high-dimensional space, thus it is unbearable for finishing a few iterations; or even both of the reasons. The unsatisfactory scalability is one of the main bottlenecks of these methods. Related Work. Recently, there are some studies focusing on improving the scalability of derivative-free methods. The two major directions are decomposition and embedding. Decomposition methods extract sub-problems from the original optimization problem, and by solving the sub- problems the original problem will be solved. In [Kandasamy

Derivative-Free Optimization of High-Dimensional Non-Convex … › ... › ijcai16-sre.pdf · 2016-04-20 · Derivative-Free Optimization of High-Dimensional Non-Convex Functions

  • Upload
    others

  • View
    4

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Derivative-Free Optimization of High-Dimensional Non-Convex … › ... › ijcai16-sre.pdf · 2016-04-20 · Derivative-Free Optimization of High-Dimensional Non-Convex Functions

Derivative-Free Optimization of High-Dimensional Non-Convex Functionsby Sequential Random Embeddings∗

Hong Qian and Yi-Qi Hu and Yang YuNational Key Laboratory for Novel Software Technology,

Nanjing University, Nanjing 210023, China{qianh,huyq,yuy}@lamda.nju.edu.cn

AbstractDerivative-free optimization methods are suitablefor sophisticated optimization problems, while arehard to scale to high dimensionality (e.g., largerthan 1,000). Previously, the random embeddingtechnique has been shown successful for solvinghigh-dimensional problems with low effective di-mensions. However, it is unrealistic to assume alow effective dimension in many applications. Thispaper turns to study high-dimensional problemswith low optimal ε-effective dimensions, which al-low all dimensions to be effective but many of themonly have a small bounded effect. We character-ize the properties of random embedding for thiskind of problems, and propose the sequential ran-dom embeddings (SRE) to reduce the embeddinggap while running optimization algorithms in thelow-dimensional spaces. We apply SRE to severalstate-of-the-art derivative-free optimization meth-ods, and conduct experiments on synthetic func-tions as well as non-convex classification tasks withup to 100,000 variables. Experiment results verifythe effectiveness of SRE.

1 IntroductionSolving sophisticated optimization problems plays an impor-tant role in artificial intelligence. Let f : RD → R be a func-tion of which we assume that a global minimizer x∗ alwaysexists. An optimization problem can be formally written as

x∗ = argminx∈RD f(x).

We assume that the optimization problems discussed in thispaper are deterministic, i.e., every call of f(x) returns thesame value for the same x.

In this paper, we focus on derivative-free optimizationmethods, which regard f as a black-box function that canonly be evaluated point-wisely, i.e., they perform optimiza-tion based on the function values f(x) for the sampled solu-tions and other information like gradient is not used. Because

∗This research was supported by the NSFC (61375061,61223003), Foundation for the Author of National Excellent Doc-toral Dissertation of China (201451), and 2015 Microsoft ResearchAsia Collaborative Research Program.

these methods do not rely on derivatives, they are suitable foroptimization problems that are, e.g., with many local optima,non-differentiable, and discontinuous, which are often en-countered in a wide range of applications. The performanceof a derivative-free optimization algorithm can be evaluatedby the simple regret [Bubeck et al., 2009], i.e., given n func-tion evaluations, for minimization,

rn = f(x(n))−minx∈RD f(x),

where x(n) ∈ RD is the solution with the lowest functionvalue found by the algorithm when the budget of n functionevaluations is used up. The simple regret measures the differ-ence between the best function value found by the algorithmand the minimum of f .

Many derivative-free optimization methods have been de-signed under various principles. They can be roughly cate-gorized into three kinds: model-based methods, determinis-tic Lipschitz optimization methods and meta-heuristic search.Model-based methods, such as Bayesian optimization meth-ods [Brochu et al., 2010; Snoek et al., 2012; Kawaguchi etal., 2015] and classification-based methods [Yu et al., 2016],learn a model from the solutions and the model is then appliedto guide sampling of solutions for the next round. Determinis-tic Lipschitz optimization methods need Lipschitz continuityassumption on f , such as [Jones et al., 1993; Pinter, 1996;Bubeck et al., 2011; Munos, 2014]. Meta-heuristic search isdesigned with inspired heuristics, such as evolutionary strate-gies [Hansen et al., 2003].

Problem. Almost all derivative-free methods are effectiveand efficient in low-dimensional problems (usually less than100 dimensions), but are hard to scale to high dimensionality(say, larger than 1,000 dimensions). This is mainly due to ei-ther the low convergence rate in high-dimensional space, thusunbearably many iterations are required; or the per-iterationcomputational cost is very high in high-dimensional space,thus it is unbearable for finishing a few iterations; or evenboth of the reasons. The unsatisfactory scalability is one ofthe main bottlenecks of these methods.

Related Work. Recently, there are some studies focusingon improving the scalability of derivative-free methods. Thetwo major directions are decomposition and embedding.

Decomposition methods extract sub-problems from theoriginal optimization problem, and by solving the sub-problems the original problem will be solved. In [Kandasamy

Page 2: Derivative-Free Optimization of High-Dimensional Non-Convex … › ... › ijcai16-sre.pdf · 2016-04-20 · Derivative-Free Optimization of High-Dimensional Non-Convex Functions

et al., 2015], additive functions were considered, i.e., thefunction value f(x) is the sum of several sub-functions withsmaller dimensions, and there is no variable overlaps be-tween any two sub-functions. In [Kandasamy et al., 2015],via employing a Bayesian optimization method, it was shownthat using the additive structure can effectively accelerate theBayesian optimization method. In [Friesen and Domingos,2015], a recursive decomposition method was proposed forapproximately locally decomposable problems. These meth-ods, however, rely on the (mostly axis-parallel) decompos-ability, which may restrict their applications.

Embedding methods assume that in the high-dimensionalspace, only a small subspace effects the function value.Therefore, optimization only in the effective subspace cansave a lot of efforts. In [Carpentier and Munos, 2012], com-pressed sensing was employed to deal with linear bandit prob-lems with low-dimensional effective subspaces. In [Chenet al., 2012], a variable selection method was proposed toidentify the effective axis-parallel subspace. In [Djolongaet al., 2013], a low-rank matrix recovery technique was em-ployed to learn the effective subspace. In [Wang et al., 2013;Qian and Yu, 2016], the random embedding based on the ran-dom matrix theory was employed to identify the underlyinglinear effective subspace. However, real-world problems maynot have a clear effective subspace, also it is hard to verify theexistence of the effective subspace.

Our Contributions. In this paper, we study high-dimensional problems with low optimal ε-effective dimen-sions (see Definition 1). In these problems, any (linear trans-formed) variable is allowed to effect the function value, how-ever, only a small linear subspace has a large impact onthe function value, and the orthogonal complement subspacemakes only a small bounded effect.

Firstly, we characterize the property of random embeddingfor this kind of problems. We find that, given optimal ε-effective dimension, single random embedding bears 2ε em-bedding gap. Note that this embedding gap cannot be com-pensated by the optimization algorithm.

We then propose the sequential random embeddings (SRE)to overcome the embedding gap. SRE applies the randomembedding several times sequentially, and in each subspace,SRE employs an optimization algorithm to reduce the residueof the previous solution. Therefore, SRE can also be viewedas a combination of decomposition and embedding, as eachrandom embedding defines a sub-problem. We also disclosethe condition under which SRE could improve the optimiza-tion quality for a large class of problems.

In experiments, we apply SRE to several state-of-the-artderivative-free optimization methods, and conduct experi-ments on synthetic functions as well as classification tasksusing the non-convex Ramp loss. Experiment results showthat SRE can significantly improve the performance of theoptimization methods in high-dimensional problems. More-over, comparing with previous related studies where testingfunctions are mostly up to 1,000 variables, the derivative-freemethods with SRE are tested for up to 100,000 variables, inreal-world data sets.

The consequent sections respectively introduce the optimalε-effective dimension problems and present the property of

random embedding, describe the proposed SRE technique aswell as its theoretical property, present the empirical results,and finally conclude the paper.

2 Optimal ε-Effective Dimension andRandom Embedding

Optimal ε-Effective DimensionEffective dimension defined in [Wang et al., 2013] requiresthe existence of a non-effective linear subspace, which hasexactly zero effect on the function value. It is often unrealis-tic to make such an assumption. We thus make a relaxation tothis assumption as the optimal ε-effective dimension in Defi-nition 1.

Note that a function with optimal ε-effective dimension canhave no low-dimensional effective subspace according to thedefinition given in [Wang et al., 2013], i.e., no linear subspacethat has exactly zero effect on the function value. Instead, ithas a linear subspace that makes at most ε contribution to thefunction value. Therefore, this kind of problems may still beefficiently tackled when ε is not so large.

DEFINITION 1 (Optimal ε-Effective Dimension)For any ε > 0, a function f : RD → R is said to havean ε-effective subspace Vε, if there exists a linear subspaceVε ⊆ RD s.t. for all x ∈ RD, we have |f(x) − f(xε)| ≤ ε,where xε ∈ Vε is the orthogonal projection of x onto Vε.Let Vε denote the collection of all the ε-effective subspacesof f , and dim(V) denote the dimension of a linear subspaceV . We define the optimal ε-effective dimension of f asdε = minVε∈Vε

dim(Vε).In the definition above, it should be noted that ε and dε

are related variables, commonly, a small dε implies a large ε,while a small ε implies a large dε.

Random EmbeddingGiven the definition of optimal ε-effective dimension,Lemma 1 below shows the effect of random embedding forsuch functions. For simplicity, let N denote the Gaussiandistribution with zero mean and σ2 variance.

LEMMA 1Given a function f : RD → R with optimal ε-effective di-mension dε, and any random matrix A ∈ RD×d (d ≥ dε)with independent entries sampled from N , then, with prob-ability 1, for any x ∈ RD, there exists y ∈ Rd s.t.|f(x)− f(Ay)| ≤ 2ε.Proof. We borrow the idea of constructing such y as in[Wang et al., 2013]. Since f has the optimal ε-effective di-mension dε, there exists an ε-effective subspace Vε ⊆ RDs.t. dim(Vε) = dε. Besides, any x ∈ RD can be decom-posed as x = xε + x⊥ε , where xε ∈ Vε, x⊥ε ∈ V⊥ε and V⊥εis the orthogonal complement of Vε. By the definition of ε-effective subspace, we have |f(x) − f(xε)| ≤ ε. Hence, itsuffices to show that, for any xε ∈ Vε, there exists y ∈ Rds.t. |f(xε)− f(Ay)| ≤ ε.

Let Φ ∈ RD×dε be a matrix whose columns form a stan-dard orthonormal basis for Vε. For any xε ∈ Vε, there ex-ists c ∈ Rdε s.t. xε = Φc. Let us for now assume that

Page 3: Derivative-Free Optimization of High-Dimensional Non-Convex … › ... › ijcai16-sre.pdf · 2016-04-20 · Derivative-Free Optimization of High-Dimensional Non-Convex Functions

Φ>A has rank dε. If rank(Φ>A) = dε, there must ex-ist y ∈ Rd s.t. (Φ>A)y = c, because rank(Φ>A) =rank([Φ>A, c]). The orthonormal projection of Ay onto Vεis given by ΦΦ>Ay = Φc = xε. Thus, Ay = xε+x wherex ∈ V⊥ε for xε is the orthonormal projection of Ay onto Vε.Since Ay ∈ RD, by the definition of ε-effective subspace, wehave |f(xε)− f(Ay)| ≤ ε, and thus |f(x)− f(Ay)| ≤ 2ε.

Similar to the proof of Theorem 2 in [Wang et al., 2013],we have that rank(Φ>A) = dε with probability 1. �

Lemma 1 implies that, given a random embedding matrixA ∈ RD×d, for any minimizer x∗ ∈ RD, there exists y ∈ Rdsuch that f(Ay)−f(x∗) ≤ 2ε. Note that this embedding gapgrows twice as fast as ε.

Optimization with Random EmbeddingVia random embedding, current derivative-free optimizationalgorithms can be applied to solve high-dimensional opti-mization problems by running in a low-dimensional solutionspace. Given a high-dimensional function f and a randommatrix A ∈ RD×d, we construct a new optimization functiong(y) = f(Ay), where the solution space for g only has a di-mension d. Note that every solution is evaluated by mappingback to the original high-dimensional space.

For functions with a low optimal ε-effective dimension, wecan bound the gap between the optimal function values of gand f based on Lemma 1, which is stated in Theorem 1. Weomit the proof since it can be verified directly.

THEOREM 1Given a function f : RD → R with optimal ε-effective di-mension dε, and a random matrix A ∈ RD×d (d ≥ dε) withindependent entries sampled fromN , let x∗ be a global min-imizer of f and g(y) = f(Ay) where y ∈ Rd, then, withprobability 1, we have

infy∈Rd

g(y)− f(x∗) ≤ 2ε.

We then look into the simple regret of a derivative-free op-timization algorithm using a random embedding. The opti-mization algorithm works in the low-dimensional space, andusually, it can only approximate the optimal solution, thusthere is a gap between the found solution y and the optimalsolution in the low-dimensional space infy∈Rd g(y). This ap-proximation gap is related to dimension size, function com-plexity, and optimization budget. We assume that the approx-imation gap is upper bounded by θ. Furthermore, as disclosedby Theorem 1, there exists an embedding gap 2ε, which can-not be compensated by the optimization algorithm. We thenhave that the simple regret of the algorithm is upper boundedby the approximation gap and the embedding gap,

g(y)− f(x∗) = g(y)− infy∈Rd

g(y) + infy∈Rd

g(y)− f(x∗)

≤ θ + 2ε. (1)

3 Sequential Random EmbeddingsFrom Eq.(1), in order to improve the simple regret, i.e.,g(y)− f(x∗), we need to reduce the approximation gap andthe embedding gap. Unfortunately, this two factors are con-flicting: In order to reduce the embedding gap, we need to

use a large dε to trade for a smaller ε, but with a large dε, theoptimization algorithm needs to deal with more dimensions,which will badly increase the approximation gap.

Let Si = {A(i)y |y ∈ Rd} denote the subspace defined bythe random matrix, where i = 1, . . . ,m. In spired by [Zhanget al., 2013], we propose the technique of sequentially usingmultiple random embeddings to reduce the embedding gapwhile keeping the approximation gap almost unaffected:• In the first step, it generates a random matrix A(1) whichdefines a subspace S1, and applies the chosen optimizationalgorithm to find a near-optimal solution in the subspace, i.e.,y1 = argminy f(A

(1)y). Let x2 = A(1)y1 be the high-dimensional solution.• In the second step, it generates another random matrix A(2)

which defines a subspace S2, and applies the algorithm to op-timize the residue of the current solution x2 in the subspace,i.e., y2 = argminy f(x2 +A(2)y). Then update the currentsolution x3 = x2 +A(2)y2.• In the following steps, it acts like the second step that per-forms the optimization to reduce the residue.

By this simple sequence of applying random embeddings,we show that, for a large class of functions with optimalε-effective dimension, an upper bound of the solution gapf(xi)− f(x∗) reduces as the step i increases.

Theoretical PropertyLet x1 = 0, and ‖ · ‖ represent ‖ · ‖2 for simplicity. In thestep i of the sequential random embeddings, the residue so-lution to be approximated is x∗ − xi, let xi be the orthonor-mal projection of x∗ − xi onto the subspace Si. Thus, wecall ‖xi‖/‖x∗− xi‖ as the embedding ratio which is smallerthan 1, and ‖xi −A(i)yi‖/‖xi‖ as the optimization ratio forthis step. We show in Proposition 1 that, when the optimiza-tion ratio is less than a fraction of the embedding ratio, thesolution is closer to the optimal solution.

PROPOSITION 1Given f ∈ F , and a sequence of random matrices {A(i)}i ⊆RD×d (d ≥ dε) each with independent entries sampled fromN , for all i = 1, . . . ,m, if

‖xi −A(i)yi‖/‖xi‖ ≤ (1/5) · ‖xi‖/‖x∗ − xi‖,

it holds that ‖x∗ − xi‖ > ‖x∗ − xi+1‖ .

Proof. For any i = 1, . . . ,m, since xi is the orthonormalprojection of x∗ − xi onto Si, we have that

‖x∗ − xi‖2 = ‖x∗ − xi − xi‖2 + ‖xi‖2

≥ (‖x∗ − xi −A(i)yi‖ − ‖xi −A(i)yi‖)2 + ‖xi‖2

= ‖x∗ − xi+1‖2 + ‖xi −A(i)yi‖2 + ‖xi‖2

− 2‖x∗ − xi −A(i)yi‖ · ‖xi −A(i)yi‖≥ ‖x∗ − xi+1‖2 + (‖xi −A(i)yi‖ − ‖xi‖)2

− 2(‖x∗ − xi‖+ ‖A(i)yi‖) · ‖xi −A(i)yi‖+ 2‖xi‖ · ‖xi −A(i)yi‖≥ ‖x∗ − xi+1‖2 + (‖xi −A(i)yi‖ − ‖xi‖)2

Page 4: Derivative-Free Optimization of High-Dimensional Non-Convex … › ... › ijcai16-sre.pdf · 2016-04-20 · Derivative-Free Optimization of High-Dimensional Non-Convex Functions

− 2‖x∗ − xi‖ · ‖xi −A(i)yi‖ − 2‖xi −A(i)yi‖2,

where the last inequality is by ‖A(i)yi‖ − ‖xi‖ ≤ ‖xi −A(i)yi‖.

Since ‖xi−A(i)yi‖·‖x∗−xi‖/‖xi‖2 ≤ 1/5 and ‖xi‖ ≤‖x∗− xi‖, we have that (‖xi−A(i)yi‖− ‖xi‖)2− 2‖x∗−xi‖ · ‖xi − A(i)yi‖ − 2‖xi − A(i)yi‖2 > 0. Therefore,‖x∗ − xi‖ > ‖x∗ − xi+1‖ for all i = 1, . . . ,m. �

We then consider a very general class of problems, i.e.,functions with local Holder continuity as in Definition 2.

DEFINITION 2 (Local Holder Continuity)There exists L,α > 0 s.t., for all x ∈ RD, f(x) − f(x∗) ≤L · ‖x− x∗‖α2 , where x∗ is a global minimizer of f .

Intuitively, Definition 2 means that the rate of increase of faround x∗ is bounded. Note that a function with local Holdercontinuity can have many local optima, or non-differentiable.

For any function with local Holder continuity, since

f(xi)− f(x∗) ≤ L · ‖xi − x∗‖α2 ,

Proposition 1 implies that the sequential random embeddingscan reduce this upper bound in a mild condition.

Less GreedyBy the method of sequential random embeddings (SRE), eachsub-problem in the subspace is solved greedily. However, aperfect solution for one sub-problem may not be good glob-ally, and once an unsatisfied solution is found, it is hard to becorrected by the later solutions because of the greedy process.Therefore, we further introduce a withdraw variable α to theprevious solution such that it is possible to eliminate the pre-vious solution if necessary, and the optimization problem ineach step becomes

miny,α

f(αxi +A(i)y).

Since derivative-free optimization methods are employed,which make few requirements on the optimization problems,we can simply let the algorithm optimize α together with y.

The full algorithm named SRE is depicted in Algorithm 1.Given a derivative-free optimization algorithm M, in eachstep of SRE,M optimizes a low-dimensional function gi(y)as defined in line 4, and returns the best solution as well asthe withdraw variable found byM (line 5). Finally, the bestsolution throughout the procedure is returned as the output.

4 ExperimentsFor the practical issues in experiments, we set the randommatrix A ∈ RD×d with independent entries sampled fromN (0, 1/d), and we consider the high-dimensional solutionspace X = [−u, u]D instead of RD and low-dimensional so-lution space Y = [−l, l]d instead of Rd with u, l > 0 for sim-plicity. To implement the derivative-free algorithm in Y , notethat there may exist y′ ∈ Y s.t. Ay′ /∈ X and thus f cannotbe evaluated at point Ay′. To tackle this problem, we em-ploy Euclidean projection, i.e., Ay′ is projected to X whenit is outside X by PX (Ay′) = argminx∈X ‖x−Ay′‖2. Letf(PX (Ay′)) + ‖PX (Ay′)−Ay′‖1 be the function value of

Algorithm 1 Sequential Random Embeddings (SRE)Input:

Objective function f ;Derivative-free optimization algorithmM;Number of function evaluation budget n;Upper bound of the optimal ε-effective dimension d;Number of sequential random embeddings m.

Procedure:1: x1 = 0.2: for i = 1 to m do3: Sample a random matrix A(i)∈RD×d with Ai,j ∼ N .4: Apply M to optimize the low-dimensional function

gi(y) = f(αxi+A(i)y) with nm function evaluations.

5: Obtain the solution yi and αi for gi(y) byM.6: xi+1 = αixi +A(i)yi.7: end for8: return argmini=2,...,m+1 f(xi).

Ay′ where Ay′ /∈ X . Let [x]i denote the i-th coordinate ofx. Since X = [−u, u]D, the closed form of Euclidean pro-jection is [PX (Ay)]i = −u if [Ay]i < −u; [PX (Ay)]i =[Ay]i if −u ≤ [Ay]i ≤ u; and [PX (Ay)]i = u otherwise.

We employ three state-of-the-art derivative-free optimiza-tion methods: IMGPO [Kawaguchi et al., 2015], which isa combination of Bayesian optimization and optimistic opti-mization, CMAES [Hansen et al., 2003], which is a repre-sentative of evolutionary algorithms, and RACOS [Yu et al.,2016], which is a model-based optimization algorithm pro-posed recently. All implementations of them are by their au-thors. We also employ random search as a reference baseline.

When we apply the single random embedding with theseoptimization algorithms, we denote by the prefix “RE-”,and when we apply the SRE, we denote by the prefix“SRE-”. Thus we have combinations including RE-IMGPO,RE-CMAES, RE-RACOS, SRE-IMGPO, SRE-CMAES, andSRE-RACOS.

On Synthetic FunctionsWe first test the algorithms on two synthetic testing functions.Based on the convex Sphere function and the highly non-convex Ackley function, we construct the high-dimensionalSphere function and Ackley function to meet the optimal ε-effective dimension assumption.

The high-dimensional Sphere function is constructed as

f1(x) =∑10i=1 ([x]i − 0.2)

2+ 1

D

∑Di=11 ([x]i − 0.2)

2,where the dimensions except the first 10 ones have limitedimpact on the function value. The high-dimensional Ackleyfunction is similarly constructed as

f2(x) = −20 exp(− 1

5

√110

∑10i=1 ([x]i − 0.2)2

)− exp

(110

∑10i=1 cos 2π([x]i − 0.2)

)+exp(1)+20

+ 1D

∑Di=11 ([x]i − 0.2)

2.

Noted that the optimal solutions of these functions are(0.2, . . . , 0.2), in order to avoid the all-zero optimal solu-tion which is in all the linear subspaces. These functions are

Page 5: Derivative-Free Optimization of High-Dimensional Non-Convex … › ... › ijcai16-sre.pdf · 2016-04-20 · Derivative-Free Optimization of High-Dimensional Non-Convex Functions

minimized within the solution space X = [−1, 1]D. We setY = [−1, 1]d, α ∈ [−1, 1]. Each algorithm is run 30 timesindependently, and the average performance is reported.On the number of random embeddings m. We first studythe effect of SRE iteration number m, i.e., the number of se-quential random embeddings. We choose D = 10000, set thetotal number of function evaluations n = 10000 and the sub-space size d = 10, and choose the number of sequential ran-dom embeddings m = {1, 2, 5, 8, 10, 20}. When m = 1, al-gorithms with SRE degenerate into algorithms with RE. Theachieved objective function values are shown in Figure 1.

0 5 10 15 20Number of random embeddings m

0

0.1

0.2

0.3

Obj

ectiv

e fu

nctio

n va

lue SRE-IMGPO

SRE-CMAESSRE-RACOS

(a) on Sphere function

0 5 10 15 20 Number of random embeddings m

0

0.5

1

1.5

Obj

ectiv

e fu

nctio

n va

lue SRE-IMGPO

SRE-CMAES SRE-RACOS

(b) on Ackley function

Figure 1: On the effect of the number of random embeddingsm.

Figure 1 (a) and (b) show that, if the total number of func-tion evaluations is fixed, we should choose a compromisedvalue for m. Because if m is large then the budget for eachstep of SRE is limited, and ifm is small then the steps in SREis limited, and both of them can lead to unsatisfied optimiza-tion performance.On subspace dimension d. To study how low-dimensionalsize d affects optimization performance, we only adopt thealgorithms with SRE (SRE-IMGPO, SRE-CMAES, SRE-RACOS). We choose D = 10000, set the total num-ber of function evaluations n = 10000, choose d ={1, 5, 8, 10, 12, 15, 20}, and set the number of sequential ran-dom embeddings m = 5. The achieved objective functionvalues are shown in Figure 2.

0 5 10 15 20Subspace dimension d

0

0.2

0.4

0.6

Obj

ectiv

e fu

nctio

n va

lue SRE-IMGPO

SRE-CMAES SRE-RACOS

(a) on Sphere function

0 5 10 15 20Subspace dimension d

0

0.5

1

1.5

2

Obj

ectiv

e fu

nctio

n va

lue SRE-IMGPO

SRE-CMAES SRE-RACOS

(b) on Ackley function

Figure 2: On the effect of the subspace dimension d.

Figure 2 (a) and (b) show that, for algorithms with SRE, inmost cases the closer the d to dε the better the optimizationperformance, indicating that a good estimate of the optimalε-effective dimension is desirable. Besides, we can observethat, even if d < dε = 10 but close to dε, the performancesof algorithms with SRE are still satisfied.On scalability. We then study the scalability w.r.t.the solution space dimensions D, we choose D =

{100, 500, 1000, 5000, 10000}, set the total number of func-tion evaluations n = 10000 for all algorithms, set d = 10 foralgorithms with RE and SRE, and set the number of sequen-tial random embeddingsm = 5 for algorithms with SRE. Theachieved objective function values are shown in Figure 3.

SRE-IMGPO SRE-CMAES SRE-RACOS

Random search CMAES RACOS

RE-IMGPO RE-CMAES RE-RACOS

102 104103 0

0.05

0.1

0.15

Obj

ectiv

e fu

nctio

n va

lue

0.50.6

0.7

0.8

0.9

1

Dimension D

(a) on Sphere function

102 104103 0

0.1

0.2

0.3

0.4

Obj

ectiv

e fu

nctio

n va

lue

1

1.5

2

2.5

3

Dimension D

(b) on Ackley function

Figure 3: Comparing the scalability with n = 10000 functionevaluations.

Figure 3 (a) and (b) show that the algorithms with SREhave the lowest growing rate, while the algorithms withoutRE have the highest growing rate as the dimension increases,indicating that SRE can scale the derivative-free algorithmsto high-dimensional problems better than the compared algo-rithms.On convergence rate. To study the convergence rate w.r.t.the number of function evaluations, we set D = 10000,and set the total number of function evaluations n ={2000, 4000, 6000, 8000, 10000}, set low-dimensional sized = 10 for RE as well as SRE, and set m = 5 for SRE. Theachieved objective function values are shown in Figure 4.

2000 4000 6000 8000 10000Number of function evaluations

10-1

100

Obj

ectiv

e fu

nctio

n va

lue

(a) on Sphere function

2000 4000 6000 8000 10000Number of function evaluations

10-1

100

Obj

ectiv

e fu

nctio

n va

lue

(b) on Ackley function

Figure 4: Comparing the convergence rate with D = 10000.The Y-axis is in log-scale. The legend is shared with Figure 3.

Figure 4 (a) and (b) show that, in most cases, algorithmswith SRE reduce the objective function value with the highestrate, while the algorithms without RE reduce the objectivefunction value with the lowest rate, indicating that algorithmswith SRE converge faster in general than the others.

On Classification with Ramp LossWe finally study on a classification task with Ramp loss [Col-lobert et al., 2006]. The Ramp loss is defined as Rs(z) =

Page 6: Derivative-Free Optimization of High-Dimensional Non-Convex … › ... › ijcai16-sre.pdf · 2016-04-20 · Derivative-Free Optimization of High-Dimensional Non-Convex Functions

0 2 4 6 8 10C

0

2

4

Obj

ectiv

e fu

nctio

n va

lue

6 *104

CCCP

(a) on Gisette, s = −1

0 2 4 6 8 10C

0

200

400

600

800

1000

Obj

ectiv

e fu

nctio

n va

lue CCCP

(b) on Arcene, s = −1

0 2 4 6 8 10C

0

1000

2000

3000

Obj

ectiv

e fu

nctio

n va

lue CCCP

(c) on Dexter, s = −1

0 2 4 6 8 10C

0

2000

4000

6000

8000

Obj

ectiv

e fu

nctio

n va

lue CCCP

(d) on Dorothea, s = −1

0 2 4 6 8 10C

0

1

2

3

Obj

ectiv

e fu

nctio

n va

lue

4 *104

CCCP

(e) on Gisette, s = 0

0 2 4 6 8 10C

0

200

400

600

800

1000

Obj

ectiv

e fu

nctio

n va

lue CCCP

(f) on Arcene, s = 0

0 2 4 6 8 10C

0

1000

2000

3000

Obj

ectiv

e fu

nctio

n va

lue CCCP

(g) on Dexter, s = 0

0 2 4 6 8 10C

0

2000

4000

6000

8000

Obj

ectiv

e fu

nctio

n va

lue CCCP

(h) on Dorothea, s = 0

Figure 5: Comparing the achieved objective function values against the parameter C of the classification with Ramp loss. Thelegend is shared with Figure 3 except CCCP.

H1(z) −Hs(z) with s < 1, where Hs(z) = max{0, s − z}is the Hinge loss with s being the Hinge point. The task is tofind a vector w and a scalar b to minimizef(w, b) = 1

2‖w‖22 + C

∑L`=1Rs

(y`(w

>v` + b)),

where v` is the training instance and y` ∈ {−1,+1} is itslabel. This objective function is similar to that of supportvector machines (SVM) [Vapnik, 2000], but the loss functionof SVM is the Hinge loss. Due to the convexity of the Hingeloss, the number of support vectors increases linearly withthe number of training instances in SVM, which is undesiredwith respect to scalability. While this problem can be relievedby employing the Ramp loss [Collobert et al., 2006].

The applied algorithms here are RE-IMGPO, RE-CMAES,RE-RACOS, SRE-IMGPO, SRE-CMAES, SRE-RACOS andthe concave-convex procedure (CCCP) [Yuille and Rangara-jan, 2001] which is a gradient-based non-convex optimiza-tion approach for objective functions that can be decomposedinto convex sub-function plus concave sub-function. We em-ploy four binary class UCI datasets [Blake et al., 1998],Gisette, Arcene, Dexter and Dorothea. The feature dimen-sion (D) of which are 5 × 103, 104, 2 × 104, and 105, re-spectively. Since there are two hyper-parameters in the opti-mization formulation, i.e., C and s, to study the effectivenessof the compared algorithms under different hyper-parameters,we test s ∈ {−1, 0} and C ∈ {0.1, 0.5, 1, 2, 5, 10}. Weset d = 20, the number of function evaluations n = 3D,and X = [−10, 10]D,Y = [−10, 10]d, α ∈ [−10, 10] forall applied algorithms except for CCCP. For CCCP, we setX = [−10, 10]D and let CCCP run until it converges. For al-gorithms with SRE, we set the number of sequential randomembeddings m = 5. Each randomized algorithm is repeated30 times independently. Since we focus on the optimizationperformance, the achieved objective function values on eachdataset are reported in Figure 5.

We should note that, in Figure 5, the lines of RE-IMGPOoften overlap with those of SRE-IMPGO, and in sub-figure

(c), the line of CCCP overlaps with SRE-IMPGO. As shownin Figure 5, except in the dataset Arcene, algorithm with SREhas the best performance for any setting of s and C. Andin the dataset Arcene, algorithm with RE achieves the bestperformance. This verifies the effectiveness of SRE and RE,and indicates that SRE is more effective than RE. Compar-ing CCCP to the derivative-free algorithms with SRE, we canobserve that the objective function value of CCCP is approx-imately as twice as that of the best algorithm with SRE. Thisimplies that algorithms with SRE can be significantly betterthan CCCP with respect to optimization performance.

5 ConclusionThis paper investigates high-dimensional problems in a moregeneral situation compared with previous works, i.e., all vari-ables can be effective to the function value, while many ofthem only have a small bounded effect. We define this kind ofproblems as functions with a low optimal ε-effective dimen-sion, and find that, for this kind of problems, single randomembedding makes 2ε loss that can not be compensated by thefollowing optimization algorithm. Thus, we propose the se-quential random embeddings (SRE) that employ the randomembedding several times, and the condition under which SREcan reduce the embedding loss strictly in each step of SRE isdisclosed. To study the effectiveness of SRE, we apply SREto several state-of-the-art derivative-free algorithms. Exper-iment results on optimization testing functions and classifi-cation with the non-convex Ramp loss (where the number ofvariables are up to 100,000) indicate that SRE can scale thesederivative-free methods to high-dimensional problems signif-icantly.

References[Blake et al., 1998] C. L. Blake, E. Keogh, and C. J.

Merz. UCI Repository of machine learning databases.

Page 7: Derivative-Free Optimization of High-Dimensional Non-Convex … › ... › ijcai16-sre.pdf · 2016-04-20 · Derivative-Free Optimization of High-Dimensional Non-Convex Functions

[http://www.ics.uci.edu/∼mlearn/MLRepository.html],1998.

[Brochu et al., 2010] E. Brochu, V. M. Cora, and N. de Fre-itas. A tutorial on bayesian optimization of expensive costfunctions, with application to active user modeling and hi-erarchical reinforcement learning. CoRR, abs/1012.2599,2010.

[Bubeck et al., 2009] S. Bubeck, R. Munos, and G. Stoltz.Pure exploration in multi-armed bandits problems. In Pro-ceedings of the 20th International Conference on Algorith-mic Learning Theory, pages 23–37, Porto, Portugal, 2009.

[Bubeck et al., 2011] S. Bubeck, G. Stoltz, and J. Y. Yu. Lip-schitz bandits without the lipschitz constant. In Proceed-ings of the 22nd International Conference on AlgorithmicLearning Theory, pages 144–158, Espoo, Finland, 2011.

[Carpentier and Munos, 2012] A. Carpentier and R. Munos.Bandit theory meets compressed sensing for high dimen-sional stochastic linear bandit. In Proceedings of the15th International Conference on Artificial Intelligenceand Statistics, pages 190–198, La Palma, Canary Islands,2012.

[Chen et al., 2012] B. Chen, R. M. Castro, and A. Krause.Joint optimization and variable selection of high-dimensional gaussian processes. In Proceedings of the29th International Conference on Machine Learning,pages 1423–1430, Edinburgh, Scotland, 2012.

[Collobert et al., 2006] R. Collobert, F. Sinz, J. Weston, andL. Bottou. Trading convexity for scalability. In Pro-ceedings of the 23rd International Conference on MachineLearning, pages 201–208, Pittsburgh, Pennsylvania, 2006.

[Djolonga et al., 2013] J. Djolonga, A. Krause, andV. Cevher. High-dimensional gaussian process bandits. InAdvances in Neural Information Processing Systems 26,pages 1025–1033, Lake Tahoe, Nevada, 2013.

[Friesen and Domingos, 2015] A. L. Friesen and P. M.Domingos. Recursive decomposition for nonconvex op-timization. In Proceedings of the 24th InternationalJoint Conference on Artificial Intelligence, pages 253–259, 2015.

[Hansen et al., 2003] N. Hansen, S. D. Muller, andP. Koumoutsakos. Reducing the time complexity ofthe derandomized evolution strategy with covariancematrix adaptation (CMA-ES). Evolutionary Computation,11(1):1–18, 2003.

[Jones et al., 1993] D. R. Jones, C. D. Perttunen, and B. E.Stuckman. Lipschitzian optimization without the lipschitzconstant. Journal of Optimization Theory and Applica-tions, 79(1):157–181, 1993.

[Kandasamy et al., 2015] K. Kandasamy, J. Schneider, andB. Poczos. High dimensional bayesian optimisation andbandits via additive models. In Proceedings of the 32nd In-ternational Conference on Machine Learning, pages 295–304, Lille, France, 2015.

[Kawaguchi et al., 2015] K. Kawaguchi, L. P. Kaelbling, andT. Lozano-Perez. Bayesian optimization with exponen-tial convergence. In Advances in Neural Information Pro-cessing Systems 28, pages 2791–2799, Montreal, Canada,2015.

[Munos, 2014] R. Munos. From bandits to Monte-Carlo TreeSearch: The optimistic principle applied to optimizationand planning. Foundations and Trends in Machine Learn-ing, 7(1):1–130, 2014.

[Pinter, 1996] J. D. Pinter. Global Optimization in Action,Continuous and Lipschitz Optimization: Algorithms, Im-plementations and Applications. Nonconvex Optimizationand Its Applications. Kluwer Academic Publishers, Dor-drecht, Boston, London, 1996.

[Qian and Yu, 2016] H. Qian and Y. Yu. Scaling simulta-neous optimistic optimization for high-dimensional non-convex functions with low effective dimensions. In Pro-ceedings of the 30th AAAI Conference on Artificial Intelli-gence, Phoenix, AZ, 2016.

[Snoek et al., 2012] J. Snoek, H. Larochelle, and R. P.Adams. Practical bayesian optimization of machine learn-ing algorithms. In Advances in Neural InformationProcessing Systems 25, pages 2960–2968, Lake Tahoe,Nevada, 2012.

[Vapnik, 2000] V. Vapnik. The Nature of Statistical LearningTheory. Springer Science & Business Media, 2000.

[Wang et al., 2013] Z. Wang, M. Zoghi, F. Hutter, D. Math-eson, and N. De Freitas. Bayesian optimization in highdimensions via random embeddings. In Proceedings ofthe 23rd International Joint Conference on Artificial Intel-ligence, pages 1778–1784, Beijing, China, 2013.

[Yu et al., 2016] Y. Yu, H. Qian, and Y.-Q. Hu. Derivative-free optimization via classification. In Proceedings of the30th AAAI Conference on Artificial Intelligence, Phoenix,AZ, 2016.

[Yuille and Rangarajan, 2001] A. L. Yuille and A. Rangara-jan. The concave-convex procedure (CCCP). In Ad-vances in Neural Information Processing Systems 14,pages 1033–1040, Vancouver, British Columbia, Canada,2001.

[Zhang et al., 2013] L. Zhang, M. Mahdavi, R. Jin, T. Yang,and S. Zhu. Recovering the optimal solution by dual ran-dom projection. In Proceedings of the 26th Conference onLearning Theory, pages 135–157, Princeton, NJ, 2013.