Parallel and Distributed Sparse Optimization

Links

Background

Modern datasets usually have a large number of features or training samples, and they are usually stored in a distributed manner. Motivated by the need of solving sparse optimization problems with large datasets, we propose two approaches including (i) distributed implementations of prox-linear algorithms and (ii) GRock, a parallel greedy coordinate descent method.

Codes and demos

Three parallel C solvers for LASSO

More codes and applications will be uploaded in the coming weeks

Solving LASSO with a 170GB matrix on Amazon EC2

Test dataset

Requested systems on Amazon EC2

LASSO

We tried to recover x^o by solving

min_x lambda|x|_1 + frac{1}{2}|Ax - b|_2^2

in which lambda and b were set so that x^o is the exact solution

Tested algorithms

parallel 

stopping creterions include:

Sparse logistic regression on a cluster

Test dataset

System

Sparse logistic regression model

We tried to recover a hyperplane h(x)=w^Tx+c for classification with a sparse normal vector w. The model is

min_{w,c} lambda|w|_1 + frac{1}{m}sum_{i=1}^mlogleft(1+expleft(-b_i(w^Ta_i + cright)right)

for given feature data points a_i and labels b_i, i=1,ldots,m.

Tested algorithms

cores p FISTA seconds GRock seconds
1 8.74 1.74
8 2.24 0.48
16 0.6 0.13
64 0.24 0.05
128 0.24 0.07

Citation

Z. Peng, M. Yan, and W. Yin. Parallel and Distributed Sparse Optimization, preprint, 2013


« Back