EURO 2024 Copenhagen
Abstract Submission

EURO-Online login

2402. Distributed Interpretable Machine Learning on GPUs

Invited abstract in session WC-4: Applications of Mixed-Integer and Nonconvex Optimization 1, stream MINLP.

Wednesday, 12:30-14:00
Room: 1001 (building: 202)

Authors (first author is the speaker)

1. Alireza Olama
Department of Information Technology, Åbo Akademi University
2. Andreas Lundell
Department of Information Technologies, Åbo Akademi University
3. Jan Kronqvist
Mathematics, KTH Royal Institute of Technology

Abstract

The research introduces the Bi-linear consensus Alternating Direction Method of Multipliers (Bi-cADMM) to address large-scale regularized Sparse Machine Learning (SML) problems over a network of computational nodes. These problems involve minimizing local convex loss functions over a global decision vector with an explicit $\ell_0$ norm constraint to ensure sparsity. This approach generalizes various sparse regression and classification models, including sparse linear and logistic regression, sparse softmax regression, and sparse support vector machines. Bi-cADMM reformulates the original non-convex SML problem using bi-linear consensus and employs a hierarchical decomposition strategy. This strategy splits the problem into smaller, parallel-computable sub-problems through a two-phase approach: initial sample decomposition and distribution of local datasets across nodes, followed by a delayed feature decomposition on GPUs available to each node. GPUs handle data-intensive computations, while CPUs manage less demanding tasks. The algorithm is implemented in an open-source Python package called Parallel Sparse Fitting Toolbox (PsFiT). Computational experiments validate the efficiency and scalability of Bi-cADMM through numerical benchmarks on various SML problems with distributed datasets.

Keywords

Status: accepted


Back to the list of papers