Learning Hierarchical Interactions at Scale: A Convex Optimization Approach

In many learning settings, it is beneficial toaugment the main features with pairwise in-teractions. Such interaction models can beoften enhanced by performing variable selec-tion under the so-calledstrong hierarchycon-straint: an interaction is non-zero only if itsassociated main features...

Full description

Bibliographic Details
Main Authors: Hazimeh, Hussein, Mazumder, Rahul
Other Authors: Sloan School of Management
Format: Article
Language:English
Published: International Machine Learning Society 2021
Online Access:https://hdl.handle.net/1721.1/130384
_version_ 1826201222517882880
author Hazimeh, Hussein
Mazumder, Rahul
author2 Sloan School of Management
author_facet Sloan School of Management
Hazimeh, Hussein
Mazumder, Rahul
author_sort Hazimeh, Hussein
collection MIT
description In many learning settings, it is beneficial toaugment the main features with pairwise in-teractions. Such interaction models can beoften enhanced by performing variable selec-tion under the so-calledstrong hierarchycon-straint: an interaction is non-zero only if itsassociated main features are non-zero. Ex-isting convex optimization-based algorithmsface difficulties in handling problems wherethe number of main featuresp∼103(withtotal number of features∼p2). In this pa-per, we study a convex relaxation which en-forces strong hierarchy and develop a highlyscalable algorithm based on proximal gradi-ent descent. We introduce novel screeningrules that allow for solving the complicatedproximal problem in parallel. In addition,we introduce a specialized active-set strategywith gradient screening for avoiding costlygradient computations. The framework can handle problems having dense design matri-ces, withp= 50,000 (∼109interactions)—instances that are much larger than state ofthe art. Experiments on real and syntheticdata suggest that our toolkithierScaleout-performs the state of the art in terms of pre-diction and variable selection and can achieveover a 4900x speed-up.
first_indexed 2024-09-23T11:48:08Z
format Article
id mit-1721.1/130384
institution Massachusetts Institute of Technology
language English
last_indexed 2024-09-23T11:48:08Z
publishDate 2021
publisher International Machine Learning Society
record_format dspace
spelling mit-1721.1/1303842022-10-01T06:09:25Z Learning Hierarchical Interactions at Scale: A Convex Optimization Approach Hazimeh, Hussein Mazumder, Rahul Sloan School of Management Massachusetts Institute of Technology. Operations Research Center In many learning settings, it is beneficial toaugment the main features with pairwise in-teractions. Such interaction models can beoften enhanced by performing variable selec-tion under the so-calledstrong hierarchycon-straint: an interaction is non-zero only if itsassociated main features are non-zero. Ex-isting convex optimization-based algorithmsface difficulties in handling problems wherethe number of main featuresp∼103(withtotal number of features∼p2). In this pa-per, we study a convex relaxation which en-forces strong hierarchy and develop a highlyscalable algorithm based on proximal gradi-ent descent. We introduce novel screeningrules that allow for solving the complicatedproximal problem in parallel. In addition,we introduce a specialized active-set strategywith gradient screening for avoiding costlygradient computations. The framework can handle problems having dense design matri-ces, withp= 50,000 (∼109interactions)—instances that are much larger than state ofthe art. Experiments on real and syntheticdata suggest that our toolkithierScaleout-performs the state of the art in terms of pre-diction and variable selection and can achieveover a 4900x speed-up. United States. Office of Naval Research (Grants ONR-N000141512342, ONR-N000141812298) National Science Foundation (U.S.) (Grant NSF-IIS1718258) 2021-04-06T13:49:15Z 2021-04-06T13:49:15Z 2020-06 2021-04-05T15:15:07Z Article http://purl.org/eprint/type/ConferencePaper 2640-3498 https://hdl.handle.net/1721.1/130384 Hazimeh, Hussein and Rahul Mazumder. “Learning Hierarchical Interactions at Scale: A Convex Optimization Approach.” Paper in the Proceedings of Machine Learning Research, 108, 23rd International Conference on Artificial Intelligence and Statistics (AISTATS) 2020, Palermo,Italy, June 3-5 2020, International Machine Learning Society © 2020 The Author(s) en http://proceedings.mlr.press/v108/hazimeh20a.html Proceedings of Machine Learning Research Article is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use. application/pdf International Machine Learning Society Proceedings of Machine Learning Research
spellingShingle Hazimeh, Hussein
Mazumder, Rahul
Learning Hierarchical Interactions at Scale: A Convex Optimization Approach
title Learning Hierarchical Interactions at Scale: A Convex Optimization Approach
title_full Learning Hierarchical Interactions at Scale: A Convex Optimization Approach
title_fullStr Learning Hierarchical Interactions at Scale: A Convex Optimization Approach
title_full_unstemmed Learning Hierarchical Interactions at Scale: A Convex Optimization Approach
title_short Learning Hierarchical Interactions at Scale: A Convex Optimization Approach
title_sort learning hierarchical interactions at scale a convex optimization approach
url https://hdl.handle.net/1721.1/130384
work_keys_str_mv AT hazimehhussein learninghierarchicalinteractionsatscaleaconvexoptimizationapproach
AT mazumderrahul learninghierarchicalinteractionsatscaleaconvexoptimizationapproach