Heterogeneous Treatment Effect with Trained Kernels of the Nadaraya–Watson Regression
A new method for estimating the conditional average treatment effect is proposed in this paper. It is called TNW-CATE (the Trainable Nadaraya–Watson regression for CATE) and based on the assumption that the number of controls is rather large and the number of treatments is small. TNW-CATE uses the N...
প্রধান লেখক: | , , |
---|---|
বিন্যাস: | প্রবন্ধ |
ভাষা: | English |
প্রকাশিত: |
MDPI AG
2023-04-01
|
মালা: | Algorithms |
বিষয়গুলি: | |
অনলাইন ব্যবহার করুন: | https://www.mdpi.com/1999-4893/16/5/226 |
_version_ | 1827742329629310976 |
---|---|
author | Andrei Konstantinov Stanislav Kirpichenko Lev Utkin |
author_facet | Andrei Konstantinov Stanislav Kirpichenko Lev Utkin |
author_sort | Andrei Konstantinov |
collection | DOAJ |
description | A new method for estimating the conditional average treatment effect is proposed in this paper. It is called TNW-CATE (the Trainable Nadaraya–Watson regression for CATE) and based on the assumption that the number of controls is rather large and the number of treatments is small. TNW-CATE uses the Nadaraya–Watson regression for predicting outcomes of patients from control and treatment groups. The main idea behind TNW-CATE is to train kernels of the Nadaraya–Watson regression by using a weight sharing neural network of a specific form. The network is trained on controls, and it replaces standard kernels with a set of neural subnetworks with shared parameters such that every subnetwork implements the trainable kernel, but the whole network implements the Nadaraya–Watson estimator. The network memorizes how the feature vectors are located in the feature space. The proposed approach is similar to transfer learning when domains of source and target data are similar, but the tasks are different. Various numerical simulation experiments illustrate TNW-CATE and compare it with the well-known T-learner, S-learner, and X-learner for several types of control and treatment outcome functions. The code of proposed algorithms implementing TNW-CATE is publicly available. |
first_indexed | 2024-03-11T04:00:28Z |
format | Article |
id | doaj.art-431028024daa43a192c0acee0cc7f70b |
institution | Directory Open Access Journal |
issn | 1999-4893 |
language | English |
last_indexed | 2024-03-11T04:00:28Z |
publishDate | 2023-04-01 |
publisher | MDPI AG |
record_format | Article |
series | Algorithms |
spelling | doaj.art-431028024daa43a192c0acee0cc7f70b2023-11-18T00:08:28ZengMDPI AGAlgorithms1999-48932023-04-0116522610.3390/a16050226Heterogeneous Treatment Effect with Trained Kernels of the Nadaraya–Watson RegressionAndrei Konstantinov0Stanislav Kirpichenko1Lev Utkin2Institute of Computer Science and Technology, Peter the Great St. Petersburg Polytechnic University, Polytechnicheskaya, 29, 195251 St. Petersburg, RussiaInstitute of Computer Science and Technology, Peter the Great St. Petersburg Polytechnic University, Polytechnicheskaya, 29, 195251 St. Petersburg, RussiaInstitute of Computer Science and Technology, Peter the Great St. Petersburg Polytechnic University, Polytechnicheskaya, 29, 195251 St. Petersburg, RussiaA new method for estimating the conditional average treatment effect is proposed in this paper. It is called TNW-CATE (the Trainable Nadaraya–Watson regression for CATE) and based on the assumption that the number of controls is rather large and the number of treatments is small. TNW-CATE uses the Nadaraya–Watson regression for predicting outcomes of patients from control and treatment groups. The main idea behind TNW-CATE is to train kernels of the Nadaraya–Watson regression by using a weight sharing neural network of a specific form. The network is trained on controls, and it replaces standard kernels with a set of neural subnetworks with shared parameters such that every subnetwork implements the trainable kernel, but the whole network implements the Nadaraya–Watson estimator. The network memorizes how the feature vectors are located in the feature space. The proposed approach is similar to transfer learning when domains of source and target data are similar, but the tasks are different. Various numerical simulation experiments illustrate TNW-CATE and compare it with the well-known T-learner, S-learner, and X-learner for several types of control and treatment outcome functions. The code of proposed algorithms implementing TNW-CATE is publicly available.https://www.mdpi.com/1999-4893/16/5/226treatment effectNadaraya–Watson regressionneural networkshared weightsmeta-learnerregression |
spellingShingle | Andrei Konstantinov Stanislav Kirpichenko Lev Utkin Heterogeneous Treatment Effect with Trained Kernels of the Nadaraya–Watson Regression Algorithms treatment effect Nadaraya–Watson regression neural network shared weights meta-learner regression |
title | Heterogeneous Treatment Effect with Trained Kernels of the Nadaraya–Watson Regression |
title_full | Heterogeneous Treatment Effect with Trained Kernels of the Nadaraya–Watson Regression |
title_fullStr | Heterogeneous Treatment Effect with Trained Kernels of the Nadaraya–Watson Regression |
title_full_unstemmed | Heterogeneous Treatment Effect with Trained Kernels of the Nadaraya–Watson Regression |
title_short | Heterogeneous Treatment Effect with Trained Kernels of the Nadaraya–Watson Regression |
title_sort | heterogeneous treatment effect with trained kernels of the nadaraya watson regression |
topic | treatment effect Nadaraya–Watson regression neural network shared weights meta-learner regression |
url | https://www.mdpi.com/1999-4893/16/5/226 |
work_keys_str_mv | AT andreikonstantinov heterogeneoustreatmenteffectwithtrainedkernelsofthenadarayawatsonregression AT stanislavkirpichenko heterogeneoustreatmenteffectwithtrainedkernelsofthenadarayawatsonregression AT levutkin heterogeneoustreatmenteffectwithtrainedkernelsofthenadarayawatsonregression |