Data normalization in machine learning
In machine learning, the input data is often given in different dimensions. As a result of the scientific papers review, it is shown that the initial data described in different types of scales and units of measurement should be converted into a single representation by normalization or standardizat...
Main Authors: | , |
---|---|
Format: | Article |
Language: | Russian |
Published: |
The United Institute of Informatics Problems of the National Academy of Sciences of Belarus
2021-09-01
|
Series: | Informatika |
Subjects: | |
Online Access: | https://inf.grid.by/jour/article/view/1156 |
_version_ | 1797877140681654272 |
---|---|
author | V. V. Starovoitov Yu. I. Golub |
author_facet | V. V. Starovoitov Yu. I. Golub |
author_sort | V. V. Starovoitov |
collection | DOAJ |
description | In machine learning, the input data is often given in different dimensions. As a result of the scientific papers review, it is shown that the initial data described in different types of scales and units of measurement should be converted into a single representation by normalization or standardization. The difference between these operations is shown. The paper systematizes the basic operations presented in these scales, as well as the main variants of the function normalization. A new scale of parts is suggested and examples of the data normalization for correct analysis are given. Analysis of publications has shown that there is no universal method of data normalization, but normalization of the initial data makes it possible to increase the accuracy of their classification. It is better to perform data clustering by methods using distance functions after converting all features into a single scale. The results of classification and clustering by different methods can be compared with different scoring functions, which often have different ranges of values. To select the most accurate function, it is reasonable to normalize several functions and to compare their estimates on a single scale. The rules for separating features of tree-like classifiers are invariant to scales of quantitative features. Only comparison operation is used. Perhaps due to this property, the random forest classifier, as a result of numerous experiments, is recognized as one of the best classifiers in the analysis of data of different nature. |
first_indexed | 2024-04-10T02:13:37Z |
format | Article |
id | doaj.art-0f7da5371aa14a469fa1e29590b07205 |
institution | Directory Open Access Journal |
issn | 1816-0301 |
language | Russian |
last_indexed | 2024-04-10T02:13:37Z |
publishDate | 2021-09-01 |
publisher | The United Institute of Informatics Problems of the National Academy of Sciences of Belarus |
record_format | Article |
series | Informatika |
spelling | doaj.art-0f7da5371aa14a469fa1e29590b072052023-03-13T08:32:25ZrusThe United Institute of Informatics Problems of the National Academy of Sciences of BelarusInformatika1816-03012021-09-01183839610.37661/1816-0301-2021-18-3-83-96982Data normalization in machine learningV. V. Starovoitov0Yu. I. Golub1The United Institute of Informatics Problems, National Academy of Sciences of BelarusThe United Institute of Informatics Problems, National Academy of Sciences of BelarusIn machine learning, the input data is often given in different dimensions. As a result of the scientific papers review, it is shown that the initial data described in different types of scales and units of measurement should be converted into a single representation by normalization or standardization. The difference between these operations is shown. The paper systematizes the basic operations presented in these scales, as well as the main variants of the function normalization. A new scale of parts is suggested and examples of the data normalization for correct analysis are given. Analysis of publications has shown that there is no universal method of data normalization, but normalization of the initial data makes it possible to increase the accuracy of their classification. It is better to perform data clustering by methods using distance functions after converting all features into a single scale. The results of classification and clustering by different methods can be compared with different scoring functions, which often have different ranges of values. To select the most accurate function, it is reasonable to normalize several functions and to compare their estimates on a single scale. The rules for separating features of tree-like classifiers are invariant to scales of quantitative features. Only comparison operation is used. Perhaps due to this property, the random forest classifier, as a result of numerous experiments, is recognized as one of the best classifiers in the analysis of data of different nature.https://inf.grid.by/jour/article/view/1156object classificationclusteringdata normalizationfunction normalizationsigmoidhyperbolic tangentrandom forest |
spellingShingle | V. V. Starovoitov Yu. I. Golub Data normalization in machine learning Informatika object classification clustering data normalization function normalization sigmoid hyperbolic tangent random forest |
title | Data normalization in machine learning |
title_full | Data normalization in machine learning |
title_fullStr | Data normalization in machine learning |
title_full_unstemmed | Data normalization in machine learning |
title_short | Data normalization in machine learning |
title_sort | data normalization in machine learning |
topic | object classification clustering data normalization function normalization sigmoid hyperbolic tangent random forest |
url | https://inf.grid.by/jour/article/view/1156 |
work_keys_str_mv | AT vvstarovoitov datanormalizationinmachinelearning AT yuigolub datanormalizationinmachinelearning |