Opinion - (2022) Volume 13, Issue 11
Received: 03-Nov-2022, Manuscript No. jbmbs-23-90219;
Editor assigned: 05-Nov-2022, Pre QC No. P-90219;
Reviewed: 18-Nov-2022, QC No. Q-90219;
Revised: 22-Nov-2022, Manuscript No. R-90219;
Published:
28-Nov-2022
, DOI: 10.37421/-2155-6180.2022.13.140
Citation: Mosy, Chand. “Quantitative Data Transformation Methods Used in Bio Statistical Research.” J Biom Biosta 13 (2022): 140.
Copyright: © 2022 Mosy C. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Normalization is the process of creating shifted and scaled versions of statistics with the goal of eliminating the effects of certain gross influences, like in an anomaly time series, by comparing the corresponding normalized values of different datasets (heterogenic data). As a result, the procedure for heterogeneous data transformation brings all attributes to the same scale. Indeed, the decimal scaling method is one of the quantitative data normalization techniques that moves the decimal point of the data's values. We divide each data value by the maximum absolute to normalize the data using this method. The original data are subjected to linear transformation in the minimum-maximum (Min-Max) data normalization method, whereas in the z-score data normalization procedure, values are normalized using the mean and standard deviation parameters. On the basis of these evidences, quantitative data standardization and normalization procedures may have distinct parametric distribution, such as the normal distribution, and data variability reduction capabilities.
Quantitative data used for the present study were drawn from previous experiments as described. Briefly, collected data included four growth parameters (diameter, plant height, leaf length and leaf number) of two maize varieties, treated by both rhizobacteria and foliar bio-fertilizing. Further, collected data for each treatment were summarized in a matrix including four columns describing variables parameters (two maize varieties growth parameters) and ninety-six rows corresponding to the observation number. Next, we submitted the above-mentioned data matrix to Box-Cox, Logarithm, Square Root, Inverse and Z-score, Minimum, Exponential and Minimum-Maximum quantitative data standardization as well as normalization (data transformation) procedures. Biometric verification is a method for checking a person's personality by using a piece of their identity, like their finger impression, facial features, or iris design. These features contain unique information that can't be duplicated. Despite their numerous benefits, certain biometrics, particularly facial recognition, have recently come under fire for being an infringement on privacy. Considering everything, your "face print" is your information, and many people don't like the idea that their face prints could be used or shared without their consent. This may eliminate the obscurity that many people anticipate in open areas, such as online. Even the idea of "connecting" a person's face to yet another source of personal data has been floated [1-5].
We focused on eight quantitative data transformation systems in the present comparative study. Processed quantitative data standardization and/or normalization procedures are as following Box-Cox (Box), Exponential (Expo), Inverse, Logarithmic normalization, Maximum, Minimum-Maximum, Square Root and Z-score. Above-mentioned data transformation systems was applied to the same data matrix (collected data) generating a new data set for each standardization and/or normalization methods. The present study provided a systematic comparative study that highlighted difference as well as similitude between eight quantitative data standardization methodologies providing useful tool to researchers, in choosing adequately data transformation methodologies that well fitting for their investigations. The same survey displayed smaller bias transformation by using the Box-Cox transformation as opposite to logarithm transformation. The same study revealed that the mean squared error of estimation is smaller with the Box-Cox transformation; and as well, the Box- Cox transformation leads to systematically higher estimated values than Logarithmic transformation. Hence, the Box-Cox transformation should be considered as a viable alternative in statistical modelling if the transformation of variables is required. Low aptitude with regard Exponential and Inverse data transformation in reducing data variability as well as in adjusting data normality could be due to processed positive value of analysed data. Indeed, our analysis suspected Exponential data transformation as a potential source of transformed data variability.
We thank the anonymous reviewers for their constructive criticisms of the manuscript. The support from ROMA (Research Optimization and recovery in the Manufacturing industry), of the Research Council of Norway is highly appreciated by the authors.
The authors declare that there was no conflict of interest in the present study.
Google Scholar, Crossref, Indexed at
Google Scholar, Crossref, Indexed at
Google Scholar, Crossref, Indexed at
Google Scholar, Crossref, Indexed at
Journal of Biometrics & Biostatistics received 3254 citations as per Google Scholar report