GET THE APP

A Transformational Viewpoint on Conditional and Marginal Models
..

Journal of Biometrics & Biostatistics

ISSN: 2155-6180

Open Access

Opinion - (2023) Volume 14, Issue 1

A Transformational Viewpoint on Conditional and Marginal Models

Mindi Mossy*
*Correspondence: Mindi Mossy, Department of Biostatistics, Science and Technology of New York, New York, USA, Email:
Department of Biostatistics, Science and Technology of New York, New York, USA

Received: 02-Jan-2023, Manuscript No. jbmbs-23-90611; Editor assigned: 03-Jan-2023, Pre QC No. P-90611; Reviewed: 16-Jan-2023, QC No. Q-90611; Revised: 20-Jan-2023, Manuscript No. R-90611; Published: 27-Jan-2023 , DOI: 10.37421/-2155-6180.2023.14.148
Citation: Mossy, Mindi. “A Transformational Viewpoint on Conditional and Marginal Models.” J Biom Biosta 14 (2023): 148.
Copyright: © 2023 Mossy M. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.

Introduction

Normalization is the process of creating shifted and scaled versions of statistics with the goal of eliminating the effects of certain gross influences, like in an anomaly time series, by comparing the corresponding normalized values of different datasets (heterogenic data). As a result, the procedure for heterogeneous data transformation brings all attributes to the same scale. Indeed, the decimal scaling method is one of the quantitative data normalization techniques that moves the decimal point of the data's values. We divide each data value by the maximum absolute to normalize the data using this method. The original data are subjected to linear transformation in the minimum-maximum (Min-Max) data normalization method, whereas in the z-score data normalization procedure, values are normalized using the mean and standard deviation parameters. On the basis of these evidences, quantitative data standardization and normalization procedures may have distinct parametric distribution, such as the normal distribution, and data variability reduction capabilities [1].

Description

Quantitative data used for the present study were drawn from previous experiments as described. Briefly, collected data included four growth parameters (diameter, plant height, leaf length and leaf number) of two maize varieties, treated by both rhizobacteria and foliar bio-fertilizing. Further, collected data for each treatment were summarized in a matrix including four columns describing variables parameters (two maize varieties growth parameters) and ninety-six rows corresponding to the observation number. Next, we submitted the abovementioned data matrix to Box-Cox, Logarithm, Square Root, Inverse and Z-score, Minimum, Exponential and Minimum-Maximum quantitative data standardization as well as normalization (data transformation) procedures. Biometric verification is a method for checking a person's personality by using a piece of their identity, like their finger impression, facial features, or iris design. These features contain unique information that can't be duplicated. Despite their numerous benefits, certain biometrics, particularly facial recognition, have recently come under fire for being an infringement on privacy. Considering everything, your "face print" is your information, and many people don't like the idea that their face prints could be used or shared without their consent. This may eliminate the obscurity that many people anticipate in open areas, such as online. Even the idea of "connecting" a person's face to yet another source of personal data has been floated [2,3].

The same survey displayed smaller bias transformation by using the Box- Cox transformation as opposite to logarithm transformation. The same study revealed that the mean squared error of estimation is smaller with the Box-Cox transformation; and as well, the Box-Cox transformation leads to systematically higher estimated values than Logarithmic transformation. Hence, the Box- Cox transformation should be considered as a viable alternative in statistical modelling if the transformation of variables is required. Low aptitude with regard Exponential and Inverse data transformation in reducing data variability as well as in adjusting data normality could be due to processed positive value of analysed data. Indeed, our analysis suspected Exponential data transformation as a potential source of transformed data variability [4,5].

Conclusion

We focused on eight quantitative data transformation systems in the present comparative study. Processed quantitative data standardization and/or normalization procedures are as following Box-Cox (Box), Exponential (Expo), Inverse, Logarithmic normalization, Maximum, Minimum-Maximum, Square Root and Z-score. Above-mentioned data transformation systems was applied to the same data matrix (collected data) generating a new data set for each standardization and/or normalization methods. The present study provided a systematic comparative study that highlighted difference as well as similitude between eight quantitative data standardization methodologies providing useful tool to researchers, in choosing adequately data transformation methodologies that well fitting for their investigations.

Acknowledgement

We thank the anonymous reviewers for their constructive criticisms of the manuscript. The support from ROMA (Research Optimization and recovery in the Manufacturing industry), of the Research Council of Norway is highly appreciated by the authors.

Conflict of Interest

The authors declare that there was no conflict of interest in the present study.

References

  1. Miguel-Alvarez, Marina, Alejandro Santos-Lozano, Fabian Sanchis-Gomar and Carmen Fiuza-Luces, et al. "Non-steroidal anti-inflammatory drugs as a treatment for Alzheimer’s disease: A systematic review and meta-analysis of treatment effect."Drugs Aging32 (2015): 139-147.
  2. Google Scholar, Crossref, Indexed at

  3. Chow, Shein-Chung.Innovative methods for rare disease drug development. Chapman and Hall/CRC (2020).
  4. Google Scholar, Crossref, Indexed at

  5. Goodman, Steven N. "A comment on replication, p‐values and evidence."Stat Med11 (1992): 875-879.
  6. Google Scholar, Crossref, Indexed at

  7. Shao, Jun and Shein‐Chung Chow. "Reproducibility probability in clinical trials."Stat Med 21 (2002): 1727-1742.
  8. Google Scholar, Crossref, Indexed at

  9. He, Jiang and Paul K. Whelton. "Elevated systolic blood pressure and risk of cardiovascular and renal disease: overview of evidence from observational epidemiologic studies and randomized controlled trial.’’ Am Heart J. 138 (1999): 211–219.
  10. Google Scholar, Crossref, Indexed at

Google Scholar citation report
Citations: 3496

Journal of Biometrics & Biostatistics received 3496 citations as per Google Scholar report

Journal of Biometrics & Biostatistics peer review process verified at publons

Indexed In

 
arrow_upward arrow_upward