Course Hero Logo

Below table 1 description of the attributes table 1

This preview shows page 15 - 18 out of 36 pages.

Below Table 1 description of the attributes.Table 1: Attribute descriptionAttributesDescriptionfixed acidityFixed acids, numeric from 3.8 to 15.9volatile acidityVolatile acids, numeric from 0.1 to 1.6citric acidCitric acids, numeric from 0.0 to 1.7residual sugarresidual sugar, numeric from 0.6 to 65.8chloridesChloride, numeric from 0.01 to 0.61free sulfur dioxideFree sulfur dioxide, numeric: from 1 to 289total sulfur dioxideTotal sulfur dioxide, numeric: from 6 to 440densityDensity, numeric: from 0.987 to 1.039pHpH, numeric: from 2.7 to 4.0sulfatesSulfates, numeric: from 0.2 to 2.0alcoholAlcohol, numeric: from 8.0 to 14.9qualityQuality, numeric: from 0 to 10, the output target
104.2.Feature selectionFeature selection is the method of selection of the best subset offeatures that will be used for classification (Fauzi et al., 2017). Mostof the feature selection method is divided into a filter and wrapper, thefilter uses the public features work individually from the learningalgorithm and the wrapper evaluates the features and choosesattributes based on the estimation of the accuracy by using a searchalgorithm and specific learning model (Onan andKorukoğlu,2017).In this study, for a better understanding of the features and to examinesthe correlation between the features. The Pearson correlationcoefficient is calculated for each feature in Table 1, this shows thepairwise person correlation coefficient P, which is calculated by usingthe below formula (Dastmard, 2013).𝑃?,?=cov (X, Y)𝜎𝑋, σYWhere the𝜎is the standard deviation of the features X and Y and covis the covariance. The range of the correlation coefficient from -1 to1. Point 1 value implies linear equation is describes the correlationbetween X and Y strong positive, which is all data points are lying ona line for Y increases as X increases. Where point -1 value indicatesthat strong negative correlations between data points. All data pointslie on a line in which Y decreases as X increases. And point 0 indicatesthat there is an absence of correlation between the points (Dastmard,2013).4.3.HyperParametr tuningThe grid search is a basic method for hyperparameter tuning. Performan inclusive search on the hyperparameter set specified by the user.Grid search is suitable for several hyperparameters with limited searchspace. The grid search algorithm is straightforward with enoughresources, the most accurate prediction can be drawn and users can
11always find the best combination (Joseph, 2018). Running grid searchin parallel is easy because each test is run separately without affectedby the time series. The results of one experiment are independent ofthe results of other experiments. Computing resources can be allottedin a very flexible way. In addition, grid search can accept a limitedsampling range, because too many settings are not suitable. Inpractice, grid search is almost preferable only when the user hasenough knowledge with these hyperparameters to allow the definitionof a narrow search space, and it is not necessary to adjust more thanthree hyperparameters simultaneously. Although other searchalgorithms may have more useful features, grid search is still the mostwidely used method due to its mathematical simplicity (Yu and Zhu,2020).

Upload your study docs or become a

Course Hero member to access this document

Upload your study docs or become a

Course Hero member to access this document

End of preview. Want to read all 36 pages?

Upload your study docs or become a

Course Hero member to access this document

Term
Fall
Professor
NoProfessor
Tags
Machine Learning, Artificial neural network, Statistical classification, Precision

Newly uploaded documents

Show More

  • Left Quote Icon

    Student Picture

  • Left Quote Icon

    Student Picture

  • Left Quote Icon

    Student Picture