Lecture 13: Case Study on Logistic Regression – October 0613-213.2.1Preliminaries on Classification ModelWe consider a binary classification problem: Given (x1, y1),(x2, y2), ...,(xn, yn),xi∈Rd,yi∈ {-1,1}. Notethat inputxiis called feature vector, outputyiis label. Suppose there is a family of functionf(x, w), binaryclassification aims to specify the bestwsuch that x and y are related.Based on the input feature vectors and certain type of function we adopt, the output can be predicted basedon certain prediction rules. For binary classification problem, the prediction rule can be shown as follows:y=1,f(x, w)≥0-1, f(x, w)<0An error term can be defined as follows to denote whether the output is successfully predicted or not:error =1,yf(x, w)<00,yf(x, w)≥0Obviously our goal is to find the best parameterwin order to minimize the total error in the process ofprediction. The optimization step in the binary classification model can be extracted below: