Artificial Intelligence and Quantum Computing for Advanced Wireless Networks. Savo G. Glisic. Читать онлайн. Newlib. NEWLIB.NET

Автор: Savo G. Glisic
Издательство: John Wiley & Sons Limited
Серия:
Жанр произведения: Программы
Год издания: 0
isbn: 9781119790310
Скачать книгу
to the odds ratio. The converse relationship is

      (2.4)upper O upper R Subscript 1 comma 2 Baseline equals e Superscript l 1 minus l 2

      In logistic regression, a categorical dependent variable Y having G (usually G = 2) unique values is regressed on a set of p independent variables X1, X2, …, Xp.

      Let X = (X1, X2, …, Xp) and Bg = (βg1, …, βgp)T; then the logistic regression model is given by the G equations ln(pg /p1) = ln(Pg /P1) + βg1 X1 + βg2 X2 + …. + βgp Xp = ln(Pg/P1) + XBg. Here, pg is the probability that an individual with values X1, X2, …, Xp is in outcome g. That is, pg = Pr(Y = g ∣ X). Usually, X1 ≡ 1 (that is, an intercept is included), but this is not necessary. The quantities P1, P2, …, PG represent the prior probabilities of outcome membership. If these prior probabilities are assumed equal, then the term ln (Pg/P1) becomes zero and drops out. If the priors are not assumed equal, they change the values of the intercepts in the logistic regression equation.

      The first outcome is called the reference value. The regression coefficients β1 , β2 , …, βp for the reference value are set to zero. The choice of the reference value is arbitrary. Usually, it is the most frequent value or a control outcome to which the other outcomes are to be compared. This leaves G − 1 logistic regression equations in the logistic model.

Schematic illustration of the regression line for predicting Y star from X star is not the 45 degrees line.

      since e Superscript normal upper X normal upper B 1 Baseline equals 1 because all of its regression coefficients are zero. Using the fact that ea + b = (ea)(eb), eXB may be reexpressed as follows: eXB = exp(β1 X1 + β2 X2 + ⋯ + βρ Xp) = eβ1 X1 e β2 X2 …e βp Xp . This shows that the final value is the product of its individual terms.

      2.1.3 Decision Tree: Regression Trees Versus Classification Trees

Schematic illustration of decision tree. Schematic illustration of tree terminology.

      Regression trees versus classification trees: From Figure 2.4 we can see that the terminal nodes (or leaves) lie at the bottom of the decision tree. This means that decision trees are typically drawn upside down such that leaves are the bottom and the roots are the top.

      Both the trees work almost similar to each other. Let us look at the primary differences and similarities between classification and regression trees:

       Regression trees are used when the dependent variable is continuous. Classification trees are used when dependent variable is categorical.

       In the case of regression trees, the value obtained by terminal nodes in the training data is the mean response of observations falling in that region. Thus, if an unseen data observation falls in that region, we will make its prediction with the mean value.

       In the case of classification trees, the value (class) obtained by terminal nodes in the training data is the mode of observations falling in that region. Thus, if an unseen data observation falls in that region, we will make its prediction with the mode value.

       Both the trees divide the predictor space (independent variables) into distinct and non‐overlapping regions. For the sake of simplicity, we can think of these regions as high‐dimensional boxes.

       Both the trees follow a top‐down greedy approach known as recursive binary splitting. We call it “top‐down” because it begins from the top of the tree when all the observations are available in a single region and successively splits the predictor space into two new branches down the tree. It is known as “greedy” because the algorithm cares about (looks for the best variable available) only the current split, and not about future splits which will lead to a better tree.

       This splitting process is continued until a user‐defined stopping criterion is reached. For example, we can tell the algorithm to stop once the number of observations per node becomes less than 50.

       In both the cases, the splitting process results in fully grown trees until the stopping criterion is reached. However, the fully grown tree is likely to overfit data, leading to poor accuracy on unseen data. This is handled by “pruning,” which is one of the techniques used to tackle overfitting.