Open main menu

Information gain in decision trees

In information theory and machine learning, information gain is a synonym for Kullback–Leibler divergence; the amount of information gained about a random variable or signal from observing another random variable. However, in the context of decision trees, the term is sometimes used synonymously with mutual information, which is the conditional expected value of the Kullback–Leibler divergence of the univariate probability distribution of one variable from the conditional distribution of this variable given the other one.

The information gain of a random variable X obtained from an observation of a random variable A taking value is defined

the Kullback-Leibler divergence of the prior distribution for x from the posterior distribution for x given a.

The expected value of the information gain is the mutual information of X and A – i.e. the reduction in the entropy of X achieved by learning the state of the random variable A.

In machine learning, this concept can be used to define a preferred sequence of attributes to investigate to most rapidly narrow down the state of X. Such a sequence (which depends on the outcome of the investigation of previous attributes at each stage) is called a decision tree and applied in the area of machine learning known as decision tree learning. Usually an attribute with high mutual information should be preferred to other attributes.[why?]

Contents

General definitionEdit

In general terms, the expected information gain is the change in information entropy Η from a prior state to a state that takes some information as given:

 

where   is the conditional entropy of   given the value of attribute  .

Formal definitionEdit

Let   denote a set of training examples, each of the form   where   is the value of the   attribute or feature of example   and y is the corresponding class label. The information gain for an attribute   is defined in terms of Shannon entropy   as follows. For a value   taken by attribute  , let

 
be defined as the set of training inputs of   such for which attribute   is equal to  . Then the information gain of   for attribute   is the difference between the a priori Shannon entropy   of the training set and the conditional entropy  .

 
 

The mutual information is equal to the total entropy for an attribute if for each of the attribute values a unique classification can be made for the result attribute. In this case, the relative entropies subtracted from the total entropy are 0. In particular, the values   defines a partition of the training set data   into mutually exclusive and all-inclusive subsets, inducing a categorical probability distribution   on the values   of attribute  . The distribution is given  . In this representation, the information gain of   given   can be defined as the difference between the unconditional Shannon entropy of   and the expected entropy of   conditioned on  , where the expectation value is taken with respect to the induced distribution on the values of  .

 

DrawbacksEdit

Although information gain is usually a good measure for deciding the relevance of an attribute, it is not perfect. A notable problem occurs when information gain is applied to attributes that can take on a large number of distinct values. For example, suppose that one is building a decision tree for some data describing the customers of a business. Information gain is often used to decide which of the attributes are the most relevant, so they can be tested near the root of the tree. One of the input attributes might be the customer's credit card number. This attribute has a high mutual information, because it uniquely identifies each customer, but we do not want to include it in the decision tree: deciding how to treat a customer based on their credit card number is unlikely to generalize to customers we haven't seen before (overfitting).

To counter this problem, Ross Quinlan proposed[1] to instead choose the attribute with highest information gain ratio from among the attributes whose information gain is average or higher. This biases the decision tree against considering attributes with a large number of distinct values, while not giving an unfair advantage to attributes with very low information value, as the information value is higher or equal to the information gain.[2]

See alsoEdit

ReferencesEdit

  1. ^ Quinlan, J. Ross. "Induction of decision trees." Machine learning 1.1 (1986): 81-106.
  2. ^ "What is the range of information gain ratio?". Cross Validated. Retrieved 2018-10-09.

Further readingEdit