<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://en.formulasearchengine.com/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=106.219.46.22</id>
	<title>formulasearchengine - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://en.formulasearchengine.com/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=106.219.46.22"/>
	<link rel="alternate" type="text/html" href="https://en.formulasearchengine.com/wiki/Special:Contributions/106.219.46.22"/>
	<updated>2026-05-04T13:23:24Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.43.0-wmf.28</generator>
	<entry>
		<id>https://en.formulasearchengine.com/index.php?title=Oil_in_place&amp;diff=10636</id>
		<title>Oil in place</title>
		<link rel="alternate" type="text/html" href="https://en.formulasearchengine.com/index.php?title=Oil_in_place&amp;diff=10636"/>
		<updated>2013-12-24T05:13:43Z</updated>

		<summary type="html">&lt;p&gt;106.219.46.22: /* Calculation of STOOIP */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{other uses2|Gain}}&lt;br /&gt;
{{Refimprove|date=December 2009}}&lt;br /&gt;
In [[information theory]] and [[machine learning]], &#039;&#039;&#039;information gain&#039;&#039;&#039; is a synonym for &#039;&#039;[[Kullback–Leibler divergence]]&#039;&#039;. However, in the context of decision trees, the term is sometimes used synonymously with [[mutual information]], which is the expectation value of the Kullback–Leibler divergence.&lt;br /&gt;
&lt;br /&gt;
In particular, the information gain about a random variable &#039;&#039;X&#039;&#039; obtained from an observation that a random variable &#039;&#039;A&#039;&#039; takes the value &#039;&#039;A=a&#039;&#039; is the Kullback-Leibler divergence &#039;&#039;D&#039;&#039;&amp;lt;sub&amp;gt;KL&amp;lt;/sub&amp;gt;(&#039;&#039;p&#039;&#039;(&#039;&#039;x&#039;&#039; | &#039;&#039;a&#039;&#039;) || &#039;&#039;p&#039;&#039;(&#039;&#039;x&#039;&#039; | I)) of the [[prior distribution]] &#039;&#039;p&#039;&#039;(&#039;&#039;x&#039;&#039; | I) for x from the [[posterior distribution]] &#039;&#039;p&#039;&#039;(&#039;&#039;x&#039;&#039; | &#039;&#039;a&#039;&#039;) for &#039;&#039;x&#039;&#039; given &#039;&#039;a&#039;&#039;.  &lt;br /&gt;
&lt;br /&gt;
The [[expected value]] of the information gain is the mutual information &#039;&#039;I(X; A)&#039;&#039; of &#039;&#039;X&#039;&#039; and &#039;&#039;A&#039;&#039; – i.e. the reduction in the [[information entropy|entropy]] of &#039;&#039;X&#039;&#039; achieved by learning the state of the random variable &#039;&#039;A&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
In machine learning, this concept can be used to define a preferred sequence of attributes to investigate to most rapidly narrow down the state of &#039;&#039;X&#039;&#039;.  Such a sequence (which depends on the outcome of the investigation of previous attributes at each stage) is called a [[Decision tree learning|decision tree]]. Usually an attribute with high mutual information should be preferred to other attributes.&lt;br /&gt;
&lt;br /&gt;
==General definition==&lt;br /&gt;
In general terms, the [[expectation value|expected]] information gain is the change in [[information entropy]]  &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; from a prior state to a state that takes some information as given: &lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt; IG(T,a) = H(T) - H(T|a) &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Formal definition==&lt;br /&gt;
Let &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt; denote a [[training set|set of training examples]], each of the form &amp;lt;math&amp;gt;(\textbf{x},y) = (x_1, x_2, x_3, ..., x_k, y)&amp;lt;/math&amp;gt;&lt;br /&gt;
where &amp;lt;math&amp;gt;x_a\in vals(a)&amp;lt;/math&amp;gt; is the value of the &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt;th attribute of example &amp;lt;math&amp;gt;\textbf{x}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; is the corresponding class label. The information gain for an attribute &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; is defined in terms of entropy &amp;lt;math&amp;gt;H()&amp;lt;/math&amp;gt; as follows:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;IG(T,a) = H(T)-\sum_{v\in vals(a)}\frac{|\{\textbf{x}\in T|x_a=v\}|}{|T|} \cdot H(\{\textbf{x}\in T|x_a=v\}) &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The mutual information is equal to the total entropy for an attribute if for each of the attribute values a unique [[statistical classification|classification]] can be made for the result attribute. In this case, the relative entropies subtracted from the total entropy are 0.&lt;br /&gt;
&lt;br /&gt;
==Drawbacks==&lt;br /&gt;
Although information gain is usually a good measure for deciding the [[relevance]] of an attribute, it is not perfect. A notable problem occurs when information gain is applied to attributes that can take on a large number of distinct values. For example, suppose that one is building a decision tree for some data describing the customers of a business. Information gain is often used to decide which of the attributes are the most relevant, so they can be tested near the root of the tree. One of the input attributes might be the customer&#039;s credit card number. This attribute has a high mutual information, because it uniquely identifies each customer, but we do &#039;&#039;not&#039;&#039; want to include it in the decision tree: deciding how to treat a customer based on their credit card number is unlikely to generalize to customers we haven&#039;t seen before ([[overfitting]]).&lt;br /&gt;
&lt;br /&gt;
[[Information gain ratio]] is sometimes used instead. This biases the decision tree against considering attributes with a large number of distinct values.  However, attributes with very low information values then appeared to receive an unfair advantage. In addition, methods such as [[permutation tests]] have been proposed to correct the bias.&amp;lt;ref&amp;gt;{{cite conference&lt;br /&gt;
|author=Deng,H.|coauthors=Runger, G.; Tuv, E.&lt;br /&gt;
|title=Bias of importance measures for multi-valued attributes and solutions&lt;br /&gt;
|conference=Proceedings of the 21st International Conference on Artificial Neural Networks (ICANN2011)&lt;br /&gt;
|year=2011|pages= 293-300|doi=10.1007/978-3-642-21738-8_38}}&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Notes ==&lt;br /&gt;
{{Reflist}}&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
* {{cite book|authorlink=Tom Mitchell|author=Mitchell, Tom M.|title=Machine Learning|publisher=The Mc-Graw-Hill Companies, Inc.|year=1997|url=https://www.cs.cmu.edu/~tom/mlbook.html|isbn=0070428077}}&lt;br /&gt;
&lt;br /&gt;
{{DEFAULTSORT:Information Gain In Decision Trees}}&lt;br /&gt;
[[Category:Decision trees]]&lt;br /&gt;
[[Category:Classification algorithms]]&lt;/div&gt;</summary>
		<author><name>106.219.46.22</name></author>
	</entry>
</feed>