Oops! It appears that you have disabled your Javascript. In order for you to see this page as it is meant to appear, we ask that you please re-enable your Javascript!

Decision Tree Induction and Entropy in data mining

Last modified on December 9th, 2018 at 9:17 pm

Decision Tree Induction

Decision tree is a tree-like structure and consists of following parts(discussed in Figure 1);

  1. Root node:
    • age is the root node
  2. Branches:
    • Following are the branches;
      • <20
      • 21…50
      • >50
      • USA
      • PK
      • High
      • Low
  3. Leaf node:
    • Following are the leaf nodes;
      • Yes
      • No
[quads id=1]

decision tree induction examplesdata mining



Entropy is a method to measure the uncertainty.

  • Entropy can be measured in between 0 and 1.
  • High entropy represents that data have more variance with each other.
  • Low entropy represents that data have less variance with each other.

P = Total yes = 9

N = Total no = 5

Note that to calculate the  logof a number, we can do the following procedure.

For example;

what is  log of 0.642?

Ans: log (0.642) / log (2)

=9/14 * log2(9/14)  –  5/14 * log2 (5/14)

=-9/14 * log2(0.642)  –  5/14 * log2 (0.357)

=-9/14 * (0.639)  –  5/14 * (-1.485)


[quads id=2]

For Age:

agePiNiInfo(Pi, Ni)
<20 2 YES 3 NO 0.970
21…50 4 YES0 NO 0
>50 3 YES 2 NO  0.970


Note: if yes =2 and No=3 then entropy is 0.970 and it is same  0.970 if yes=3 and No=2

So here when we calculate the entropy for age<20, then there is no need to calculate the entropy for age >50 because the total number of Yes and No is same.


The gain of Age0.2480.248 is a greater value than income, Credit Rating, and Region. So Age will be considered as the root node.
Gain of Income0.029 
Gain of Credit Rating0.048 
Gain of  Region0.151 
[quads id=3]

decision tree .pdf

Note that

  • if yes and no are in the following sequence like (0, any number) or (any number, 0) then entropy is always 0.
  • If yes and no are occurring in such a sequence (3,5) and (5, 3) then both have same entropy.
  • Entropy calculates impurity or uncertainty of data.
  • If the coin is fair (1/2, head and tail have equal probability, represent maximum uncertainty because it is difficult to guess that head occurs or tails occur) and suppose coin has the head on both sides then the probability is 1/1, and uncertainty or entropy is less.
  • if p is equal to q then more uncertainty
  • if p is not equal to q then less uncertainty

Now again calculate entropy for;

  1. Income
  2. Region
  3. Credit

For Income:

IncomePiNiInfo(Pi, Ni)
High0 YES2 NO0
Medium 1 YES1 NO1
Low 1 YES0 NO0

For Region:

RegionPiNiInfo(Pi, Ni)

For Credit Rating:

Credit RatingPiNiInfo(Pi, Ni)
Low1 YES2 NO0
High1 YES1 NO0


[quads id=4]
The gain of Region0.9700.970 is a greater value than income, Credit Rating, and Region. So Age will be considered as the root node.
Gain of Credit Rating0.02 
Gain of Income0.57 

Similarly, you can calculate for all.

Prof. Fazal Rehman Shamil
Researcher, Publisher of International Journal Of Software Technology & Science ISSN: 2616-5325
Instructor, SEO Expert, Web Programmer and poet.
Feel free to contact.