Introduction to Create a Decision Tree
With the recent rapid growth of the amount of data generated by information systems, to handle large data sets, there is a dominant need for the decision tree to reduce computation complexity. A decision tree can be considered the most important approach for representing classifiers. In other words, we can say the data are structured using a divide and conquer strategy. up to know we have only explored. A decision tree is structured as a framework to accurate the values and probability of outcomes decisions fro
m each level of the node, helping decision-makers to choose correct predictions among the various inappropriate data. In this article, you will walk through how to create a decision tree based on sample data in a simple way.
What is Decision Tree?
A decision tree is a binary hierarchical structure that identifies the way where each node split a data set based on different conditions. To construct an optimal tree with a model approach to classify a response variable that predicts the value of a target variable with simple decision rules (if-then-else statements). The approach is supervised learning mostly used in classification problems and considered to be a very effective predictive model. They are used in different application domains like Game theory, Artificial Intelligence, Machine Learning, Data Mining and areas like security and medicine.
How to Create a Decision Tree?
A decision tree is created in simple ways with the top-down manner; they consist of nodes that form a directed node which has root nodes with no incoming edges all other nodes are called decision -nodes (internal node and leaf nodes which corresponds to attribute and class labels) with at least one incoming edges. The main goal from the data sets is to minimize the generalization errors by finding the optimal solution in the decision tree.
An example of a decision tree is explained below with a sample data set. The goal is to predict whether a profit is down or up using the attributes of life and competition. Here the decision tree variables are categorical (Yes, No).
The Data Set
From the above data set: life, competition, Type are the predictors and the attribute profit is the target. There are various algorithms to implement a decision tree, but the best algorithm used to build a decision tree is ID3 which emphasis on greedy search approach. The decision tree follows decision inference rule or disjunctive normal form(^).
4.7 (4,439 ratings)
Initially, all the training attribute is considered to be the root. The order priority for placing the attributes as root is done by the following approach. This process is known to attribute selection to identify which attribute is made to be a root node at each level. The tree follows two steps: construction of a tree, tree pruning. And the data’s are been split in all the decision nodes.
It is the measure of the change in entropy based on the independent variable. The decision tree must find the highest information gain.
Entropy is defined as for the finite set, the measure of randomness in data or event predictability, if the sample is similar values then entropy is zero and if it is equally divided with the sample then it is one.
Entropy for the Class
Where p is the probability of getting profit to say ‘yes’ and N is loss say ‘No’.
therefore, entropy =1
Once entropy value is calculated it is necessary to decide a root node from the attribute.
Entropy of Age
According to the data set for Life attribute we have old =3 down, mid = 2 down and one up concerning profit label.
Gain = Class Entropy – Entropy of Life = 1 – 0.4 = 0.6
Entropy (competition) = 0.87
Gain = Class Entropy – Entropy of Life = 1 – 0.87 = 0.12
Now the problem arises in the attribute Life where the mid has an equal probability on both up and down. therefore, entropy is 1. similarly, it is calculated for type attribute again the entropy is 1 and gain is 0. Now a complete decision has been created to get an accurate result for mid-value.
Advantages of Decision Tree
- They are easy to understand and the rules generated are flexible. Has little effort for data preparation.
- A visual approach to represent decisions and outcomes is very helpful.
- The decision tree handles the training data set with errors and missing values.
- They can handle discrete value and a numerical attribute. It works categorical and continuous variables for input and output.
- They are a useful tool for the business domain who has to take decisions after analyzing under certain conditions.
Disadvantages of Decision Tree
- Learners can create a complex decision tree depending on trained data. this process is termed as overfitting, a difficult process in decision tree models.
- The values preferred to be is categorical, if it is continuous, the decision tree loses information which leads to error-prone. Exponential calculation growth is higher while analyzing.
- Many class labels lead to incorrect complex calculations and give low prediction accuracy of the dataset.
- Information gained in the DT algorithm gives a biased response to categorical higher values.
Therefore, to conclude, decision trees provide a practical and easy method for learning and strongly known as efficient tools for machine learning as in a short time they perform well with large datasets. It’s a learning task that uses a statistical approach to make a generalized conclusion. Now its better understood why the decision tree is used in predictive modeling and for the data scientists they are the powerful tool.
This is a guide to Create a Decision Tree. Here we discuss how to Create a Decision Tree along with various advantages and disadvantages. You can also go through our other suggested articles to learn more –