MLT Unit 3
MLT Unit 3
MLT Unit 3
Decision Learning
& INSTANCE-BASED LEARNING
Decision Tree
• Decision Tree is a Supervised learning technique that can be used for both classification and
Regression problems, but mostly it is preferred for solving Classification problems.
• It is a tree-structured classifier, where internal nodes represent the features of a dataset, branches
represent the decision rules and each leaf node represents the outcome.
• In a Decision tree, there are two nodes, which are the Decision Node and Leaf Node. Decision nodes
are used to make any decision and have multiple branches, whereas Leaf nodes are the output of those
decisions and do not contain any further branches.
• The decisions or the test are performed on the basis of features of the given dataset.
Decision Tree (Cont.)
It is a graphical representation for getting all the possible solutions to a problem/decision based on
given conditions.
• It is called a decision tree because, similar to a tree, it starts with the root node, which expands on
further branches and constructs a tree-like structure.
• In order to build a tree, we use the CART algorithm, which stands for Classification and
Regression Tree algorithm.
• A decision tree simply asks a question, and based on the answer (Yes/No), it further split the tree
into subtrees.
• NOTE: A decision tree can contain categorical data (YES/NO) as well as numeric data.
INDUCTIVE BIAS IN DECISION TREE
• The inductive bias refers to the assumption or constraints that shape how the decision
tree algorithm builds to the tree and makes predictions.
• Decision Tree algorithm has an inductive bias towards using hierarchical if-else rules
to represent the relationships between features and the target variable.
• It uses binary splitting to divide the data.
Inductive Inference
• Inductive inference refers to the process of drwaing
general conclusions of making predictions based on
limited observation .
• We start with a set of specific observations or example
and aim to derive general principles or rules that can be
applies to new, unseen situation.
• The goal is to make reliable predictions or generalizations
beyond the observed data.
ID3 ALGORITHM
• Iterative Dichotomiser 3 or commonly known as ID3. ID3 was invented by Ross Quinlan.
• It is a classification algorithm that follows a greedy approach of building a decision tree by
selecting a best attribute that yields maximum Information Gain (IG) or minimum Entropy (H).
• Decision Tree is most effective if the problem characteristics look like the following points:
• Here, the attribute with maximum information gain is Outlook. So, the decision tree built so far -
• ID3
– Searches a complete hypothesis space incompletely
– Inductive bias is solely a consequence of the ordering of
hypotheses by its search strategy
• Candidate-Elimination
– Searches an incomplete hypothesis space completely
– Inductive bias is solely a consequence of the expressive
power of its hypothesis representation
ISSUES IN DECISION TREE
7 4 False
1 4 True
Euclidean Distance =
• D(x,i)= + = 4
• D(x,ii)= + = 5
• D(x,iii)= + = 3
• D(x,iv)= + = 3.6
We, need to find out the three nearest neighbors that means, the
distance having the lowest value: 3, 3.6 and 4
TRUE FALSE
TRUE
In locally weighted regression, points are weighted by proximity to the current x in question using
a kernel. A regression is then computed using the weighted points.
CASE BASED LEARNING