Sent Successfully.
Home / Blog / Interview Questions / Decision Tree Interview Q&A : Key Concepts in 2024
Decision Tree Interview Q&A : Key Concepts in 2024
Table of Content
- Decision tree algorithm belongs to the__________family.
- True/False. Can we use a decision tree for both Regressions as well as Classification?
- ______ is to create a training model that can be used to predict the class or value of the target variable by learning simple decision rules.
- Identify the type of a decision tree________.
- How many nodes are there in a decision tree?
- True/False. Are parent node and root node the both same in the decision tree?
- True/False. Are child node and branch nodes the same in the Decision tree?
- ______is used for cutting or trimming the tree in Decision trees.
- Does the decision of making strategic splits heavily affects a tree’s accuracy?
- ______ measure of the randomness in the information being processed in the Decision Tree.
- ______is a statistical property that measures how well a given attribute separates the training examples according to their target classification.
- _________computes the difference between entropy before the split and average entropy after the split of the dataset based on given attribute values.
- Information gain is biased towards choosing attributes with a large number of values as ______.
- Are leaf nodes terminal nodes?
- ________is an algorithm used for continuous target variables that are used for regression problems
- Decision tree is a ______.
- Choose the correct statement from below –
- Decision tree is also referred to as________ algorithm.
- Decision tree is used for _______.
- In the Decision tree, one rule is applied after another, resulting in a hierarchy of segments within segments. The hierarchy is called a ______, and each segment is called a ______.
- Choose the correct sequence of typical decision tree structure –
- _______ denotes the entire population or sample and it further divides into two or more homogeneous sets.
- ______ Node are those that do not split into parts.
- The Process of removing sub-nodes from a decision node is called ______.
- Decision tree classifier is achieved by _______ splitting criteria.
- Decision tree regressor is achieved by ______ splitting criteria
- ______ is a measure of uncertainty of a random variable.
- Information gain is required to decide _______.
- _______ is a measurement of likelihood of an incorrect classification of a new instance for a random variable, if the new instance is randomly classified as per the distribution of class labels from the data set.
- In Decision tree pruning methods include
- What is the regularization parameter in Decision tree?
- Which argument we need to pass in decision tree to make the algorithm boosting algorithm?
- Which nodes have the maximum Gini impurity in a decision tree?
- In decision tree we only use discrete data ?
- Which algorithm is most prune to overfitting?
- In which pruning method we allow our tree to grow till last including all the observations and variables?
- According the value of ------ we split the node and split the decision tree?
- Entropy is similar to?
- Gene index is almost similar to
- Pre-pruning the decision tree may results in
- Statement : Missing data can be handled by the DT.
reason : classification is done by the yes or no condition. - Leaf node in a decision tree will have entropy value
- Entropy value for the data sample that has 50-50 split belonging to two categories is
- Which measure select the best attribute to split the records?
- To plot the tree we use?
-
Decision tree algorithm belongs to the__________family.
- a) Supervised
- b) Unsupervised
Answer - a) Supervised
-
True/False. Can we use a decision tree for both Regressions as well as Classification?
- a) True
- b) False
Answer - a) True
-
______ is to create a training model that can be used to predict the class or value of the target variable by learning simple decision rules.
- a) Random Forest
- b) Decision Tree
- c) Both a and b
- d) None
Answer - b) Decision Tree
-
Identify the type of a decision tree________.
- a) Categorical
- b) Continuous
- c) Both a and b
- d) None
Answer - c) Both a and b
-
How many nodes are there in a decision tree?
- a) 2
- b) 3
- c) 4
- d) Only 1
Answer - c) 4
-
True/False. Are parent node and root node the both same in the decision tree?
- a) Yes
- b) No
Answer - a) Yes
-
True/False. Are child node and branch nodes the same in the Decision tree?
- a) Yes
- b) No
Answer - a) Yes
-
______is used for cutting or trimming the tree in Decision trees.
- a) Pruning
- b) Stemming
Answer - a) Pruning
-
Does the decision of making strategic splits heavily affects a tree’s accuracy?
- a) No
- b) Yes
- c) Maybe
- d) Don’t know
Answer - d) Don’t know
-
______ measure of the randomness in the information being processed in the Decision Tree.
- a) Entropy
- b) Information Gain
Answer - a) Entropy
-
______is a statistical property that measures how well a given attribute separates the training examples according to their target classification.
- a) Entropy
- b) Information Gain
Answer - b) Information Gain
-
_________computes the difference between entropy before the split and average entropy after the split of the dataset based on given attribute values.
- a) Information gain
- b) Gini ratio
- c) Pruning
Answer - a) Information gain
-
Information gain is biased towards choosing attributes with a large number of values as ______.
- a) Branch nodes
- b) Root nodes
- c) Leaf nodes
Answer - b) Root nodes
-
Are leaf nodes terminal nodes?
- a) Yes
- b) No
Answer - a) Yes
-
________is an algorithm used for continuous target variables that are used for regression problems in Decision Tree.
- a) Reduction in Variance
- b) Collinearity
- c) Correlation
Answer - a) Reduction in Variance
-
Decision tree is a ___________.
- a) Non-linear ML technique.
- b) Non-Parametric technique.
- c) Supervised Learning technique.
- d) All of the above.
Answer - d) All of the above
-
Choose the correct statement from below –
- a) A decision tree is a graphical representation of all the possible solutions to a decision based on certain conditions.
- b) Decision Trees usually mimic human thinking ability while making a decision, so it is easy to understand.
- c) A decision tree model consists of a set of rules for dividing a large heterogeneous population into smaller, more homogenous (mutually exclusive) classes.
- d) All of the above.
Answer - d) All of the above
-
Decision tree is also referred to as__________________ algorithm.
- a) Recursive partitioning.
- b) Non-Recursive partitioning.
- c) Variable partitioning.
- d) None of the above.
Answer - a) Recursive partitioning
-
Decision tree is used for __________________.
- a) Only Regression.
- b) Only Classification.
- c) Both (a) and (b).
- d) None of the above.
Answer - c) Both (a) and (b)
-
In the Decision tree, one rule is applied after another, resulting in a hierarchy of segments within segments. The hierarchy is called a __________, and each segment is called a ____________.
- a) Node, Tree.
- b) Tree, Node.
- c) Branch, Node.
- d) None of the above.
Answer - b) Tree, Node
-
Choose the correct sequence of typical decision tree structure –
(I) Take the entire data set as input
(II) Divide the input data into two part
(III) Reapply the split to every part recursively
(IV) Stop when meeting desired criteria
(V) Cut the tree when we went too far while doing splits(pruning)- a) (I), (II),(V),(IV),(III).
- b) (V),(I),(III),(II),(IV).
- c) (I),(III),(II),(V),(IV).
- d) (I),(II),(III),(IV),(V).
Answer - d) (I),(II),(III),(IV),(V)
-
___________ denotes the entire population or sample and it further divides into two or more homogeneous sets.
- a) Leaf node.
- b) Terminal node.
- c) Root node.
- d) None of the above.
Answer - c) Root node
-
___________ Node are those that do not split into parts.
- a) Root.
- b) Leaf / terminal.
- c) Branch.
- d) None of the above.
Answer - b) Leaf / terminal
-
The Process of removing sub-nodes from a decision node is called ___________.
- a) Splitting.
- b) Breaking.
- c) Pruning.
- d) None of the above.
Answer - c) Pruning
-
Decision tree classifier is achieved by __________ splitting criteria.
- a) Entropy.
- b) Information Gain.
- c) Gini Index.
- d) All of the above.
Answer - d) All of the above
-
Decision tree regressor is achieved by __________ splitting criteria
- a) Loss in the mean squared error.
- b) Entropy.
- c) Information Gain.
- d) None of the above.
Answer - a) Loss in the mean squared error
-
___________ is a measure of uncertainty of a random variable.
- a) Information gain.
- b) Entropy.
- c) Gini Index.
- d) None of the above.
Answer - b) Entropy
-
Information gain is required to decide ___________.
- a) Which feature to split on at every step in building the tree.
- b) Possible combinations of variables.
- c) No. of possible branches.
- d) All of the above.
Answer - a) Which feature to split on at every step in building the tree
-
____________ is a measurement of likelihood of an incorrect classification of a new instance for a random variable, if the new instance is randomly classified as per the distribution of class labels from the data set.
- a) Gini impurity.
- b) Entropy.
- c) Information gain.
- d) None of the above.
Answer - a) Gini impurity
-
In Decision tree pruning methods include
- a) No pruning.
- b) Reduced error pruning.
- c) Bagging.
- d) All of the above.
Answer - d) All of the above
-
What is the regularization parameter in Decision tree?
- a. Regularization
- b. max_depth
- c. min_length
- d. None of the above
Answer - a. Regularization
-
Which argument we need to pass in decision tree to make the algorithm boosting algorithm?
- a.Entropy
- b. Trial
- c. Gini
- d. None of the above
Answer - b. Trial
-
Which nodes have the maximum Gini impurity in a decision tree?
- a. 0.6
- b. 0.1
- c. 0.5
- d. none of the above
Answer - c. 0.5
-
In decision tree we only use discrete data ?
- a. True
- b.False
Answer - a. True
-
Which algorithm is most prune to overfitting?
- a. Link analysis
- b. Neural Analysis
- c. Decision Tree
- d. None of the above
Answer - c. Decision Tree
-
In which pruning method we allow our tree to grow till last including all the observations and variables?
- a. post prunning
- b. pre prunning
- c. both a & b
- d. None of the above
Answer - a. post prunning
-
According the value of ------ we split the node and split the decision tree?
- a. Gini Index
- b. Entropy
- c. Information Gain
- d. None of the above
Answer - c. Information Gain
-
Entropy is similar to
a.estimated mean
b.estimated information event- a. both a and b
- b. only b
- c. only a
- d. neither a and b
Answer - b. only b
-
Gene index is almost similar to
a. estimated mean
b. estimated information event- a. both a and b
- b. only b
- c. only a
- d. neither a and b
Answer - c. only a
-
Pre-pruning the decision tree may results in
- a. Overfitting
- b. Underfitting
Answer - b. Underfitting
-
Statement : Missing data can be handled by the DT.
reason : classification is done by the yes or no condition.- a. statment is True, Reason gives the correct explanation of statement
- b. statement is false, Reason is correct
- c. statement is false, reason is incorrect
- d. statement is true, reason is false explaination of statement
Answer - a. statment is True, Reason gives the correct explanation of statement
-
Leaf node in a decision tree will have entropy value
- a. equal to zero
- b. equal to 1
- c. none
Answer - a. equal to zero
-
Entropy value for the data sample that has 50-50 split belonging to two categories is
- a. 1
- b. 0
- c. None
Answer - a.1
-
Which measure select the best attribute to split the records?
- a.export_text
- b. Attribute Selection Measures(ASM)
- c. tree.plot_tree(clf)
- d. None of these
Answer - b. Attribute Selection Measures(ASM)
-
To plot the tree we use
- a.export_text
- b. Attribute Selection Measures(ASM)
- c. tree.plot_tree(clf)
- d. None of these
Answer - c. tree.plot_tree(clf)
Navigate to Address
360DigiTMG - Data Analytics, Data Science Course Training Hyderabad
2-56/2/19, 3rd floor, Vijaya Towers, near Meridian School, Ayyappa Society Rd, Madhapur, Hyderabad, Telangana 500081
099899 94319