Difference between revisions of "Python-for-Machine-Learning/C3/Decision-Tree/English"

From Script | Spoken-Tutorial
Jump to: navigation, search
Line 1: Line 1:
 
 
 
  
 
{| border="1"
 
{| border="1"
Line 300: Line 297:
 
|| The output displays the '''multi class ROC curve''' of our classifier.
 
|| The output displays the '''multi class ROC curve''' of our classifier.
  
'''DrugA, DrugB, DrugC''' have an '''AUC score''' of '''1''', indicating perfect classification.
+
Let us see about the AUC score that is area under the curve.
 +
 
 +
'''DrugA, DrugB, DrugC''' have an '''AUC score''' of 1, indicating perfect classification.
  
 
'''DrugX''' and''' DrugY '''have''' AUC scores '''of '''0.96 '''and''' 0.98''', which are very high.
 
'''DrugX''' and''' DrugY '''have''' AUC scores '''of '''0.96 '''and''' 0.98''', which are very high.
Line 315: Line 314:
  
 
'''plt.figure(figsize=(29, 10)) '''
 
'''plt.figure(figsize=(29, 10)) '''
|| Then we extract the column names excluding '''Drug '''for tree '''visualization.'''
+
|| Then we extract the column names excluding '''Drug ''' for tree '''visualization.'''
  
 
Next, we set the figure size and plot the '''decision tree'''.
 
Next, we set the figure size and plot the '''decision tree'''.
Line 338: Line 337:
  
 
'''Summary'''
 
'''Summary'''
|| This brings us to the end of the tutorial. Let us summarize.
+
 
  
 
In this tutorial, we have learnt about
 
In this tutorial, we have learnt about
 
* '''Decision Tree'''
 
* '''Decision Tree'''
 
* '''Decision Tree Structure and Nodes'''
 
* '''Decision Tree Structure and Nodes'''
 +
 +
|| This brings us to the end of the tutorial. Let us summarize.
 +
  
 
|-  
 
|-  

Revision as of 13:31, 28 June 2025

Visual Cue Narration
Show slide:

Welcome

Welcome to the Spoken Tutorial on Decision Tree.
Show Slide:

Learning Objectives

In this tutorial, we will learn about
  • Decision Tree
  • Decision Tree Structure and Nodes
Show Slide:

System Requirements

To Record this tutorial, I am using
  • Ubuntu Linux OS 24.04
  • Jupyter Notebook IDE
Show Slide:

Prerequisite

To follow this tutorial,

To follow this tutorial,
  • The learner must have basic knowledge of Python.
  • For prerequisite Python tutorials, please visit this website.
Show Slide:

Code files

  • The files used in this tutorial are provided in the Code files link.
  • Please download and extract the files.
  • Make a copy and then use them while practicing.
Show Slide:

Decision Tree

  • A decision tree is a tool used in machine learning that helps make decisions.
  • It uses a tree like structure to make predictions.
Show Slide:

Working of Decision Tree

Show dt.png img

  • The root node starts the decision tree with a question or condition.
  • Based on the answer, we follow a branch to another node.
  • A branch connects nodes, where each node represents a condition with outcomes.
Show dt.png img
  • This new node poses another question or condition.
  • We repeat this process of asking questions and following branches.
  • Finally, we reach a leaf node, which gives us the final decision or outcome.
Hover over the files I have created required files for the demonstration of Decision Tree.
Open the file drug200.csv and point to the fields as per narration. To implement the Decision Tree model, we use the drug200 dot csv dataset.

Here, we analyze patient’s data to predict the most suitable drug for them.

drug200 dataset contains Age, Sex, BP, Cholesterol, Na to K ratio and Drug.

Point to the DecisionTree.pynb DecisionTree dot ipynb is the ipython notebook file for this demonstration.
Press Ctrl,Alt and T keys

Type conda activate ml

Press Enter

Let us open the Linux terminal by pressing Ctrl, Alt and T keys together.

Activate the machine learning environment by typing conda space activate space ml Press Enter.

Go to the Downloads folder

Type cd Downloads Press Enter Type jupyter notebook

Press Enter

I have saved my code file in the Downloads folder.

Please navigate to the respective folder of your code file location.

Type, jupyter space notebook and press Enter to open Jupyter Notebook.

Show Jupyter Notebook Home page:

Click on DecisionTree.ipynb file

We can see the Jupyter Notebook Home page has opened in the web browser.

Click the DecisionTree dot ipynb file to open it.

Note that each cell will have the output displayed in this file.

Highlight The lines:

import numpy as np import pandas as pd from sklearn.model_selection import train_test_split from sklearn.tree import DecisionTreeClassifier Press Shift+Enter

These are the necessary libraries to be imported for the Decision Tree.

Please remember to Execute each cell by pressing Shift and Enter to get output.

Highlight the lines:

df_drug=pd.read_csv("drug200.csv") df_drug.head() Press Shift+Enter

We start by loading the dataset from a CSV file and display the first few rows.


Highlight The lines:

print("Length of Dataset:", len(df_drug)) print("Dataset Shape:", df_drug.shape) Press Shift+Enter

Then we print the number of rows and the shape of the dataset.
Highlight the lines:

le_sex = LabelEncoder() le_BP = LabelEncoder() Press Shift+Enter

Next, we encode the categorical variables like Sex and BP into numerical values.
Highlight The lines:

x=df_drug.drop(columns=['Drug']).values y = df_drug['Drug'].values Press Shift+Enter

We then separate the features x and the target variable y.
Highlight the lines:

x

Now we print the values of features.
Highlight the lines:

y

Similarly, we print the values of the target.
Highlight the lines:

print("\nNumber of Duplicate Rows:", df_drug.duplicated().sum()) df = df_drug.drop_duplicates() print("Dataset Shape After Removing Duplicates:", df.shape)

Next, we check for duplicate rows and remove them if found.

We also print the dataset shape after removing the duplicates.

Highlight the lines:

numerical_columns = df.select_dtypes(include=['int64', 'float64']).columns scaler = StandardScaler()

Now we use StandardScaler to standardize the numerical columns.
Highlight the lines:

df.hist(figsize=(10, 8), bins=20, color='skyblue', edgecolor='black')

plt.suptitle("Distribution of Numerical Features")

plt.show()

We then visualize the data distribution for numerical columns.
Highlight the lines:

x_train, x_test, y_train, y_test = train_test_split(x, y, test_size=0.3, random_state=100)

Press Shift+Enter

Now, we split the data into training and testing sets.
Highlight the lines:

clf_entropy = DecisionTreeClassifier(criterion="entropy", random_state=5, max_depth=4, min_samples_leaf=5)

Press Shift+Enter

After that we create a decision tree classifier using entropy as the criterion.

Entropy is the measure of disorder in the dataset. It helps to classify the features into root and branches of the decision tree.

By navigating through the root and branches, we arrive at a decision of classes.

Highlight the lines:

clf_entropy.fit(x_train, y_train) y_pred_train = clf_entropy.predict(x_train)

We then fit the classifier to the training data.

Once trained, we predict the target values for the training set.

Highlight the lines

print("Training Accuracy is", accuracy_score(y_train, y_pred_train) * 100 )

Now we print the training accuracy.
Highlight the lines:y_pred_en = clf_entropy.predict(x_test)

y_pred_en

Next we make predictions on the test data.
Highlight The lines:

accuracy = round(accuracy_score(y_test, y_pred_en) * 100, 3)

print("Accuracy is", accuracy)

Press Shift+Enter

Then we print the test accuracy
Highlight the output The accuracy is 98.33 which indicates the model performs very well.
Highlight the lines:

cm = confusion_matrix(y_test, y_pred_en) plt.figure(figsize=(10, 7)) Press Shift+Enter

To analyze model performance further, we create and display a confusion matrix.

It shows how well the model is correctly classifying the instances.

Highlight the lines:

report = classification_report(y_test, y_pred_en,zero_division=0)

print("Classification Report:")

Press Shift+Enter

We also generate and print the classification report.

This report gives precision, recall, f1-score, and support for each class.

Highlight the lines:

classes = np.unique(y) y_test_bin = label_binarize(y_test, classes=classes) y_score = clf_entropy.predict_proba(x_test)

Now we get all unique target classes from the dataset.

Next, we binarize y underscore test for multi class ROC plotting.

We binarize to handle multi class ROC as it needs binary format.

Then, we predict class probabilities using the model.

Highlight the lines:

fpr = dict() tpr = dict() roc_auc = dict() n_classes = len(classes) for i in range(n_classes): plt.legend(loc="lower right") plt.grid(True) plt.show()

We plot the ROC curve for each class using the predicted probabilities.

The ROC curve shows how well the model distinguishes between classes.

Show the output: The output displays the multi class ROC curve of our classifier.

Let us see about the AUC score that is area under the curve.

DrugA, DrugB, DrugC have an AUC score of 1, indicating perfect classification.

DrugX and DrugY have AUC scores of 0.96 and 0.98, which are very high.

The closer the curve is to the top left, the better the model performs.

Here, all curves are close to the top left corner of the plot.

So, our classifier performs very well for all classes.

Highlight the lines:

feature_names = df_drug.columns[:-1]

plt.figure(figsize=(29, 10))

Then we extract the column names excluding Drug for tree visualization.

Next, we set the figure size and plot the decision tree.

Then we save and display the tree visualization as a PNG image file.

Show the output The tree helps classify which drug to give based on patient features.

It first checks the Na to K value, then splits further using BP, Age, and Cholesterol.

Each colored box shows sample count and predicted drug type.

The tree splits until leaves mostly contain one drug, showing zero entropy.

Narration Thus, we built a decision tree to predict drug types based on patient data.

The model showed high accuracy, indicating strong predictive performance.

Show slide:

Summary


In this tutorial, we have learnt about

  • Decision Tree
  • Decision Tree Structure and Nodes
This brings us to the end of the tutorial. Let us summarize.


Show Slide:

Assignment

As an assignment, please do the following

As an assignment, please do the following:

Replace the max underscore depth as shown here.

Observe the change in Testing accuracy.

Show Slide:

Assignment Solution

Show x img

After completing the assignment, the output should match as the expected result.
Show Slide:

FOSSEE Forum

For any general or technical questions on Python for Machine Learning, visit the FOSSEE forum and post your question
Show Slide:

Thank you

This is Harini Theiveegan, a FOSSEE Summer Fellowship 2025, IIT Bombay signing off

Thanks for joining

Contributors and Content Editors

Madhurig, Nirmala Venkat

Retrieved from "https://script.spoken-tutorial.org/index.php?title=Python-for-Machine-Learning/C3/Decision-Tree/English&oldid=56999"