Course Content

  • 5_4_Decision_Tree_Regressor

Course Content


Decision tree uses a tree structure to develop regression or classification models. It incrementally divides a dataset into smaller and smaller sections while also developing an associated decision tree. The end output is a tree with decision and leaf nodes.

A regression tree is constructed using binary recursive partitioning, which is an iterative procedure that divides the data into partitions or branches and then divides each partition into smaller groups as the approach progresses up each branch.

The major distinction between classification and regression decision trees is that classification decision trees are formed using unordered values and dependent variables, whereas regression decision trees are built with ordered values and dependent variables. The regression decision trees accept both ordered and continuous values.

  • Simple to comprehend and interpret.
  • Little data preparation is required.
  • The cost of using the tree (predicting data) is proportional to the quantity of data points needed to train the tree.
  • Capable of dealing with both numerical and category data.
  • Capable of dealing with multi-output difficulties.

Recommended Courses

Share With Friend

Have a friend to whom you would want to share this course?

Download LearnVern App

App Preview Image
App QR Code Image
Code Scan or Download the app
Google Play Store
Apple App Store
598K+ Downloads
App Download Section Circle 1
4.57 Avg. Ratings
App Download Section Circle 2
15K+ Reviews
App Download Section Circle 3
  • Learn anywhere on the go
  • Get regular updates about your enrolled or new courses
  • Share content with your friends
  • Evaluate your progress through practice tests
  • No internet connection needed
  • Enroll for the webinar and join at the time of the webinar from anywhere