The document describes a project to develop a regression model to predict house prices using data on house attributes. It outlines data processing steps including variable creation, outlier treatment, and splitting data into training and validation sets. Random forest variable selection identified important predictors, which were input sequentially into a linear regression model. The model explained 90.76% of price variation and had good accuracy on training and validation data based on error rates and MAPE. Random forest accuracy was lower, so the linear regression model was selected.
This project aims to determine the housing prices of California properties for new sellers and also for buyers to estimate the profitability of the deal using various regression models.
Below are the details of the models implemented and their performance score:
Linear Regression: RMSE- 68321.7051304
Decision Tree Regressor: RMSE- 70269.5738668
Random Forest Regressor: RMSE- 52909.1080535
Support Vector Regressor: RMSE- 110914.791356
Fine Tuning the Hyperparameters for Random Forest Regressor: RMSE- 49261.2835608
Prediction of house price using multiple regressionvinovk
- Constructed a mathematical model using Multiple Regression to estimate the Selling price of the house based on a set of predictor variables.
- SAS was used for Variable profiling, data transformations, data preparation, regression modeling, fitting data, model diagnostics, and outlier detection.
ABSTRACT
House Price Index is commonly used to estimate the changes in housing price. Since housing price is strongly correlated to other factors such as location, area, population, it requires other information apart from House price prediction to predict individual housing price. There has been a considerably large number of papers adopting traditional machine learning approaches to predict housing prices accurately, but they rarely concern about the performance of individual models and neglect the less popular yet complex models. As a result, to explore various impacts of features on prediction methods, this paper will apply both traditional and advanced machine learning approaches to investigate the difference among several advanced models. This paper will also comprehensively validate multiple techniques in model implementation on regression and provide an optimistic result for housing price prediction.
INTODUCTION
House price prediction is great project to learn and apply the machine learning algorithm. The basic idea behind this project is we are training the machine using the machine learning algorithm from the data set.
In this busy world it is very difficult to find a house according to our need and budget. It becomes more difficult to find the house in metropolitan cities like Mumbai, Kolkata, Delhi, etc. This project uses the data of Mumbai city in order to train and test the machine so that it become capable of predicting the price of house. Machine learning algorithm makes it easy to know the price of houses depending on the location, area, number of bedrooms, etc.
In this project Random Forest Regression, Linear Regression, and Decision Tree Machine learning algorithm has been used to compare the efficiency of the algorithm. Based on comparison we predict which algorithm best suits for the prediction of price of house in Mumbai.
CONCLUSION AND FUTURE SCOPE
The model designed accuracy depends on the dataset selected, better the dataset better will be the accuracy. Best suited model applied is Random Forest. This can be applied to datset of any city for their house price prediction. The project can be enhanced by UI designing through they can predict the price in more easier and interactive way. In this busy world it will be of immense use to search for a house at near to our workplace.
DATASET LINK
https://www.kaggle.com/
This project aims to determine the housing prices of California properties for new sellers and also for buyers to estimate the profitability of the deal using various regression models.
Below are the details of the models implemented and their performance score:
Linear Regression: RMSE- 68321.7051304
Decision Tree Regressor: RMSE- 70269.5738668
Random Forest Regressor: RMSE- 52909.1080535
Support Vector Regressor: RMSE- 110914.791356
Fine Tuning the Hyperparameters for Random Forest Regressor: RMSE- 49261.2835608
Prediction of house price using multiple regressionvinovk
- Constructed a mathematical model using Multiple Regression to estimate the Selling price of the house based on a set of predictor variables.
- SAS was used for Variable profiling, data transformations, data preparation, regression modeling, fitting data, model diagnostics, and outlier detection.
ABSTRACT
House Price Index is commonly used to estimate the changes in housing price. Since housing price is strongly correlated to other factors such as location, area, population, it requires other information apart from House price prediction to predict individual housing price. There has been a considerably large number of papers adopting traditional machine learning approaches to predict housing prices accurately, but they rarely concern about the performance of individual models and neglect the less popular yet complex models. As a result, to explore various impacts of features on prediction methods, this paper will apply both traditional and advanced machine learning approaches to investigate the difference among several advanced models. This paper will also comprehensively validate multiple techniques in model implementation on regression and provide an optimistic result for housing price prediction.
INTODUCTION
House price prediction is great project to learn and apply the machine learning algorithm. The basic idea behind this project is we are training the machine using the machine learning algorithm from the data set.
In this busy world it is very difficult to find a house according to our need and budget. It becomes more difficult to find the house in metropolitan cities like Mumbai, Kolkata, Delhi, etc. This project uses the data of Mumbai city in order to train and test the machine so that it become capable of predicting the price of house. Machine learning algorithm makes it easy to know the price of houses depending on the location, area, number of bedrooms, etc.
In this project Random Forest Regression, Linear Regression, and Decision Tree Machine learning algorithm has been used to compare the efficiency of the algorithm. Based on comparison we predict which algorithm best suits for the prediction of price of house in Mumbai.
CONCLUSION AND FUTURE SCOPE
The model designed accuracy depends on the dataset selected, better the dataset better will be the accuracy. Best suited model applied is Random Forest. This can be applied to datset of any city for their house price prediction. The project can be enhanced by UI designing through they can predict the price in more easier and interactive way. In this busy world it will be of immense use to search for a house at near to our workplace.
DATASET LINK
https://www.kaggle.com/
A ppt based on predicting prices of houses. Also tells about basics of machine learning and the algorithm used to predict those prices by using regression technique.
House Price Estimates Based on Machine Learning Algorithmijtsrd
Housing prices are increasing every year, necessitating the creation of a long term housing price strategy. Predicting a homes price will assist a developer in determining a homes purchase price, as well as a consumer in determining the best time to buy a home. The sale price of real estate in major cities depends on the specific circumstances. Housing prices are constantly changing from day to day and are sometimes fired rather than based on estimates. Predicting real estate prices by real factors is a key element as part of our analysis. We want to make our test dependent on all of the simple metrics that are taken into account when deciding the significance. In this research we use linear regression techniques pathway and our results are not self inflicted process rather is a weighted method of various techniques to give the most accurate results. There are fifteen features in the data collection. In this research. There has been an effort to build a forecasting model for determining the price based on the variables that influence the price.The results have proven to be effective lower error and higher accuracy than individual algorithms are used. Jakir Khan | Dr. Ganesh D "House Price Estimates Based on Machine Learning Algorithm" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-4 , June 2021, URL: https://www.ijtsrd.compapers/ijtsrd42367.pdf Paper URL: https://www.ijtsrd.comcomputer-science/other/42367/house-price-estimates-based-on-machine-learning-algorithm/jakir-khan
Predicting Moscow Real Estate Prices with Azure Machine LearningLeo Salemann
With only three months' instruction, a five-person team uses Azure Machine Learning Studio to predict Moscow real estate prices based on property descriptors, macroeconomic indicators, and geospatial data.
House Price Prediction An AI Approach.Nahian Ahmed
Suppose you have a house. And you want to sell it. Through House Price Prediction project you can predict the price from previous sell history.
And we make this prediction using Machine Learning.
Basics of machine learning including architecture, types, various categories, what does it takes to be an ML engineer. pre-requisites of further slides.
Abstract: This PDSG workshop introduces basic concepts of multiple linear regression in machine learning. Concepts covered are Feature Elimination and Backward Elimination, with examples in Python.
Level: Fundamental
Requirements: Should have some experience with Python programming.
A Support Vector Machine (SVM) is a discriminative classifier formally defined by a separating hyperplane. In other words, given labeled training data (supervised learning), the algorithm outputs an optimal hyperplane which categorizes new examples. In two dimentional space this hyperplane is a line dividing a plane in two parts where in each class lay in either side.
Decision tree is a type of supervised learning algorithm (having a pre-defined target variable) that is mostly used in classification problems. It is a tree in which each branch node represents a choice between a number of alternatives, and each leaf node represents a decision.
A ppt based on predicting prices of houses. Also tells about basics of machine learning and the algorithm used to predict those prices by using regression technique.
House Price Estimates Based on Machine Learning Algorithmijtsrd
Housing prices are increasing every year, necessitating the creation of a long term housing price strategy. Predicting a homes price will assist a developer in determining a homes purchase price, as well as a consumer in determining the best time to buy a home. The sale price of real estate in major cities depends on the specific circumstances. Housing prices are constantly changing from day to day and are sometimes fired rather than based on estimates. Predicting real estate prices by real factors is a key element as part of our analysis. We want to make our test dependent on all of the simple metrics that are taken into account when deciding the significance. In this research we use linear regression techniques pathway and our results are not self inflicted process rather is a weighted method of various techniques to give the most accurate results. There are fifteen features in the data collection. In this research. There has been an effort to build a forecasting model for determining the price based on the variables that influence the price.The results have proven to be effective lower error and higher accuracy than individual algorithms are used. Jakir Khan | Dr. Ganesh D "House Price Estimates Based on Machine Learning Algorithm" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-4 , June 2021, URL: https://www.ijtsrd.compapers/ijtsrd42367.pdf Paper URL: https://www.ijtsrd.comcomputer-science/other/42367/house-price-estimates-based-on-machine-learning-algorithm/jakir-khan
Predicting Moscow Real Estate Prices with Azure Machine LearningLeo Salemann
With only three months' instruction, a five-person team uses Azure Machine Learning Studio to predict Moscow real estate prices based on property descriptors, macroeconomic indicators, and geospatial data.
House Price Prediction An AI Approach.Nahian Ahmed
Suppose you have a house. And you want to sell it. Through House Price Prediction project you can predict the price from previous sell history.
And we make this prediction using Machine Learning.
Basics of machine learning including architecture, types, various categories, what does it takes to be an ML engineer. pre-requisites of further slides.
Abstract: This PDSG workshop introduces basic concepts of multiple linear regression in machine learning. Concepts covered are Feature Elimination and Backward Elimination, with examples in Python.
Level: Fundamental
Requirements: Should have some experience with Python programming.
A Support Vector Machine (SVM) is a discriminative classifier formally defined by a separating hyperplane. In other words, given labeled training data (supervised learning), the algorithm outputs an optimal hyperplane which categorizes new examples. In two dimentional space this hyperplane is a line dividing a plane in two parts where in each class lay in either side.
Decision tree is a type of supervised learning algorithm (having a pre-defined target variable) that is mostly used in classification problems. It is a tree in which each branch node represents a choice between a number of alternatives, and each leaf node represents a decision.
Determining the sales price of a house based on suitable predictor variables (24) of 71 variables, best describing the house to predict the sales price
본 자료에서는 이러한 고민에 다소나마 도움을 드리고자 중국의 ICT시장과 모바일 시장을 보다 심도 있게 정리하였으며, 기 진출을 시도한 스타트업(말랑스튜디오, VCNC, 피키캐스트)의 사례를 통해 국내 스타트업이 중국에 진출 시 참고할만한 사항들을 도출했습니다.
또한 중국어라는 동일한 언어와 유사한 문화권이면서도 중국 외 시장인 대만, 싱가포르,홍콩에 대한 정보를 정리하여, 중국 외 중화권 국가를 통해 중국에 진출하거나, 반대로 중국에서의 성과에 힘입어 동남아를 가는 길목에 있는 국가에 들어가고자 하는 기업들에게 도움이 되는 내용을 전달하고자 하였습니다.
중화권 ICT 시장 진출의 첫걸음(2014 중화권 ICT 시장조사 보고서)은 미래창조과학부와 정보통신산업진흥원(NIPA)가 주관한 '스마트콘텐츠 중화권 진출 지원 사업' 중 플래텀이 시장 정보 컨설팅 연구를 맡아 제작한 자료입니다.
본 자료가 국내 스타트업이 중화권 시장에 진출하는 데 실용적으로 도움이 되기를 바라며, 대한민국 스타트업의 성공적인 중화권 진출을 진심으로 기원합니다.
본문 이미지는 연구, 분석 목적으로 쓰여 졌으며 출처를 표기하였습니다. 이 보고서의 내용을 대외적으로 사용하실 때에는 반드시 정보통신산업진흥원(NIPA) 및 플래텀의 연구 결과임을 밝혀야 합니다.
그 밖에 저작권관련 별도 협의가 필요한 경우 정보통신산업진흥원(NIPA) 및 플래텀에 연락 주시기 바랍니다.
contact@platum.kr
MM School Brochure | International School in ChhattisgarhMM School
MM School is an establishment that spotlights on building up the fundamental aptitudes and demeanor required in the present century. The school intensely holds onto innovation as an indispensable piece of scholastic practices and makes a benchmark in innovation empowered learning background.
Lecture 10 - Model Testing and Evaluation, a lecture in subject module Statis...Maninda Edirisooriya
Model Testing and Evaluation is a lesson where you learn how to train different ML models with changes and evaluating them to select the best model out of them. This was one of the lectures of a full course I taught in University of Moratuwa, Sri Lanka on 2023 second half of the year.
Семинар «Использование современных информационных технологий для решения современных задач физики частиц» в московском офисе Яндекса, 3 июля 2012
Андрей Гулин, разработчик, Яндекс
The objective of the project is to use the dataset 'Factor-Hair-Revised.csv' to build an optimum regression model to predict satisfaction.
Perform exploratory data analysis on the dataset. Showcase some charts, graphs. Check for outliers and missing values.
Check evidence of multicollinearity.
Perform simple linear regression for the dependent variable with every independent variable.
Perform PCA/Factor analysis by extracting 4 factors. Interpret the output and name the Factors.
Perform Multiple linear regression with customer satisfaction as dependent variables and the four factors as independent variables.
Unleashing the Power of Data_ Choosing a Trusted Analytics Platform.pdfEnterprise Wired
In this guide, we'll explore the key considerations and features to look for when choosing a Trusted analytics platform that meets your organization's needs and delivers actionable intelligence you can trust.
Enhanced Enterprise Intelligence with your personal AI Data Copilot.pdfGetInData
Recently we have observed the rise of open-source Large Language Models (LLMs) that are community-driven or developed by the AI market leaders, such as Meta (Llama3), Databricks (DBRX) and Snowflake (Arctic). On the other hand, there is a growth in interest in specialized, carefully fine-tuned yet relatively small models that can efficiently assist programmers in day-to-day tasks. Finally, Retrieval-Augmented Generation (RAG) architectures have gained a lot of traction as the preferred approach for LLMs context and prompt augmentation for building conversational SQL data copilots, code copilots and chatbots.
In this presentation, we will show how we built upon these three concepts a robust Data Copilot that can help to democratize access to company data assets and boost performance of everyone working with data platforms.
Why do we need yet another (open-source ) Copilot?
How can we build one?
Architecture and evaluation
Learn SQL from basic queries to Advance queriesmanishkhaire30
Dive into the world of data analysis with our comprehensive guide on mastering SQL! This presentation offers a practical approach to learning SQL, focusing on real-world applications and hands-on practice. Whether you're a beginner or looking to sharpen your skills, this guide provides the tools you need to extract, analyze, and interpret data effectively.
Key Highlights:
Foundations of SQL: Understand the basics of SQL, including data retrieval, filtering, and aggregation.
Advanced Queries: Learn to craft complex queries to uncover deep insights from your data.
Data Trends and Patterns: Discover how to identify and interpret trends and patterns in your datasets.
Practical Examples: Follow step-by-step examples to apply SQL techniques in real-world scenarios.
Actionable Insights: Gain the skills to derive actionable insights that drive informed decision-making.
Join us on this journey to enhance your data analysis capabilities and unlock the full potential of SQL. Perfect for data enthusiasts, analysts, and anyone eager to harness the power of data!
#DataAnalysis #SQL #LearningSQL #DataInsights #DataScience #Analytics
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Discussion on Vector Databases, Unstructured Data and AI
https://www.meetup.com/unstructured-data-meetup-new-york/
This meetup is for people working in unstructured data. Speakers will come present about related topics such as vector databases, LLMs, and managing data at scale. The intended audience of this group includes roles like machine learning engineers, data scientists, data engineers, software engineers, and PMs.This meetup was formerly Milvus Meetup, and is sponsored by Zilliz maintainers of Milvus.
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Round table discussion of vector databases, unstructured data, ai, big data, real-time, robots and Milvus.
A lively discussion with NJ Gen AI Meetup Lead, Prasad and Procure.FYI's Co-Found
Analysis insight about a Flyball dog competition team's performanceroli9797
Insight of my analysis about a Flyball dog competition team's last year performance. Find more: https://github.com/rolandnagy-ds/flyball_race_analysis/tree/main
The Building Blocks of QuestDB, a Time Series Databasejavier ramirez
Talk Delivered at Valencia Codes Meetup 2024-06.
Traditionally, databases have treated timestamps just as another data type. However, when performing real-time analytics, timestamps should be first class citizens and we need rich time semantics to get the most out of our data. We also need to deal with ever growing datasets while keeping performant, which is as fun as it sounds.
It is no wonder time-series databases are now more popular than ever before. Join me in this session to learn about the internal architecture and building blocks of QuestDB, an open source time-series database designed for speed. We will also review a history of some of the changes we have gone over the past two years to deal with late and unordered data, non-blocking writes, read-replicas, or faster batch ingestion.
2. Outline
■ Project Objective
■ Data Source and Variables
■ Data Processing
■ Method of Analysis
■ Result
■ Predicted House Prices
All coding and model building is done using R software
3. Objective
■ Create an analytical framework to understand
– Key factors impacting house price
■ Develop a modeling framework
– To estimate the price of a house that is up for sale
4. Data Source and Variables
■ Kaggle competition - “House Prices: Advanced Regression Techniques”
– Dataset prepared by Dean De Cock
■ Variables:
– 79 variables present in the dataset
■ Variable named “SalePrice”
– Dependent variable
– Represent final price at which the house was sold
■ Remaining 78 variables
– Represent different attributes of the house like area, car parking, number of
fireplaces, etc.
5. Data Processing
■ Normalizing Response Variable
■ Training Vs Validation split
– Train data – 75%
– Validation Data – 25%
■ Data cleansing
– Variable treatments
■ Missing value treatment:
– Continuous variables
– Character variables
■ Outlier treatment
■ Variable creations:
– Character variables were converted to indicators
– Based on train data, further grouping of character variables were done and new indicators were created
6. Data Processing – Normalizing
Response Variable
■ The response variable is converted to
its logarithmic form to normalize it.
■ Underlying reason:
– Satisfying the basic assumption of
Ordinary least square
7. Data Processing – Training Vs
Validation split
■ Training Data
– Containing 75% of the total observations picked up at a random.
– Model is developed on this dataset.
■ Validation Data
– Containing remaining 25% of the total observations.
– Validation of the model is carried out based on this data
– Model tuning, if required, is carried based on the model performance on the
validation data
8. Data Processing – Missing value
treatment
■ Continuous Variable
– Missing values are replaced by the median of the corresponding variable.
■ Why median and not mean?
– Mean is more prone and get highly impacted by outliers
– Median is a more stable measure
■ Character Variable
– A separate category is created for missing values
■ This helps us retaining the prediction power of a variable
■ Also impact of missing values on the dependent variable can be established
9. Data Processing – Outlier treatment
■ Upper tailed values
– Cut-off value: sum of 99th percentile and
1.5 times of IQR of the corresponding
variables.
■ Replaced by 99th percentile
■ Lower tailed values
– Cut-off value: value of 1st percentile
point less 1.5 times IQR
■ Replaced by 1st percentile
10. Data Processing – Variable creation
■ Character Variables
– (n – 1) indicators are created for a
character variable containing n
different categories
– Separate indicator created for missing
value
■ Additional Indicator Variables
– Based on bivariate plots: if two or more
categories contains similar level of
value of dependent variable, they are
combined and converted into an
indicator
■ For example, Alley has three levels:
Grvl, Missing and Pave. A new
variable was created for Missing and
Pave category as they both have a high
median value of the dependent
variable.
11. Method of Analysis
■ Variable Selection Using Random Forest
■ Multiple Linear Regression
– Significance
■ t value and probability of t value
– Goodness of fit
■ Adjusted R-square
– Multicollinearity
■ Random Forest
■ Model Accuracy
– Error rate
– MAPE
12. Introduction – Random Forest
• Random Forest operate by constructing multitude of
decision trees at training time and outputting the mean
prediction of the individual trees. It also correct the
decision trees’ habit of over fitting to the training set.
• It is also been used to rank the importance of variables
in a regression problem in a natural way.
• In a regression tree, for each independent variable, the
data is split at several split points. Sum of Squared
Error(SSE) at each split point between the predicted
value and the actual values is calculated. The variable
resulting in minimum SSE is selected for the node.
Then this process is recursively continued till the
entire data is covered.
13. Method of Analysis – Variable
Selection
■ Random Forest Model
– Model
■ A model has been built on train data
using random forest – number of
trees:100
– Importance of Variables
■ Importance of Variables were extracted
■ Variables are sorted descending based
on the importance measure
– Variables selected in order of
importance measure are introduced into
the OLS model Introduce one-by-one variable into model 2
Variables sorted in descending order of their importance
Table of variable importance
Model1- Random Forest
14. Method of Analysis – Multiple Linear
Regression
■ Ordinary Least Square (OLS) :
– Simplest method regression in which the
unknown coefficients of features are estimated
with the goal to minimize the sum square
errors. i.e.
𝑚𝑖𝑛 (𝑌 − 𝑌)
2
– Visually this is seen as the sum of the vertical
distances between each data point in the set
and the corresponding point on the regression
line
15. Method of Analysis – Multiple Linear
Regression
■ Iteration
– Select one variable at a time from the variable importance table created using random forest
■ Significance
– Check significance of new variable along with existing variables by its t-value and probability of t-
statistics.
– If R-square is improved, keep the variable, else drop it
■ Multicollinearity
– Multicollinearity is checked at each step – ensuring the maximum value is < 4
– If new variable has multicollinearity above threshold value – drop it
– If introducing the new variable increases the multicollinearity of any existing variable – then the variable
with lowest t-value is dropped
■ Model Accuracy
– After adding the new variable, the model accuracy on train and test data using error rate and MAPE is
checked.
– Drop the new variable if the model accuracy falls.
16. Method of Analysis – Random Forest
• Variables
• Using the same variable used in the linear regression
• Trees and Nodes
• Checking various combinations of number of trees and maximum number of nodes to get the best
result.
• Using number of trees = 100 and maximum nodes = 10 for best fitted model.
• Model Accuracy
• Checking model accuracy using error rate and MAPE
• Decision
• Drop the variable if the model accuracy falls or remains same.
17. Method of Analysis – Model Accuracy
• Error rate
• Calculated as : Error Rate = 1 −
𝑌
𝑌
∗ 100
• Calculated minimum error rate and maximum error rate for train and test data.
• Aim is to reduce the error rate
• Difference between minimum and maximum error rate between training and validation.
• MAPE
• Mean Absolute Percentage Error is calculated by :
M𝐴𝑃𝐸 = 𝑀𝑒𝑎𝑛 1 −
𝑌
𝑌
∗ 100
• Aim is to minimize the MAPE
19. Result – Interpretation
■ Linear Regression Model
– All the variables taken in the final
model.
– Adjusted R-square is: 90.76
■ i.e. these variables together are
explaining 90.76% variability of
SalePrice.
– All of the variables are significant.
– Multicollinearity is not severe. All
VIFs’are below 4.
20. Result – Interpretation
■ Linear Regression Model
– Train Data
■ Minimum error rate is -9.3%
■ Maximum error rate is 3.35%
■ MAPE is 0.697
– Test Data
■ Minimum error rate is -8.68%
■ Maximum error rate is 3.41%
■ MAPE is 0.778
■ Random Forest Model
– Train Data
■ Minimum error rate is -6.9%
■ Maximum error rate is 4.49%
■ MAPE is 0.993
– Test Data
■ Minimum error rate is -7.29%
■ Maximum error rate is 3.37%
■ MAPE is 1.02
21. Result
■ MAPE
– Low for Linear Regression
– High for Random Forest
■ Linear Regression Model Chosen – based on minimum MAPE