Next Article in Journal
Learning to Rank Sports Teams on a Graph
Previous Article in Journal
Pharmacological and Cosmeceutical Potential of Seaweed Beach-Casts of Macaronesia
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Machine Learning Models and Actual Transaction Data for Predicting Real Estate Prices

Department of Information Management, National Chi Nan University, 1 University Rd., Puli, Nantou 54561, Taiwan
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(17), 5832; https://doi.org/10.3390/app10175832
Submission received: 22 July 2020 / Revised: 18 August 2020 / Accepted: 20 August 2020 / Published: 23 August 2020
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
Real estate price prediction is crucial for the establishment of real estate policies and can help real estate owners and agents make informative decisions. The aim of this study is to employ actual transaction data and machine learning models to predict prices of real estate. The actual transaction data contain attributes and transaction prices of real estate that respectively serve as independent variables and dependent variables for machine learning models. The study employed four machine learning models-namely, least squares support vector regression (LSSVR), classification and regression tree (CART), general regression neural networks (GRNN), and backpropagation neural networks (BPNN), to forecast real estate prices. In addition, genetic algorithms were used to select parameters of machine learning models. Numerical results indicated that the least squares support vector regression outperforms the other three machine learning models in terms of forecasting accuracy. Furthermore, forecasting results generated by the least squares support vector regression are superior to previous related studies of real estate price prediction in terms of the average absolute percentage error. Thus, the machine learning-based model is a substantial and feasible way to forecast real estate prices, and the least squares support vector regression can provide relatively competitive and satisfactory results.

1. Introduction

The real estate market is one of the most crucial components of any national economy. Hence, observations of the real estate market and accurate predictions of real estate prices are helpful for real estate buyers and sellers as well as economic specialists. However, real estate forecasting is a complicated and difficult task owing to many direct and indirect factors that inevitably influence the accuracy of predictions. In general, factors influencing real estate prices could be quantitative or qualitative [1]. The quantitative factors possibly include macroeconomic factors [2], business cycles [3], and real estate attributes [4]. The macroeconomic factors contain unemployment rates, share index, current account of a country, industrial production, and gross domestic product [2]. Attributes of real estate, for example, includes past sale prices, land area, years of constructions, floor space, surface area, number of floors and building conditions [1,4]. The qualitative factors refer to subject preferences of decision makers, such as views [5], building styles, and living environment [1]. However, some difficulties arise in data collection for qualitative factors. For qualitative factors, sometimes these data are suffering from lack of measurements. Data of qualitative factors sometimes are suffering from lack of measurements. Thus, qualitative factors are hard to measure [1]. Therefore, this study did not take qualitative factors influencing real estate prices into considerations and used quantitative data gathered from actual transaction data recording details of real estate transaction data in Taiwan. Four machine learning models were used to forecast real estate prices accordingly. The rest of this study is organized as follows. Section 2 presents literature review of real estate prediction. Section 3 illustrates methods used in this study. Section 4 introduces the proposed real estate appraising system. The numerical results are depicted in Section 5. Section 6 provides conclusions.

2. The Literature Review of Real Estate Price Predictions

Some studies of real estate prices predictions are presented as follows. Singh et al. [6] employed the concept of big data to predict housing sale data in Iowa, using three models to forecast house sale prices: linear regression, random forest, and gradient boosting. The numerical results indicated that the gradient boosting model outperforms the other forecasting models in terms of forecasting accuracy. Segnon et al. [7] presented a logistic smooth transition autoregressive fractionally integrated process to predict housing price volatility in the U.S.A. and analyzed complicated statistical models based on assumptions of the variance process. The numerical results revealed that the Markov-switching multifractal and fractionally integrated generalized autoregressive conditional heteroscedastic models provide satisfied forecasting accuracy.
Kang [8] developed a news article-based forecast model to predict Jeonse prices in South Korea. The Internet search intensity of keywords from news was treated as the independent variable. The numerical results showed that the designed models obtain more accurate results than time series techniques. Giudice et al. [9] used genetic algorithms to forecast real estate rental prices when geographic locations and four real estate attributes are considered. The multivariate regression technique was performed to forecast the same data. Numerical results indicated that the genetic algorithms are superior to the multivariate regression in term of prediction accuracy. Park and Bae [10] designed a house price prediction system by machine learning approaches to help house sellers or real estate agents with house price evaluations. In their investigation, data were collected from the Multiple Listing Service of the Metropolitan Regional Information Systems. The numerical results indicated that repeated incremental pruning to produce error reduction (RIPPER) algorithm obtains more accurate forecasting results than the other forecasting models.
Bork and Møller [11] employed dynamic model averaging and dynamic model selection to forecast house price growth rates in the 50 states of the U.S.A. The presented forecasting system captures house price growth rates by varying the model and the coefficients’ change over time and across locations. Thus, the forecasting results provided by the proposed system are substantial. Plakandaras et al. [12] designed a hybrid model as an early warning system, including ensemble empirical mode decomposition and support vector regression, for predicting sudden house price drops. The proposed forecasting approach generates superior results over other forecasting models in terms of forecasting accuracy. Chen et al. [13] developed a housing price analysis system using information from public websites to analyze the total ratio of the average value and standard deviation of housing prices. They reported that the designed housing price analysis system is a helpful way for obtaining insights into housing prices. Lee et al. [14] employed fuzzy adaptive networks to forecast pre-owned housing prices in Taipei by taking both objective variables and subjective variables into considerations. The empirical results indicated that the fuzzy adaptive networks outperform back-propagation neural networks and the adaptive network fuzzy inference system in terms of forecasting accuracy.
Antipov and Pokryshevskaya [15] used random forest to appraise residential estate of Saint Petersburg, Russia and reported that the random forest approach outperforms the other forecasting methods in prediction accuracy. In addition, the authors claimed that the random forest approach is capable of dealing with missing values and categorical variables. Kontrimas and Verikas [16] presented an ensemble learning system integrating ordinary least squares linear regression and support vector regression to appraise real estate. In addition, weights based on value zones provided by experts of the register center and weights generated by the self-organizing map (SOM) were used by the ensemble learning system. The numerical results revealed that the ensemble learning system with weights generated by SOM outperform the ensemble learning systems with weights provided by experts. Furthermore, the ensemble systems can reach more accurate forecasting results than the other single forecasting models.
Gupta et al. [17] employed time series models with or without the information content of 10 or 120 additional quarterly macroeconomic series to forecast the U.S. real house price index. Their study concluded that the utilization of fundamental economic variables could increase the forecasting accuracy and especially be effective for the 10 fundamental economic variables in the dynamic stochastic general equilibrium model. Kusan et al. [18] developed a grading fuzzy logic model to forecast house selling prices. Many housing attributes, such as public transportations systems and environmental factors, served as inputs of the fuzzy logic system. The numerical results indicated that the proposed fuzzy logic system captures the patterns of house selling prices and lead to satisfied forecasting accuracy. The collection of actual transaction data of real estate has been performed since August 2012 by the Ministry of the Interior of Taiwan. Thus, the motivation of this study is to examine the performance of machine learning models with actual transaction data in forecasting real estate prices.

3. Methods

3.1. Least Squares Support Vector Regression

The support vector machines [19,20] technique was originally designed for classification problems. For dealing with regression problems, support vector regression [21,22,23] was proposed and has become a popular alternative for estimating linear or non-linear prediction problems. However, owing to solving quadratic functions in the procedure of support vector regression, the computation burden is quite challenging. Thus, the least squares support vector regression [24] was designed to decrease the computation load by converting a quadratic programming problem into a linear problem. For an input-output dataset { X i , Y i } , i = 1…N, the LSSVR model can be illustrated as Equation (1) [24].
Min :   f ( W , τ ) = 1 2 W T W + 1 2 Ω i = 1 N τ i 2
Subject to
y i = W T · ψ ( x i ) + δ + τ i , i = 1 , , N
where W represents the weighted vector,   Ω denotes the penalty factors controlling the trade-off between the optimization of approximation error and flatness of the approximated function, τ i is the ith error vector, ψ ( x i ) is the nonlinear mapping function transferring the original input space into a high dimension input space, δ indicates the bias parameter. Using the Lagrange function, the Lagrange form of Equation (1) is depicted as Equation (2)
L ( W , δ , α ,   τ ) = f ( W , τ ) + i = 1 N α i ( y i W T · ψ ( x i ) δ τ i )
where α i represent Lagrange multipliers.
Applying the Karush–Kuhn–Tucker conditions [25,26,27] and setting derivatives with respect to four variables, W , δ , α ,   and   τ , equal to zero, Equations (3)–(6) can be obtained.
L ( W , δ , α ,   τ ) W = 0 ,   W = i = 1 N δ i ψ ( x i )
L ( W , δ , α ,   τ ) δ   ,   i = 1 N α i = 0
L ( W , δ , α ,   τ ) τ   ,   α i = Ω τ i
L ( W , δ , α ,   τ ) δ   ,   W T ψ ( x i ) + δ + τ i y i = 0
Then, by the least squares method, the LSSVR model can be reformed as Equation (7):
y ( x ) = i = 1 N α i K ( x , x i ) + δ
K ( x , x i )   =   ψ ( x ) T · ψ ( x i ) T
where K ( x , x i ) represents the kernel function satisfying the Mercer’s principle [28]. The radial basis function with a variance σ 2 expressed by Equation (9) is specified here as a kernel function.
K ( x , x i ) = e x p ( x x i 2 2 σ 2 )

3.2. Classification and Regression Trees

Proposed by Breiman et al. [29], the classification and regression tree is one of the most popular techniques in dealing with classification or regression problems. The Gini measurement and least-squared deviation measurement are performed by CART models for categorical and numerical problems correspondingly [29,30]. Let the pth sample be illustrated as ( I p , 1 , I p , 2 ,   I p , n . O p ) , where I p , n is the value of the pth sample with n features, and O p is the corresponding output value of the sample. For a regression problem of CART, the minimization of the least-squared deviation measure of impurity [29] represented by Equation (10) serves as a decision to determine the split-up of trees into branches.
1 N V U r ( O p O ¯ r ) 2 + 1 N V U l ( O p O ¯ l ) 2
where N is the total number of training samples, U r and U l are training data sets directing to the right child node and the left child node correspondingly.   O ¯ r and O ¯ l are mean output values of the right node and the left node respectively.

3.3. General Regression Neural Networks

Based on the Parzen window non-parametric estimation [31] of a probability density function, the general regression neural network [32] is a probabilistic neural network that is able to cope with linear or non-linear forecasting problems with continuous outcome values. Suppose a probability density function f ( I , O ) is associated with input vectors I and output vectors O. The regression of O on i is illustrated as Equation
E ( o | I ) = o f ( I , o ) d o f ( I , o ) d o
where I is the expected value of i.
Furthermore, the process of conducting a general regression neural network model can be treated as dealing with the kernel regression problems expressed as Equation (12), where ( I p   , O p ) is a data pair of (I,O), and σ is the smoothing parameter.
O   ( I )   =   p = 1 q O p exp [   ( I   I p   ) T ( I     I p   ) 2 σ 2 ] p = 1 q exp [   ( I   I p   ) T ( I     I p   ) 2 σ 2 ]
where ( I p   , O p ) is a data pair of (I, O), σ is the smoothing parameter.

3.4. Backpropagation Neural Networks

The backpropagation learning algorithms [33] make up a powerful and popular method to train multilayer perceptron neural networks. The backpropagation neural network containing one or more hidden layers delivers data patterns forward from the input layer through hidden layers to the output layer and generates output values. Feedback errors represented by the difference between the actual output values and the output values of the networks are then sent backward from the direction of the output layer to the input layer. During the process errors’ propagation, a chain rule is applied to obtain the updated weights between neuros. By passing data forward and transmitting errors backward iteratively, the training error value decreases. For simplicity and generality, three layers MLP is illustrated for addressing the learning procedures of backpropagation neural networks [34]. Suppose the input data set is represented as I, the h-th hidden neuro obtains an input depicted as Equation (13):
H N I h = f ( p = 1 l U h p I p + B h ) = f ( N E T h ) , p = 1 l , h = 1 m
The f ( · ) is the activation function, B h is the bias of the h-th hidden neuro, I p is the p-th input value, U h p is the weight between the p-th input neuro and the h-th hidden neuro.
The output of the h-th hidden neuro is represented as Equation (14).
H N O q = h = 1 m V q h f ( p = 1 l U h p I p + B h ) , h = 1 m , q = 1 n
Then, the output of the q-th output neuro is represented as Equation (15).
Y q = f ( h = 1 l V q h H N I h + B q ) , h = 1 m , q = 1 n
where n is the number of output nodes, Y q is the output value of the q-th output neuro, A q is the q-th actual value.
The training error is expressed as Equation (16).
E = 1 2 q = 1 n ( A q Y q ) 2
Then, by the gradient-descent method and chain rules, updated weights between the output layer and the hidden layer can be obtained and illustrated as Equation (17) and Equation (18).
Δ V q h ( t ) = ( E V q h ( t ) )   =   ( E Y q ) ( Y q H N O q ) ( H N O q U h p ( t ) )   =   ( A q Y q ) f ´ ( H N O q ) ( H N I h )
V q h ( t + 1 ) = Ω V q h ( t ) + γ Δ V q h ( t )
where V q h ( t + 1 ) is the weight connecting hidden neuro h and output neuro q for (t + 1)-th epoch, Ω is the momentum, γ denotes the learning rate.
Then, the learning algorithm between the input layer and the hidden layer can be illustrated as Equations (19) and (20).
Δ U h p ( t ) = ( E U h p ( t ) )   =   ( E H N I h ) ( H N I h N E T h ) ( N E T h U h p ( t ) ) =   q = 1 n [   ( A q Y q ) f ´ ( H N O q ) V q h ( t ) ]   f ´ ( N E T h )   I p
U h p ( t + 1 ) = Ω U h p ( t ) + γ Δ U h p ( t )
where U h p ( t + 1 ) is the weight between input neuro p and hidden neuro h for (t+1)-th epoch.

4. Data Collection and the Proposed Real Estate Appraising Framework

4.1. Data Collection

The data were collected from actual transaction price data of Taichung, Taiwan (https://lvr.land.moi.gov.tw/) from 2016/04 to 2019/04. Real estates with buildings served as objectives of this study. Due to blank, unclear and outlier data, data cleansing was conducted before using the data. Blank and unclear data were deleted, and the interquartile range technique was performed to deal with outlier data. After data cleansing, in total 32,215 data observations were used in this study. In addition, the real estate attributes were rearranged, and we finally utilized a total of twenty-three independent variables and one dependent variable to forecast real estate transaction prices. For example, the real estate addresses were transformed into geographical coordinates by the Taiwan Geospatial One-Stop system (https://www.tgos.tw/tgos/Web/Address/TGOS_Address.aspx) provided by the Ministry of the Interior of Taiwan. Three attributes namely “non-urban use land zone”, “urban use land zone”, and “non-urban land use compilation” were integrated into an attribute of “purpose of land use”. Ages of real estate were generated from attributes of transaction dates and completion dates of buildings. Table 1 depicts variables used in this study. Moreover, the Pearson correlation coefficient was used to select independent variables with significant coefficient of correlation. Pearson correlation coefficients of independent variables to the real estate price are listed in Table 2. Independent variables with absolute values of Pearson correlation coefficients larger than 0.1 were left. Totally, eleven independent variables were selected.

4.2. The Proposed Real Estate Appraising Framework

Figure 1 illustrates the proposed real estate appraising framework in this study. After completing the preparation of data, data were spitted into a training data set and a testing data set with data sizes of eighty percent and twenty percent of the total data correspondingly. Then, a 5-fold cross validation was performed to examine the robustness of machine learning models in forecasting real estate pierces. In this study, genetic algorithms (GA) [35] were used to tune parameters of machine learning models. The average absolute percentage error served as the objective function of genetic algorithms. In this study, parameters of machine learning models were expressed by a chromosome including ten genes in a binary-coded forms, and the population size is twenty. A single point crossover was conducted and the crossover and mutation rates were 0.6 and 0.6, correspondingly. In the parameter tuning procedure, tentative models with different parameters were iteratively generated by genetic algorithms. Parameters generated by genetic algorithms for four machine learning models are specified as follows. Two parameters, penalty factors and widths of Gaussian functions, were selected for the least squares support vector regression models. For the classification and regression tree models, the maximum depth of the tree, the minimum numbers of leaf nodes, and the numbers of predictors were adjusted. The smoothing parameter was determined for general regression neural networks. Learning rates and momentums were altered for backpropagation neural networks.

5. Numerical Results

Two indices, namely the average absolute percentage error (MAPE) and the normalized mean absolute error (NMAE), were employed to evaluate performances of machine learning models in real estate prices forecasting. The mathematical forms of MAPE and NMAE are expressed as follows.
MAPE ( % ) = 100 N t = 1 N | R t P t R t |
NMAE = 1 R h R l [ 1 N t = 1 N | R t P t   | ]
where N is the amount of forecasting periods, At is the real value at period t, and Pt is the predicting value at period t, R h is the highest actual value, and R l is the lowest actual value. Table 3 and Table 4 show average MAPE and average NMAE values of machines learning models without and with attribute selection correspondingly. The computation results revealed that four machine learning models can result in better results with selected independent variables. According to Lewis [36], MAPE values less than 10 percent are highly accurate predictions; and values between 10 percent and 20 percent are good predictions. Thus, with attribute selection, LSSVR, GRNN and CART are highly accurate predictions and BPNN is a good forecasting in this study. Moreover, real estate prices can be expressed in different currencies or units. To avoid influences of real estate prices in various currencies or units, MAPE is specified when comparing the prediction accuracy of this study with forecasting results of previous studies. This study collected previous related studies which employed MAPE as measurements. Table 5 lists MAPE values of previous studies and the LSSVR models in this study for real estate prices forecasting. Table 5 revealed that the LSSVR models are superior to the other forecasting models of previous studies in terms of MAPE. However, the performance differences may come from many variances of real estate such as countries, cultures, market trends, and economic conditions. Furthermore, buyers’ and owners’ anticipations are continually varying owning to changes in lifestyles, materials of buildings, environmental legislation, regulations on the rational use of energy [5]. Thus, this study provides a feasible and comparative alternative in forecasting real estate. Forecasting models should be kept adjusted and improved to maintain stability and feasibility over different time periods. In this study, the number of residential properties is 31,397 and the proportion of residential properties is about 97.46 percent of the total data. The other 2.54 percent is commercial real estate. Commercial real estate appraisal uses factors such as spatial autocorrelation which is different from residential real estate appraisal [37,38,39]. Therefore, the focus of this study is on the residential aspect.

6. Conclusions

The decision to purchase real estate is undeniably very essential in the life of most adults. Thus, the appraisal and prediction of real estate can provide useful information to help facilitate real estate transactions. Real estate prices vary due to a wide variety of attributes. Machine learning models, including least squares support vector regression, classification and regression tree, general regression neural networks, and backpropagation neural networks, were used in this investigation to forecast real estate transaction prices with actual transaction data in Taichung, Taiwan. Genetic algorithms were employed to determine parameters of machine learning models. Empirical results revealed that attribute selection for machine learning models in this study does improve performances of four forecasting models in forecasting accuracy. With attribute selection, three machine learning models offer highly accurate predictions, and one machine learning model presents good prediction. Thus, four machines model with features selection used in this study are appropriate for forecasting real estate prices. Furthermore, the least squares support vector regression outperformed the other three forecasting models and obtains more accurate results than some previous studies in terms of MAPE. Thus, the least squares support vector regression with genetic algorithms is a feasible and promising machine learning technique in forecasting real estate prices.
For future work, diverse data types such as comments of real estate attributes, prices from social media, images from Google maps, and economic indicators are possible sources added as inputs for machine learning models to improve forecasting accuracy. In this study, general regression neural networks and backpropagation neural networks seems not to generate satisfying results when compared with results provided by the least squares support vector regression and the classification and regression tree. Thus, another potential opportunity for future research might be the use of deep learning techniques to forecast real estate prices.

Author Contributions

Conceptualization, P.-F.P.; data curation; W.-C.W.; formal analysis, P.-F.P. and W.-C.W.; funding acquisition, P.-F.P.; methodology, P.-F.P. and W.-C.W.; software, W.-C.W.; visualization, P.-F.P.; writing-original draft, review and editing, P.-F.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Science and Technology, Taiwan under the Contract Number MOST 109-2410-H-260-023.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ahn, J.J.; Byun, H.W.; Oh, K.J.; Kim, T.Y. Using ridge regression with genetic algorithm to enhance real estate appraisal forecasting. Expert Syst. Appl. 2012, 39, 8369–8379. [Google Scholar] [CrossRef]
  2. Gruma, B.; Govekar, D.K. Influence of Macroeconomic Factors on Prices of Real Estate in Various Cultural Environments: Case of Slovenia, Greece, France, Poland and Norway. Procedia Econ. Financ. 2016, 39, 597–604. [Google Scholar] [CrossRef] [Green Version]
  3. Leamer, E.E. Housing is the Business Cycle. NBER Working Paper No. 13428. 2007. Available online: http://www.nber.org/papers/w13428 (accessed on 9 August 2020).
  4. Beimer, W.; Maennig, W. Noise effects and real estate prices: A simultaneous analysis of different noise sources. Transp. Res. Part D 2017, 54, 282–286. [Google Scholar] [CrossRef]
  5. Ferlan, N.; Bastic, M.; Psunder, I. Influential Factors on the Market Value of Residential Properties. Inz. Ekon. Eng. Econ. 2017, 28, 135–144. [Google Scholar] [CrossRef]
  6. Singh, A.; Sharma, A.; Dubey, G. Big data analytics predicting real estate prices. Int. J. Syst. Assur. Eng. Manag. 2020. [Google Scholar] [CrossRef]
  7. Segnon, M.; Gupta, R.; Lesame, K.; Wohar, M.E. High-Frequency Volatility Forecasting of US Housing Markets. J. Real Estate Finance Econ. 2020. [Google Scholar] [CrossRef] [Green Version]
  8. Kang, H.; Lee, K.; Shin, D.H. Short-Term Forecast Model of Apartment Jeonse Prices Using Search Frequencies of News Article Keywords. Ksce J. Civ. Eng. 2019, 23, 4984–4991. [Google Scholar] [CrossRef]
  9. Giudice, V.D.; Paola, P.D.; Forte, F. Using Genetic Algorithms for Real Estate Appraisals. Buildings 2017, 7, 31. [Google Scholar] [CrossRef] [Green Version]
  10. Park, B.; Bae, P.J. Using machine learning algorithms for housing price prediction: The case of Fairfax County, Virginia housing data. Expert Syst. Appl. 2015, 42, 2928–2934. [Google Scholar] [CrossRef]
  11. Bork, L.; Møller, S.V. Forecasting house prices in the 50 states using Dynamic Model Averaging and Dynamic Model Selection. Int. J. Forecast. 2015, 31, 63–78. [Google Scholar] [CrossRef]
  12. Plakandaras, V.; Gupta, R.; Gogas, P.; Papadimitriou, T. Forecasting the U.S. real house price index. Econ. Model. 2015, 45, 259–267. [Google Scholar] [CrossRef] [Green Version]
  13. Chen, Z.-H.; Tsai, C.-T.; Yuan, S.-M.; Chou, S.-H.; Chern, J. Big data: Open data and realty website analysis. In Proceedings of the 8th International Conference on Ubi-Media Computing, Colombo, Sri Lanka, 24–26 August 2015; pp. 84–88. [Google Scholar]
  14. Lee, W.-T.; Chen, J.; Chen, K. Determination of Housing Price in Taipei City Using Fuzzy Adaptive Networks. In Proceedings of the International Multiconference of Engineers and Computer Scientists, Hong Kong, China, 13–15 March 2013. [Google Scholar]
  15. Antipov, E.A.; Pokryshevskaya, E.B. Mass appraisal of residential apartments: An application of Random forest for valuation and a CART-based approach for model diagnostics. Expert Syst. Appl. 2012, 39, 1772–1778. [Google Scholar] [CrossRef] [Green Version]
  16. Kontrimas, V.; Verikas, A. The mass appraisal of the real estate by computational intelligence. Appl. Soft. Comput. 2011, 11, 443–448. [Google Scholar] [CrossRef]
  17. Gupta, R.; Kabundi, A.; Miller, S.M. Forecasting the US real house price index: Structural and non-structural models with and without fundamentals. Econ. Model. 2011, 28, 2013–2021. [Google Scholar] [CrossRef] [Green Version]
  18. Kusan, H.; Aytekin, O.; Özdemir, I. The use of fuzzy logic in predicting house selling price. Expert Syst. Appl. 2010, 37, 1808–1813. [Google Scholar] [CrossRef]
  19. Cortes, C.; Vapnik, V. Support-vector networks. MLear 1995, 20, 273–297. [Google Scholar] [CrossRef]
  20. Vapnik, V. The Nature of Statistical Learning Theory; Springer: New York, NY, USA, 1995. [Google Scholar]
  21. Mukherjee, S.; Osuna, E.; Girosi, F. Nonlinear prediction of chaotic time series using support vector machines. In Proceedings of the IEEE Signal Processing Society Workshop, Amelia Island, FL, USA, 24–26 September 1997; pp. 511–520. [Google Scholar]
  22. Müller, K.-R.; Smola, A.J.; Rätsch, G.; Schölkopf, B.; Kohlmorgen, J.; Vapnik, V. Predicting time series with support vector machines. In Proceedings of the International Conference on Artificial Neural Networks, Munich, Germany, 17–19 September 2019; pp. 999–1004. [Google Scholar]
  23. Vapnik, V.; Golowich, S.E.; Smola, A.J. Support vector method for function approximation, regression estimation and signal processing. In Proceedings of the Advances Neural Information Processing System, Denver, CO, USA, 2–6 December 1997; pp. 281–287. [Google Scholar]
  24. Suykens, J.A.; Vandewalle, J. Least squares support vector machine classifiers. Neural Process. Lett. 1999, 9, 293–300. [Google Scholar] [CrossRef]
  25. Fletcher, R. Practical Methods of Optimization; Wiley: Hoboken, NJ, USA, 1987; pp. 80–94. [Google Scholar]
  26. Karush, W. Minima of Functions of Several Variables with Inequalities as Side Conditions. Master’s Thesis, University of Chicago, Chicago, IL, USA, 1939. [Google Scholar]
  27. Kuhn, H.W.; Tucker, A.W. Nonlinear programming. In Proceedings of the 2nd Berkeley Symposium on Mathematical Statistics and Probabilities, Berkeley, CA, USA, 31 July–12 August 1951; pp. 481–492. [Google Scholar]
  28. Mercer, J. Functions of Positive and Negative Type and Their Connection with the Theory of Integral Equations. Philos. Trans. R. Soc. A Math. Phys. Eng. Sci. 1909, 209, 415–446. [Google Scholar]
  29. Breiman, L.; Friedman, J.; Olshen, R.; Stone, C. Classification and Regression Trees; Chapman and Hall, Wadsworth: New York, NY, USA, 1984. [Google Scholar]
  30. Liu, Y.Y.; Yang, M.; Ramsay, M.; Li, X.S.; Coid, J.W. A Comparison of Logistic Regression, Classification and Regression Tree, and Neural Networks Models in Predicting Violent Re-Offending. J. Quant. Criminol. 2011, 27, 547–573. [Google Scholar] [CrossRef]
  31. Parzen, E. On Estimation of a Probability Density Function and Mode. Ann. Math. Statist. 1962, 33, 1065–1076. [Google Scholar] [CrossRef]
  32. Specht, D.F. A general regression neural network. IEEE Trans. Neural Netw. 1991, 2, 568–576. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  34. Lin, C.T.; Lee, C.G. Neural Fuzzy Systems: A Neuro-Fuzzy Synergism to Intelligent Systems; Prentice Hall: Upper Saddle River, NJ, USA, 1996. [Google Scholar]
  35. Holland, J. Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence; University of Michigan Press: Ann Arbor, MI, USA, 1975; pp. 439–444. [Google Scholar]
  36. Lewis, C.D. Industrial and Business Forecasting Methods; Butterworth Scientific: London, UK, 1982. [Google Scholar]
  37. Zhang, R.; Du, Q.; Geng, J.; Liu, B.; Huang, Y. An improved spatial error model for the mass appraisal of commercial real estate based on spatial analysis: Shenzhen as a case study. Habitat Int. 2015, 46, 196–205. [Google Scholar] [CrossRef]
  38. Kato, T. Prediction in the lognormal regression model with spatial error dependence. J. Hous. Econ. 2012, 21, 66–76. [Google Scholar] [CrossRef]
  39. Seya, H.; Yamagata, Y.; Tsutsumi, M. Automatic selection of a spatial weight matrix in spatial econometrics: Application to a spatial hedonic approach. Reg. Sci. Urban. Econ. 2013, 43, 429–444. [Google Scholar] [CrossRef]
Figure 1. The proposed real estate appraising framework for real estate appraisal.
Figure 1. The proposed real estate appraising framework for real estate appraisal.
Applsci 10 05832 g001
Table 1. Depictions of variables.
Table 1. Depictions of variables.
Types of VariablesCodes of VariablesDescriptions VariablesData Types
Independent VariablesX1City or TownshipCategorical
X2With or without parking spaceCategorical
X3LongitudeNumerical
X4LatitudeNumerical
X5Transaction area of landNumerical
X6Purpose of land useCategorical
X7Ages of buildingsNumerical
X8Transaction amount of propertyNumerical
X9Transaction floorsNumerical
X10Total floors of buildingsNumerical
X11Types of buildingsCategorical
X12Use of buildingsCategorical
X13Materials of buildingsCategorical
X14Total transaction areas of buildingsNumerical
X15Number of bedroomsNumerical
X16Number of living roomsNumerical
X17Number of bathroomsNumerical
X18With or without compartmentsCategorical
X19With or without management committeeCategorical
X20Prices per square meterNumerical
X21Types of parking spaceCategorical
X22Area of parking spaceNumerical
X23Prices of parking spaceNumerical
The Dependent VariableYTransaction pricesNumerical
Table 2. Pearson correlation coefficient of independent variables to the real estate price.
Table 2. Pearson correlation coefficient of independent variables to the real estate price.
Independent Variables (Pearson Correlation Coefficient) X1(0.01158), X2(0.4097), X3(−0.043), X4(−0.0184), X5(0.3518),
X6(−0.0319), X7(−0.2897), X8(0.00465), X9(0.03199), X10(0.09087), X11(−0.0271), X12(0.0884), X13(−0.0382), X14(0.7923), X15(0.4766), X16(0.4391), X17(0.4024), X18(0.05456), X19(−0.0201), X20(0.5299), X21(0.3871), X22(0.3686), X23(0.1932)
Table 3. Average MAPE and average NMAE values of machines learning models without attribute selection.
Table 3. Average MAPE and average NMAE values of machines learning models without attribute selection.
ModelsLSSVRCARTGRNNBPNN
MAPE (%)1.6762.294422.893615.0357
NMAE4.13 × 10−36.86 × 10−31.07 × 1063.82 × 10−2
Table 4. Average MAPE and average NMAE values of machines learning models with attribute selection.
Table 4. Average MAPE and average NMAE values of machines learning models with attribute selection.
ModelsLSSVRCARTGRNNBPNN
MAPE (%)0.2282.2788.73814.424
NMAE8.11 × 10−46.76 × 10−34.52 × 1054.13 × 10−2
Table 5. A list of MAPE values of previous studies and this study.
Table 5. A list of MAPE values of previous studies and this study.
Forecasting Models MAPE (%)
Kang et al. [8]3.84
Giudice et al. [9]10.62
Plakandaras et al. [12]2.151
Lee et al. [14]4.54
Antipov and Pokryshevskaya [15]13.95
Kusan et al. [18]3.65
*LSSVR 1 1.676
**LSSVR 2 0.228
*LSSVR1: LSSVR models without attribute selection; **LSSVR 2: LSSVR models with attribute selection.

Share and Cite

MDPI and ACS Style

Pai, P.-F.; Wang, W.-C. Using Machine Learning Models and Actual Transaction Data for Predicting Real Estate Prices. Appl. Sci. 2020, 10, 5832. https://doi.org/10.3390/app10175832

AMA Style

Pai P-F, Wang W-C. Using Machine Learning Models and Actual Transaction Data for Predicting Real Estate Prices. Applied Sciences. 2020; 10(17):5832. https://doi.org/10.3390/app10175832

Chicago/Turabian Style

Pai, Ping-Feng, and Wen-Chang Wang. 2020. "Using Machine Learning Models and Actual Transaction Data for Predicting Real Estate Prices" Applied Sciences 10, no. 17: 5832. https://doi.org/10.3390/app10175832

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop