Traffic Flow Prediction With Big Data - A Deep Learning Approach

1y ago
10 Views
2 Downloads
1.48 MB
9 Pages
Last View : 14d ago
Last Download : 3m ago
Upload by : Karl Gosselin
Transcription

IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. 16, NO. 2, APRIL 2015865Traffic Flow Prediction With Big Data:A Deep Learning ApproachYisheng Lv, Yanjie Duan, Wenwen Kang, Zhengxi Li, and Fei-Yue Wang, Fellow, IEEEAbstract—Accurate and timely traffic flow information is important for the successful deployment of intelligent transportationsystems. Over the last few years, traffic data have been exploding,and we have truly entered the era of big data for transportation.Existing traffic flow prediction methods mainly use shallow trafficprediction models and are still unsatisfying for many real-worldapplications. This situation inspires us to rethink the traffic flowprediction problem based on deep architecture models with bigtraffic data. In this paper, a novel deep-learning-based traffic flowprediction method is proposed, which considers the spatial andtemporal correlations inherently. A stacked autoencoder modelis used to learn generic traffic flow features, and it is trained ina greedy layerwise fashion. To the best of our knowledge, this isthe first time that a deep architecture model is applied using autoencoders as building blocks to represent traffic flow features forprediction. Moreover, experiments demonstrate that the proposedmethod for traffic flow prediction has superior performance.Index Terms—Deep learning, stacked autoencoders (SAEs),traffic flow prediction.I. I NTRODUCTIONACCURATE and timely traffic flow information is currently strongly needed for individual travelers, businesssectors, and government agencies [1]. It has the potential tohelp road users make better travel decisions, alleviate trafficcongestion, reduce carbon emissions, and improve traffic operation efficiency. The objective of traffic flow prediction is toprovide such traffic flow information. Traffic flow predictionhas gained more and more attention with the rapid developmentand deployment of intelligent transportation systems (ITSs). Itis regarded as a critical element for the successful deploymentof ITS subsystems, particularly advanced traveler informationsystems, advanced traffic management systems, advanced public transportation systems, and commercial vehicle operations.Manuscript received May 24, 2014; revised July 10, 2014; accepted July 20,2014. Date of publication September 9, 2014; date of current version March27, 2015. This work was supported in part by the National Natural ScienceFoundation of China under Grants 61233001, 61203166, 71232006, 61104054,and 61273326. The Associate Editor for this paper was J. Zhang.Y. Lv, Y. Duan, W. Kang, and F.-Y. Wang are with State Key Laboratoryof Management and Control for Complex Systems, Institute of Automation,Chinese Academy of Sciences, Beijing 100190, China (e-mail: yisheng.lv@ia.ac.cn; duanyanjie2012@ia.ac.cn; kangwenwen2012@ia.ac.cn; feiyue@ieee.org).Z. Li is with North China University of Technology, Beijing 100144, China(e-mail: lzx@ncut.edu.cn).Color versions of one or more of the figures in this paper are available onlineat http://ieeexplore.ieee.org.Digital Object Identifier 10.1109/TITS.2014.2345663Traffic flow prediction heavily depends on historical andreal-time traffic data collected from various sensor sources,including inductive loops, radars, cameras, mobile Global Positioning System, crowd sourcing, social media, etc. With thewidespread traditional traffic sensors and new emerging trafficsensor technologies, traffic data are exploding, and we haveentered the era of big data transportation. Transportation management and control is now becoming more data driven [2], [3].Although there have been already many traffic flow predictionsystems and models, most of them use shallow traffic modelsand are still somewhat unsatisfying. This inspires us to rethinkthe traffic flow prediction problem based on deep architecturemodels with such rich amount of traffic data.Recently, deep learning, which is a type of machine learningmethod, has drawn a lot of academic and industrial interest[4]. It has been applied with success in classification tasks,natural language processing, dimensionality reduction, objectdetection, motion modeling, and so on [5]–[9]. Deep learningalgorithms use multiple-layer architectures or deep architectures to extract inherent features in data from the lowest levelto the highest level, and they can discover huge amounts ofstructure in the data. As a traffic flow process is complicated innature, deep learning algorithms can represent traffic featureswithout prior knowledge, which has good performance fortraffic flow prediction.In this paper, we propose a deep-learning-based traffic flowprediction method. Herein, a stacked autoencoder (SAE) modelis used to learn generic traffic flow features, and it is trainedin a layerwise greedy fashion. To the best of the authors’knowledge, it is the first time that the SAE approach is usedto represent traffic flow features for prediction. The spatial andtemporal correlations are inherently considered in the modeling. In addition, it demonstrates that the proposed method fortraffic flow prediction has superior performance.The rest of this paper is organized as follows. Section IIreviews the studies on short-term traffic flow prediction.Section III presents the deep learning approach with autoencoders as building blocks for traffic flow prediction. Section IVdiscusses the experimental results. Concluding remarks aredescribed in Section V.II. L ITERATURE R EVIEWTraffic flow prediction has been long regarded as a keyfunctional component in ITSs. Over the past few decades, anumber of traffic flow prediction models have been developedto assist in traffic management and control for improving transportation efficiency ranging from route guidance and vehicle1524-9050 2014 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.See http://www.ieee.org/publications standards/publications/rights/index.html for more information.

866IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. 16, NO. 2, APRIL 2015routing to signal coordination. The evolution of traffic flow canbe considered a temporal and spatial process. The traffic flowprediction problem can be stated as follows. Let Xit denote theobserved traffic flow quantity during the tth time interval at theith observation location in a transportation network. Given asequence {Xit } of observed traffic flow data, i 1, 2, . . . , m,t 1, 2, . . . , T , the problem is to predict the traffic flow at timeinterval (t Δ) for some prediction horizon Δ.As early as 1970s, the autoregressive integrated moving average (ARIMA) model was used to predict short-term freewaytraffic flow [10]. Since then, an extensive variety of modelsfor traffic flow prediction have been proposed by researchersfrom different areas, such as transportation engineering, statistics, machine learning, control engineering, and economics.Previous prediction approaches can be grouped into three categories, i.e., parametric techniques, nonparametric methods,and simulations. Parametric models include time-series models,Kalman filtering models, etc. Nonparametric models includek-nearest neighbor (k-NN) methods, artificial neural networks(ANNs), etc. Simulation approaches use traffic simulation toolsto predict traffic flow.A widely used technique to the problem of traffic flow prediction is based on time-series methods. Levin and Tsao appliedBox–Jenkins time-series analyses to predict expressway trafficflow and found that the ARIMA (0, 1, 1) model was the moststatistically significant for all forecasting [11]. Hamed et al.applied an ARIMA model for traffic volume prediction inurban arterial roads [12]. Many variants of ARIMA wereproposed to improve prediction accuracy, such as KohonenARIMA (KARIMA) [13], subset ARIMA [14], ARIMA withexplanatory variables (ARIMAX) [15], vector autoregressivemoving average (ARMA) and space–time ARIMA [16], andseasonal ARIMA (SARIMA) [17]. Except for the ARIMA-liketime-series models, other types of time-series models were alsoused for traffic flow prediction [18].Due to the stochastic and nonlinear nature of traffic flow,researchers have paid much attention to nonparametric methodsin the traffic flow forecasting field. Davis and Nihan used thek-NN method for short-term freeway traffic forecasting andargued that the k-NN method performed comparably with butnot better than the linear time-series approach [19]. Chang et al.presented a dynamic multiinterval traffic volume predictionmodel based on the k-NN nonparametric regression [20].El Faouzi developed a kernel smoother for the autoregressionfunction to do short-term traffic flow prediction, in which functional estimation techniques were applied [21]. Sun et al. useda local linear regression model for short-term traffic forecasting[22]. A Bayesian network approach was proposed for trafficflow forecasting in [23]. An online learning weighted supportvector regression (SVR) was presented in [24] for short-termtraffic flow predictions. Various ANN models were developedfor predicting traffic flow [25]–[34].To obtain adaptive models, some works explore hybrid methods, in which they combine several techniques. Tan et al.proposed an aggregation approach for traffic flow predictionbased on the moving average (MA), exponential smoothing(ES), ARIMA, and neural network (NN) models. The MA, ES,and ARIMA models were used to obtain three relevant timeseries that were the basis of the NN in the aggregation stage[35]. Zargari et al. developed different linear genetic programming, multilayer perceptron, and fuzzy logic (FL) models forestimating 5-min and 30-min traffic flow rates [36]. Cetin andComert combined the ARIMA model with the expectation—maximization and cumulative sum algorithms [37]. An adaptivehybrid fuzzy rule-based system approach was proposed formodeling and predicting urban traffic flow [38].In addition to the methods aforementioned, the Kalmanfiltering method [39], [40], stochastic differential equations[41], the online change-point-based model [42], the type-2FL approach [43], the variational infinite-mixture model [44],simulations [45], and dynamic traffic assignment [46], [47]were also applied in predicting short-term traffic flow.Comparison studies of traffic flow prediction models havebeen reported in literature. The linear regression, the historical average, the ARIMA, and the SARIMA were assessed in[48], in which it was concluded that these algorithms performreasonably well during normal operating conditions but do notrespond well to external system changes. The SARIMA modelsand the nonparametric regression forecasting methods wereevaluated in [49]. It was found that the proposed heuristic forecast generation methods improved the performance of nonparametric regression, but they did not equal the performance of theSARIMA models. The multivariate state-space models andthe ARIMA models were compared in [50], and it showedthat the performance of the multivariate state-space modelsis better than that of the ARIMA models. Stathopoulos andKarlaftis [50] also pointed out that different model specifications are appropriate for different time periods of the day.Lippi et al. [51] compared SVR models and SARIMA models,and they concluded that the proposed seasonal support vectorregressor is highly competitive when performing forecasts during the most congested periods. Chen et al. [52] reported theperformance results for the ARMA, ARIMA, SARIMA, SVR,Bayesian network, ANN, k-NN, Naïve I, and Naïve II modelsat different aggregation time scales, which were set at 3, 5, 10,and 15 min, respectively. A series of research is dedicated tothe comparison of NNs and other techniques such as the historical average, the ARIMA models, and the SARIMA models[53]–[55]. Interestingly, it could be found that nonparametrictechniques obviously outperform simple statistical techniquessuch as the historical average and smoothing techniques, butthere are contradicting results on whether nonparametric methods can yield better or comparable results compared with theadvanced forms of statistical approaches such as the SARIMA.Detailed reviews on the short-term traffic flow forecast can befound in [56] and [57].In summary, a large number of traffic flow prediction algorithms have been developed due to the growing need for realtime traffic flow information in ITSs, and they involve varioustechniques in different disciplines. However, it is difficult tosay that one method is clearly superior over other methods inany situation. One reason for this is that the proposed modelsare developed with a small amount of separate specific trafficdata, and the accuracy of traffic flow prediction methods isdependent on the traffic flow features embedded in the collectedspatiotemporal traffic data. Moreover, in general, literature

LV et al.: TRAFFIC FLOW PREDICTION WITH BIG DATA: DEEP LEARNING APPROACHFig. 1.867Autoencoder.shows promising results when using NNs, which have goodprediction power and robustness.Although the deep architecture of NNs can learn morepowerful models than shallow networks, existing NN-basedmethods for traffic flow prediction usually only have one hiddenlayer. It is hard to train a deep-layered hierarchical NN witha gradient-based training algorithm. Recent advances in deeplearning have made training the deep architecture feasible sincethe breakthrough of Hinton et al. [58], and these show that deeplearning models have superior or comparable performance withstate-of-the-art methods in some areas. In this paper, we explorea deep learning approach with SAEs for traffic flow prediction.Fig. 2. Layerwise training of SAEs.restrictions such as sparsity constraints are imposed, this is nota problem [60]. When sparsity constraints are added to theobjective function, an autoencoder becomes a sparse autoencoder, which considers the sparse representation of the hiddenlayer. To achieve the sparse representation, we will minimizethe reconstruction error with a sparsity constraint asIII. M ETHODOLOGYHere, a SAE model is introduced. The SAE model is a stackof autoencoders, which is a famous deep learning model. It usesautoencoders as building blocks to create a deep network [59].A. AutoencoderAn autoencoder is an NN that attempts to reproduce its input,i.e., the target output is the input of the model. Fig. 1 gives anillustration of an autoencoder, which has one input layer, onehidden layer, and one output layer. Given a set of training samples {x(1) , x(2) , x(3) , . . .}, where x(i) Rd , an autoencoderfirst encodes an input x(i) to a hidden representation y(x(i) )based on (1), and then it decodes representation y(x(i) ) backinto a reconstruction z(x(i) ) computed as in (2), as shown iny(x) f (W1 x b)z(x) g (W2 y(x) c)(1)(2)where W1 is a weight matrix, b is an encoding bias vector, W2 isa decoding matrix, and c is a decoding bias vector; we considerlogistic sigmoid function 1/(1 exp( x)) for f (x) and g(x)in this paper.By minimizing reconstruction error L(X, Z), we can obtainthe model parameters, which are here denoted as θ, asθ arg min L(X, Z) arg minθθN 21 (i) x z x(i) . (3)2 i 1One serious issue concerned with an autoencoder is that ifthe size of the hidden layer is the same as or larger than theinput layer, this approach could potentially learn the identityfunction. However, current practice shows that if nonlinearautoencoders have more hidden units than the input or if otherSAO L(X, Z) γHD KL(ρ ρ̂j )(4)j 1where γ is the weight of the sparsity term, HD is the number ofhidden units, ρ is a sparsity parameter and is(i)typically a smallvalue close to zero, ρ̂j (1/N ) Ni 1 yj (x ) is the averageactivation of hidden unit j over the training set, and KL(ρ ρ̂j )is the Kullback–Leibler (KL) divergence, which is defined asKL(ρ ρ̂j ) ρ logρ1 ρ (1 ρ) log.ρ̂j1 ρ̂jThe KL divergence has the property that KL(ρ ρ̂j ) 0if ρ ρ̂j . It provides the sparsity constraint on the coding.The backpropagation (BP) algorithm can be used to solve thisoptimization problem.B. SAEsA SAE model is created by stacking autoencoders to form adeep network by taking the output of the autoencoder foundon the layer below as the input of the current layer [59].More clearly, considering SAEs with l layers, the first layer istrained as an autoencoder, with the training set as inputs. Afterobtaining the first hidden layer, the output of the kth hiddenlayer is used as the input of the (k 1)th hidden layer. In thisway, multiple autoencoders can be stacked hierarchically. Thisis illustrated in Fig. 2.To use the SAE network for traffic flow prediction, we needto add a standard predictor on the top layer. In this paper, we puta logistic regression layer on top of the network for supervisedtraffic flow prediction. The SAEs plus the predictor comprisethe whole deep architecture model for traffic flow prediction.This is illustrated in Fig. 3.

868IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. 16, NO. 2, APRIL 2015— Use the output of the kth hidden layer as theinput of the (k 1)th hidden layer. For the firsthidden layer, the input is the training set.l 1}k 0— Find encoding parameters {W1k 1 , bk 11for the (k 1)th hidden layer by minimizing theobjective function.Step 2) Fine-tuning the whole network— Initialize {W1l 1 , bl 11 } randomly or by supervisedtraining.— Use the BP method with the gradient-based optimization technique to change the whole network’sparameters in a top–down fashion.Fig. 3. Deep architecture model for traffic flow prediction. A SAE model isused to extract traffic flow features, and a logistic regression layer is applied forprediction.C. Training AlgorithmIt is straightforward to train the deep network by applyingthe BP method with the gradient-based optimization technique.Unfortunately, it is known that deep networks trained in thisway have bad performance. Recently, Hinton et al. have developed a greedy layerwise unsupervised learning algorithm thatcan train deep networks successfully. The key point to using thegreedy layerwise unsupervised learning algorithm is to pretrainthe deep network layer by layer in a bottom–up way. After thepretraining phase, fine-tuning using BP can be applied to tunethe model’s parameters in a top–down direction to obtain betterresults at the same time. The training procedure is based on theworks in [58] and [59], which can be stated as follows.1) Train the first layer as an autoencoder by minimizing theobjective function with the training sets as the input.2) Train the second layer as an autoencoder taking the firstlayer’s output as the input.3) Iterate as in 2) for the desired number of layers.4) Use the output of the last layer as the input for theprediction layer, and initialize its parameters randomly orby supervised training.5) Fine-tune the parameters of all layers with the BP methodin a supervised way.IV. E XPERIMENTSA. Data DescriptionThe proposed deep architecture model was applied to thedata collected from the Caltrans Performance MeasurementSystem (PeMS) database as a numerical example. The trafficdata are collected every 30 s from over 15 000 individualdetectors, which are deployed statewide in freeway systemsacross California [61]. The collected data are aggregated 5-mininterval each for each detector station. In this paper, the trafficflow data collected in the weekdays of the first three monthsof the year 2013 were used for the experiments. The data ofthe first two months were selected as the training set, and theremaining one month’s data were selected as the testing set.For freeways with multiple detectors, the traffic data collectedby different detectors are aggregated to get the average trafficflow of this freeway. Note that we separately treat two directionsof the same freeway among all the freeways, in which three areone-way. Fig. 4 is a plot of a typical freeway’s traffic flow overtime for weekdays of some week.B. Index of PerformanceTo evaluate the effectiveness of the proposed model, we usethree performance indexes, which are the mean absolute error(MAE), the mean relative error (MRE), and the RMS error(RMSE). They are defined asMAE 1 fi fˆi n i 1MRE n1 fi fˆi n i 1fiThis procedure is summarized in Algorithm 1.nAlgorithm 1. Training SAEsGiven training samples X and the desired number of hiddenlayers l,Step 1) Pretrain the SAE— Set the weight of sparsity γ, sparsity parameter ρ,initialize weight matrices and bias vectors randomly.— Greedy layerwise training hidden layers. RMSE 21 fi fˆi n i 1n 12where fi is the observed traffic flow, and fˆi is the predictedtraffic flow.

LV et al.: TRAFFIC FLOW PREDICTION WITH BIG DATA: DEEP LEARNING APPROACHFig. 4.869Typical daily traffic flow pattern. (a) Monday. (b) Tuesday. (c) Wednesday. (d) Thursday. (e) Friday.C. Determination of the Structure of a SAE ModelWith regard to the structure of a SAE network, we need todetermine the size of the input layer, the number of hiddenlayers, and the number of hidden units in each hidden layer.For the input layer, we use the data collected from all freewaysas the input; thus, the model can be built from the perspective ofa transportation network considering the spatial correlations oftraffic flow. Furthermore, considering the temporal relationshipof traffic flow, to predict the traffic flow at time interval t, weshould use the traffic flow data at previous time intervals, i.e.,X t 1 , X t 2 , . . . , X t r. Therefore, the proposed model accountsfor the spatial and temporal correlations of traffic flow inherently.The dimension of the input space is mr, whereas the dimensionof the output is m, where m is the number of freeways.In this paper, we used the proposed model to predict 15-mintraffic flow, 30-min traffic flow, 45-min traffic flow, and 60-mintraffic flow. We choose r from 1 to 12, the hidden layer sizefrom 1 to 6, and the number of hidden units from {100, 200,300, 400, 500, 600, 700, 800, 900, 1000}. After performinggrid search runs, we obtained the best architecture for differentprediction tasks, which is shown in Table I. For the 15-mintraffic flow prediction, our best architecture consists of threehidden layers, and the number of hidden units in each hiddenlayer is [400, 400, 400], respectively. For the 30-min trafficflow prediction, our best architecture consists of three hiddenlayers, and the number of hidden units in each hidden layeris [200, 200, 200], respectively. For the 45-min traffic flowprediction, our best architecture consists of two hidden layers,

870IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. 16, NO. 2, APRIL 2015TABLE IS TRUCTURE OF SAE S FOR T RAFFIC F LOW P REDICTIONFig. 5. Traffic flow prediction of roads with different traffic volume. (a) Roadwith heavy traffic flow. (b) Road with medium traffic flow. (c) Road with lowtraffic flow.and the number of hidden units in each hidden layer is [500,500], respectively. For the 60-min traffic flow prediction, ourbest architecture consists of four hidden layers, and the numberof hidden units in each hidden layer is [300, 300, 300, 300],respectively. From the results, we can see that the best numberof hidden layers is at least two and no more than five for ourexperiments. Lessons learned from experience indicate that thenumber of hidden layers of an NN should be neither too smallnor too large. Our results confirmed these lessons.D. ResultsFig. 5 presents the output of the proposed model for thetraffic flow prediction of typical roads with heavy, medium,and low traffic loads. The observed traffic flow is also includedin Fig. 5 for comparison. In Fig. 5, it is shown that the predicted traffic flow has similar traffic patterns with the observedtraffic flow. In addition, it matches well in heavy and mediumtraffic flow conditions. However, the proposed model does notperform well in low traffic flow conditions, which is the sameas existing traffic flow prediction methods. The reason for thisphenomenon is that small differences between the observedflow and the predicted flow can cause a bigger relative errorwhen the traffic flow rate is small. In fact, we are more focusedon the traffic flow prediction results under heavy and mediumtraffic flow conditions; hence, the proposed method is effectiveand promising for traffic flow prediction in practice.We compared the performance of the proposed SAE modelwith the BP NN, the random walk (RW) forecast method,the support vector machine (SVM) method, and the radialbasis function (RBF) NN model. Among these four competingmethods, the RW method is a simple baseline that predictstraffic in the future as equal to the current traffic flow (X t 1 X t ), the NN methods have good performance for the trafficflow forecast, as aforementioned in Section II, and the SVMmethod is a relatively advanced model for prediction. In allcases, we used the same data set. The prediction results on thetest data sets for freeways with the average 15-min traffic flowrate larger than 450 vehicles are given in Table II. In Table II,we can see that the average accuracy (1-MRE) of the SAE isover 93% for all the four tasks and has low MAE values. Thisprediction accuracy is promising, robust, and comparable withthe reported results. Notice that we only use the traffic volumedata as the input for prediction without considering handengineering factors, such as weather conditions, accidents, andother traffic flow parameters (density and speed), that have arelationship with the traffic volume.In Table II, we can also find that the SAE proved to bemore accurate than the BP NN model, the RW, the SVM, andthe RBF NN model for the short-term prediction of the trafficvolume. For the BP NN, the prediction performance is relativelystationary, which is from 88% to 90% or so. For the RW,the SVM, and the RBF, the average prediction accuracy dropsmuch with the aggregate time interval of the traffic flow dataincreasing. For the 15-min traffic flow prediction, the averageaccuracy of the RW, the SVM, and the RBF is 7.8%, 8.0%,and 7.4%, respectively. However, for the 60-min traffic flowprediction, the average accuracy of the RW, the SVM, and theRBF has a large drop to 22.3%, 22.1%, and 26.4%, respectively.The maximum average prediction accuracy improvement ofthe SAE is up to 4.8% compared with the BP NN, over 16%compared with the RW, over 15% compared with the SVM, andover 20% compared with the RBF.A visual display of the performance of the MRE derived withthe SAE, the BP NN model, the RW, the SVM, and the RBF NNmodel is given in Fig. 6. It displays for each method the cumulative distribution function (cdf) of the MRE, which describes thestatistical results on freeways with the average 15-min trafficflow larger than 450 vehicles. The method that uses SAEsleads to improved traffic flow prediction performance whencompared with the BP NN, the RW, the SVM, and the RBFNN model. For the 15-min traffic flow prediction, over 86% offreeways with the average 15-min traffic flow larger than 450vehicles have an accuracy of more than 90%. For the 30-mintraffic flow prediction, over 88% of freeways with the average15-min traffic flow larger than 450 vehicles have an accuracy ofmore than 90%. For the 45-min traffic flow prediction and for

LV et al.: TRAFFIC FLOW PREDICTION WITH BIG DATA: DEEP LEARNING APPROACH871TABLE IIP ERFORMANCE C OMPARISON OF THE MAE, THE MRE, AND THE RMSE FOR SAE S , THE BP NN, THE RW, THE SVM, AND THE RBF NNFig. 6. Empirical cdf of the MRE for freeways with the average 15-min traffic flow larger than 450 vehicles. (a) 15-min traffic flow prediction. (b) 30-min trafficflow prediction. (c) 45-min traffic flow prediction. (d) 60-min traffic flow prediction.the 60-min traffic flow prediction, over 90% of freeways withthe average 15-min traffic flow larger than 450 vehicles have anaccuracy of more than 90%. Thus, the effectiveness of the SAEmethod for traffic flow prediction is promising and manifested.V. C ONCLUSIONWe propose a deep learning approach with a SAE modelfor traffic flow prediction. Unlike the previous methods thatonly consider the shallow structure of traffic data, the proposedmethod can successfully discover the latent traffic flow featurerepresentation, such as the nonlinear spatial and temporal correlations from the traffic data. We applied the greedy layerwiseunsupervised learning algorithm to pretrain the deep network,and then, we did the fine-tuning process to update the model’sparameters to improve the prediction performance. We evaluated the performance of the proposed method on a PeMS dataset and compared it with the BP NN, the RW, the SVM, and theRBF NN model, and the results show that the proposed methodis superior to the competing methods.

872IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. 16, NO. 2, APRIL 2015For future work, it would be interesting to investigate otherdeep learning algorithms for traffic flow prediction and to applythese algorithms on different public open traffic data sets toexamine their effectiveness. Furthermore, the prediction layerin our paper has been just a logistic regression. Extending itto more powerful predictors may make further performanceimprovement.ACKNOWLEDGMENTThe authors would like to thank the anonymous referees fortheir invaluable insights.R EFERENCES[1] N. Zhang, F.-Y. Wang, F. Zhu, D. Zhao, and S. Tang, “DynaCAS: Computational experiments and decision support for ITS,” IEEE Intell. Syst.,vol. 23, no. 6, pp. 19–23, Nov./Dec. 2008.[2] J. Zhang et al., “Data-driven intelligent transportation systems: A survey,”IEEE Trans. Intell. Transp. Syst., vol. 12, no. 4, pp. 1624–1639, Dec. 2011.[3] C. L. Philip Chen and C.-Y. Zhang, “Data-intensive applications, challenges, techniques and technologies: A survey on Big Data,” Inf. Sci.,vol. 275, pp. 314–347, Aug. 2014.[4] Y. Bengio, “Learning deep architectures for AI,” Found. Trends Mach.Learn., vol. 2, no. 1, pp. 1–127, Jan. 20

nature, deep learning algorithms can represent traffic features without prior knowledge, which has good performance for traffic flow prediction. In this paper, we propose a deep-learning-based traffic flow del is used to learn generic traffic flow features, and it is trained

Related Documents:

SA Learner Driver Manual Road Traffic Signs Version: Draft Page 1 of 56 2. ROAD TRAFFIC SIGNS, SIGNALS AND MARKINGS The purpose of road traffic signs is to regulate traffic in such a way that traffic flow and road traffic safety are promoted. 1. SIGNS IN GENERAL Road traffic signs can be divided into the following six main groups:

Abstract: Traffic flow forecasting is a popular research topic of Intelligent Transportation Systems (ITS). With the development of information technology, lots of history electronic traffic flow data are collected. How to take full use of the history traffic flow data to improve the traffic flow forecasting precision is an important issue.

This product provides a local flow indication and automatically signals the operator or PLC if flow is too high or too low. Uses of the Flow-Alert flow meter include: bearing lubrication, case drain verification, gun drill cooling, pump flow confirmation, etc. FLOW-ALERT FLOW SWITCH 6000 PSI / 414 BARS MAX GPM LPM OIL FIGURE 1- Flow-AlERt Flow .

Intelligent application traffic management F5 Product Guide EMEA 4. BIG-IP Virtual Edition Local Traffic Manager (LTM) . BIG-IP Virtual Edition Local Traffic Manager High Performance (8 vCPUs, v13.x - v18.x) F5-BIG-LTM-VE-08-V18 BIG-IP Virtual Edition Local Traffic Manager 10 Gbps (v

CIV2282 Transport & Traffic Engineering (Part 1) . Traffic Flow Fundamentals Cannot efficiently describe traffic flow has each driver behaves differently hence we model traffic flow like fluid but still need quantitative techniques to assess operational measures of highway performance, so we can

the destination. The traffic light system designed by Salim Bin Islam provided a design and development of a microcontroller based intelligent traffic control system. He proposed a new intelligent traffic control system that is to control the traffic system through traffic signal on the basis of current traffic density.

Traffic light controller, Real-time traffic signaling, congestion, ZigBee communication board, Google Traffic API, Agent-based traffic modeling. ABSTRACT: Controlling of traffic signals optimally helps in avoiding traffic jams as vehicle volume density changes on temporally short and spatially small scales.

BAB 1 Akuntansi Keuangan & Standar Akuntansi Keuangan 1 BAB 2 Laporan Laba Rugi, Neraca dan Arus Kas 11 BAB 3 Pengawasan Terhadap Kas 25 BAB 4 P i u t a n g 33 BAB 5 Wesel dan Promes 47 BAB 6 Persediaan Barang Dagang 53 BAB 7 Penilaian Persediaan Berdasarkan Selain Harga Pokok 71 BAB 8 Amortisasi Aktiva Tak Berwujud 81 . Modul Akuntansi Keuangan 1 Dy Ilham Satria 1 1 AKUNTANSI KEUANGAN DAN .