Engineering, Technology & Applied Science Research Vol. 9, No. 2, 2019, 3881-3886 3881 www.etasr.com Belhaj Salah & Fourati: Systems Modeling Using Deep Elman Neural Network Systems Modeling Using Deep Elman Neural Network Latifa Belhaj Salah Control and Energy Management Laboratory (CEM-Lab) University of Gabes, Tunisia latifa.belhadjsalah@enis.tn Fathi Fourati Control and Energy Management Laboratory (CEM-Lab) University of Sfax, Tunisia fethi.fourati@ipeis.rnu.tn Abstractโ€”In this paper, the modeling of complex systems using deep Elman neural network architecture is improved. The emphasis is to retrieve better deep Elman structure that emulates perfectly such dynamic systems. To achieve this goal, sigmoid activation functions in the hidden and output layer nodes are chosen and data files on considered systems for modeling and validation steps are given. Simulation results prove the ability and the efficiency of a deep Elman neural network with two hidden layers in this task. Keywords-Elman neural network; recurrent neural network; deep learning; complex systems; modeling I. INTRODUCTION Recently, deep neural networks (including recurrent networks) have been successfully applied in several areas [1]: modeling and control of complex systems, fault detection, text understanding [2, 3], speech recognition [4, 5] and computer vision [6, 7]. The advantage of deep learning is reflected on the modeling of high-level abstractions from the data by its architecture which is consisted of several non-linear learning layers. Each layer replies to several levels of nonlinear abstraction. [8]. In [9], authors used recurrent neural networks with long-term memory units (LSTMs) to find a solution to the vanishing gradient problem observed in simple recurrent neural networks (RNNs) [9-10]. Authors in [11] used the pre-training and fine-tuning steps to ensure the effective training of deep learning. Deep learning has a significant success in big data since it can recover valuable information from complex processes [12]. In [13], authors have described a combination between deep learning and reinforcement learning for the prediction and control of intelligent laser welding. The effectiveness of using deep Elman RNN for modeling complex systems was demonstrated in [14]. In order to improve the performance of deep learning with multiplayer Perceptron (MLP), a new technique was used that requires a combination of adaptive learning rate and Laplacian score concept to vary the weights. In [15], authors proposed a new method for automatic modulation classification (AMC) based on unsupervised feed-forward deep learning. The performance of this approach was compared to conventional AMC techniques. In [16], authors used the application of end-to-end deep learning to solve the classification problem of speech emotion recognition. They showed the limitations and benefits of these architectures in speech recognition. In [17], authors used deep Elman RNN (ERNN) for acoustic modeling. They compared the performance of this technique with other RNN architectures like LSTM, GRU, and simplified LSTM [18-20]. In [21], authors developed two novel deep RNN models with LSTM units to predict building electricity consumption. In [22], authors proposed a new learning algorithm based on simplified convolutional neural network to achieve visual tracking with adaptive filtering of particles. In [23], authors used convolutional deep learning neural networks to detect and diagnose plant diseases from leaves images of healthy and diseased plants. In [24], authors exploited deep learning to solve the problem of large data analysis which is found in several areas. In [25], authors proposed the use of a new method based on deep belief networks and multiple models (DBNs-MMs) to detect faults of complex systems. The novelty in this paper is the use of a deep neural network that improves modeling systems tasks. In fact, comparing to results given in [26], an Elman neural network with two hidden layers gives more accuracy when modeling complex systems. II. ELMAN NEURAL NETWORK The Elman network is a type of recurrent network [27]. It has been applied in many areas such as dynamic system identification [28] and financial prediction [29]. In [27], a simple recurrent neural network was proposed: the input and output units are in contact with the external environment, in contrast to the context and hidden units. This network is characterized by context units that are used to save the previous hidden unit activations. Hidden unit activations at time k are returned to the context units and stored for the next training step. The Elman network is named partially recurrent network because the feed forward connections are modifiable and the recurrent connections are fixed. To train this kind of neural network, the back propagation algorithm can then be used and the hidden units activation functions can be linear or non-linear [17, 27, 30]. Figure 1 represents an Elman neural network architecture. The l-th input unit to the network is represented by ๐‘‰๐ธ๐‘™(๐‘˜) and the m-th network output unit by ๐‘›๐‘’๐‘ก๐‘œ๐‘š(๐‘˜). The total input to the i-th first hidden layer unit is denoted as ๐‘›๐‘’๐‘ก๐‘– 1(k). The output of the i-th first hidden layer unit is denoted as ๐‘›๐‘’๐‘ก๐‘– 11(k). The output of the j-th context layer unit is ๐‘‰๐‘— ๐‘(๐‘˜). The total input to the ii-th second hidden layer unit is denoted Corresponding author: Latifa Belhaj Salah Engineering, Technology & Applied Science Research Vol. 9, No. 2, 2019, 3881-3886 3882 www.etasr.com Salah & Fourati: Systems Modeling Using Deep Elman Neural Network as ๐‘›๐‘’๐‘ก๐‘–๐‘– 2(๐‘˜). The output of the ii-th second hidden layer unit is denoted as ๐‘›๐‘’๐‘ก๐‘–๐‘– 22(๐‘˜). The output of the jj-th second context layer unit is ๐‘‰๐‘—๐‘— ๐‘1(๐‘˜). The total input to the i..i-th (n-1-th) hidden layer unit is denoted as ๐‘›๐‘’๐‘ก๐‘–..๐‘– ๐‘›โˆ’1(๐‘˜).The output of the i..i-th (n- 1-th) hidden layer unit is denoted as ๐‘›๐‘’๐‘ก๐‘–..๐‘– (๐‘›โˆ’1)(๐‘›โˆ’1) (๐‘˜). The output of the j..j-th (n-1-th) context layer unit is ๐‘‰๐‘—..๐‘— ๐‘๐‘›โˆ’1(๐‘˜). The total input to the i..ii-th (n-th) hidden layer unit is denoted as ๐‘›๐‘’๐‘ก๐‘–..๐‘–๐‘– ๐‘› (๐‘˜). The output of the i..ii-th (n-th) hidden layer unit is denoted as ๐‘›๐‘’๐‘ก๐‘–..๐‘–๐‘– ๐‘›๐‘›(๐‘˜). The output of the j..jj-th (n-th) context layer unit is ๐‘‰๐‘—..๐‘—๐‘— ๐‘๐‘› (๐‘˜). The total input to the m-th output layer unit is denoted as ๐‘›๐‘’๐‘ก๐‘š ๐‘  (๐‘˜) . ๐‘ค๐‘—,๐‘– ๐‘ (. ) , ๐‘ค๐‘™,๐‘– ๐‘‰๐ธ(. ) , ๐‘ค๐‘–..๐‘–๐‘–,๐‘š ๐‘›๐‘’๐‘ก๐‘œ (. ) , ๐‘ค๐‘–,๐‘–๐‘– ๐‘›๐‘’๐‘ก2(. ) , ๐‘ค๐‘—๐‘—,๐‘–๐‘– ๐‘1 (. ), ๐‘ค๐‘–..๐‘–,๐‘–..๐‘–๐‘– ๐‘›๐‘’๐‘ก๐‘› (. ) , ๐‘ค๐‘—..๐‘—๐‘—,๐‘–..๐‘–๐‘– ๐‘๐‘› (. ) are the weights of the links, respectively between the first context layer and the first hidden layer, the input layer and the first hidden layer, the i..ii-th (n-th) hidden layer and the output layer, the first hidden layer and the second hidden layer, the second context layer and the second hidden layer, the (n-1)-th hidden layer and the n-th hidden layer and the n-th context layer and the n-th hidden layer. Fig. 1. Elman neural network architecture. III. ELMAN NEURAL NETWORK LEARNING STEP The training of the Elman neural network to emulate direct dynamics of a system amounts to minimize the squared error criterion defined as: 2 1 1 ( ( ) ( )) 2 = = โˆ’๏ƒฅ sn k m m m E VS k neto k (1) where VS(k) is the desired output vector. In the training step the back propagation algorithm is used to adjust the Elman neural network connection weights, in order to emulate the dynamics of a system. In the case of many hidden layers, the Elman neural network is governed by the following equations: 1 , , 1 1 ( ) ( 1) ( ) ( 1) ( ) = = = โˆ’ + โˆ’๏ƒฅ ๏ƒฅ c VEn n c c VE i j i j l i l j l net k w k V k w k VE k (2) 11 1 ( ) ( )= i i net k f net (3) 11 ( ) ( 1)= โˆ’ c j j V k net k (4) 1 2 1 12 11 , , 1 1 ( ) ( 1) ( ) ( 1) ( ) = = = โˆ’ + โˆ’๏ƒฅ ๏ƒฅ ch nn c cnet ii i ii i jj ii jj i jj net k w k net k w k V k (5) 22 2 ( ) ( ( ))= ii ii net k f net k (6) 1 22( ) ( 1)= โˆ’ c jj jj V k net k (7) 1 ( 1)( 1) .. .. , .. .. .. 1 .. , .. .. .. 1 ( ) ( 1) ( ) ( 1) ( ) โˆ’ โˆ’ โˆ’ = = = โˆ’ + โˆ’ ๏ƒฅ ๏ƒฅ hn n cn n n n n net n n i ii i i i ii i i i i n c c j jj i ii j jj j jj net k w k net k w k V k (8) .. .. ( ) ( ( ))= nn n i ii i ii net k f net k (9) .. .. ( ) ( 1)= โˆ’n c nn j jj j jj V k net k (10) .. , .. .. 1 ( ) ( 1) ( ) = = โˆ’๏ƒฅ hn n s neto nn m i ii m i ii i ii net k w k net k (11) ( ) ( )= s m m neto k f net (12) where , , c h VE n n n , 1 1 , h c n n , 1nh n โˆ’ , , n nh c n n represent the number of units respectively in the first context layer, the first hidden layer, the input layer, the second hidden layer, the second context layer , the (n-1)th hidden layer, the n-th hidden layer and the n-th context layer. The squared error at the network output is defined as in (1). For .. , ( 1)โˆ’ neto i ii m w k the error gradient is: .. .. , ( ( ) ( )) ( ) ( 1) ๏‚ถ ๏‚ข= โˆ’ โˆ’ ๏‚ถ โˆ’ s m nnk m m i iineto net i ii m E VS k neto k f net k w k (13) for , ( 1)โˆ’ VE l i w k , , ( 1)โˆ’ c j i w k , 2 , ( 1)โˆ’ net i ii w k , 1 , ( 1)โˆ’ c jj ii w k , .. , .. ( 1)โˆ’ n net i i i ii w k and .. , .. ( 1)โˆ’n c j jj i ii w k . 2 1 2 1 .. .. .. , , .. , .. , ( ( ) ( )) ( 1) ( 1) ( 1) ... ( 1) ( )โˆ’ ๏‚ถ ๏‚ข= โˆ’ โˆ’ โˆ’ ๏‚ถ โˆ’ ๏‚ข ๏‚ข ๏‚ข ๏‚ขโˆ’ โˆ’ s m n n n i ii i i ii i netok m m i ii mVE net l i net net i i i ii i ii lnet net net net E VS k neto k f w k w k f w k f f w k f VE k (14) 2 1 2 1 .. .. .. , , .. , .. , ( ( ) ( )) ( 1) ( 1) ( 1) ... ( 1) ( )โˆ’ ๏‚ถ ๏‚ข= โˆ’ โˆ’ โˆ’ ๏‚ถ โˆ’ ๏‚ข ๏‚ข ๏‚ข ๏‚ขโˆ’ โˆ’ s m n n n i ii i i ii i netok m m i ii mc net j i net net c i i i ii i ii jnet net net net E VS k neto k f w k w k f w k f f w k f V k (15) 1 2 .. .. .. ,2 , 11 .. , .. ( ( ) ( )) ( 1) ( 1) ( 1) ... ( )โˆ’ ๏‚ถ ๏‚ข= โˆ’ โˆ’ โˆ’ ๏‚ถ โˆ’ ๏‚ข ๏‚ข ๏‚ขโˆ’ s m n n n i ii i i ii netok m m i ii mnet net i ii net i i i ii inet net net E VS k neto k f w k w k f w k f f net k (16) 1 1 1 2 .. .. .. , , .. , .. ( ( ) ( )) ( 1) ( 1) ( 1) ... ( )โˆ’ ๏‚ถ ๏‚ข= โˆ’ โˆ’ โˆ’ ๏‚ถ โˆ’ ๏‚ข ๏‚ข ๏‚ขโˆ’ s m n n n i ii i i ii netok m m i ii mc net jj ii cnet i i i ii jjnet net net E VS k neto k f w k w k f w k f f V k (17) .. .. , .. , .. ( 1)( 1) .. ( ( ) ( )) ( 1) ( 1) โˆ’ โˆ’ ๏‚ถ ๏‚ข= โˆ’ โˆ’ โˆ’ ๏‚ถ โˆ’ ๏‚ข sn m n i ii netok m m i ii mnetnet i i i ii n n i inet E VS k neto k f w k w k f net (18) Engineering, Technology & Applied Science Research Vol. 9, No. 2, 2019, 3881-3886 3883 www.etasr.com Salah & Fourati: Systems Modeling Using Deep Elman Neural Network .. .. , .. , .. ( 1)( 1) .. ( ( ) ( )) ( 1) ( 1) โˆ’ โˆ’ ๏‚ถ ๏‚ข= โˆ’ โˆ’ โˆ’ ๏‚ถ โˆ’ ๏‚ข sn m n i ii netok m m i ii mnetnet i i i ii n n i inet E VS k neto k f w k w k f net (19) In (13) ๏‚ข s mnet f denotes the derivative of f representing s m net . The general weight modification in the gradient descent method is: ๏ญ ๏‚ถ ๏„ = โˆ’ ๏‚ถ k E w w (20) Where: โ€ข The weight adjustment between the output and the n-th hidden layer are: .. , .. ( ) ( ( ) ( )) ( )s m neto nn i ii m m m i iinet w k VS k neto k f net k๏ญ ๏‚ข๏„ = โˆ’ (21) โ€ข The weights adjustment between the first hidden layer and the input layer are: .. 2 1 2 1 .. , .. , .. , .. , ( ) ( ( ) ( )) ( 1) ( 1) ... ( 1) ( ) ๏ญ โˆ’ ๏‚ข ๏‚ข๏„ = โˆ’ โˆ’ ๏‚ข ๏‚ข ๏‚ขโˆ’ โˆ’ s n m i ii n n i i ii i VE neto l i m m i ii mnet net net net i i i ii i ii lnet net net w k VS k neto k f w k f w k f f w k f VE k (22) โ€ข The weights adjustment between the first hidden and the context layers are: .. 2 1 2 1 .. , .. , .. , .. , ( ) ( ( ) ( )) ( 1) ( 1) ... ( 1) ( ) ๏ญ โˆ’ ๏‚ข ๏‚ข๏„ = โˆ’ โˆ’ ๏‚ข ๏‚ข ๏‚ขโˆ’ โˆ’ s n m i ii n n i i ii i c neto j i m m i ii mnet net net net c i i i ii i ii jnet net net w k VS k neto k f w k f w k f f w k f V k (23) โ€ข The weights adjustment between the first hidden layer and the second hidden layer are: 2 1 2 .. .. , .. , 11 .. , .. ( ) ( ( ) ( )) ( 1) ( 1) ... ( ) ๏ญ โˆ’ ๏‚ข๏„ = โˆ’ โˆ’ ๏‚ข ๏‚ข ๏‚ขโˆ’ s m n n n i ii i i ii net neto i ii m m i ii mnet net i i i ii inet net net w k VS k neto k f w k f w k f f net k (24) โ€ข The weights adjustment between the second hidden and the context layers are: 1 1 1 2 .. .. , .. , .. , .. ( 1) ( ( ) ( )) ( 1) ( 1) ... ( ) ๏ญ โˆ’ ๏‚ข๏„ โˆ’ = โˆ’ โˆ’ ๏‚ข ๏‚ข ๏‚ขโˆ’ s m n n n i ii i i ii c neto jj ii m m i ii mnet cnet i i i ii jjnet net net w k VS k neto k f w k f w k f f V k (25) โ€ข The weights adjustment between the n-th hidden and context layers are: .. .. , .. .. , .. ( ) ( ( ) ( )) ( 1) ๏ญ๏„ = โˆ’ ๏‚ข ๏‚ขโˆ’ n n s n m i ii c j jj i ii m m cneto i ii m j jjnet net w k VS k neto k f w k f V (26) โ€ข The weights adjustment between the n-th hidden layer and the (n-1-th) hidden layer are: .. .. , .. ( 1)( 1) .. , .. ( ) ( ( ) ( )) ( 1) ๏ญ โˆ’ โˆ’ ๏‚ข๏„ = โˆ’ ๏‚ขโˆ’ n s m n i ii net i i i ii m m net neto n n i ii m i inet w k VS k neto k f w k f net (27) IV. CONSIDERED SYSTEMS Complex systems are generally multi-variable, described by non-linearity, characterized by uncertainties and subject to disturbances [31]. The goal of modeling a complex system consists of constructing a concise and efficient neural Elman structure to emulate the dynamics of a system [32, 33]. At first we consider a single input and a single output (SISO) system with a non-linear function f (.), as in Figure 2. Fig. 2. A SISO sytem Here, sin( ) ( )= =y u f u (28) The complex system to be modeled is a greenhouse with multi-inputs, multi-outputs (MIMO), disturbances and uncertainty. It is composed of sensors to measure internal and external climate. The greenhouse outputs are defined by the internal climate (โ€œinternal temperature and internal hygrometryโ€). It is equipped with actuators to control the internal climate, consisted with a sliding shutter with an opening between 0ยฐ and 35ยฐ, a heater which operating in on/off mode with a power of 5kW, a sprayer and a curtain with a length varying between 0 and 3m. The considered greenhouse is a classical one, it is characterized by physical quantities that constitute its functioning [34, 35]: โ€ข Measurable but not controllable input: Te (external temperature (in ยบC), He (external hygrometry in %), Rg (global radiant in W/m2), Vv (wind speed in km/h). โ€ข Measurable and controllable input: Ch (heating input varying between 0 and 1), Ov (sliding shutter in degrees), Br (sprayer varying between 0 and 1), Om (curtain in m). โ€ข Outputs: Ti (internal temperature in ยบC), Hi (internal hygrometry in %). Fig. 3. Greenhouse functional bloc diagram. V. SIMULATION RESULTS Three neural network structures are considered for both systems: the first is with one hidden and context layers, the second one is constituted of two hidden and context layers and the third one has three hidden and context layers. Table I presents the deep Elman neural network caracteristics modeling the non-linear process. Engineering, Technology & Applied Science Research Vol. 9, No. 2, 2019, 3881-3886 3884 www.etasr.com Salah & Fourati: Systems Modeling Using Deep Elman Neural Network TABLE I. DEEP ELMAN NEURAL NETWORK -NON-LINEAR PROCESS. Parameters One hidden and context layers Two hidden and context layers Three hidden and context layers ๐’๐Ÿ 1 1 1 ๐’๐’” 1 1 1 ๐’๐’„๐Ÿ 4 4 4 ๐’๐’‰๐Ÿ 4 4 4 ๐’๐’„๐Ÿ 4 4 ๐’๐’‰๐Ÿ 4 4 ๐’๐’„๐Ÿ‘ 4 ๐’๐’‰๐Ÿ‘ 4 iterations 3000 3000 10000 Learning coefficient 0.2 0.2 0.4 Figures 4-6 represent the evolution of the criterion (1) for the first non-linear process. Table II presents deep Elman neural network characteristics modeling the greenhouse. Figures 7-9 present the evolution of the criterion Ek in the case of the second process (greenhouse). From the Figures and for the two considered systems we conclude that the error Ek is lower in the case of two hidden and context layers network. Thus, the convergence of the error is faster in the case of the neural structure with a single hidden and context layer than the other neural structures. Fig. 4. Evolution of Ek in the case of one hidden layer network. Fig. 5. Evolution of Ek in the case of two hidden layers network. Fig. 6. Evolution of Ek in the case of three hidden layers network. TABLE II. DEEP ELMAN NEURAL NETWORKW-GREENHOUSE Parameters One hidden and context layers Two hidden and context layers Three hidden and context layers ๐’๐Ÿ 8 8 8 ๐’๐’” 2 2 2 ๐’๐’„๐Ÿ 4 4 4 ๐’๐’‰๐Ÿ 4 4 4 ๐’๐’„๐Ÿ 4 4 ๐’๐’‰๐Ÿ 4 4 ๐’๐’„๐Ÿ‘ 4 ๐’๐’‰๐Ÿ‘ 4 iterations 10000 10000 20000 Learning coefficient 0.1 0.2 0.4 Fig. 7. Evolution of Ek in the case of one hidden layer network. Fig. 8. Evolution of Ek in the case of two hidden layers network. Fig. 9. Evolution of Ek in the case of three hidden layers network. In order to compare and validate the three neural models the criterion (29) is considered: ( ) 1 1 ( ) ( ) = = = โˆ’๏ƒฅ๏ƒฅ snnb m t i i k i E abs y k y k (29) where, nb is the operating interval, k is the sample time, ns is the number of outputs, ( ) i y k is the i-th output of the system at time k and ( )m i y k is the i-th output of the neural model at time k. In the case of the first system, nb=1000. Figures 10-12 represent the evolution of the process output (continuous line) and the neural model output (dashed lines). Engineering, Technology & Applied Science Research Vol. 9, No. 2, 2019, 3881-3886 3885 www.etasr.com Salah & Fourati: Systems Modeling Using Deep Elman Neural Network Fig. 10. Evolution of the process and the one hidden layer neural model outputs. Et=59.496. Fig. 11. Evolution of the process and the two hidden layer neural model outputs. Et=57.8332. Fig. 12. Evolution of the process and the three hidden layer neural model outputs. Et=58.9947. In the case of the second system (greenhouse), we have a data file of the whole parametersโ€™ values acting on the greenhouse during one day. The sampling time is one minute, so in one day (24 hours), we obtain a data file with 1440 lines. We divided the data file in two parts, each part constituted of 720 lines. The first part was used for learning (training) and the second was used for the validation. In our case: VEl(k)=[Ov(k),Ch(k),Br(k),Om(k),Te(k),He(k),Rg(k),Vv(k)] T ,netom(k)=[Ti(k) Hi(k)]. Here nb=720. Figures 13 and 14 represent the evolution of the real internal climate (temperature and hygrometry) with continuous lines and the one hidden layer neural model outputs with dashed lines and using the validation data file part. Here Et=88.7962. Fig. 13. Evolution of the internal temperature. Fig. 14. Evolution of the internal hygrometry. Figures 15 and 16 represent the evolution of the real internal climate (temperature and hygrometry) with continuous lines and the two hidden layers neural model outputs with dashed lines and using the validation data file part (Et=82.4499). Figures 17 and 18 represent the evolution of the real internal climate with continuous lines and the three hidden layers neural model outputs with dashed lines and using the validation data file part (Et=90.5219). Fig. 15. Evolution of the internal temperature. Fig. 16. Evolution of the internal hygrometry. Fig. 17. Evolution of the internal temperature. Fig. 18. Evolution of the internal hygrometry. Engineering, Technology & Applied Science Research Vol. 9, No. 2, 2019, 3881-3886 3886 www.etasr.com Salah & Fourati: Systems Modeling Using Deep Elman Neural Network From the previous figures and results, we conclude that the total error is lower in the case of two hidden and context layers than the one of the other neural networks. This Elman neural structure gives better performance results for the modeling and validation steps. VI. CONCLUSION In this paper, we trained a deep Elman neural network to improve the modeling of complex systems. After observing the simulation results, we concluded that the direct model reproduces the dynamic behavior of two processes with acceptable performance. We showed that an Elman network with two hidden and two context layers is the best and the most efficient structure for modeling the complex process. The obtained model will be used in a control oriented task. REFERENCES [1] J. Schmidhuber, โ€œDeep learning in neural networks: An overviewโ€, Neural Networks, Vol. 61, pp. 85-117, 2015 [2] N. Majumder, S. Poria, A. Gelbukh, E. Cambria, โ€œDeep learning-based document modeling for personality detection from textโ€, IEEE Intelligent Systems, Vol. 32, pp. 74-79, 2017 [3] Z. Jiang, L. Li, D. Huang, L. Lin, โ€œTraining word embeddings for deep learning in biomedical text mining tasksโ€, IEEE International Conference on Bioinformatics and Biomedicine, Washington, DC, USA, November 9-12, 2015 [4] L. Deng, G. Hinton, B. Kingsbury, โ€œNew types of deep neual network learning for speech recognition and related applications: an overviewโ€, IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, Canada, May 26-31, 2013 [5] D. Chen, B. K. W. Mak, โ€œMultitask learning of deep neural networks for low-resource speech recognitionโ€, IEEE/ACM Transactions on Audio, Speech, and Language Processing, Vol. 23, No. 7, pp. 1172-1183, 2015 [6] S. M. S. Islam, S. Rahman, M. M. Rahman, E. K. Dey, M. Shoyaib, โ€œApplication of deep learning to computer vision: a comprehensive studyโ€, 5th International Conference on Informatics, Electronics and Vision, Dhaka, Bangladesh, May 13-14, 2016 [7] N. Kruger, P. Janssen, S. Kalkan, M. Lappe, A. Leonardis, J. Piater, A. J. Rodriguez-Sanchez, L. Wiskott, โ€œDeep hierarchies in the primate visual cortex: what can we learn for computer visionโ€, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 35, No. 8, pp. 1847-1871, 2013 [8] M. Chengcai, G. Xiaodong, W. Yuanyuan, โ€œFault diagnosis of power electronic system based on fault gradation and neural network groupโ€, Neurocomputing, Vol. 72, pp. 2909-2914, 2009 [9] S. Hochreiter, J. Schmidhuber, โ€œLong short-term memoryโ€, Neural Computing, Vol. 91, pp. 735-780, 1997 [10] Z. C. Lipton, J. Berkowitz, C. Elkan, โ€œA critical review of recurrent neural networks for sequence learningโ€, available at: https://arxiv.org/abs/1506.00019, 2015 [11] G. E. Hinton, R. R. Salakhutdinov, โ€œReducing the dimensionality of data with neural networksโ€, Science, Vol. 313, pp. 504-507, 2006 [12] X. Chen, X. Lin, โ€œBig data deep learning: challenges and perspectivesโ€, IEEE Access, Vol. 2, pp. 514-525, 2014 [13] J. Gunther, P. M. Pilarski, G. Helfrich, H. Shen, K. Diepold, โ€œIntelligent laser welding through representation, prediction, and control learning: An architecture with deep neural networks and reinforcement learningโ€, Mechatronics, Vol. 34, pp. 1-11, 2016 [14] B. Chandra, R. K. Sharma, โ€œDeep learning with adaptive learning rate using Laplacian scoreโ€, Expert Systems with Applications, Vol. 63, pp. 1-7, 2016 [15] A. Ali, F. Yangyu, โ€œUnsupervised feature learning and automatic modulation classification using deep learning modelโ€, Physical Communication, Vol. 25, No. 1, pp. 75-84, 2017 [16] H. M. Fayek, M. Lech, L. Cacedon, โ€œEvaluating deep learning architectures for Speech Emotion Recognitionโ€, Neural Networks, Vol. 92, pp. 60-68, 2017 [17] S. Achanta, S. V. Gangashetty, โ€œDeep Elman Recurrent Neural Networks for Statistical Parametric Speech Synthesisโ€, Speech Communicationโ€, Vol. 93, pp. 31-42, 2017 [18] H. Zen, H. Sak, โ€œUnidirectional long short-term memory recurrent neural network with recurrent output layer for low-latency speech synthesisโ€, in: 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 4470-4474, IEEE, 2015 [19] Y. Fan, Y. Qian, F. L. Xie, F. K. Soong, โ€œTTS synthesis with bidirectional LSTM based recurrent neural networksโ€, Interspeech 2014, Singapore, September 14-18, 2014 [20] Z. Wu, S. King, โ€œInvestigating gated recurrent networks for speech synthesisโ€, 2016 IEEE International Conference on Acoustics, Speech and Signal Processing, Shanghai, China, March 20-25, 2016 [21] A. Rahman, V. Srikumar, A. D. Smith, โ€œPredicting electricity consumption for commercial and residential buildings using deep recurrent neural networksโ€, Applied Energy, Vol. 212, pp. 372-385, 2018 [22] X. Qian, L. Han, Y. Wang, M. Ding, โ€œDeep learning assisted robust visual tracking with adaptive particle filteringโ€, Signal Processing: Image Communication, Vol. 60, pp. 183-192, 2018 [23] K. P. Ferentinos, โ€œDeep learning models for plant disease detection and diagnosisโ€, Computers and Electronics in Agriculture, Vol. 145, pp. 311- 318, 2018 [24] Q. Zhang, L. T. Yang, Z. Chen, P. Li, โ€œA survey on deep learning for big dataโ€, Information Fusion, Vol. 42, pp. 146-157, 2018 [25] R. Hao, C. Yi, J. Qu, Y. Xin, T. Qiu, โ€œA novel adaptive fault detection methodology for complex system using deep belief networks and multiple models: A case study on cryogenic propellant loading systemโ€, Neurocomputing, Vol. 275, pp. 2111-2125, 2018 [26] F. Fourati, M. Chtourou, โ€œA greenhouse control with feed-forward and recurrent neural networksโ€, Simulation Modelling Practice and Theory, Vol. 15, No. 8, pp. 1016-1028, 2007 [27] J. L. Elman, โ€œFinding structure in timeโ€, Cognitive Science, Vol. 14, No. 2, pp. 179-211, 1990 [28] D. T. Pham, X. Liu, โ€œDynamic system modeling using partially recurrent neural networksโ€, Journal of Systems Engineereingโ€, Vol. 2, No. 2, pp. 90-97, 1992 [29] K. Kamijo, T. Tanigawa, โ€œStock price pattern recognition-a recurrent neural network approachโ€, International Joint Conference on Neural Networks, San Diego, CA, USA, June 17-21, 1990 [30] D. T. Pham, X. Liu, โ€œTraining of Elman networks and dynamic system modellingโ€, International Journal of Systems Science Vol. 27, No. 2, pp. 221-226, 1996 [31] A. Yan, W. Wang, C. Zhang, H. Zhao, โ€œA fault prediction method that uses improved case-based reasoning to continuously predict the status of a shaft furnaceโ€, Information Sciences, Vol. 259, pp. 269-281, 2014 [32] F. Baghernezhad, K. Khorasani, โ€œComputationally intelligent strategies for robust fault detection, isolation, and identification of mobile robotsโ€, Neurocomputing, Vol. 171, pp. 335-346, 2016 [33] H. B. Huang, X. R. Huang, R. X. Li, T. C. Lim, W. P. Ding, โ€œSound quality prediction of vehicle interior noise using deep belief networksโ€, Applied Acoustics,Vol. 113, pp. 149-161, 2016 [34] D. Psaltis, A. Sideris, A. A. Yamamura, โ€œA multilayer neural network Controllerโ€, IEEE International Conference on Neural Networks, San Diego, California, June 21-24, 1987 [35] M. Souissi, Modelisation et Commande du Climat dโ€™une Serre Agricole, PhD Thesis, University of Tunis, Tunis, 2002 (in French) https://sci-hub.tw/https:/www.sciencedirect.com/science/article/abs/pii/S0957415815001555/#! https://sci-hub.tw/https:/www.sciencedirect.com/science/article/abs/pii/S0957415815001555/#! https://sci-hub.tw/https:/www.sciencedirect.com/science/article/abs/pii/S0957415815001555/#! https://sci-hub.tw/https:/www.sciencedirect.com/science/article/abs/pii/S0957415815001555/#! https://sci-hub.tw/https:/www.sciencedirect.com/science/article/abs/pii/S0957415815001555/#! https://sci-hub.tw/https:/www.sciencedirect.com/science/article/abs/pii/S0957415815001555/#! https://sci-hub.tw/https:/www.sciencedirect.com/science/article/abs/pii/S0957415815001555/#! https://sci-hub.tw/https:/www.sciencedirect.com/science/article/abs/pii/S0957415815001555/#! https://sci-hub.tw/https:/www.sciencedirect.com/science/journal/09574158 https://sci-hub.tw/https:/www.sciencedirect.com/science/journal/09574158 https://www.sciencedirect.com/science/journal/09574174 https://www.sciencedirect.com/science/journal/09574174 https://www.sciencedirect.com/science/journal/09574174/63/supp/C https://www.sciencedirect.com/science/journal/09574174/63/supp/C https://www.sciencedirect.com/science/journal/18744907 https://www.sciencedirect.com/science/journal/18744907 https://www.sciencedirect.com/science/journal/18744907 https://www.sciencedirect.com/science/journal/18744907 https://www.sciencedirect.com/science/journal/08936080 https://www.sciencedirect.com/science/journal/08936080 https://www.sciencedirect.com/science/journal/08936080 https://www.sciencedirect.com/science/journal/08936080/92/supp/C https://www.sciencedirect.com/science/journal/08936080/92/supp/C https://www.sciencedirect.com/science/journal/08936080/92/supp/C https://www.sciencedirect.com/science/journal/08936080/92/supp/C https://www.sciencedirect.com/science/journal/08936080/92/supp/C https://www.sciencedirect.com/science/journal/01676393 https://www.sciencedirect.com/science/journal/01676393 https://www.sciencedirect.com/science/journal/01676393 https://www.sciencedirect.com/science/journal/01676393 https://www.sciencedirect.com/science/journal/09235965/60/supp/C https://www.sciencedirect.com/science/journal/09235965/60/supp/C https://www.sciencedirect.com/science/journal/01681699/145/supp/C https://www.sciencedirect.com/science/journal/01681699/145/supp/C https://www.sciencedirect.com/science/journal/09252312 https://www.sciencedirect.com/science/journal/09252312 https://www.sciencedirect.com/science/article/pii/S0020025513003290#! https://www.sciencedirect.com/science/article/pii/S0020025513003290#! https://www.sciencedirect.com/science/article/pii/S0003682X16301785#! https://www.sciencedirect.com/science/article/pii/S0003682X16301785#! https://www.sciencedirect.com/science/article/pii/S0003682X16301785#! https://www.sciencedirect.com/science/article/pii/S0003682X16301785#! https://www.sciencedirect.com/science/article/pii/S0003682X16301785#! https://www.sciencedirect.com/science/article/pii/S0003682X16301785#!