long short-term memory projection recurrent neural network

8
Research Article Long Short-Term Memory Projection Recurrent Neural Network Architectures for Piano’s Continuous Note Recognition YuKang Jia, 1 Zhicheng Wu, 1 Yanyan Xu, 1 Dengfeng Ke, 2 and Kaile Su 3 1 School of Information Science and Technology, Beijing Forestry University, No. 35 Qinghuadong Road, Haidian District, Beijing 100083, China 2 National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy of Sciences, No. 95 Zhongguancundong Road, Haidian District, Beijing 100190, China 3 College of Information Science and Technology, Jinan University, No. 601, West Huangpu Avenue, Guangzhou, Guangdong 510632, China Correspondence should be addressed to Yanyan Xu; [email protected] Received 10 May 2017; Revised 30 July 2017; Accepted 6 August 2017; Published 12 September 2017 Academic Editor: Keigo Watanabe Copyright © 2017 YuKang Jia et al. is is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Long Short-Term Memory (LSTM) is a kind of Recurrent Neural Networks (RNN) relating to time series, which has achieved good performance in speech recogniton and image recognition. Long Short-Term Memory Projection (LSTMP) is a variant of LSTM to further optimize speed and performance of LSTM by adding a projection layer. As LSTM and LSTMP have performed well in pattern recognition, in this paper, we combine them with Connectionist Temporal Classification (CTC) to study piano’s continuous note recognition for robotics. Based on the Beijing Forestry University music library, we conduct experiments to show recognition rates and numbers of iterations of LSTM with a single layer, LSTMP with a single layer, and Deep LSTM (DLSTM, LSTM with multilayers). As a result, the single layer LSTMP proves performing much better than the single layer LSTM in both time and the recognition rate; that is, LSTMP has fewer parameters and therefore reduces the training time, and, moreover, benefiting from the projection layer, LSTMP has better performance, too. e best recognition rate of LSTMP is 99.8%. As for DLSTM, the recognition rate can reach 100% because of the effectiveness of the deep structure, but compared with the single layer LSTMP, DLSTM needs more training time. 1. Introduction Piano’s continuous note recognition is important for a robot, whether it is a bionic robot, a dance robot, or a music robot. ere have been companies researching on music robots. For example, Vadi produced by Vatti is able to identify a voiceprint. Most of the existing piano’s note recognition techniques use Hidden Markov Model (HMM) and Radial Basis Func- tion (RBF) to recognize musical notes with one musical note at a time and therefore are not suitable for continuous note recognition. Fortunately, in the field of pattern recognition, Deep Neural Networks (DNNs) have shown great advantages. DNNs are used to recognize features extracted from a large number of hidden nodes [1] and seek reverse partial guidance through the chain rule and at the same time make the neural network weight matrix convergence through training data iteratively and then achieve recognition [2]. RNN adds a time series based on DNN [3], which makes features have time continuity [4, 5]. However, in experiments, we find that RNN’s time characteristics will disappear completely aſter four iterations [6], and a music note is generally longer than a frame [7], so RNN is not suitable for piano’s continuous note recognition [8]. Fortunately, a variant of RNN, named LSTM, is proposed [9–12], in which an input gate, an output gate, and a forgotten gate are added to memorize a long- term cell state to maintain long-term memory [8, 9, 13–16]. Furthermore, LSTMP adds a projection layer to LSTM to increase its efficiency and effectiveness. is paper studies LSTM and LSTMP for piano’s contin- uous note recognition, and in order to solve the temporal classification problem, we combine LSTM and LSTMP with Hindawi Journal of Robotics Volume 2017, Article ID 2061827, 7 pages https://doi.org/10.1155/2017/2061827

Upload: others

Post on 16-Jan-2022

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Long Short-Term Memory Projection Recurrent Neural Network

Research ArticleLong Short-Term Memory Projection Recurrent Neural NetworkArchitectures for Pianorsquos Continuous Note Recognition

YuKang Jia1 ZhichengWu1 Yanyan Xu1 Dengfeng Ke2 and Kaile Su3

1School of Information Science and Technology Beijing Forestry University No 35 Qinghuadong Road Haidian DistrictBeijing 100083 China2National Laboratory of Pattern Recognition Institute of Automation Chinese Academy of Sciences No 95 Zhongguancundong RoadHaidian District Beijing 100190 China3College of Information Science and Technology Jinan University No 601 West Huangpu Avenue GuangzhouGuangdong 510632 China

Correspondence should be addressed to Yanyan Xu xuyanyanbjfueducn

Received 10 May 2017 Revised 30 July 2017 Accepted 6 August 2017 Published 12 September 2017

Academic Editor Keigo Watanabe

Copyright copy 2017 YuKang Jia et al This is an open access article distributed under the Creative Commons Attribution Licensewhich permits unrestricted use distribution and reproduction in any medium provided the original work is properly cited

Long Short-TermMemory (LSTM) is a kind of Recurrent Neural Networks (RNN) relating to time series which has achieved goodperformance in speech recogniton and image recognition Long Short-Term Memory Projection (LSTMP) is a variant of LSTMto further optimize speed and performance of LSTM by adding a projection layer As LSTM and LSTMP have performed well inpattern recognition in this paper we combine themwith Connectionist Temporal Classification (CTC) to study pianorsquos continuousnote recognition for robotics Based on the Beijing Forestry University music library we conduct experiments to show recognitionrates and numbers of iterations of LSTM with a single layer LSTMP with a single layer and Deep LSTM (DLSTM LSTM withmultilayers) As a result the single layer LSTMP proves performing much better than the single layer LSTM in both time and therecognition rate that is LSTMP has fewer parameters and therefore reduces the training time and moreover benefiting from theprojection layer LSTMP has better performance tooThe best recognition rate of LSTMP is 998 As for DLSTM the recognitionrate can reach 100 because of the effectiveness of the deep structure but compared with the single layer LSTMP DLSTM needsmore training time

1 Introduction

Pianorsquos continuous note recognition is important for a robotwhether it is a bionic robot a dance robot or a music robotThere have been companies researching on music robotsFor example Vadi produced by Vatti is able to identify avoiceprint

Most of the existing pianorsquos note recognition techniquesuse Hidden Markov Model (HMM) and Radial Basis Func-tion (RBF) to recognize musical notes with one musical noteat a time and therefore are not suitable for continuous noterecognition Fortunately in the field of pattern recognitionDeepNeuralNetworks (DNNs) have shown great advantagesDNNs are used to recognize features extracted from a largenumber of hidden nodes [1] and seek reverse partial guidancethrough the chain rule and at the same time make the neural

network weight matrix convergence through training dataiteratively and then achieve recognition [2] RNN adds atime series based on DNN [3] which makes features havetime continuity [4 5] However in experiments we find thatRNNrsquos time characteristics will disappear completely afterfour iterations [6] and a music note is generally longer thana frame [7] so RNN is not suitable for pianorsquos continuousnote recognition [8] Fortunately a variant of RNN namedLSTM is proposed [9ndash12] in which an input gate an outputgate and a forgotten gate are added to memorize a long-term cell state to maintain long-term memory [8 9 13ndash16]Furthermore LSTMP adds a projection layer to LSTM toincrease its efficiency and effectiveness

This paper studies LSTM and LSTMP for pianorsquos contin-uous note recognition and in order to solve the temporalclassification problem we combine LSTM and LSTMP with

HindawiJournal of RoboticsVolume 2017 Article ID 2061827 7 pageshttpsdoiorg10115520172061827

2 Journal of Robotics

Lastoutput

Input Newinput

Input_gate

Output_gate

Cell_value

Forget_gate

Last_cell_recent

Thisinput

Thisinput

This_cell_value Cell_memory OutputActive

ActiveActiveActive

1-input_gate

Figure 1 The LSTM network structure

a method named CTC [17] In experiments we test theperformance of a single layer LSTM Deep LSTM and asingle layer LSTMP with different parameters Comparedwith the traditional pianorsquos note recognition methods LSTMand LSTMP can recognize continuous notes that is somesimple piano music The experimental results show that asingle layer LSTMP can attain a recognition rate of 998 andDeep LSTM can reach 100 which proves that our methodsare quite effective

The rest of this paper is organised as follows In Section 2we first introduce the LSTM network architecture and thenDeep LSTM LSTMP is illustrated in Section 3 In Section 4we discuss CTC The experimental results are presented inSection 5 and finally in Section 6 we draw conclusions andgive our future work

2 LSTM

21 The LSTMNetwork Architecture LSTM is a kind of RNNwhich succeeds to keep memory for a period of time byadding a ldquomemory cellrdquoThememory cell ismainly controlledby ldquothe input gaterdquo ldquothe forgetting gaterdquo and ldquothe outputgaterdquo The input gate activates the input of information tothememory cell and the forgetting gate selectively obliteratessome information in the memory cell and activates thestorage to the next input [18] Finally the output gate decideswhat information will be outputted by the memory cell [19]

The LSTM network structure is illustrated in Figure 1Each box represents different data and the lines with arrowsmean data flow among these data From Figure 1 we canunderstand how LSTM stores memory for a long period oftime

The recognition procedure of LSTM begins with a set ofinput sequences 119909 = (1199091 1199092 119909119905) (119909119894 is a vector) and finally

outputs a set of 119910 = (1199101 1199102 119910119905) (119910119894 is also a vector) whichis calculated according to the following equations

119894119905 = 120590 (119882119894119909119909119905 +119882119894119898119898119905minus1 + 119887119894) (1)

119900119905 = 120590 (119882119900119909119909119905 +119882119900119898119898119905minus1 + 119887119900) (2)

119891119905 = 1 minus 119894119905 (3)

119905119888119905 = 119892 (119882119888119909119909119905 +119882119888119898119898119905minus1 + 119887119888) (4)

119888119905 = 119891119905 ⊙ 119888119905minus1 + 119894119905 ⊙ 119905119888119905 (5)

119898119905 = 119900119905 ⊙ ℎ119888119905 (6)

119910119905 = 120601 (119882119910119898119898119905 + 119887119910) (7)

In these equations 119894 means the input gate and 119900 and 119891are the output gate and the forget gate respectively 119905119888 is theinformation input to the memory cell and 119888 includes cellactivation vectors and119898 is the information the memory celloutputs 119882 represents weight matrices (eg 119882119894119909 representstheweightmatrix from input119909 to the input gate 119894) 119887 is the bias(119887119894 is the input gate bias vector) and 119892 and ℎ are the activationfunction of cell inputting and cell outputting respectivelyregarded as 119905119886119899ℎ and 119897119894119899119890119886119903 in most of the models and alsoin this paper ⊙ is the point multiplication in a matrix 120601 is theactivation function of the neural network output and we use119904119900119891119905119898119886119909 in this paper

After conducting some experiments we find that com-pared with the 119891119905 standard equation (3) is more simple andeasier to converge Not only does the training time becomeless but also the number of iterations becomes smallerTherefore in the neural networks in this paper we use (3) tocalculate 119891119905 instead of the 119891119905 standard equation

22 Deep LSTM In pianorsquos continuous note recognition wealso build a multilayer neural network to further increase the

Journal of Robotics 3

recognition rate Deep LSTM adds an LSTM after anotherand so on [10] The added LSTMs have the same structureas the original one Each layer regards the output fromthe last layer as the input of the next layer We hope thatthe neural networks in different LSTM layers will learndifferent characteristics so as to learn the various features ofmusical notes from different aspects and therefore improvethe recognition rate

3 LSTMP-LSTM with a Projection Layer

In LSTM there are a large number of calculations in thevarious gates calculating the number of parameters119873 in theneural network The weight matrix dimension input by theinput gate the output gate and the cell state at this time is119899119894 lowast 119899119888 and the weight matrix dimension at the last time is119899119888 lowast 119899119888 and the output matrix dimension connected to theoutput of the neural network is 119899119900lowast119899119888 where 119899119894 and 119899119900 are thedimensions of the input and the output respectively and 119899119888 isthe number of memory cells We can easily get the followingformula

119873LSTM = 3 lowast 119899119894 lowast 119899119888 + 3 lowast 119899119888 lowast 119899119888 + 119899119900 lowast 119899119888 (8)

that is

119873LSTM = 3 lowast 119899119894 lowast 119899119888 + 119899119888 lowast (3 lowast 119899119888 + 119899119900) (9)

As we increase 119899119888 119873LSTM grows in a square patternTherefore increasing the number of memory cells to increasethe amount of memory costs a lot but a smaller cellnumber will bring a lower recognition rate so we propose anarchitecture named LSTMP which can not only improve theaccuracy but also effectively reduce the computations

In the output layer of the neural network LSTM outputsa matrix of 1198991198882 lowast 119898119905 Then 119898119905 is sent into the output matrixto be outputted and also serves as the input to the neuralnetwork at the next time We add a 119901119903119900119895119890119888119905119894119900119899 layer to theLSTM architecture and after passing this layer 119898119905 becomesan 119899119888 lowast 119899119903matrix called 119903119905 which replaces 119898119905 as the input ofthe next neural network When the memory cell number ofthe neural network increases the number of parameters inthe neural network is

119873LSTMP = 3 lowast 119899119894 lowast 119899119888 + 3 lowast 119899119888 lowast 119899119903 + 119899119900 lowast 119899119903 + 119899119888

lowast 119899119903(10)

that is

119873LSTMP = 3 lowast 119899119894 lowast 119899119888 + 119899119903 lowast (4 lowast 119899119888 + 119899119900) (11)

Calculating119873LSTM minus 119873LSTMP we have

119873LSTM minus 119873LSTMP = 119899119888 lowast (3 lowast 119899119888 + 119899119900) minus 119899119903

lowast (4 lowast 119899119888 + 119899119900) (12)

Therefore in LSTMP the factor that affects the totalnumber of parameters changes from 119899119888 lowast 119899119888 to 119899119888 lowast 119899119903 Wecan change the value of 119899119888119899119903 to reduce the computational

complexity When 3 lowast 119899119888 gt 4 lowast 119899119903 LSTMP can speed up thetraining model Moreover with the projection layer LSTMPcan converge faster to ensure the convergence of the modelThe mathematical formulae of LSTMP are as follows

119894119905 = 120590 (119882119894119909119909119905 +119882119894119903119903119905minus1 + 119887119894)

119900119905 = 120590 (119882119900119909119909119905 +119882119900119903119903119905minus1 + 119887119900)

119891119905 = 1 minus 119894119905

119905119888119905 = 119892 (119882119888119909119909119905 +119882119888119903119903119905minus1 + 119887119888)

119888119905 = 119891119905 ⊙ 119888119905minus1 + 119894119905 ⊙ 119905119888119905119898119905 = 119900119905 ⊙ ℎ119888119905119903119905 = 119882119903119898119898119905

119910119905 = 120601 (119882119910119903119903119905 + 119887119910)

(13)

In these formulae 119903119905 represents the 119901119903119900119895119890119888119905119894119900119899 layer andthe other equations are the same as LSTM

Figure 2 is the structure of LSTMP and the part markedwith red dashed lines is the projection By comparing Figure 1with Figure 2 we can see that LSTMP is LSTM with aprojection layer

Algorithm 1 is the pseudocode of LSTMP 119908 is the inputweight matrix and 119906 is the weight matrix of the last result 119887is bias and 119903 is the projection matrix We put the extractedmusical notes features into the neural network and thealgorithm executes until we get an acceptable recognitionrate

4 CTC

Theoutput layer of our LSTM and LSTMP is called CTC [20]We use CTC because it does not need presegmented trainingdata or external postprocessing to extract label sequencesfrom the network outputs

To be the same as many latest neural networks CTChas forward and backward algorithms When it comes tothe forward algorithm the key point is to estimate thedistribution through probabilities Given the length 119879 theinput sequence 119909 and the training set 119878 at time 119905 theactivation 119910119905119896 of the output unit 119896 at time 119905 is interpreted asthe probability of observing label 119896 (119900119903 119887119897119886119899119896 119894119891 119896 = |119871| + 1)

119901 (120587 | 119909 119878) =119879

prod119905=1

119910119905120587119905 (14)

We refer to the elements 120587 isin 1198711015840119879 as paths where 1198711015840119879

is the set of the length 119879 sequences over the alphabet 1198711015840 =119871 cup 119887119897119886119899119896 Then we define a many-to-one map 120601 to removefirst the repeated labels and then the blanks from the pathsWith glance at the paths will find they aremutually exclusiveAccording to the characteristic the conditional probability ofsome labelling 119897 isin 119871⩽119879 can be calculated by summing theprobabilities of all the paths mapped onto it by 120601

119901 (119897 | 119909) = sum120587isin120601minus1(119897)

119901 (120587 | 119909) (15)

4 Journal of Robotics

Lastoutput

Input Newinput

Input_gate

Output_gate

Cell_value

Forget_gate

Last_cell_recent

Thisinput

Thisinput

This_cell_value

Cell_memory Cell_recent Output

Active

Active

Active

1-input_gateProjection

Projection

Active

Figure 2 The LSTMP structure

Require 119894119899119901119906119905 119877119890119904119906119897119905119897119886119904119905 119862119890119897119897119897119886119904119905 119908 119906 119887 119903Ensure 119862119890119897119897V119886119897119906119890 119877119890119904119906119897119905while119882ℎ119890119899 ℎ119886V119894119899119892 119894119899119901119906119905 do119905119890119898119901119894119899119901119906119905 = (119877119890119904119906119897119905119897119886119904119905 lowast 119906119894119899119901119906119905) + (119894119899119901119906119905 lowast 119908119894119899119901119906119905) + 119887119905119890119898119901119900119906119905119901119906119905 = (119877119890119904119906119897119905119897119886119904119905 lowast 119906119900119906119905119901119906119905) + (119894119899119901119906119905 lowast 119908119900119906119905119901119906119905) + 119887119905119890119898119901119888119890119897119897 = (119877119890119904119906119897119905119897119886119904119905 lowast 119906119888119890119897119897) + (119894119899119901119906119905 lowast 119908119888119890119897119897) + 119887119894119899119901119906119905119892119886119905119890 = 119904119894119892119898119900119894119889(119905119890119898119901119894119899119901119906119905)119900119906119905119901119906119905119892119886119905119890 = 119904119894119892119898119900119894119889(119905119890119898119901119900119906119905119901119906119905)119891119900119903119892119890119905119892119886119905119890 = 1 minus 119894119899119901119906119905119892119886119905119890119862119890119897119897V119886119897119906119890 = 119886119888119905119894V119886119905119890(119905119890119898119901119888119890119897119897) lowast 119894119899119901119906119905119892119886119905119890 + 119862119890119897119897119897119886119904119905 lowast 119891119900119903119892119890119905119892119886119905119890119877119890119904119906119897119905 = 119900119906119905119901119906119905119892119886119905119890 lowast 119886119888119905119894V119886119905119890(119888119890119897119897V119886119897119906119890)119862119890119897119897V119886119897119906119890 = (119877119890119904119906119897119905 119903)

end while

Algorithm 1 LSTM-projection

After all these procedures CTCwill complete its classificationtask

5 Experiments

We conduct all our experiments on a server with 4 IntelXeon E5-2620 CPUs and 512GB memories A NVIDIA TeslaM2070-Q graphics card is used to train all the models Theprogramming language we use is python 35

We choose the piano as our instrument We record 445note sequences as our dataset and the length of each sequenceis around 8 seconds

In the extraction of features we carry out Hammingwindow processing and then take Fast Fourier Transform(FFT) for the real part and the imaginary part of eachwindowThen we let the FFT result to be orthogonal by adding thesquare of the real part and that of the imaginary part together

Apart from that we gain the log of the quadratic sum Finallythe normalization of the input data is performed

In the experiments the number of kinds of notes is 8 andthe number of input nodes is 9 We try different numbers ofcell units in our models from 20 to 320 The initial value ofthe neural network is set as a random value within [minus02 02]and the learning rate is 0001 In terms of the structures all theneural networks are connected to a single layer CTC As forthe dataset we choose 80 of the samples as the developmentset and 20 as the test set

51 Experimental Results Table 1 shows the recognition ratesand how many times LSTM DLSTM and LSTMP withdifferent parameters need to iterate until their recognitionrates are stable and the best results are in bold

In Table 1 ldquoLSTMP-80 to 20rdquo means the LSTMP modelprojecting 80 cell states to 20 cell states From Table 1 we see

Journal of Robotics 5

Table 1 The number of iterations and the recognition rates of different models

Model Number of iterations Recognition rateLSTM-20 300 772LSTM-40 300 871LSTM-80 300 907LSTM-160 400 825DLSTM-two layers 61 857DLSTM-three layers 71 995DLSTM-four layers 109 1000DLSTM-five layers 76 975DLSTM-six layers 103 1000LSTMP-10 to 20 43 56LSTMP-30 to 20 35 942LSTMP-40 to 20 45 960LSTMP-40 to 30 30 940LSTMP-60 to 30 22 980LSTMP-80 to 20 58 998LSTMP-80 to 30 27 973LSTMP-80 to 40 29 973LSTMP-160 to 40 39 990LSTMP-160 to 80 50 950LSTMP-320 to 160 93 934

LSTMP with different parametersDLSTM with different layers

LSTMP (30 to 20)LSTMP (40 to 20)LSTMP (60 to 30)

LSTMP (80 to 20)LSTMP (160 to 40)

35302520151050

00

02

04

06

08

10

250200150100500

00

02

04

06

08

10

LSTM_layer 1Deep LSTM_layer 3

Deep LSTM_layer 4Deep LSTM_layer 6

300

Figure 3 The recognition rates of LSTMP with different parameters and DLSTM with different layers

that DLSTM and LSTMP perform much better than LSTMand their best recognition rates are almost the same whichare 100 and 998 respectively As for the numbers ofiterations LSTMP needs much less iterations than LSTMand DLSTM which makes LSTMP more suitable for pianorsquoscontinuous note recognition for robotics considering theefficiency

52 LSTMP and DLSTM with Different Parameters Figure 3illustrates LSTMP with different parameters and DLSTMwith different layers The 119909 axis means the number of

iterations and the 119910 axis means the recognition rate We seethat for LSTMP the model projecting 80 cell states to 20 cellstates has the best result but all LSTMP results are very closeAs for DLSTM we see clearly that Deep LSTM is much betterthan LSTM with only one layer

53 Comparisons of LSTM LSTMP and DLSTM We com-pare LSTM LSTMP and DLSTM in Figure 4 Given the sameparameters LSTMP performs much better than LSTM Asfor LSTMP and DLSTM we find that when the number of

6 Journal of Robotics

LSTM and LSTMP LSTMP and DLSTM

Deep LSTM (4 layers)

6050403020100

00

02

04

06

08

10

LSTM (80)LSTMP (80 to 20)

LSTMP (80 to 20)

100806040200

00

02

04

06

08

10

Figure 4 Comparisons of LSTM LSTMP and DLSTM

iterations is small LSTMP has great advantages but as thenumber of iterations increases DLSTM becomes better

6 Conclusions and Future Work

In this paper we have used neural network structures calledLSTM with CTC to recognize continuous musical notes Onthe basis of LSTM we have also tried LSTMP and DLSTMAmong them LSTMP worked best when projecting 80 cellstates to 20 cell states which needed much less iterationsthan LSTM and DLSTM making it most suitable for pianorsquoscontinuous note recognition

In the future we will use LSTM LSTMP and DLSTMto recognize more complex continuous chord music suchas piano music violin pieces or even symphony which willgreatly improve the development of music robots

Conflicts of Interest

The authors declare that they have no conflicts of interest

Acknowledgments

Thanks are due to Yanlin Yang for collecting the record-ing materials This work is supported by the FundamentalResearch Funds for the Central Universities (no 2016JX06)and the National Natural Science Foundation of China (no61472369)

References

[1] A K Jain J Mao and K M Mohiuddin ldquoArtificial neural net-works a tutorialrdquo IEEE Computational Science amp Engineeringvol 29 no 3 pp 31ndash44 1996

[2] J R Zhang T M Lok andM R Lyu ldquoA hybrid particle swarmoptimization-back-propagation algorithm for feedforwardneu-ral network trainingrdquo Applied Mathematics and Computationvol 185 no 2 pp 1026ndash1037 2007

[3] L Liang Y Tang L Bin and W Xiaohua Artificial neuralnetwork based universal time series May 11 2004 US Patent6735580

[4] R J Williams and D Zipser ldquoA learning algorithm forcontinually running fully recurrent neural networksrdquo NeuralComputation vol 1 no 2 pp 270ndash280 1989

[5] L R Medsker and L C Jain ldquoRecurrent neural networksrdquoDesign and Applications vol 5 2001

[6] G Shalini O Vinyals B Strope R Scott T Dean and LHeck ldquoContextual lstm (clstm)models for large scale nlp tasksrdquohttpsarxivorgabs160206291

[7] A Karpathy The unreasonable effectiveness of recurrentneural networks httpkarpathygithubio20150521rnn-effectiveness 2015

[8] D Turnbull and C Elkan ldquoFast recognition of musical genresusingRBFnetworksrdquo IEEETransactions onKnowledge andDataEngineering vol 17 no 4 pp 580ndash584 2005

[9] H Sak A Senior and F Beaufays ldquoLong short-term memoryrecurrent neural network architectures for large scale acousticmodelingrdquo in Proceedings of the 15th Annual Conference of theInternational Speech Communication Association Celebratingthe Diversity of Spoken Languages INTERSPEECH 2014 pp338ndash342 September 2014

[10] AGraves A-RMohamed andGHinton ldquoSpeech recognitionwith deep recurrent neural networksrdquo in Proceedings of the 38thIEEE International Conference on Acoustics Speech and SignalProcessing (ICASSP rsquo13) pp 6645ndash6649 May 2013

[11] J Schmidhuber ldquoDeep learning in neural networks anoverviewrdquo Neural Networks vol 61 pp 85ndash117 2015

[12] K M Hermann T Kocisky E Grefenstette et al ldquoTeachingmachines to read and comprehendrdquo in Proceedings of the 29thAnnual Conference on Neural Information Processing SystemsNIPS 2015 pp 1693ndash1701 December 2015

[13] B Bakker ldquoReinforcement learning with long short-termmem-oryrdquo in In Advances in Neural Information Processing Systemspp 1475ndash1482

[14] S Hochreiter and J Schmidhuber ldquoLong short-term memoryrdquoNeural Computation vol 9 no 8 pp 1735ndash1780 1997

Journal of Robotics 7

[15] P Goelet V F Castellucci S Schacher and E R KandelldquoThe long and the short of long-term memory - A molecularframeworkrdquo Nature vol 322 no 6078 pp 419ndash422 1986

[16] Q Lyu Z Wu and J Zhu ldquoPolyphonic music modelling withLSTM-RTRBMrdquo in Proceedings of the 23rd ACM InternationalConference onMultimedia MM 2015 pp 991ndash994 aus October2015

[17] S Haim A Senior R Kanishka and F Beaufays ldquoFast andaccurate recurrent neural network acoustic models for speechrecognitionrdquo httpsarxivorgabs150706947

[18] F A Gers J Schmidhuber and F Cummins ldquoLearning toforget Continual prediction with LSTMrdquo Neural Computationvol 12 no 10 pp 2451ndash2471 2000

[19] F A Gers N N Schraudolph and J Schmidhuber ldquoLearningprecise timing with LSTM recurrent networksrdquo Journal ofMachine Learning Research (JMLR) vol 3 no 1 pp 115ndash1432003

[20] A Graves S Fernandez F Gomez and J SchmidhuberldquoConnectionist temporal classification Labelling unsegmentedsequence data with recurrent neural networksrdquo in Proceedingsof the 23rd International Conference onMachine Learning ICML2006 pp 369ndash376 Pittsburgh Pa USA June 2006

RoboticsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Active and Passive Electronic Components

Control Scienceand Engineering

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

RotatingMachinery

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporation httpwwwhindawicom

Journal of

Volume 201

Submit your manuscripts athttpswwwhindawicom

VLSI Design

Hindawi Publishing Corporationhttpwwwhindawicom Volume 201

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Shock and Vibration

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawi Publishing Corporation httpwwwhindawicom

Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

SensorsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Modelling amp Simulation in EngineeringHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Navigation and Observation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

DistributedSensor Networks

International Journal of

Page 2: Long Short-Term Memory Projection Recurrent Neural Network

2 Journal of Robotics

Lastoutput

Input Newinput

Input_gate

Output_gate

Cell_value

Forget_gate

Last_cell_recent

Thisinput

Thisinput

This_cell_value Cell_memory OutputActive

ActiveActiveActive

1-input_gate

Figure 1 The LSTM network structure

a method named CTC [17] In experiments we test theperformance of a single layer LSTM Deep LSTM and asingle layer LSTMP with different parameters Comparedwith the traditional pianorsquos note recognition methods LSTMand LSTMP can recognize continuous notes that is somesimple piano music The experimental results show that asingle layer LSTMP can attain a recognition rate of 998 andDeep LSTM can reach 100 which proves that our methodsare quite effective

The rest of this paper is organised as follows In Section 2we first introduce the LSTM network architecture and thenDeep LSTM LSTMP is illustrated in Section 3 In Section 4we discuss CTC The experimental results are presented inSection 5 and finally in Section 6 we draw conclusions andgive our future work

2 LSTM

21 The LSTMNetwork Architecture LSTM is a kind of RNNwhich succeeds to keep memory for a period of time byadding a ldquomemory cellrdquoThememory cell ismainly controlledby ldquothe input gaterdquo ldquothe forgetting gaterdquo and ldquothe outputgaterdquo The input gate activates the input of information tothememory cell and the forgetting gate selectively obliteratessome information in the memory cell and activates thestorage to the next input [18] Finally the output gate decideswhat information will be outputted by the memory cell [19]

The LSTM network structure is illustrated in Figure 1Each box represents different data and the lines with arrowsmean data flow among these data From Figure 1 we canunderstand how LSTM stores memory for a long period oftime

The recognition procedure of LSTM begins with a set ofinput sequences 119909 = (1199091 1199092 119909119905) (119909119894 is a vector) and finally

outputs a set of 119910 = (1199101 1199102 119910119905) (119910119894 is also a vector) whichis calculated according to the following equations

119894119905 = 120590 (119882119894119909119909119905 +119882119894119898119898119905minus1 + 119887119894) (1)

119900119905 = 120590 (119882119900119909119909119905 +119882119900119898119898119905minus1 + 119887119900) (2)

119891119905 = 1 minus 119894119905 (3)

119905119888119905 = 119892 (119882119888119909119909119905 +119882119888119898119898119905minus1 + 119887119888) (4)

119888119905 = 119891119905 ⊙ 119888119905minus1 + 119894119905 ⊙ 119905119888119905 (5)

119898119905 = 119900119905 ⊙ ℎ119888119905 (6)

119910119905 = 120601 (119882119910119898119898119905 + 119887119910) (7)

In these equations 119894 means the input gate and 119900 and 119891are the output gate and the forget gate respectively 119905119888 is theinformation input to the memory cell and 119888 includes cellactivation vectors and119898 is the information the memory celloutputs 119882 represents weight matrices (eg 119882119894119909 representstheweightmatrix from input119909 to the input gate 119894) 119887 is the bias(119887119894 is the input gate bias vector) and 119892 and ℎ are the activationfunction of cell inputting and cell outputting respectivelyregarded as 119905119886119899ℎ and 119897119894119899119890119886119903 in most of the models and alsoin this paper ⊙ is the point multiplication in a matrix 120601 is theactivation function of the neural network output and we use119904119900119891119905119898119886119909 in this paper

After conducting some experiments we find that com-pared with the 119891119905 standard equation (3) is more simple andeasier to converge Not only does the training time becomeless but also the number of iterations becomes smallerTherefore in the neural networks in this paper we use (3) tocalculate 119891119905 instead of the 119891119905 standard equation

22 Deep LSTM In pianorsquos continuous note recognition wealso build a multilayer neural network to further increase the

Journal of Robotics 3

recognition rate Deep LSTM adds an LSTM after anotherand so on [10] The added LSTMs have the same structureas the original one Each layer regards the output fromthe last layer as the input of the next layer We hope thatthe neural networks in different LSTM layers will learndifferent characteristics so as to learn the various features ofmusical notes from different aspects and therefore improvethe recognition rate

3 LSTMP-LSTM with a Projection Layer

In LSTM there are a large number of calculations in thevarious gates calculating the number of parameters119873 in theneural network The weight matrix dimension input by theinput gate the output gate and the cell state at this time is119899119894 lowast 119899119888 and the weight matrix dimension at the last time is119899119888 lowast 119899119888 and the output matrix dimension connected to theoutput of the neural network is 119899119900lowast119899119888 where 119899119894 and 119899119900 are thedimensions of the input and the output respectively and 119899119888 isthe number of memory cells We can easily get the followingformula

119873LSTM = 3 lowast 119899119894 lowast 119899119888 + 3 lowast 119899119888 lowast 119899119888 + 119899119900 lowast 119899119888 (8)

that is

119873LSTM = 3 lowast 119899119894 lowast 119899119888 + 119899119888 lowast (3 lowast 119899119888 + 119899119900) (9)

As we increase 119899119888 119873LSTM grows in a square patternTherefore increasing the number of memory cells to increasethe amount of memory costs a lot but a smaller cellnumber will bring a lower recognition rate so we propose anarchitecture named LSTMP which can not only improve theaccuracy but also effectively reduce the computations

In the output layer of the neural network LSTM outputsa matrix of 1198991198882 lowast 119898119905 Then 119898119905 is sent into the output matrixto be outputted and also serves as the input to the neuralnetwork at the next time We add a 119901119903119900119895119890119888119905119894119900119899 layer to theLSTM architecture and after passing this layer 119898119905 becomesan 119899119888 lowast 119899119903matrix called 119903119905 which replaces 119898119905 as the input ofthe next neural network When the memory cell number ofthe neural network increases the number of parameters inthe neural network is

119873LSTMP = 3 lowast 119899119894 lowast 119899119888 + 3 lowast 119899119888 lowast 119899119903 + 119899119900 lowast 119899119903 + 119899119888

lowast 119899119903(10)

that is

119873LSTMP = 3 lowast 119899119894 lowast 119899119888 + 119899119903 lowast (4 lowast 119899119888 + 119899119900) (11)

Calculating119873LSTM minus 119873LSTMP we have

119873LSTM minus 119873LSTMP = 119899119888 lowast (3 lowast 119899119888 + 119899119900) minus 119899119903

lowast (4 lowast 119899119888 + 119899119900) (12)

Therefore in LSTMP the factor that affects the totalnumber of parameters changes from 119899119888 lowast 119899119888 to 119899119888 lowast 119899119903 Wecan change the value of 119899119888119899119903 to reduce the computational

complexity When 3 lowast 119899119888 gt 4 lowast 119899119903 LSTMP can speed up thetraining model Moreover with the projection layer LSTMPcan converge faster to ensure the convergence of the modelThe mathematical formulae of LSTMP are as follows

119894119905 = 120590 (119882119894119909119909119905 +119882119894119903119903119905minus1 + 119887119894)

119900119905 = 120590 (119882119900119909119909119905 +119882119900119903119903119905minus1 + 119887119900)

119891119905 = 1 minus 119894119905

119905119888119905 = 119892 (119882119888119909119909119905 +119882119888119903119903119905minus1 + 119887119888)

119888119905 = 119891119905 ⊙ 119888119905minus1 + 119894119905 ⊙ 119905119888119905119898119905 = 119900119905 ⊙ ℎ119888119905119903119905 = 119882119903119898119898119905

119910119905 = 120601 (119882119910119903119903119905 + 119887119910)

(13)

In these formulae 119903119905 represents the 119901119903119900119895119890119888119905119894119900119899 layer andthe other equations are the same as LSTM

Figure 2 is the structure of LSTMP and the part markedwith red dashed lines is the projection By comparing Figure 1with Figure 2 we can see that LSTMP is LSTM with aprojection layer

Algorithm 1 is the pseudocode of LSTMP 119908 is the inputweight matrix and 119906 is the weight matrix of the last result 119887is bias and 119903 is the projection matrix We put the extractedmusical notes features into the neural network and thealgorithm executes until we get an acceptable recognitionrate

4 CTC

Theoutput layer of our LSTM and LSTMP is called CTC [20]We use CTC because it does not need presegmented trainingdata or external postprocessing to extract label sequencesfrom the network outputs

To be the same as many latest neural networks CTChas forward and backward algorithms When it comes tothe forward algorithm the key point is to estimate thedistribution through probabilities Given the length 119879 theinput sequence 119909 and the training set 119878 at time 119905 theactivation 119910119905119896 of the output unit 119896 at time 119905 is interpreted asthe probability of observing label 119896 (119900119903 119887119897119886119899119896 119894119891 119896 = |119871| + 1)

119901 (120587 | 119909 119878) =119879

prod119905=1

119910119905120587119905 (14)

We refer to the elements 120587 isin 1198711015840119879 as paths where 1198711015840119879

is the set of the length 119879 sequences over the alphabet 1198711015840 =119871 cup 119887119897119886119899119896 Then we define a many-to-one map 120601 to removefirst the repeated labels and then the blanks from the pathsWith glance at the paths will find they aremutually exclusiveAccording to the characteristic the conditional probability ofsome labelling 119897 isin 119871⩽119879 can be calculated by summing theprobabilities of all the paths mapped onto it by 120601

119901 (119897 | 119909) = sum120587isin120601minus1(119897)

119901 (120587 | 119909) (15)

4 Journal of Robotics

Lastoutput

Input Newinput

Input_gate

Output_gate

Cell_value

Forget_gate

Last_cell_recent

Thisinput

Thisinput

This_cell_value

Cell_memory Cell_recent Output

Active

Active

Active

1-input_gateProjection

Projection

Active

Figure 2 The LSTMP structure

Require 119894119899119901119906119905 119877119890119904119906119897119905119897119886119904119905 119862119890119897119897119897119886119904119905 119908 119906 119887 119903Ensure 119862119890119897119897V119886119897119906119890 119877119890119904119906119897119905while119882ℎ119890119899 ℎ119886V119894119899119892 119894119899119901119906119905 do119905119890119898119901119894119899119901119906119905 = (119877119890119904119906119897119905119897119886119904119905 lowast 119906119894119899119901119906119905) + (119894119899119901119906119905 lowast 119908119894119899119901119906119905) + 119887119905119890119898119901119900119906119905119901119906119905 = (119877119890119904119906119897119905119897119886119904119905 lowast 119906119900119906119905119901119906119905) + (119894119899119901119906119905 lowast 119908119900119906119905119901119906119905) + 119887119905119890119898119901119888119890119897119897 = (119877119890119904119906119897119905119897119886119904119905 lowast 119906119888119890119897119897) + (119894119899119901119906119905 lowast 119908119888119890119897119897) + 119887119894119899119901119906119905119892119886119905119890 = 119904119894119892119898119900119894119889(119905119890119898119901119894119899119901119906119905)119900119906119905119901119906119905119892119886119905119890 = 119904119894119892119898119900119894119889(119905119890119898119901119900119906119905119901119906119905)119891119900119903119892119890119905119892119886119905119890 = 1 minus 119894119899119901119906119905119892119886119905119890119862119890119897119897V119886119897119906119890 = 119886119888119905119894V119886119905119890(119905119890119898119901119888119890119897119897) lowast 119894119899119901119906119905119892119886119905119890 + 119862119890119897119897119897119886119904119905 lowast 119891119900119903119892119890119905119892119886119905119890119877119890119904119906119897119905 = 119900119906119905119901119906119905119892119886119905119890 lowast 119886119888119905119894V119886119905119890(119888119890119897119897V119886119897119906119890)119862119890119897119897V119886119897119906119890 = (119877119890119904119906119897119905 119903)

end while

Algorithm 1 LSTM-projection

After all these procedures CTCwill complete its classificationtask

5 Experiments

We conduct all our experiments on a server with 4 IntelXeon E5-2620 CPUs and 512GB memories A NVIDIA TeslaM2070-Q graphics card is used to train all the models Theprogramming language we use is python 35

We choose the piano as our instrument We record 445note sequences as our dataset and the length of each sequenceis around 8 seconds

In the extraction of features we carry out Hammingwindow processing and then take Fast Fourier Transform(FFT) for the real part and the imaginary part of eachwindowThen we let the FFT result to be orthogonal by adding thesquare of the real part and that of the imaginary part together

Apart from that we gain the log of the quadratic sum Finallythe normalization of the input data is performed

In the experiments the number of kinds of notes is 8 andthe number of input nodes is 9 We try different numbers ofcell units in our models from 20 to 320 The initial value ofthe neural network is set as a random value within [minus02 02]and the learning rate is 0001 In terms of the structures all theneural networks are connected to a single layer CTC As forthe dataset we choose 80 of the samples as the developmentset and 20 as the test set

51 Experimental Results Table 1 shows the recognition ratesand how many times LSTM DLSTM and LSTMP withdifferent parameters need to iterate until their recognitionrates are stable and the best results are in bold

In Table 1 ldquoLSTMP-80 to 20rdquo means the LSTMP modelprojecting 80 cell states to 20 cell states From Table 1 we see

Journal of Robotics 5

Table 1 The number of iterations and the recognition rates of different models

Model Number of iterations Recognition rateLSTM-20 300 772LSTM-40 300 871LSTM-80 300 907LSTM-160 400 825DLSTM-two layers 61 857DLSTM-three layers 71 995DLSTM-four layers 109 1000DLSTM-five layers 76 975DLSTM-six layers 103 1000LSTMP-10 to 20 43 56LSTMP-30 to 20 35 942LSTMP-40 to 20 45 960LSTMP-40 to 30 30 940LSTMP-60 to 30 22 980LSTMP-80 to 20 58 998LSTMP-80 to 30 27 973LSTMP-80 to 40 29 973LSTMP-160 to 40 39 990LSTMP-160 to 80 50 950LSTMP-320 to 160 93 934

LSTMP with different parametersDLSTM with different layers

LSTMP (30 to 20)LSTMP (40 to 20)LSTMP (60 to 30)

LSTMP (80 to 20)LSTMP (160 to 40)

35302520151050

00

02

04

06

08

10

250200150100500

00

02

04

06

08

10

LSTM_layer 1Deep LSTM_layer 3

Deep LSTM_layer 4Deep LSTM_layer 6

300

Figure 3 The recognition rates of LSTMP with different parameters and DLSTM with different layers

that DLSTM and LSTMP perform much better than LSTMand their best recognition rates are almost the same whichare 100 and 998 respectively As for the numbers ofiterations LSTMP needs much less iterations than LSTMand DLSTM which makes LSTMP more suitable for pianorsquoscontinuous note recognition for robotics considering theefficiency

52 LSTMP and DLSTM with Different Parameters Figure 3illustrates LSTMP with different parameters and DLSTMwith different layers The 119909 axis means the number of

iterations and the 119910 axis means the recognition rate We seethat for LSTMP the model projecting 80 cell states to 20 cellstates has the best result but all LSTMP results are very closeAs for DLSTM we see clearly that Deep LSTM is much betterthan LSTM with only one layer

53 Comparisons of LSTM LSTMP and DLSTM We com-pare LSTM LSTMP and DLSTM in Figure 4 Given the sameparameters LSTMP performs much better than LSTM Asfor LSTMP and DLSTM we find that when the number of

6 Journal of Robotics

LSTM and LSTMP LSTMP and DLSTM

Deep LSTM (4 layers)

6050403020100

00

02

04

06

08

10

LSTM (80)LSTMP (80 to 20)

LSTMP (80 to 20)

100806040200

00

02

04

06

08

10

Figure 4 Comparisons of LSTM LSTMP and DLSTM

iterations is small LSTMP has great advantages but as thenumber of iterations increases DLSTM becomes better

6 Conclusions and Future Work

In this paper we have used neural network structures calledLSTM with CTC to recognize continuous musical notes Onthe basis of LSTM we have also tried LSTMP and DLSTMAmong them LSTMP worked best when projecting 80 cellstates to 20 cell states which needed much less iterationsthan LSTM and DLSTM making it most suitable for pianorsquoscontinuous note recognition

In the future we will use LSTM LSTMP and DLSTMto recognize more complex continuous chord music suchas piano music violin pieces or even symphony which willgreatly improve the development of music robots

Conflicts of Interest

The authors declare that they have no conflicts of interest

Acknowledgments

Thanks are due to Yanlin Yang for collecting the record-ing materials This work is supported by the FundamentalResearch Funds for the Central Universities (no 2016JX06)and the National Natural Science Foundation of China (no61472369)

References

[1] A K Jain J Mao and K M Mohiuddin ldquoArtificial neural net-works a tutorialrdquo IEEE Computational Science amp Engineeringvol 29 no 3 pp 31ndash44 1996

[2] J R Zhang T M Lok andM R Lyu ldquoA hybrid particle swarmoptimization-back-propagation algorithm for feedforwardneu-ral network trainingrdquo Applied Mathematics and Computationvol 185 no 2 pp 1026ndash1037 2007

[3] L Liang Y Tang L Bin and W Xiaohua Artificial neuralnetwork based universal time series May 11 2004 US Patent6735580

[4] R J Williams and D Zipser ldquoA learning algorithm forcontinually running fully recurrent neural networksrdquo NeuralComputation vol 1 no 2 pp 270ndash280 1989

[5] L R Medsker and L C Jain ldquoRecurrent neural networksrdquoDesign and Applications vol 5 2001

[6] G Shalini O Vinyals B Strope R Scott T Dean and LHeck ldquoContextual lstm (clstm)models for large scale nlp tasksrdquohttpsarxivorgabs160206291

[7] A Karpathy The unreasonable effectiveness of recurrentneural networks httpkarpathygithubio20150521rnn-effectiveness 2015

[8] D Turnbull and C Elkan ldquoFast recognition of musical genresusingRBFnetworksrdquo IEEETransactions onKnowledge andDataEngineering vol 17 no 4 pp 580ndash584 2005

[9] H Sak A Senior and F Beaufays ldquoLong short-term memoryrecurrent neural network architectures for large scale acousticmodelingrdquo in Proceedings of the 15th Annual Conference of theInternational Speech Communication Association Celebratingthe Diversity of Spoken Languages INTERSPEECH 2014 pp338ndash342 September 2014

[10] AGraves A-RMohamed andGHinton ldquoSpeech recognitionwith deep recurrent neural networksrdquo in Proceedings of the 38thIEEE International Conference on Acoustics Speech and SignalProcessing (ICASSP rsquo13) pp 6645ndash6649 May 2013

[11] J Schmidhuber ldquoDeep learning in neural networks anoverviewrdquo Neural Networks vol 61 pp 85ndash117 2015

[12] K M Hermann T Kocisky E Grefenstette et al ldquoTeachingmachines to read and comprehendrdquo in Proceedings of the 29thAnnual Conference on Neural Information Processing SystemsNIPS 2015 pp 1693ndash1701 December 2015

[13] B Bakker ldquoReinforcement learning with long short-termmem-oryrdquo in In Advances in Neural Information Processing Systemspp 1475ndash1482

[14] S Hochreiter and J Schmidhuber ldquoLong short-term memoryrdquoNeural Computation vol 9 no 8 pp 1735ndash1780 1997

Journal of Robotics 7

[15] P Goelet V F Castellucci S Schacher and E R KandelldquoThe long and the short of long-term memory - A molecularframeworkrdquo Nature vol 322 no 6078 pp 419ndash422 1986

[16] Q Lyu Z Wu and J Zhu ldquoPolyphonic music modelling withLSTM-RTRBMrdquo in Proceedings of the 23rd ACM InternationalConference onMultimedia MM 2015 pp 991ndash994 aus October2015

[17] S Haim A Senior R Kanishka and F Beaufays ldquoFast andaccurate recurrent neural network acoustic models for speechrecognitionrdquo httpsarxivorgabs150706947

[18] F A Gers J Schmidhuber and F Cummins ldquoLearning toforget Continual prediction with LSTMrdquo Neural Computationvol 12 no 10 pp 2451ndash2471 2000

[19] F A Gers N N Schraudolph and J Schmidhuber ldquoLearningprecise timing with LSTM recurrent networksrdquo Journal ofMachine Learning Research (JMLR) vol 3 no 1 pp 115ndash1432003

[20] A Graves S Fernandez F Gomez and J SchmidhuberldquoConnectionist temporal classification Labelling unsegmentedsequence data with recurrent neural networksrdquo in Proceedingsof the 23rd International Conference onMachine Learning ICML2006 pp 369ndash376 Pittsburgh Pa USA June 2006

RoboticsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Active and Passive Electronic Components

Control Scienceand Engineering

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

RotatingMachinery

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporation httpwwwhindawicom

Journal of

Volume 201

Submit your manuscripts athttpswwwhindawicom

VLSI Design

Hindawi Publishing Corporationhttpwwwhindawicom Volume 201

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Shock and Vibration

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawi Publishing Corporation httpwwwhindawicom

Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

SensorsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Modelling amp Simulation in EngineeringHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Navigation and Observation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

DistributedSensor Networks

International Journal of

Page 3: Long Short-Term Memory Projection Recurrent Neural Network

Journal of Robotics 3

recognition rate Deep LSTM adds an LSTM after anotherand so on [10] The added LSTMs have the same structureas the original one Each layer regards the output fromthe last layer as the input of the next layer We hope thatthe neural networks in different LSTM layers will learndifferent characteristics so as to learn the various features ofmusical notes from different aspects and therefore improvethe recognition rate

3 LSTMP-LSTM with a Projection Layer

In LSTM there are a large number of calculations in thevarious gates calculating the number of parameters119873 in theneural network The weight matrix dimension input by theinput gate the output gate and the cell state at this time is119899119894 lowast 119899119888 and the weight matrix dimension at the last time is119899119888 lowast 119899119888 and the output matrix dimension connected to theoutput of the neural network is 119899119900lowast119899119888 where 119899119894 and 119899119900 are thedimensions of the input and the output respectively and 119899119888 isthe number of memory cells We can easily get the followingformula

119873LSTM = 3 lowast 119899119894 lowast 119899119888 + 3 lowast 119899119888 lowast 119899119888 + 119899119900 lowast 119899119888 (8)

that is

119873LSTM = 3 lowast 119899119894 lowast 119899119888 + 119899119888 lowast (3 lowast 119899119888 + 119899119900) (9)

As we increase 119899119888 119873LSTM grows in a square patternTherefore increasing the number of memory cells to increasethe amount of memory costs a lot but a smaller cellnumber will bring a lower recognition rate so we propose anarchitecture named LSTMP which can not only improve theaccuracy but also effectively reduce the computations

In the output layer of the neural network LSTM outputsa matrix of 1198991198882 lowast 119898119905 Then 119898119905 is sent into the output matrixto be outputted and also serves as the input to the neuralnetwork at the next time We add a 119901119903119900119895119890119888119905119894119900119899 layer to theLSTM architecture and after passing this layer 119898119905 becomesan 119899119888 lowast 119899119903matrix called 119903119905 which replaces 119898119905 as the input ofthe next neural network When the memory cell number ofthe neural network increases the number of parameters inthe neural network is

119873LSTMP = 3 lowast 119899119894 lowast 119899119888 + 3 lowast 119899119888 lowast 119899119903 + 119899119900 lowast 119899119903 + 119899119888

lowast 119899119903(10)

that is

119873LSTMP = 3 lowast 119899119894 lowast 119899119888 + 119899119903 lowast (4 lowast 119899119888 + 119899119900) (11)

Calculating119873LSTM minus 119873LSTMP we have

119873LSTM minus 119873LSTMP = 119899119888 lowast (3 lowast 119899119888 + 119899119900) minus 119899119903

lowast (4 lowast 119899119888 + 119899119900) (12)

Therefore in LSTMP the factor that affects the totalnumber of parameters changes from 119899119888 lowast 119899119888 to 119899119888 lowast 119899119903 Wecan change the value of 119899119888119899119903 to reduce the computational

complexity When 3 lowast 119899119888 gt 4 lowast 119899119903 LSTMP can speed up thetraining model Moreover with the projection layer LSTMPcan converge faster to ensure the convergence of the modelThe mathematical formulae of LSTMP are as follows

119894119905 = 120590 (119882119894119909119909119905 +119882119894119903119903119905minus1 + 119887119894)

119900119905 = 120590 (119882119900119909119909119905 +119882119900119903119903119905minus1 + 119887119900)

119891119905 = 1 minus 119894119905

119905119888119905 = 119892 (119882119888119909119909119905 +119882119888119903119903119905minus1 + 119887119888)

119888119905 = 119891119905 ⊙ 119888119905minus1 + 119894119905 ⊙ 119905119888119905119898119905 = 119900119905 ⊙ ℎ119888119905119903119905 = 119882119903119898119898119905

119910119905 = 120601 (119882119910119903119903119905 + 119887119910)

(13)

In these formulae 119903119905 represents the 119901119903119900119895119890119888119905119894119900119899 layer andthe other equations are the same as LSTM

Figure 2 is the structure of LSTMP and the part markedwith red dashed lines is the projection By comparing Figure 1with Figure 2 we can see that LSTMP is LSTM with aprojection layer

Algorithm 1 is the pseudocode of LSTMP 119908 is the inputweight matrix and 119906 is the weight matrix of the last result 119887is bias and 119903 is the projection matrix We put the extractedmusical notes features into the neural network and thealgorithm executes until we get an acceptable recognitionrate

4 CTC

Theoutput layer of our LSTM and LSTMP is called CTC [20]We use CTC because it does not need presegmented trainingdata or external postprocessing to extract label sequencesfrom the network outputs

To be the same as many latest neural networks CTChas forward and backward algorithms When it comes tothe forward algorithm the key point is to estimate thedistribution through probabilities Given the length 119879 theinput sequence 119909 and the training set 119878 at time 119905 theactivation 119910119905119896 of the output unit 119896 at time 119905 is interpreted asthe probability of observing label 119896 (119900119903 119887119897119886119899119896 119894119891 119896 = |119871| + 1)

119901 (120587 | 119909 119878) =119879

prod119905=1

119910119905120587119905 (14)

We refer to the elements 120587 isin 1198711015840119879 as paths where 1198711015840119879

is the set of the length 119879 sequences over the alphabet 1198711015840 =119871 cup 119887119897119886119899119896 Then we define a many-to-one map 120601 to removefirst the repeated labels and then the blanks from the pathsWith glance at the paths will find they aremutually exclusiveAccording to the characteristic the conditional probability ofsome labelling 119897 isin 119871⩽119879 can be calculated by summing theprobabilities of all the paths mapped onto it by 120601

119901 (119897 | 119909) = sum120587isin120601minus1(119897)

119901 (120587 | 119909) (15)

4 Journal of Robotics

Lastoutput

Input Newinput

Input_gate

Output_gate

Cell_value

Forget_gate

Last_cell_recent

Thisinput

Thisinput

This_cell_value

Cell_memory Cell_recent Output

Active

Active

Active

1-input_gateProjection

Projection

Active

Figure 2 The LSTMP structure

Require 119894119899119901119906119905 119877119890119904119906119897119905119897119886119904119905 119862119890119897119897119897119886119904119905 119908 119906 119887 119903Ensure 119862119890119897119897V119886119897119906119890 119877119890119904119906119897119905while119882ℎ119890119899 ℎ119886V119894119899119892 119894119899119901119906119905 do119905119890119898119901119894119899119901119906119905 = (119877119890119904119906119897119905119897119886119904119905 lowast 119906119894119899119901119906119905) + (119894119899119901119906119905 lowast 119908119894119899119901119906119905) + 119887119905119890119898119901119900119906119905119901119906119905 = (119877119890119904119906119897119905119897119886119904119905 lowast 119906119900119906119905119901119906119905) + (119894119899119901119906119905 lowast 119908119900119906119905119901119906119905) + 119887119905119890119898119901119888119890119897119897 = (119877119890119904119906119897119905119897119886119904119905 lowast 119906119888119890119897119897) + (119894119899119901119906119905 lowast 119908119888119890119897119897) + 119887119894119899119901119906119905119892119886119905119890 = 119904119894119892119898119900119894119889(119905119890119898119901119894119899119901119906119905)119900119906119905119901119906119905119892119886119905119890 = 119904119894119892119898119900119894119889(119905119890119898119901119900119906119905119901119906119905)119891119900119903119892119890119905119892119886119905119890 = 1 minus 119894119899119901119906119905119892119886119905119890119862119890119897119897V119886119897119906119890 = 119886119888119905119894V119886119905119890(119905119890119898119901119888119890119897119897) lowast 119894119899119901119906119905119892119886119905119890 + 119862119890119897119897119897119886119904119905 lowast 119891119900119903119892119890119905119892119886119905119890119877119890119904119906119897119905 = 119900119906119905119901119906119905119892119886119905119890 lowast 119886119888119905119894V119886119905119890(119888119890119897119897V119886119897119906119890)119862119890119897119897V119886119897119906119890 = (119877119890119904119906119897119905 119903)

end while

Algorithm 1 LSTM-projection

After all these procedures CTCwill complete its classificationtask

5 Experiments

We conduct all our experiments on a server with 4 IntelXeon E5-2620 CPUs and 512GB memories A NVIDIA TeslaM2070-Q graphics card is used to train all the models Theprogramming language we use is python 35

We choose the piano as our instrument We record 445note sequences as our dataset and the length of each sequenceis around 8 seconds

In the extraction of features we carry out Hammingwindow processing and then take Fast Fourier Transform(FFT) for the real part and the imaginary part of eachwindowThen we let the FFT result to be orthogonal by adding thesquare of the real part and that of the imaginary part together

Apart from that we gain the log of the quadratic sum Finallythe normalization of the input data is performed

In the experiments the number of kinds of notes is 8 andthe number of input nodes is 9 We try different numbers ofcell units in our models from 20 to 320 The initial value ofthe neural network is set as a random value within [minus02 02]and the learning rate is 0001 In terms of the structures all theneural networks are connected to a single layer CTC As forthe dataset we choose 80 of the samples as the developmentset and 20 as the test set

51 Experimental Results Table 1 shows the recognition ratesand how many times LSTM DLSTM and LSTMP withdifferent parameters need to iterate until their recognitionrates are stable and the best results are in bold

In Table 1 ldquoLSTMP-80 to 20rdquo means the LSTMP modelprojecting 80 cell states to 20 cell states From Table 1 we see

Journal of Robotics 5

Table 1 The number of iterations and the recognition rates of different models

Model Number of iterations Recognition rateLSTM-20 300 772LSTM-40 300 871LSTM-80 300 907LSTM-160 400 825DLSTM-two layers 61 857DLSTM-three layers 71 995DLSTM-four layers 109 1000DLSTM-five layers 76 975DLSTM-six layers 103 1000LSTMP-10 to 20 43 56LSTMP-30 to 20 35 942LSTMP-40 to 20 45 960LSTMP-40 to 30 30 940LSTMP-60 to 30 22 980LSTMP-80 to 20 58 998LSTMP-80 to 30 27 973LSTMP-80 to 40 29 973LSTMP-160 to 40 39 990LSTMP-160 to 80 50 950LSTMP-320 to 160 93 934

LSTMP with different parametersDLSTM with different layers

LSTMP (30 to 20)LSTMP (40 to 20)LSTMP (60 to 30)

LSTMP (80 to 20)LSTMP (160 to 40)

35302520151050

00

02

04

06

08

10

250200150100500

00

02

04

06

08

10

LSTM_layer 1Deep LSTM_layer 3

Deep LSTM_layer 4Deep LSTM_layer 6

300

Figure 3 The recognition rates of LSTMP with different parameters and DLSTM with different layers

that DLSTM and LSTMP perform much better than LSTMand their best recognition rates are almost the same whichare 100 and 998 respectively As for the numbers ofiterations LSTMP needs much less iterations than LSTMand DLSTM which makes LSTMP more suitable for pianorsquoscontinuous note recognition for robotics considering theefficiency

52 LSTMP and DLSTM with Different Parameters Figure 3illustrates LSTMP with different parameters and DLSTMwith different layers The 119909 axis means the number of

iterations and the 119910 axis means the recognition rate We seethat for LSTMP the model projecting 80 cell states to 20 cellstates has the best result but all LSTMP results are very closeAs for DLSTM we see clearly that Deep LSTM is much betterthan LSTM with only one layer

53 Comparisons of LSTM LSTMP and DLSTM We com-pare LSTM LSTMP and DLSTM in Figure 4 Given the sameparameters LSTMP performs much better than LSTM Asfor LSTMP and DLSTM we find that when the number of

6 Journal of Robotics

LSTM and LSTMP LSTMP and DLSTM

Deep LSTM (4 layers)

6050403020100

00

02

04

06

08

10

LSTM (80)LSTMP (80 to 20)

LSTMP (80 to 20)

100806040200

00

02

04

06

08

10

Figure 4 Comparisons of LSTM LSTMP and DLSTM

iterations is small LSTMP has great advantages but as thenumber of iterations increases DLSTM becomes better

6 Conclusions and Future Work

In this paper we have used neural network structures calledLSTM with CTC to recognize continuous musical notes Onthe basis of LSTM we have also tried LSTMP and DLSTMAmong them LSTMP worked best when projecting 80 cellstates to 20 cell states which needed much less iterationsthan LSTM and DLSTM making it most suitable for pianorsquoscontinuous note recognition

In the future we will use LSTM LSTMP and DLSTMto recognize more complex continuous chord music suchas piano music violin pieces or even symphony which willgreatly improve the development of music robots

Conflicts of Interest

The authors declare that they have no conflicts of interest

Acknowledgments

Thanks are due to Yanlin Yang for collecting the record-ing materials This work is supported by the FundamentalResearch Funds for the Central Universities (no 2016JX06)and the National Natural Science Foundation of China (no61472369)

References

[1] A K Jain J Mao and K M Mohiuddin ldquoArtificial neural net-works a tutorialrdquo IEEE Computational Science amp Engineeringvol 29 no 3 pp 31ndash44 1996

[2] J R Zhang T M Lok andM R Lyu ldquoA hybrid particle swarmoptimization-back-propagation algorithm for feedforwardneu-ral network trainingrdquo Applied Mathematics and Computationvol 185 no 2 pp 1026ndash1037 2007

[3] L Liang Y Tang L Bin and W Xiaohua Artificial neuralnetwork based universal time series May 11 2004 US Patent6735580

[4] R J Williams and D Zipser ldquoA learning algorithm forcontinually running fully recurrent neural networksrdquo NeuralComputation vol 1 no 2 pp 270ndash280 1989

[5] L R Medsker and L C Jain ldquoRecurrent neural networksrdquoDesign and Applications vol 5 2001

[6] G Shalini O Vinyals B Strope R Scott T Dean and LHeck ldquoContextual lstm (clstm)models for large scale nlp tasksrdquohttpsarxivorgabs160206291

[7] A Karpathy The unreasonable effectiveness of recurrentneural networks httpkarpathygithubio20150521rnn-effectiveness 2015

[8] D Turnbull and C Elkan ldquoFast recognition of musical genresusingRBFnetworksrdquo IEEETransactions onKnowledge andDataEngineering vol 17 no 4 pp 580ndash584 2005

[9] H Sak A Senior and F Beaufays ldquoLong short-term memoryrecurrent neural network architectures for large scale acousticmodelingrdquo in Proceedings of the 15th Annual Conference of theInternational Speech Communication Association Celebratingthe Diversity of Spoken Languages INTERSPEECH 2014 pp338ndash342 September 2014

[10] AGraves A-RMohamed andGHinton ldquoSpeech recognitionwith deep recurrent neural networksrdquo in Proceedings of the 38thIEEE International Conference on Acoustics Speech and SignalProcessing (ICASSP rsquo13) pp 6645ndash6649 May 2013

[11] J Schmidhuber ldquoDeep learning in neural networks anoverviewrdquo Neural Networks vol 61 pp 85ndash117 2015

[12] K M Hermann T Kocisky E Grefenstette et al ldquoTeachingmachines to read and comprehendrdquo in Proceedings of the 29thAnnual Conference on Neural Information Processing SystemsNIPS 2015 pp 1693ndash1701 December 2015

[13] B Bakker ldquoReinforcement learning with long short-termmem-oryrdquo in In Advances in Neural Information Processing Systemspp 1475ndash1482

[14] S Hochreiter and J Schmidhuber ldquoLong short-term memoryrdquoNeural Computation vol 9 no 8 pp 1735ndash1780 1997

Journal of Robotics 7

[15] P Goelet V F Castellucci S Schacher and E R KandelldquoThe long and the short of long-term memory - A molecularframeworkrdquo Nature vol 322 no 6078 pp 419ndash422 1986

[16] Q Lyu Z Wu and J Zhu ldquoPolyphonic music modelling withLSTM-RTRBMrdquo in Proceedings of the 23rd ACM InternationalConference onMultimedia MM 2015 pp 991ndash994 aus October2015

[17] S Haim A Senior R Kanishka and F Beaufays ldquoFast andaccurate recurrent neural network acoustic models for speechrecognitionrdquo httpsarxivorgabs150706947

[18] F A Gers J Schmidhuber and F Cummins ldquoLearning toforget Continual prediction with LSTMrdquo Neural Computationvol 12 no 10 pp 2451ndash2471 2000

[19] F A Gers N N Schraudolph and J Schmidhuber ldquoLearningprecise timing with LSTM recurrent networksrdquo Journal ofMachine Learning Research (JMLR) vol 3 no 1 pp 115ndash1432003

[20] A Graves S Fernandez F Gomez and J SchmidhuberldquoConnectionist temporal classification Labelling unsegmentedsequence data with recurrent neural networksrdquo in Proceedingsof the 23rd International Conference onMachine Learning ICML2006 pp 369ndash376 Pittsburgh Pa USA June 2006

RoboticsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Active and Passive Electronic Components

Control Scienceand Engineering

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

RotatingMachinery

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporation httpwwwhindawicom

Journal of

Volume 201

Submit your manuscripts athttpswwwhindawicom

VLSI Design

Hindawi Publishing Corporationhttpwwwhindawicom Volume 201

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Shock and Vibration

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawi Publishing Corporation httpwwwhindawicom

Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

SensorsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Modelling amp Simulation in EngineeringHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Navigation and Observation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

DistributedSensor Networks

International Journal of

Page 4: Long Short-Term Memory Projection Recurrent Neural Network

4 Journal of Robotics

Lastoutput

Input Newinput

Input_gate

Output_gate

Cell_value

Forget_gate

Last_cell_recent

Thisinput

Thisinput

This_cell_value

Cell_memory Cell_recent Output

Active

Active

Active

1-input_gateProjection

Projection

Active

Figure 2 The LSTMP structure

Require 119894119899119901119906119905 119877119890119904119906119897119905119897119886119904119905 119862119890119897119897119897119886119904119905 119908 119906 119887 119903Ensure 119862119890119897119897V119886119897119906119890 119877119890119904119906119897119905while119882ℎ119890119899 ℎ119886V119894119899119892 119894119899119901119906119905 do119905119890119898119901119894119899119901119906119905 = (119877119890119904119906119897119905119897119886119904119905 lowast 119906119894119899119901119906119905) + (119894119899119901119906119905 lowast 119908119894119899119901119906119905) + 119887119905119890119898119901119900119906119905119901119906119905 = (119877119890119904119906119897119905119897119886119904119905 lowast 119906119900119906119905119901119906119905) + (119894119899119901119906119905 lowast 119908119900119906119905119901119906119905) + 119887119905119890119898119901119888119890119897119897 = (119877119890119904119906119897119905119897119886119904119905 lowast 119906119888119890119897119897) + (119894119899119901119906119905 lowast 119908119888119890119897119897) + 119887119894119899119901119906119905119892119886119905119890 = 119904119894119892119898119900119894119889(119905119890119898119901119894119899119901119906119905)119900119906119905119901119906119905119892119886119905119890 = 119904119894119892119898119900119894119889(119905119890119898119901119900119906119905119901119906119905)119891119900119903119892119890119905119892119886119905119890 = 1 minus 119894119899119901119906119905119892119886119905119890119862119890119897119897V119886119897119906119890 = 119886119888119905119894V119886119905119890(119905119890119898119901119888119890119897119897) lowast 119894119899119901119906119905119892119886119905119890 + 119862119890119897119897119897119886119904119905 lowast 119891119900119903119892119890119905119892119886119905119890119877119890119904119906119897119905 = 119900119906119905119901119906119905119892119886119905119890 lowast 119886119888119905119894V119886119905119890(119888119890119897119897V119886119897119906119890)119862119890119897119897V119886119897119906119890 = (119877119890119904119906119897119905 119903)

end while

Algorithm 1 LSTM-projection

After all these procedures CTCwill complete its classificationtask

5 Experiments

We conduct all our experiments on a server with 4 IntelXeon E5-2620 CPUs and 512GB memories A NVIDIA TeslaM2070-Q graphics card is used to train all the models Theprogramming language we use is python 35

We choose the piano as our instrument We record 445note sequences as our dataset and the length of each sequenceis around 8 seconds

In the extraction of features we carry out Hammingwindow processing and then take Fast Fourier Transform(FFT) for the real part and the imaginary part of eachwindowThen we let the FFT result to be orthogonal by adding thesquare of the real part and that of the imaginary part together

Apart from that we gain the log of the quadratic sum Finallythe normalization of the input data is performed

In the experiments the number of kinds of notes is 8 andthe number of input nodes is 9 We try different numbers ofcell units in our models from 20 to 320 The initial value ofthe neural network is set as a random value within [minus02 02]and the learning rate is 0001 In terms of the structures all theneural networks are connected to a single layer CTC As forthe dataset we choose 80 of the samples as the developmentset and 20 as the test set

51 Experimental Results Table 1 shows the recognition ratesand how many times LSTM DLSTM and LSTMP withdifferent parameters need to iterate until their recognitionrates are stable and the best results are in bold

In Table 1 ldquoLSTMP-80 to 20rdquo means the LSTMP modelprojecting 80 cell states to 20 cell states From Table 1 we see

Journal of Robotics 5

Table 1 The number of iterations and the recognition rates of different models

Model Number of iterations Recognition rateLSTM-20 300 772LSTM-40 300 871LSTM-80 300 907LSTM-160 400 825DLSTM-two layers 61 857DLSTM-three layers 71 995DLSTM-four layers 109 1000DLSTM-five layers 76 975DLSTM-six layers 103 1000LSTMP-10 to 20 43 56LSTMP-30 to 20 35 942LSTMP-40 to 20 45 960LSTMP-40 to 30 30 940LSTMP-60 to 30 22 980LSTMP-80 to 20 58 998LSTMP-80 to 30 27 973LSTMP-80 to 40 29 973LSTMP-160 to 40 39 990LSTMP-160 to 80 50 950LSTMP-320 to 160 93 934

LSTMP with different parametersDLSTM with different layers

LSTMP (30 to 20)LSTMP (40 to 20)LSTMP (60 to 30)

LSTMP (80 to 20)LSTMP (160 to 40)

35302520151050

00

02

04

06

08

10

250200150100500

00

02

04

06

08

10

LSTM_layer 1Deep LSTM_layer 3

Deep LSTM_layer 4Deep LSTM_layer 6

300

Figure 3 The recognition rates of LSTMP with different parameters and DLSTM with different layers

that DLSTM and LSTMP perform much better than LSTMand their best recognition rates are almost the same whichare 100 and 998 respectively As for the numbers ofiterations LSTMP needs much less iterations than LSTMand DLSTM which makes LSTMP more suitable for pianorsquoscontinuous note recognition for robotics considering theefficiency

52 LSTMP and DLSTM with Different Parameters Figure 3illustrates LSTMP with different parameters and DLSTMwith different layers The 119909 axis means the number of

iterations and the 119910 axis means the recognition rate We seethat for LSTMP the model projecting 80 cell states to 20 cellstates has the best result but all LSTMP results are very closeAs for DLSTM we see clearly that Deep LSTM is much betterthan LSTM with only one layer

53 Comparisons of LSTM LSTMP and DLSTM We com-pare LSTM LSTMP and DLSTM in Figure 4 Given the sameparameters LSTMP performs much better than LSTM Asfor LSTMP and DLSTM we find that when the number of

6 Journal of Robotics

LSTM and LSTMP LSTMP and DLSTM

Deep LSTM (4 layers)

6050403020100

00

02

04

06

08

10

LSTM (80)LSTMP (80 to 20)

LSTMP (80 to 20)

100806040200

00

02

04

06

08

10

Figure 4 Comparisons of LSTM LSTMP and DLSTM

iterations is small LSTMP has great advantages but as thenumber of iterations increases DLSTM becomes better

6 Conclusions and Future Work

In this paper we have used neural network structures calledLSTM with CTC to recognize continuous musical notes Onthe basis of LSTM we have also tried LSTMP and DLSTMAmong them LSTMP worked best when projecting 80 cellstates to 20 cell states which needed much less iterationsthan LSTM and DLSTM making it most suitable for pianorsquoscontinuous note recognition

In the future we will use LSTM LSTMP and DLSTMto recognize more complex continuous chord music suchas piano music violin pieces or even symphony which willgreatly improve the development of music robots

Conflicts of Interest

The authors declare that they have no conflicts of interest

Acknowledgments

Thanks are due to Yanlin Yang for collecting the record-ing materials This work is supported by the FundamentalResearch Funds for the Central Universities (no 2016JX06)and the National Natural Science Foundation of China (no61472369)

References

[1] A K Jain J Mao and K M Mohiuddin ldquoArtificial neural net-works a tutorialrdquo IEEE Computational Science amp Engineeringvol 29 no 3 pp 31ndash44 1996

[2] J R Zhang T M Lok andM R Lyu ldquoA hybrid particle swarmoptimization-back-propagation algorithm for feedforwardneu-ral network trainingrdquo Applied Mathematics and Computationvol 185 no 2 pp 1026ndash1037 2007

[3] L Liang Y Tang L Bin and W Xiaohua Artificial neuralnetwork based universal time series May 11 2004 US Patent6735580

[4] R J Williams and D Zipser ldquoA learning algorithm forcontinually running fully recurrent neural networksrdquo NeuralComputation vol 1 no 2 pp 270ndash280 1989

[5] L R Medsker and L C Jain ldquoRecurrent neural networksrdquoDesign and Applications vol 5 2001

[6] G Shalini O Vinyals B Strope R Scott T Dean and LHeck ldquoContextual lstm (clstm)models for large scale nlp tasksrdquohttpsarxivorgabs160206291

[7] A Karpathy The unreasonable effectiveness of recurrentneural networks httpkarpathygithubio20150521rnn-effectiveness 2015

[8] D Turnbull and C Elkan ldquoFast recognition of musical genresusingRBFnetworksrdquo IEEETransactions onKnowledge andDataEngineering vol 17 no 4 pp 580ndash584 2005

[9] H Sak A Senior and F Beaufays ldquoLong short-term memoryrecurrent neural network architectures for large scale acousticmodelingrdquo in Proceedings of the 15th Annual Conference of theInternational Speech Communication Association Celebratingthe Diversity of Spoken Languages INTERSPEECH 2014 pp338ndash342 September 2014

[10] AGraves A-RMohamed andGHinton ldquoSpeech recognitionwith deep recurrent neural networksrdquo in Proceedings of the 38thIEEE International Conference on Acoustics Speech and SignalProcessing (ICASSP rsquo13) pp 6645ndash6649 May 2013

[11] J Schmidhuber ldquoDeep learning in neural networks anoverviewrdquo Neural Networks vol 61 pp 85ndash117 2015

[12] K M Hermann T Kocisky E Grefenstette et al ldquoTeachingmachines to read and comprehendrdquo in Proceedings of the 29thAnnual Conference on Neural Information Processing SystemsNIPS 2015 pp 1693ndash1701 December 2015

[13] B Bakker ldquoReinforcement learning with long short-termmem-oryrdquo in In Advances in Neural Information Processing Systemspp 1475ndash1482

[14] S Hochreiter and J Schmidhuber ldquoLong short-term memoryrdquoNeural Computation vol 9 no 8 pp 1735ndash1780 1997

Journal of Robotics 7

[15] P Goelet V F Castellucci S Schacher and E R KandelldquoThe long and the short of long-term memory - A molecularframeworkrdquo Nature vol 322 no 6078 pp 419ndash422 1986

[16] Q Lyu Z Wu and J Zhu ldquoPolyphonic music modelling withLSTM-RTRBMrdquo in Proceedings of the 23rd ACM InternationalConference onMultimedia MM 2015 pp 991ndash994 aus October2015

[17] S Haim A Senior R Kanishka and F Beaufays ldquoFast andaccurate recurrent neural network acoustic models for speechrecognitionrdquo httpsarxivorgabs150706947

[18] F A Gers J Schmidhuber and F Cummins ldquoLearning toforget Continual prediction with LSTMrdquo Neural Computationvol 12 no 10 pp 2451ndash2471 2000

[19] F A Gers N N Schraudolph and J Schmidhuber ldquoLearningprecise timing with LSTM recurrent networksrdquo Journal ofMachine Learning Research (JMLR) vol 3 no 1 pp 115ndash1432003

[20] A Graves S Fernandez F Gomez and J SchmidhuberldquoConnectionist temporal classification Labelling unsegmentedsequence data with recurrent neural networksrdquo in Proceedingsof the 23rd International Conference onMachine Learning ICML2006 pp 369ndash376 Pittsburgh Pa USA June 2006

RoboticsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Active and Passive Electronic Components

Control Scienceand Engineering

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

RotatingMachinery

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporation httpwwwhindawicom

Journal of

Volume 201

Submit your manuscripts athttpswwwhindawicom

VLSI Design

Hindawi Publishing Corporationhttpwwwhindawicom Volume 201

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Shock and Vibration

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawi Publishing Corporation httpwwwhindawicom

Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

SensorsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Modelling amp Simulation in EngineeringHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Navigation and Observation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

DistributedSensor Networks

International Journal of

Page 5: Long Short-Term Memory Projection Recurrent Neural Network

Journal of Robotics 5

Table 1 The number of iterations and the recognition rates of different models

Model Number of iterations Recognition rateLSTM-20 300 772LSTM-40 300 871LSTM-80 300 907LSTM-160 400 825DLSTM-two layers 61 857DLSTM-three layers 71 995DLSTM-four layers 109 1000DLSTM-five layers 76 975DLSTM-six layers 103 1000LSTMP-10 to 20 43 56LSTMP-30 to 20 35 942LSTMP-40 to 20 45 960LSTMP-40 to 30 30 940LSTMP-60 to 30 22 980LSTMP-80 to 20 58 998LSTMP-80 to 30 27 973LSTMP-80 to 40 29 973LSTMP-160 to 40 39 990LSTMP-160 to 80 50 950LSTMP-320 to 160 93 934

LSTMP with different parametersDLSTM with different layers

LSTMP (30 to 20)LSTMP (40 to 20)LSTMP (60 to 30)

LSTMP (80 to 20)LSTMP (160 to 40)

35302520151050

00

02

04

06

08

10

250200150100500

00

02

04

06

08

10

LSTM_layer 1Deep LSTM_layer 3

Deep LSTM_layer 4Deep LSTM_layer 6

300

Figure 3 The recognition rates of LSTMP with different parameters and DLSTM with different layers

that DLSTM and LSTMP perform much better than LSTMand their best recognition rates are almost the same whichare 100 and 998 respectively As for the numbers ofiterations LSTMP needs much less iterations than LSTMand DLSTM which makes LSTMP more suitable for pianorsquoscontinuous note recognition for robotics considering theefficiency

52 LSTMP and DLSTM with Different Parameters Figure 3illustrates LSTMP with different parameters and DLSTMwith different layers The 119909 axis means the number of

iterations and the 119910 axis means the recognition rate We seethat for LSTMP the model projecting 80 cell states to 20 cellstates has the best result but all LSTMP results are very closeAs for DLSTM we see clearly that Deep LSTM is much betterthan LSTM with only one layer

53 Comparisons of LSTM LSTMP and DLSTM We com-pare LSTM LSTMP and DLSTM in Figure 4 Given the sameparameters LSTMP performs much better than LSTM Asfor LSTMP and DLSTM we find that when the number of

6 Journal of Robotics

LSTM and LSTMP LSTMP and DLSTM

Deep LSTM (4 layers)

6050403020100

00

02

04

06

08

10

LSTM (80)LSTMP (80 to 20)

LSTMP (80 to 20)

100806040200

00

02

04

06

08

10

Figure 4 Comparisons of LSTM LSTMP and DLSTM

iterations is small LSTMP has great advantages but as thenumber of iterations increases DLSTM becomes better

6 Conclusions and Future Work

In this paper we have used neural network structures calledLSTM with CTC to recognize continuous musical notes Onthe basis of LSTM we have also tried LSTMP and DLSTMAmong them LSTMP worked best when projecting 80 cellstates to 20 cell states which needed much less iterationsthan LSTM and DLSTM making it most suitable for pianorsquoscontinuous note recognition

In the future we will use LSTM LSTMP and DLSTMto recognize more complex continuous chord music suchas piano music violin pieces or even symphony which willgreatly improve the development of music robots

Conflicts of Interest

The authors declare that they have no conflicts of interest

Acknowledgments

Thanks are due to Yanlin Yang for collecting the record-ing materials This work is supported by the FundamentalResearch Funds for the Central Universities (no 2016JX06)and the National Natural Science Foundation of China (no61472369)

References

[1] A K Jain J Mao and K M Mohiuddin ldquoArtificial neural net-works a tutorialrdquo IEEE Computational Science amp Engineeringvol 29 no 3 pp 31ndash44 1996

[2] J R Zhang T M Lok andM R Lyu ldquoA hybrid particle swarmoptimization-back-propagation algorithm for feedforwardneu-ral network trainingrdquo Applied Mathematics and Computationvol 185 no 2 pp 1026ndash1037 2007

[3] L Liang Y Tang L Bin and W Xiaohua Artificial neuralnetwork based universal time series May 11 2004 US Patent6735580

[4] R J Williams and D Zipser ldquoA learning algorithm forcontinually running fully recurrent neural networksrdquo NeuralComputation vol 1 no 2 pp 270ndash280 1989

[5] L R Medsker and L C Jain ldquoRecurrent neural networksrdquoDesign and Applications vol 5 2001

[6] G Shalini O Vinyals B Strope R Scott T Dean and LHeck ldquoContextual lstm (clstm)models for large scale nlp tasksrdquohttpsarxivorgabs160206291

[7] A Karpathy The unreasonable effectiveness of recurrentneural networks httpkarpathygithubio20150521rnn-effectiveness 2015

[8] D Turnbull and C Elkan ldquoFast recognition of musical genresusingRBFnetworksrdquo IEEETransactions onKnowledge andDataEngineering vol 17 no 4 pp 580ndash584 2005

[9] H Sak A Senior and F Beaufays ldquoLong short-term memoryrecurrent neural network architectures for large scale acousticmodelingrdquo in Proceedings of the 15th Annual Conference of theInternational Speech Communication Association Celebratingthe Diversity of Spoken Languages INTERSPEECH 2014 pp338ndash342 September 2014

[10] AGraves A-RMohamed andGHinton ldquoSpeech recognitionwith deep recurrent neural networksrdquo in Proceedings of the 38thIEEE International Conference on Acoustics Speech and SignalProcessing (ICASSP rsquo13) pp 6645ndash6649 May 2013

[11] J Schmidhuber ldquoDeep learning in neural networks anoverviewrdquo Neural Networks vol 61 pp 85ndash117 2015

[12] K M Hermann T Kocisky E Grefenstette et al ldquoTeachingmachines to read and comprehendrdquo in Proceedings of the 29thAnnual Conference on Neural Information Processing SystemsNIPS 2015 pp 1693ndash1701 December 2015

[13] B Bakker ldquoReinforcement learning with long short-termmem-oryrdquo in In Advances in Neural Information Processing Systemspp 1475ndash1482

[14] S Hochreiter and J Schmidhuber ldquoLong short-term memoryrdquoNeural Computation vol 9 no 8 pp 1735ndash1780 1997

Journal of Robotics 7

[15] P Goelet V F Castellucci S Schacher and E R KandelldquoThe long and the short of long-term memory - A molecularframeworkrdquo Nature vol 322 no 6078 pp 419ndash422 1986

[16] Q Lyu Z Wu and J Zhu ldquoPolyphonic music modelling withLSTM-RTRBMrdquo in Proceedings of the 23rd ACM InternationalConference onMultimedia MM 2015 pp 991ndash994 aus October2015

[17] S Haim A Senior R Kanishka and F Beaufays ldquoFast andaccurate recurrent neural network acoustic models for speechrecognitionrdquo httpsarxivorgabs150706947

[18] F A Gers J Schmidhuber and F Cummins ldquoLearning toforget Continual prediction with LSTMrdquo Neural Computationvol 12 no 10 pp 2451ndash2471 2000

[19] F A Gers N N Schraudolph and J Schmidhuber ldquoLearningprecise timing with LSTM recurrent networksrdquo Journal ofMachine Learning Research (JMLR) vol 3 no 1 pp 115ndash1432003

[20] A Graves S Fernandez F Gomez and J SchmidhuberldquoConnectionist temporal classification Labelling unsegmentedsequence data with recurrent neural networksrdquo in Proceedingsof the 23rd International Conference onMachine Learning ICML2006 pp 369ndash376 Pittsburgh Pa USA June 2006

RoboticsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Active and Passive Electronic Components

Control Scienceand Engineering

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

RotatingMachinery

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporation httpwwwhindawicom

Journal of

Volume 201

Submit your manuscripts athttpswwwhindawicom

VLSI Design

Hindawi Publishing Corporationhttpwwwhindawicom Volume 201

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Shock and Vibration

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawi Publishing Corporation httpwwwhindawicom

Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

SensorsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Modelling amp Simulation in EngineeringHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Navigation and Observation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

DistributedSensor Networks

International Journal of

Page 6: Long Short-Term Memory Projection Recurrent Neural Network

6 Journal of Robotics

LSTM and LSTMP LSTMP and DLSTM

Deep LSTM (4 layers)

6050403020100

00

02

04

06

08

10

LSTM (80)LSTMP (80 to 20)

LSTMP (80 to 20)

100806040200

00

02

04

06

08

10

Figure 4 Comparisons of LSTM LSTMP and DLSTM

iterations is small LSTMP has great advantages but as thenumber of iterations increases DLSTM becomes better

6 Conclusions and Future Work

In this paper we have used neural network structures calledLSTM with CTC to recognize continuous musical notes Onthe basis of LSTM we have also tried LSTMP and DLSTMAmong them LSTMP worked best when projecting 80 cellstates to 20 cell states which needed much less iterationsthan LSTM and DLSTM making it most suitable for pianorsquoscontinuous note recognition

In the future we will use LSTM LSTMP and DLSTMto recognize more complex continuous chord music suchas piano music violin pieces or even symphony which willgreatly improve the development of music robots

Conflicts of Interest

The authors declare that they have no conflicts of interest

Acknowledgments

Thanks are due to Yanlin Yang for collecting the record-ing materials This work is supported by the FundamentalResearch Funds for the Central Universities (no 2016JX06)and the National Natural Science Foundation of China (no61472369)

References

[1] A K Jain J Mao and K M Mohiuddin ldquoArtificial neural net-works a tutorialrdquo IEEE Computational Science amp Engineeringvol 29 no 3 pp 31ndash44 1996

[2] J R Zhang T M Lok andM R Lyu ldquoA hybrid particle swarmoptimization-back-propagation algorithm for feedforwardneu-ral network trainingrdquo Applied Mathematics and Computationvol 185 no 2 pp 1026ndash1037 2007

[3] L Liang Y Tang L Bin and W Xiaohua Artificial neuralnetwork based universal time series May 11 2004 US Patent6735580

[4] R J Williams and D Zipser ldquoA learning algorithm forcontinually running fully recurrent neural networksrdquo NeuralComputation vol 1 no 2 pp 270ndash280 1989

[5] L R Medsker and L C Jain ldquoRecurrent neural networksrdquoDesign and Applications vol 5 2001

[6] G Shalini O Vinyals B Strope R Scott T Dean and LHeck ldquoContextual lstm (clstm)models for large scale nlp tasksrdquohttpsarxivorgabs160206291

[7] A Karpathy The unreasonable effectiveness of recurrentneural networks httpkarpathygithubio20150521rnn-effectiveness 2015

[8] D Turnbull and C Elkan ldquoFast recognition of musical genresusingRBFnetworksrdquo IEEETransactions onKnowledge andDataEngineering vol 17 no 4 pp 580ndash584 2005

[9] H Sak A Senior and F Beaufays ldquoLong short-term memoryrecurrent neural network architectures for large scale acousticmodelingrdquo in Proceedings of the 15th Annual Conference of theInternational Speech Communication Association Celebratingthe Diversity of Spoken Languages INTERSPEECH 2014 pp338ndash342 September 2014

[10] AGraves A-RMohamed andGHinton ldquoSpeech recognitionwith deep recurrent neural networksrdquo in Proceedings of the 38thIEEE International Conference on Acoustics Speech and SignalProcessing (ICASSP rsquo13) pp 6645ndash6649 May 2013

[11] J Schmidhuber ldquoDeep learning in neural networks anoverviewrdquo Neural Networks vol 61 pp 85ndash117 2015

[12] K M Hermann T Kocisky E Grefenstette et al ldquoTeachingmachines to read and comprehendrdquo in Proceedings of the 29thAnnual Conference on Neural Information Processing SystemsNIPS 2015 pp 1693ndash1701 December 2015

[13] B Bakker ldquoReinforcement learning with long short-termmem-oryrdquo in In Advances in Neural Information Processing Systemspp 1475ndash1482

[14] S Hochreiter and J Schmidhuber ldquoLong short-term memoryrdquoNeural Computation vol 9 no 8 pp 1735ndash1780 1997

Journal of Robotics 7

[15] P Goelet V F Castellucci S Schacher and E R KandelldquoThe long and the short of long-term memory - A molecularframeworkrdquo Nature vol 322 no 6078 pp 419ndash422 1986

[16] Q Lyu Z Wu and J Zhu ldquoPolyphonic music modelling withLSTM-RTRBMrdquo in Proceedings of the 23rd ACM InternationalConference onMultimedia MM 2015 pp 991ndash994 aus October2015

[17] S Haim A Senior R Kanishka and F Beaufays ldquoFast andaccurate recurrent neural network acoustic models for speechrecognitionrdquo httpsarxivorgabs150706947

[18] F A Gers J Schmidhuber and F Cummins ldquoLearning toforget Continual prediction with LSTMrdquo Neural Computationvol 12 no 10 pp 2451ndash2471 2000

[19] F A Gers N N Schraudolph and J Schmidhuber ldquoLearningprecise timing with LSTM recurrent networksrdquo Journal ofMachine Learning Research (JMLR) vol 3 no 1 pp 115ndash1432003

[20] A Graves S Fernandez F Gomez and J SchmidhuberldquoConnectionist temporal classification Labelling unsegmentedsequence data with recurrent neural networksrdquo in Proceedingsof the 23rd International Conference onMachine Learning ICML2006 pp 369ndash376 Pittsburgh Pa USA June 2006

RoboticsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Active and Passive Electronic Components

Control Scienceand Engineering

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

RotatingMachinery

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporation httpwwwhindawicom

Journal of

Volume 201

Submit your manuscripts athttpswwwhindawicom

VLSI Design

Hindawi Publishing Corporationhttpwwwhindawicom Volume 201

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Shock and Vibration

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawi Publishing Corporation httpwwwhindawicom

Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

SensorsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Modelling amp Simulation in EngineeringHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Navigation and Observation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

DistributedSensor Networks

International Journal of

Page 7: Long Short-Term Memory Projection Recurrent Neural Network

Journal of Robotics 7

[15] P Goelet V F Castellucci S Schacher and E R KandelldquoThe long and the short of long-term memory - A molecularframeworkrdquo Nature vol 322 no 6078 pp 419ndash422 1986

[16] Q Lyu Z Wu and J Zhu ldquoPolyphonic music modelling withLSTM-RTRBMrdquo in Proceedings of the 23rd ACM InternationalConference onMultimedia MM 2015 pp 991ndash994 aus October2015

[17] S Haim A Senior R Kanishka and F Beaufays ldquoFast andaccurate recurrent neural network acoustic models for speechrecognitionrdquo httpsarxivorgabs150706947

[18] F A Gers J Schmidhuber and F Cummins ldquoLearning toforget Continual prediction with LSTMrdquo Neural Computationvol 12 no 10 pp 2451ndash2471 2000

[19] F A Gers N N Schraudolph and J Schmidhuber ldquoLearningprecise timing with LSTM recurrent networksrdquo Journal ofMachine Learning Research (JMLR) vol 3 no 1 pp 115ndash1432003

[20] A Graves S Fernandez F Gomez and J SchmidhuberldquoConnectionist temporal classification Labelling unsegmentedsequence data with recurrent neural networksrdquo in Proceedingsof the 23rd International Conference onMachine Learning ICML2006 pp 369ndash376 Pittsburgh Pa USA June 2006

RoboticsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Active and Passive Electronic Components

Control Scienceand Engineering

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

RotatingMachinery

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporation httpwwwhindawicom

Journal of

Volume 201

Submit your manuscripts athttpswwwhindawicom

VLSI Design

Hindawi Publishing Corporationhttpwwwhindawicom Volume 201

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Shock and Vibration

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawi Publishing Corporation httpwwwhindawicom

Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

SensorsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Modelling amp Simulation in EngineeringHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Navigation and Observation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

DistributedSensor Networks

International Journal of

Page 8: Long Short-Term Memory Projection Recurrent Neural Network

RoboticsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Active and Passive Electronic Components

Control Scienceand Engineering

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

RotatingMachinery

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporation httpwwwhindawicom

Journal of

Volume 201

Submit your manuscripts athttpswwwhindawicom

VLSI Design

Hindawi Publishing Corporationhttpwwwhindawicom Volume 201

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Shock and Vibration

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Civil EngineeringAdvances in

Acoustics and VibrationAdvances in

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Electrical and Computer Engineering

Journal of

Advances inOptoElectronics

Hindawi Publishing Corporation httpwwwhindawicom

Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

SensorsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Modelling amp Simulation in EngineeringHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Chemical EngineeringInternational Journal of Antennas and

Propagation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Navigation and Observation

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

DistributedSensor Networks

International Journal of