Please wait a minute...
Frontiers of Computer Science

ISSN 2095-2228

ISSN 2095-2236(Online)

CN 10-1014/TP

Postal Subscription Code 80-970

2018 Impact Factor: 1.129

Front. Comput. Sci.    2016, Vol. 10 Issue (1) : 96-112    https://doi.org/10.1007/s11704-015-4478-2
RESEARCH ARTICLE
Exploiting multi-channels deep convolutional neural networks for multivariate time series classification
Yi ZHENG1,3,Qi LIU1,Enhong CHEN1,*(),Yong GE2,J. Leon ZHAO3
1. School of Computer Science and Technology, University of Science and Technology of China, Hefei 230027, China
2. Department of Computer Science, University of North Carolina at Charlotte, Charlotte 28223, USA
3. Department of Information Systems, City University of Hong Kong, Hong Kong, China
 Download: PDF(779 KB)  
 Export: BibTeX | EndNote | Reference Manager | ProCite | RefWorks
Abstract

Time series classification is related to many different domains, such as health informatics, finance, and bioinformatics. Due to its broad applications, researchers have developed many algorithms for this kind of tasks, e.g., multivariate time series classification. Among the classification algorithms, k-nearest neighbor (k-NN) classification (particularly 1-NN) combined with dynamic time warping (DTW) achieves the state of the art performance. The deficiency is that when the data set grows large, the time consumption of 1-NN with DTWwill be very expensive. In contrast to 1-NN with DTW, it is more efficient but less effective for feature-based classification methods since their performance usually depends on the quality of hand-crafted features. In this paper, we aim to improve the performance of traditional feature-based approaches through the feature learning techniques. Specifically, we propose a novel deep learning framework, multi-channels deep convolutional neural networks (MC-DCNN), for multivariate time series classification. This model first learns features from individual univariate time series in each channel, and combines information from all channels as feature representation at the final layer. Then, the learnt features are applied into a multilayer perceptron (MLP) for classification. Finally, the extensive experiments on real-world data sets show that our model is not only more efficient than the state of the art but also competitive in accuracy. This study implies that feature learning is worth to be investigated for the problem of time series classification.

Keywords convolutional neural networks      time series classification      feature learning      deep learning     
Corresponding Author(s): Enhong CHEN   
Just Accepted Date: 05 June 2015   Issue Date: 06 January 2016
 Cite this article:   
Yi ZHENG,Qi LIU,Enhong CHEN, et al. Exploiting multi-channels deep convolutional neural networks for multivariate time series classification[J]. Front. Comput. Sci., 2016, 10(1): 96-112.
 URL:  
https://academic.hep.com.cn/fcs/EN/10.1007/s11704-015-4478-2
https://academic.hep.com.cn/fcs/EN/Y2016/V10/I1/96
1 Xing Z, Pei J, Keogh E. A brief survey on sequence classification.ACM SIGKDD Explorations Newsletter, 2010, 12(1): 40–48
https://doi.org/10.1145/1882471.1882478
2 Ding H, Trajcevski G, Scheuermann P, Wang X, Keogh E. Querying and mining of time series data: experimental comparison of representations and distance measures. Proceedings of the VLDB Endowment,2008, 1(2): 1542–1552
https://doi.org/10.14778/1454159.1454226
3 Orsenigo C, Vercellis C. Combining discrete svm and fixed cardinality warping distances for multivariate time series classification. Pattern Recognition, 2010, 43(11): 3787–3794
https://doi.org/10.1016/j.patcog.2010.06.005
4 Batal I, Sacchi L, Bellazzi R, Hauskrecht M. Multivariate time series classification with temporal abstractions. In: Proceedings of FLAIRS Conference. 2009
5 Haselsteiner E, Pfurtscheller G. Using time-dependent neural networks for EEG classification. IEEE Transactions on Rehabilitation Engineering,2000, 8(4): 457–463
https://doi.org/10.1109/86.895948
6 Kampouraki A, Manis G, Nikou C. Heartbeat time series classification with support vector machines. IEEE Transactions on Information Technology in Biomedicine, 2009, 13(4): 512–518
https://doi.org/10.1109/TITB.2008.2003323
7 Reiss A, Stricker D. Introducing a modular activity monitoring system.In: Proceedings of IEEE Annual International Conference on Engineering in Medicine and Biology Society. 2011, 5621–5624
https://doi.org/10.1109/iembs.2011.6091360
8 Batista G E A P A, Wang X, Keogh E J. A complexity-invariant distance measure for time series. In: Proceedings of SIAM Conference on Data Mining. 2011
https://doi.org/10.1137/1.9781611972818.60
9 Rakthanmanon T, Campana B, Mueen A, Batista G, Westover B, Zhu Q, Zakaria J, Keogh E. Searching and mining trillions of time series subsequences under dynamic time warping. In: Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 2012, 262–270
https://doi.org/10.1145/2339530.2339576
10 Xi X, Keogh E J, Shelton C R, Wei L, Ratanamahatana C A. Fast time series classification using numerosity reduction. In: Proceedings of the 23rd International Conference on Machine Learning. 2006, 1033–1040
https://doi.org/10.1145/1143844.1143974
11 Bengio Y, Courville A, Vincent P. Representation learning: a review and new perspectives. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2013, 35(8): 1798–1828
https://doi.org/10.1109/TPAMI.2013.50
12 LeCun Y, Bengio Y. Convolutional networks for images, speech, and time series. The Handbook of Brain Theory and Neural Networks,1995, 3361(10)
13 LeCun Y, Kavukcuoglu K, Farabet C. Convolutional networks and applications in vision. In: Proceedings of IEEE International Symposium on Circuits and Systems. 2010, 253–256
https://doi.org/10.1109/iscas.2010.5537907
14 Zheng Y, Liu Q, Chen E, Ge Y, Zhao J. Time series classification using multi-channels deep convolutional neural networks. In: Proceedings of the 15th International Conference on Web-Age Information Management.2014, 298–310
https://doi.org/10.1007/978-3-319-08010-9_33
15 Hu B, Chen Y, Keogh E. Time Series Classification under More Realistic Assumptions. In: Proceedings of SIAM International Conference on Data Mining. 2013, 578
https://doi.org/10.1137/1.9781611972832.64
16 Goldberger A L, Amaral L A N, Glass L, Hausdorff J M, Ivanov P C,Mark R G, Mietus J E, Moody G B, Peng C K, Stanley H E. Physiobank,Physiotoolkit, and Physionet omponents of a new research resource for complex physiologic signals. Circulation, 2000, 101(23):e215–e220
https://doi.org/10.1161/01.CIR.101.23.e215
17 Ye L, Keogh E. Time series shapelets: a new primitive for data mining.In: Proceedings of the 15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 2009, 947–956
https://doi.org/10.1145/1557019.1557122
18 Ratanamahatana C A, Keogh E.Making time-series classification more accurate using learned constraints. In: Proceedings of SIAM International Conference on Data Mining. 2004
https://doi.org/10.1137/1.9781611972740.2
19 Ratanamahatana C A, Keogh E. Three myths about dynamic time warping data mining. In: Proceedings of SIAM International Conference on Data Mining. 2005, 506–510
https://doi.org/10.1137/1.9781611972757.50
20 Yu D, Yu X, Hu Q, Liu J, Wu A. Dynamic time warping constraint learning for large margin nearest neighbor classification. Information Sciences, 2011, 181(13): 2787–2796
https://doi.org/10.1016/j.ins.2011.03.001
21 LeCun Y, Bottou L, Bengio Y, Haffner P. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 1998, 86(11):2278–2324
https://doi.org/10.1109/5.726791
22 Simard P Y, Steinkraus D, Platt J C. Best practices for convolutional neural networks applied to visual document analysis. In: Proceedings of the 7th International Conference on Document Analysis and Recognition.2003, 958–962
https://doi.org/10.1109/icdar.2003.1227801
23 Nair V, Hinton G E. Rectified linear units improve restricted boltzmann machines. In: Proceedings of the 27th International Conference onMachine Learning. 2010, 807–814
24 Zeiler M D, Ranzato M, Monga R, Mao M, Yang K, Le Q, Nguyen P,Senior A, Vanhoucke V, Dean J, Hinton G E. On rectified linear units for speech processing. In: Proceedings of IEEE International Conference on Acoustics, Speech and Signal Processing. 2013, 3517–3521
https://doi.org/10.1109/icassp.2013.6638312
25 Scherer D, Müller A, Behnke S. Evaluation of pooling operations in convolutional architectures for object recognition. In: Proceedings of the 20th International Conference on Artificial Neural Networks. 2010,92–101
https://doi.org/10.1007/978-3-642-15825-4_10
26 Nagi J, Ducatelle F, Di Caro G A, Ciresan D,Meier U, Giusti A, Nagi F,Schmidhuber J, Gambardella L M. Max-pooling convolutional neural networks for vision-based hand gesture recognition. In: Proceedings of IEEE International Conference on Signal and Image Processing Applications.2011, 342–347
https://doi.org/10.1109/icsipa.2011.6144164
27 LeCun Y, Bottou L, Orr G B, Müller K R. Efficient backprop. Lecture Notes in Computer Science, 2012, 7700: 9–48
https://doi.org/10.1007/978-3-642-35289-8_3
28 Bouvrie J. Notes on convolutional neural networks. Technical Report.2006
29 Krizhevsky A, Sutskever I, Hinton G. Imagenet classification with deep convolutional neural networks. In: Proceedings of Advances in Neural Information Processing Systems. 2012, 1106–1114
30 Sutskever I, Martens J, Dahl G, Hinton G. On the importance of initialization and momentum in deep learning. In: Proceedings of the 30th International Conference on Machine Learning. 2013, 1139–1147
31 Erhan D, Bengio Y, Courville A, Manzagol P A, Vincent P, Bengio S.Why does unsupervised pre-training help deep learning? The Journal of Machine Learning Research, 2010, 11: 625–660
32 Hinton G E, Salakhutdinov R R. Reducing the dimensionality of data with neural networks. Science, 2006, 313(5786): 504–507
https://doi.org/10.1126/science.1127647
33 Masci J, Meier U, Cire¸ssan D, Schmidhuber J. Stacked convolutional auto-encoders for hierarchical feature extraction. Lecture Notes in Computer Science, 2011, 6791: 52–59
https://doi.org/10.1007/978-3-642-21735-7_7
34 Pinto N, Cox D D, DiCarlo J J. Why is real-world visual object recognition hard? PLoS Computational Biology, 2008, 4(1): e27
https://doi.org/10.1371/journal.pcbi.0040027
35 Cire¸ssan D C, Meier U, Masci J, Gambardella L M, Schmidhuber J.Flexible, high performance convolutional neural networks for image classification. In: Proceedings of the 22nd International Joint Conference on Artificial Intelligence. 2011, 1237–1242
36 Cire¸ssan D, Meier U, Masci J, Schmidhuber J. Multi-column deep neural network for traffic sign classification. Neural Networks, 2012, 32:333–338
https://doi.org/10.1016/j.neunet.2012.02.023
37 Lines J, Davis L M, Hills J, Bagnall A. A shapelet transform for time series classification. In: Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 2012,289–297
https://doi.org/10.1145/2339530.2339579
38 Nanopoulos A, Alcock R O B, Manolopoulos Y. Feature-based classification of time-series data. International Journal of Computer Research,2001, 10(3)
39 Lee H, Grosse R, Ranganath R, Ng A Y. Convolutional deep belief networks for scalable unsupervised learning of hierarchical representations.In: Proceedings of the 26th Annual International Conference on Machine Learning. 2009, 609–616
https://doi.org/10.1145/1553374.1553453
40 Lee H, Largman Y, Pham P, Ng A Y. Unsupervised feature learning for audio classification using convolutional deep belief networks. In:Proceedings of Advances in Neural Information Processing Systems.2009, 1096–1104
41 Waibel A, Hanazawa T, Hinton G, Shikano K, Lang K J. Phoneme recognition using time-delay neural networks. IEEE Transactions on Acoustics, Speech and Signal Processing, 1989, 37(3): 328–339
https://doi.org/10.1109/29.21701
[1] Supplementary Material-Highlights in 3-page ppt Download
[1] Huiying ZHANG, Yu ZHANG, Xin GENG. Practical age estimation using deep label distribution learning[J]. Front. Comput. Sci., 2021, 15(3): 153318-.
[2] Syed Farooq ALI, Muhammad Aamir KHAN, Ahmed Sohail ASLAM. Fingerprint matching, spoof and liveness detection: classification and literature review[J]. Front. Comput. Sci., 2021, 15(1): 151310-.
[3] Chune LI, Yongyi MAO, Richong ZHANG, Jinpeng HUAI. A revisit to MacKay algorithm and its application to deep network compression[J]. Front. Comput. Sci., 2020, 14(4): 144304-.
[4] Guijuan ZHANG, Yang LIU, Xiaoning JIN. A survey of autoencoder-based recommender systems[J]. Front. Comput. Sci., 2020, 14(2): 430-450.
[5] Anna ZHU, Seiichi UCHIDA. Scene word recognition from pieces to whole[J]. Front. Comput. Sci., 2019, 13(2): 292-301.
[6] Jun ZHANG, Bineng ZHONG, Pengfei WANG, Cheng WANG, Jixiang DU. Robust feature learning for online discriminative tracking without large-scale pre-training[J]. Front. Comput. Sci., 2018, 12(6): 1160-1172.
[7] Qianjun ZHANG, Lei ZHANG. Convolutional adaptive denoising autoencoders for hierarchical feature extraction[J]. Front. Comput. Sci., 2018, 12(6): 1140-1148.
[8] Lili HUANG, Jiefeng PENG, Ruimao ZHANG, Guanbin LI, Liang LIN. Learning deep representations for semantic image parsing: a comprehensive overview[J]. Front. Comput. Sci., 2018, 12(5): 840-857.
[9] Xin LIU,Meina KAN,Wanglong WU,Shiguang SHAN,Xilin CHEN. VIPLFaceNet: an open source deep face recognition SDK[J]. Front. Comput. Sci., 2017, 11(2): 208-218.
[10] Feifei ZHANG,Yongbin YU,Qirong MAO,Jianping GOU,Yongzhao ZHAN. Pose-robust feature learning for facial expression recognition[J]. Front. Comput. Sci., 2016, 10(5): 832-844.
[11] Wenge RONG,Baolin PENG,Yuanxin OUYANG,Chao LI,Zhang XIONG. Structural information aware deep semi-supervised recurrent neural network for sentiment analysis[J]. Front. Comput. Sci., 2015, 9(2): 171-184.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed