Abstract
Over the past decade, time series clustering has become an increasingly important research topic in data mining community. Most existing methods for time series clustering rely on distances calculated from the entire raw data using the Euclidean distance or Dynamic Time Warping distance as the distance measure. However, the presence of significant noise, dropouts, or extraneous data can greatly limit the accuracy of clustering in this domain. Moreover, for most real world problems, we cannot expect objects from the same class to be equal in length. As a consequence, most work on time series clustering only considers the clustering of individual time series “behaviors,” e.g., individual heart beats or individual gait cycles, and contrives the time series in some way to make them all equal in length. However, automatically formatting the data in such a way is often a harder problem than the clustering itself. In this work, we show that by using only some local patterns and deliberately ignoring the rest of the data, we can mitigate the above problems and cluster time series of different lengths, e.g., cluster one heartbeat with multiple heartbeats. To achieve this, we exploit and extend a recently introduced concept in time series data mining called shapelets. Unlike existing work, our work demonstrates the unintuitive fact that shapelets can be learned from unlabeled time series. We show, with extensive empirical evaluation in diverse domains, that our method is more accurate than existing methods. Moreover, in addition to accurate clustering results, we show that our work also has the potential to give insight into the domains to which it is applied. While a brute-force algorithm to discover shapelets in an unsupervised way could be untenably slow, we introduce two novel optimization procedures to significantly speed up the unsupervised-shapelet discovery process and allow it to be cast as an anytime algorithm.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Andino SLG et al (2000) Measuring the complexity of time series: an application to neurophysiological signals. Human Brain Mapp 11:46–57
Aziz W, Arif M (2006) Complexity analysis of stride interval time series by threshold dependent symbolic entropy. Euro J Appl Phys 98:30–40
Batista G, Wang X, Keogh E (2011) A complexity-invariant distance measure for time seies. In: SDM
Cerra D, Bieniarz J, Avbelj J, Reinartz P, Mueller R (2011) Compression-based unsupervised clustering of spectral signatures. In: Hyperspectral image and signal processing: evolution in remote sensing (WHISPERS 2011)
Ding H, Trajcevski G, Scheuermann P, Wang X, Keogh E (2008) Querying and mining of time series data: experimental comparison of representations and distance measures. Proc VLDB Endow 1:1542–52
Garilov M, Anguelov D, Indyk P, Motwani R (2000) Mining the stock market: which measure is best? In: Proceedings of the ACM KDD
Goldberger AL et al (1997) PhysioBank, PhysioToolkit, and PhysioNet: circulation. Discovery 101:1
Halkidi M, Batistakis Y, Vazirgiannis M (2001) On clustering validation techniques. J Intell Inform Syst 17:107–145
Hartmann B, Schwab I, Link N (2010) Prototype optimization for temporarily and spatially distorted time series. In: AAAI spring symposium series (SSS 2010), pp 15–20
Hirano S, Tsumoto S (2006) Cluster analysis of time-series medical data based on the trajectory representation and multiscale comparison techniques. In: Proceedings of the ICDM
Hu B, Chen Y, Keogh E (2013) Time series classification under more realistic assumptions. In: SDM, pp 578–586
Hu B, Rakthanmanon T, Hao Y, Evans S, Lonardi S, Keogh E (2011) Discovering the intrinsic cardinality and dimensionality of time series using MDL. In: Proceedings of the ICDM
Jiang D, Tang C, Zhang A (2004) Cluster analysis for gene expression data: a survey. IEEE Trans. Knowl Data Eng 16(11):1370–1386
Jesin’s Webpage (2013) https://sites.google.com/a/ucr.edu/clusteringtsusingushapelet/. Accessed 19 April 2015
Kalpakis K, Gada D, Puttagunta V (2001) Distance measures for effective clustering of ARIMA time-series. In: ICDM
Keogh E, Kasetty S (2002) On the need for time series data mining benchmarks: a survey and empirical demonstration. In: Proceedings of the ACM KDD, pp 102–111
Keogh E, Lin J, Truppel W (2003) Clustering of time series subsequences is meaningless: implications for past and future research. In: Proceedings of the IEEE ICDM, pp 115–122
Keogh E, Zhu Q, Hu B, Hao Y, Xi X, Wei L, Ratanamahatana CA (2011) The UCR time series classification/clustering homepage. http://www.cs.ucr.edu/~eamonn/time_series_data/. Accessed 19 April 2015
Kosala R, Blockeel H (2000) Web mining research: a survey. In: ACM SIGKDD
Li M, Vitanyi P (1997) An introduction to Kolmogorov complexity and its applications, 2nd edn. Springer, Berlin
Liao TW (2005) Clustering of time series data—a survey. Pattern Recognit 38:1857–1874
Lines J, Bagnall A, Smith PC, Anderson S (2011) Classification of household devices by electricity usage profiles. In: IDEAL, LNCS, vol 6936
Lin J, Khade R, Li Y (2012) Rotation-invariant similarity in time series using bag-of-patterns representation. J Intell Inform Syst 39:287–315
Möerchen F (2003) Time series feature extraction for data mining using DWT and DFT. Technical report no. 33. Philipps-University Marburg, Marburg
Muata K (2007) Post-pruning in decision tree induction using multiple performance measures. Comput Oper Res 34:3331–3345
Mueen A, Keogh E, Young N (2011) Logical-Shapelets: an expressive primitive for time series classification. In: Proceedings of the ACM SIGKDD, pp 1154–1162
Rakthanmanon T, Keogh E, Lonardi S, Evans S (2011) Time series epenthesis: clustering time series streams requires ignoring some data. In: Proceedings of the IEEE ICDM
Rand WM (1971) Objective criteria for the evaluation of clustering methods. J Am Stat Assoc 66:846–850
Ratanamahatna CA, Keogh E (2004) Making time-series classification more accurate using learned constraints. In: SDM
Ruiz EJ, Hristidis V, Castillo C, Gionis A (2012) Correlating financial time series with micro-blogging activity. In: WSDM
Salton G, Wong A, Yang CS (1975) A vector space model for automatic indexing. Commun ACM 19:613–620
Shariat S, Pavlovic V (2011) Isotonic CCA for sequence alignment and activity recognition. In: Proceedings of the IEEE international conference on computer vision (ICCV 2011), pp 2572–2578
Silva C, Ribeiro B (2003) The importance of stop word removal on recall values in text categorization. In: Proceedings of the international joint conference on neural networks 2003
Wang X, Smith K, Hyndman R (2006) Characteristic-based clustering for time series data. Data Min Knowl Discov 13:335–364
Xing Z, Pei J, Yu P, Wang K (2011) Extracting interpretable features for early classification on time series. In: Proceedings of the SDM
Ye L, Keogh E (2009) Time series shapelets: a new primitive for data mining. In: Proceedings of the ACM SIGKDD, pp 947–956
Zakaria J, Mueen A, Keogh E (2012) Clustering time series using unsupervised-shapelets. In: Proceedings of the IEEE ICDM, pp 785–794
Zakaria J, Rotschafer S, Mueen A, Razak KA, Keogh E (2012) Mining massive archives of mice sounds with symbolized representations. In: Proceedings of the SDM
Zhang H, Ho TB, Zhang Y, Lin MS (2005) Unsupervised feature extraction for time series clustering using orthogonal wavelet transform. J Inform 30:305–319
Zhang M, Sawchuk AA (2012) Motion primitive-based human activity recognition using a bag-of-features approach. In: ACM SIGHIT international health informatics symposium (IHI 2012), pp 1–10
Zilberstein S (1996) Using anytime algorithms in intelligent systems. AI Mag 17:73–83
Acknowledgments
Thanks to all the donors of the datasets. This work was funded by NSF IIS—1161997 and a gift from Siemens.
Author information
Authors and Affiliations
Corresponding author
Additional information
Responsible editor: Ian Davidson
Appendix: Proof of Theorem 1
Appendix: Proof of Theorem 1
Theorem 1
We will prove (10).
Proof
Without loss of generality, assume
(If not, reverse both \(A\) and \(B\). That is, replace \(A\) with (\(A_{n}, A_{n-1},{\ldots }, A_{1}\)), and likewise for \(B\). Note that (10) holds for \(A\) and \(B\) iff it holds for \(A\) and \(B\) reversed.)
Note that the inequality (10) is equivalent to
where \(d(X, Y)\) is Euclidean distance, i.e., \(d\left( {X,Y} \right) =\sqrt{{\sum }_i^n (X_i -Y_i )^{2}}\), and
So, to prove (10), it suffices to prove (12). Here is a proof of (10).
The triangle inequality holds for Euclidean distance. Applying it twice shows
Rearranging gives
By inspection, \(d(A^{{\prime }},B^{\prime })^{2}=d(A,B)^{2}-\left( {A_n -B_n } \right) ^{2}+(A_1 -B_1 )^{2}\), so
Together with (11), this gives
Substituting this into (13) gives,
Rights and permissions
About this article
Cite this article
Zakaria, J., Mueen, A., Keogh, E. et al. Accelerating the discovery of unsupervised-shapelets. Data Min Knowl Disc 30, 243–281 (2016). https://doi.org/10.1007/s10618-015-0411-4
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10618-015-0411-4