Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3533271.3561709acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicaifConference Proceedingsconference-collections
research-article

Core Matrix Regression and Prediction with Regularization

Published: 26 October 2022 Publication History

Abstract

Many finance time-series analyses often track a matrix of variables at each time and study their co-evolution over a long time. The matrix time series is overly sparse, involves complex interactions among latent matrix factors, and demands advanced models to extract dynamic temporal patterns from these interactions. This paper proposes a Core Matrix Regression with Regularization algorithm (CMRR) to capture spatiotemporal relations in sparse matrix-variate time series. The model decomposes each matrix into three factor matrices of row entities, column entities, and interactions between row entities and column entities, respectively. Subsequently, it applies recurrent neural networks on interaction matrices to extract temporal patterns. Given the sparse matrix, we design an element-wise orthogonal matrix factorization that leverages the Stochastic Gradient Descent (SGD) in a deep learning platform to overcome the challenge of the sparsity and large volume of complex data. The experiment confirms that combining orthogonal matrix factorization with recurrent neural networks is highly effective and outperforms existing graph neural networks and tensor-based time series prediction methods. We apply CMRR in three real-world financial applications: firm earning forecast, predicting firm fundamentals, and firm characteristics, and demonstrate its consistent performance superiority: reducing error by 23%-53% over other state-of-the-art high-dimensional time series prediction algorithms.

References

[1]
Miguel Araújo, Pedro Ribeiro, and Christos Faloutsos. 2018. TensorCast: Forecasting Time-Evolving Networks with Contextual Information. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, IJCAI-18. International Joint Conferences on Artificial Intelligence Organization, 5199–5203.
[2]
Ryan T Ball and Eric Ghysels. 2018. Automated Earnings Forecasts: Beat Analysts or Combine and Conquer?Management Science 64, 10 (2018), 4936–4952.
[3]
V. Banzon, T. M. Smith, T. M. Chin, C. Liu, and W. Hankins. 2016. A long-term record of blended satellite and in situ sea-surface temperature for climate monitoring, modeling and environmental studies. Earth System Science Data 8, 1 (2016), 165–176.
[4]
Robert Bell, Yehuda Koren, and Chris Volinsky. 2007. Modeling Relationships at Multiple Scales to Improve Accuracy of Large Recommender Systems. In Proceedings of the 13th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. New York, NY, USA, 95–104.
[5]
Dimitris Bertsimas and Martin S Copenhaver. 2018. Characterization of the equivalence of robustification and regularization in linear and matrix regression. European Journal of Operational Research 270, 3 (2018), 931–942.
[6]
Gary Chamberlain and Michael Rothschild. 1982. Arbitrage, Factor Structure, and Mean-Variance Analysis on Large Asset Markets. Working Paper 996. National Bureau of Economic Research.
[7]
Elynn Y. Chen, Ruey S. Tsay, and Rong Chen. 2020. Constrained Factor Models for High-Dimensional Matrix-Variate Time Series. J. Amer. Statist. Assoc. 115, 530 (2020), 775–793.
[8]
Junyoung Chung, Caglar Gulcehre, Kyunghyun Cho, and Yoshua Bengio. 2014. Empirical evaluation of gated recurrent neural networks on sequence modeling. In NIPS 2014 Workshop on Deep Learning, December 2014.
[9]
Guillaume Coqueret. 2021. Persistence in factor-based supervised learning models. The Journal of Finance and Data Science(2021).
[10]
Guillaume Coqueret and Tony Guida. 2020. Machine Learning for Factor Investing: R Version. CRC Press.
[11]
Daniel M. Dunlavy, Tamara G. Kolda, and Evrim Acar. 2011. Temporal Link Prediction Using Matrix and Tensor Factorizations. ACM Trans. Knowl. Discov. Data 5, 2, Article 10 (feb 2011), 27 pages.
[12]
Eugene F Fama and Kenneth R French. 2006. Profitability, investment and average returns. Journal of financial economics 82, 3 (2006), 491–518.
[13]
Felix A. Gers, Nicol N. Schraudolph, and Jürgen Schmidhuber. 2003. Learning Precise Timing with Lstm Recurrent Networks. J. Mach. Learn. Res. 3, null (mar 2003), 115–143.
[14]
Kewei Hou, Mathijs A Van Dijk, and Yinglei Zhang. 2012. The implied cost of capital: A new approach. Journal of Accounting and Economics 53, 3 (2012), 504–526.
[15]
Baoyu Jing, Hanghang Tong, and Yada Zhu. 2021. Network of Tensor Time Series. In Proceedings of the Web Conference 2021(Ljubljana, Slovenia). 2425–2437.
[16]
Dohyun Kim and Bong-Jin Yum. 2005. Collaborative filtering based on iterative principal component analysis. Expert Systems with Applications 28, 4 (2005), 823–830.
[17]
Diederik P. Kingma and Jimmy Ba. 2015. Adam: A Method for Stochastic Optimization. In 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9.
[18]
Clifford Lam and Qiwei Yao. 2012. Factor modeling for high-dimensional time series: Inference for the number of factors. The Annals of Statistics 40, 2 (2012), 694 – 726.
[19]
Clifford Lam, Qiwei Yao, and Neil Bathia. 2011. Estimation of latent factors for high-dimensional time series. Biometrika 98, 4 (10 2011), 901–918.
[20]
Aldo Pareja, Giacomo Domeniconi, Jie Chen, Tengfei Ma, Toyotaro Suzumura, Hiroki Kanezashi, Tim Kaler, and Charles E Leisersen. 2020. Evolvegcn: Evolving graph convolutional networks for dynamic graphs. AAAI (2020).
[21]
Adam Paszke and et al.2019. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems 32, H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett (Eds.). Curran Associates, Inc., 8024–8035.
[22]
Rajat Sen, Hsiang-Fu Yu, and Inderjit S. Dhillon. 2019. Think Globally, Act Locally: A Deep Neural Network Approach to High-Dimensional Time Series Forecasting. In Neural Information Processing Systems (NIPS).
[23]
Shikhar Srivastava and Stefan Lessmann. 2018. A comparative study of LSTM neural networks in forecasting day-ahead global horizontal irradiance with satellite data. Solar Energy 162(2018), 232–247. https://doi.org/10.1016/j.solener.2018.01.005
[24]
Chenhao Su, Sheng Gao, and Si Li. 2020. GATE: Graph-Attention Augmented Temporal Neural Network for Medication Recommendation. IEEE Access 8(2020), 125447–125458.
[25]
Ajim Uddin, Xinyuan Tao, Chia-Ching Chou, and Dantong Yu. 2020. Nonlinear Tensor Completion Using Domain Knowledge: An Application in Analysts’ Earnings Forecast. In 2020 International Conference on Data Mining Workshops (ICDMW). IEEE.
[26]
Ajim Uddin, Xinyuan Tao, Chia-Ching Chou, and Dantong Yu. 2021. Are missing values important for earnings forecasts? A machine learning perspective. Quantitative Finance(2021), 1–20.
[27]
Dong Wang, Xialu Liu, and Rong Chen. 2019. Factor models for matrix-valued high-dimensional time series. Journal of Econometrics 208, 1 (2019), 231–248. Special Issue on Financial Engineering and Risk Management.
[28]
Jianxin Yin and Hongzhe Li. 2012. Model selection and estimation in the matrix normal graphical model. Journal of Multivariate Analysis 107 (2012), 119–140. https://doi.org/10.1016/j.jmva.2012.01.005
[29]
Hsiang-Fu Yu, Nikhil Rao, and Inderjit S Dhillon. 2016. Temporal Regularized Matrix Factorization for High-dimensional Time Series Prediction. In Advances in Neural Information Processing Systems.
[30]
Junlong Zhao and Chenlei Leng. 2014. STRUCTURED LASSO FOR REGRESSION WITH MATRIX COVARIATES. Statistica Sinica (2014).
[31]
Ling Zhao, Yujiao Song, Chao Zhang, Yu Liu, Pu Wang, Tao Lin, Min Deng, and Haifeng Li. 2020. T-GCN: A Temporal Graph Convolutional Network for Traffic Prediction. IEEE Transactions on Intelligent Transportation Systems 21, 9(2020), 3848–3858.
[32]
Pengpeng Zhao, Anjing Luo, Yanchi Liu, Fuzhen Zhuang, Jiajie Xu, Zhixu Li, Victor S. Sheng, and Xiaofang Zhou. 2020. Where to Go Next: A Spatio-Temporal Gated Network for Next POI Recommendation. IEEE Transactions on Knowledge and Data Engineering (2020), 1–1.
[33]
Hua Zhou and Lexin Li. 2014. Regularized matrix regression. Journal of the Royal Statistical Society. Series B, Statistical methodology 76, 2 (March 2014), 463—483. https://doi.org/10.1111/rssb.12031

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
ICAIF '22: Proceedings of the Third ACM International Conference on AI in Finance
November 2022
527 pages
ISBN:9781450393768
DOI:10.1145/3533271
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 26 October 2022

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. Tensor algorithm
  2. matrix factorization
  3. matrix-variate time series prediction
  4. recurrent neural networks

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

ICAIF '22
Sponsor:

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 142
    Total Downloads
  • Downloads (Last 12 months)29
  • Downloads (Last 6 weeks)2
Reflects downloads up to 10 Nov 2024

Other Metrics

Citations

View Options

Get Access

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format.

HTML Format

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media