Prior knowledge distillation based on financial time series

by   Jie Fang, et al.

One of the major characteristics of financial time series is that they contain a large amount of non-stationary noise, which is challenging for deep neural networks. People normally use various features to address this problem. However, the performance of these features depends on the choice of hyper-parameters. In this paper, we propose to use neural networks to represent these indicators and train a large network constructed of smaller networks as feature layers to fine-tune the prior knowledge represented by the indicators. During back propagation, prior knowledge is transferred from human logic to machine logic via gradient descent. Prior knowledge is the deep belief of neural network and teaches the network to not be affected by non-stationary noise. Moreover, co-distillation is applied to distill the structure into a much smaller size to reduce redundant features and the risk of overfitting. In addition, the decisions of the smaller networks in terms of gradient descent are more robust and cautious than those of large networks. In numerical experiments, we find that our algorithm is faster and more accurate than traditional methods on real financial datasets. We also conduct experiments to verify and comprehend the method.


Comparative study of Financial Time Series Prediction by Artificial Neural Network with Gradient Descent Learning

Financial forecasting is an example of a signal processing problem which...

Bilinear Input Normalization for Neural Networks in Financial Forecasting

Data normalization is one of the most important preprocessing steps when...

Alpha Discovery Neural Network based on Prior Knowledge

In financial automatic feature construction task, genetic programming is...

Channel Planting for Deep Neural Networks using Knowledge Distillation

In recent years, deeper and wider neural networks have shown excellent p...

Evaluating data augmentation for financial time series classification

Data augmentation methods in combination with deep neural networks have ...

Knowledge Representing: Efficient, Sparse Representation of Prior Knowledge for Knowledge Distillation

Despite the recent works on knowledge distillation (KD) have achieved a ...

DIVERSIFY: A General Framework for Time Series Out-of-distribution Detection and Generalization

Time series remains one of the most challenging modalities in machine le...

Please sign up or login with your details

Forgot password? Click here to reset