Improving modeling of stochastic processes by smart denoising
Abstract
This paper proposes a novel method for modeling stochastic processes, which are knownto be notoriously hard to predict accurately. State of the art methods quickly overfitand create big differences between train and test datasets. We present a method basedon smart noise addition to the data obtained from unknown stochastic process, whichis capable of reducing data overfitting. The proposed method works as an addition tothe current state of the art methods in both supervised and unsupervised setting. Weevaluate the method on equities and cryptocurrency datasets, specifically chosen fortheir chaotic and unpredictable nature. We show that with our method we significantlyreduce overfitting and increase performance, compared to several commonly used machinelearning algorithms: Random forest, General linear model and LSTM deep learning model.References
Kraken exchange.
Yahoo Finance.om
A. Creswell, T. White, V. Dumoulin,
K. Arulkumaran, B. Sengupta, and A. A.
Bharath. Generative adversarial networks:
An overview. IEEE Signal Processing Maga-
zine, 35(1):53–65, 2018.
P. Jaworski, F. Durante, W. K. Hardle, and
T. Rychlik. Copula theory and its applications,
volume 198. Springer, 2010.
J. Jelencic and D. Mladenic. Modeling stochas-
tic processes by simultaneous optimization of latent representation and target variable. 2020.
H. Joe. Dependence Modeling with Copulas.
CRC Press, 2014.
D. Kingma and J. Ba. Adam: A Method
for Stochastic Optimization. 2014. https:
//arxiv.org/abs/1412.6980.
K. Y. Levy. The power of normalization:
Faster evasion of saddle points. arXiv preprint
arXiv:1611.04831, 2016.
M. Liu, W. Wu, Z. Gu, Z. Yu, F. Qi, and Y. Li.
Deep learning based on batch normalization
for p300 signal detection. Neurocomputing,
:288–297, 2018.
R. C. Merton. Option pricing when underlying
stock returns are discontinuous. Journal of
financial economics, 3(1-2):125–144, 1976.
J. J. Murphy. Technical Analysis of the Fi-
nancial Markets: A Comprehensive Guide to
Trading Methods and Applications. New York
Institute of Finance Series. New York Insti-
tute of Finance, 1999.
T. Salimans and D. P. Kingma. Weight nor-
malization: A simple reparameterization to
accelerate training of deep neural networks.
Advances in neural information processing sys-
tems, 29:901–909, 2016.
B. W. Turnbull. The empirical distribution
function with arbitrarily grouped, censored
and truncated data. Journal of the Royal
Statistical Society: Series B (Methodological),
(3):290–295, 1976.
R. Vidal, J. Bruna, R. Giryes, and S. Soatto.
Mathematics of deep learning. arXiv preprint
arXiv:1712.04741, 2017.
P. Vincent, H. Larochelle, Y. Bengio, and
P.-A. Manzagol. Extracting and composing
robust features with denoising autoencoders.
In Proceedings of the 25th international con-
ference on Machine learning, pages 1096–1103,
N. Wax. Selected papers on noise and stochas-
tic processes. Courier Dover Publications,
DOI:
https://doi.org/10.31449/inf.v46i1.3875Downloads
Published
How to Cite
Issue
Section
License
Authors retain copyright in their work. By submitting to and publishing with Informatica, authors grant the publisher (Slovene Society Informatika) the non-exclusive right to publish, reproduce, and distribute the article and to identify itself as the original publisher.
All articles are published under the Creative Commons Attribution license CC BY 3.0. Under this license, others may share and adapt the work for any purpose, provided appropriate credit is given and changes (if any) are indicated.
Authors may deposit and share the submitted version, accepted manuscript, and published version, provided the original publication in Informatica is properly cited.







