联系方式

  • QQ:99515681
  • 邮箱:99515681@qq.com
  • 工作时间:8:00-23:00
  • 微信:codinghelp

您当前位置:首页 >> Java编程Java编程

日期:2021-06-08 10:09

IS71083A Financial Data Modelling

Coursework Assignment

Implementing an Incremental (online) Training Algorithm

for Density Neural Networks

Design and implement an incremental backpropagation algorithm for training density neural networks

(DNN). DNN networks compute not only the mean of the target distribution, but also the variance with a

second output node. That is why you have to develop the backprop algorithm to train all weights leading

to each of the two output nodes in DNN given in the figure below. Start the training process using

plausible randomly generated weights. After the design of the training algorithm, apply it to model and

forecast a time series, generated using the Mackey-Glass equation (available in the lecture handouts).

Use more lagged inputs and hidden units if necessary to achieve better predictions.

sum

Design a working prototype of the incremental backpropagation for DNN density networks in Matlab.

The prototype development should include data structures for the input-to-hidden and hidden-to-output

connections, and loops for the forward and backward pass.

References:

D.A.Nix and E.Weigend (1994). Estimating the mean and variance of the target probability distribution,

In: Proc. 1994 IEEE Int. Conf. on Neural Networks (ICNN'94), Orlando, FL, pp. 55-60.

Nikolay Nikolaev and Hitoshi Iba (2006).

"Adaptive Learning of Polynomial Networks: Genetic Programming, Backpropagation and Bayesian Methods",

Springer, New York, pp. 256 - 262.


版权所有:编程辅导网 2020 All Rights Reserved 联系方式:QQ:99515681 电子信箱:99515681@qq.com
免责声明:本站部分内容从网络整理而来,只供参考!如有版权问题可联系本站删除。