A Representational MDL Framework for Improving Learning Power of Neural Network Formalisms - Artificial Intelligence Applications and Innovations - Part I (AIAI 2012)
Conference Papers Year : 2012

A Representational MDL Framework for Improving Learning Power of Neural Network Formalisms

Abstract

Minimum description length (MDL) principle is one of the well-known solutions for overlearning problem, specifically for artificial neural networks (ANNs). Its extension is called representational MDL (RMDL) principle and takes into account that models in machine learning are always constructed within some representation. In this paper, the optimization of ANNs formalisms as information representations using the RMDL principle is considered. A novel type of ANNs is proposed by extending linear recurrent ANNs with nonlinear “synapse to synapse” connections. Most of the elementary functions are representable with these networks (in contrast to classical ANNs) and that makes them easily learnable from training datasets according to a developed method of ANN architecture optimization. Methodology for comparing quality of different representations is illustrated by applying developed method in time series prediction and robot control.
Fichier principal
Vignette du fichier
978-3-642-33409-2_8_Chapter.pdf (85.52 Ko) Télécharger le fichier
Origin Files produced by the author(s)
Loading...

Dates and versions

hal-01521390 , version 1 (11-05-2017)

Licence

Identifiers

Cite

Alexey Potapov, Maxim Peterson. A Representational MDL Framework for Improving Learning Power of Neural Network Formalisms. 8th International Conference on Artificial Intelligence Applications and Innovations (AIAI), Sep 2012, Halkidiki, Greece. pp.68-77, ⟨10.1007/978-3-642-33409-2_8⟩. ⟨hal-01521390⟩
130 View
307 Download

Altmetric

Share

More