Browse wiki

From Navigators

Jump to: navigation, search
Abstract A method based on the information theory c A method based on the information theory concept of entropy is presented for selecting subsets of data for off-line model identification. By using entropy-based data selection instead of random equiprobable sampling before training models, significant improvements are achieved in parameter convergence, accuracy and generalisation ability. Furthermore, model evaluation metrics exhibit less variance, therefore allowing faster convergence when multiple modelling trials have to be executed. These features are experimentally demonstrated by the results of an extensive number of neural network predictive modelling experiments, where the single difference in the identification of pairs of models was the data set used to tune model parameters. Unlike most active learning and instance selection procedures, the method is not iterative, does not rely on an existing model, and does not require a specific modelling technique. Instead, it selects data points in one unsupervised step relying solely on Shannon’s information measure. g solely on Shannon’s information measure.
Address Vancouver, Canada  +
Author Pedro M. Ferreira +
Booktitle 2016 International Joint Conference on Neural Networks (IJCNN) (World Congress on Computational Intelligence)  +
Key Ferreira2016-Entropy  +
Month jul  +
NumPubDate 2,016.07  +
Pages 3330–3337  +
ResearchLine Timeliness and Adaptation in Dependable Systems (TADS) +
Title Unsupervised Entropy-Based Selection of Data Sets for Improved Model Fitting  +
Type inproceedings  +
Url  +
Year 2016  +
Categories Publication  +
Modification dateThis property is a special property in this wiki. 27 December 2016 15:55:54  +
hide properties that link here 
  No properties link to this page.


Enter the name of the page to start browsing from.
Personal tools
Navigators toolbox