Recurrent neural networks and exponential PAA for virtual marine sensors

Stefan Oehmcke, Oliver Zielinski, Oliver Kramer

    4 Citations (Scopus)

    Abstract

    Virtual sensors are getting more and more important as replacement and quality control tool for expensive and fragile hardware sensors. We introduce a virtual sensor application with marine sensor data from two data sources. The virtual sensor models are built upon recurrent neural networks (RNNs). To take full advantage of past data, we employ the time dimensionality reduction method piecewise approximate aggregation (PAA). We present an extension of this method, called exponential PAA (ExPAA) that pulls finer details from recent values, but preserves less exact information about the past. Experimental results demonstrate that RNNs benefit from this extension and confirm the stability and usability of our virtual sensor models over a five-month period of multivariate marine time series data.

    Original languageEnglish
    Title of host publication2017 International Joint Conference on Neural Networks, IJCNN 2017 - Proceedings
    Number of pages8
    PublisherInstitute of Electrical and Electronics Engineers Inc.
    Publication date30 Jun 2017
    Pages4459-4466
    Article number7966421
    ISBN (Electronic)9781509061815
    DOIs
    Publication statusPublished - 30 Jun 2017
    Event2017 International Joint Conference on Neural Networks, IJCNN 2017 - Anchorage, United States
    Duration: 14 May 201719 May 2017

    Conference

    Conference2017 International Joint Conference on Neural Networks, IJCNN 2017
    Country/TerritoryUnited States
    CityAnchorage
    Period14/05/201719/05/2017
    SponsorBrain-Mind Institute (BMI), Budapest Semester in Cognitive Science (BSCS), Intel
    SeriesProceedings of the International Joint Conference on Neural Networks
    Volume2017-May

    Fingerprint

    Dive into the research topics of 'Recurrent neural networks and exponential PAA for virtual marine sensors'. Together they form a unique fingerprint.

    Cite this