Dalarna University's logo and link to the university's website

du.sePublications
System disruptions
We are currently experiencing disruptions on the search portals due to high traffic. We are working to resolve the issue, you may temporarily encounter an error message.
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • chicago-author-date
  • chicago-note-bibliography
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Temporal Localization of Representations in Recurrent Neural Networks
Dalarna University, School of Information and Engineering.
2023 (English)Independent thesis Advanced level (degree of Master (Two Years)), 10 credits / 15 HE creditsStudent thesis
Abstract [en]

Recurrent Neural Networks (RNNs) are pivotal in deep learning for time series prediction, but they suffer from 'exploding values' and 'gradient decay,' particularly when learning temporally distant interactions. Long Short-Term Memory (LSTM) and Gated Recurrent Units (GRU) have addressed these issues to an extent, but the precise mitigating mechanisms remain unclear. Moreover, the success of feedforward neural networks in time series tasks using an 'attention mechanism' raises questions about the solutions offered by LSTMs and GRUs. This study explores an alternative explanation for the challenges faced by RNNs in learning long-range correlations in the input data. Could the issue lie in the movement of the representations - how hidden nodes store and process information - across nodes instead of localization? Evidence presented suggests that RNNs can indeed possess "moving representations," with certain training conditions reducing this movement. These findings point towards the necessity of further research on localizing representations.

Place, publisher, year, edition, pages
2023.
Keywords [en]
Recurrent Neural Networks (RNNs), Deep Learning, Time Series Prediction, Exploding Values, Gradient Decay, Long Short-Term Memory (LSTMs), Gated Recurrent Units (GRUs), Attention Mechanism, Moving Representations, Localizing Representations
National Category
Computer and Information Sciences
Identifiers
URN: urn:nbn:se:du-46777OAI: oai:DiVA.org:du-46777DiVA, id: diva2:1790974
Subject / course
Microdata Analysis
Available from: 2023-08-24 Created: 2023-08-24Bibliographically approved

Open Access in DiVA

fulltext(1166 kB)106 downloads
File information
File name FULLTEXT01.pdfFile size 1166 kBChecksum SHA-512
b45549a9dfeaefa1854a3f2695c74345f69d68b5e8d2385e4020a9b80effd8bf48477533a8839cdf32dd36a9302c2a6f7166ba38a136800a6863cabc926e38eb
Type fulltextMimetype application/pdf

By organisation
School of Information and Engineering
Computer and Information Sciences

Search outside of DiVA

GoogleGoogle Scholar
Total: 110 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

urn-nbn

Altmetric score

urn-nbn
Total: 196 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • chicago-author-date
  • chicago-note-bibliography
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf