Nalazite se na CroRIS probnoj okolini. Ovdje evidentirani podaci neće biti pohranjeni u Informacijskom sustavu znanosti RH. Ako je ovo greška, CroRIS produkcijskoj okolini moguće je pristupi putem poveznice www.croris.hr
izvor podataka: crosbi

Benchmarking attention-based interpretability of deep learning in multivariate time series predictions (CROSBI ID 290633)

Prilog u časopisu | izvorni znanstveni rad | međunarodna recenzija

Barić, Domjan ; Fumić, Petar ; Horvatić, Davor ; Lipić, Tomislav Benchmarking attention-based interpretability of deep learning in multivariate time series predictions // Entropy (Basel. Online), 23 (2021), 2; 143, 23. doi: 10.3390/e23020143

Podaci o odgovornosti

Barić, Domjan ; Fumić, Petar ; Horvatić, Davor ; Lipić, Tomislav

engleski

Benchmarking attention-based interpretability of deep learning in multivariate time series predictions

The adaptation of deep learning models within safety-critical systems cannot rely only on good prediction performance but needs to provide interpretable and robust explanations for their decisions. When modeling complex sequences, attention mechanisms are regarded as the established approach to support deep neural networks with intrinsic interpretability. This paper focuses on the emerging trend of specifically designing diagnostic datasets for understanding the inner workings of attention mechanism based deep learning models for multivariate forecasting tasks. We design a novel benchmark of synthetically designed datasets with the transparent underlying generating process of multiple time series interactions with increasing complexity. The benchmark enables empirical evaluation of the performance of attention based deep neural networks in three different aspects: (i) prediction performance score, (ii) interpretability correctness, (iii) sensitivity analysis. Our analysis shows that although most models have satisfying and stable prediction performance results, they often fail to give correct interpretability. The only model with both a satisfying performance score and correct interpretability is IMV-LSTM, capturing both autocorrelations and crosscorrelations between multiple time series. Interestingly, while evaluating IMV-LSTM on simulated data from statistical and mechanistic models, the correctness of interpretability increases with more complex datasets.

multivariate time series ; attention mechanism ; interpretability ; synthetically designed datasets

nije evidentirano

nije evidentirano

nije evidentirano

nije evidentirano

nije evidentirano

nije evidentirano

Podaci o izdanju

23 (2)

2021.

143

23

objavljeno

1099-4300

10.3390/e23020143

Povezanost rada

Fizika, Interdisciplinarne tehničke znanosti, Računarstvo

Poveznice
Indeksiranost