Deep learning for time series data

"Arun Kejariwal (Independent) and Ira Cohen (Anodot) share a novel two-step approach for building more reliable prediction models by integrating anomalies in them. The first step uses anomaly detection algorithms to discover anomalies in a time series in the training data. In the second, multip...

Descripción completa

Detalles Bibliográficos
Autor Corporativo: O'Reilly Artificial Intelligence Conference (-)
Otros Autores: Kejariwal, Arun, on-screen presenter (onscreen presenter), Cohen, Ira M., on-screen presenter
Formato: Vídeo online
Idioma:Inglés
Publicado: [Place of publication not identified] : O'Reilly 2019.
Materias:
Ver en Biblioteca Universitat Ramon Llull:https://discovery.url.edu/permalink/34CSUC_URL/1im36ta/alma991009822781606719
Descripción
Sumario:"Arun Kejariwal (Independent) and Ira Cohen (Anodot) share a novel two-step approach for building more reliable prediction models by integrating anomalies in them. The first step uses anomaly detection algorithms to discover anomalies in a time series in the training data. In the second, multiple prediction models, including time series models and deep networks, are trained, enriching the training data with the information about the anomalies discovered in the first step. Anomaly detection for individual time series is a necessary but insufficient step due to the fact that anomaly detection over a set of live data streams may result in anomaly fatigue, thereby limiting effective decision making. One way to address the above is to carry out anomaly detection in a multidimensional space. However, this is typically very expensive computationally and hence not suitable for live data streams. Another approach is to carry out anomaly detection on individual data streams and then leverage correlation analysis to minimize false positives, which in turn helps in surfacing actionable insights faster. Anomaly detection for individual time series is a necessary but insufficient step due to the fact that anomaly detection over a set of live data streams may result in anomaly fatigue, thereby limiting effective decision making. One way to address the above is to carry out anomaly detection in a multidimensional space. However, this is typically very expensive computationally and hence not suitable for live data streams. Another approach is to carry out anomaly detection on individual data streams and then leverage correlation analysis to minimize false positives, which in turn helps in surfacing actionable insights faster."--Resource description page.
Notas:Title from title screen (viewed November 14, 2019).
Descripción Física:1 online resource (1 streaming video file (42 min., 29 sec.)) : digital, sound, color