TY - JOUR
T1 - Interpretable transformation and analysis of timelines through learning via surprisability
AU - Mokryn, Osnat
AU - Lazebnik, Teddy
AU - Ben-Shoshan, Hagit
N1 - Publisher Copyright:
© 2025 Author(s).
PY - 2025/7/1
Y1 - 2025/7/1
N2 - The analysis of high-dimensional timeline data and the identification of outliers and anomalies are critical across diverse domains, including sensor readings, biological and medical data, historical records, and global statistics. However, conventional analysis techniques often struggle with challenges such as high dimensionality, complex distributions, and sparsity. These limitations hinder the ability to extract meaningful insights from complex temporal datasets, making it difficult to identify trending features, outliers, and anomalies effectively. Inspired by surprisability—a cognitive science concept describing how humans instinctively focus on unexpected deviations—we propose Learning via Surprisability (LvS), a novel approach for transforming high-dimensional timeline data. LvS quantifies and prioritizes anomalies in time-series data by formalizing deviations from the expected behavior. LvS bridges cognitive theories of attention with computational methods, enabling the detection of anomalies and shifts in a way that preserves critical context, offering a new lens for interpreting complex datasets. We demonstrate the usefulness of LvS on three high-dimensional timeline use cases: a time series of sensor data, a global dataset of mortality causes over multiple years, and a textual corpus containing over two centuries of the State of the Union Addresses by U.S. presidents. Our results show that LvS transformation enables efficient and interpretable identification of outliers, anomalies, and the most variable features along the timeline.
AB - The analysis of high-dimensional timeline data and the identification of outliers and anomalies are critical across diverse domains, including sensor readings, biological and medical data, historical records, and global statistics. However, conventional analysis techniques often struggle with challenges such as high dimensionality, complex distributions, and sparsity. These limitations hinder the ability to extract meaningful insights from complex temporal datasets, making it difficult to identify trending features, outliers, and anomalies effectively. Inspired by surprisability—a cognitive science concept describing how humans instinctively focus on unexpected deviations—we propose Learning via Surprisability (LvS), a novel approach for transforming high-dimensional timeline data. LvS quantifies and prioritizes anomalies in time-series data by formalizing deviations from the expected behavior. LvS bridges cognitive theories of attention with computational methods, enabling the detection of anomalies and shifts in a way that preserves critical context, offering a new lens for interpreting complex datasets. We demonstrate the usefulness of LvS on three high-dimensional timeline use cases: a time series of sensor data, a global dataset of mortality causes over multiple years, and a textual corpus containing over two centuries of the State of the Union Addresses by U.S. presidents. Our results show that LvS transformation enables efficient and interpretable identification of outliers, anomalies, and the most variable features along the timeline.
UR - https://www.scopus.com/pages/publications/105011406097
U2 - 10.1063/5.0269365
DO - 10.1063/5.0269365
M3 - ???researchoutput.researchoutputtypes.contributiontojournal.article???
C2 - 40690607
AN - SCOPUS:105011406097
SN - 1054-1500
VL - 35
JO - Chaos
JF - Chaos
IS - 7
M1 - 073134
ER -