Search Results

Now showing 1 - 7 of 7
  • Item
    Topological data analysis of contagion maps for examining spreading processes on networks
    ([London] : Nature Publishing Group UK, 2015) Taylor, Dane; Klimm, Florian; Harrington, Heather A.; Kramár, Miroslav; Mischaikow, Konstantin; Porter, Mason A.; Mucha, Peter J.
    Social and biological contagions are influenced by the spatial embeddedness of networks. Historically, many epidemics spread as a wave across part of the Earth’s surface; however, in modern contagions long-range edges—for example, due to airline transportation or communication media—allow clusters of a contagion to appear in distant locations. Here we study the spread of contagions on networks through a methodology grounded in topological data analysis and nonlinear dimension reduction. We construct ‘contagion maps’ that use multiple contagions on a network to map the nodes as a point cloud. By analysing the topology, geometry and dimensionality of manifold structure in such point clouds, we reveal insights to aid in the modelling, forecast and control of spreading processes. Our approach highlights contagion maps also as a viable tool for inferring low-dimensional structure in networks.
  • Item
    Sleep apnea-hypopnea quantification by cardiovascular data analysis
    (San Francisco, CA : Public Library of Science (PLoS), 2014) Camargo, S.; Riedl, M.; Anteneodo, C.; Kurths, J.; Penzel, T.; Wessel, N.
    Sleep disorders are a major risk factor for cardiovascular diseases. Sleep apnea is the most common sleep disturbance and its detection relies on a polysomnography, i.e., a combination of several medical examinations performed during a monitored sleep night. In order to detect occurrences of sleep apnea without the need of combined recordings, we focus our efforts on extracting a quantifier related to the events of sleep apnea from a cardiovascular time series, namely systolic blood pressure (SBP). Physiologic time series are generally highly nonstationary and entrap the application of conventional tools that require a stationary condition. In our study, data nonstationarities are uncovered by a segmentation procedure which splits the signal into stationary patches, providing local quantities such as mean and variance of the SBP signal in each stationary patch, as well as its duration L. We analysed the data of 26 apneic diagnosed individuals, divided into hypertensive and normotensive groups, and compared the results with those of a control group. From the segmentation procedure, we identified that the average duration 〈L〉, as well as the average variance 〈σ2〉, are correlated to the apnea-hypoapnea index (AHI), previously obtained by polysomnographic exams. Moreover, our results unveil an oscillatory pattern in apneic subjects, whose amplitude S∗ is also correlated with AHI. All these quantities allow to separate apneic individuals, with an accuracy of at least 79%. Therefore, they provide alternative criteria to detect sleep apnea based on a single time series, the systolic blood pressure.
  • Item
    Probing the Statistical Properties of Unknown Texts: Application to the Voynich Manuscript
    (San Francisco, CA : Public Library of Science (PLoS), 2013) Amancio, D.R.; Altmann, E.G.; Rybski, D.; Oliveira Jr., O.N.; da Costa, L.F.
    While the use of statistical physics methods to analyze large corpora has been useful to unveil many patterns in texts, no comprehensive investigation has been performed on the interdependence between syntactic and semantic factors. In this study we propose a framework for determining whether a text (e.g., written in an unknown alphabet) is compatible with a natural language and to which language it could belong. The approach is based on three types of statistical measurements, i.e. obtained from first-order statistics of word properties in a text, from the topology of complex networks representing texts, and from intermittency concepts where text is treated as a time series. Comparative experiments were performed with the New Testament in 15 different languages and with distinct books in English and Portuguese in order to quantify the dependency of the different measurements on the language and on the story being told in the book. The metrics found to be informative in distinguishing real texts from their shuffled versions include assortativity, degree and selectivity of words. As an illustration, we analyze an undeciphered medieval manuscript known as the Voynich Manuscript. We show that it is mostly compatible with natural languages and incompatible with random texts. We also obtain candidates for keywords of the Voynich Manuscript which could be helpful in the effort of deciphering it. Because we were able to identify statistical measurements that are more dependent on the syntax than on the semantics, the framework may also serve for text analysis in language-dependent applications.
  • Item
    Change in the embedding dimension as an indicator of an approaching transition
    (San Francisco, CA : Public Library of Science (PLoS), 2014) Neuman, Y.; Marwan, N.; Cohen, Y.
    Predicting a transition point in behavioral data should take into account the complexity of the signal being influenced by contextual factors. In this paper, we propose to analyze changes in the embedding dimension as contextual information indicating a proceeding transitive point, called OPtimal Embedding tRANsition Detection (OPERAND). Three texts were processed and translated to time-series of emotional polarity. It was found that changes in the embedding dimension proceeded transition points in the data. These preliminary results encourage further research into changes in the embedding dimension as generic markers of an approaching transition point.
  • Item
    Spatiotemporal data analysis with chronological networks
    ([London] : Nature Publishing Group UK, 2020) Ferreira, Leonardo N.; Vega-Oliveros, Didier A.; Cotacallapa, Moshé; Cardoso, Manoel F.; Quiles, Marcos G.; Zhao, Liang; Macau, Elbert E. N.
    The number of spatiotemporal data sets has increased rapidly in the last years, which demands robust and fast methods to extract information from this kind of data. Here, we propose a network-based model, called Chronnet, for spatiotemporal data analysis. The network construction process consists of dividing a geometric space into grid cells represented by nodes connected chronologically. Strong links in the network represent consecutive recurrent events between cells. The chronnet construction process is fast, making the model suitable to process large data sets. Using artificial and real data sets, we show how chronnets can capture data properties beyond simple statistics, like frequent patterns, spatial changes, outliers, and spatiotemporal clusters. Therefore, we conclude that chronnets represent a robust tool for the analysis of spatiotemporal data sets.
  • Item
    Testing the detectability of spatio-temporal climate transitions from paleoclimate networks with the start model
    (Göttingen : Copernicus, 2014) Rehfeld, K.; Molkenthin, N.; Kurths, J.
    A critical challenge in paleoclimate data analysis is the fact that the proxy data are heterogeneously distributed in space, which affects statistical methods that rely on spatial embedding of data. In the paleoclimate network approach nodes represent paleoclimate proxy time series, and links in the network are given by statistically significant similarities between them. Their location in space, proxy and archive type is coded in the node attributes. We develop a semi-empirical model for Spatio- Temporally AutocoRrelated Time series, inspired by the interplay of different Asian Summer Monsoon (ASM) systems. We use an ensemble of transition runs of this START model to test whether and how spatio-temporal climate transitions could be detectable from (paleo)climate networks. We sample model time series both on a grid and at locations at which paleoclimate data are available to investigate the effect of the spatially heterogeneous availability of data. Node betweenness centrality, averaged over the transition region, does not respond to the transition displayed by the START model, neither in the grid-based nor in the scattered sampling arrangement. The regionally defined measures of regional node degree and cross link ratio, however, are indicative of the changes in both scenarios, although the magnitude of the changes differs according to the sampling. We find that the START model is particularly suitable for pseudo-proxy experiments to test the technical reconstruction limits of paleoclimate data based on their location, and we conclude that (paleo)climate networks are suitable for investigating spatio-temporal transitions in the dependence structure of underlying climatic fields.
  • Item
    Bayesian Data Analysis for Revealing Causes of the Middle Pleistocene Transition
    ([London] : Macmillan Publishers Limited, part of Springer Nature, 2019) Mukhin, Dmitry; Gavrilov, Andrey; Loskutov, Evgeny; Kurths, Juergen; Feigin, Alexander
    Currently, causes of the middle Pleistocene transition (MPT) – the onset of large-amplitude glacial variability with 100 kyr time scale instead of regular 41 kyr cycles before – are a challenging puzzle in Paleoclimatology. Here we show how a Bayesian data analysis based on machine learning approaches can help to reveal the main mechanisms underlying the Pleistocene variability, which most likely explain proxy records and can be used for testing existing theories. We construct a Bayesian data-driven model from benthic δ18O records (LR04 stack) accounting for the main factors which may potentially impact climate of the Pleistocene: internal climate dynamics, gradual trends, variations of insolation, and millennial variability. In contrast to some theories, we uncover that under long-term trends in climate, the strong glacial cycles have appeared due to internal nonlinear oscillations induced by millennial noise. We find that while the orbital Milankovitch forcing does not matter for the MPT onset, the obliquity oscillation phase-locks the climate cycles through the meridional gradient of insolation.