Unevenly spaced time series

Last updated

In statistics, signal processing, and econometrics, an unevenly (or unequally or irregularly) spaced time series is a sequence of observation time and value pairs (tn, Xn) in which the spacing of observation times is not constant.

Contents

Unevenly spaced time series naturally occur in many industrial and scientific domains: natural disasters such as earthquakes, floods, or volcanic eruptions typically occur at irregular time intervals. In observational astronomy, measurements such as spectra of celestial objects are taken at times determined by weather conditions, availability of observation time slots, and suitable planetary configurations. In clinical trials (or more generally, longitudinal studies), a patient's state of health may be observed only at irregular time intervals, and different patients are usually observed at different points in time. Wireless sensors in the Internet of things often transmit information only when a state changes to conserve battery life. There are many more examples in climatology, ecology, high-frequency finance, geology, and signal processing.

Analysis

A common approach to analyzing unevenly spaced time series is to transform the data into equally spaced observations using some form of interpolation - most often linear - and then to apply existing methods for equally spaced data. However, transforming data in such a way can introduce a number of significant and hard to quantify biases, [1] [2] [3] [4] [5] especially if the spacing of observations is highly irregular.

Ideally, unevenly spaced time series are analyzed in their unaltered form. However, most of the basic theory for time series analysis was developed at a time when limitations in computing resources favored an analysis of equally spaced data, since in this case efficient linear algebra routines can be used and many problems have an explicit solution. As a result, fewer methods currently exist specifically for analyzing unevenly spaced time series data. [5] [6] [7] [8] [9] [10] [11]

The least-squares spectral analysis methods are commonly used for computing a frequency spectrum from such time series without any data alterations.

Software

See also

Related Research Articles

Digital signal processing (DSP) is the use of digital processing, such as by computers or more specialized digital signal processors, to perform a wide variety of signal processing operations. The digital signals processed in this manner are a sequence of numbers that represent samples of a continuous variable in a domain such as time, space, or frequency. In digital electronics, a digital signal is represented as a pulse train, which is typically generated by the switching of a transistor.

<span class="mw-page-title-main">Fourier analysis</span> Branch of mathematics

In mathematics, Fourier analysis is the study of the way general functions may be represented or approximated by sums of simpler trigonometric functions. Fourier analysis grew from the study of Fourier series, and is named after Joseph Fourier, who showed that representing a function as a sum of trigonometric functions greatly simplifies the study of heat transfer.

Harmonic analysis is a branch of mathematics concerned with investigating the connections between a function and its representation in frequency. The frequency representation is found by using the Fourier transform for functions on the real line, or by Fourier series for periodic functions. Generalizing these transforms to other domains is generally called Fourier analysis, although the term is sometimes used interchangeably with harmonic analysis. Harmonic Analysis has become a vast subject with applications in areas as diverse as number theory, representation theory, signal processing, quantum mechanics, tidal analysis and neuroscience.

<span class="mw-page-title-main">Signal processing</span> Analysing, modifying and creating signals

Signal processing is an electrical engineering subfield that focuses on analyzing, modifying and synthesizing signals, such as sound, images, potential fields, seismic signals, altimetry processing, and scientific measurements. Signal processing techniques are used to optimize transmissions, digital storage efficiency, correcting distorted signals, subjective video quality and to also detect or pinpoint components of interest in a measured signal.

<span class="mw-page-title-main">Cepstrum</span>

In Fourier analysis, the cepstrum is the result of computing the inverse Fourier transform (IFT) of the logarithm of the estimated signal spectrum. The method is a tool for investigating periodic structures in frequency spectra. The power cepstrum has applications in the analysis of human speech.

<span class="mw-page-title-main">Wavelet</span> Function for integral Fourier-like transform

A wavelet is a wave-like oscillation with an amplitude that begins at zero, increases or decreases, and then returns to zero one or more times. Wavelets are termed a "brief oscillation". A taxonomy of wavelets has been established, based on the number and direction of its pulses. Wavelets are imbued with specific properties that make them useful for signal processing.

Spectral analysis or spectrum analysis is analysis in terms of a spectrum of frequencies or related quantities such as energies, eigenvalues, etc. In specific areas it may refer to:

<span class="mw-page-title-main">Frequency domain</span> Signal representation

In mathematics, physics, electronics, control systems engineering, and statistics, the frequency domain refers to the analysis of mathematical functions or signals with respect to frequency, rather than time. Put simply, a time-domain graph shows how a signal changes over time, whereas a frequency-domain graph shows how much of the signal lies within each given frequency band over a range of frequencies. A frequency-domain representation can also include information on the phase shift that must be applied to each sinusoid in order to be able to recombine the frequency components to recover the original time signal.

<span class="mw-page-title-main">Time series</span> Sequence of data points over time

In mathematics, a time series is a series of data points indexed in time order. Most commonly, a time series is a sequence taken at successive equally spaced points in time. Thus it is a sequence of discrete-time data. Examples of time series are heights of ocean tides, counts of sunspots, and the daily closing value of the Dow Jones Industrial Average.

In signal processing, a periodogram is an estimate of the spectral density of a signal. The term was coined by Arthur Schuster in 1898. Today, the periodogram is a component of more sophisticated methods. It is the most common tool for examining the amplitude vs frequency characteristics of FIR filters and window functions. FFT spectrum analyzers are also implemented as a time-sequence of periodograms.

A cyclostationary process is a signal having statistical properties that vary cyclically with time. A cyclostationary process can be viewed as multiple interleaved stationary processes. For example, the maximum daily temperature in New York City can be modeled as a cyclostationary process: the maximum temperature on July 21 is statistically different from the temperature on December 20; however, it is a reasonable approximation that the temperature on December 20 of different years has identical statistics. Thus, we can view the random process composed of daily maximum temperatures as 365 interleaved stationary processes, each of which takes on a new value once per year.

<span class="mw-page-title-main">Volatility (finance)</span> Degree of variation of a trading price series over time

In finance, volatility is the degree of variation of a trading price series over time, usually measured by the standard deviation of logarithmic returns.

In statistical signal processing, the goal of spectral density estimation (SDE) or simply spectral estimation is to estimate the spectral density of a signal from a sequence of time samples of the signal. Intuitively speaking, the spectral density characterizes the frequency content of the signal. One purpose of estimating the spectral density is to detect any periodicities in the data, by observing peaks at the frequencies corresponding to these periodicities.

<span class="mw-page-title-main">Least-squares spectral analysis</span> Periodicity computation method

Least-squares spectral analysis (LSSA) is a method of estimating a frequency spectrum based on a least-squares fit of sinusoids to data samples, similar to Fourier analysis. Fourier analysis, the most used spectral method in science, generally boosts long-periodic noise in the long and gapped records; LSSA mitigates such problems. Unlike in Fourier analysis, data need not be equally spaced to use LSSA.

The Hilbert–Huang transform (HHT) is a way to decompose a signal into so-called intrinsic mode functions (IMF) along with a trend, and obtain instantaneous frequency data. It is designed to work well for data that is nonstationary and nonlinear. In contrast to other common transforms like the Fourier transform, the HHT is an algorithm that can be applied to a data set, rather than a theoretical tool.

<span class="mw-page-title-main">Kolmogorov–Zurbenko filter</span>

Within statistics, the Kolmogorov–Zurbenko (KZ) filter was first proposed by A. N. Kolmogorov and formally defined by Zurbenko. It is a series of iterations of a moving average filter of length m, where m is a positive, odd integer. The KZ filter belongs to the class of low-pass filters. The KZ filter has two parameters, the length m of the moving average window and the number of iterations k of the moving average itself. It also can be considered as a special window function designed to eliminate spectral leakage.

Multidimension spectral estimation is a generalization of spectral estimation, normally formulated for one-dimensional signals, to multidimensional signals or multivariate data, such as wave vectors.

High frequency data refers to time-series data collected at an extremely fine scale. As a result of advanced computational power in recent decades, high frequency data can be accurately collected at an efficient rate for analysis. Largely used in financial analysis and in high frequency trading, high frequency data provides intraday observations that can be used to understand market behaviors, dynamics, and micro-structures.

<span class="mw-page-title-main">Directional-change intrinsic time</span>

Directional-change intrinsic time is an event-based operator to dissect a data series into a sequence of alternating trends of defined size .

References

  1. Myron Scholes; Joseph Williams (1977). "Estimating betas from nonsynchronous data". Journal of Financial Economics. 5 (3): 309–327. doi:10.1016/0304-405X(77)90041-1.
  2. Mark C. Lundin; Michel M. Dacorogna; Ulrich A. Müller (1999). "Chapter 5: Correlation of High Frequency Financial Time Series". In Pierre Lequex (ed.). The Financial Markets Tick by Tick. pp. 91–126.
  3. Takaki Hayashi; Nakahiro Yoshida (2005). "On covariance estimation of non-synchronously observed diffusion processes". Bernoulli. 11 (2): 359–379. doi: 10.3150/bj/1116340299 .
  4. K. Rehfeld; N. Marwan; J. Heitzig; J. Kurths (2011). "Comparison of correlation analysis techniques for irregularly sampled time series" (PDF). Nonlinear Processes in Geophysics. 18 (3): 389–404. doi: 10.5194/npg-18-389-2011 .
  5. 1 2 Andreas Eckner (2014), A Framework for the Analysis of Unevenly-Spaced Time Series Data (PDF)
  6. Ulrich A. Müller (1991). "Specially Weighted Moving Averages with Repeated Application of the EMA Operator" (PDF). Working Paper, Olsen and Associates, Zurich, Switzerland.
  7. Gilles Zumbach; Ulrich A. Müller (2001). "Operators on Inhomogeneous Time Series". International Journal of Theoretical and Applied Finance. 4: 147–178. doi:10.1142/S0219024901000900. Preprint
  8. Michel M. Dacorogna; Ramazan Gençay; Ulrich A. Müller; Richard B. Olsen; Olivier V. Pictet (2001). An Introduction to High-Frequency Finance (PDF). Academic Press.
  9. Andreas Eckner (2017), Algorithms for Unevenly-Spaced Time Series: Moving Averages and Other Rolling Operators (PDF)
  10. Andreas Eckner (2017), A Note on Trend and Seasonality Estimation for Unevenly-Spaced Time Series (PDF)
  11. Mehmet Süzen; Alper Yegenoglu (13 December 2021). "Generalised learning of time-series: Ornstein-Uhlenbeck processes". arXiv: 1910.09394 [stat.ML].