# Time Series Analysis

With its broad coverage of methodology, this comprehensive book is a useful learning and reference tool for those in applied sciences where analysis and research of time series is useful. Its plentiful examples show the operational details and purpose of a variety of univariate and multivariate time series methods. Numerous figures, tables and real-life time series data sets illustrate the models and methods useful for analyzing, modeling, and forecasting data collected sequentially in time. The text also offers a balanced treatment between theory and applications. Time Series Analysis is a thorough introduction to both time-domain and frequency-domain analyses of univariate and multivariate time series methods, with coverage of the most recently developed techniques in the field.

- Author:
- William W. S. Wei
- Format:
- Paperback

Out of stock

#### Description

With its broad coverage of methodology, this comprehensive book is a useful learning and reference tool for those in applied sciences where analysis and research of time series is useful. Its plentiful examples show the operational details and purpose of a variety of univariate and multivariate time series methods. Numerous figures, tables and real-life time series data sets illustrate the models and methods useful for analyzing, modeling, and forecasting data collected sequentially in time. The text also offers a balanced treatment between theory and applications. Time Series Analysis is a thorough introduction to both time-domain and frequency-domain analyses of univariate and multivariate time series methods, with coverage of the most recently developed techniques in the field.

#### Product details

- Author:
- William W. S. Wei
- Publisher:
- Addison Wesley Longman
- Edition:
- Revised edition (2)
- ISBN:
- 9780321322166
- Audience:
- Professional
- Pages:
- 624
- Width (mm):
- 172
- Length (mm):
- 228
- Table of Contents:
- 1: Overview 1.1 Introduction 1.2 Examples and Scope of This Book 2: Fundamental Concepts 2.1 Stochastic Processes 2.2 The Autocovariance and Autocorrelation Functions 2.3 The Partial Autocorrelation Function 2.4 White Noise Processes 2.5 Estimation of the Mean, Autocovariances, and Autocorrelations 2.5.1 Sample Mean 2.5.2 Sample Autocovariance Function 2.5.3 Sample Autocorrelation Function 2.5.4 Sample Partial Autocorrelation Function 2.6 Moving Average and Autoregressive Representations of Time Series Processes 2.7 Linear Difference Equations 3: Stationary Time Series Models 3.1 Autoregressive Processes 3.1.1 The First-Order Autoregressive AR(1) Process 3.1.2 The Second-Order Autoregressive AR(2) Process 3.1.3 The General pth-Order Autoregressive AR(p) Process 3.2 Moving Average Processes 3.2.1 The First-Order Moving Average MA(1) Process 3.2.2 The Second-Order Moving Average MA(2) Process 3.2.3 The General qth-Order Moving Average MA(q) Process 3.3 The Dual Relationship Between AR(p) and MA(q) Processes 3.4 Autoregressive Moving Average ARMA(p, q) Processes 3.4.1 The General Mixed ARMA(p, q) Process 3.4.2 The ARMA(1, 1) Process 4: Nonstationary Time Series Models 4.1 Nonstationarity in the Mean 4.1.1 Deterministic Trend Models 4.1.2 Stochastic Trend Models and Differencing 4.2 Autoregressive Integrated Moving Average (ARIMA) Models 4.2.1 The General ARIMA Model 4.2.2 The Random Walk Model 4.2.3 The ARIMA(0, 1, 1) or IMA(1, 1) Model 4.3 Nonstationarity in the Variance and the Autocovariance 4.3.1 Variance and Autocovariance of the ARIMA Models 4.3.2 Variance Stabilizing Transformations 5: Forecasting 5.1 Introduction 5.2 Minimum Mean Square Error Forecasts 5.2.1 Minimum Mean Square Error Forecasts for ARMA Models 5.2.2 Minimum Mean Square Error Forecasts for ARIMA Models 5.3 Computation of Forecasts 5.4 The ARIMA Forecast as a Weighted Average of Previous Observations 5.5 Updating Forecasts 5.6 Eventual Forecast Functions 5.7 A Numerical Example 6: Model Identification 6.1 Steps for Model Identification 6.2 Empirical Examples 6.3 The Inverse Autocorrelation Function (IACF) 6.4 Extended Sample Autocorrelation Function and Other Identification Procedures 6.4.1 The Extended Sample Autocorrelation Function (ESACF) 6.4.2 Other Identification Procedures 7: Parameter Estimation, Diagnostic Checking, and Model Selection 7.1 The Method of Moments 7.2 Maximum Likelihood Method 7.2.1 Conditional Maximum Likelihood Estimation 7.2.2 Unconditional Maximum Likelihood Estimation and Backcasting Method 7.2.3 Exact Likelihood Functions 7.3 Nonlinear Estimation 7.4 Ordinary Least Squares (OLS) Estimation in Time Series Analysis 7.5 Diagnostic Checking 7.6 Empirical Examples for Series W1-W7 7.7 Model Selection Criteria 8: Seasonal Time Series Models 8.1 General Concepts 8.2 Traditional Methods 8.2.1 Regression Method 8.2.2 Moving Average Method 8.3 Seasonal ARIMA Models 8.4 Empirical Examples 9: Testing for a Unit Root 9.1 Introduction 9.2 Some Useful Limiting Distributions 9.3 Testing for a Unit Root in the AR(1) Model 9.3.1 Testing the AR(1) Model without a Constant Term 9.3.2 Testing the AR(1) Model with a Constant Term 9.3.3 Testing the AR(1) Model with a Linear Time Trend 9.4 Testing for a Unit Root in a More General Model 9.5 Testing for a Unit Root in Seasonal Time Series Models 9.5.1 Testing the Simple Zero Mean Seasonal Model 9.5.2 Testing the General Multiplicative Zero Mean Seasonal Model 10: Intervention Analysis and Outlier Detection 10.1 Intervention Models 10.2 Examples of Intervention Analysis 10.3 Time Series Outliers 10.3.1 Additive and Innovational Outliers 10.3.2 Estimation of the Outlier Effect When the Timing of the Outlier Is Known 10.3.3 Detection of Outliers Using an Iterative Procedure 10.4 Examples of Outlier Analysis 10.5 Model Identification in the Presence of Outliers 11: Fourier Analysis 11.1 General Concepts 11.2 Orthogonal Functions 11.3 Fourier Representation of Finite Sequences 11.4 Fourier Representation of Periodic Sequences 11.5 Fourier Representation of Nonperiodic Sequences: The Discrete-Time Fourier Transform 11.6 Fourier Representation of Continuous-Time Functions 11.6.1 Fourier Representation of Periodic Functions 11.6.2 Fourier Representation of Nonperiodic Functions: The Continuous-Time Fourier Transform 11.7 The Fast Fourier Transform 12: Spectral Theory of Stationary Processes 12.1 The Spectrum 12.1.1 The Spectrum and Its Properties 12.1.2 The Spectral Representation of Autocovariance Functions: The Spectral Distribution Function 12.1.3 Wold's Decomposition of a Stationary Process 12.1.4 The Spectral Representation of Stationary Processes 12.2 The Spectrum of Some Common Processes 12.2.1 The Spectrum and the Autocovariance Generating Function 12.2.2 The Spectrum of ARMA Models 12.2.3 The Spectrum of the Sum of Two Independent Processes 12.2.4 The Spectrum of Seasonal Models 12.3 The Spectrum of Linear Filters 12.3.1 The Filter Function 12.3.2 Effect of Moving Average 12.3.3 Effect of Differencing 12.4 Aliasing 13: Estimation of the Spectrum 13.1 Periodogram Analysis 13.1.1 The Periodogram 13.1.2 Sampling Properties of the Periodogram 13.1.3 Tests for Hidden Periodic Components 13.2 The Sample Spectrum 13.3 The Smoothed Spectrum 13.3.1 Smoothing in the Frequency Domain: The Spectral Window 13.3.2 Smoothing in the Time Domain: The Lag Window 13.3.3 Some Commonly Used Windows 13.3.4 Approximate Confidence Intervals for Spectral Ordinates 13.4 ARMA Spectral Estimation 14: Transfer Function Models 14.1 Single-Input Transfer Function Models 14.1.1 General Concepts 14.1.2 Some Typical Impulse Response Functions 14.2 The Cross-Correlation Function and Transfer Function Models 14.2.1 The Cross-Correlation Function (CCF) 14.2.2 The Relationship between the Cross-Correlation Function and the Transfer Function 14.3 Construction of Transfer Function Models 14.3.1 Sample Cross-Correlation Function 14.3.2 Identification of Transfer Function Models 14.3.3 Estimation of Transfer Function Models 14.3.4 Diagnostic Checking of Transfer Function Models 14.3.5 An Empirical Example 14.4 Forecasting Using Transfer Function Models 14.4.1 Minimum Mean Square Error Forecasts for Stationary Input and Output Series 14.4.2 Minimum Mean Square Error Forecasts for Nonstationary Input and Output Series 14.4.3 An Example 14.5 Bivariate Frequency-Domain Analysis 14.5.1 Cross-Covariance Generating Functions and the Cross-Spectrum 14.5.2 Interpretation of the Cross-Spectral Functions 14.5.3Examples 14.5.4 Estimation of the Cross-Spectrum 14.6 The Cross-Spectrum and Transfer Function Models 14.6.1 Construction of Transfer Function Models through Cross-Spectrum Analysis 14.6.2 Cross-Spectral Functions of Transfer Function Models 14.7 Multiple-Input Transfer Function Models 15: Time Series Regression and GARCH Models 15.1 Regression with Autocorrelated Errors 15.2 ARCH and GARCH Models 15.3 Estimation of GARCH Models 15.3.1 Maximum Likelihood Estimation 15.3.2 Iterative Estimation 15.4 Computation of Forecast Error Variance 15.5 Illustrative Examples 16: Vector Time Series Models 16.1 Covariance and Correlation Matrix Functions 16.2 Moving Average and Autoregressive Representations of Vector Processes 16.3 The Vector Autoregressive Moving Average Process 16.3.1 Covariance Matrix Function for the Vector AR(1) Model 16.3.2 Vector AR(p) Models 16.3.3 Vector MA(1) Models 16.3.4 Vector MA(q) Models 16.3.5 Vector ARMA(1, 1) Models 16.4 Nonstationary Vector Autoregressive Moving Average Models 16.5 Identification of Vector Time Series Models 16.5.1 Sample Correlation Matrix Function 16.5.2 Partial Autoregression Matrices 16.5.3 Partial Lag Correlation Matrix Function 16.6 Model Fitting and Forecasting 16.7 An Empirical Example 16.7.1 Model Identification 16.7.2 Parameter Estimation 16.7.3 Diagnostic Checking 16.7.4 Forecasting 16.7.5 Further Remarks 16.8 Spectral Properties of Vector Processes Supplement 16.A Multivariate Linear Regression Models 17: More on Vector Time Series 17.1 Unit Roots and Cointegration in Vector Processes 17.1.1 Representations of Nonstationary Cointegrated Processes 17.1.2 Decomposition of Zt 17.1.3 Testing and Estimating Cointegration 17.2 Partial Process and Partial Process Correlation Matrices 17.2.1 Covariance Matrix Generating Function 17.2.2 Partial Covariance Matrix Generating Function 17.2.3 Partial Process Sample Correlation Matrix Functions 17.2.4 An Empirical Example: The U.S. Hog Data 17.3 Equivalent Representations of a Vector ARMA Model 17.3.1 Finite-Order Representations of a Vector Time Series Process 17.3.2 Some Implications 18: State Space Models and the Kalman Filter 18.1 State Space Representation 18.2 The Relationship between State Space and ARMA Models 18.3 State Space Model Fitting and Canonical Correlation Analysis 18.4 Empirical Examples 18.5 The Kalman Filter and Its Applications Supplement 18.A Canonical Correlations 19: Long Memory and Nonlinear Processes 19.1 Long Memory Processes and Fractional Differencing 19.1.1 Fractionally Integrated ARMA Models and Their ACF 19.1.2 Practical Implications of the ARFIMA Processes 19.1.3 Estimation of the Fractional Difference 19.2 Nonlinear Processes 19.2.1 Cumulants, Polyspectrum, and Tests for Linearity and Normality 19.2.2 Some Nonlinear Time Series Models 19.3 Threshold Autoregressive Models 19.3.1 Tests for TAR Models 19.3.2 Modeling TAR Models 20: Aggregation and Systematic Sampling in Time Series 20.1 Temporal Aggregation of the ARIMA Process 20.1.1 The Relationship of Autocovariances between the Nonaggregate and Aggregate Series 20.1.2 Temporal Aggregation of the IMA(d, q) Process 20.1.3 Temporal Aggregation of the AR(p) Process 20.1.4 Temporal Aggregation of the ARIMA(p, d, q) Process 20.1.5 The Limiting Behavior of Time Series Aggregates 20.2 The Effects of Aggregation on Forecasting and Parameter Estimation 20.2.1 Hilbert Space 20.2.2 The Application of Hilbert Space in Forecasting 20.2.3 The Effect of Temporal Aggregation on Forecasting 20.2.4 Information Loss Due to Aggregation in Parameter Estimation 20.3 Systematic Sampling of the ARIMA Process 20.4 The Effects of Systematic Sampling and Temporal Aggregation on Causality 20.4.1 Decomposition of Linear Relationship between Two Time Series 20.4.2 An Illustrative Underlying Model 20.4.3 The Effects of Systematic Sampling and Temporal Aggregation on Causality 20.5 The Effects of Aggregation on Testing for Linearity and Normality 20.5.1 Testing for Linearity and Normality 20.5.2 The Effects of Temporal Aggregation on Testing for Linearity and Normality 20.6 The Effects of Aggregation on Testing for a Unit Root 20.6.1 The Model of Aggregate Series 20.6.2 The Effects of Aggregation on the Distribution of the Test Statistics 20.6.3 The Effects of Aggregation on the Significance Level and the Power of the Test 20.6.4Examples 20.6.5 General Cases and Concluding Remarks 20.7 Further Comments References Appendix Time Series Data Used for Illustrations Statistical Tables Author Index Subject Index
- Weight (g):
- 1020

### Look for Similar Items by Category

### Your Recent History

**Recently Viewed Items**