Exogeneity in Error Correction Models

Exogeneity in Error Correction Models
Author: Jean-Pierre Urbain
Publisher: Springer Science & Business Media
Total Pages: 201
Release: 2012-12-06
Genre: Business & Economics
ISBN: 3642957064

In the recent years, the study of cointegrated time series and the use of error correction models have become extremely popular in the econometric literature. This book provides an analysis of the notion of (weak) exogeneity, which is necessary to sustain valid inference in sub-systems, inthe framework of error correction models (ECMs). In many practical situations, the applied econometrician wants to introduce "structure" on his/her model in order to get economically meaningful coefficients. For thispurpose, ECMs in structural form provide an appealing framework, allowing the researcher to introduce (theoretically motivated) identification restrictions on the long run relationships. In this case, the validity of the inference will depend on a number of conditions which are investigated here. In particular,we point out that orthogonality tests, often used to test for weak exogeneity or for general misspecification, behave poorly in finite samples and are often not very useful in cointegrated systems.

Testing Exogeneity

Testing Exogeneity
Author: Neil R. Ericsson
Publisher:
Total Pages: 436
Release: 1994
Genre: Business & Economics
ISBN: 9780198774044

This book discusses the nature of exogeneity, a central concept in standard econometrics texts, and shows how to test for it through numerous substantive empirical examples from around the world, including the UK, Argentina, Denmark, Finland, and Norway. Part I defines terms and provides the necessary background; Part II contains applications to models of expenditure, money demand, inflation, wages and prices, and exchange rates; and Part III extends various tests of constancy and forecast accuracy, which are central to testing super exogeneity. About the Series Advanced Texts in Econometrics is a distinguished and rapidly expanding series in which leading econometricians assess recent developments in such areas as stochastic probability, panel and time series data analysis, modeling, and cointegration. In both hardback and affordable paperback, each volume explains the nature and applicability of a topic in greater depth than possible in introductory textbooks or single journal articles. Each definitive work is formatted to be as accessible and convenient for those who are not familiar with the detailed primary literature.

Econometric Analysis of Financial Markets

Econometric Analysis of Financial Markets
Author: Jürgen Kaehler
Publisher: Springer Science & Business Media
Total Pages: 232
Release: 2012-12-06
Genre: Business & Economics
ISBN: 3642486665

This collection of papers represents the state of the art in the applicationof recent econometric methods to the analysis of financial markets. From a methodological point of view the main emphasis is on cointegration analysis and ARCH modelling. In cointegration analysis the links between long-runcomponents of time series are studied. The methods used can be applied to the determination of equilibrium relationships between the variables, whereas ARCH models are concerned with the measurement and analysis of changing variances in time series. These econometric models have been the most significant innovations for the empirical analysis of financial time series in recent years. Other econometric methods and models applied in the papers include factor analysis, vector autoregressions, and Markov-switching models. The papers cover a wide range of issues and theories in financial and international economics: the term structure ofinterest rates, exchange-rate determination, target-zone dynamics, stock-market efficiency, and option pricing.

Co-integration, Error Correction, and the Econometric Analysis of Non-Stationary Data

Co-integration, Error Correction, and the Econometric Analysis of Non-Stationary Data
Author: Anindya Banerjee
Publisher: Oxford University Press
Total Pages: 344
Release: 1993-05-27
Genre: Business & Economics
ISBN: 0191638919

This book provides a wide-ranging account of the literature on co-integration and the modelling of integrated processes (those which accumulate the effects of past shocks). Data series which display integrated behaviour are common in economics, although techniques appropriate to analysing such data are of recent origin and there are few existing expositions of the literature. This book focuses on the exploration of relationships among integrated data series and the exploitation of these relationships in dynamic econometric modelling. The concepts of co-integration and error-correction models are fundamental components of the modelling strategy. This area of time-series econometrics has grown in importance over the past decade and is of interest to econometric theorists and applied econometricians alike. By explaining the important concepts informally, but also presenting them formally, the book bridges the gap between purely descriptive and purely theoretical accounts of the literature. The asymptotic theory of integrated processes is described and the tools provided by this theory are used to develop the distributions of estimators and test statistics. Practical modelling advice, and the use of techniques for systems estimation, are also emphasized. A knowledge of econometrics, statistics, and matrix algebra at the level of a final-year undergraduate or first-year undergraduate course in econometrics is sufficient for most of the book. Other mathematical tools are described as they occur.

Econometric Modelling with Time Series

Econometric Modelling with Time Series
Author: Vance Martin
Publisher: Cambridge University Press
Total Pages: 925
Release: 2013
Genre: Business & Economics
ISBN: 0521139813

"Maximum likelihood estimation is a general method for estimating the parameters of econometric models from observed data. The principle of maximum likelihood plays a central role in the exposition of this book, since a number of estimators used in econometrics can be derived within this framework. Examples include ordinary least squares, generalized least squares and full-information maximum likelihood. In deriving the maximum likelihood estimator, a key concept is the joint probability density function (pdf) of the observed random variables, yt. Maximum likelihood estimation requires that the following conditions are satisfied. (1) The form of the joint pdf of yt is known. (2) The specification of the moments of the joint pdf are known. (3) The joint pdf can be evaluated for all values of the parameters, 9. Parts ONE and TWO of this book deal with models in which all these conditions are satisfied. Part THREE investigates models in which these conditions are not satisfied and considers four important cases. First, if the distribution of yt is misspecified, resulting in both conditions 1 and 2 being violated, estimation is by quasi-maximum likelihood (Chapter 9). Second, if condition 1 is not satisfied, a generalized method of moments estimator (Chapter 10) is required. Third, if condition 2 is not satisfied, estimation relies on nonparametric methods (Chapter 11). Fourth, if condition 3 is violated, simulation-based estimation methods are used (Chapter 12). 1.2 Motivating Examples To highlight the role of probability distributions in maximum likelihood estimation, this section emphasizes the link between observed sample data and 4 The Maximum Likelihood Principle the probability distribution from which they are drawn"-- publisher.

International Trade Modelling

International Trade Modelling
Author: m Dagenais
Publisher: CRC Press
Total Pages: 386
Release: 1992-06-01
Genre: Business & Economics
ISBN: 9780412450006

In recent years, international trade has become a subject of increaed practical importance and also one of the most intellectually exciting parts of economics. In his introduction to this volume, Paul Krugman outlines why this is so, by analysing the original contribution of the New Trade Theory in interpreting and explaining the observed trade behaviour of the past twenty years. Then follow sections which discuss: formal tests of the New Trade Theory, Price Discrimination and Exchange Rate, as well as New Protectionism, measures of Comparative Advantages and Import Demand in industrialized and developing countries. Some chapters also use GCE models to evaluate Trade Protectionism, while others encompass External Trade within aggregate Disequilibrium Models.