Bayesian Econometrics / Edition 1

Paperback (Print)
Buy New
Buy New from BN.com
$87.98
Used and New from Other Sellers
Used and New from Other Sellers
from $30.75
Usually ships in 1-2 business days
(Save 67%)
Other sellers (Paperback)
  • All (18) from $30.75   
  • New (8) from $56.07   
  • Used (10) from $30.75   

Overview

Researchers in many fields are increasingly finding the Bayesian approach to statistics to be an attractive one. This book introduces the reader to the use of Bayesian methods in the field of econometrics at the advanced undergraduate or graduate level. The book is self-contained and does not require that readers have previous training in econometrics. The focus is on models used by applied economists and the computational techniques necessary to implement Bayesian methods when doing empirical work. Topics covered in the book include the regression model (and variants applicable for use with panel data), time series models, models for qualitative or censored data, nonparametric methods and Bayesian model averaging. The book includes numerous empirical examples and the website associated with it contains data sets and computer programs to help the student develop the computational skills of modern Bayesian econometrics.

Read More Show Less

Product Details

  • ISBN-13: 9780470845677
  • Publisher: Wiley
  • Publication date: 7/16/2003
  • Edition description: New Edition
  • Edition number: 1
  • Pages: 376
  • Product dimensions: 9.61 (w) x 6.69 (h) x 0.78 (d)

Meet the Author

Gary Koop is Professor of Economics at the University of Glasgow.

Read More Show Less

Read an Excerpt

Bayesian Econometrics


By Gary Koop

John Wiley & Sons

ISBN: 0-470-84567-8


Chapter One

An Overview of Bayesian Econometrics

1.1 BAYESIAN THEORY

Bayesian econometrics is based on a few simple rules of probability. This is one of the chief advantages of the Bayesian approach. All of the things that an econometrician would wish to do, such as estimate the parameters of a model, compare different models or obtain predictions from a model, involve the same rules of probability. Bayesian methods are, thus, universal and can be used any time a researcher is interested in using data to learn about a phenomenon.

To motivate the simplicity of the Bayesian approach, let us consider two random variables, A and B. The rules of probability imply:

p(A, B) = p(A|B)p(B) where p(A, B) is the joint probability of A and B occurring, p(A|B) is the probability of A occurring conditional on B having occurred (i.e. the conditional probability of A given B), and p(B) is the marginal probability of B. Alternatively, we can reverse the roles of A and B and find an expression for the joint probability of A and B:

p(A, B) = p(A|B)p(B)

Equating these two expressions for p(A, B) and rearranging provides us with Bayes'rule, which lies at the heart of Bayesian econometrics:

(1.1) p(B|A) = p(A|B)p(B) / p(A)

Econometrics is concerned with using data to learn about something the researcher is interested in. Just what the 'something' is depends upon the context. However, in economics we typically work with models which depend upon parameters. For the reader with some previous training in econometrics, it might be useful to have in mind the regression model. In this model interest often centers on the coefficients in the regression, and the researcher is interested in estimating these coefficients. In this case, the coefficients are the parameters under study. Let y be a vector or matrix of data and [theta] be a vector or matrix which contains the parameters for a model which seeks to explain y. We are interested in learning about [theta] based on the data, y. Bayesian econometrics uses Bayes' rule to do so. In other words, the Bayesian would replace B by [theta] and A by y in (1.1) to obtain:

(1.2) p([theta]|y) = p([y]|theta) p([theta]) / p(y)

Bayesians treat p([theta]|y) as being of fundamental interest. That is, it directly addresses the question "Given the data, what do we know about [theta]?". The treatment of [theta] as a random variable is controversial among some econometricians. The chief competitor to Bayesian econometrics, often called frequentist econometrics, says that [theta] is not a random variable. However, Bayesian econometrics is based on a subjective view of probability, which argues that our uncertainty about anything unknown can be expressed using the rules of probability. In this book, we will not discuss such methodological issues (see Poirier (1995) for more detail). Rather, we will take it as given that econometrics involves learning about something unknown (e.g. coefficients in a regression) given something known (e.g. data) and the conditional probability of the unknown given the known is the best way of summarizing what we have learned.

Having established that p([theta]|y) is of fundamental interest for the econometrician interested in using data to learn about parameters in a model, let us now return to (1.2). Insofar as we are only interested in learning about [theta], we can ignore the term p(y), since it does not involve [theta]. We can then write:

(1.3) p([theta]|y) [varies] p(y|[theta]) p([theta]) The term p([theta]|y) is referred to as the posterior density, the p.d.f. for the data given the parameters of the model, p(y|[theta]), as the likelihood function and p([theta]) as the prior density. You often hear this relationship referred to as "posterior is proportional to likelihood times prior". At this stage, this may seem a little abstract, and the manner in which priors and likelihoods are developed to allow for the calculation of the posterior may be unclear. Things should become clearer to you in the following chapters, where we will develop likelihood functions and priors in specific contexts. Here we provide only a brief general discussion of what these are.

The prior, p([theta]), does not depend upon the data. Accordingly, it contains any non-data information available about [theta]. In other words, it summarizes what you know about [theta] prior to seeing the data. As an example, suppose [theta] is a parameter which reflects returns to scale in a production process. In many cases, it is reasonable to assume that returns to scale are roughly constant. Thus, before you look at the data, you have prior information about [theta], in that you would expect it to be approximately one. Prior information is a controversial aspect of Bayesian methods. In this book, we will discuss both informative and noninformative priors for various models. In addition, in later chapters, we will discuss empirical Bayes methods. These use data-based information to choose the prior and, hence, violate a basic premise of Bayesian methods. Nevertheless, empirical Bayes methods are becoming increasingly popular for the researcher who is interested in practical, objective, tools that seem to work well in practice.

The likelihood function, p(y|[theta]), is the density of the data conditional on the parameters of the model. It is often referred to as the data generating process. For instance, in the linear regression model (which will be discussed in the next chapter), it is common to assume that the errors have a Normal distribution. This implies that p(y|[theta]) is a Normal density, which depends upon parameters (i.e. the regression coefficients and the error variance).

The posterior, p([theta]|y), is the density which is of fundamental interest. It summarizes all we know about [theta] after (i.e. posterior to) seeing the data. Equation (1.3) can be thought of as an updating rule, where the data allows us to update our prior views about [theta]. The result is the posterior which combines both data and non-data information.

In addition to learning about parameters of a model, an econometrician might be interested in comparing different models. A model is formally defined by a likelihood function and a prior. Suppose we have m different models, [M.sub.i] for i = 1; ..., m, which all seek to explain y. [M.sub.i] depends upon parameters [[theta].sup.i]. In cases where many models are being entertained, it is important to be explicit about which model is under consideration. Hence, the posterior for the parameters calculated using [M.sub.i] is written as

(1.4) p([[theta].sup.i]|y, [M.sub.i]) = p(y|[[theta].sup.i], [M.sub.i]p([[theta].sup.i]|[M.sub.i] / p(y|[M.sub.i]) (1.4)

and the notation makes clear that we now have a posterior, likelihood, and prior for each model.

The logic of Bayesian econometrics suggests that we use Bayes' rule to derive a probability statement about what we do not know (i.e. whether a model is a correct one or not) conditional on what we do know (i.e. the data). This means the posterior model probability can be used to assess the degree of support for [M.sub.i] . Using (1.1) with B = [M.sub.i] and A = y, we obtain

(1.5) p([M.sub.i]|y) = p(y|[M.sub.i]) p([M.sub.i]) / p(y)

Of the terms in (1.5), p(M.sub.i) is referred to as the prior model probability. Since it does not involve the data, it measures how likely we believe [M.sub.i] to be the correct one before seeing the data. p(y| [M.sub.i]) is called the marginal likelihood, and is calculated using (1.4) and a few simple manipulations. In particular, if we integrate both sides of (1.4) with respect to [[theta].sup.i], use the fact that [integral] p([[theta].sup.i]|y, [M.sub.i])d[[theta].sup.i] = 1 (since probability density functions integrate to one), and rearrange, we obtain:

(1.6) p(y|[M.sub.i]) = [integral] p(y|[[theta].sup.i], [M.sup.i])p([[theta].sup.i] |[M.sub.i])d[[theta].sup.i] (1.6)

Note that the marginal likelihood depends only upon the prior and the likelihood. In subsequent chapters, we discuss how (1.6) can be calculated in practice.

Since the denominator in (1.5) is often hard to calculate directly, it is common to compare two models, i and j, using the posterior odds ratio, which is simply the ratio of their posterior model probabilities:

(1.7) P[O.sub.ij] = p([M.sub.i]|y) / p([M.sub.j]|y) = p(y|[M.sub.i] p([M.sub.i] p(y|[M.sub.j] p([M.sub.j]

Note that, since p(y) is common to both models, it cancels out when we take the ratio. As we will discuss in subsequent chapters, there are special techniques in many cases for calculating the posterior odds ratio directly. If we calculate the posterior odds ratio comparing every pair of models, and we assume that our set of models is exhaustive (in that p([M.sub.1]|y) + p([M.sub.2]|y) + ... + p([M.sub.m]|y = 1), then we can use posterior odds ratios to calculate the posterior model probabilities given in (1.5). For instance, if we have m = 2 models then we can use the two equations

p([M.sub.1]|y) + p([M.sub.2]|y = 1

and

P[O.sub.12] = p([M.sub.1]|y) / p([M.sub.2]|y

to work out

p([M.sub.1]|y) = P]O.sub.12] / 1 + P]O.sub.12]

and

p([M.sub.2]|y = 1 - p([M.sub.1]|y)

Thus, knowledge of the posterior odds ratio allows us to figure out the posterior model probabilities.

To introduce some more jargon, econometricians may be interested in model comparison when equal prior weight is attached to each model. That is, p([M.sub.i]) = p([M.sub.j]) or, equivalently, the prior odds ratio which is p([M.sub.i]) / p([M.sub.j]) / is set to one. In this case, the posterior odds ratio becomes simply the ratio of marginal likelihoods, and is given a special name, the Bayes Factor, defined as:

(1.8) B[F.sub.ij] = p(y|[M.sub.i]) / p(y|[M.sub.j])

Finally, econometricians are often interested in prediction. That is, given the observed data, y, the econometrician may be interested in predicting some future unobserved data y*. Our Bayesian reasoning says that we should summarize our uncertainty about what we do not know (i.e. y*) through a conditional probability statement. That is, prediction should be based on the predictive density p(y*|y) (or, if we have many models, we would want to make explicit the dependence of a prediction on a particular model, and write p(y*|y, [M.sub.i])). Using a few simple rules of probability, we can write p(y|y*) in a convenient form. In particular, since a marginal density can be obtained from a joint density through integration (see Appendix B), we can write:

p(y*|y) = [integral] p(y*, [theta]|y/d[theta]

However, the term inside the integral can be rewritten using another simple rule of probability:

(1.9) p(y*|y) = [integral] p(y*|y, [theta]) p([theta]|y) d[theta]

As we shall see in future chapters, the form for the predictive in (1.9) is quite convenient, since it involves the posterior.

On one level, this book could end right here. These few pages have outlined all the basic theoretical concepts required for the Bayesian to learn about parameters, compare models and predict. We stress what an enormous advantage this is. Once you accept that unknown things (i.e. [theta], [M.sub.i] and y*) are random variables, the rest of Bayesian approach is non-controversial. It simply uses the rules of probability, which are mathematically true, to carry out statistical inference. A benefit of this is that, if you keep these simple rules in mind, it is hard to lose sight of the big picture. When facing a new model (or reading a new chapter in the book), just remember that Bayesian econometrics requires selection of a prior and a likelihood. These can then be used to form the posterior, (1.3), which forms the basis for all inference about unknown parameters in a model. If you have many models and are interested in comparing them, you can use posterior model probabilities (1.5), posterior odds ratios (1.7), or Bayes Factors (1.8). To obtain any of these, we usually have to calculate the marginal likelihood (1.6).

Continues...


Excerpted from Bayesian Econometrics by Gary Koop Excerpted by permission.
All rights reserved. No part of this excerpt may be reproduced or reprinted without permission in writing from the publisher.
Excerpts are provided by Dial-A-Book Inc. solely for the personal use of visitors to this web site.

Read More Show Less

Table of Contents

Preface.

1.  An Overview of Bayesian Econometrics.

2.  The Normal Linear Regression Model with Natural Conjugate Prior and a Single Explanatory Variable.

3.  The Normal Linear Regression Model with Natural Conjugate Prior and Many Explanatory Variables.

4.  The Normal Linear Regression Model with Other Priors.

5.  The Nonlinear Regression Model.

6.  The Linear Regression Model with General Error Covariance Matrix.

7.  The Linear Regression Model with Panel Data.

8.  Introduction to Time Series: State Space Models.

9.  Qualitative and Limited Dependent Variable Models.

10.  Flexible Models: Nonparametric and Semi-Parametric Methods.

11.  Bayesian Model Averaging.

12.  Other Models, Methods and Issues.

Appendix A: Introduction to Matrix Algebra.

Appendix B: Introduction to Probability and Statistics.

Bibliography.

Index.

Read More Show Less

Customer Reviews

Be the first to write a review
( 0 )
Rating Distribution

5 Star

(0)

4 Star

(0)

3 Star

(0)

2 Star

(0)

1 Star

(0)

Your Rating:

Your Name: Create a Pen Name or

Barnes & Noble.com Review Rules

Our reader reviews allow you to share your comments on titles you liked, or didn't, with others. By submitting an online review, you are representing to Barnes & Noble.com that all information contained in your review is original and accurate in all respects, and that the submission of such content by you and the posting of such content by Barnes & Noble.com does not and will not violate the rights of any third party. Please follow the rules below to help ensure that your review can be posted.

Reviews by Our Customers Under the Age of 13

We highly value and respect everyone's opinion concerning the titles we offer. However, we cannot allow persons under the age of 13 to have accounts at BN.com or to post customer reviews. Please see our Terms of Use for more details.

What to exclude from your review:

Please do not write about reviews, commentary, or information posted on the product page. If you see any errors in the information on the product page, please send us an email.

Reviews should not contain any of the following:

  • - HTML tags, profanity, obscenities, vulgarities, or comments that defame anyone
  • - Time-sensitive information such as tour dates, signings, lectures, etc.
  • - Single-word reviews. Other people will read your review to discover why you liked or didn't like the title. Be descriptive.
  • - Comments focusing on the author or that may ruin the ending for others
  • - Phone numbers, addresses, URLs
  • - Pricing and availability information or alternative ordering information
  • - Advertisements or commercial solicitation

Reminder:

  • - By submitting a review, you grant to Barnes & Noble.com and its sublicensees the royalty-free, perpetual, irrevocable right and license to use the review in accordance with the Barnes & Noble.com Terms of Use.
  • - Barnes & Noble.com reserves the right not to post any review -- particularly those that do not follow the terms and conditions of these Rules. Barnes & Noble.com also reserves the right to remove any review at any time without notice.
  • - See Terms of Use for other conditions and disclaimers.
Search for Products You'd Like to Recommend

Recommend other products that relate to your review. Just search for them below and share!

Create a Pen Name

Your Pen Name is your unique identity on BN.com. It will appear on the reviews you write and other website activities. Your Pen Name cannot be edited, changed or deleted once submitted.

 
Your Pen Name can be any combination of alphanumeric characters (plus - and _), and must be at least two characters long.

Continue Anonymously

    If you find inappropriate content, please report it to Barnes & Noble
    Why is this product inappropriate?
    Comments (optional)