WHAT ARE TIMES SERIES








Read More Add your Comment 0 comments


INTRODUCTION TO OPERATION RESEARCH



INTRODUCTION
Although it is a distinct discipline in its own right, Operations Research (O.R.) has also become an integral part of the Industrial Engineering (I.E.) profession.  This is hardly a matter of surprise when one considers that they both share many of the same objectives, techniques and application areas.  O.R. as a formal subject is about fifty years old and its origins may be traced to the latter half of World War II.  Most of the O.R. techniques that are commonly used today were developed over (approximately) the first twenty years following its inception.  During the next thirty or so years the pace of development of fundamentally new O.R. methodologies has slowed somewhat.  However, there has been a rapid expansion in (1) the breadth of problem areas to which O.R. has been applied, and (2) in the magnitudes of the problems that can be addressed using O.R. methodologies.  Today, operations research is a mature, well-developed field with a sophisticated array of techniques that are used routinely to solve problems in a wide range of application areas

This chapter will provide an overview of O.R. from the perspective of an Industrial Engineer.  A brief review of its historical origins is first provided.  This is followed by a detailed discussion of the basic philosophy behind O.R. and the so-called “O.R. approach.”  The chapter concludes with several examples of successful applications to typical problems that might be faced by an Industrial Engineer.  Broadly speaking, an O.R. project comprises three steps: (1) building a model, (2) solving it, and (3) implementing the results.  The emphasis of this chapter is on the first and third steps.  The second step typically involves specific methodologies or techniques, which could be quite sophisticated and require significant mathematical development.  Several important methods are overviewed elsewhere in this handbook.  The reader who has an interest in learning more about these topics is referred to one of the many excellent texts on O.R. that are available today and that are listed under "Further Reading" at the end of this chapter, e.g., Hillier and Lieberman (1995), Taha (1997) or Winston (1994).
       A HISTORICAL PERSPECTIVE
While there is no clear date that marks the birth of O.R., it is generally accepted that the field originated in England during World War II.  The impetus for its origin was the development of radar defense systems for the Royal Air Force, and the first recorded use of the term Operations Research is attributed to a British Air Ministry official named A. P. Rowe who constituted teams to do “operational researches” on the communication system and the control room at a British radar station.  The studies had to do with improving the operational efficiency of systems (an objective which is still one of the cornerstones of modern O.R.).  This new approach of picking an “operational” system and conducting “research” on how to make it run more efficiently soon started to expand into other arenas of the war.  Perhaps the most famous of the groups involved in this effort was the one led by a physicist named P. M. S. Blackett which included physiologists, mathematicians, astrophysicists, and even a surveyor.  This multifunctional team focus of an operations research project group is one that has carried forward to this day.  Blackett’s biggest contribution was in convincing the authorities of the need for a scientific approach to manage complex operations, and indeed he is regarded in many circles as the original operations research analyst.
O.R. made its way to the United States a few years after it originated in England.  Its first presence in the U.S. was through the U.S. Navy’s Mine Warfare Operations Research Group; this eventually expanded into the Antisubmarine Warfare Operations Research Group that was led by Phillip Morse, which later became known simply as the Operations Research Group.  Like Blackett in Britain, Morse is widely regarded as the “father” of O.R. in the United States, and many of the distinguished scientists and mathematicians that he led went on after the end of the war to become the pioneers of O.R. in the United States.
In the years immediately following the end of World War II, O.R. grew rapidly as many scientists realized that the principles that they had applied to solve problems for the military were equally applicable to many problems in the civilian sector.  These ranged from short-term problems such as scheduling and inventory control to long-term problems such as strategic planning and resource allocation.  George Dantzig, who in 1947 developed the simplex algorithm for Linear Programming (LP), provided the single most important impetus for this growth.   To this day, LP remains one of the most widely used of all O.R. techniques and despite the relatively recent development of interior point methods as an alternative approach, the simplex algorithm (with numerous computational refinements) continues to be widely used.  The second major impetus for the growth of O.R. was the rapid development of digital computers over the next three decades.  The simplex method was implemented on a computer for the first time in 1950, and by 1960 such implementations could solve problems with about 1000 constraints.  Today, implementations on powerful workstations can routinely solve problems with hundreds of thousands of variables and constraints.  Moreover, the large volumes of data required for such problems can be stored and manipulated very efficiently.
Once the simplex method had been invented and used, the development of other methods followed at a rapid pace. The next twenty years witnessed the development of most of the O.R. techniques that are in use today including nonlinear, integer and dynamic programming, computer simulation, PERT/CPM, queuing theory, inventory models, game theory, and sequencing and scheduling algorithms.  The scientists who developed these methods came from many fields, most notably mathematics, engineering and economics.  It is interesting that the theoretical bases for many of these techniques had been known for years, e.g., the EOQ formula used with many inventory models was developed in 1915 by Harris, and many of the queuing formulae were developed by Erlang in 1917.  However, the period from 1950 to 1970 was when these were formally unified into what is considered the standard toolkit for an operations research analyst and successfully applied to problems of industrial significance.  The following section describes the approach taken by operations research in order to solve problems and explores how all of these methodologies fit into the O.R. framework


Read More Add your Comment 0 comments


What is Econometrics



Introduction




1.1   What is Econometrics?

The  term  “econometrics” is believed  to  have  been  crafted  by  Ragnar  Frisch  (1895-1973)  of Norway, one of the three principle founders of the Econometric Society, …rst editor of the journal Econometrica, and  co-winner  of the  …rst  Nobel Memorial  Prize  in Economic Sciences in 1969.  It is therefore …tting that we turn to Frisch’s own words in the introduction to the …rst issue of Econometrica for an explanation of the discipline.

A word of explanation regarding  the term  econometrics  may be in order.  Its de…ni- tion is implied in the statement of the scope of the [Econometric]  Society, in Section I of the Constitution, which reads:  “The  Econometric Society is an international society for the advancement of economic theory  in its relation  to statistics and mathematics.... Its main object shall be to promote studies that aim at a uni…cation of the theoretical- quantitative and the empirical-quantitative approach to economic problems....”
But there are several aspects of the quantitative approach to economics, and no single one of these aspects, taken by itself, should be confounded with econometrics.  Thus, econometrics  is by no means  the  same as economic statistics.  Nor is it identical  with what we call general economic theory,  although  a considerable portion  of this theory has a de…ninitely quantitative character. Nor should econometrics  be taken  as synonomous with  the  application of mathematics to  economics.   Experience  has  shown  that each of these  three  view-points,  that of statistics, economic  theory and  mathematics, is a necessary, but not by itself a su¢cient, condition for a real understanding of the quantitative relations in modern economic life.  It is the uni…cation of all three that is powerful.  And it is this uni…cation that constitutes econometrics.

Ragnar  Frisch,  Econometrica, (1933), 1, pp.  1-2.


This de…nition remains valid today, although some terms have evolved somewhat in their usage. Today,  we would  say  that econometrics  is the  uni…ed  study  of economic  models,  mathematical statistics, and economic data.
Within  the …eld of econometrics there are sub-divisions and specializations. Econometric theory concerns  the  development  of tools and  methods,  and  the  study  of the  properties  of econometric methods.    Applied  econometrics  is a term  describing  the  development  of quantitative  economic models and the application of econometric  methods  to these models using economic data.


1. The Probability Approach to Econometrics

The unifying methodology  of modern econometrics  was articulated by Trygve Haavelmo (1911-
1999) of Norway,  winner  of the  1989 Nobel Memorial  Prize  in Economic  Sciences, in his seminal

1



paper “The probability approach in econometrics”, Econometrica (1944).  Haavelmo argued that quantitative economic models  must  necessarily  be probability  models  (by  which  today  we would mean  stochastic).  Deterministic models are blatently inconsistent  with  observed  economic quan- tities,  and  it  is incoherent  to  apply  deterministic models  to  non-deterministic data.    Economic models should  be explicitly  designed  to  incorporate randomness; stochastic errors  should  not  be simply added  to deterministic models to make them  random.   Once we acknowledge that an eco- nomic model is a probability model, it follows naturally that an appropriate tool way to quantify, estimate, and  conduct  inferences  about  the  economy  is through  the  powerful  theory  of mathe- matical statistics.  The appropriate method for a quantitative economic analysis follows from the probabilistic construction of the economic model.
Haavelmos probability approach was quickly embraced by the economics profession. Today no quantitative work in economics shuns its fundamental vision.
While all economists embrace the probability approach, there has been some evolution in its implementation.
The structural approach is the closest to Haavelmo’s original idea.  A probabilistic economic model is speci…ed, and the quantitative analysis performed under the assumption that the economic model is correctly speci…ed. Researchers often describe this as “taking their model seriously.” The structural approach typically leads to likelihood-based analysis, including maximum likelihood and Bayesian  estimation.
A  criticism  of the  structural  approach is that it  is misleading  to  treat an  economic  model as correctly speci…ed.   Rather, it is more accurate to view a model as a useful abstraction or approximation. In this case, how should we interpret structural econometric analysis? The quasi- structural approach to inference views a structural economic model as an approximation rather than  the truth. This theory  has led to the concepts  of the pseudo-true value (the  parameter value de…ned by the estimation problem), the quasi-likelihood function, quasi-MLE, and quasi-likelihood inference.
Closely related is the semiparametric approach. A probabilistic economic model is partially speci…ed but some features are left unspeci…ed.  This approach typically leads to estimation methods such as least-squares and the Generalized Method of Moments.   The semiparametric approach dominates  contemporary econometrics,  and is the main focus of this textbook.
Another  branch  of quantitative structural economics is the  calibration approach.  Similar to the quasi-structural approach, the calibration approach interprets structural models as approx- imations and hence inherently false.   The di¤erence is that the calibrationist literature rejects mathematical statistics as inappropriate for approximate models,  and  instead  selects parameters by matching  model and data  moments  using non-statistical ad hoc1  methods.


1. Econometric Terms and Notation

In a typical application, an econometrician has a set of repeated measurements on a set of vari- ables. For example, in a labor application the variables could include weekly earnings, educational attainment, age, and other descriptive  characteristics. We call this information  the data, dataset, or sample.
We use the term observations to refer to the distinct repeated  measurements on the variables. An individual observation often corresponds to a speci…c economic unit, such as a person, household, corporation, …rm, organization, country, state, city or other geographical region.  An individual observation could  also be a measurement  at  a point  in time,  such  as quarterly GDP  or a daily interest rate.
Economists typically denote variables by the italicized roman characters y, x; and/or z: The convention  in econometrics  is to use the  character y to denote  the  variable  to be explained,  while


1 Ad hoc mean“for this purpose  a method designed for a specic problem   and  not based  on a generalizable principle.



the characters x and z are used to denote  the conditioning  (explaining)  variables.
Following mathematical convention,  real numbers  (elements  of the real line R) are written using lower case italics such as y, and vectors (elements  of Rk ) by lower case bold italics such as x; e.g.
  0 x 1

x = B x 




B     C      

   xk    


Upper case bold italics such as X are used for matrices.
We typically  denote  the  number  of observations by the  natural number  n; and  subscript  the variables  by the index i to denote  the individual  observation, e.g.  yi; xi  and zi. In some contexts we use indices other  than  i, such as in time-series  applications where the  index t is common,  and in panel studies  we typically use the double index it to refer to individual  i at a time period t.



The i’th observation is the set (yi; xi; zi):




It is proper  mathematical practice  to use upper  case X  for random  variables  and lower case x for realizations  or speci…c values.  This practice  is not commonly followed in econometrics  because instead  we use upper  case to denote  matrices.   Thus  the  notation yi will in some places refer to a random  variable,  and in other  places a speci…c realization. Hopefully there  will be no confusion as the use should be evident from the context.
We typically use Greek letters  such as   ;   and    2  to denote  unknown  parameters of an econo-
metric  model,  and  will use boldface,  e.g.      or    , when  these  are  vector-valued.  Estimates are
typically denoted  by putting a hat  “^”, tilde  “~”  or bar  “-” over the  corresponding  letter,  e.g.   ^
and ~ are estimates  of  :
The  covariance  matrix  of an econometric  estimator will typically be written using the  capital

boldface V ; often with  a subscript  to denote  the  estimator, e.g.  V  b


= var   pn  b           as the

covariance  matrix  for pn  b         : Hopefully without causing confusion, we will use the notation

V   = avar( b ) to  denote  the  asymptotic covariance  matrix  of pn  b         (the  variance  of the



asymptotic distribution).  Estimates will be denoted  by appending  hats  or tildes,  e.g. estimate V   is an of V   .







Read More Add your Comment 0 comments


 

About Me

Fan

© 2010 statistics for all All Rights Reserved Converted into Blogger Template by Hack Tutors.info