Jump to content

The last person to reply to this thread...


Seffiks

Who Will be The Winner?????  

1,235 members have voted

You do not have permission to vote in this poll, or see the poll results. Please sign in or register to vote in this poll.

Recommended Posts

Standardised outcome measure

 

 

Individual results have to be expressed in a standardised format to allow for comparison between studies. If the end point is continuous—for example, blood pressure—the mean difference between the treatment and control groups is used. The size of a difference, however, is influenced by the underlying population value. An antihypertensive drug, for example, is likely to have a greater absolute effect on blood pressure in overtly hypertensive patients than in borderline hypertensive patients. Differences are therefore often presented in units of standard deviation. If the end point is binary—for example, disease versus no disease, or dead versus alive) then odds ratios or relative risks are often calculated (box). The odds ratio has convenient mathematical properties, which allow for ease in combining data and testing the overall effect for significance. Absolute measures, such as the absolute risk reduction or the number of patients needed to be treated to prevent one event,17 are more helpful when applying results in clinical practice (see below).

 

 

Odds ratio or relative risk?

 

Odds and odds ratio

 

The odds is the number of patients who fulfil the criteria for a given endpoint divided by the number of patients who do not. For example, the odds of diarrhoea during treatment with an antibiotic in a group of 10 patients may be 4 to 6 (4 with diarrhoea divided by 6 without, 0.66); in a control group the odds may be 1 to 9 (0.11) (a bookmaker would refer to this as 9 to 1). The odds ratio of treatment to control group would be 6 (0.66÷0.11).

 

Risk and relative risk

 

The risk is the number of patients who fulfil the criteria for a given end point divided by the total number of patients. In the example above the risks would be 4 in 10 in the treatment group and 1 in 10 in the control group, giving a risk ratio, or relative risk, of 4 (0.4÷0.1).

 

Statistical methods for calculating overall effect

 

The last step consists in calculating the overall effect by combining the data. A simple arithmetic average of the results from all the trials would give misleading results. The results from small studies are more subject to the play of chance and should therefore be given less weight. Methods used for meta-analysis use a weighted average of the results, in which the larger trials have more influence than the smaller ones. The statistical techniques to do this can be broadly classified into two models,18 the difference consisting in the way the variability of the results between the studies is treated. The "fixed effects" model considers, often unreasonably, that this variability is exclusively due to random variation.19 Therefore, if all the studies were infinitely large they would give identical results. The "random effects" model20 assumes a different underlying effect for each study and takes this into consideration as an additional source of variation, which leads to somewhat wider confidence intervals than the fixed effects model. Effects are assumed to be randomly distributed, and the central point of this distribution is the focus of the combined effect estimate. Although neither of two models can be said to be "correct," a substantial difference in the combined effect calculated by the fixed and random effects models will be seen only if studies are markedly heterogeneous.

 

Bayesian meta-analysis

 

Some statisticians feel that other statistical approaches are more appropriate than either of the above. One approach uses Bayes's theorem, named after an 18th century English clergyman.21 Bayesian statisticians express their belief about the size of an effect by specifying some prior probability distribution before seeing the data, and then they update that belief by deriving a posterior probability distribution, taking the data into account.22 Bayesian models are available under both the fixed and random effects assumption.23 The confidence interval (or more correctly in bayesian terminology, the 95% credible interval, which covers 95% of the posterior probability distribution) will often be wider than that derived from using the conventional models because another component of variability, the prior distribution, is introduced. Bayesian approaches are controversial because the definition of prior probability will often be based on subjective assessments and opinion.

 

Heterogeneity between study results

 

If the results of the studies differ greatly then it may not be appropriate to combine the results. How to ascertain whether it is appropriate, however, is unclear. One approach is to examine statistically the degree of similarity in the studies' outcomes—in other words, to test for heterogeneity across studies. In such procedures, whether the results of a study reflect a single underlying effect, rather than a distribution of effects, is assessed. If this test shows homogeneous results then the differences between studies are assumed to be a consequence of sampling variation, and a fixed effects model is appropriate. If, however, the test shows that significant heterogeneity exists between study results then a random effects model is advocated. A major limitation with this approach is that the statistical tests lack power—they often fail to reject the null hypothesis of homogeneous results even if substantial differences between studies exist. Although there is no statistical solution to this issue, heterogeneity between study results should not be seen as purely a problem for meta-analysis—it also provides an opportunity for examining why treatment effects differ in different circumstances. Heterogeneity should not simply be ignored after a statistical test is applied; rather, it should be scrutinised, with an attempt to explain it.

 

Graphic display

 

Results from each trial are usefully graphically displayed, together with their confidence intervals. Figure 3 represents a meta-analysis of 17 trials of ß blockers in secondary prevention after myocardial infarction. Each study is represented by a black square and a horizontal line, which correspond to the point estimate and the 95% confidence intervals of the odds ratio. The 95% confidence intervals would contain the true underlying effect in 95% of the occasions if the study was repeated again and again. The solid vertical line corresponds to no effect of treatment (odds ratio 1.0). If the confidence interval includes 1, then the difference in the effect of experimental and control treatment is not significant at conventional levels (P>0.05). The area of the black squares reflects the weight of the study in the meta-analysis. The confidence interval of all but two studies cross this line, indicating that the effect estimates were non-significant (P>0.05).

 

The diamond represents the combined odds ratio, calculated using a fixed effects model, with its 95% confidence interval. The combined odds ratio shows that oral ß blockade starting a few days to a few weeks after the acute phase reduces subsequent mortality by an estimated 22% (odds ratio 0.78; 95% confidence interval 0.71 to 0.87). A dashed line is plotted vertically through the combined odds ratio. This line crosses the horizontal lines of all individual studies except one (N). This indicates a fairly homogenous set of studies. Indeed, the test for heterogeneity gives a non-significant P value of 0.2.

 

A logarithmic scale was used for plotting the odds ratios in figure 3. There are several reasons that ratio measures are best plotted on logarithmic scales.25 Most importantly, the value of an odds ratio and its reciprocal—for example, 0.5 and 2—which represent odds ratios of the same magnitude but opposite directions, will be equidistant from 1.0. Studies with odds ratios below and above 1.0 will take up equal space on the graph and thus look equally important. Also, confidence intervals will be symmetrical around the point estimate.

 

Relative and absolute measures of effect

 

Repeating the analysis by using relative risk instead of the odds ratio gives an overall relative risk of 0.80 (95% confidence interval 0.73 to 0.88). The odds ratio is thus close to the relative risk, as expected when the outcome is relatively uncommon (see box). The relative risk reduction, obtained by subtracting the relative risk from 1 and expressing the result as a percentage, is 20% (12% to 27%). The relative measures used in this analysis ignore the absolute underlying risk. The risk of death among patients who have survived the acute phase of myocardial infarction, however, varies widely.26 For example, among patients with three or more cardiac risk factors the probability of death at two years after discharge ranged from 24% to 60%.26 Conversely, two year mortality among patients with no risk factors was less than 3%. The absolute risk reduction or risk difference reflects both the underlying risk without treatment and the risk reduction associated with treatment. Taking the reciprocal of the risk difference gives the "number needed to treat" (the number of patients needed to be treated to prevent one event).17

 

For a baseline risk of 1% a year, the absolute risk difference shows that two deaths are prevented per 1000 patients treated (1). This corresponds to 500 patients (1÷0.002) treated for one year to prevent one death. Conversely, if the risk is above 10%, less than 50 patients have to be treated to prevent one death. Many clinicians would probably decide not to treat patients at very low risk, given the large number of patients that have to be exposed to the adverse effects of ß blockade to prevent one death. Appraising the number needed to treat from a patient's estimated risk without treatment and the relative risk reduction with treatment is a helpful aid when making a decision in an individual patient. A nomogram that facilitates calculation of the number needed to treat at the bedside has recently been published.27

 

 

 

View this table:

[in this window]

[in a new window]

ß Blockade in secondary prevention after myocardial infarction—absolute risk reductions and numbers needed to treat for one year to prevent one death for different levels of mortality in control group

 

Meta-analysis using absolute effect measures such as the risk difference may be useful to illustrate the range of absolute effects across studies. The combined risk difference (and the number needed to treat calculated from it) will, however, be essentially determined by the number and size of trials in patients at low, intermediate, or high risk. Combined results will thus be applicable only to patients at levels of risk corresponding to the average risk of the trials included. It is therefore generally more meaningful to use relative effect measures for summarising the evidence and absolute measures for applying it to a concrete clinical or public health situation.

Link to comment
Share on other sites

This forum is supported by the 12ozProphet Shop, so go buy a shirt and help support!
This forum is brought to you by the 12ozProphet Shop.
This forum is brought to you by the 12oz Shop.
  • Replies 8.3k
  • Created
  • Last Reply

Top Posters In This Topic

  • -Tib3-

    1297

  • !RENOBOF

    1255

  • CALIgula

    1180

  • watson

    809

Top Posters In This Topic

Posted Images

Sensitivity analysis

 

Opinions will often diverge on the correct method for performing a particular meta-analysis. The robustness of the findings to different assumptions should therefore always be examined in a thorough sensitivity analysis. This is illustrated in figure 4 for the meta-analysis of ß blockade after myocardial infarction. Firstly, the overall effect was calculated by different statistical methods, by using both a fixed and a random effects model. The 4 shows that the overall estimates are virtually identical and that confidence intervals are only slightly wider with the random effects model. This is explained by the relatively small amount of variation between trials in this meta-analysis.

 

Secondly, methodological quality was assessed in terms of how patients were allocated to active treatment or control groups, how outcome was assessed, and how the data were analysed.6 The maximum credit of nine points was given if patient allocation was truly random, if assessment of vital status was independent of treatment group, and if data from all patients initially included were analysed according to the intention to treat principle. Figure 4 shows that the three low quality studies (<=7 points) showed more benefit than the high quality trials. Exclusion of these three studies, however, leaves the overall effect and the confidence intervals practically unchanged.

 

Thirdly, significant results are more likely to get published than non-significant findings,28 and this can distort the findings of meta-analyses. The presence of such publication bias can be identified by stratifying the analysis by study size—smaller effects can be significant in larger studies. If publication bias is present, it is expected that, of published studies, the largest ones will report the smallest effects. Figure 4 shows that this is indeed the case, with the smallest trials (50 or fewer deaths) showing the largest effect. However, exclusion of the smallest studies has little effect on the overall estimate.

 

Finally, two studies (J and N; see 3) were stopped earlier than anticipated on the grounds of the results from interim analyses. Estimates of treatment effects from trials that were stopped early are liable to be biased away from the null value. Bias may thus be introduced in a meta-analysis that includes such trials.29 Exclusion of these trials, however, affects the overall estimate only marginally.

 

The sensitivity analysis thus shows that the results from this meta-analysis are robust to the choice of the statistical method and to the exclusion of trials of poorer quality or of studies stopped early. It also suggests that publication bias is unlikely to have distorted its findings.

 

 

Conclusions

 

Meta-analysis should be seen as structuring the processes through which a thorough review of previous research is carried out. The issues of completeness and combinability of evidence, which need to be considered in any review,30 are made explicit. Was it sensible to have combined the individual trials that comprise the meta-analysis? How robust is the result to changes in assumptions? Does the conclusion reached make clinical and pathophysiological sense? Finally, has the analysis contributed to the process of making rational decisions about the management of patients? It is these issues that we explore further in later articles in this series.

 

 

 

 

CONGRATS

you just learnt about meta-analysis

Link to comment
Share on other sites

The linear regression model

In the linear regression model, the dependent variable is assumed to be a linear function of one or more independent variables plus an error introduced to account for all other factors:

 

In the above regression equation, y_i is the dependent variable, x_i1, ...., x_iK are the independent or explanatory variables, and u_i is the disturbance or error term. The goal of regression analysis is to obtain estimates of the unknown parameters Beta_1, ..., Beta_K which indicate how a change in one of the independent variables affects the values taken by the dependent variable.

 

Applications of regression analysis exist in almost every field. In economics, the dependent variable might be a family's consumption expenditure and the independent variables might be the family's income, number of children in the family, and other factors that would affect the family's consumption patterns. In political science, the dependent variable might be a state's level of welfare spending and the independent variables measures of public opinion and institutional variables that would cause the state to have higher or lower levels of welfare spending. In sociology, the dependent variable might be a measure of the social status of various occupations and the independent variables characteristics of the occupations (pay, qualifications, etc.). In psychology, the dependent variable might be individual's racial tolerance as measured on a standard scale and with indicators of social background as independent variables. In education, the dependent variable might be a student's score on an achievment test and the independent variables characteristics of the student's family, teachers, or school.

 

The common aspect of the applications described above is that the dependent variable is a quantitative measure of some condition or behavior. When the dependent variable is qualitative or categorical, then other methods (such as logit or probit analysis, described in Chapter 7) might be more appropriate.

 

Ordinary least squares estimation

The usual method of estimation for the regression model is ordinary least squares (OLS). Let b_1, ..., b_K denote the OLS estimates of Beta_1, ..., Beta_K. The predicted value of y_i is:

 

The error in the OLS prediction of y_i, called the residual, is:

 

The basic idea of ordinary least squares estimation is to choose estimates Beta_1, ..., Beta_K to minimize the sum of squared residuals:

 

It can be shown that:

 

where X is an n * k matrix with (i,k)th element x_ki, y is an n * k vector with typical element y_i, and b is a k * 1 vector with typical element b_k.

 

Assumptions for regression analysis

The least squares fitting procedure described below can be used for data analysis as a purely descriptive technique. However, the procedure has strong theoretical justification if a few assumptions are made about how the data are generated. The starting point is the regression equation presented above which describes some causal or behavioral process. The independent variables play the role of experimental or treatment variables, though in few social science applications will the investigator actually have control over the values of the independent variables. The error term captures the effects of all omitted variables. In an experiment, randomization of the treatments (independent variables) ensures that the omitted factors (the disturbances) are uncorrelated with the treatments. This greatly simplifies inference. Non-experimental researchers, however, must substitute assumptions for experimental controls. The validity of non-experimental results therefore depends critically upon the accuracy of the assumptions. We present one set of assumptions, known as the Gauss-Markov assumptions, that are sufficient to guarantee that ordinary regression estimates will have good properties.

 

First, we assume that the errors u_i have an expected value of zero: E(u_i ) = 0 This means that on average the errors balance out.

 

Second, we assume that the independent variables are non-random. In an experiment, the values of the independent variable would be fixed by the experimenter and repeated samples could be drawn with the independent variables fixed at the same values in each sample. As a consequence of this assumption, the indenpendent variables will in fact be independent of the disturbance. For non-experimental work, this will need to be assumed directly along with the assumption that the independent variables have finite variances.

 

Third, we assume that the independent variables are linearly independent. That is, no independent variable can be expressed as a (non-zero) linear combination of the remaining independent variables. The failure of this assumption, known as multicollinearity, clearly makes it infeasible to disentangle the effects of the supposedly independent variables. If the independent variables are linearly dependent, SST will produce an error message (singularity in independent variables) and abort the REG command.

 

Fourth, we assume that the disturbances u_i are homoscedastic:

 

This means that the variance of the disturbance is the same for each observation.

 

Fifth, we assume that the disturbances are not autocorrelated:

 

This means disturbances associated with different observations are uncorrelated.

 

Properties of the OLS estimator

If the first three assumptions above are satisfied, then the ordinary least squares estimator b will be unbiased: E(b) = beta Unbiasedness means that if we draw many different samples, the average value of the OLS estimator based on each sample will be the true parameter value beta. Usually, however, we have only one sample, so the variance of the sampling distribution of b is an important indicator of the quality of estimates obtained.

 

If all five of the assumptions above hold, then it can be shown that the variance of the OLS estimator is given by:

If the independent variables are highly intercorrelated, then the matrix X' X will be nearly singular and the element of (X' X)^-1 will be large, indicating that the estimates of beta may be imprecise.

 

To estimate Var(b) we require an estimator of sigma^2. It can be shown that:

 

is an unbiased estimator of sigma^2. The square root of (sigma hat)^2 is called the standard error of the regression. It is just the standard deviation of the residuals e_i.

 

There are two important theorems about the properties of the OLS estimators. The Gauss-Markov theorem states that under the five assumptions above, the OLS estimator b is best linear unbiased. That is, the OLS estimator has smaller variance than any other linear unbiased estimator. (One covariance matrix is said to be larger than another if their difference is positive semi-definite.) If we add the assumption that the disturbances u_i have a joint normal distribution, then the OLS estimator has minimum variance among all unbiased estimators (not just linear unbiased estimators).

 

Although the preceding theorems provide strong justification for using the OLS estimator, it should be realized that OLS is rather sensitive to departures from the assumptions. A few outliers (stray observations generated by a different process) can strongly influence the OLS estimates. SST provides useful diagnostic tools for detecting data problems that we discuss below.

 

Use of the REG command

To estimate a regression in SST, you need to specify one or more dependent variables (in the DEP subop) and one or more independent variables (in the IND subop). Unlike some other programs, SST does not automatically add a constant to your independent variables. If you want one, you should create a constant and add it to the list of your independent variables. For example, to regress the variable y on x with an intercept:

 

set one=1

reg dep[y] ind[one x]

 

SST will produce two coefficients: an intercept and a slope parameter. The corresponding regression line passes through the point (0,b_0) and has slope equal to b_1:

 

where b_0 is the coefficient of one and b_1 is the coefficient of the variable x. If, on the other hand, you had omitted the variable one from the IND subop:

 

reg dep[y] ind[x]

 

SST would produce a "regression through the origin". That is, the regression line would pass through the point (0,0) with slope equal to the coefficient of x (b):

 

For most purposes you will want to include a constant, but SST allows you the flexibility to decide otherwise.

 

The IF and OBS subops can be used to restrict the range of observations used in the regression. Only the subset of observations activated by the current RANGE statement that meet the criteria set in the IF and OBS subops will be used. If any of the variables specified in the IND or DEP subops have missing data for an observation, the entire observation is deleted from the estimation range for that regression. For example, to run a regression of y on x (and a constant) including only observations one through ten:

 

reg dep[y] ind[one x] obs[1-10]

 

Using the OBS subop does not affect the observation range for subsequent commands.

 

SST also allows you to specify multiple dependent variables in the DEP subop. If one variable is specified in the DEP subop, it will be regressed on the variables specified in the IND subop. If more than one variable is specified in the DEP subop, separate regressions will be run for each of these variables on the variables listed in the IND subop. The same observation range will be used for all regressions.

 

An example

We illustrate use of the REG command using an example taken from David A. Belsey, Edwin Kuh, and Roy E. Welsch, Regression Diagnostics (Wiley, 1980). The variables in the data set are macroeconomic and demographic indicators for fifty countries for the decade of the 1960's and are used to test a simply life cycle savings model.

 

A text file containing the data is supplied with your SST program disk (bkw.dat). The following variables are in the file:

 

SR Personal savings rate

POP15 Percentage of population under age 15

POP75 Percentage of population over age 75

PDI Personal disposable income per capita (constant dollars)

DELPDI Percentage growth rate of PDI from 1960 to 1970

 

According to the life cycle savings model, savings rates will be highest among middle-aged individuals. Younger individuals, anticipating higher incomes as they become older, will have low savings rates. On the other hand, older individuals will tend to consume whatever savings they accumulated during middle age. The following regression equation is proposed to test this theory:

 

To replicate the Belsey, Kuh, and Welsch analysis, first we READ the data file bkw.dat as described in Chapter 2. Then we LABEL it, SET a variable one equal to a vector of ones, and SAVE the entire data set. The commands are the following:

 

range obs[1-100]

read to[sr pop15 pop75 dpi deldpi] file[bkw.dat]

label var[sr] lab[average personal savings rate]

label var[pop15] lab[percentage population under 15]

label var[pop75] lab[percentage population over 75]

label var[dpi] lab[real disposable income per capita]

label var[deldpi] lab[real disposable income growth rate]

save file[bkw]

set one=1

 

Now we are ready to regress sr on one, pop15, pop75, dpi, and deldpi:

 

reg dep[sr] ind[one pop15 pop75 dpi deldpi]

 

The output produced is:

 

********** ORDINARY LEAST SQUARES **********

Dependent Variable: sr

 

Independent Estimated Standard t-

Variable Coefficient Error Statistic

 

one 28.5662941 7.3544917 3.8841969

pop15 -0.4611972 0.1446415 -3.1885533

pop75 -1.6914257 1.0835935 -1.5609412

dpi -0.0003371 0.0009311 -0.3620211

deldpi 0.4096868 0.1961958 2.0881534

 

Number of Observations 50

R-squared 0.338458

Corrected R-squared 0.279655

Sum of Squared Residuals 6.51e+002

Standard Error of the Regression 3.8026627

Mean of Dependent Variable 9.6710000

 

The OLS estimates provide some support for the life cycle model. The t-statistic for the coefficient of pop15 is -3.18, which would enable us to reject the null hypothesis beta_2 = 0 at conventional significance levels. The coefficient of pop75, however, does not achieve significance at 0.05 level. Notice, also, that the income effect is small (a thousand dollars of income is associated with only a 0.3 percent rise in the savings rate) and statistically insignificant, while the income change variable has a positive and statistically significant impact on the savings rate.

 

Regression diagnostics

The REG procedure also allows you to produce diagnostic statistics to evaluate the regression estimates including:

 

* predicted values (PRED)

* residuals (RSD)

* studentized residuals (SRSD)

* diagonal elements of the "hat" matrix (HAT)

* estimated coefficients (COEF)

* covariance matrix (COVMAT)

 

Most of these will be familiar, but we discuss in some detail some of the less well known diagnostics: studentized residuals and the hat matrix. These two diagnostics are discussed in detail in Regression Diagnostics.

 

Studentized residuals and the hat matrix

Studentized residuals are helpful in identify outliers which do not appear to be consistent with the rest of the data. The hat matrix is used to identify "high leverage" points which are outliers among the independent variables. The two concepts are related. In the case of studentized residuals, large deviations from the regression line are identified. Since the residuals from a regression will generally not be independently or identically distributed (even if the disturbances in the regression model are), it is advisable to weight the residuals by their standard deviations (this is what is meant by studentization). A similar idea motivates the calculation of the hat matrix (see Regression Diagnostics, p. 17).

 

The hat matrix H is given by: H = X(X' X)^-1 X' Note that since: b = (X' X)^-1 X' y and by definition: y hat = Xb it follows that: y hat = Hy Since the hat matrix is of dimension n * n, the number of elements in it can become quite large. Usually it suffices to work with only the diagonal elements h_1, ..., h_n:

where x_i is the ith row of the matrix X. Note that:

 

so that:

 

since I-H is idempotent. It follows therefore:

 

so that the diagonal elements of the hat matrix are closely related to the variances of the residuals. To compute the studentized residuals, we divide e_i by an estimate of its variance. Rather than using

 

, we recompute the regression deleting the ith observation. Denote the corresponding estimate of sigma^2 with the ith observation deleted by s^2 (i) and the corresponding diagonal element of the hat matrix from the regression with the ith observation deleted by h_i tilde. The formula for the studentized residual for the ith observation is:

 

Use of the hat matrix diagonal elements

Since y hat = Hy, the diagonal elements of H, the h_i, indicate the effect of a given observation. There are a few useful facts about the diagonal elements of the hat matrix:

 

where K is the number of independent variables, including the constant if there is one. Belsley, Kuh, and Welch suggest 2p/n as a rough cutoff for determing high leverage points, terming the ith observation a leverage point when h_i exceeds 2p/n.

 

Use of studentized residuals

Belsley, Kuh, and Welch point out that the studentized residuals have an approximate t-distribution with n-p-1 degrees of freedom. This means we can assess the significance of any single studentized residual using a t-table (or, equivalently, a table of the standard normal distribution if n is moderately large). For example, we might rerun the above regression and save the studentized residuals by specifying a variable name in the SRSD subop:

 

reg dep[sr] ind[one pop15 pop75 dpi deldpi] srsd[studrsd]

 

Next, we might investigate which observations have large studentized residuals:

 

print var[studrsd] if[abs(studrsd) > 1.96]

 

OBS VARIABLES

studrsd

7: -2.3134

46: 2.8536

 

Next, we could rerun the regression omitting those observations with large studentized residuals:

 

reg dep[sr] ind[one pop15 pop75 dpi deldpi] if[abs(studrsd) <= 1.96]

 

********** ORDINARY LEAST SQUARES **********

Dependent Variable: sr

 

Independent Estimated Standard t-

Variable Coefficient Error Statistic

 

one 28.8187711 6.5324171 4.4116551

pop15 -0.4683572 0.1280318 -3.6581323

pop75 -1.5778925 0.9686178 -1.6290146

dpi -0.0003989 0.0008229 -0.4846829

deldpi 0.3480148 0.1740605 1.9993897

 

Number of Observations 48

R-squared 0.410031

Corrected R-squared 0.355150

Sum of Squared Residuals 4.85e+002

Standard Error of the Regression 3.3589505

Mean of Dependent Variable 9.6747917

 

In this case the coefficient estimates seem relatively stable with the outliers removed.

 

Instrumental variables estimation

SST also allows easy estimation of simulataneous equations models using two stage least squares. Models where some of the independent variables are correlated with the disturbances will be inconsistently estimated by OLS. If, however, a set of instrumental variables (which are correlated with the independent variables, but uncorrelated with the disturbances) is available, it is possible to "purge" the independent variables of their correlation with the disturbance. The instrumental variables are specified in the IV subop. Variables in the IV subop can overlap with those in the IND subop if their are included exogenous variables in the equation. The number of instrumental variables (including included exogenous variables) must be at least as large as the number of independent variables (or else the order condition for identification will not be met).

 

A simple example of simultaneous equations estimation occurs in estimating a market supply or demand equation. Consider, for example, the supply function for an agricultural commodity. Let price be the market price of the commodity in each period and quantity the quantity supplied of the commodity. In equilibrium, quantity demanded and quantity supplied are equal. Moreover, price and quantity are simultaneously determined in the market, so it does not make sense to regress quantity on price to obtain an estimate of the price elasticity of supply. Suppose, we believe that the supply function also depends on the weather (measured, say, by rainfall) while the demand function also depends on population (populat) and aggregate personal disposable income (pdi), but that these variables are exogenous to the market for this particular commodity. To estimate the supply equation by two stage least squares, give the command:

 

reg dep[quantity] ind[one price weather] iv[one weather populat pdi]

 

Remember to include the constant one in the IV subop, since it is certainly exogenous. For details of simultaneous equations estimation, consult any econometrics text, e.g. H. Theil, Principles of Econometrics (Wiley, 1971), chapters 9-10.

Link to comment
Share on other sites

Guest
This topic is now closed to further replies.

×
×
  • Create New...