4.2. Poisson Models for Count Data with Two Observations Per Individual

When there are only two observations per individual, we saw previously that a linear or logistic fixed effects analysis could be done using simplified methods with conventional software. This is also true for count data. In fact, a fixed effects Poisson regression model can be estimated with an ordinary logistic regression program.

For the patent data, let's ignore the intervening years and focus only on 1975 and 1979. Let yi1 be the patent count for firm i in 1975 and yi2 the patent count in 1979. We assume that each of these variables has a Poisson distribution with parameter λit. That is, the probability that yit = r is given by


Why a Poisson distribution? Well, the Poisson distribution is perhaps the simplest probability distribution that is appropriate for count data. It can be derived from a stochastic process in which it is assumed that events (in this case patents) cannot occur simultaneously, and that events are independent (Cameron and Trivedi 1998). That is, the occurrence of an event neither raises nor lowers the probability of future events. Note that we are not assuming that there is a single Poisson distribution for the entire sample. Instead, each firm's patent count is drawn from a different Poisson distribution whose parameter λit varies across both firms and time.

An important property of the Poisson distribution is that its mean and variance are equal, and both are equal to the Poisson parameter:


Next, we let λit be a log-linear function of the predictor variables:


As in earlier chapters, xit represents the time-varying predictor variables, zi denotes the time-invariant predictors, and αi denotes the unobserved fixed effects. The vector xit includes the R & D expenditures in the current year t and in each of the preceding five years.

Our goal is to estimate the parameters in equation (4.3). To do this, we shall use conditional maximum likelihood, the same method used in chapter 3 to estimate the fixed effects logistic regression model. Consider the distribution of yi2 conditional on the total event count for the two time periods combined, denoted by wi = yi1 + yi2. It can be shown that yi2|wi ~ B (pi, wi).

That is, conditional on the total count wi, the 1979 count yi2 has a binomial distribution with parameters pi and wi, where


It follows, after a bit of algebra, that


Thus, we have converted our Poisson regression model into a logistic regression model in which the predictor variables are difference scores for the original predictors. Note that, as in earlier applications, both αi and γzi drop out of equation (4.5).

To implement this conditional approach in SAS, we may use any SAS procedure that does logistic regression for grouped data, including LOGISTIC, GENMOD, CATMOD and PROBIT. Here's how to do it in GENMOD. First, we create a new data set that contains the total count for each firm and the difference scores for the research and development variables:

DATA patents;
   SET my.patents;
   total=pat75+pat79;
   rd_0=logr79-logr75;
   rd_1=logr78-logr74;
   rd_2=logr77-logr73;
   rd_3=logr76-logr72;
   rd_4=logr75-logr71;
   rd_5=logr74-logr70;
RUN;

RD_0 is the difference score for the same years in which the patents were counted, whereas RD_1 through RD_5 are difference scores for lags of one to five years.

Let's first estimate a model with no covariates:

PROC GENMOD DATA=patents;
   MODEL pat79/total = / DIST=B;
RUN;

Note that the dependent variable is expressed with the events/trials syntax, which tells SAS that PAT79 events occurred out of a possible TOTAL. As in chapter 3, DIST=B specifies that PAT79 has a binomial distribution whose default link function is logit (i.e., logistic). Results are shown in Output 4.1.

Table 4.1. Output 4.1 Conditional Poisson Regression Model for Patents, with No Covariates
Model Information
Data SetWORK.PATENTS
DistributionBinomial
Link FunctionLogit
Response Variable (Events)pat79
Response Variable (Trials)total
Number of Observations Read346
Number of Observations Used300
Number of Events11107
Number of Trials23865
Number of Invalid Responses46
Criteria For Assessing Goodness Of Fit
CriterionDFValueValue/DF
Deviance2991001.36563.3490
Scaled Deviance2991001.36563.3490
Pearson Chi-Square299938.24583.1379
Scaled Pearson X2299938.24583.1379
Log Likelihood −16484.8031 
Analysis Of Parameter Estimates
ParameterDFEstimateStandard ErrorWald 95% Confidence Limits Chi-SquarePr>ChiSq
Intercept1−0.13860.0130−0.1640−0.1131114.03<.0001
Scale01.00000.00001.00001.0000  

Under "Model Information" we see that 46 firms had invalid response values. These are firms that had 0 patents in both 1975 and 1979, so their total for the two years was also 0. Of course, the binomial distribution is undefined when the number of trials is 0, which is why these firms are excluded. This points out a more general characteristic of Poisson regression that extends to the next section as well. Whenever you condition on the total count, those cases that have a total count of 0 are effectively removed from the likelihood function. If the total is 0, then each component must also be 0, leaving no within-individual variability to analyze.

In the next panel, "Criteria for Goodness of Fit," we see that both the deviance and the Pearson chi-square statistics are more than three times their degrees of freedom. For a model to have a good fit, these statistics should be close to their degrees of freedom. However, because many of the expected counts generated by this model are small (near 0), the chi-square distribution may not be a good approximation. For that reason, it's probably not a good idea to compute a p-value. Nevertheless, the magnitude of these ratios suggests that there is a problem with overdispersion, about which I'll have more to say as the chapter progresses.

Finally, we get to the "Analysis of Parameter Estimates." The only estimate is the intercept, with a value of −;.1386. What does this tell us? Well, if m1 is the mean number of patents in year 1 and m2 is the mean for year 2, the intercept is log(m1/m2). If the mean number of patents were exactly the same in both years, the intercept would be 0. The fact that it's negative indicates that the mean went down over time. More specifically, if we calculate 100(exp(−.1386) − 1) = − 12.9%, we get the percentage decrease in the mean from 1975 to 1979. Furthermore, because the chi-square for the intercept is so large, we can reject the null hypothesis that the means for the two years are the same. In effect, what we have here is the count data analog of the paired-comparisons t-test discussed in chapter 1, or the McNemar test for dichotomous variables discussed in chapter 3.

Now let's add the lagged variables for research and development expenditures as covariates, with results shown in Output 4.2. Again we see that the ratio of the goodness-of-fit chi-square statistics to their degrees of freedom is above 3, suggesting that we really need to do something about overdispersion. But let's postpone that issue for a moment. Examining the parameter estimates and their associated statistics, we see that RD_0, the contemporaneous measure of research and development expenditures, has a highly significant effect on the patent count, with a coefficient of .5214. To interpret this, keep in mind that both the dependent variable (expected number of patents) and the independent variable (research and development expenditures) are logged (see equation (4.3)). In that case, we can say that a 1% increase in R & D expenditures is associated with a .52% increase in the expected number of patents in the same year, controlling for the lagged R & D measures. The effects of the lagged measures are much smaller.

Table 4.2. Output 4.2 Conditional Poisson Regression Model for Patents, with Covariates
Criteria For Assessing Goodness Of Fit
CriterionDFValueValue/DF
Deviance293949.30313.2399
Scaled Deviance293949.30313.2399
Pearson Chi-Square293890.29033.0385
Scaled Pearson X2293890.29033.0385
Log Likelihood −16458.7718 
Analysis Of Parameter Estimates
ParameterDFEstimateStandard ErrorWald 95% Confidence Limits Chi-SquarePr>ChiSq
Intercept1−0.22250.0178−0.2573−0.1876156.50<.0001
rd_010.52140.08440.35610.686838.19<.0001
rd_11−0.20670.1129−0.42800.01463.350.0671
rd_21−0.11790.1110−0.33550.09961.130.2880
rd_310.06010.0958−0.12770.24780.390.5305
rd_410.18060.09000.00420.35694.030.0448
rd_51−0.09320.0690−0.22840.04201.830.1765
Scale01.00000.00001.00001.0000  

Now let's attend to the overdispersion problem. The big danger with overdispersion is that the standard errors may be underestimated, leading to chi-squares that are too large and p-values that are too low. There are several possible solutions to this problem, one of which is to formulate and estimate a model that directly builds in the overdispersion. One such model is the negative binomial model that will be discussed later in this chapter. But a simpler, though less elegant, approach is to correct the standard errors and chi-squares based on the goodness-of-fit ratios that alerted us to the problem. In PROC GENMOD, this is accomplished by using the DSCALE or PSCALE options on the MODEL statement. For example,

PROC GENMOD DATA=patents;
   MODEL pat79/total = rd_0-rd_5 / DIST=B DSCALE;
RUN;

The DSCALE option uses the deviance chi-square to make the adjustment while PSCALE uses the Pearson chi-square. The adjustment is very simple: Calculate the square root of the ratio of the chi-square statistic to its degrees of freedom. In Output 4.3 this number is reported as the "Scale" parameter in the last line. All standard errors are then multiplied by the scale parameter, which in turn attenuates the chi-squares and the p-values, as shown in Output 4.3.

Table 4.3. Output 4.3 Conditional Poisson Regression Model with Overdispersion Adjustment
Criteria For Assessing Goodness Of Fit
CriterionDFValueValue/DF
Deviance293949.30313.2399
Scaled Deviance293293.00001.0000
Pearson Chi-Square293890.29033.0385
Scaled Pearson X2293274.78580.9378
Log Likelihood −5079.9585 
ParameterDFEstimateStandard ErrorWald 95% ConfidenceLimitsChi-SquarePr>ChiSq
Intercept1−0.22250.0320−0.2852−0.159748.30<.0001
rd_010.52140.15190.22380.819111.790.0006
rd_11−0.20670.2032−0.60500.19161.030.3091
rd_21−0.11790.1998−0.50950.27360.350.5550
rd_310.06010.1724−0.27790.39800.120.7275
rd_410.18060.1620−0.13690.49801.240.2649
rd_51−0.09320.1241−0.33650.15010.560.4527
Scale01.80000.00001.80001.8000  

When this is done for the patent data, we find that only RD_0 retains its statistical significance, and even for this variable the chi-square is greatly reduced. Note also that the coefficients are not modified at all by this overdispersion correction. Other approaches to overdispersion—such as estimating a negative binomial model—might produce different coefficient estimates.

It's also possible to include predictor variables that do not vary with time, although the interpretation of their coefficients is not always straightforward. Output 4.4, for example, shows results for a model that includes the dummy variable for SCIENCE sector and the variable for LOGSIZE of the firm (while deleting the nonsignificant lagged R & D measures). Neither variable approaches statistical significance. Their coefficients can be interpreted as measuring interactions between each variable and time. Like all interactions, these coefficients can be interpreted in two ways. For example, the coefficient of .0275 for SCIENCE represents the difference between the coefficient for SCIENCE in 1979 and the coefficient in 1975. The fact that it is far from statistically significant suggests that this variable has the same the effect in both years. Alternatively, we can interpret .0275 as the increment in the effect of time for firms in the science sector, relative to those not in the science sector. Again, because it is far from significant, we may conclude that the rate of change in the number of patents from 1975 through 1979 is essentially the same for the two sectors.

Table 4.4. Output 4.4 Conditional Poisson Model with Time-Invariant Covariates
Analysis Of Parameter Estimates
ParameterDFEstimateStandard ErrorWald 95% Confidence Limits Chi-SquarePr>ChiSq
Intercept1−0.33350.1100−0.5490−0.11809.200.0024
rd_010.37700.10250.17610.577813.530.0002
science10.02750.0482−0.06700.12190.330.5686
logsize10.01610.0146−0.01250.04481.220.2700
Scale01.79610.00001.79611.7961  

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
13.59.100.205