Dear Maria,

On 08-04-2009 23:06, [email protected] wrote:
The DV of my expt  is a target response (poresp)  which can be either
PO (coded as 1) or DO (coded as 0). So PO is a success and DO is a
failure. The predictors are prime and nounrep (nounrepetition), each
with 2 levels. I have effect coded the predictors (-.5 and +.5
respectively for each of the two levels of the predictors, following
Barr, 2008): with this coding, which I believe is a way of centering,
the intercept should correspond to the grand mean and the coefficients
to the differences between the means etc., like in an ANOVA. Please
correct me if I am wrong. I tried this coding in the past where the
outcome was a continuous variable and indeed when I checked the means
and differences in my data , I found a correspondence between the ANOVA
measures and the regression intercept and coefficients; however, now I
am dealing  with binomially distributed data, where the model parameters
are in log odds space, so it is a bit more complicated, because I have
to do transformations).
The problem is that when I check the output of the mixed  logit
regression (using lmer and family = binomial ) and try to make sense of
the coefficients by checking, for example, whether the intercept
corresponds to the grand mean, in log odds, of course, things do not
match. I checked and re-checked, and I do not know where I am going
wrong.
==========

head(verbdiff)
     X subject cond item myscore poresp nounrep prime primec nounrepc
1 1069       1    2   19       p      1       0     1    0.5     -0.5
2 1070       1    3    6       p      1       1     0   -0.5      0.5
3 1071       1    3    5       p      1       1     0   -0.5      0.5
4 1072       1    2   12       p      1       0     1    0.5     -0.5
5 1073       1    4   16       p      1       1     1    0.5      0.5
6 1074       1    3   14       p      1       1     0   -0.5      0.5

#primec and nounrepc are effect coded



COUNTS OF 0s and 1s:

xtabs (~ poresp +prime +nounrep, data=verbdiff)

     nounrep = 0

      prime
poresp   0   1
     0  89  58
     1 207 235

, , nounrep = 1

      prime
poresp   0   1
     0  91  64
     1 202 228


MODEL:

verbdiff.lmer = lmer(poresp ~ primec *nounrepc + (1|subject) +
(1|item), data= verbdiff, family="binomial")

print (verbdiff.lmer)


Generalized linear mixed model fit using Laplace Formula: poresp ~ primec * nounrepc + (1 | subject) + (1 | item) Data: verbdiff Family: binomial(logit link)
  AIC  BIC logLik deviance
 1092 1122   -540     1080
Random effects:
 Groups  Name        Variance Std.Dev.
item (Intercept) 1.2143 1.1020 subject (Intercept) 1.8470 1.3590 number of obs: 1174, groups: item, 40; subject, 32

Estimated scale (compare to 1 ) 0.917418
Fixed effects:
Estimate Std. Error z value Pr(>|z|) (Intercept) 1.6605 0.3121 5.320 1.04e-07 ***
primec            0.7854     0.1645   4.773 1.81e-06 ***
nounrepc -0.1054 0.1612 -0.653 0.513 primec:nounrepc -0.2138 0.3224 -0.663 0.507 ---
Signif. codes:  0 ?***? 0.001 ?**? 0.01 ?*? 0.05 ?.?
0.1 ? ? 1
Correlation of Fixed Effects:
            (Intr) primec nonrpc
primec 0.053 nounrepc -0.009 -0.023 primc:nnrpc -0.012 -0.028 0.128

=============

QUESTIONS:

1. THE ESTIMATED INTERCEPT IN THE LOGIT MODEL IS 1.66 IN LOG ODDS,
WHICH should correspond to grand mean, i.e. likelihood of successes (=PO
responses coded as "1") independent of the predictors.

THERE WERE OVERALL 872 SUCCESSES AND 302 FAILURES IN THE EXPT, SO ODDS
SHOULD BE 872/302=2.88 or (in probability space) .74/.26 = 2.85;
THIS SHOULD GIVE A LOG OF ODDS OF APPROX 1.05, BUT THE INTERCEPT
PREDICTED BY THE MODEL  IS MUCH HIGHER (1.66)

Indeed:
> logit(872/(872+302))
[1] 1.060362

This may be because with your coding, you are estimating the intercept if primec==0 & nounrepc==0. This means that if prime and DV are related (e.g. because there are more hits if prime==1), or likewise if noun and DV are related, then the estimated intercept is no longer the grand mean, but rather the estimated value of your DV if all predictors would be zero.


SAME THING WHEN I USE THE CORRESPONDING DUMMY CODED (0-1) V
ARIABLES
-INTERCEPT IS MUCH HIGHER THAN WHAT MY DATA SUGGEST


In my experience, models are far easier to interpret if you do NOT centralize the binary dummy predictors. The estimated intercept then equals the mean logit for the particular cell where all dummies are zero. So, my advice is to use your dummy coded variables and not the centered dummies.

The main point I guess is that the intercept does not reflect the grand mean, in your model, but it reflects the estimated DV if all predictors would be zero, which is different from the grand mean IF predictors and DV are related.

Did you try this model...
>> verbdiff.m0.lmer = lmer(poresp ~ 1+(1|subject) +
> (1|item), data= verbdiff, family="binomial")

It should give you ONLY the intercept which should then be about 1.06.

More background and details are available at
  http://www.hugoquene.nl/mixedeffects/
and references in the articles mentioned there.

Hope this helps, with kind regards, Hugo Quene


--
Dr Hugo Quené | assoc prof Phonetics | Utrecht inst of Linguistics OTS | Utrecht University | Trans 10 | 3512 JK Utrecht | The Netherlands | T +31 30 253 6070 | F +31 30 253 6000 | [email protected] | www.hugoquene.nl | www.hum.uu.nl


_______________________________________________
R-lang mailing list
[email protected]
http://pidgin.ucsd.edu/mailman/listinfo/r-lang

Reply via email to