Select Page
Poker Forum
Over 1,292,000 Posts!
Poker ForumFTR Community

**** Elections thread *****

Results 1 to 75 of 8309

Hybrid View

Previous Post Previous Post   Next Post Next Post
  1. #1
    Quote Originally Posted by Poopadoop View Post
    The D swings are the outcome variable, the predictor variable is 2016 vs. 2017. If, say, you measured the swings across a number of years, say 2008, 2010...2016, and the variability in the swings differed depending on which two years you were comparing (i.e., which two years you examined a number of swings over) that would be an example of heteroscedacity.
    I'm referring to the change in the D candidates themselves. Lamb is a vastly different D than the 2016 "D" in the sample. Is that heteroskedastic?

    That's your subjective opinion.
    Another subjective opinion I hold is that scientism is a big problem.

    The claim though is a bit more general than this particular special election in PA. To 'debunk' it you would typically be expected to provide an alternative explanation for every one of those 6 swings. Arguably, these would become more and more complicated until your model includes several extra variables to explain each case. My model only has one variable, Trump. Doesn't mean the more complex model can't be true, but in scientific inference generally the simpler model is preferred.
    Are you suggesting that your model reliably approximates the truth?
  2. #2
    Quote Originally Posted by wufwugy View Post
    I'm referring to the change in the D candidates themselves. Lamb is a vastly different D than the 2016 "D" in the sample. Is that heteroskedastic?
    Your argument is that, in PA, something besides Trump being president changed between 2016 and 2017 that resulted in the D swing. That's got nothing to do with heteroscedascity. Even if it were possible for one predictor and one outcome varaible to result in heteroscedascity (it isn't), the binomial analysis doesn't assume homoscedascity, so it's not relevant. If it were an ANOVA then that assumption would have to be upheld.

    My response to your argument is that explanation only works in PA, it can't explain the other five data points.



    Quote Originally Posted by wufwugy View Post
    Are you suggesting that your model reliably approximates the truth?
    I'm suggesting in the absence of a more compelling explanation it does pretty good, yes. For each election, it's obvious that different things are going to factor into what happens. We can't quantify what those things are or how they affect the swings. Certainly Conor's a different candidate than the last guy and that makes a difference. But it only makes a difference in PA, and we don't know how much of a difference it made. Further, all kinds of other things will have effects as well, because lots of things change between 2016 and 2017, not just the candidates in those special elections. Without any systematic reason to think those changes all favoured D candidates, then the appropriate assumption is that the effects of those unknown variables would tend to self-correct, i.e., cancel each other out.

    The one common denominator in all these special elections is that they took place since Trump was elected. So that is one variable that arguably should impact all their outcomes. Further a change in the PA S.E should not affect the five special elections other states, and certainly not when those elections occurred BEFORE the one yesterday. So to exclude Trump as a cause, you need to go through each S.E. and make separate arguments for why each of those had the swing it had. Maybe you can do that convincingly, I don't know. But until someone does, I'm holding the simplest, one variable explanation as my model.
  3. #3
    Quote Originally Posted by Poopadoop View Post
    Your argument is that, in PA, something besides Trump being president changed between 2016 and 2017 that resulted in the D swing. That's got nothing to do with heteroscedascity. Even if it were possible for one predictor and one outcome varaible to result in heteroscedascity (it isn't),
    Do there need to be two+ input variables for heteroskedasticity to arise? For example, if the regression is income = age, the variation in income over the range of age doesn't result in heteroskedasticity, yet if the regression is income = age + gender, then the variation in income over the range of age does result in heteroskedasticity?

    My response to your argument is that explanation only works in PA, it can't explain the other five data points.
    That's fine. I'm not trying to opine on the other ones. I shouldn't have quoted them.
  4. #4
    Quote Originally Posted by wufwugy View Post
    Do there need to be two+ input variables for heteroskedasticity to arise? For example, if the regression is income = age, the variation in income over the range of age doesn't result in heteroskedasticity, yet if the regression is income = age + gender, then the variation in income over the range of age does result in heteroskedasticity?
    It's a bit complicated to explain (and a bit early in the morning) and not relevant to the binomial test I did because that's a non-parametric test that doesn't make any assumptions regarding how the variance is distributed. I will try to get back to this later.
  5. #5
    Quote Originally Posted by wufwugy View Post
    Do there need to be two+ input variables for heteroskedasticity to arise?
    Sorry, I was thinking in terms of the analysis I had just done when I said one predictor and one outcome variable can't result in heteroscedascity. What I should have done, to be completely correct, is preface that with 'assuming you treat both variables as dichotomous' as in the case of an 'election being held either in 2016 or 2017', where that predictor variable is dichotomous because it has only one of two values (2016 or 2017), and the outcome variable being dichotomous as in either swing(D) or swing (R).

    However, if you wanted to do a t-test, you would be treating the outcome variable as continuous (% swing in either direction). So you would have to jump through certain hoops to fulfill the assumptions of normality, including possibly transforming the data, and you would as a matter of course use a pooled estimate of the variance which generally speaking should adequately address any issues around heteroscedascity.

    If you can't be arsed to do that and just want to do a quick back-of-envelope calculation, you would just choose a non-parametric test which has less power but also fewer assumptions to worry about, which is why I did the binomial one. My guess is that the more powerful t-test would have returned a likelihood ratio closer to 1000:1 in favour of the model assuming an increase in D support from 2016 to 2017 in those six election districts. This is because of the generally large effect (mean = 17.7%) closely clustered around the mean. The binomial test ignores the size of the individual values and only considers whether they are positive or negative, so the evidence it gives in this case, while still strong, is not overwhelming.

    If you have a lot of experience with numbers, you can also use the interocular trauma test, whereby if the data hits you between the eyes you can glean the existence of an effect without carrying out a formal test. A layperson's version of the interocular trauma test would be to believe that the house next to theirs is closer to them than the moon 365 days a year, without carrying out any formal measurements or doing a statistical test.




    Quote Originally Posted by wufwugy View Post
    For example, if the regression is income = age, the variation in income over the range of age doesn't result in heteroskedasticity, yet if the regression is income = age + gender, then the variation in income over the range of age does result in heteroskedasticity?
    Both income and age in the example you give are variables that can have continuous values. If, e.g., you draw a scatterplot of income and age, you'd expect there to be a tight cluster around people 0-1 yrs old having an income of 0, and as age increases, the spread of incomes around age would increase, so that the overall impression would be of a cone-shaped distribution. That would be an example of heteroscedascity, because age would strongly predict income at age 0-1 but not so well at age 50-51 (or whatever). The Pearson correlation coeffecient of that analysis based on a normal distribution would be problematic, and a sensible statistician would not do a Pearson correlation but would do something where normality is not one of the assumptions, known as a non-parametric test, such as a Spearman's rho correlation. Moreover, things would get more complicated if your data included people over 65 who are retired and generally not having a lot of income.

    If you added in gender as a coefficient, its interaction with the other variables would be another potential source of violation of normality.
  6. #6
    Quote Originally Posted by Poopadoop View Post
    I'm holding the simplest, one variable explanation as my model.
    In a way I'm using a simple, one variable explanation as my model too (regarding PA, the only one I've opined on).

    That model is that the D ran as an (effective) R (he really did, pro-gun, anti-abortion, etc.), so change in vote from R to D in this instance doesn't tell us much about what would happen in a hypothetical future election when the D runs as an effective D.
  7. #7
    Quote Originally Posted by wufwugy View Post
    In a way I'm using a simple, one variable explanation as my model too (regarding PA, the only one I've opined on).

    That model is that the D ran as an (effective) R (he really did, pro-gun, anti-abortion, etc.), so change in vote from R to D in this instance doesn't tell us much about what would happen in a hypothetical future election when the D runs as an effective D.

    Then our models are explaining different levels of things.

Posting Permissions

  • You may not post new threads
  • You may not post replies
  • You may not post attachments
  • You may not edit your posts
  •