Nate Silver at 538 talks about the whole dissecting polls deal, or "unskewing" them. All polls must make methodological choices, and all of those choices have advantages and disadvantages. Spending a lot of time trying to dissect those choices and passing judgment on them is not as productive as:<p>1. Looking at aggregates of lots of polls.<p>2. Looking at the variance that a poll captures from one iteration of it to another.<p>Or at least, so he claims. Obviously, he runs a poll aggregator, using a model that heavily weights the trendline of individual polls, so he has a dog in this fight.
One thing I noted is that NY Times says that most polls have four categories of age and five categories of education; except these aren't categories, they are ordinal variables.<p>Age and level of education are slightly co-variant (you don't get many 18 year olds who have a PhD). Because the age classification and education levels are ordinal you should use an ordinal smoothing [0] function to turn them into pseudo continuous variables. Given the continuous and co-variant independent variables (as well as other categorical independent variables) and a categorical dependent variable the best analysis is probably to use a quadratic discriminant analysis (QDA).<p>[0] <a href="http://epub.ub.uni-muenchen.de/2100/1/tr015.pdf" rel="nofollow">http://epub.ub.uni-muenchen.de/2100/1/tr015.pdf</a> and <a href="http://cran.r-project.org/web/packages/ordPens/ordPens.pdf" rel="nofollow">http://cran.r-project.org/web/packages/ordPens/ordPens.pdf</a>
Side rant: what the heck did they implement on NYtimes.com ?! When I quickly click the text (which I compulsively do to select a paragraph etc.), it <i>changes the font size</i> (sigh)
This is really fascinating. I get why the poll creators made these decisions, but the results of the weighting lead to a ridiculous result compared to other polls. Supposedly this poll was extremely accurate in 2012, so who knows?
I'm not sure if he doesn't know his role, and I'm curious how tracking polls like this try to account for the large media attention paid to the poll and its methodology. This guy is known to stats nerds, and they've been tracking his moves and (rather mean-spiritedly) calling him "Carlton" for a while now.
TLDR; The 19-year black man is very small demographic and would have small sample size. Apparently, the sample for LA Times poll includes and outlier who favors Trump which then gets weighted disproportionately to arrive at conclusion that trump is favored by young black voters.
Let's go find the really heavily weighted members of _all_ the polls and dox them too. This way we can screw them all up. Not just one that is influenced by a potential Trump voter.
Regarding the <i>Times</i>' decision to run this article, I wonder how much of it was based upon "hey, polling is kind of goofy" and how much of was "look! Here's another way we can show that Trump isn't really resonating with voters!".