What’s dumber than these surveys which are always filled with loaded questions and questionable sampling methods, is the fact people take them as fact without even looking into them. I spent so much time trying to track down where the original survey is, and it doesn’t exist. It’s not on the Change Research website, it’s not linked by any article that mentions it, and it’s not reported by any credible outlet. It’s reported on by 5 or so tabloids like The NY Post, The Daily Mail, and Teen Vogue and nothing more.
We used the following sources to recruit respondents:
● targeted advertisements using the Meta advertising >platform
● SMS text messages
Regardless of which of these sources a respondent came from, they were directed to a survey hosted on
SurveyMonkey’s website.
Ads placed on social media targeted likely voters nationwide. Those who indicated that they were not
registered to vote were terminated. Those who indicated they were over the age of 34 were terminated. As the survey fielded, Change Research used dynamic online sampling: adjusting ad budgets, lowering budgets for ads targeting groups that were overrepresented, raising budgets for ads targeting groups that were underrepresented. The survey was conducted in English.
So this is a self reported online poll with 84 oddly phrased questions that was advertised primarily on Instagram and Facebook and conducted on a redirected website. The methodology seems dodgy, and even the people conducting it agree:
We adopt The Pew Research Center’s convention for the term “modeled margin of error”(1) (mMOE) to indicate that
our surveys are not simple random samples in the pure sense, similar to any survey that has either non-response bias
or for which the general population was not invited at random. A common, if imperfect, convention for reporting survey
results is to use a single, survey-level mMOE based on a normal approximation. This is a poor approximation for
proportion estimates close to 0 or 1. However, it is a useful communication tool in many settings and is reasonable in
places where the proportion of interest is close to 50%. We report this normal approximation for our surveys assuming a proportion estimate of 50%
What’s dumber than these surveys which are always filled with loaded questions and questionable sampling methods, is the fact people take them as fact without even looking into them. I spent so much time trying to track down where the original survey is, and it doesn’t exist. It’s not on the Change Research website, it’s not linked by any article that mentions it, and it’s not reported by any credible outlet. It’s reported on by 5 or so tabloids like The NY Post, The Daily Mail, and Teen Vogue and nothing more.
2 minutes of Googling led me to the poll: https://changeresearch.com/post/poll-young-voters-facing-tough-economic-reality/
The question asked was “Do you consider each of the following behaviors or traits a green, beige, or red flag in a partner?”.
Hmm, not seeing that in the link provided…maybe it’s been updated, or an issue on my end?
It’s question 78. You have to actually go into the sample and methodology document.
Thanks kind internet stranger! Also, am I just obtuse, or did the last questions not have response data?
As an unrelated aside, who the hell has that kind of time for a survey?!
Okay, mild progress… But where’s the data for it?
Also, this is from the methodology pdf:
So this is a self reported online poll with 84 oddly phrased questions that was advertised primarily on Instagram and Facebook and conducted on a redirected website. The methodology seems dodgy, and even the people conducting it agree:
Thanks for the link regardless