Hi, I'm Adriene Hill and welcome back to Crash Course Statistics.
In our last episode we talked about how we use experiments to imitate having two parallel
universes to test things.
But sometimes you can't do certain experiments without becoming an all-powerful and evil
dictator, and since it's statistically unlikely that any of you are evil dictators, today,
we'll explore those methods.
Like we mentioned at the beginning of the series, you're not always able to answer
the questions you really want to answer using statistics.
For example, it would be great to experimentally test whether getting married increases your
lifespan, but you can't randomly assign some people to be married and force another
group to be single.
Not only would that be difficult to enforce, it would also be pretty unethical, though
I suppose you being evil takes care of that particular concern.
Similarly we can't assign someone to be a twin, or a Democrat, or a smoker.
But that doesn't mean we should just give up and stop trying to find out more about
these topics.
Not at all.
Instead we just need a different method to collect data.
Enter Non-Experimental methods.
INTRO
One of the most common non-experimental method is the survey.
From user experience surveys on websites, to political polls, to health questionnaires
at the doctor's office, you've probably taken hundreds of surveys in your lifetime.
There are two things that can make or break a survey: the questions, and who the researcher
gives the questions to.
The goal of a survey is to get specific information.
Say you're walking your dog in a local park, and someone approaches you and asks you to
take a survey on local businesses in your town.
When you look at the questions you notice that none of them are about local businesses,
instead you find yourself answering questions about your politics and religious beliefs.
Unless the surveyor was lying to you about their purposes, this is not a very good survey….It's
also not a very good lie A survey should measure what it claims to
measure.
It might seem obvious that having only unrelated questions on your survey is problematic, and
there are even more subtle ways a question can be biased.
Let's take a look at a few questions from a health survey you might take at a doctor's office.
The first question asks you how often you exercise: never, less than 30 minutes a week
or 30 minutes a day.
So what do you answer if you exercise for half an hour twice a week?
Or if you're on swim team and exercise for at least an hour a day?
And does walking count as exercise?
Multiple choice questions that don't offer all possible options and/or an "Other"
option can cause respondents to either skip the question, or feel forced to choose an
answer that isn't accurate.
Claims made using these questions aren't as strong as they could be if people were
offered a full range of choices.
The next question asks you "Answer yes or no: I don't smoke because I know it's
damaging to my health" this is a leading question since the wording leads to towards
the quote "desired" answer.
This is especially effective when a question deals with sensitive issues like smoking,
politics, or religion.
People answering the questions want to be seen in a positive light, and so they tend
to give the answer they think is "appropriate".
While having people fill surveys out anonymously by themselves can help, it can sometimes be
the case that respondents don't want to admit things--even to themselves--that are
socially undesirable.
In general terms, good survey questions are worded in a neutral way such as asking "how
often do you exercise" or "describe your smoking habits" instead of using wording
or options that push survey takers in a certain direction.
And while your doctor wouldn't...or shouldn't...do this...sometimes groups purposely use biased
questions in their surveys to get the results that they want.
Apparently, back in 1972, Virginia Slims conducted a poll asking respondents if they would agree
with the statement: "There won't be a woman President of the United States for a
long time and that's probably just as well."
Not a well-written question.
Biased questions can be more subtle...and can lead to skewed reports of very serious
things like sexual assault, or mental health conditions.
It's important to always look for biased questions in surveys, especially when the
people giving the survey stand to benefit from a certain response.
Even when researchers have created a non-biased survey, they still need to get it into the
right hands.
Ideally, a survey should go to a random sample of the population that they're interested
in.
Usually this means using a random number generator to pick who gets the survey.
We do Simple Random Sampling so that there's no pattern or system for selecting respondents
and each respondent has an equal chance of being selected.
For example, telephone surveys often use Random Digit Dialing which selects 7 random digits
and dials them.
When someone picks up, they're asked to take a survey.
But here's where we hit our first issue.
If people aren't forced to respond to the survey, we might experience something called
Non-Response Bias in which the people who are most likely to complete a survey are systematically
different from those who don't.
For example, people with non-traditional working schedules like retirees, stay at home parents,
or people who work from home might be more likely to answer a middle-of-the-day phone
survey.
This is a huge problem if those groups are different than the population as a whole.
If your survey was on health insurance plans, or political opinions, it's likely that
these three groups would have different opinions than the population, but they represent the
majority of survey responses, which means your data won't represent the total population
very well.
This is also related to Voluntary Response Bias in which people who choose to respond
to voluntary surveys they see on Facebook...or Twitter... are people who again, are different
than the broad population.
This is especially true with things like customer service surveys.
People who respond tend to have either very positive or very negative opinions.
See the comment section below.
The majority of customers with an average experience tend not to respond because service
wasn't noteworthy.
Wait.
Does that mean I'm not noteworthy?
Another source of bias is just plain underrepresentation.
If a group of interest is a minority in the population, random sampling paired with response
biases might mean that that minority isn't represented at all in the sample.
Let's say there is a city where 5% of the population is single mothers, it's entirely possible
that the sample will contain no single moms.
To overcome these issues, we have a couple options.
We could weight people's responses so that they match the population (like, counting
the few single mothers who do respond multiple times so that they count for 5% of the total sample).
But, this can be problematic for the same reasons that response bias is problematic.
If the few single mothers who respond don't represent all single mothers, our data is
still biased.
In a 2016 LA Times/USC political tracking poll, a 19-year-old black man was one of 3,000
panelists who was interviewed week after week about the upcoming presidential election.
Because he was a member of more than one group that was underrepresented in this poll, his
response was weighted 30x more than the average respondent.
According to the New York Times, his survey boost his candidate's margins by an entire
percentage point.
Stratified Random Sampling is another option.
It splits the population into groups of interest and randomly selects people from each of the
"stratas" so that each group in the overall sample is represented appropriately.
Researchers have used stratified sampling to study differences in the way same-sex and
different-sex couples parent their kids.
They randomly select people from the same-sex parenting group and... randomly select people
from a different-sex group of parents to make sure that they're well represented in the
sample.
Another issue is that getting surveys to people can be expensive.
If a cereal company wants to see how families react to their new cereal, it would be costly
to send some cereal to a random sample of all families in the country.
Instead they use Cluster Sampling which create clusters (not Honey Nut Clusters) that are naturally occuring (like
schools, or cities) and randomly select a few clusters to survey instead of randomly
selecting individuals.
For this to work, clusters cannot be systematically different than the population as a whole and
and they should about equally represent all groups.
Issues can also arise when the population being surveyed is very small or difficult
to reach, like children with rare genetic disorders, or people addicted to certain drugs.
In this case, surveyors may choose to not use randomness at all, and instead use Snowball Sampling.
That's when current respondents are asked to help recruit people they know from the
population of interest... since people tend to know others in their communities and can
help researchers get more responses.
And note that these sampling techniques can and are used in experiments as well as surveys.
There are also non-experimental data collection methods like a Census.
A Census is a survey that samples an ENTIRE population.
The United States conducts a Census every 10 years, with the next one scheduled to be
done in 2020.
It attempts to collect data from every.single.resident of the United States (even undocumented residents,
and homeless residents).
As you can imagine, this is hard, and it is not without error.
In Medieval Europe, William the I of England conducted a census in order to properly tax
the people he had conquered.
In fact a lot of rulers tended to use censuses to know just how much money they should be
demanding.
Until the widespread availability of computers, the US census data took almost 10 years to
collect and analyze.
Meaning that the data from the last census wasn't even available until right before
the next census.
The length of time it took to complete the census is part of the reason we even have
computers...check out our CompSci series for more on that.
So why collect census data--instead of just sampling the population?
In the US--the Census could cost more than 15 Billion dollars in 2020.
There are a lot of reasons.
The constitution says we have to, but also the census provides the truest measure of
the population we can get.
It minimizes sampling error.
It also functions as a benchmark for future studies.
And a census can give researchers really specific information about small groups of the population--information
that might be hard to gather with regular sampling methods.
Doing statistics on Census data is different, because most statistical inference aims to
take a small sample and use it to make guesses about that population.
But with a census we already have data from the entire population, we don't need to
guess if there are differences, we can just see them.
Analysis on Census data is usually more concerned with whether differences we see are large
enough to make a difference in everyday life, rather than guessing IF there is a relationship.
The census as we said can take years.
And entire countries to fund.
That doesn't discount the value of sampling.
But we should be cautious...Badly worded polls, fake polls, and biased polls are common.
So are the results of those polls.
The statistics-friendly website FiveThirtyEight put together a great list of advice on how
not-to-fall for a fake poll.
Among its advice--Ask yourself if it seems professional.
Check to see who conducted the poll--and if you trust them.
See how the poll was conducted.
Check out the questions they asked...and who they asked.
If it seems fishy.
It probably is fishy.
That said, well done surveys are essential.
They allow us to get information without all the trouble of doing an experiment, and since
they're comparatively easy, they're popular ways for businesses, countries, and even Youtube
channels to collect information.
In fact Crash Course Statistics has its own survey! The Link is in the description.
And it takes way less time than the Nerdfighteria one. I promise.
Thanks for watching. I'll see you next time.
No comments:
Post a Comment