Talk of a silly game show leads to a good question about surveying methodology. Namely, could we consider a survey of 100 Americans to be a fair representation of America at large? This question, although focused on a game show, is relevant today for many reasons. Consider political poling,
Even though many utilities do not compete for their customer
base, it can be very useful to understand their customer base,
their attitudes toward energy consumption, their use of energy
consuming equipment in the home, or their awareness and
understanding of utility programs. Sure, a utility carefully
designed a program that is to the benefit of their retail
customers and the utility, and then promoted it through a bill
insert and in a couple of public forums. But what good is that
program if only a few of the customers know about it? A survey
can help answer that question. Think about it for a moment: is
asking 100 customers a fair representation of your entire
customer base? If not, how many do you think you need to
survey to gain representative information?
It turns out there are several major elements to determining the
appropriate sample size. I’ll touch on each briefly, without using
a single math equation.
The first key point to establishing a good survey sample
size is to identify what is being measured with the survey. There
are two types of information we are interested in measuring. The first is knowing what percentage of the population responds in a certain way. This is the Family Feud model, in which we are
asking, what percentage of Americans say “car” when asked,
“Name something with wheels”. The second type of
measurement is the value of a continuous variable. For instance,
trying to measure the peak demand for a typical home in a
service territory. The necessary statistical formulas are different
for determining the appropriate sample size.
It’s hard to escape the concept of confidence when doing statistical experiments. It’s natural to want to draw one sample (run a single survey) and make inferences about the entire population from that single sample. The level of confidence in the survey results is established at the outset and has an impact on sample size. Confidence levels typically range from 85% to 99%, and the greater the confidence desired, the larger the
The level of precision is reported after the survey has been conducted and is often reported as the margin of error. The precision determines how narrow the margin of error is. When designing a survey sample size, it’s necessary to assume a targeted level of precision, usually 5% or 10%. Once the survey
is conducted, though, the actual precision will be measured when summarizing responses. Initially, the survey may have a
targeted precision of +/- 10%, but enough survey responses are
received that the precision is +/- 6%. The more precise the survey needs to be, the larger the sample size of the survey.
Another key factor is the size of the population we are trying to represent in the survey. For a political poll, pollsters are trying to represent the full population of voters. An extremely challenging population to pin down since voter turnout is difficult to predict. For a utility, things get a little easier because the utility already knows the number of retail customers on its system. The smaller the population, the lower the sample size needs to be to draw the same level of precision and with the same confidence.
We have touched on the major factors that impact sample size and it has made me thirsty. Well, one of the other major factors in sample size calculations is attributable to one tasty beverage
The answer, it turns out, is maybe. For example, a survey needs 68 responses from a very large population to achieve a 90% confidence and +/- 10% precision. But to get more precision requires more responses, for example, a 95% confidence and +/-5% precision level requires just under 400 responses.
Just as important as the sample size is ensuring the sample is
truly representative of the population of interest. This is the part
political pollsters try very hard to accurately capture, by trying to
poll likely voters with an appropriate mix of political persuasions. Errors in the representation of their polls will lead to big variations between election day results and their polls.
Likewise, if the utility only receives responses from a certain segment of their customers, then it may interpret the results incorrectly. Therefore, survey design should also carefully consider the representative nature of the sample drawn.
That’s probably more information about sample size than anyone other than an ardent stats nerd or quant would want to
know. The point is surveying is easy and I encourage utilities to
take advantage of seeking information about their customers.
With internet and email, it can be simple to run a survey and
collect information that can be useful for various planning efforts, including developing load forecasts, fine tuning utility
programs, or just understanding customer attitudes about the
utility. However, easy execution of a survey should be balanced
with using the tools correctly to avoid making the mistake of
drawing incorrect inferences from survey data collected.
For more information or to comment on this article, please contact:
GDS Associates, Inc.
Marietta, GA
770-799-2381 or
jacob.thomas@gdsassociates.com