|
|
|
| Subscribe |
Sign up here to receive the TCEC newsletter direct to your inbox every quarter!
| |
|
Welcome
|
|
A recent posting on the American Evaluation Association listserve said, "Evaluation...is not just a matter of finding and reproducing information, but of doing something with it!"
If that's not already how you think of evaluation -- in terms of how it can help guide and inform your program -- then we hope the articles in this and future newsletters lead you there.
As many of you wrap up your final round of data collection activities for the 2007-2010 contract period, we've focused this issue on ways to analyze, report and use your data.
Hope you find it helpful,
Robin Kipke
Editor
| |
|
|
|
Creating Beautiful Evaluation Results Source: WestEd
Maybe you have a few hundred public opinion polls, some
store surveys, or observations from parks or apartment complexes. It was hard work to collect that data, and
now you want to share the results. But
how do you do that? More precisely, how
do you present the evaluation results in ways that are interesting and
engaging? Edward Tufte, an expert on the visual display of
quantitative data, had good reason to name one of his books Beautiful Evidence. Each table, chart, graph or map in his book
tells an amazing story. With some
planning and commitment to simple design, your evaluation results can tell a beautiful
story, too.
Why Make Tables and Charts?
Most of the results from your evaluation data can be
summarized in the text of your progress reports and final evaluation reports,
and in fact sheets, press releases and other documents you use to communicate
with community stakeholders. In the case
of a public opinion survey, you could write about the number of surveys completed,
or how many were compl  eted in each location respondents were recruited into the
sample. You could also summarize the
survey results on each topic or question, for example, describing the
percentage of smokers versus non-smokers who favored smoke-free multi-unit
housing. But there are times when you will want to make the extra
effort to create a table or chart to highlight a point that you want
to make. Usually, that's a good idea
when you want to show particularly interesting, important or unexpected results that really
need to stand out. Another instance is
when the data are so complex or comparisons so numerous that a visual
representation would illustrate the comparison more clearly than mere statistics. How many tables or figures should you include? Try to strike a balance between giving your
readers enough information and not overwhelming them. In a short document like a press release or
brief, one table or chart is about right, since this will focus the readers'
attention on the main point you are making.
In a longer document like a final evaluation report, you can use five
to seven figures to illustrate a longer "story" you are telling.
Beautiful Tables
Tables are one way to visually display data. The key is to summarize the raw data into
meaningful categories, and to present the results in a way that are clear. The reader should be able to see the table and
understand exactly what you are trying to say without reading the accompanying
text. Of course, you should still
summarize the main results of the table in the narrative section of the final
evaluation report.
Here are some key features of a well designed table. - A
title. The title describes what the
table is displaying, along with a location and year the data was collected.
- Simple
formatting. Edward Tufte recommends
using the newspaper, especially the sports section, as a model for data
tables. They are simply formatted
and designed to communicate effectively to a wide audience.
- Label
each column and row, and avoid overlapping categories -- for instance
"convenience stores" and "family owned convenience stores" (are these
distinct categories, or is one a subset of the other?). Show totals in the last row.
Use tables rather than graphs if you want to do a comparison
of multiple data because few graphs can accommodate such
comparisons, and they are very difficult to decipher. Here is an example of a table that shows cities
as rows and the number of stores surveyed by year as columns. The table is easy to read and very
informative:
Beautiful Graphs and
Charts
Charts and graphs serve as "pictures" to illustrate
relationships. Some people find them
easier to interpret than a narrative description or table in the same way a
photograph can portray a thousand words.
Some key elements of good graphs and charts are: - A
title. Just as with a table, the
title describes what the graph or chart is displaying, along with a
location and year the data was collected. If you use multiple graphs and charts,
begin the title with "Figure 1" and then number each figure in sequence so
that you can refer to them accurately in the text.

- Keep
the format simple. Pie and bar charts
will probably meet most of your needs.
Pie charts show the distribution of a characteristic; bar charts
can show that as well as changes over time. Some experts say that it's harder to
"see" differences in pie charts than in bar charts. Don't use cones, pyramids, or other
shapes, even if Excel allows you to.
Unusual shapes, along with three-dimensions, only add clutter and do
not communicate any additional information.
- Label
clearly so the reader does not have to guess. Each bar or pie slice should have a name
and a value associated with it. Bar
charts should have labels for both axes.
Even if it's obvious to you that you are showing the percent of
stores selling to minors, for instance, the reader might not know that. Use a legend if many different categories
are displayed; otherwise, there is no need for a legend.
- As
with tables, state the total number of people or things you are counting,
and the number and/or percent in each category.
For More Information
Edward Tufte's website includes a forum and a schedule of
classes he teaches around the country. http://www.edwardtufte.com/tufte/. Dr. Tufte's books can be ordered from that website. The books are available free in most
university libraries, also. Just Plain Data
Analysis is a book and website hosted by Gary Klass at Illinois State
University. http://lilt.ilstu.edu/jpda/. The website has great tips on formatting bar
charts, pie charts and other means of displaying results, along with basics on
using Microsoft Excel (2003 and 2007 versions) to create charts and graphs.
TCEC staff are also available to answer questions about the
display of data. So contact us! Beautiful graphics created by WestEd back to top
|
|
Upcoming Regional Trainings on Writing Great Final Evaluation Reports
The Tobacco Control Evaluation Center will be hitting the
road to offer trainings around the state on evaluation report writing.
Open
to project directors, evaluators and staff of all TCP-funded programs
in CA, these interactive trainings will cover the components of a final
evaluation report, how your reports are scored, tips on writing the
results section, and ways you can repackage pieces of the report for
use with different audiences. 
Trainings will be offered in the following locations: Gold Country (Sacramento) - April 7
Monterey Coast (Paso Robles) - April 13
Bay Area (Oakland) - April 15
Central Valley (Fresno) - April 19
San Diego Region (San Diego marina) - April 28
Greater Los Angeles Area (downtown LA) - April 30
Northern CA Region (Redding) - May 3
The sessions will run from 10 a.m. to 3:30 p.m., with
an hour afterward for "office hours" with TCEC staff if you would like individualized consultation.
Registration is free, and lunch is included.
Travel reimbursements are available for CTCP grantees traveling 100+ miles one-way to the conference location.
Contact Diana Dmitrevsky at ddmitrevsky@ucdavis.edu or (530) 752-9951 for
more information. Prior approval for travel reimbursement is required.
Seating is limited, so register now by clicking this link! Once you register, we'll send you a confirmation with the date and
specific location, along with a map and agenda.
Please help us build our image archive of tobacco control activities by sharing your photos with us. We are looking for shots of people conducting surveys, interviews,
observations or evaluation trainings. We also could use photos of
apartment complexes, parks, outdoor dining, rodeos, tobacco
advertising, smoking areas, signage, parks, beaches, casinos, etc. in
your community. For more submission details, click here.
Stay tuned for the next issue of our quarterly newsletter
coming in May which will cover topics like: - Telling a story with your results
- Reporting strategies
- Writing final evaluation reports
back to top
|
|
More than Luck: Producing Usable Data for a Smoke-free Casino Objective
The following is an
excerpt from an interview with Stephanie Taylor, internal evaluator and
epidemiologist for Shasta County conducted by TCEC evaluation associate Travis
Satterlund as part of the January 28th webinar on "Journey of a
Survey: Roadmap to Usable Data." To hear
a full recording of the entire webinar and interview, go to the TCEC website.
T: Hi Stephanie, we
really do appreciate you joining us today.
One of the reasons we asked you to do this is because of your compelling
story at Shasta County.
With that said, do you mind giving us an overview of your data
collection activity and what you did up in Shasta regarding the casino?
S: We originally
didn't have this activity in our scope of work.
The casino came to us. We had
worked with them previously on an objective several years ago and built a good
relationship with them...over the years. In
Spring or Summer of 2008, they wanted information about the effects of secondhand
smoke because the casino manager was concerned for the health of
employees. That snowballed into a data
gathering effort to support possible smoke-free casinos. Previously, when they tried to go smoke-free,
they didn't have any data to show if there was really any support for it. So what we tried to do is to gather that data
to gauge the level of support.
T: What did you ask
in the surveys?
S: In the casino
employee survey, we asked them about what kind of work they do so we could
determine their level of secondhand smoke exposure, how often patrons smoke
around them, whether they were bothered by it, and whether or not they were a
smoker. We also tried to gauge their
level of support by asking questions about if they would prefer to work in a
non-smoking room or in a non-smoking casino.
In the patron survey, we asked about smoking status, how often they came
to the casino, why they come to the casino, and how often they would come if
there was a smoke-free policy. In the
tribal membership survey and in the survey of those that worked for the
Rancheria, we asked about their current and former smoking status, their level
of concern for the casino employees with regard to how much secondhand smoke
they are being exposed to, whether they would support a smoke-free casino, and
whether a policy would conflict with their beliefs and values.
T: When all these
surveys came in, how did you go about analyzing them?
S: We used
SurveyMonkey to do the basic frequency analysis for us. So it did most of the work, but I did a
couple of cross tabs by downloading the raw data.
T: What variables were
you looking at?
S: I looked at how
often the employees who worked in certain areas of the casino like the back of
the house [administrative offices] verses the floor [food, gaming or security] were bothered or exposed to secondhand smoke.
I also wanted to combine the surveys to look at the three ways that the
respondents could express support. Let
me explain. We asked the Rancheria
employees and the tribal membership directly if they [would] support a
smoke-free casino, but on the other two [surveys], we didn't ask directly. We asked the patrons how often they would
come (less often, more often or about the same), and the casino employees if
they would prefer to work in a non - smoking casino. We combined the three surveys to look at the
level of support across all respondents.
T: So what did you
find in the results? 
S: It looks like we
had good level of support of non-smokers across the board in all the surveys
for a non-smoking casino, and also not a lot of opposition. The employees that were more exposed to
secondhand smoke were more inclined to want to work in a smoke-free casino than
those who work in the back of the house where some had more of an ambivalent
response, or no opinion. For the patron
survey, we did get a lot of smokers that responded to the survey, so that
demonstrated some [selection] bias....We
did some air quality monitoring data at the same time as we did the surveys and
that included smoker density counts. From
those, we estimated that anywhere between 18-33 percent of the patrons were smokers. Yet about half of the patron survey [respondents]
were smokers. But surprisingly we still
didn't see overwhelming opposition.
T: In terms of your results,
were you able to share them with the casino management? If so, what was their reaction?
S: ...It helped that we had a champion from inside
and the casino manager who has always been concerned about the health of the
employees. But after we shared the
results with them, it kind of solidified their resolve knowing that they had a
good amount of backing for the smoke-free casino. That was a good tool for them to use when
they took it to tribal council.
T: Thank you very much
Stephanie, this really was a tremendous success story and we thought our tobacco
partners would want to hear this.
Stay tuned on PARTNERS
to hear how the casino chooses to move forward!
Photos by Jim G and twoGiraffe
back to top
|
|
Analyze This! Doing Quantitative Analysis
You have piles of surveys on your desk, and your computer
practically spills over with data. Now what? To make sense of all of those
responses and figure out what they can tell you will require quantitative analysis.
Quantitative refers to measurable numbers and percentages, including relationships
between variables and statistical significance.
Let's look at why the numbers are important. An example of a
quantitative evaluation activity often used in process evaluation is a public
opinion survey. The results will be used to help convince policy makers, so the
main purpose is to gather answers to the question "Are you in favor of XYZ
policy?" Numbers matter in this case because the more people voice that they
are in favor of a policy, the stronger the campaign becomes.
But chances  are you will also ask additional questions that can
inform your campaign -- questions that measure awareness of the issue or explore
various options for the way the policy could be framed, for example. What's key
is that you can quantify the answers and show how many people answered one way
or another. If you are also interested in showing how people with different
characteristics answered the questions, you can add some demographic questions
and analyze results by grouping responses according to demographic
characteristics.
In outcome evaluation, quantitative data can show how much
change occurred because of an intervention. Is there less cigarette litter now
than before the non-smoking policy? Are fewer participants smoking than before
they took a cessation class? Are more apartments designated as non-smoking than
before the housing campaign? And did the difference occur across the board or
only for segments of the population or the group of interest?
Although statistical analysis of quantitative data can get
rather complicated, most Tobacco Control evaluation activities require only the
most basic statistical concepts, which are percentages, means, and median. These calculations can be fairly easily
arrived at, especially with the help of software like Excel or an online
program like SurveyMonkey which does the analysis for you.
Percentages
Once data are entered, you can start calculating. If
calculating percentages, make sure you look at how many people answered each of
the questions. For instance, "30% of respondents said 'yes' when asked if they
smoked." Note that the percentage is of respondents rather than the total
number of people asked. If your sample size was 120 but only 100 of
those approached actually answered, then the percentages are
calculated using the respondents only (n=100).
Means and Medians
Let's say you did a survey of bar night participants where the
average age was important to know. Using
the data from the chart below, the mean (or average) age of respondents was 28
years. The mean is easily calculated by adding all the ages and dividing by the
number of resp  ondents. But looking at this scatter plot graph which shows
us each of the data points, we can easily see that a few respondents were much
older than the rest. This shifted the
average age higher than the age of most participants. The actual midpoint, or
median, of the ages of all participants is 26 years. This difference between
the 26 years of the median and the 28 years average shows us what effect a few
outliers, or atypical cases, can have on the mean. For this reason, it is a good idea to report both
the mean and median and/or standard deviation (a measure of the spread of data points from the mean) because a sophisticated critic
might challenge results that show only means.
Relationships between
Variables
If you are interested in showing relationships between
variables (the things that were measured) or showing statistical significance (which
can be useful if only a small percentage of difference is expected between
categories or in change over time), a statistical program like SPSS is best
suited for calculation.
Let's look at an example: Casino employees and patrons answered
the same questions about smoking preference in the casino. When analyzing
results, you want to see if employees answered a particular question
differently from patrons. A statistical program lets you compare the variables
and will tell you whether or not the difference is "statistically significant,"
meaning that it is not only accidentally different, but that the difference is
big enough given the sample size that it can be considered to have something
to do with the variable (e.g., the status of the person answering -- employee or
patron) and not by chance.
Statistical significance is especially relevant for comparisons
over time or between groups; for instance you may want to show whether or not
and to what extent a difference occurred after your intervention. When the
difference is very obvious, it is not necessary to figure out statistical
significance. If before a smoke-free parks policy was passed 200 pieces of cigarette
litter were collected compared to only 10 pieces after its passage, the
difference is obvious. But what if smoking incidences among students at a
school declined by 1.4 percent? Is that statistically significant or just a
matter of chance? Would another round of the intervention yield
the same results? A statistical program can easily tell us whether mere chance was
at work or if we are actually on to something.
Be aware, though, that statistical significance only shows that there
is a difference. It cannot tell us whether or not the difference occurred
because of the intervention.
Using Results to
Guide Program Efforts
So what if a person's status (employee or patron) predicts their
smoking preference? What does this mean for the intervention? Let's say most
patrons want to smoke but the employees don't. The purpose of pursuing a
smoke-free casino is to protect employees and patrons from secondhand smoke, but casinos are concerned about losing patrons whose desire to smoke is not accommodated. One
recommendation could be to launch an education campaign directed towards
patrons who may become more supportive if they understand that a smoke-free casino policy is an attempt to protect a group of people rather than to take
freedoms away from smokers.
If even the patrons remain non-supportive, the campaign could
take other tacks like collecting information about the experience of casinos that have gone smoke-free, exploring the financial benefit of reduced
employee sick time or cleaning costs, or touting the likely economic advantage of
marketing to patrons who would be attracted to a smoke-free venue.
So whatever your quantitative data reveals -- whether support
or opposition for your campaign -- it can be useful to your program in deciding
what needs to be done next.
For more ideas on how to analyze quantitative data with
descriptive statistics, check out Tips & Tools #10 on the TCEC website. Photo by Robyn Lee
back to top
|
|
New Evaluation Associate Joins Our Team
The Tobacco
Control Evaluation
Center would like to introduce
the newest member of our team, evaluation associate Travis Satterlund. Travis comes to us from UC Berkeley, where he
recently completed a two-year post-doctoral fellowship in the School of Public Health. As a researc  h scientist, he combined his
interests in qualitative research methods, public policy and law by conducting
hundreds of hours of observations and interviews in Bay Area bars and taverns to evaluate California's smoke-free
workplace law. (If you ever need a
recommendation for a good pub in the City, he's your man!) As part of these studies, Travis was able to
see firsthand the tobacco-related norm changes that have taken place in bars
due to the hard work of people like you. Travis is excited be a part of the TCEC team
and looks forward to assisting all of our Tobacco Control partners. For more info on our newest evaluator, check
out the TCEC website!
back to top
|
|
Wading through the Words: Deriving Meaning with Qualitative Analysis
How can you synthesize a lengthy conversation into usable
data? With qualitative information
gathered from key informant interviews, focus groups or some record
reviews, the aim is not to just tally the number of similar responses like with
a survey. Instead, you look for themes
and patterns in order to understand what is meaningful and what is
extraneous. This is what thematic or content analysis consists of. These
patterns and themes tell a story, and this story should be used to inform your program efforts.
So how
do you go about analyzing qualitative data? Typically, you develop
what is called a "coding manual" based on the main topics that are
central to the interview questions or the objective. Once a preliminary list of themes is
created, actual data (quotes from interviews or notes from observations, etc.)
can be coded. Quite often you will have
data relating to something you set out to evaluate as well as data you
were not necessarily expecting to find.
For
instance, if you were evaluating the implementation and enforcement of a
smoke-free park policy, your preliminary list of broad themes might
include: barriers to implementation,
public viewpoints, policy makers' rationale, factors related to smoking in
parks, reasons for successful implementation, and so on. Because of the emergent nature of analyzing
and organizing the data, your code list will evolve over time as you find more sub-themes
or categories in the data.
The
next step is the  coding. There are many
ways to sort and code the data, although the idea remains in the same regardless
of the method used. Some people use
qualitative statistical software programs like ATLAS.ti, The Ethnograph, and
Nud*ist. However, by no means are such
software packages necessary. In fact,
unless you have conducted hundreds of interviews, coding the data can be done
by simply sorting the data into separate Word documents based on the respective
codes, or even by cutting out the quotations and creating separate piles based
on the codes. What is key is that you
read the transcripts, notes or documents several times, allowing the
information to sink in and percolate.
Regardless
of how you decide to code the data, during this stage the data
get separated into categories. For
example, here is an excerpt of a transcript of a county health administrator: "Yeah, last year our city passed a law regarding smoke-free parks. I thought it was a great
law, and much needed. I know for a fact
that one of the children's parks was a frequent hangout for homeless
and vagrant folk, and they all seemed to smoke and drink there. Anyway, the law was passed and went into
effect earlier this year. While I think
the community supports it, I don't think the law really gets enforced. I've gotten phone calls from citizens who
report smoking. So, I write a report and
send it to the local police department, which is supposed to be in charge of
enforcing the law. However, I'm pretty
sure they don't care; or, they certainly don't make it a priority. I never hear from them, and people are still
smoking in some of those parks."
In
coding this paragraph, you might first label it based on the demographics of the
person interviewed. In this case: white,
male, county health administrator, etc.
The demographics provide context and may, in fact, play a major part of
the overall story. Next you would code according to theme. You might code this
entire paragraph under the theme of "Smok  e-free
Park Law" since this was
the topic of the quote. However, this
same quote, or parts of it, may also be placed in other category codes as
well. For example, "homeless," "park
smoking," "children's park," "alcohol," "enforcement," "written report," and/or
"police." Once
all the data are coded, the real analysis begins. In reviewing the coded data, decide whether the
original themes and categories you started with were salient to the story that
has emerged from the data. Some people
make tallies of how often a code or theme might come up in the data, while
others prefer to use of method of weighting based on theory or relevance. How you decide to do this is up to you, but
you'll need to be consistent and transparent by describing the process you used
in the write up of your results. You will
also most likely collapse many of the codes or categories into new broader
themes. This is where it gets tricky for
most newcomers, because there is no set way to do this. But, it's helpful to think of the data as
main points in a story about this issue. What do the
data say that will enable you to communicate this story? Again, this is not easy, and it can be a
painstakingly slow process, even for experts in the field (and why many people
prefer quantitative data collection activities). Nonetheless, the more you become intimately familiar with your data, the more a
pronounced story will emerge. It simply takes
time.
So
how would you write up this data? If a pattern emerged across several
interviews, you could write about how one of the barriers to implementing a successful
smoke-free park law or policy is the lack of enforcement. You could further say that it was suggested
that law enforcement officials do not view this policy or law as a priority to
enforce. Add a few direct quotes in the
write up to document these findings. And
voila, you then have one part of the story!
The
most important part of any story, though, is the ending where the reader is told what it all means. The real heart of your write up of any evaluation activity is where you interpret the findings and make recommendations for
next steps. So in this case, the data
indicated that your project may need to do more education and collaboration
with local law enforcement.
Ultimately, qualitative analysis is all about making sense of the various pieces of data so that your project has a clear picture about what needs to happen next (or what has happened and why).
For more ideas about how to analyze qualitative data, check
out Tips & Tools #7 and Tips & Tools #4 on the TCEC website.
Photos by: Robin Kipke, The Fluid Project
back to top
|
|
|
 Your evaluation technical assistance resource: WE'RE HERE TO HELP!
back to top
|
|
|
|
|