Archive for the ‘News’ Category
Attention iPad and iPhone users: SurveyMonkey recently launched a mobile app so you can create, send, and monitor your surveys from your phone or tablet. The app is free, although you need a SurveyMonkey account to use it.
With the SurveyMonkey app, you no longer have to rely on your computer to design and manage a survey. The app also allows you to conveniently view your data from any location with Internet access. I think the most notable benefit is that the analytic reports are optimized for mobile devices and are easy to read on small screens.
I have been asked how this app compares to QuickTapSurvey (see my previous blog entry). In my opinion, the app does not make SurveyMonkey comparable to QuickTapSurvey, which is designed specifically to collect onsite visitor feedback in informal settings such as exhibits and museums. SurveyMonkey, by comparison, is designed to collect data through email, web sites, or social media. Both apps work best in their respective settings. I think you could adapt SurveyMonkey to collect data at face-to-face events (if there is onsite Internet access), but it probably won’t work as smoothly as QuickTapSurvey.
For more information about the Survey Monkey mobile app, click here.
Did you know that the American Medical Association has a specific recommendation for its authors about questionnaire response rate? Here it is, from the JAMA Instructions for Authors:
Manuscripts reporting survey data, such as studies involving patients, clinicians, the public, or others, should report data collected as recently as possible, ideally within the past 2 years. Survey studies should have sufficient response rates (generally at least 60%) and appropriate characterization of nonresponders to ensure that nonresponse bias does not threaten the validity of the findings. For most surveys, such as those conducted by telephone, personal interviews (eg, drawn from a sample of households), mail, e-mail, or via the web, authors are encouraged to report the survey outcome rates using standard definitions and metrics, such as those proposed by the American Association for Public Opinion Research.
Meanwhile, response rates to questionnaires have been declining over the past 20 years, as reported by the Pew Research Center in “The Problem of Declining Response Rates.” Why should we care about the AMA’s recommendation regarding questionnaire response rates? Many of us will send questionnaires to health care professionals who, like physicians, are very busy and might not pay attention to our efforts to learn about them. Even JAMA authors such as Johnson and Wislar have pointed out that “60% is only a “rule of thumb” that masks a more complex issue.” (Johnson TP; Wislar JS. “Response Rates and Nonresponse Errors in Surveys.” JAMA, May 2, 2012—Vol 307, No. 17, p.1805) These authors recommend that we evaluate nonresponse bias in order to characterize differences between those who respond and those who don’t. These standard techniques include:
- Conduct a follow-up survey with nonrespondents
- Use data about your sampling frame and study population to compare respondents to nonrespondents
- Compare the sample with other data sources
- Compare early and late respondents
Johnson and Wislar’s article is not open access, unfortunately, but you can find more suggestions about increasing response rates to your questionnaires in two recent AEA365 blog posts that are open access:
Find more useful advice (e.g., make questionnaires short, personalize your mailings, send full reminder packs to nonrespondents) at this open access article: Sahlqvist S, et al., “Effect of questionnaire length, personalisation and reminder type on response rate to a complex postal survey: randomised controlled trial.” BMC Medical Research Methodology 2011, 11:62
The National Network of Libraries of Medicine Outreach Evaluation Resource Center (OERC) offers a range of webinars and workshops upon request by network members and coordinators from the various regions. Take a look at the list and see if one of the options appeals to you. To request a workshop or webinar, contact Susan Barnes.
The workshops were designed as face-to-face learning opportunities but we can tailor them to meet distance learning needs by distilling them to briefer webinars or offering them in series of 1-hour webinars.
Don’t see what you’re looking for on this list? Then please contact Susan and let her know!
We’re looking forward to hearing from you.
The Nielsen Norman Group (NNG) conducts research and publishes information about user experience with interfaces. NNG was an early critic of the troubled “healthcare.gov” web site: “Healthcare.gov’s Account Setup: 10 Broken Usability Guidelines.” recent post (“Talking with participants during a usability test”) provided tips for facilitating usability tests that could be very useful whenever you’re facilitating a discussion or conducting an observation. When in doubt about whether to speak to a participant, count to 10 and decide whether to say something. Consider using “Echo” or “Boomerang” or “Columbo” approaches:
- Echo–repeat the last words or phrase, using an interrogatory tone.
- Boomerang–formulate a nonthreatening question that “pushes” a user’s comment back and causes them to think of a response for you, such as “What would you do if you were on your own?”
- Columbo–be smart but don’t act that way, as in the “Columbo” TV series from the 1960’s and 1970’s starring Peter Falk.
The full article “Talking with participants during a usability test” provides audio examples of these techniques that you can listen to. You can find a large amount of additional information about usability testing on the Nielsen Norman Group’s web site, such as “How to Conduct Usability Studies” and “Usability 101: Introduction to Usability.”
So how are those New Year’s resolutions going?
Many of us like to start the year resolving to clean up some part of our lives. Our diet. Our spending habits. The five years of magazine subscriptions sitting by our recliner.
Here’s another suggestion: Resolve to clean up “chart junk” in the charts you add to PowerPoint presentations or written reports.
Now I can pack information into a bar chart with the best of them. But it is no longer in vogue to clutter charts with data labels, gridlines, and detailed legends. This is not just a fashion statement, either. Design experts point out that charts should make their point without the inclusion of a bunch of distracting details. If the main point of your chart is not visually obvious, you either have not designed it correctly or you are presenting a finding that is not particularly significant.
So the next time you create a chart, consider these suggestions:
- Use your title to communicate the main point of the chart. Take a tip from newspaper headlines and make your title a complete sentence.
- Don’t use three-dimensional displays. It interferes with people’s comprehension of charts.
- Ditch the gridlines or make them faint so they don’t clutter the view.
- Use contrast to make your point. Add a bright color to the bar or line that carries the main point and use gray or another faint color for the comparison bars or lines.
- Be careful in picking colors. Use contrasting colors that are distinguishable to people with colorblindness. If your report is going to be printed, be sure the contrast still shows up when presented in black-and-white.
- Consider not using data labels, or just label the bar or line associated with your main point.
- Remove legends and apply legend labels inside the bars or at the end of lines.
For more comprehensive information on eliminating chart junk, check out this article:
Evergreen S, Mezner C. Design principles for data visualization in evaluation. In Azzam T, Evergreen S. (eds). Data visualization, part 2. New Directions in Evaluation. Winter 2013, 5-20.
In our health information outreach work we are expected to provide evidence of the value of our work, but there are varying definitions of the word “evidence.” The classical evidence-based medicine approach (featuring results from randomized controlled clinical trials) is a model that is not always relevant in our work. At the 2013 EBLIP7 meeting in Saskatoon, Saskatchewan, Canada, Denise Kaufogiannakis presented a keynote address that is now available as an open-access article on the web:
“What We Talk About When We Talk About Evidence” Evidence-Based Library and Information Practice 2013 8.4
This article looks at various interpretations of what it means to provide “evidence” such as
theoretical (ideas, concepts and models to explain how and why something works),
empirical (measuring outcomes and effectiveness via empirical research), and
experiential (people’s experiences with an intervention).
Kaufogiannakis points out that academic librarians’ decisions are usually made in groups of people working together and she proposes a new model for evidence-based library and information practice:
1) Articulate – come to an understanding of the problem and articulate it. Set boundaries and clearly articulate a problem that requires a decision.
2) Assemble – assemble evidence from multiple sources that are most appropriate to the problem at hand. Gather evidence from appropriate sources.
3) Assess – place the evidence against all components of the wider overarching problem. Assess the evidence for its quantity and quality. Evaluate and weigh evidence sources. Determine what the evidence says as a whole.
4) Agree – determine the best way forward and if working with a group, try to achieve consensus based on the evidence and organizational goals. Determine a course of action and begin implementation of the decision.
5) Adapt – revisit goals and needs. Reflect on the success of the implementation. Evaluate the decision and how it has worked in practice. Reflect on your role and actions. Discuss the situation
with others and determine any changes required.
Kaufogiannakis concludes by reminding us that “Ultimately, evidence, in its many forms, helps us find answers. However, we can’t just accept evidence at face value. We need to better understand evidence – otherwise we don’t really know what ‘proof’ the various pieces of evidence provide.”
Do you want to learn about how your user groups and communities find and use information? Do you want to gather evidence to demonstrate that your work is making a difference?
Exciting news! You can work on these questions, and questions like them, June 16-26, 2014!
The Institute for Research Design in Librarianship is a great opportunity for an academic librarian who is interested in conducting research. Research and evaluation are not necessarily identical, although they do employ many of the same methods and are closely related. This Institute is open to academic librarians from all over the country. If your proposal is accepted, your attendance at the Institute will be paid for, as will your travel, lodging, and food expenses.
The William H. Hannon Library has received a three-year grant from the Institute for Museum and Library Services (IMLS) to offer a nine-day continuing education opportunity for academic and research librarians. Each year 21 librarians will receive instruction in research design and a full year of support to complete a research project at their home institutions. The summer Institute for Research Design in Librarianship (IRDL) is supplemented with pre-institute learning activities and a personal learning network that provides ongoing mentoring. The institutes will be held on the campus of Loyola Marymount University in Los Angeles, California.
The Institute is particularly interested in applicants who have identified a real-world research question and/or opportunity. It is intended to
“bring together a diverse group of academic and research librarians who are motivated and enthusiastic about conducting research but need additional training and/or other support to perform the steps successfully. The institute is designed around the components of the research process, with special focus given to areas that our 2010 national survey of academic librarians identified as the most troublesome; the co-investigators on this project conducted the survey to provide a snapshot of the current state of academic librarian confidence in conducting research. During the nine-day institute held annually in June, participants will receive expert instruction on research design and small-group and one-on-one assistance in writing and/or revising their own draft research proposal. In the following academic year, participants will receive ongoing support in conducting their research and preparing the results for dissemination.”
Your proposal is due by February 1, 2014. Details are available at the Institute’s Prepare Your Proposal web site.
Factoid: Loyola Marymount is on a bluff above the Pacific Ocean, west of central LA.
The 2nd Edition of the Planning and Evaluating Health Information Outreach Projects series of 3 booklets is now available online:
Getting Started with Community-Based Outreach (Booklet 1)
What’s new? More emphasis and background on the value of health information outreach, including its relationship to the Healthy People 2020 Health Communication and Health Information Technology topic areas
Planning Outcomes-Based Outreach Projects (Booklet 2)
What’s new? Focus on uses of the logic model planning tool beyond project planning, such as providing approaches to writing proposals and reports.
Collecting and Analyzing Evaluation Data (Booklet 3)
What’s new? Step-by-step guide to collecting, analyzing, and assessing the validity (or trustworthiness) of quantitative and qualitative data, using questionnaires and interviews as examples.
These are all available free to NN/LM regional offices and network members. To request printed copies, send an email to firstname.lastname@example.org.
Non-508 compliant pdf versions of all three booklets are available here: http://nnlm.gov/evaluation/guides.html#A2 .
The Planning and Evaluating Health Information Outreach series, by Cynthia Olney and Susan Barnes, supplements and summarizes material in Cathy Burroughs’ groundbreaking work from 2000, Measuring the Difference: Guide to Planning and Evaluating Health Information Outreach. Printed copies of Burroughs’ book are also available free—just send an email request to email@example.com.
An interesting study by clinicians of how clinicians use information resources has appeared in a recent issue of JAMA Internal Medicine:
Cook DA; Sorensen KJ; Wilkinson JM; and Berger RA. “Barriers and decisions when answering clinical questions at the point of care: A grounded theory study.” JAMA Intern Med, published online August 26, 2013. [epub ahead of print PMID: 23979118] This article provides details about steps that the researchers took in their qualitative study of how 50 primary care and subspecialist internal medicine and family medicine physicians use online information resources (such as UpToDate, MD Consult, Micromedex, and publicly available Internet resources) to answer clinical questions at the point of care. You can find details here about how the focus groups were conducted, how the participants were selected, and how data was collected and analyzed. This article provides a great template for an approach to collecting qualitative information via focus groups, and ends with an unsurprising conclusion:
“Physicians perceive that insufficient time is the greatest barrier to point-of-care learning, and efficiency is the most important determinant in selecting an information source.”
A picture’s worth a thousand words, and a method called photovoice takes advantage of pictures’ compelling qualities by incorporating photography into research and evaluation. Photovoice is a participatory evaluation method in which program participants are given cameras to capture images that convey their feelings, beliefs and experiences about an issue. The method is used frequently in advocacy projects, allowing the less powerful stakeholders to communicate about issues that impact their lives.
Photovoice seems to be a particularly popular way to engage youth in projects or in evaluation. For examples of photovoice projects with teenagers, check out the two articles listed at the end of this blog entry. The project described in Necheles et al. used photovoice to engage teenagers in identifying influences over their own health behavior. These teens then developed materials such as posters to advocate for healthier lifestyles among their peers. The article by Strack, Magill and McDonagh presents a project in which teens identified problems in their neighborhoods through photovoice. Both articles provide abundant advice for conducting photovoice projects, including how to engage youth in analyzing photos and ideas for presenting results.
Some photovoice projects carry potential risk for participants. Participants also must be taught how to get and document consent from others who appear in their photos. Consequently, photovoice projects require above-average planning and management. For an excellent resource on managing photovoice projects, check out photovoice.org.
Necheles JW et al. The Teen Photovoice Project: A pilot study to promote Health through Advocacy. Prog Community Health Partnersh 2007 Fall; 1(3): 221–229. (available at PubMedCentral, http://www.ncbi.nlm.nih.gov/pmc/articles/PMC2837515/)
Strack RW, Magill C, McDonagh K. Engaging youth through photovoice. Health Promot Pract 2004;5:49–58. Available at http://www.ncbi.nlm.nih.gov/pubmed/14965435
Photovoice.org, particularly the organization’s methodology section at http://www.photovoice.org/shop/info/methodology-series