Category Archives: Data collection

Testing the Test (or an intro to “Does the measurement measure up?”)

When reading a research article, you may be tempted only to read the Introduction & Background, then go straight to the Discussion, Implications, and Conclusions at the end. You skip all those pesky, procedures, numbers, and p levels in the Methods & Results sections.

Perhaps you are intimidated by all those “research-y” words like content validity, construct validity, test-retest reliability, and Cronbach’s alpha because they just aren’t part of your vocabulary….YET!

WHY should you care about those terms, you ask? Well…let’s start with an example. If your bathroom scale erratically measured your weight each a.m., you probably would toss it and find a more reliable and valid bathroom scale. The quality of the data from that old bathroom scale would be useless in learning how much you weighed. Similarly in research, the researcher wants useful outcome data. And to get that quality data the person must collect it with a measurement instrument that consistently (reliably) measures what it claims to measure (validity). A good research instrument is reliable and valid. So is a good bathroom scale.

Let’s start super-basic: Researchers collect data to answer their research question using an instrument. That test or tool might be a written questionnaire, interview questions, an EKG machine, an observation checklist, or something else. And whatever instrument the researcher uses needs to give them correct data answers.

For example, if I want to collect BP data to find out how a new med is working, I need a BP cuff that collects systolic and diastolic BP without a lot of artifacts or interference. That accuracy in measuring BP only is called instrument validity. Then if I take your BP 3 times in a row, I should get basically the same answer and that consistency is called instrument reliability. I must also use the cuff as intended–correct cuff size and placement–in order to get quality data that reflects the subject’s actual BP.

The same thing is true with questionnaires or other measurement tools. A researcher must use an instrument for the intended purpose and in the correct way. For example, a good stress scale should give me accurate data about a person’s stress level (not their pain, depression, or anxiety)–in other words it should have instrument validity. It should measure stress without a lot of artifacts or interference from other states of mind.

NO instrument is 100% valid–it’s a matter of degree. To the extent that a stress scale measures stress, it is valid. To the extent that it also measures other things besides stress–and it will–it is less valid. The question you should ask is, “How valid is the instrument?” often on a 0 to 1 scale with 1 being unachievable perfection. The same issue and question applies to reliability.

Reliability & validity are interdependent. An instrument that yields inconsistent results under the same circumstances cannot be valid (accurate). Or, an instrument can consistently (reliably) measure the wrong thing–that is, it can measure something other than what the researcher intended to measure. Research instruments need both strong reliability AND validity to be most useful; they need to measure the outcome variable of interest consistently.

Valid for a specific purpose: Researchers must also use measurement instruments as intended. First, instruments are often validated for use with a particular population; they may not be valid for measuring the same variable in other populations. For example, different cultures, genders, professions, and ages may respond differently to the same question. Second, instruments may be valid in predicting certain outcomes (e.g., SAT & ACT have higher validity in predicting NCLEX success than does GPA). As Sullivan (2011) wrote: “Determining validity can be viewed as constructing an evidence-based argument regarding how well a tool measures what it is supposed to do. Evidence can be assembled to support, or not support, a specific use of the assessment tool.”

In summary….

  1. Instrument validity = how accurate the tool is in measuring a particular variable
  2. Instrument reliability = how consistently the tool measures whatever it measures

Fun Practice: In your own words relate the following article excerpt to the concept of validity? “To assess content validity [of the Moral Distress Scale], 10 nurses were asked to provide comments on grammar, use of appropriate words, proper placement of phrases, and appropriate scoring. From p.3, Ghafouri et al. (2021). Psychometrics of the moral distress scale in Iranian mental health nurses. BMC Nursing. https://doi.org/10.1186/s12912-021-00674-4

Pilot sTUdies: Look before you leap! (a priori vs. posthoc)

Why does it matter if a study is labeled a “pilot”?

SHORT ANSWER: …Because a pilot is about testing research methods,….not about answering research questions.

If a project has “pilot” in the title, then you as a reader should expect a study that examines whether certain research methods work (methodologic research). Methods include things like timing of data collection, sampling strategies, length of questionnaire, and so on. Pilots suggest what methods will effectively to answer researchers’ questions. Advance prep in methods makes for a smooth research landing.

Small sample = Pilot? A PILOT is related to study goals and design–not sample size. Of course pilots typically have small samples, but a small sample does not a pilot study make. Sometimes journals may tempt a researcher to call their study a pilot because of small samples. Don’t go there. Doing so means after-the-fact, posthoc changes that were Not the original, a priori goals and design.

Practical problems? If researchers label a study a “pilot” after it is completed (post hoc), they raise practical & ethical issues. At a practical level, researchers must create feasibility questions & answers. (See NIH.) The authors should drop data analysis that answers their original research questions.

Ethics? This ethically requires researchers 1) to say they planned something that they didn’t or 2) to take additional action. Additional action may be complete transparency about the change and seeking modification to original human subjects’ committee approvals. An example of one human subjects issue is that you informed your subjects that their data would answer a particular research question, and now you want to use their data to answer something else–methods questions!

Options? You can just learn from your small study and go for a bigger one, including improving methods. Some journals will consider publication of innovative studies even when small.

Look first, then leap: Better to look a priori, before leaping. If you think you might have trouble with your methods, design a pilot. If you made the unpleasant discovery that your methods didn’t work as you hoped, you can 1) disseminate your results anyway or 2) rethink ethical and practical issues.

Who’s with me? The National Institutes of Health agree: https://nccih.nih.gov/grants/whatnccihfunds/pilot_studies . NIH notes that common misuses of “pilots” are determining safety, efficacy of intervention, and effect size.

Who disagrees? McGrath argues that clinical pilots MAY test safety and efficacy, as well as feasibility. (See McGrath, J. M. (2013). Not all studies with small samples are pilot studies, Journal of Perinatal & Neonatal Nursing, 27(4): 281-283. doi: 10.1097/01.JPN.0000437186.01731.bc )

Trial Balloons & Pilot Studies

A pilot study is to research what a trial balloon is to politics

In politics, a trial balloon is communicating a law or policy idea via media to see how the intended audience reacts to it.  A trial balloon does not answer the question, “Would this policy (or law) work?” Instead a trial balloon answers questions like “Which people hate the idea of the policy/law–even if it would work?” or “What problems might enacting it create?” In other words, a trial balloon answers questions that a politician wants to know BEFORE implementing a policy so that the policy or law can be tweaked to be successfully put in place.

meeting2

In research, a pilot study is sort of like a trial balloon. It is “a small-scale test of the methods and procedures” of a planned full-scale study (Porta, Dictionary of Epidemiology, 5th edition, 2008). A pilot study answers questions that we want to know BEFORE doing a larger study, so that we can tweak the study plan and have a successful full-scale research project. A pilot study does NOT answer research questions or hypotheses, such as “Does this intervention work?”  Instead a pilot study answers the question “Are these research procedures workable?”

A pilot study asks & answers:Can I recruit my target population? Can the treatments be delivered per protocol? Are study conditions acceptable to participants?” and so on.   A pilot study should have specific measurable benchmarks for feasibility testing. For example if the pilot is finding out whether subjects will adhere to the study, then adherence might be defined as  “70 percent of participants in each [group] will attend at least 8 of 12 scheduled group sessions.”  Sample size is based on practical criteria such as  budget, participant flow, and the number needed to answer feasibility questions (ie. questions about whether the study is workable).

A pilot study does NOT Test hypotheses (even preliminarily); Use inferential statistics; Assess safety of a treatment; Estimate effect size; Demonstrate safety of an intervention.

A pilot study is not just a small study.

Next blog: Why this matters!!

For more info read the source of all quotes in this blog: Pilot Studies: Common Uses and Misuses @ https://nccih.nih.gov/grants/whatnccihfunds/pilot_studies

Is History “Bunk”? We report. You Decide.

History?  Really?  Fascinating!  Ever thought about all the stories behind your own present life?

Check out this youtube dramatized documentary about Nurse Mary Seacole.  I promise – you’ll enjoy: https://www.youtube.com/watch?v=RIrim4r-LbY   

You can be a part of documenting such stories, including your own.  Can I pique your interest with these examples about historical research?

1. Artifacts: Example = http://acif.org/ The American Collectors of Infant Feeders:

Infant feeder
CREDIT http://acif.org/

The American Collectors of Infant Feeders is a non-profit organization whose primary purpose is to gather and publish information pertaining to the feeding of infants throughout history. The collecting of infant feeders and related items is promoted.

2. Interviews: Example = http://www.oralhistory.org/  Want to do interviews of interesting faculty, students, leaders, “ordinary” nurses?  Check out the Oral History Association    In addition to fostering communication among its members, the OHA encourages standards of excellence in the collection, preservation, dissemination and uses of oral testimony.

scrapbook
CREDIT https://archives.mc.duke.edu/blog/nursing-materials-displa

3. Stories from the “ordinary: Example: http://www.murphsplace.com/mother/main.html My Mother’s War – “Helen T.Burrey was an American nurse who served as a Red Cross Nurse during World War I. She documented her experience in both a journal and a scrapbook which has been treasured by her daughter, Mary Murphy. Ms Murphy has placed many of these items on the Internet for people to access and it provides a first-hand account of that experience. Additionally she has a variety of links to other WWI resources.” (quoted from AAHN Resources online)

Army history
CREDIT http://e-anca.org/

4. Ethnic studies: Example=https://libguides.rowan.edu/blacknurses  Black Nurses in History “This is a ‘bibliography and guide to web resources’ from the UMDNJ and Coriell Research Library. Included are Mamie O. Hail, Mary Eliza Mahoney, Jessie Sleet Scales, Mary Seacole, Mabel Keaton Staupers, Susie King Taylor, Sojourner Truth, Harriet Tubman.” (quoted from AAHN Resources online)

Want more?  

Critical thinking:  Don’t forget to save your own materials.  Your life is history!  What in your life is most interesting?  Have you written it down or dictated it into your iphone voice memo? There is GREAT interest in “ordinary” men and women.  Many times items are tossed because they are “just letters” or “only old records,” or “stuff.” Just Don’t Do It.

 

“Please answer….” (cont.)

What do people HATE about online surveys?   If you want to improve your response rates, check out SurveyMonkey Eric V’s (May Mail2017)  Eliminate survey fatigue: Fix 3 things your respondents hate 

For more info: Check out my earlier post “Please Answer!”

Nightingale: Avant garde in meaningful data

In honor of Nurse Week, I offer this tribute to the avant garde research work of Florence Nightingale in the Crimea that saved lives and set a precedent worth following.

Nightingale was a “passionate statistician” knowing that outcome data are convincing when one wants to change the world.  She did not merely collect the data, but also documented it in a way that revealed its critical meaning for care.

As noted by John H. Lienhard (1998-2002): Nightingale coxcombchart“Once you see Nightingale’s graph, the terrible picture is clear. The Russians were a minor enemy. The real enemies were cholera, typhus, and dysentery. Once the military looked at that eloquent graph, the modern army hospital system was inevitable.  You and I are shown graphs every day. Some are honest; many are misleading….So you and I could use a Florence Nightingale today, as we drown in more undifferentiated data than anyone could’ve imagined during the Crimean War.” (Source: Leinhard, 1998-2002)

As McDonald (2001) writes in the BMJ free, full-text,  Nightingale was “a systemic thinker and a “passionate statistician.”  She insisted on improving care by making policy & care decisions based on “the best available government statistics and expertise, and the collection of new material where the existing stock was inadequate.”(p.68)

Moreover, her display of the data brought its message home through visual clarity!

Thus while Nightingale adhered to some well-accepted, but mistaken, scientific theories of the time (e.g., miasma) her work was superb and scientific in the best sense of the word.   We could all learn from Florence.

CRITICAL THINKING:   What issue in your own practice could be solved by more data?  How could you collect that data?   If you have data already, how can you display it so that it it meaningful to others and “brings the point home”?

FOR MORE INFO:

HAPPY NURSE WEEK TO ALL MY COLLEAGUES.  

MAY YOU GO WHERE THE DATA TAKES YOU!

Listen up! Don’t interrupt!

Researchers collect two types of data in their studiescounting-sheetword-art

  1. Numbers (called quantitative data)
  2. Words & narratives (called qualitative data)

StorytellerOne source of rich word or narrative (qualitative) data for answering nursing questions is nurses’ stories.  Dr. Pat Benner RN, author of Novice to Expert explains two things we can do to help nurses fully tell their stories so we can learn the most from their practice.

  1. Listen well without interrupting
  2. Help nurses ‘unpack’ their stories 

Check out this excellent 2:59 video of Dr. Benner’s and revolutionize how you learn about nursing from nursing stories:  Preview: The use of Narratives 

Critical thinking:  For a study using narratives in research see  Leboul et al. (2017).  Palliative sedation challenging the professional competency of health care providers and staff: A qualitative focus group and personal written narrative study.  [full text available thru PubMed at https://www.ncbi.nlm.nih.gov/pubmed/28399846].    1) Do you think the authors listened and unpacked information from the focus groups & written narratives; 2)  Do you think there might be a difference in the way people write narratives and verbally tell narratives?   3) How might that difference if any affect the research findings?

For more information:  Check out The Power of Story  by Wang & Geale (2015) at http://www.sciencedirect.com/science/article/pii/S2352013215000496

 

Direct speaking about INdirect outcomes: HCAHPS as a measurement

When you first plan a project, you need to know what OUTCOMES you want to achieve.  You need STRONG outcomes to show your project worked! imagesCALQ0QK9

Outcome measures are tricky & can be categorized into Indirect & Direct measures:

  1. INDIRECT outcome measures are often affected by many factors, not just your innovation
  2. DIRECT outcome measures are specific to what you are trying to accomplish.

For example: If you want to know your patient’s weight, you put them on the scale (direct). weight-scaleYou don’t merely ask them how much they weigh (indirect).

Another example?  If you planned music to reduce pain, you might a) measure how many patients were already using music and their pain scores (& perhaps those not using music and their pain scores), b) begin your music intervention, and c) thmusicen directly measure how many patients started using it after you started your intervention and their pain scores.  These data DIRECTLY target your inpatient outcomes versus looking at INDIRECT HCAHPS answers of discharged patients’ feelings after the fact in response to “During this hospital stay, how often was your pain well controlled?”

Nurses often decide to measure their project outcomes ONLY with indirect HCAHPS scores.  I hope you can see this is not as good as DIRECT measures.

So why use HCAHPS at all?measuring-tape

  • They reflect institutional priorities related to quality and reimbursement
  • Data are already collected for you
  • Data are available for BEFORE and AFTER comparisons of your project outcomes
  • It doesn’t cost you any additional time or money to get the data

Disadvantages of indirect HCAHPS measures?

  • HCAHPS data are indirect measures that are affected by lots of different things, and so they may have little to do with effect of your project.
  • HCAHPS responders often do Not represent all patients because the number responding is so small–sometimes just 1 or 2

Still, I think it’s good to include HCAHPS.  Just don’t limit yourself to that. Include also a DIRECT measure of outcomethat targets the precisely what you hope will be the result of your study.

imagesCALQ0QK9You need STRONG outcomes to convince others that your project works to improve care!

CRITICAL THINKING:  McClelland, L.E., &  Vogus, T.J. (2014) used HCHAPS as an outcome measure in their study, Compassion practices & HCAHPS: Does rewarding and supporting questionworkplace compassion influence patient perceptions?    What were the strengths & weaknesses of using HCHAPS in this study? [hint: check out the discussion section]  What would be a good direct measure that you could add to HCAHPS outcomes to improve the study?

FOR MORE INFORMATION:  Whole books of measurement instruments are available through the library or a librarian can help you search for something that will measure motivation, pain, anxiety, medication compliance, or whatever it is you are looking for!!  You can limit your own literature searches by selecting “instrument” as part of your search, or you can consult with a nurse researcher for more help.

DATA COLLECTION SECTION! (Methods in the Madness)

Key point! The data collection section of a research article includes: who collects what data when, where & how.

In previous blogs we’ve looked at title, introduction, and other elements of methods section (design, sample, & setting). In this one let’s take a look at data collection.

Data are a collection of measurements. For example, student scores on a classroom test might be 97, 90, 88, 85, & so on. Each single score is a datum; collectively they are data.

What data are collected is answered in this section. The data (or measurements) can be counting-hashmarksnumbers OR words. For example, numbers data might include patient ratings of their pain on a 0-10 scale. An example of word data would asking participants to describe something in words without counting the words or anything else.  For example, word data might include patient descriptions pain in words, like word-art“stabbing,”  “achy,” and so on.  Sometimes a researcher collects both number and word data in the same study to give a more complete description.  You can see how knowing the patient’s pain rating and hearing a description would give you a much clearer picture of pain.

  • Studies reporting data in numbers are called quantitative studies
  • Studies reporting data in words/descriptions are called qualitative studies
  • Studies reporting number & word data are called mixed methods studies

How the data are collected includes what instrument or tool was used to gather data (e.g., observation, biophysical measure, or self-report) and how consistently & accurately that tool measures what it is supposed to measure (e.g., reliability & validity). Also included is who collected the data and the procedures that they followed—how did they obtain consent, interaction with subjects, timing of data collection and so on.

Now you know!

Critical thinking question: Did these authors use qualitative or quantitative data collection methods?  Coelho, A., Parola, V., Escobar-Bravo, M., & Apostolo, J. (2016). Comfort experience in palliative care, BMD Palliative care, 15(71). doi: 10.1186/s12904-016-0145-0.  Explain your answer.

Self-Report Data: “To use or not to use. That is the question.”

[Note: The following was inspired by and benefited from Rob Hoskin’s post at http://www.sciencebrainwaves.com/the-dangers-of-self-report/]Penguins

If you want to know what someone thinks or feels, you ask them, right?

The same is true in research, but it is good to know the pros and cons of using the “self-report method” of collecting data in order to answer a research question.  Most often self-report is done in ‘paper & pencil’ or SurveyMonkey form, but it can be done by interview.

Generally self-report is easy and inexpensive, and sometimes facilitates research that might otherwise be impossible.  To answer well, respondents must be honest, have insight into themselves, and understand the questions.  Self-report is an important tool in much behavioral research.

But, using self-report to answer a research question does have its limits. People may tend to answer in ways that make themselves look good (social desirability bias), agree with whatever is presented (social acquiescence bias), or answer in either extreme terms (extreme response set bias) or always pick the non-commital middle Hypothesisnumbers.  Another problem will occur if the reliability  and validity of the self-report questionnaire is not established.  (Reliability is consistency in measurement and validity is the accuracy of measuring what it purports to measure.) Additionally, self-reports typically provide only a)ordinal level data, such as on a 1-to-5 scale, b) nominal data, such as on a yes/no scale, or c) qualitative descriptions in words without categories or numbers.  (Ordinal data=scores are in order with some numbers higher than others, and nominal data = categories. Statistical calculations are limited for both and not possible for qualitative data unless the researcher counts themes or words that recur.)

Gold_BarsAn example of a self-report measure that we regard as a gold standard for clinical and research data = 0-10 pain scale score.   An example of a self-report measure that might be useful but less preferred is a self-assessment of knowledge (e.g., How strong on a 1-5 scale is your knowledge of arterial blood gas interpretation?)  The use of it for knowledge can be okay as long as everyone understands that it is perceived level of knowledge.

Critical Thinking: What was the research question in this study? Malaria et al. (2016) Pain assessment in elderly with behavioral and psychological symptoms of dementia. Journal of Alzheimer’s Disease as posted on PubMed.gov questionat http://www.ncbi.nlm.nih.gov/pubmed/26757042 with link to full text.  How did the authors use self-report to answer their research question?  Do you see any of the above strengths & weaknesses in their use?

For more information: Be sure to check out Rob Hoskins blog: http://www.sciencebrainwaves.com/the-dangers-of-self-report/