Cannabis Ruderalis

In social science research, social-desirability bias is a type of response bias that is the tendency of survey respondents to answer questions in a manner that will be viewed favorably by others.[1] It can take the form of over-reporting "good behavior" or under-reporting "bad", or undesirable behavior. The tendency poses a serious problem with conducting research with self-reports. This bias interferes with the interpretation of average tendencies as well as individual differences.

Topics subject to social-desirability bias[edit]

Topics where socially desirable responding (SDR) is of special concern are self-reports of abilities, personality, sexual behavior, and drug use. When confronted with the question "How often do you masturbate?," for example, respondents may be pressured by a social taboo against masturbation, and either under-report the frequency or avoid answering the question. Therefore, the mean rates of masturbation derived from self-report surveys are likely to be severely underestimated.

When confronted with the question, "Do you use drugs/illicit substances?" the respondent may be influenced by the fact that controlled substances, including the more commonly used marijuana, are generally illegal. Respondents may feel pressured to deny any drug use or rationalize it, e.g. "I only smoke marijuana when my friends are around." The bias can also influence reports of number of sexual partners. In fact, the bias may operate in opposite directions for different subgroups: Whereas men tend to inflate the numbers, women tend to underestimate theirs. In either case, the mean reports from both groups are likely to be distorted by social desirability bias.

Other topics that are sensitive to social-desirability bias include:

  • Self-reported personality traits will correlate strongly with social desirability bias[2]
  • Personal income and earnings, often inflated when low and deflated when high
  • Feelings of low self-worth and/or powerlessness, often denied
  • Excretory functions, often approached uncomfortably, if discussed at all
  • Compliance with medicinal-dosing schedules, often inflated
  • Family planning, including use of contraceptives and abortion[3][4]
  • Religion, often either avoided or uncomfortably approached[5]
  • Patriotism, either inflated or, if denied, done so with a fear of other party's judgment
  • Bigotry and intolerance, often denied, even if it exists within the responder
  • Intellectual achievements, often inflated
  • Physical appearance, either inflated or deflated
  • Acts of real or imagined physical violence, often denied
  • Indicators of charity or "benevolence," often inflated
  • Illegal acts, often denied
  • Voter turnout[6][7][8][9]

Individual differences in socially desirable responding[edit]

In 1953, Allen L. Edwards introduced the notion of social desirability to psychology, demonstrating the role of social desirability in the measurement of personality traits. He demonstrated that social desirability ratings of personality trait descriptions are very highly correlated with the probability that a subsequent group of people will endorse these trait self-descriptions. In his first demonstration of this pattern, the correlation between one group of college students’ social desirability ratings of a set of traits and the probability that college students in a second group would endorse self-descriptions describing the same traits was so high that it could distort the meaning of the personality traits. In other words, do these self-descriptions describe personality traits or social desirability?[10]

Edwards subsequently developed the first Social Desirability Scale, a set of 39, true-false questions extracted from the Minnesota Multiphasic Personality Inventory (MMPI), questions that judges could, with high agreement, order according to their social desirability.[2] These items were subsequently found to be very highly correlated with a wide range of measurement scales, MMPI personality and diagnostic scales.[11] The SDS is also highly correlated with the Beck Hopelessness Inventory.[12]

The fact that people differ in their tendency to engage in socially desirable responding (SDR) is a special concern to those measuring individual differences with self-reports. Individual differences in SDR make it difficult to distinguish those people with good traits who are responding factually from those distorting their answers in a positive direction.

When SDR cannot be eliminated, researchers may resort to evaluating the tendency and then control for it. A separate SDR measure must be administered together with the primary measure (test or interview) aimed at the subject matter of the research/investigation. The key assumption is that respondents who answer in a socially desirable manner on that scale are also responding desirably to all self-reports throughout the study.

In some cases, the entire questionnaire package from high scoring respondents may simply be discarded. Alternatively, respondents' answers on the primary questionnaires may be statistically adjusted commensurate with their SDR tendencies. For example, this adjustment is performed automatically in the standard scoring of MMPI scales.

The major concern with SDR scales is that they confound style with content. After all, people actually differ in the degree to which they possess desirable traits (e.g. nuns versus criminals). Consequently, measures of social desirability confound true differences with social-desirability bias.

Standard measures of individual SDR[edit]

Until the 1990s, the most commonly used measure of socially desirable responding was the Marlowe–Crowne Social Desirability Scale.[13] The original version comprised 33 True-False items. A shortened version, the Strahan–Gerbasi only comprises ten items, but some have raised questions regarding the reliability of this measure.[14]

In 1991, Delroy L. Paulhus published the Balanced Inventory of Desirable Responding (BIDR): a questionnaire designed to measure two forms of SDR.[15] This forty-item instrument provides separate subscales for "impression management," the tendency to give inflated self-descriptions to an audience; and self-deceptive enhancement, the tendency to give honest but inflated self-descriptions. The commercial version of the BIDR is called the "Paulhus Deception Scales (PDS)."[16]

Scales designed to tap response styles are available in all major languages, including Italian[17][18] and German.[19]

Techniques to reduce social-desirability bias[edit]

Anonymity and confidentiality[edit]

Anonymous survey administration, compared with in-person or phone-based administration, has been shown to elicit higher reporting of items with social-desirability bias.[20] In anonymous survey settings, the subject is assured that their responses will not be linked to them, and they are not asked to divulge sensitive information directly to a surveyor. Anonymity can be established through self-administration of paper surveys returned by envelope, mail, or ballot boxes, or self-administration of electronic survey via computer, smartphone, or tablet.[1][21] Audio-assisted electronic surveys have also been established for low-literacy or non-literate study subjects.[1][22]

Confidentiality can be established in non-anonymous settings by ensuring that only study staff are present and by maintaining data confidentiality after surveys are complete. Including assurances of data confidentiality in surveys has a mixed effect on sensitive-question response; it may either increase response due to increased trust, or decrease response by increasing suspicion and concern.[1]

Specialized questioning techniques[edit]

Several techniques have been established to reduce bias when asking questions sensitive to social desirability.[20] Complex question techniques may reduce social-desirability bias, but may also be confusing or misunderstood by respondents.

Beyond specific techniques, social-desirability bias may be reduced by neutral question and prompt wording.[1]

Ballot Box Method[edit]

The Ballot Box Method (BBM) provides survey respondents anonymity by allowing them to respond in private by self-completing their responses to the sensitive survey questions on a secret ballot and submitting them to a locked box. The interviewer has no knowledge of what is recorded on the secret ballot and does not have access to the lock on the box, providing obscurity to the responses and limiting the potential for SDB. However, a unique control number on each ballot allows the answers to be reunited with a corresponding questionnaire that contains less sensitive questions.[23][24] The BBM has been used successfully to obtain estimates of sensitive sexual behaviours during an HIV prevention study,[23] as well as illegal environmental resource use.[24][25] In a validation study where observed behaviour was matched to reported behaviour using various SDB control methods, the BBM was by far the most accurate bias reduction method, performing significantly better than the Randomized Response Technique (RRT).[24]

Randomized response techniques[edit]

The randomized response technique asks a participant to respond with a fixed answer or to answer truthfully based on the outcome of a random act.[22] For example, respondents secretly throw a coin and respond "yes" if it comes up heads (regardless of their actual response to the question), and are instructed to respond truthfully if it comes up tails. This enables the researcher to estimate the actual prevalence of the given behavior among the study population without needing to know the true state of any one individual respondent. Research shows that the validity of the randomized response technique is limited.[26] Validation research has shown that the RRT actually performs worse than direct questioning for some sensitive behaviours and care should be taken when considering its use.[24]

Nominative and best-friend techniques[edit]

The nominative technique asks a participant about the behavior of their close friends, rather than about their own behavior.[27] Participants are asked how many close friends they know have done for certain a sensitive behavior and how many other people they think know about that behavior. Population estimates of behaviors can be derived from the response.

The similar best-friend methodology asks the participant about the behavior of one best friend.[28]

Unmatched-count technique[edit]

The unmatched-count technique asks respondents to indicate how many of a list of several items they have done or are true for them.[29] Respondents are randomized to receive either a list of non-sensitive items or that same list plus the sensitive item of interest. Differences in the total number of items between the two groups indicate how many of those in the group receiving the sensitive item said yes to it.

Grouped-answer method[edit]

The grouped-answer method, also known as the two-card or three-card method, combines answer choices such that the sensitive response is combined with at least one non-sensitive response option.[30]

Crosswise, triangular, and hidden-sensitivity methods[edit]

These methods ask participants to select one response based on two or more questions, only one of which is sensitive.[31] For example, a participant will be asked whether their birth year is even and whether they have performed an illegal activity; if yes to both or no to both, to select A, and if yes to one but no to the other, select B. By combining sensitive and non-sensitive questions, the participant's response to the sensitive item is masked. Research shows that the validity of the crosswise model is limited.[32]

Bogus pipeline[edit]

Bogus-pipeline techniques are those in which a participant believes that an objective test, like a lie detector, will be used along with survey response, whether or not that test or procedure is actually used.[1]

Other response styles[edit]

"Extreme-response style" (ERS) takes the form of exaggerated-extremity preference, e.g. for '1' or '7' on 7-point scales. Its converse, 'moderacy bias' entails a preference for middle-range (or midpoint) responses (e.g. 3–5 on 7-point scales).

"Acquiescence" (ARS) is the tendency to respond to items with agreement/affirmation independent of their content ("yea"-saying).

These kinds of response styles differ from social-desirability bias in that they are unrelated to the question's content and may be present in both socially neutral and in socially favorable or unfavorable contexts, whereas SDR is, by definition, tied to the latter.

See also[edit]

References[edit]

  1. ^ a b c d e f Krumpal, Ivar (2013). "Determinants of social desirability bias in sensitive surveys: a literature review". Quality & Quantity. 47 (4): 2025–2047. doi:10.1007/s11135-011-9640-9. S2CID 143045969.
  2. ^ a b Edwards, Allen (1957). The social desirability variable in personality assessment and research. New York: The Dryden Press.
  3. ^ Stuart, Gretchen S.; Grimes, David A. (2009). "Social desirability bias in family planning studies: A neglected problem". Contraception. 80 (2): 108–112. doi:10.1016/j.contraception.2009.02.009. PMID 19631784.
  4. ^ Sedgh, Gilda; Keogh, Sarah C. (2019-04-18). "Novel approaches to estimating abortion incidence". Reproductive Health. 16 (1): 44. doi:10.1186/s12978-019-0702-0. PMC 6472065. PMID 30999917.
  5. ^ Presser, Stanley; Stinson, Linda (1998). "Data Collection Mode and Social Desirability Bias in Self-Reported Religious Attendance". American Sociological Review. 63 (1): 137–145. doi:10.2307/2657486. JSTOR 2657486.
  6. ^ Brian, Duff; Hanmer, Michael J.; Park, Won-Ho; White, Ismail K. (2007). "Good Excuses: Understanding Who Votes With An Improved Turnout Question". Public Opinion Quarterly. 71 (1): 67–90. doi:10.1093/poq/nfl045.
  7. ^ Hanmer, Michael J.; Banks, Antoine J.; White, Ismail K. (2013). "Experiments to reduce the over-reporting of voting: A pipeline to the truth". Political Analysis. 22 (1): 130–141. doi:10.1093/pan/mpt027.
  8. ^ Morin-Chassé, Alexandre; Bol, Damien; Stephenson, Laura B.; Labbé St-Vincent, Simon (2017). "How to survey about electoral turnout? The efficacy of the face-saving response items in 19 different contexts" (PDF). Political Science Research and Methods. 5 (3): 575–584. doi:10.1017/psrm.2016.31. S2CID 148277964.
  9. ^ Morin-Chassé, Alexandre (2018). "How to Survey About Electoral Turnout? Additional Evidence". Journal of Experimental Political Science. 5 (3): 230–233. doi:10.1017/XPS.2018.1. S2CID 158608425.
  10. ^ Edwards, Allen (1953). "The relationship between the judged desirability of a trait and the probability that the trait will be endorsed". Journal of Applied Psychology. 37 (2): 90–93. doi:10.1037/h0058073.
  11. ^ Fordyce, William (1956). "Social desirability in the MMPI". Journal of Consulting Psychology. 20 (3): 171–175. doi:10.1037/h0048547. PMID 13357640.
  12. ^ Linehan, Marsha (1981). "Assessment of suicide ideation and parasuicide: Hopelessness and social desirability". Journal of Consulting and Clinical Psychology. 49 (5): 773–775. doi:10.1037/0022-006X.49.5.773. PMID 7287996.
  13. ^ Crowne, Douglas P.; Marlowe, David (1960). "A new scale of social desirability independent of psychopathology". Journal of Consulting Psychology. 24 (4): 349–354. doi:10.1037/h0047358. PMID 13813058. S2CID 9781635.
  14. ^ Thompson, Edmund R.; Phua, Florence T. T. (2005). "Reliability among Senior Managers of the Marlowe–Crowne Short-Form Social Desirability Scale". Journal of Business and Psychology. 19 (4): 541–554. doi:10.1007/s10869-005-4524-4. S2CID 143818289.
  15. ^ Paulhus, D.L. (1991). Measurement and control of response biases. In J.P. Robinson et al. (Eds.), Measures of personality and social psychological attitudes. San Diego: Academic Press
  16. ^ Paulhus D.L., (1998) Paulhus Deception Scales (PDS) is published by Multi-Health Systems of Toronto.
  17. ^ Roccato M., (2003) Desiderabilità Sociale e Acquiescenza. Alcune Trappole delle Inchieste e dei Sondaggi. LED Edizioni Universitarie, Torino. ISBN 88-7916-216-0
  18. ^ Corbetta P., (2003) La ricerca sociale: metodologia e tecniche. Vol. I-IV. Il Mulino, Bologna.
  19. ^ Stöber, Joachim (2001). "The Social Desirability Scale-17 (SDS-17)" (PDF). European Journal of Psychological Assessment. 17 (3): 222–232. doi:10.1027//1015-5759.17.3.222. S2CID 51780539.
  20. ^ a b Nederhof, Anton J. (1985-07-01). "Methods of coping with social desirability bias: A review". European Journal of Social Psychology. 15 (3): 263–280. doi:10.1002/ejsp.2420150303.
  21. ^ McBurney D.H., (1994) Research Methods. Brooks/Cole, Pacific Grove, California.
  22. ^ a b Tourangeau, R.; Yan, T. (2007). "Sensitive questions in surveys". Psychological Bulletin. 133 (5): 859–83. CiteSeerX 10.1.1.563.2414. doi:10.1037/0033-2909.133.5.859. PMID 17723033. S2CID 7160451.
  23. ^ a b Lewis, James JC; Ronsmans, Carine; Ezeh, Alex; Gregson, Simon (June 2004). "The population impact of HIV on fertility in sub-Saharan Africa". AIDS. 18 (Supplement 2): S35–S43. doi:10.1097/00002030-200406002-00005. ISSN 0269-9370. PMID 15319742.
  24. ^ a b c d Bova, Christopher S.; Aswani, Shankar; Farthing, Matthew W.; Potts, Warren M. (2018-12-01). "Limitations of the random response technique and a call to implement the ballot box method for estimating recreational angler compliance using surveys". Fisheries Research. 208: 34–41. doi:10.1016/j.fishres.2018.06.017. ISSN 0165-7836. S2CID 92793552.
  25. ^ Arias, Melissa; Hinsley, Amy; Nogales-Ascarrunz, Paola; Negroes, Nuno; Glikman, Jenny Anne; Milner-Gulland, E. J. (July 2021). "Prevalence and characteristics of illegal jaguar trade in north-western Bolivia". Conservation Science and Practice. 3 (7). Bibcode:2021ConSP...3E.444A. doi:10.1111/csp2.444. ISSN 2578-4854.
  26. ^ John, Leslie K.; Loewenstein, George; Acquisti, Alessandro; Vosgerau, Joachim (September 2018). "When and why randomized response techniques (fail to) elicit the truth". Organizational Behavior and Human Decision Processes. 148: 101–123. doi:10.1016/j.obhdp.2018.07.004. S2CID 52263233.
  27. ^ Miller, J.D. (1985). "The nominative technique: a new method of estimating heroin prevalence" (PDF). NIDA Research Monograph. 54: 104–124. PMID 3929108.
  28. ^ Yeatman, Sara; Trinitapoli, Jenny (2011-09-01). "Best-Friend Reports: A Tool for Measuring the Prevalence of Sensitive Behaviors". American Journal of Public Health. 101 (9): 1666–1667. doi:10.2105/AJPH.2011.300194. PMC 3154247. PMID 21778489.
  29. ^ Droitcour, Judith; Caspar, Rachel A.; Hubbard, Michael L.; Parsley, Teresa L.; Visscher, Wendy; Ezzati, Trena M. (2011), "The Item Count Technique as a Method of Indirect Questioning: A Review of Its Development and a Case Study Application", Measurement Errors in Surveys, John Wiley & Sons, Ltd, pp. 185–210, doi:10.1002/9781118150382.ch11, ISBN 9781118150382
  30. ^ Droitcour, Judith A.; Larson, Eric M. (2016-07-22). "An Innovative Technique for Asking Sensitive Questions: the Three-Card Method". Bulletin of Sociological Methodology/Bulletin de Méthodologie Sociologique. 75: 5–23. doi:10.1177/075910630207500103. S2CID 73189531.
  31. ^ Yu, Jun-Wu; Tian, Guo-Liang; Tang, Man-Lai (2007-04-18). "Two new models for survey sampling with sensitive characteristic: design and analysis". Metrika. 67 (3): 251. doi:10.1007/s00184-007-0131-x. S2CID 122941401.
  32. ^ Schnapp, Patrick (2019). "Sensitive Question Techniques and Careless Responding: Adjusting the Crosswise Model for Random Answers". Methods, Data, Analyses. 13: 307–320. doi:10.12758/mda.2019.03. S2CID 88504613.

Leave a Reply