13 Sep 2013: (2013) Correction: Ineffectiveness of Reverse Wording of Questionnaire Items: Let’s Learn from Cows in the Rain. PLoS ONE 8(9): 10.1371/annotation/af78b324-7b44-4f89-b932-e851fe04a8e5. doi: 10.1371/annotation/af78b324-7b44-4f89-b932-e851fe04a8e5 | View correction
We examined the effectiveness of reverse worded items as a means of reducing or preventing response bias. We first distinguished between several types of response bias that are often confused in literature. We next developed arguments why reversing items is probably never a good way to address response bias. We proposed testing whether reverse wording affects response bias with item-level data from the Multidimensional Fatigue Inventory (MFI-20), an instrument that contains reversed worded items.
With data from 700 respondents, we compared scores on items that were similar with respect either to content or to direction of wording. Psychometric properties of sets of these items worded in the same direction were compared with sets consisting of both straightforward and reversed worded items.
We did not find evidence that ten reverse-worded items prevented response bias. Instead, the data suggest scores were contaminated by respondent inattention and confusion.
Using twenty items, balanced for scoring direction, to assess fatigue did not prevent respondents from inattentive or acquiescent answering. Rather, fewer mistakes are made with a 10-item instrument with items posed in the same direction. Such a format is preferable for both epidemiological and clinical studies.
Citation: Sonderen Ev, Sanderman R, Coyne JC (2013) Ineffectiveness of Reverse Wording of Questionnaire Items: Let’s Learn from Cows in the Rain. PLoS ONE 8(7): e68967. doi:10.1371/journal.pone.0068967
Editor: Hamid Reza Baradaran, Iran University of Medical Sciences, Iran (Republic of Islamic)
Received: November 28, 2012; Accepted: June 9, 2013; Published: July 31, 2013
Copyright: © 2013 Sonderen et al. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Funding: The work was done as part of a regular appointment at the University Medical Center Groningen. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have read the journal's policy and have the following conflicts: The third author, James Coyne, is a PLOS ONE Academic Editor. This does not alter the authors' adherence to all the PLOS ONE policies on sharing data and materials.
The use of both positively and negatively worded items in questionnaires was introduced decades ago in order to prevent response bias. Response bias refers to answer patterns on questionnaires that do not reflect the respondents’ actual state or opinion , and that thus can pose a serious threat to the validity of self-report instruments .
We have three goals with this paper. First, we want to distinguish among several types of response bias that are confused in literature. Next, we develop arguments why reversing items is probably almost never a good way to prevent or deal with response bias. Finally, as an example, we tested reverse wording of items with data from the Multidimensional Fatigue Inventory (MFI-20), an instrument on which ten of the items are reverse worded.
The process of obtaining survey data is complex, with many possibilities of discrepancies arising between the state or opinion the researcher wants to elicit and the answer given by the respondent . Usually this discrepancy is called response bias . Based on Rorer , Weijters  distinguishes two main types of response bias: response set and response style. He defines response set as bias related to the content of the items and response style as a tendency to answer items regardless their content. The best known type of response set is social desirability, in which a person’s response is a function of the desirability of the response rather than its veracity. Three types of response styles can be distinguished.
Respondents may have read and understood completely the question and answer categories, but nonetheless be inclined to agree with statements in general (acquiescence), to disagree (disacquiescence), or to give extreme answers, or, alternatively, less extreme answers. Consistent with Swain et al. , we will take acquiescence as an example of this type of response styles.
Respondents may also lack sufficient attention to carefully read both the question and answer categories, and thus, by missing the intended meaning of an item, give a response that may differ from the true value. Krosnick  mentions a satisficing response style, whereby the respondent deliberately makes less effort to understand all subtleties of the question. We will call this style inattention, irrespective whether the respondent is aware of it or not.
Finally, the question in combination with answer categories may be too difficult for a respondent to comprehend. Sometimes the respondent may think the item is well understood, but still an error can be made, due to a high level of difficulty . The respondent may however be aware of this difficulty, and thus the answer can be considered a ‘best guess’. We will call both varieties of this type confusion.
According to Weijters et al.  we focus on response style in discussing the consequences of reversed worded items. We consider different types of response set, like social desirability, to be less sensitive for reversed worded items.
It is the challenge for a researcher to deal with all these threats and still obtain an optimal answer, that is, one that consistently resembles the true value.
Reversing a portion of the items is often intended to reduce the effects of response styles, although there is no consensus that this is an effective strategy. In general, two strategies are available for item reversal . The first consists of adding negative particles: words like ‘not’ or ‘no’ or affixal morphemes like ‘un-‘, ‘non-‘, ‘dis-‘ or ‘-less’. In this case the direction of the item is changed without changing substantially item wording. The new item is considered to be ‘reverse oriented’. The second strategy is using words with an opposite meaning. For example, the opposite of “I feel fit” is “I feel fatigue”. In this case the direction of the new item is changed by means of ‘reverse wording’. Swain et al.  analyzed nearly 2000 items and found that 81% of the reversed items were negations, i.e. items created by the first strategy.
We will now discuss the consequences of item reversal in the light of the three response styles we have introduced, acquiescence, inattention, and confusion.
Reversal and Acquiescence
Although many researchers advocate the use of reversed items in order to address acquiescence, it is doubtful whether this response style will be affected by the direction in which the items are formulated. Considering acquiescent respondents as people who carefully read each question, they will, when confronted with a reversed item, still agree, and thus leave the researcher with an uninterpretable patterning of answers (someone who is both ‘tired’ and ‘fit’). If a respondent agrees with an ordinary item and disagrees with a reversed worded item, according to the true state, then by definition this person is not an acquiescent respondent from whom we expect only affirmative answers on all items.
Reversal and Inattention
Several types of inattention can be distinguished. First, a respondent may miss the presence of a negative particle or an affixal morphem, for example s/he may read ‘I am healthy’ instead of ‘I am not healthy’ or ‘I am unhealthy’. This type of inattention relates to any individual item. A respondent may also miss the fact that a consecutive item is formulated in a reversed way, compared to the previous one. Finally, a respondent can miss the difference in content between two consecutive items.
As Drolet and Morrison  demonstrate, respondents sometimes just answer the first item and assume that this answer also holds for subsequent (considered similar or even identical) items. Sometimes respondents do not bother to endorse an answer for each item individually, but draw one large circle around the same response for all items. The risk for this type of inattention grows with the extent to which items resemble each other, and when scales are longer. Only the second type of inattention, missing subtle differences with respect to content, can be challenged by reversing some items, provided that respondents are not inattentive to the reversal, in which case reversing will be counterproductive.
Reversal and Confusion
The last response style that has to be considered when reversing items, is confusion. As Swain et al.  demonstrated, item verification difficulty, that is respondents’ difficulty interpreting items, increases when reverse oriented items rely on negative particles or affixal morphemes. Modifying an item by reverse wording, and thus inquiring about the opposite state, will only lead to more difficulties in interpretation, if the described state is not in accordance with the respondents actual state. For example, for a tired person, the item ‘I am tired’ is easy, and ‘I am fit’ is a bit more difficult since this item has to be denied. For a fit person however, the opposite holds.
Demonstrating or Preventing Response Styles
With these arguments in mind it does not seem advisable in scale construction to reverse a portion of the items. Schriesheim and Hill  previously concluded that reversed worded items, when used in an effort to control for acquiescence, lowered questionnaires’ validity. Yet, many instrument developers persist in adhering to this strategy. Usually their intention is to assess a one-dimensional construct and reversing some items is seen as limiting the influence of response styles, especially acquiescence. Yet, the unintended consequence is the emergence of two factors in subsequent factor analyses, commonly precipitating a debate in the literature whether or not two meaningful concepts can be distinguished where only one was intended. Eventually papers are published concluding that the second factor is just a methodological artifact, caused by the use of reversed worded items. We mention some examples:
- Meyer and Allen’s  scale to assess affective commitment, of which Merritt  demonstrated that the answers on reversed items were prone to careless responding and cognitive fatigue.
- Roszkowski and Soven  examined a questionnaire used in student evaluations, that contained two reversed worded items among mainly positively worded items, and concluded that replacing the reversed worded items by ordinary items improved the internal consistency
- Meyer et al.’s  scale to assess worry contains 11 positively and five reversed worded items, leading Fresco et al.  to the conclusion of two distinct concepts, Worry Engagement and Absence of Worry. Hazlett-Stevens et al.  however conclude that the reversed worded items caused a method factor.
- Bradley et al.  inspected the behavior of ten items from a scale used to ask students to evaluate courses and instructors. This questionnaire contains five pairs of items that were worded in opposite directions (e.g. “the professor was unprepared for class” versus “the professor was prepared for class”). They concluded that the use of opposite items introduced noise, known as measurement error.
Summarizing, we see no convincing arguments that reversing part of the items will prevent response styles of any kind. When no items are reversed, the answer pattern of an acquiescent person cannot be distinguished from the pattern of a person who intends to agree with all the items. Confronting some reversed items, the acquiescent person will make mistakes that can thus be detected. Unfortunately, the opposite does not hold: agreeing with all items, including the reversed ones, may not only be caused by acquiescence, but also by inattention or by confusion. Hinz et al. reported that an acquiescence response style can be detected in instruments with some reverse worded items . Although they stated that in their sample 6% of the respondents on the MFI-20, showed ‘pronounced acquiescence’, they also admitted that this ‘acquiescence’ could be due to ‘differences in the verbal diction of the items’. Woods  found that carelessness among 10% of the respondents on a scale containing reversed worded items, would already lead to an artificially created factor, inducing researchers to erroneously reject unidimensionality. More researchers attribute the claim of multidimensionality that is often made in scales with reversed worded items, not to conceptual differences, but to artificial factors resulting from response bias –.
Data were collected in a multi-center study from 700 patients with inflammatory bowel disease. Participating centers were Department of Gastroenterology and Hepatology, Maastricht University Medical Center, Department of Internal Medicine and Gastroenterology, Orbis Medical Centre Sittard, and Department of Internal Medicine and Gastroenterology, Atrium Medical Centre Heerlen, all from the Netherlands. The Ethical Committees of all participating centers approved the protocol and written informed consent was obtained from all patients at inclusion in the IBD-SL registry for future analysis and publication of data. In the few patients below 18 years of age their parents or a legal representative signed an informed consent .
We will now examine the effect of reversing items on response bias with data collected with the Multidimensional Fatigue Inventory (MFI-20) . The MFI-20 is intended to measure fatigue with twenty items, and consists of five four-item subscales assessing General Fatigue, Physical Fatigue, Reduced Activity, Reduced Motivation, and Mental Fatigue. There is some inconsistency regarding the number of answer categories used in the MFI-20. In 1995 Smets et al.  constructed a version with seven categories, but in accordance with their adaptation in 1996 , a version is used with five answer categories, ranging from 1 (‘yes, that is true’) to 5 (‘no, that is not true’). The three answer categories in between are lacking a description. Smets et al.  used ‘reversed wording’, the second strategy mentioned by Swain et al. , to reverse ten of the items, thus measuring fitness. All items measuring one of the five types of fatigue, original or reversed, are presented to the respondent in a mixed order (see table 1).
Table 1. MFI-20 items.doi:10.1371/journal.pone.0068967.t001
The MFI is intended to assess fatigue. Therefore, to avoid misunderstandings, items about fatigue are considered to be straightforward, negatively formulated, and items about fitness to be the reversed worded, positively formulated, items.
The developers of the MFI explicitly implemented use of reverse worded items to prevent response set, and thus are not seeking to assess two slightly different aspects of fatigue. Although the developers of the MFI-20 state they reversed items to prevent response set , it is more likely they intended to prevent response styles. Which of the aforementioned response style(s) they try to prevent remains however unclear.
Data were collected in a multi-center study from 700 patients with inflammatory bowel disease (IBD). The aim of that study was to investigate the prevalence and severity of fatigue and the impact on health-related quality of life in patients included in a population-based IBD cohort in the Netherlands . We are aware of the special characteristics of this sample, and it may not be representative of the general population. We can expect this sample to score higher on fatigue than a healthy sample. However, we are not interested in the average level of fatigue and therefor do not consider any differences with the general population to be a problem in the use of the data set as an example. On the contrary, we wanted to analyze a sample with higher average levels of fatigue, in order to avoid highly skewed distributions of answers.
Respondents were send a questionnaire by post. Of the 707 patients that were asked to fill in the MFI-20, seven had missing data on one or more of the twenty items. For this study, all patients with missing data were deleted. Of the remaining 700 patients, 311 were males with mean age 51.1 years (sd 15.2). The 389 females have a mean age of 44.0 (sd 13.7).
Each subscale of the MFI-20 consists of four items, two of them reversed worded. First, we will create pairs of original and reversed worded items, that resemble each other maximally with respect to content, except for their direction.
We will then examine the inter-item correlations for each subset of four items. If reversing items reduces response bias, we would expect two items that are identical with respect to content but different in direction, to be stronger related than two items formulated in the same direction, but with a slightly different content. We consider this a variety of the Multi-Trait-Multi-Method matrix discussed by Campbell & Fiske .
Although suspicious answer patterns may become visible through lower correlation coefficients, the opposite does not hold. Low correlation coefficients can also occur in a homogeneous sample with small item variances. We will therefor do an additional analyses by checking the percentage of suspicious answer patterns for each pair of items within a subset. We will consider a difference of at least three points to be indicative for a wrong answer on at least one of the two items. The choice for a three points criterion is a bit arbitrary, but still well defensible. Since a 5-point scale is used, there are two options left or right from the middle one, indicating (strong) agreement or (strong) disagreement. We did not want an extreme answer (1 or 5) on one item, together with a 3 on the reversed version, to be considered suspicious. Since this would be a difference of 2 points, we choose a three points criterion, that can only be met by an extreme answer on one item (1 or 5) and an extreme or nearby extreme answer on the opposite item (4 or 5 respectively 1 or 2).
Finally, we will assess the psychometric qualities of the ten items measuring fatigue and the ten items measuring fitness separately. The results will be compared with the scales using both types of items simultaneously. If the assumption holds, that reversing part of the items leads to less response bias, we expect instruments that contain both original and reversed items, to have better psychometric properties than instruments containing only original or reversed items.
The scores on negatively formulated items, i.e. items asking for fatigue, are reversed in order to have higher scores indicating higher levels of fatigue. All analyses were done with SPSS 15.
Correlations between Pairs of Items
Spearman correlations between pairs of items were computed in order to compare the effect of content with the effect of direction. Results are presented in table 2.
Table 2. Spearman correlations between itempairs (item numbers between brackets).doi:10.1371/journal.pone.0068967.t002
If reversal of items would have had no effect at all, we would expect the highest correlations between itempairs that resemble each other maximally with respect to content (despite the opposite direction), the pairs in the first column. Also we would expect lower correlations between items that measure (subtle) different aspects of fatigue, irrespective whether this is done with two items formulated in the same direction (column 2) or in opposite directions (column 3).
Overall, the correlations between item pairs formulated in the same direction, shown in the first column, are not consistently higher than in the other columns, indicating an adverse effect of reversed wording.
Suspicious Answer Patterns
In table 3 percentages of respondents with unlikely combinations of answers are given on pairs of items that (should) measure about the same aspect of fatigue. Again, in column 1 item pairs are presented that resemble each other maximally with respect to content, except for different direction. Ideally we would expect no respondent to give unlikely answers, but if they occur, we would expect them to represent apparently important differences in the content of two related items. Thus, we would expect the highest percentages in column 2 and comparable percentages in column 3. Again the percentages in column 2 are not higher than those in column 1, rather a bit lower, indicating an adverse effect of reversed wording.
Table 3. Percentage of large discrepancies (at least three points) between answers on ‘similar’ items.doi:10.1371/journal.pone.0068967.t003
Scale Properties for Positive and Negatively Worded Items Separately
Cronbach’s alpha and the mean inter-item correlation were computed for the complete sets of items as well as for the five dimensions of fatigue. This was also done for the positively and negatively worded items separately. Since Cronbach’s alpha is dependent on scale length, smaller alpha’s can be expected when the scale length is reduced by 50%. Results are presented in table 4.
Table 4. Cronbach’s alpha (mean inter-item correlation between brackets).doi:10.1371/journal.pone.0068967.t004
The mean inter-item correlations of the (sub)scales with combined items are in general lower than of (sub)scales with items all stated in the same direction. The alpha’s of the combined (sub)scales are all above.80 and for the overall scale even.95. Considering the reduction of scale length, lower alpha’s are to be expected with the smaller (sub)scales containing only positive or negative items. The reduction of alpha is however, especially for the negative items rather small.
In this paper we addressed a strategy that is adopted regularly with multi-item questionnaires, namely the use of reversed worded items. Many developers of questionnaires adopt this strategy with the intention of avoiding response bias, particularly acquiescence. We found evidence that this goal is not met. We also discussed an often unintended consequence of reversing some items, namely the debate arising in literature about the presence of two related but different conceptual features, like for example fitness and fatigue. We will discuss first the goal of reversing items and later the consequence. Response bias is a difficult phenomenon to detect as well as to prevent. We distinguished two types of response bias, response set and response style, and focused on characteristics of three types of response styles, that are often confused or overlooked in literature: acquiescence, inattention and confusion due to item verification difficulty. An acquiescent respondent has the tendency to affirm all statements. Pure acquiescence cannot be prevented by reversing half of the items. When all items assess fatigue and are worded in the same direction, an acquiescent person who is fatigued will receive a sumscore close to the true state. An acquiescent person who is not at all fatigued, however, will receive a biased sumscore. Thus, reversing half of the items will lead to a less biased sumscore for the latter respondent, but only at the cost of a more biased sumscore for the fatigue respondent. Therefore, reversing some items cannot be presumed to lead to better assessments in case of acquiescent respondents.
Inattention can relate to several different aspects of an item, like the specific content (is it about mental or physical fatigue), the time frame (is it about today, or the last week), but also the direction of the item (is it about being fatigue or not being fatigue) or the answer categories (do they range from always to never, or the other way around).
On closer examination, several characteristics can be distinguished, all covered by the concept inattention. Some respondents may be less precise in general. Others may become more inattentive in answering, depending on the length of the questionnaire and the extent to which the items resemble each other. And some may get frustrated by having to answer more or less the same items either in the same or opposite direction. Thus, in some cases extension of the questionnaire by more items, in the same or opposite direction, may work counterproductive.
Confusion is a response style dependent on the difficulty of the item and the cognitive strategies the respondent has to employ to give an answer that is in accordance with the true state.
We argued why reversing a portion of the items is an ineffective way of dealing with response bias. Reversing items by using negative particles or affixal morphemes will lead to increased difficulty, and thus more bias, without any clear advantage. Reversing some items by reversed worded items, may decrease item difficulty for those respondents that can agree with the reversed items, but at the same time will lead to more bias due to confusion for the other respondents, together with increased bias due to inattention for all respondents. In any case, acquiescence will not be avoided, at best detected. To a great extent, the confusion that is caused by reversing items, is due to the custom to present both original and reversed items mixed up.
Finally, we demonstrated that a particular instrument, the MFI-20, designed to prevent response bias using reverse wording of half the items, does not achieve this goal. The MFI-20 is a widely used instrument for reliable and valid assessment of fatigue in general and several types of fatigue. Results of the study raise questions whether the addition of ten reverse worded items, intended to prevent response bias, is justified. An added value of the negatively formulated items to the positive items, or vice versa, was not demonstrable. With respect to content of the items, the ten negatively and ten positively formulated items are measuring almost the same, if not exactly the same aspects of fatigue. Since the developers of the MFI explicitly stated that they added reverse worded items in order to tackle response bias, it would be useless to focus on any potential difference with respect to content or responsiveness of these ten items.
Instead of preventing response bias, the addition of ten reverse worded items appears to increase the risk of inattention and confusion. No intensive focus was put on potential subtle differences with respect to their content between two reverse worded items. Firstly, because the developers of the questionnaire explicitly stated that the purpose of adding reverse worded items was to prevent response bias. Secondly, any difference with respect to content should, if it is considered to be important, be assessed by items, all formulated in the same direction, in order to maximize opportunities to assess subtle differences and to avoid artifacts due to accidentally misreading.
The addition of ten reverse worded items did lead to slightly higher values of Cronbach’s alpha. This is however to be expected when scales are twice as long. With one exception, mean inter-item correlations decreased when adding ten reversed worded items, where and increase was to be expected, considering the reason for adding reversed items.
Considering the findings of Swain et al. , respondents seem to make less errors with items that reflect their experience or situation than with items that describe the opposite. Since this instrument is designed to measure fatigue, it will probably be used more often among persons with a certain level of fatigue. Therefore the negatively formulated items are to be preferred. The psychometric qualities of these ten items are acceptable, if not good.
A consequence of reversing items is the identification of two related but unipolar concepts where only one was intended. We expect some validity to the claim of unipolarity and thus two related concepts that are tapped by asking for both fitness and fatigue, positivism and negativism, happiness and sadness, being relaxed and nervous. However, we consider the emerging of these claims, originating from data-analysis, instead of from a theoretical position, a serious weakness. If distinguishing between two related but opposite concepts is truly relevant, it would be helpful to take precautionary actions to assess these concepts unambiguously. In accordance with Roszkowski and Soven , we suggest separate presentation of ordinary items and reversed worded items, instead of a list where these items are all mixed up.
Even when a multi-item questionnaire consists of items stated in the same direction, there are problems to be addressed that hamper an obvious relationship between the theoretical concept and the sumscore resulting from an addition of the itemscores . Some aspects, commonly seen in multi-item instruments, that deserve to be addressed are:
- Differences in item difficulty and their consequences for the interpretation of summed scores, a field that Item Response Theory is addressing.
- Sometimes in the same questionnaire some aspects are addressed with more items than others, leading to an often unknown and implicit weighing of their contribution to the total score.
- The rationale and consequences of using different answercategories for items that are supposed to belong to the same scale
- The rationale and consequences of using both items asking for frequency and items asking for intensity.
All these phenomena deserve to be addressed. This discussion will be more fruitful if it is not obscured by effects resulting from reversed worded items.
In conclusion, we consider reversing items in order to prevent response bias a counterproductive strategy. Acquiescence cannot be prevented by reversing, and more errors will be made due to inattention or confusion. An instrument with all items formulated in the same direction and referring to the intended concept (i.e. fatigue or fitness, depression or happiness) is to be preferred. If a researcher is concerned about respondents missing subtle differences between the items, other strategies are to be considered.
It is surprising that reversing items, introduced several decades ago, is still predominant in many popular questionnaires. Discussion about the pros and cons of this phenomenon should be revived. Consider, on a rainy day, all cows in a pasture tending to stand facing in the same direction, with their back pointing from where the wind comes. We admit that one cow standing in the opposite direction, would be conspicuous immediately. Unfortunately items do not have a head and tail.
The authors wish to thank Mariëlle Romberg-Camps, and Marieke Pierik, Department of Gastroenterology and Hepatology, Maastricht University Medical Center, the Netherlands, for providing the dataset. We also wish to thank the two anonymous reviewers. Their comments improved the manuscript considerably.
Conceived and designed the experiments: EVS RS JCC. Analyzed the data: EVS. Contributed reagents/materials/analysis tools: EVS. Wrote the paper: EVS RS JCC.
- 1. Rorer LG (1965) The great response-style myth. Psychol Bull 63: 129–156. doi: 10.1037/h0021888
- 2. Baumgartner H, Steenkamp JBEM (2001) Response styles in marketing research: A cross-national investigation. J Marketing Res 38: 143–156. doi: 10.1509/jmkr.220.127.116.1140
- 3. Tourangeau R, Rips LJ, Rasinski K (2000) The psychology of survey response. Cambridge, England: Cambridge University Press.
- 4. Paulhus DL (1991) Measurement and control of response bias. In: Robinson JP, Shaver PR, Wrightsman LS, editors. Measures of Personality and Social Psychological Attitudes, 17–59. San Diego, CA: Academic Press.
- 5. Weijters B (2006) Response styles in consumer research. Doctoral dissertation, Ghent University, Ghent, Belgium.
- 6. Swain SD, Weathers D, Niedrich RW (2008) Assessing three sources of misresponse to reversed Likert items. J Marketing Res 45: 116–131. doi: 10.1509/jmkr.45.1.116
- 7. Krosnick JA (1999) Survey research. Annu Rev Psychol 50: 537–567. doi: 10.1146/annurev.psych.50.1.537
- 8. Weijters B, Baumgartner H, Schillewaert N (2013, May 6) Reversed Item Bias: An Integrative Model. Psychological Methods. Advance online publication. doi: 10.1037/a0032121.
- 9. Drolet AL, Morrison DG (2001) Do we really need multiple-item measures in service research? J Serv Res-US 3: 196–204. doi: 10.1177/109467050133001
- 10. Schriesheim CA, Hill KD (1981) Controlling acquiescence response bias by item reversals: The effect on questionnaire validity. Educ Psychol Meas 41, no. 4: 1101–14.
- 11. Meyer JP, Allen NJ (1991) A three-component conceptualization of organizational commitment. Hum Resour Manage R, 1, 61–89.
- 12. Merritt SM (2012) The Two-Factor Solution to Allen and Meyer’s (1990) Affective Commitment Scale: Effects of Negatively Worded Items. J Bus Psychol, Volume 27, issue 4, 421–436.
- 13. Roszkowski MJ, Soven J (2010) Shifting gears: Consequences of including two negatively worded items in the middle of a positively worded questionnaire. Assess Eval High Educ, 35, 113–130.
- 14. Meyer TJ, Miller ML, Metzger RL, Borkovec TD (1990) Development and validation of the Penn State Worry Questionnaire. Behav Res Ther, 28, 487–495.
- 15. Fresco DM, Heimberg RG, Mennin DS, Turk CL (2002) Confirmatory factor analysis of the Penn State Worry Questionnaire. Behav Res Ther, 40, 313–323.
- 16. Hazlett-Stevens H, Ullman JB, Craske MG (2004) Factor structure of the Penn State Worry Questionnaire: Examination of a method factor. Assessment, 11(4), 361–370.
- 17. Bradley KD, Royal KD, Bradley JW (2008) An investigation of ‘honesty check’ items in higher education course evaluations. J Coll Teach Learn, 5(8), 39–48.
- 18. Hinz A, Michalski D, Schwarz R, Herzberg PY (2007) The acquiescence effect in responding to a questionnaire. GMS Psychosoc Med 4: Doc07.
- 19. Woods CM (2006) Careless responding to reverse-worded items: Implications for confirmatory factor analysis. J Psychopathol Behav 28: 189–194. doi: 10.1007/s10862-005-9004-7
- 20. Spector PE, VanKatwyk PT, Brannick MT, Chen PY (1997) When two factors don't reflect two constructs: How item characteristics can produce artifactual factors. J Manage 23: 659–677. doi: 10.1177/014920639702300503
- 21. Schmitt N, Stults DM (1985) Factors defined by negatively keyed items - the result of careless respondents. Appl Psych Meas 9: 367–373. doi: 10.1177/014662168500900405
- 22. Watson D, Tellegen A (1999) Issues in the dimensional structure of affect - Effects of descriptors, measurement error, and response formats: Comment on Russell and Carroll (1999). Psychol Bull 125: 601–610. doi: 10.1037/0033-2909.125.5.601
- 23. Russell JA, Carroll JM (1999) On the bipolarity of positive and negative affect. Psychol Bull 125: 3–30. doi: 10.1037//0033-2909.125.1.3
- 24. Russell JA, Carroll JM (1999) The phoenix of bipolarity: Reply to Watson and Tellegen (1999). Psychol Bull 125: 611–617. doi: 10.1037/0033-2909.125.5.611
- 25. Segura SL, Gonzalez-Roma V (2003) How do respondents construe ambiguous response formats of affect items? J Pers Soc Psychol 85: 956–968. doi: 10.1037/0022-3518.104.22.1686
- 26. Romberg-Camps M, Bol Y, Dagnelie PC, van der Hesselink-Kruijs MAM, Kester AD, et al.. (2010) Fatigue and health-related quality of life in inflammatory bowel disease: Results from a population-based study in the Netherlands: The IBD-South Limburg cohort. Inflamm Bowel Dis, 16(12), 2137–2147.
- 27. Smets EM, Garssen B, Cull A, de Haes JC (1996) Application of the multidimensional fatigue inventory (MFI-20) in cancer patients receiving radiotherapy. Brit J Cancer 73: 241–245.
- 28. Smets EM, Garssen B, Bonke B, de Haes JC (1995) The Multidimensional Fatigue Inventory (MFI) psychometric qualities of an instrument to assess fatigue. J Psychosom Res 39: 315–325. doi: 10.1016/0022-3999(94)00125-o
- 29. Campbell DT, Fiske DW (1959) Convergent and discriminant validation by the Multitrait-Multimethod Matrix. Psychol Bull 56: 81–105. doi: 10.1037/h0046016