Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Social distancing in America: Understanding long-term adherence to COVID-19 mitigation recommendations

  • Christopher P. Reinders Folmer ,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Resources, Software, Visualization, Writing – original draft, Writing – review & editing

    c.p.reindersfolmer@uva.nl

    Affiliation Amsterdam Law School, Center for Law and Behavior, University of Amsterdam, Amsterdam, The Netherlands

  • Megan A. Brownlee,

    Roles Investigation, Methodology, Software, Writing – original draft, Writing – review & editing

    Affiliation Amsterdam Law School, Center for Law and Behavior, University of Amsterdam, Amsterdam, The Netherlands

  • Adam D. Fine,

    Roles Methodology, Resources, Writing – original draft, Writing – review & editing

    Affiliation School of Criminology and Criminal Justice, Arizona State University, Phoenix, Arizona, United States of America

  • Emmeke B. Kooistra,

    Roles Methodology, Resources, Writing – review & editing

    Affiliation Amsterdam Law School, Center for Law and Behavior, University of Amsterdam, Amsterdam, The Netherlands

  • Malouke E. Kuiper,

    Roles Methodology, Resources, Writing – review & editing

    Affiliation Amsterdam Law School, Center for Law and Behavior, University of Amsterdam, Amsterdam, The Netherlands

  • Elke H. Olthuis,

    Roles Methodology, Resources, Writing – review & editing

    Affiliation Amsterdam Law School, Center for Law and Behavior, University of Amsterdam, Amsterdam, The Netherlands

  • Anne Leonore de Bruijn,

    Roles Methodology, Resources, Writing – review & editing

    Affiliation Amsterdam Law School, Center for Law and Behavior, University of Amsterdam, Amsterdam, The Netherlands

  • Benjamin van Rooij

    Roles Conceptualization, Funding acquisition, Methodology, Resources, Supervision, Writing – original draft, Writing – review & editing

    Affiliations Amsterdam Law School, Center for Law and Behavior, University of Amsterdam, Amsterdam, The Netherlands, School of Law, University of California, Irvine, California, United States of America

Abstract

A crucial question in the governance of infectious disease outbreaks is how to ensure that people continue to adhere to mitigation measures for the longer duration. The present paper examines this question by means of a set of cross-sectional studies conducted in the United States during the COVID-19 pandemic, in May, June, and July of 2020. Using stratified samples that mimic the demographic characteristics of the U.S. population, it seeks to understand to what extent Americans continued to adhere to social distancing measures in the period after the first lockdown ended. Moreover, it seeks to uncover which variables sustained (or undermined) adherence across this period. For this purpose, we examined a broad range of factors, relating to people’s (1) knowledge and understanding of the mitigation measures, (2) perceptions of their costs and benefits, (3) perceptions of legitimacy and procedural justice, (4) personal factors, (5) social environment, and (6) practical circumstances. Our findings reveal that adherence was chiefly shaped by three major factors: respondents adhered more when they (a) had greater practical capacity to adhere, (b) morally agreed more with the measures, and (c) perceived the virus as a more severe health threat. Adherence was shaped to a lesser extent by impulsivity, knowledge of social distancing measures, opportunities for violating, personal costs, and descriptive social norms. The results also reveal, however, that adherence declined across this period, which was partly explained by changes in people’s moral alignment, threat perceptions, knowledge, and perceived social norms. These findings show that adherence originates from a broad range of factors that develop dynamically across time. Practically these insights help to improve pandemic governance, as well as contributing theoretically to the study of compliance and the way that rules come to shape behavior.

Introduction

The global COVID-19 outbreak in 2020 has made clear that the initial defense against a new deadly infectious disease requires large scale behavioral modification. Until there is a vaccine or a cure that can halt a pandemic outbreak, the only protection that people have is to ensure that the spread of the disease is minimized. This entails a range of changes in basic human conduct, from things that have limited economic and social consequences, such as better hand hygiene and the adoption of face masks, to profound, costly changes such as social distancing, forced isolation, quarantine, and broader lockdowns. Such measures only work, however, if people effectively follow them. In this way, the 2020 pandemic has shown the importance of understanding compliance and adherence to outbreak mitigation measures.

There is now a quite well-developed body of research about what made people across the globe follow mitigation measures when they were first adopted. When many governments adopted lockdown rules and social distancing measures as compulsory mandates during this initial “first wave” period, compliance levels were high. This is demonstrated not only by drastic reductions in mobility [1], but also by consequences associated with this, such as the unprecedented event that the price of oil turned negative [2]. A recent review identifies a range of variables that predicted compliance with social distancing measures during the first pandemic wave, including psychosocial, institutional, and situational variables, as well as incentives [3]. Furthermore, this review showed that some highly important policy variables were not associated with compliance during this period. These included for instance deterrence where neither the threat of stricter punishment nor more certain punishment predicted compliance.

After the first wave, many countries lifted the most invasive restrictions, such as lockdowns, and even some of the social distancing measures. Yet as the outbreak was neither controlled nor overcome through a vaccine or medicine, mitigation measures have remained essential for keeping the virus at bay. During the fall, however, many countries found themselves faced with a second pandemic wave. This raises the question of how adherence to mitigation measures has developed during the summer months after the initial strict behavioral measures of the first wave were repealed. Is it the case that social distancing has degraded back toward pre-pandemic normality, and thus gave fertile ground for a resurgence of infections? And if so, which factors shaped such changes and caused people to abandon (or sustain) social distancing?

To understand these questions, the present research collected three cross-sectional surveys in the United States in May, June, and July of 2020. Using stratified samples that mimic the demographic characteristics of the U.S. population, we examined how Americans’ adherence to social distancing measures has developed across this period, and which factors have sustained or undermined this. To answer this question, we consider a broad range of influences, which can be arranged into six categories. First, factors related to people’s practical understanding and knowledge of the measures. Second, factors related to their perception of the costs and benefits of the measures. Third, factors related to their perceptions of the legitimacy and procedural justice of the measures and the responsible authorities. Fourth, personal factors relevant to adherence. Fifth, influences from people’s social environment. And sixth, practical circumstances that may constrain or facilitate their adherence. The paper allows us to understand how these variables shape adherence to social distancing measures in the critical period that follows as a country starts to reopen following a first wave in a pandemic outbreak. By doing so, we contribute to the overall understanding of pandemic governance, as well as to insight into the interaction between rules and human conduct most generally. We also contribute to compliance theory by illuminating how influences at each of these levels may shape adherence over a longer time period. And finally, we identify important avenues for policy, on how adherence to mitigation measures can be promoted when strict measures are lifted.

The present study

Following the initial lockdown period in the spring of 2020, the United States underwent dramatic changes, both in terms of the spread of the virus and the measures to counter it. At the beginning of April, approximately 70% of Americans were subject to stay-at-home and social distancing measures [4, 5]. However, by the end of April, infections began to decline [6], and some states began re-opening or reopened altogether, starting with the Southern and Midwestern regions [7]. During the same period, Federal social distancing guidelines were repealed [8], although the requirement remained in place at nearly every state level [9]. Infection rates strongly accelerated from mid-June to late July, however, reaching a peak of almost 75,000 new cases per day [6].

The period between May and July was also characterized by increasing controversy over mitigation measures. There was a continuation of protests against mitigation measures, where people deliberately violated social distancing and other mitigation measures [10, 11]. Furthermore, mitigation measures became increasingly politicized. Compared to Democrats, Republicans voiced greater concern over the economic costs of mitigation measures, and less concern over the threat of the virus [12, 13]. This was illustrated during the 2020 presidential election campaign, where Republican mass rallies were held and some organizers actively countered social distancing measures (e.g., by removing “do not sit here” stickers) [14].

Throughout this period, mitigation measures have remained essential for keeping the virus at bay. But to what extent have Americans followed these measures, and what factors influenced them to do so (or not)? To answer these questions, we leveraged three surveys, collected in May, June, and July of 2020, among stratified samples that mimic the demographic characteristics of the U.S. population.

Our surveys focus on adherence to social distancing recommendations. Although they became less visible in federal public health recommendations after this period, social distancing recommendations continue to exist nearly everywhere at the state level [9]. Our surveys assessed self-reported adherence to social distancing recommendations across various situations, and examine how this has developed in the period after the first wave lockdown. Furthermore, we explored a range of factors that may explain why people did, or did not adhere to these measures, derived from insights on compliance from psychology, criminology, sociology, and economics [5, 1518]. In operationalizing the present study we broadly distinguish six categories of variables:

  1. People’s practical understanding and knowledge of mitigation measures. In order to be able to adhere to mitigation measures, it is necessary that people have sufficient knowledge of what is expected from them [1921], and that the measures are clear to them [22]. Accordingly, our surveys firstly test people’s knowledge of social distancing measures, and the perceived clarity of the mitigation measures to them. Logically, a lack of knowledge about mitigation measures would be expected to reduce adherence, as would lower perceived clarity.
  2. The perceived costs and benefits of the mitigation measures. According to the rational choice theory of compliance, people’s tendency to adhere should decrease as the costs of doing so become larger, and increase as the benefits improve [23, 24]. Our surveys assess different aspects of this. A first aspect is people’s perception of the threat of the virus. Mitigation measures become more beneficial if people regard the virus as a severe threat to their own health and/or that of others. Yet the health risk of COVID-19 varies between individuals [2527], as do subjective perceptions of this risk [28]. For this reason, we expected that adherence to social distancing measures would be higher among people who perceive the virus as a greater health threat. The second aspect is the cost people personally face due to the mitigation measures. Due to the pandemic and the measures to mitigate it, many Americans have suffered decreases to their income or employment opportunities [29]. We expected that adherence would be lower among people for whom the personal costs of the mitigation measures are greater. The final aspect is fear of punishment (deterrence). Although social distancing measures were not widely enforced in the U.S., sanctions did occur during the first wave lockdown [30]; furthermore, severe sanctions were communicated for other COVID-related violations [31]. Research on perceptual deterrence suggests that subjective perceptions of punishment may also influence compliance [32]. For this reason, we also examined subjective perceptions of punishment for not following social distancing measures, separating punishment certainty and severity–the key dimensions separated by general deterrence theory [3335]. We expected that adherence would be higher among people who regarded punishment as more certain, and more severe.
  3. The perceived legitimacy and procedural justice of the mitigation measures and the responsible authorities. As Max Weber has explained: “So far as it is not derived merely from fear or from motives of expediency, a willingness to submit to an order imposed by one man or a small group, always implies a belief in the legitimate authority (Herrschaftsgewalt) of the source imposing it” (see [36] p. 37). Accordingly, we also aimed to capture such legitimacy perceptions in our study. Jackson and Gau [37] describe legitimacy as the property or quality of possessing rightful power and the acceptance of authority. To the extent that the law and legal authority are perceived to be legitimate, people will feel more obligated to obey the law. Individuals judge legal authority to be legitimate to the extent that they embody the values of being appropriate and proper [38, 39]. Our study assesses six core areas of this. First, we assess people’s moral alignment with social distancing measures; i.e., the extent to which they agree with the substance of these measures [40, 41]. During the period that preceded our study, there were clear indications that support for mitigation measures differed among Americans [42, 43]. We expected that adherence would be higher among people with greater substantive support for social distancing measures.

A second core area is people’s evaluation of the authorities’ responses. To study this, we examined whether people found the overall approach taken by authorities to be consistent and adequate. We expected that adherence would be higher among people who evaluated the authorities’ approach more favorably. Relatedly, we assessed procedural justice, or people’s perceptions of the procedural fairness through which the rules were made and enforced. The more that people see that rules are made and enforced in a procedurally fair manner, the more likely it is that they will see them as legitimate–and the more likely it becomes that they will feel bound to obey such measures and come to comply with them [40, 41, 44].

In the final area, we assessed people’s sense of duty to obey the law. Such sense of duty is a core expression or a downstream consequence of their felt legitimacy, as people with a higher sense of legitimacy, in theory, develop more of a sense of a duty to obey rules developed and enforced by authorities they view as legitimate [45]. We have used three measures to capture this. First is the normative obligation to obey the law, which captures people’s sense of duty to obey the law out of moral obligation [46]. Second is the non-normative duty to obey the law, which originates in a sense of coercion or fear, where people feel obligated to obey the law out of fear of the authorities [46]. And last is people’s obligation to obey the law in general, which captures the extent to which they feel that they should obey the law regardless of different circumstances [4749].

  1. 4. Personal factors relevant to adherence. As the fourth facet of adherence to social distancing measures, we look at personal factors that are relevant for people’s stance toward mitigation measures, or for compliance more generally. A first factor is people’s trust in science. Scientific evidence (and indeed, scientists) have played an important (and very visible) role in the public health response to COVID-19, and the measures to mitigate it. Yet trust in science varies between individuals, which may strongly affect their willingness to follow these measures [50, 51]. We expected that adherence to social distancing measures would be higher among people who have greater trust in science. A second, related factor is trust in traditional media. Research suggests that distrust in traditional media is associated with greater belief in misinformation about COVID-19 [52]. This, in turn, predicts lower adherence to measures to mitigate it [53]. Accordingly, we expected adherence to social distancing measures to be higher among people with more trust in traditional media. The third personal factor is impulsivity. To effectively distance oneself from others, it is necessary to inhibit one’s usual tendency to get close to them. However, people differ in their capacity to control their impulses, and high levels of impulsivity predict deviant and rule breaking behavior [5458]. We therefore expected adherence to be lower among more impulsive individuals. Last, we examined people’s emotional state. According to strain theory, people may cope with negative emotions through rule violating behavior [5965]. Indeed, also in context of the COVID-19 pandemic, studies show that negative emotions may lead to lower compliance with quarantine measures [66]. Thus, we expected adherence to be lower among people who experienced more negative emotions.
  2. 5. People’s social environment. As the fifth facet of adherence, we look at influences from people’s social environment–specifically descriptive social norms for adhering. In many situations, it is highly visible whether others do (or do not) adhere to social distancing measures. Research shows that perceptions of the norms for complying with particular rules or requests can have an important effect on people’s own tendency to do so: the more that they see others comply, the more likely they are to do so themselves; the more that they see others violate or disobey, the more likely they are to offend [6770]. In light of this, our surveys assessed people’s perceptions of the norms for social distancing within their community. We expected adherence to be higher among people who perceived more adherence within their community.
  3. 6. People’s practical circumstances. As the final facet, we looked at the practical circumstances that may shape people’s adherence. Whether people can adhere to social distancing measures (or conversely, can violate these) may also be contingent on the extent to which their practical circumstances allow them to do so. Our surveys looked at different aspects of this. First, people’s practical capacity to adhere. In order for people to effectively do as social distancing measures demand, it is necessary that their practical circumstances effectively allow them to do so. However, in practice, their capacity to adhere may often vary. For example, keeping a safe distance from others may be more difficult in crowded or constrained environments, or in occupations that cannot be conducted from home or at a distance. Capacity thus may strongly shape adherence, but it should be understood that these concepts are not identical. Simply having the capacity to commit a crime does not mean that one also will do so. The same applies to social distancing: being practically able to keep a distance from others does not mean that someone wishes to do so. We expected adherence with social distancing measures to be higher among people who had greater practical capacity to adhere to these measures. The second aspect is people’s opportunities for violating the measures. In order to violate social distancing recommendations, it first is necessary that there are practical opportunities to do so. However, practical circumstances may make this impossible, for example, when physical environments have been rearranged to separate people from each other. Insights from routine activities theory [7173] and situational crime prevention [74, 75] show that there is less rule breaking when there are less practical opportunities to do so. We expected greater adherence with social distancing measures among people who saw less opportunities for violating such measures by getting close to others.

Method

We obtained ethical approval for this project from the Institutional Review Board of the University of California, Irvine and by the Ethics Review Board of the University of Amsterdam. All participants provided consent before participating in the study. Participation was voluntary, and all participants could stop the survey at any time.

Participants

Participants were residents (18 years or older) of the U.S. that were recruited via the online survey platform SurveyMonkey (https://surveymonkey.com). They were recruited using a stratified sampling approach, in which the final intended sample size was divided into subgroups with the same demographic proportions (age, gender, and race/ethnicity) as the national population based on estimates from the U.S. Census Bureau (https://www.census.gov/). This stratified sampling approach mimics the demographic characteristics of the United States, though it retains the biases and characteristics of a non-probability convenience sample. Three cross-sectional surveys were administered in May, June, and July 2020, using different samples of participants. Participants were paid $3.00 for participating.

1,452 participants took part in Survey 1 (May 8–18). Here, 436 participants were excluded from the sample because they failed to complete the survey, provided incomplete responses, or failed to pass two attention checks. Six participants indicated a nonbinary gender identity; as this number was insufficient for analysis, they were also omitted. The final sample for Survey 1 consisted of 1,012 cases (56.5% women, 43.5% men; Mage = 40.32 years).

1,711 participants took part in Survey 2 (June 8–16). Here, 723 participants failed to complete the survey, provided incomplete responses, or failed to pass two attention checks; these participants were excluded. Additionally, five nonbinary participants were omitted from the sample. The final sample for Survey 2 consisted of 986 cases (54.3% women, 45.7% men; Mage = 40.17 years).

1,758 participants took part in Survey 3 (July 11–17). Here, 835 participants failed to complete the survey, provided partial responses, or failed to pass two attention checks; again, these participants were excluded. Four nonbinary participants were also omitted. As such, the final sample for Survey 3 consisted of 921 cases (52.7% women, 47.3% men; Mage = 40.17 years).

In total, the sample thus consisted of 2,919 cases across three cross-sectional survey waves (54.5% women, 45.5% men; Mage = 40.22 years). The sample thus was slightly more female and older than the general population (2019 census: 50.9% women, 49.0% men; Mage = 38.3 years) [76]. There was some variability between waves on specific variables (i.e., education, COVID care, inclusion in an ethnic minority group, insurance status, socio-economic status change, and health risk to self and others). These variables were either unrelated to adherence or controlled for in the analyses. Demographical information for all three survey waves and for the full sample is displayed in Table 1.

thumbnail
Table 1. Sample characteristics and control variables, Surveys 1 (May), 2 (June), and 3 (July), and full sample.

https://doi.org/10.1371/journal.pone.0257945.t001

Materials

Survey

Our survey (see Supporting Information) was based on our prior surveys conducted in April 2020 in the United States [5], the United Kingdom [77], the Netherlands [78], and Israel [79]. It assessed the same variables and relied on the same measures. Measures that displayed poor internal consistency in the previous surveys were revised to improve their internal consistency (e.g., adherence, social norms, capacity to adhere, and opportunity to violate); reliability of the revised measures was high (α ≥ .85, more details below). Throughout the survey, we referred to COVID-19 as “the coronavirus,” which reflects the greater usage of this name in everyday speech, especially during the early stages of the pandemic.

Control variables

The following demographic variables were recorded: age, gender, nationality, information on residency (state), inclusion in an ethnic minority group, education, employment status, insurance status, social economic status before and after COVID-19 (MacArthur Scale of Subjective Social Status [80]), and political orientation (adapted from [8183]). For political orientation, a considerable number of participants preferred to not disclose their preference (Survey 1: 12.4%; Survey 2: 10.7%; Survey 3: 12.8%). To enable such cases to be retained in the analysis, this variable was recoded into two dummy variables: one comparing conservative to progressive orientation (1 = very conservative or conservative, 0 = progressive, very progressive, or prefer not to say) and one comparing undisclosed to progressive orientation (1 = prefer not to say, 0 = very conservative, conservative, progressive, very progressive). This approach yielded the same results for adherence as the scale measure, but allowed all cases to be utilized.

Additionally, we asked several questions that probed exposure to and risk from COVID-19. Specifically, we asked participants to indicate whether they provided professional care for coronavirus patients, and whether they or anyone they knew had underlying health issues that would put them more at-risk to suffer complications from the coronavirus.

Correlations between the control variables for all three surveys are displayed in S1S3 Tables.

Adherence to social distancing measures

To assess adherence to social distancing measures, we measured participants’ self-reported tendency to keep a safe distance from others in various situations [18]. Specifically, we included seven questions that measured their tendency to keep a safe distance (six feet or more) from: (1) “others outside of my direct household,” (2) “my neighbors,” (3) “colleagues at work,” (4) “friends and family from outside of my direct household,” (5) “others when grocery shopping,” (6) “others when taking a walk or exercising,” and (7) “others when commuting or traveling” (1 = “never,” 7 = “always”). Responses were mean-scored into a single measure for each wave (Survey 1: α = .92; Survey 2: α = .92; Survey 3: α = .93), with higher scores indicating greater adherence to COVID-19 social distancing measures (see Table 2).

thumbnail
Table 2. Descriptive statistics of dependent variables, Surveys 1 (May), 2 (June), and 3 (July), and full sample.

https://doi.org/10.1371/journal.pone.0257945.t002

Practical knowledge and understanding

To assess participants’ knowledge and understanding of the mitigation measures, two variables were measured: (1) knowledge of these measures, and (2) perceived clarity of these measures.

To measure participants’ knowledge of mitigation measures, we asked them to indicate whether current COVID-19 mitigation measures required them to keep a safe distance (six feet or more) from others (1 = yes, 2 = no, 3 = don’t know). The key comparison is whether people who know that they are under social distancing measures adhere more to these recommendations than people who do not, or are unsure of this. To capture this, these responses were recoded (1 = yes, 0 = no or don’t know).

To measure the perceived clarity of mitigation measures, one item was solicited. This asked them to evaluate how clear the measures were that were taken by the authorities to reduce the spread of the coronavirus (1 = “extremely unclear;” 7 = “extremely clear”).

Costs and benefits

To assess the costs and benefits of the mitigation measures, four variables were measured: (1) the perceived health threat of COVID-19, (2) personal costs of the mitigation measures, (3) perceptions of the certainty of punishment for not following social distancing measures, and (4) perceptions of the severity of punishment for failure to do so.

The perceived health threat of COVID-19 was measured by mean-scoring three items. These asked participants to indicate to what extent they believed the coronavirus to be a major threat to (1) their own health, (2) the health of friends and relatives, and (3) the general health (1 = “strongly disagree,” 7 = “strongly agree”). Their answers were combined into a scale measure (Survey 1: α = .91; Survey 2: α = .92; Survey 3: α = .92), with higher scores indicating greater perceived health threat.

Personal costs of COVID-19 mitigation measures were assessed by means of five items. Specifically, we asked participants to indicate how likely it was that they would (1) “lose income,” (2) “lose their job,” (3) “not be able to work,” (4) “not be able to work as effectively as normal,” and (5) “experience a negative impact on their social life” as a result of the measures (1 = “extremely unlikely,” 7 = “extremely likely”). These were combined into a scale measure of personal costs (Survey 1: α = .86; Survey 2: α = .86; Survey 3: α = .86), with higher scores indicating personal greater costs of the mitigation measures.

Perceptions of punishment certainty for violating social distancing measures were measured with two questions. These assessed the perceived likelihood that the authorities would (1) “find out,” and (2) “punish you” if participants would not keep a safe distance (six feet or more) from others (1 = “extremely improbable,” 7 = “extremely probable”). Both items were highly correlated (Survey 1: r = .75; Survey 2: r = .75; Survey 3: r = .74), and hence were aggregated into a scale measure, with higher scores indicating greater perceived punishment certainty.

Perceptions of punishment severity were assessed using one item. Participants indicated how much they would “suffer” if the authorities would punish them for not keeping a safe distance (six feet or more) from others (1 = “extreme suffering;” 6 = “no suffering at all”). The item was reverse-coded so that higher scores indicate greater perceived punishment severity.

Legitimacy, procedural justice, and obligation to obey

Six variables were measured to capture participants’ perceptions of the legitimacy of the mitigation measures and the responsible authorities, and their felt obligation to follow them: (1) their moral alignment with social distancing measures, (2) their evaluation of the authority response to the pandemic, (3) their normative obligation to obey the authorities handling the pandemic, (4) their non-normative obligation to obey these authorities, (5) their general obligation to obey the law, and (6) their perception of the procedural fairness of these authorities when enforcing the measures.

Moral alignment with social distancing measures was measured by asking participants to which extent they “morally believe that people should keep a safe distance from others (six feet or more) in order to contain the coronavirus” (1 = “strongly disagree,” 7 = “strongly agree”).

Evaluation of the authority response was measured using two items. These asked to which extent participants believed the authorities to have been (1) “consistent,” and (2) “adequate” in their response to contain the coronavirus (1 = “strongly disagree,” 7 = “strongly agree”). Both items were strongly correlated (Survey 1: r = .81; Survey 2: r = .79; Survey 3: r = .80); accordingly, a scale measure was constructed from their responses, with higher scores indicating more favorable evaluations.

Participants’ normative obligation to obey the authorities handling COVID-19 was measured by mean-scoring three items (adapted for this study following [46, 84]): (1) “I feel a moral obligation to obey the authorities handling the coronavirus,” (2) “I feel a moral duty to support the decisions of the authorities handling the coronavirus, even if I disagree with them,” and (3) “I feel a moral duty to obey the instructions of the authorities handling the coronavirus, even when I don’t understand the reasons behind them” (1 = “strongly disagree,” 5 = “strongly agree”). Answers were aggregated into a scale measure (Survey 1: α = .87; Survey 2: α = .89; Survey 3: α = .90). Higher scores indicated greater normative obligation to obey.

Participants’ non-normative obligation to obey the authorities handling COVID-19 was assessed with three items (again adapted for this study following [46, 84]): (1) “people like me have no choice but to obey the authorities handling the coronavirus,” (2) “if you don’t do what the authorities handling the coronavirus tell you they will treat you badly,” and (3) “I only obey the authorities handling the coronavirus because I am afraid of them” (1 = “strongly disagree,” 5 = “strongly agree”). Responses were combined into a scale measure (Survey 1: α = .72; Survey 2: α = .73; Survey 3: α = .70), with higher scores indicating greater non-normative obligation to obey.

Participants’ general obligation to obey the law was measured using the 12-item Rule Orientation scale [47]. This instrument assesses the perceived acceptability of breaking legal rules across a range of situations (e.g., when the rule is against one’s moral principles; when the rule is not enforced; when others think that breaking the rule is justified, etc.; 1 = “strongly disagree,” 7 = “strongly agree”). Responses were mean-scored into a scale measure (Survey 1: α = .94; Survey 2: α = .94; Survey 3: α = .94), with higher scores indicating greater felt obligation to obey the law in general.

Perceptions of the authorities’ procedural fairness in enforcing the mitigation measures were measured by means of four items (adapted from [40, 8587]). These asked to which extent they expected that the authorities would: (1) “treat people with respect,” (2) “give a person the chance to tell their side of the story if the person is accused of violating measures to contain the coronavirus,” (3) “treat people fairly, despite gender, race, religion, or socioeconomic background,” and (4) “be honest in enforcing measures to contain the coronavirus” (1 = “strongly disagree,” 7 = “strongly agree”). Responses were aggregated into a scale measure (Survey 1: α = .92; Survey 2: α = .93; Survey 3: α = .92), with higher scores indicating greater perceived procedural fairness.

Personal factors

Four variables were measured to assess personal factors relevant to adherence: participants’ (1) trust in science, (2) their trust in traditional media, (3) their impulsivity, and (4) the negative emotions that they experience as a result of the pandemic.

Trust in science was measured by means of four items [88]. Participants indicated to which extent they trusted scientists to (1) “create knowledge that is unbiased and accurate,” (2) “create knowledge that is useful,” (3) “advise government officials on policy,” and (4) “inform the public on important issues” (1 = completely distrust, 5 = completely trust). Their answers were mean-scored into a scale measure (Survey 1: α = .92; Survey 2: α = .92; Survey 3: α = .92), with higher scores indicating greater trust in science.

Trust in media was assessed by means of a single item [5]: “Please indicate how much you trust traditional media (e.g., newspapers, TV news, news apps) to be unbiased and accurate”(1 = completely distrust, 5 = completely trust).

Impulsivity was measured by means of a subset of five items taken from the 8-item impulse control subscale from the Weinberger Adjustment Inventory (WAI; [89]): (1) “I should try harder to control myself when I’m having fun,” (2) “I do things without giving them enough thought,” (3) “When I’m doing something fun (like partying or acting silly), I tend to get carried away and go too far,” (4) “I say the first thing that comes to my mind without thinking enough about it,” and (5) “I stop and think things through before I act” (1 = “false,” 5 = “true;” last item reverse coded). The last item correlated poorly with the other items, and hence was eliminated. The remaining four items were combined into a scale measure (Survey 1: α = .82; Survey 2: α = .81; Survey 3: α = .82), with higher scores indicating greater impulsivity.

Negative emotional state due to COVID-19 was assessed by means of six items. Participants indicated to what extent the coronavirus made them feel (1) “angry,” (2) “scared,” (3) “powerless,” (4) “depressed,” (5) “stressed,” and (6) “lonely” (1 = “strongly disagree,” 7 = “strongly agree”). Responses were aggregated into a scale measure (Survey 1: α = .89; Survey 2: α = .91; Survey 3: α = .90), with higher scores indicating more negative emotions.

Social environment

To capture influences from the social environment, one variable was measured: perceived (descriptive) social norms for adhering to social distancing measures.

Perceived descriptive social norms regarding safe-distancing measures were measured by means of seven items, based on our measure of reported adherence. Participants were asked whether most people they knew were keeping a safe distance (six feet or more) from: (1) “others outside of their direct household,” (2) “their neighbors,” (3) “colleagues at work,” (4) “friends and family from outside of their direct household,” (5) “others when grocery shopping,” (6) “others when taking a walk or exercising,” and (7) “others in traffic or public transport” (1 = “strongly disagree,” 7 = “strongly agree”). Participants’ answers were combined into a scale measure (Survey 1: α = .94; Survey 2: α = .95; Survey 3: α = .95), with higher scores indicating greater perceived descriptive social norms for adhering.

Practical circumstances

To assess practical circumstances, two variables were measured: (1) participants’ practical capacity to adhere to social distancing measures, and (2) their perceived opportunity to violate those measures.

Participants’ practical capacity to adhere to social distancing mitigation measures was measured by means of seven items, again based on our measures of reported adherence. Participants were asked whether they were capable of keeping a safe distance (six feet or more) from: (1) “others outside of my direct household,” (2) “my neighbors,” (3) “colleagues at work,” (4) “friends and family from outside of my direct household,” (5) “others when grocery shopping,” (6) “others when taking a walk or exercising,” and (7) “others in traffic or public transport” (1 = “strongly disagree,” 7 = “strongly agree”). Responses were mean-scored into a single scale measure (Survey 1: α = .87; Survey 2: α = .85; Survey 3: α = .89), with higher scores indicating greater practical capacity to adhere.

Opportunity to violate social distancing measures was measured by means of seven items (again based on our measures of adherence). Participants were asked whether, at the present time, it was still possible for them to come within an unsafe distance (closer than six feet) from: (1) “others outside of my direct household,” (2) “my neighbors,” (3) “colleagues at work,” (4) “friends and family from outside of my direct household,” (5) “others when grocery shopping,” (6) “others when taking a walk or exercising,” and (7) “others in traffic or public transport” (1 = “strongly disagree,” 7 = “strongly agree”). Responses were aggregated into a single scale measure (Survey 1: α = .94; Survey 2: α = .94; Survey 3: α = .94), with higher scores indicating greater practical opportunity to violate.

Descriptive statistics of all independent variables are displayed for all three samples in Table 3, and correlations are shown in S4S6 Tables.

thumbnail
Table 3. Descriptive statistics of independent variables, Surveys 1 (May), 2 (June), and 3 (July), and full sample.

https://doi.org/10.1371/journal.pone.0257945.t003

Analysis plan

Our research focused on five major questions: (1) To what extent have Americans adhered to social distancing measures in the period after the first wave lockdown, between May and July 2020, (2) how have the various predictors that were hypothesized to influence adherence developed during this period, (3) which of these predictors in fact influenced adherence during this period, (4) how has the influence of these predictors on adherence changed across this period, and (5) how do the in- and decreases in the level of these predictors that occurred during this period explain the observed changes in adherence? Accordingly, our analysis consisted of three steps.

To examine the first two questions, we explored how adherence to social distancing measures, as well as the situational and motivational variables that were hypothesized to sustain it, evolved from May to July. To do so, we compare these variables between the three survey waves by means of analyses of covariance (ANCOVA), with parameter estimates with robust standard errors (HC3) to conduct pairwise comparisons between months. To illuminate the strictness with which individuals adhere to social distancing recommendations, we also compare frequencies of full adherence. This approach exploits the notion that anyone who reports anything less than full adherence (7 = “always”) in fact admits to not having followed the measures (either occasionally or more frequently); this therefore represents a stricter measure of adherence than the average. We compared the frequency of full adherence (across all seven situations) between survey waves using negative binomial regression; to compare the probability of full adherence within specific situations, logistic regression was utilized. All analyses controlled for all demographic and control variables.

To answer the third question, we examined how adherence to social distancing measures was predicted by the various predictors that were hypothesized to sustain it. To do so, we relied on linear (OLS) regression analyses, in which self-reported adherence to social distancing measures was regressed upon these variables (for a similar approach, see [5]). We estimated a hierarchical model in which the different categories of predictors were added to the model in iterated steps. To examine the fourth question, we reran the final iteration of the model expanded with an interaction term, between one of the predictors and survey wave. Separate models were estimated to test the interaction with survey wave for each of the predictors. All analyses were adjusted for heteroscedasticity using Huber/White robust standard error estimation.

Finally, to examine the fifth question, mediation analyses were conducted. These tested how the effect of survey wave on adherence was explained by its indirect effect on the key predictors that were identified in the final step of the hierarchical regression model.

Results

Development of adherence levels, May to July

First, we examined how Americans’ relative levels of adherence to social distancing measures developed from May to July by comparing average adherence levels between the surveys.

Average adherence.

Adherence levels on average as well as by situation are displayed in Fig 1. ANCOVA using parameter estimates with robust standard errors indicated that average levels of adherence among Americans declined from May to June (b = -.23, robust SE = .05, p < .001, Cohen’s d = .15), but did not change further from June to July (b = -.01, robust SE = .06, p = .797, Cohen’s d = .00). When separating the seven situations, adherence declined from May to June in all situations (outside household: b = -.19, robust SE = .07, p = .004, Cohen’s d = .11; neighbors: b = -.29, robust SE = .07, p < .001, Cohen’s d = .15; colleagues: b = -.29, robust SE = .08, p < .001, Cohen’s d = .13; friends and family: b = -.41, robust SE = .08, p < .001, Cohen’s d = .20; grocery shopping: b = -.13, robust SE = .06, p = .029, Cohen’s d = .09; walk or exercise: b = -.18, robust SE = .07, p = .005, Cohen’s d = .11; commute or travel: b = -.20, robust SE = .07, p = .003, Cohen’s d = .11). From June to July, however, no further significant changes in adherence were observed in any of the situations (all ps ≥ .269). In sum, the findings suggest a pattern where adherence to social distancing measures declined from May to June (although differences were relatively modest in terms of effect size), but not further in July.

thumbnail
Fig 1. Adherence to social distancing measures, Survey 1 (May) to Survey 3 (July).

https://doi.org/10.1371/journal.pone.0257945.g001

Full adherence.

Levels of full adherence are displayed in Fig 2. It displays the percentage of participants who reported adhering fully (7 = “always”) in each situation (grey and black lines), as well the average percentage of full adherence across all situations (red dashed line). Moreover, it displays the percentage of participants who reported full adherence in all seven situations (red solid lines). When comparing levels of full adherence averaged across all seven situations (red dashed line), negative binomial regression revealed a significant difference between the three survey waves, Wald χ2 (2) = 13.45, p = .001. Average levels of full adherence declined by from May to June (b = -.15, SE = .04, Wald χ2 (1) = 12.16, p < .001 –a reduction of 14.4% relative to May), but did not change further from June to July (b = -.03, SE = .04, Wald χ2 (1) = 0.57, p = .450). When comparing the number of participants who reported full adherence in every situation (red solid line), there also was a significant decrease from May to June (b = -.27, SE = .10, Wald χ2 (1) = 6.63, p = .01 –a reduction of 15.0% relative to May). Here also, no further changes were observed from June to July (b = -.02, SE = .11, Wald χ2 (1) = 0.05, p = .825).

thumbnail
Fig 2. Full adherence by situation, across all situations, and in every situation, Survey 1 (May) to Survey 3 (July).

https://doi.org/10.1371/journal.pone.0257945.g002

When separating the seven situations (grey and black lines), logistic regression indicated that the probability that participants fully adhered to social distancing recommendation declined significantly from May to June in all situations (outside household: b = -.23, SE = .09, Wald χ2 (1) = 6.21, p = .013; neighbors: b = -.39, SE = .09, Wald χ2 (1) = 17.26, p < .001; colleagues: b = -.39, SE = .09, Wald χ2 (1) = 17.26, p < .001; friends and family: b = -.32, SE = .09, Wald χ2 (1) = 11.22, p = .001; grocery shopping: b = -.22, SE = .09, Wald χ2 (1) = 5.78, p = .016; walk or exercise: b = -.27, SE = .09, Wald χ2 (1) = 8.20, p = .004; commute or travel: b = -.37, SE = .09, Wald χ2 (1) = 15.49, p < .001). From June to July, however, probabilities of full adherence did not change any further (all ps ≥ .116).

Development of predictor variables, May to July

Practical knowledge and understanding.

Fig 3 displays the development of participants’ knowledge of social distancing measures across the three surveys, as well as that of their perceptions of the clarity of those measures. Logistic regression indicated that levels of knowledge of social distancing measures (Table 3) declined significantly in June (b = -.71, SE = .14, Wald χ2 (1) = 26.03, p < .001), but partially recovered in July (b = .28, SE = .13, Wald χ2 (1) = 4.48, p = .034). Furthermore, ANCOVA using parameter estimates with robust standard errors indicated that relative to May, the perceived clarity of mitigation measures was significantly lower in July (b = -.34, robust SE = .08, p < .001, Cohen’s d = .17).

thumbnail
Fig 3. Practical knowledge and understanding, Survey 1 (May) to Survey 3 (July).

https://doi.org/10.1371/journal.pone.0257945.g003

Costs and benefits.

Fig 4 displays the development of the variables reflecting costs and benefits of mitigation measures across the three surveys. Threat perceptions did not change significantly between May and June (p = .073), but increased significantly from June to July (b = .20, robust SE = .07, p = .002, Cohen’s d = .11). Conversely, reported personal costs of mitigation measures decreased from May to June (b = -.22, robust SE = .07, p = .002, Cohen’s d = .11), as did perceptions of the certainty of punishment (b = -.24, robust SE = .07, p = .002, Cohen’s d = .11); neither changed significantly thereafter (both ps ≥ .392). Perceptions of the severity of punishment did not change significantly between May and July (all ps ≥ .110).

thumbnail
Fig 4. Costs and benefits of mitigation measures, Survey 1 (May) to Survey 3 (July).

https://doi.org/10.1371/journal.pone.0257945.g004

Legitimacy, procedural justice, and obligation to obey.

Fig 5 displays the development of the variables reflecting the core constructs in this area. The analyses revealed that moral alignment with social distancing measures declined significantly from May to June (b = -.14, robust SE = .05, p = .011, Cohen’s d = .09), while evaluations of the authority response declined significantly from June to July (b = -.54, robust SE = .08, p < .001, Cohen’s d = .25). Furthermore, there was a significant decline from May to June in participants’ normative obligation to obey the authorities handling COVID-19 (b = -.15, robust SE = .04, p < .001, Cohen’s d = .14), and in perceptions of their procedural fairness (b = -.21, robust SE = .07, p = .003, Cohen’s d = .11). No significant changes were observed in non-normative obligation to obey these authorities, however, or in their general obligation to obey the law (all ps ≥ .214).

thumbnail
Fig 5. Legitimacy variables, Survey 1 (May) to Survey 3 (July).

https://doi.org/10.1371/journal.pone.0257945.g005

Personal factors.

Fig 6 shows the development of personal factors relevant to adherence. The results revealed small, but significant changes in trust in science and media: trust in science decreased significantly from May to June (b = -.09, robust SE = .04, p = .032, Cohen’s d = .09), whereas trust in mainstream media showed a significant decrease from May to July (b = -.12, robust SE = .06, p = .030, Cohen’s d = .09). No significant changes were observed in impulsivity (all ps ≥ .092) or negative emotions (all ps ≥ .554).

thumbnail
Fig 6. Personal factors, Survey 1 (May) to Survey 3 (July).

https://doi.org/10.1371/journal.pone.0257945.g006

Social environment.

Fig 7 shows the development of perceived (descriptive) social norms for adhering to social distancing measures. From May to July, perceived social norms for keeping a safe distance were significantly reduced (b = -.38, robust SE = .06, p < .001, Cohen’s d = .23).

thumbnail
Fig 7. Social environment, Survey 1 (May) to Survey 3 (July).

https://doi.org/10.1371/journal.pone.0257945.g007

Practical circumstances.

Finally, Fig 8 displays the development in practical circumstances for adhering. From May to July, there was a significant decrease in respondents’ reported capacity to adhere to social distancing measures (b = -.14, robust SE = .05, p = .002, Cohen’s d = .11). Conversely, perceived opportunities for violate social distancing measures became significantly greater from May to June (b = .20, robust SE = .08, p = .013, Cohen’s d = .09).

thumbnail
Fig 8. Practical circumstances, Survey 1 (May) to Survey 3 (July).

https://doi.org/10.1371/journal.pone.0257945.g008

Understanding adherence to social distancing measures from May to July

Hierarchical regression model.

As the previous section demonstrates, adherence to social distancing measures declined significantly in the period after the initial first wave lockdown. At the same time, significant changes were observed in many of the variables that were hypothesized to shape adherence. Our next major question is to understand how these processes shaped adherence to social distancing measures during this period. To do so, we estimated a linear regression model, in which adherence was regressed upon the various predictors in a series of hierarchical steps. This model was estimated using the combined data from all three survey waves (N = 2,919), with survey wave included as an additional predictor (1 = May, 2 = June, 3 = July). Collinearity statistics indicated no issues with multicollinearity (all VIFs ≤ 2.55; all tolerances ≥ .39). Table 4 displays the results.

thumbnail
Table 4. Hierarchical linear regression (with robust standard errors), adherence to mitigation measures by predictor and control variables.

https://doi.org/10.1371/journal.pone.0257945.t004

In Step 1, the model included only the survey wave dummies and the control variables. As shown in Table 4 (column 1), relative to May, adherence decreased significantly in June and in July. Adherence was significantly higher among older participants, female participants, minority group members, people with higher levels of education, and people with higher socio-economic status. Furthermore, adherence was greater among people who suffered from a health condition that placed them at increased risk, or who knew others who suffered from such health conditions. Conversely, adherence was significantly lower among people who professionally cared for COVID patients, and among participants with more conservative political orientations. Lastly, relative to participants from the Northeast region, adherence was significantly lower among participants from the Midwest and the South regions. Together, the model explained only 8% of the variability in adherence, however.

In Step 2, predictors reflecting participants’ practical knowledge and understanding of the mitigation measures were added to the model. Adherence was significantly greater among participants who had greater knowledge of social distancing measures, and among participants who regarded these measures as more clear. Inclusion of these predictors meant that the effect of health risk to others was no longer significant. The model now explained 14% of the variance in adherence, a significant increase relative to Step 1 (χ2 (2) = 223.87, p < .001).

In Step 3, predictors related to the costs and benefits of adherence were entered into the model. Adherence was significantly greater among participants and who regarded the COVID-19 pandemic as more threatening. Also, adherence was greater among participants for whom the costs of adhering was higher. Conversely, punishment perceptions did not predict adherence. Inclusion of these variables rendered the effect of political orientation nonsignificant. The model explained 31% of the variance in adherence, a significant increase over Step 2 (χ2 (4) = 622.75, p < .001).

In Step 4, predictors reflecting participants’ legitimacy perceptions were added to the model. Adherence was significantly higher among participants who morally agreed more with the measures (i.e., moral alignment), who felt greater normative obligation to obey the COVID-19 authorities, and who felt a higher general obligation to obey the law. Inclusion of these variables rendered nonsignificant the effect of perceived clarity of mitigation measures, socio-economic status, and Midwest region. The model explained 39% of the variance in adherence, a significant increase over Step 3 (χ2 (5) = 383.19, p < .001).

In Step 5, personal factors were entered into the model. Adherence was significantly higher among participants who had greater trust in science. Conversely, adherence was significantly lower among more impulsive participants. Controlling for these variables rendered nonsignificant the effect of general obligation to obey the law, and revealed a significant effect of undisclosed political orientation, which predicted greater adherence (relative to liberals). The model explained 40% of the variance in adherence, a significant increase over Step 4 (χ2 (5) = 46.47, p < .001).

In Step 6, predictors reflecting the social environment were added to the model. Adherence was significantly higher among people who perceived stronger (descriptive) social norms for keeping a safe distance. Inclusion of this variable rendered nonsignificant the effect of South region. Additionally, the decline in adherence from May to June was reduced to nonsignificance when this variable was included. The model explained 42% of the variance in adherence, a significant increase over Step 5 (χ2 (1) = 103.62, p < .001).

Finally, in Step 7, predictors reflecting the practical circumstances were entered into the model. Adherence was significantly higher among people who had greater practical ability to keep at a safe distance from others. In contrast, adherence was significantly lower among people who saw more opportunities for violating social distancing measures. By including these variables in the model, the effects of trust in science, normative obligation to obey, and care for COVID patients were reduced to nonsignificance. Conversely, inclusion of these variables restored to significance the previously observed effect of South region, and the decline in adherence from May to June. Last, inclusion of these variables revealed a significant effect of private insurance, such that adherence was greater among participants who had private (relative to no) insurance. The final model explained 52% of the variance in adherence, a significant increase over Step 6 (χ2 (2) = 532.61, p < .001).

Change in predictors across waves.

In addition, we sought to understand how the effect of these predictors on adherence changed across survey waves. To do so, we estimated additional models based on the final step of the hierarchical regression model (step 7, Table 4). Each of these models included all the predictors and control variables from the hierarchical model (as in step 7), and one single interaction term, between survey wave and one of the 19 predictors (respectively knowledge of measures, clarity of measures, perceived health threat, personal costs, punishment certainty, punishment severity, moral alignment, authority response, normative obligation to obey, non-normative obligation to obey, general obligation to obey the law, procedural justice of enforcement, trust in science, trust in media, impulsivity, negative emotions, descriptive social norms, practical capacity to adhere, or opportunity to violate). In total, 19 interaction models therefore were estimated, each including a single interaction term. Because our interest with these models was exclusively in the interactive effects with survey waves, only the interaction terms are displayed in Table 5, for each of the 19 models.

thumbnail
Table 5. Interaction models: Interaction effects on adherence to mitigation measures of predictors by survey waves.

https://doi.org/10.1371/journal.pone.0257945.t005

The results of these analyses indicated that largely, the effect of the predictors on adherence did not vary across waves. Of the key predictors of adherence in the final hierarchical model (i.e., knowledge, perceived threat, personal costs, moral alignment, impulsivity, descriptive social norms, practical capacity to adhere, and opportunity to violate, see Table 4, step 7), only the effect of impulsivity varied across survey waves (in May: b = -.13, SE = .03, p < .001; in July: b = -.06, SE = .03, p = .049; contrast = .07, SE = .04, p = .049). The results did indicate changes across waves in the effects of normative obligation to obey (in May: b = -.06, SE = .04, p = .098; in July: b = .08, SE = .04, p = .066; contrast = .13, SE = .05, p = .006), non-normative obligation to obey (in May: b = -.07, SE = .03, p = .043; in July: b = .10, SE = .03, p = .001; contrast = .17, SE = .04, p < .001), trust in science (in May: b = -.04, SE = .03, p = .181; in July: b = .10, SE = .04, p = .008; contrast = .14, SE = .04, p = .001), and trust in media (in May: b = -.06, SE = .02, p = .011; in July: b = .02, SE = .02, p = .511; contrast = .07, SE = .03, p = .021). Thus, impulsivity became gradually less important as a predictor of adherence across waves, while particularly non-normative obligation to obey and trust in science became more influential. These changes were modest in terms of effect size, however.

Although the effect of the key predictors of adherence generally did not vary across waves (i.e., they did not become more or less predictive of adherence), the absolute levels of these variables did change significantly between May and July. Indeed, as was previously shown in the descriptive analyses of changes across survey waves, there were significant changes during this period in participants’ reported knowledge of mitigation measures, their perceptions of the threat of the virus, their personal costs of the mitigation measures, their moral alignment with those measures, their perceived social norms, their practical capacity to comply, and their perceived opportunities for violating social distancing measures. To examine how these changes contributed to the observed decline in compliance across this period, we finally conducted mediation analyses. These tested whether the effect of survey wave on compliance was mediated by the effect of survey wave on each of the key predictors of adherence.

To do so, we relied on the PARAMED module in Stata [90], which can handle both linear and categorical mediators. In these models, survey wave was the independent variable, adherence the dependent variable, and the mediator was either reported knowledge of mitigation measures, perceptions of the threat of the virus, personal costs of the mitigation measures, moral alignment with mitigation measures, perceived social norms, practical capacity to comply, or perceived opportunities for violating. The models controlled for all other predictors and control variables, and featured bias-corrected bootstrap confidence intervals (1,000 replications). In total, 14 mediation models were estimated (for 7 mediators, with two models each: one comparing wave 1 to wave 2, and one comparing wave 1 to wave 3). Results are presented in Table 6.

thumbnail
Table 6. Mediation models: Total, direct, and indirect effects per mediator by survey waves.

https://doi.org/10.1371/journal.pone.0257945.t006

The indirect effects reported in Table 6 suggest that the effect of survey wave on adherence (the total effect) was significantly reduced, and thus partially mediated (i.e., confidence interval of the indirect effect does not include zero) by the following variables: knowledge of mitigation measures, perceived health threat, moral alignment, and social norms. Conversely, personal costs and capacity to adhere did not mediate this effect. Accordingly, these findings suggest that the observed decrease in adherence from May to July was partially explained by their lower knowledge of mitigation measures, by reductions in the perceived health threat of COVID-19, by people’s lower alignment with social distancing measures, and by reduced (descriptive) social norms for keeping distance. The notion that the personal cost of mitigation measures decreased during this period, and that people became more practically capable of adhering to these, did not counter these trends.

Discussion

The results of our study show that a broad range of behavioral mechanisms has been at play in shaping adherence to pandemic mitigation measures in the period that followed the first wave lockdown against COVID-19. In the period after stricter mitigation measures were repealed, during the summer months of 2020, a significant decline in adherence was observed. Across this period, adherence to social distancing measures was shaped by a range of factors, relating to people’s practical knowledge and understanding of mitigation measures, their perceptions of their costs and benefits, their perceptions of legitimacy and procedural justice, their personal factors, their social environment, and their practical circumstances. Moreover, changes in the levels of these factors during this period explained (in part) the observed decline in adherence. These findings demonstrate that large-scale behavioral change can be accomplished through a combination of factors situated at different levels. Yet, the study also shows that some variables that have received much attention in general psychological, economic, and criminological compliance scholarship did not play a clear and consistent role in shaping adherence.

Across the different steps of our analysis, eight variables emerged as consistent predictors of adherence. Respondents adhered more when (1) they had greater knowledge of social distancing measures, (2) they perceived the virus as a more severe health threat, (3) adherence was more costly for them (possibly reflecting the reverse: that costs were higher for those who adhered more), (4) they morally agreed more with the measures, (5) they were low in impulsivity, (6) they perceived stronger (descriptive) social norms for keeping a safe distance, (7) they had greater practical ability to adhere, and (8) they perceived fewer opportunities for violating the measures. When examining their effect sizes in the final step of the regression model, however, it becomes clear that especially capacity had a critical impact on respondents’ adherence (according to Cohen’s standards, a large effect). Moral alignment and perceived threat also had a substantial, but smaller impact on adherence (according to Cohen’s standards, a small to medium effect). The impact of impulsivity, knowledge, opportunity for violating, personal costs, and social norms was only limited, however (according to Cohen’s standards, a small effect).

The impact of the predictors on adherence was largely consistent throughout this period, although the influence of impulsivity became gradually weaker (and the influence of non-normative obligation to obey and trust in science gradually stronger) as the distance from the lockdown period increased. The decline that occurred across this period in levels of knowledge, moral alignment, and perceived social norms for adhering partially explained the observed decrease in adherence. Conversely, the increase in perceived threat that was observed toward the end of this period positively affected the development of adherence. Other variables, however, failed to predict adherence, or no longer did so when other variables were taken into account. These most notably included procedural justice [40, 41], obligation to obey the law or the responsible authorities [46, 47], deterrence [3335], and trust in science [50, 51].

Theoretically, the present comparison of adherence over the summer months demonstrates that the nature of behavioral change and influence on behavior is not static. Rather, our findings show that across similar samples of people, with similar measures staying in place, key factors that sustain compliance can grow or decline even in a matter of months. Our data allow us to trace these processes more deeply by examining how the key predictors have changed over the summer months. Although the influence of these variables on adherence was largely consistent throughout this period, the data revealed significant changes in their absolute levels. People reported, for instance, to have more opportunity to violate the social distancing measures (which makes sense given that stay-at-home orders were mostly lifted in this period), lower capacity to adhere, and lower perceived social norms for adherence (consistent with the notion that there were larger crowds, and that more people were expected to resume normal work and social activities). Our mediation analyses revealed that the observed decline in adherence to social distancing measures that was observed during this period was partially explained by the decreases in people’s knowledge, moral alignment, and perceived social norms for adhering. Conversely, the increase in perceived threat that was observed toward the end of this period positively affected the development of adherence. When viewed together, these changes provide important indications of why adherence has changed over time. These processes do not seem to indicate that there was a so-called general behavioral fatigue [91, 92] at play at this time, but rather that lower adherence may have resulted from very particular and factual changes in people’s circumstances, the environment, and their motivations. By providing insight into which variables do (and do not) shape adherence, the present research offers a more practical way of assessing whether people are able to sustain behavioral change for as long as needed, compared to broad and vague concepts such as behavioral fatigue (which rely more on common-sense understanding than mechanisms from behavioral science). An important question for future research, however, is to understand more deeply how the changes that we observed across this period may be connected to local developments in policy, society, and the pandemic (e.g., see [93]). For this, a more fine-grained analysis is needed, which takes into account how these processes developed locally at the level of regions, states, counties, or even cities.

Our findings on deterrence deserve extra discussion. In light of the fact that stricter mitigation measures have been repealed, and thus are no longer widely enforced [94], it is noteworthy that Americans nevertheless reported moderately high levels (i.e., close to the scale midpoint) of perceived punishment certainty and severity. One explanation for such continuing perceptions of deterrence when there is no longer any enforcement is that there are spill-over effects. In this case, this might mean that prior enforcement continues to drive deterrence perceptions even after it has ended, or that enforcement of other measures (e.g., facemasks; quarantine) also shapes deterrence perceptions for social distancing [95]. A second, and related explanation is that people generally do not have very good perceptions of deterrence and can underestimate or overestimate both the certainty and severity of punishment [96]. Importantly, however, even though many Americans considered it quite likely that they would be punished when not keeping a safe distance, and regarded such punishment as quite severe, these beliefs did not predict greater adherence. This finding is in line with studies in other countries where there was actual enforcement of social distancing measures, where also no effects of deterrence on compliance were observed [18]. However, these conclusions clearly oppose belief in the effectiveness of strong punishment for COVID-19 violations [97, 98].

Clearly, the data allow for the exploration of many other relationships beyond those that we study in the present manuscript. For example, the data can inform about relationship between adherence and political orientation or trust in science (both singled out as important predictors of adherence in prior research [50, 51, 99, 100], yet neither a significant predictor in our final regression model), or demographic factors like ethnicity or socio-economic status. From the results of the hierarchical regression analysis, it seems plausible that these and other factors may have indirect relationships with adherence, through their effects on more proximal predictors. The data further could illuminate how specific subsets of predictors may interact with each other, or could be used to study other outcome variables (e.g., how these predictors may explain felt negative emotions, or support for authorities, etc.). The present research was primarily oriented on understanding the proximal predictors of adherence. For this reason, we feel that other relationships, such as those outlined above, are best reserved for dedicated manuscripts that are specifically oriented on these questions. We welcome further analyses of these questions, and have made our data publicly available for this purpose. Future research could also expand on these findings by zooming in further on specific variables that may directly or indirectly shape compliance (e.g., by distinguishing essential and nonessential work; by separating individuals from different generations [101]), or by identifying further variables with which our model could be expanded.

Our findings have several policy implications, which may aid authorities in the U.S. and elsewhere to sustain adherence with mitigation measures, both for the current outbreak and for future pandemics. The results of our surveys identify seven factors that influence adherence. We formulate recommendations based on the most influential of these.

First, and most critically, authorities can increase adherence by making it practically easier for citizens to do so, and by removing opportunities to offend. Indeed, in terms of effect size, people’s practical capacity to adhere was the strongest predictor of adherence, by some margin. This suggests that authorities can have an important impact on adherence by increasing citizens’ practical capacity to do so. In context of social distancing, this has included arrangements that guide crowds through public venues in ways that keep them apart as much as possible, facilitating telework where possible, instituting caps on the number of people able to enter a public space, and so forth. Conversely, authorities can also shape adherence by removing practical opportunities for not following mitigation measures. Such measures are best reserved for especially harmful offenses that are widely condemned, however, because if these are not widely supported, overly coercive measures may strongly undermine citizens’ motivation [102].

Second, our results show that individuals adhere more when they morally agree with mitigation measures. This finding suggests that authorities can increase adherence if they can effectively convince citizens of the importance and legitimacy of such measures. In the case of social distancing, this has included presenting evidence of how social distancing measures can prevent the spread of the virus, or emphasizing citizens’ shared moral duty to protect vulnerable individuals. Cultivating citizens’ support–or conversely, attuning mitigation measures to what is widely supported–will increase the chance that citizens will effectively adhere to such measures.

Third, perceptions of threat to oneself and others are an important predictor of adherence to mitigation measures. However, the present findings also demonstrate that threat perceptions are dynamic. Here, threat perceptions increased from June to July, reflecting the increase in infections that occurred during this period [6]. Findings from our studies in the Netherlands [18], however, demonstrate that threat perceptions can quickly recede as infections decline, with deleterious effects on support for, and adherence to, mitigation measures. Accordingly, to sustain adherence to mitigation measures, it is important that authorities do not give the impression that the threat is waning once infections recede [103]. Rather, authorities can sustain adherence if they successfully convince citizens of the continuing threat of the pandemic, for example to themselves or vulnerable others.

Fourth, our findings show that knowledge of mitigation measures is important for adherence. Due to the fragmented authority response in the U.S., mitigation measures may differ substantially between states, counties, and municipalities. Consequently, it can be unclear to citizens what mitigation measures require of them. Accordingly, authorities can promote adherence by clearly communicating what the measures are and what they require of citizens.

Finally, our findings demonstrate that people’s adherence to mitigation measures is influenced by the behavior of others in their community (i.e., descriptive social norms). Although this effect was modest in terms of effect size in the final regression model, effects of social norms on social distancing have also been demonstrated in other research [18, 104, 105]. Authorities thus can enhance adherence to mitigation measures by demonstrating that adherence is common and widely approved of. This also means, however, that authorities should take care to not convey the impression that violations are ubiquitous and normal. It is plausible that in the period after the initial lockdown, highly publicized instances where people widely disregarded social distancing measures may have undermined adherence, by normalizing lack of distancing. To promote adherence to mitigation measures, authorities therefore should express that doing so is the norm (or ought to be), highlight examples where many others are seen to adhere, not draw undue attention to examples where people do not, and to ensure that they are always seen to adhere to the measures themselves.

Overall, the study of adherence of social distancing measures has important implications for the study of compliance generally and the way rules shape human behavior. These questions have been studied across different academic domains, and with a focus on different mechanisms and interventions [106]. This has resulted in a patchwork of theories and approaches that are seldomly brought together, which exist in compartmentalized silos that draw on their own literatures, methods and findings. The present study brings together a broad range of variables from across these approaches, situated at different levels (i.e., the individual, the social environment, the practical circumstances), and reveals how these together shape adherence in context of social distancing measures. Although the associations that were observed here may not extend beyond this setting, the insight that adherence derived from such a diverse range of influences is nevertheless important for study of compliance. It underlines that to better understand why people comply, research can benefit from a multi-theoretical approach, in which the extant, siloed literatures are brought together and integrated.

Our study has several limitations. First, although our samples were large and stratified sampled by age, gender, and ethnicity to mimic the demographic characteristics of the United States population based on U.S. Census Bureau data, they remain non-probability convenience samples. Furthermore, there was some variability between the samples in terms of demographics, possibly due to the considerable subset of participants who failed to complete the survey or pass the attention checks. As a consequence, our samples cannot be regarded as truly nationally representative. Nevertheless, there is evidence that such convenience samples can be as accurate as random digit dial telephone surveys [107, 108], and they may reduce social desirability biases [109]. Further research therefore is needed to understand the robustness of the observed findings, although they align with evidence from other research [3]. Second, our surveys rely on self-reported measures that may be subject to imperfect recall or social desirability bias [110, 111]. We do note, however, that a recent study demonstrated that social desirability bias did not inflate the estimates of compliance with COVID-19 measures in online surveys [112], and that the finding of high self-reported adherence is in line with objective data from Google COVID19 Community Mobility [113]. Furthermore, prior research shows that there can be strong concordance between self-reported and objective compliance measures when surveys are used (see [78] p. 29). Even so, future research into these questions would benefit from methods that supplement self-reported measures with behavioral data, such as video observation [114].

Conclusion

In the summer of 2020, the Federal lockdown and stay-at-home measures against COVID-19 that were in force in spring were lifted, and in large parts of the country, society began to reopen. The present findings, based on three stratified samples collected in May, June, and July, show that Americans’ adherence to social distancing measures declined, as did several of the factors that sustained it–including people’s practical capacity to adhere, their knowledge of the measures, and social norms for adherence. Our research identifies key variables that predicted greater adherence as society reopened, and which contributed to the changes in adherence that were observed thereafter. By doing so, this research contributes to the understanding of pandemic governance and the interaction between rules and human conduct more generally. Moreover, in the current stage of the pandemic, these findings provide important directions for the public health response, by highlighting processes through which adherence to mitigation measures can be promoted, as we strive to return to normality.

Supporting information

S1 Table. Kendall’s tau correlations between demographic variables and adherence.

May 8–18 (Survey 1. N = 1012). Note. *–Correlation is significant at the .05 level. **–Correlation is significant at the .01 level. Gender–Female as reference category. Political orientation–N = 866.

https://doi.org/10.1371/journal.pone.0257945.s003

(DOCX)

S2 Table. Kendall’s tau correlations between demographic variables and adherence.

June 8–16 (Survey 2. N = 986). Note. *–Correlation is significant at the .05 level. **–Correlation is significant at the .01 level. Gender–Female as reference category. Political orientation–N = 880.  

https://doi.org/10.1371/journal.pone.0257945.s004

(DOCX)

S3 Table. Kendall’s tau correlations between demographic variables and adherence.

July 11–17 (Survey 3. N = 921). Note. *–Correlation is significant at the .05 level. **–Correlation is significant at the .01 level. Gender–Female as reference category. Political orientation–N = 803.

https://doi.org/10.1371/journal.pone.0257945.s005

(DOCX)

S4 Table. Kendall’s tau correlations between independent variables and adherence.

May 8–18 (Survey 1. N = 1012). Note. *–Correlation is significant at the .05 level. **–Correlation is significant at the .01 level.

https://doi.org/10.1371/journal.pone.0257945.s006

(DOCX)

S5 Table. Kendall’s tau correlations between independent variables and adherence.

June 8–16 (Survey 2. N = 986). Note. *–Correlation is significant at the .05 level. **–Correlation is significant at the .01 level.

https://doi.org/10.1371/journal.pone.0257945.s007

(DOCX)

S6 Table. Kendall’s tau correlations between independent variables and adherence.

July 11–17 (Survey 3. N = 921). Note. *–Correlation is significant at the .05 level. **–Correlation is significant at the .01 level.

https://doi.org/10.1371/journal.pone.0257945.s008

(DOCX)

S1 Output. Complete regression output Survey 1 (May).

https://doi.org/10.1371/journal.pone.0257945.s009

(PDF)

S2 Output. Complete regression output Survey 2 (June).

https://doi.org/10.1371/journal.pone.0257945.s010

(PDF)

S3 Output. Complete regression output Survey 3 (July).

https://doi.org/10.1371/journal.pone.0257945.s011

(PDF)

S4 Output. Complete regression output Surveys 1–3 combined (May-July).

https://doi.org/10.1371/journal.pone.0257945.s012

(PDF)

References

  1. 1. Ritchie H. Google Mobility Trends: How has the pandemic changed the movement of people around the world? [Internet]. Our World in Data; 2020 Jun 2 [cited 2020 Dec 8]. Available from: https://ourworldindata.org/covid-mobility-trends.
  2. 2. Walker A. US oild prices turn negative as demand dries up [Internet]. BBC News; 2020 Apr 20 [cited 2020 Dec 8]. Available from: https://www.bbc.com/news/business-52350082.
  3. 3. Kooistra EB, Van Rooij B. Pandemic compliance: A systematic review of influences on social distancing behavior during the first wave of the COVID-19 outbreak. PsyArXiv c5x2k [Preprint]. 2020 Nov 25 [cited 2020 Dec 8]. Available from: https://psyarxiv.com/c5x2k.
  4. 4. Haffajee RL, Mello MM. Thinking globally, acting locally—The US response to COVID-19. The New England Journal of Medicine. 2020 382(22):e75. pmid:32240580
  5. 5. Van Rooij B, de Bruijn AL, Reinders Folmer CP, Kooistra E, Kuiper ME, Brownlee M, et al. Compliance with COVID-19 mitigation measures in the United States. SSRN 3582626 [Preprint]. 2020 May 1 [cited 2020 Sep 11]. Available from: https://doi.org/10.2139/ssrn.3582626.
  6. 6. CDC COVID Data Tracker: The Centers for Disease Control and Prevention; 2020 [cited 2020 Sep 10]. Available from: https://covid.cdc.gov/covid-data-tracker/#trends.
  7. 7. Rojas R, Fausset R. Businesses Tiptoe Into a World of Masks, Gloves and Wary Customers [Internet]. The New York Times; 2020 Apr 24 [cited 2020 Sep 10]. Available from: https://www.nytimes.com/2020/04/24/us/coronavirus-georgia-oklahoma-alaska-reopen.html.
  8. 8. Freking K, Colvin J. Trump says he’s not extending social distancing guidelines [Internet]. Associated Press; 2020 Apr 30 [cited 2020 Oct 28]. Available from: https://abcnews.go.com/Politics/wireStory/trump-extending-social-distancing-guidelines-70419986.
  9. 9. See How All 50 States Are Reopening (and Closing Again) [Internet]. The New York Times; 2020 [cited 2020 Sep 10]. Available from: https://www.nytimes.com/interactive/2020/us/states-reopen-map-coronavirus.html.
  10. 10. Pitofsky M. Stylists ticketed for cutting hair on Michigan Capitol lawn to protest lockdown [Internet]. The Hill; 2020 May 21 [cited 2020 Dec 8]. Available from: https://thehill.com/blogs/blog-briefing-room/news/499031-stylists-ticketed-for-cutting-hair-on-michigan-capitol-lawn-to.
  11. 11. DeSantis M. Protesters In Commack Demand Economy Reopens [Internet]. Patch; 2020 May 1 [cited 2020 Dec 8]. Available from: https://patch.com/new-york/commack/protesters-commack-demand-economy-opens.
  12. 12. Hart PS, Chinn S, Soroka S. Politicization and polarization in COVID-19 news coverage. Science Communication. 2020 Oct 1;42(5):679–97.
  13. 13. Tyson A. Republicans remain far less likely than Democrats to view COVID-19 as a major threat to public health [Internet]. Pew Research Center; 2020 Jul 22 [cited 2020 Dec 8]. Available from: https://www.pewresearch.org/fact-tank/2020/07/22/republicans-remain-far-less-likely-than-democrats-to-view-covid-19-as-a-major-threat-to-public-health/.
  14. 14. Partlow J, Dawsey J. Workers removed thousands of social distancing stickers before Trump’s Tulsa rally, according to video and a person familiar with the set-up [Internet]. The Washington Post; 2020 Jun 27 [cited 2020 Sep 10]. Available from: https://www.washingtonpost.com/politics/workers-removed-thousands-of-social-distancing-stickers-before-trumps-tulsa-rally-according-to-video-and-a-person-familiar-with-the-set-up/2020/06/27/f429c3be-b801-11ea-9b0f-c797548c1154_story.html.
  15. 15. Feldman Y. The law of good people: Challenging states’ ability to regulate human behavior. Cambridge, UK: Cambridge University Press; 2018.
  16. 16. Friedman LM. Impact. Cambridge: Harvard University Press; 2016.
  17. 17. Van Rooij B, Sokol DD. Cambridge Handbook of Compliance. Cambridge, UK: Cambridge University Press; 2021.
  18. 18. Reinders Folmer CP, Kuiper ME, Olthuis E, Kooistra EB, De Bruijn AL, Brownlee M, et al. Maintaining compliance when the virus returns: Understanding adherence to COVID-19 social distancing measures in the Netherlands in July 2020. PsyArXiv vx3mn [Preprint]. 2020 Aug 28 [cited 2020 Sep 10]. Available from: https://psyarxiv.com/vx3mn/.
  19. 19. Darley JM, Robinson PH, Carlsmith KM. The ex ante function of the criminal law. Law and Society Review. 2001 35(1):165–90.
  20. 20. Kim PT. Norms, learning and law: Exploring the influences of workers’ legal knowledge. University of Illinois Legal Review. 1999 (2):447–516.
  21. 21. Van Rooij B. Do people know the law? Empirical evidence about legal knowledge and its implications for compliance. In: van Rooij B, Sokol DD, editors. Cambridge Handbook of Compliance. Cambridge, UK: Cambridge University Press; 2021. p. 467–88.
  22. 22. Feldman Y, Teichman D. Are all legal probabilities created equal. NYUL Review. 2009 84:980–1022.
  23. 23. Paternoster R, Simpson S. A rational choice theory of corporate crime. In: Clarke RV, Felson M, editors. Routine Activity and Rational Choice: Advances in Criminological Theory. New York: Routledge; 1993. p. 37–58.
  24. 24. Donovan JL, Blake DR. Patient non-compliance: deviance or reasoned decision-making? Social Science & Medicine. 1992 34(5):507–13.
  25. 25. Zhou F, Yu T, Du R, Fan G, Liu Y, Liu Z, et al. Clinical course and risk factors for mortality of adult inpatients with COVID-19 in Wuhan, China: a retrospective cohort study. The Lancet. 2020 395(10229):1054–62. pmid:32171076
  26. 26. Jung S-M, Akhmetzhanov AR, Hayashi K, Linton NM, Yang Y, Yuan B, et al. Real-time estimation of the risk of death from novel coronavirus (COVID-19) infection: Inference using exported cases. Journal of Clinical Medicine. 2020 9(2):523. pmid:32075152
  27. 27. Weiss P, Murdoch DR. Clinical course and mortality risk of severe COVID-19. The Lancet. 2020 395(10229):1014–5. pmid:32197108
  28. 28. Yıldırım M, Geçer E, Akgül Ö. The impacts of vulnerability, perceived risk, and fear on preventive behaviours against COVID-19. Psychology, Health & Medicine. 2021 26(1):35–43. pmid:32490689
  29. 29. Economic Attitudes as the Country Starts to Reopen [Internet]. AP NORC; 2020 [cited 2020 Dec 8]. Available from: https://apnorc.org/projects/economic-attitudes-as-the-country-starts-to-reopen/.
  30. 30. Hunter L, Pearl B, Lo K. Tracking enforcement measures for violation of stay-at-home orders [Internet]. Center for American Progress; 2020 Apr 2 [cited 2021 Jul 7]. Available from: https://www.americanprogress.org/issues/criminal-justice/news/2020/04/02/482568/tracking-enforcement-measures-violation-stay-home-orders/.
  31. 31. Janes C. Coughing ’attacks’ may be prosecuted as terrorism in a war on coronavirus [Internet]. The Washington Post; 2020 Apr 9 [cited 2021 Jul 7]. Available from: https://www.washingtonpost.com/national/health-science/coughing-attacks-may-be-prosecuted-as-terrorism-in-war-on-coronavirus/2020/04/08/b97d7f9a-790d-11ea-9bee-c5bf9d2e3288_story.html.
  32. 32. Grasmick HG, Bryjak GJ. The deterrent effect of perceived severity of punishment. Social Forces. 1980 59(2):471–91.
  33. 33. Becker GS. Crime and punishment: an economic approach. In: Fielding NG, Clarke A, Witt R, editors. The Economic Dimensions of Crime. 76. London: Palgrave Macmillan; 1968. p. 169–217.
  34. 34. Polinsky AM, Shavell S. Public Enforcement of Law. In: Bouckaert B, De Geest G, editors. Encyclopedia of Law and Economics, Volume V: The Economics of Crime and Litigation. Cheltenham: Edward Elgar; 2000. p. 307–44.
  35. 35. Shavell S. Specific versus general enforcement of law. Journal of Political Economy. 1991 99:1099–108.
  36. 36. Weber M. Economy and society: An outline of interpretive sociology. Berkeley, CA: University of California Press; 1978.
  37. 37. Jackson J, Gau JM. Carving Up Concepts? Differentiating Between Trust and Legitimacy in Public Attitudes Towards Legal Authority. In: Shockley E, Neal TMS, PytlikZillig LM, Bornstein BH, editors. Interdisciplinary Perspectives on Trust: Towards Theoretical and Methodological Integration. Cham: Springer International Publishing; 2016. p. 49–69.
  38. 38. Jackson J, Bradford B, Hough M, Myhill A, Quinton P, Tyler TR. Why do people comply with the law? Legitimacy and the influence of legal institutions. British Journal of Criminology. 2012 52(6):1051–71.
  39. 39. Jackson J, Huq AZ, Bradford B, Tyler TR. Monopolizing force? Police legitimacy and public attitudes toward the acceptability of violence. Psychology, Public Policy, and Law. 2013 19(4):479–97.
  40. 40. Tyler TR. Procedural fairness and compliance with the law. Swiss Journal of Economics and Statistics. 1997 133(2):219–40.
  41. 41. Tyler TR. Why People Obey the Law. Princeton: Princeton University Press; 2006.
  42. 42. Beaumont T, Fingerhut H. AP-NORC poll: Few Americans support easing virus protections [Internet]. AP News; 2020 Apr 22 [cited 2020 Dec 8]. Available from: https://apnews.com/article/9ed271ca13012d3b77a2b631c1979ce1.
  43. 43. Freking K, Fingerhut H. AP-NORC poll: Support for restrictions, virus worries wane [Internet]. AP news: AP news; 2020 Jun 25 [cited 2020 Dec 8]. Available from: https://apnews.com/article/915fdbccb3434fee125efaaaaefba0af.
  44. 44. Walters GD, Bolger PC. Procedural justice perceptions, legitimacy beliefs, and compliance with the law: A meta-analysis. Journal of Experimental Criminology. 2019 15(3):341–72.
  45. 45. Gau JM. Procedural justice, police legitimacy, and legal cynicism: A test for mediation effects. Police Practice and Research. 2015 16(5):402–15.
  46. 46. Posch K, Jackson J, Bradford B, Macqueen S. "Truly Free Consent”? Clarifying the nature of police legitimacy using causal mediation analysis. Journal of Experimental Criminology. 2020.
  47. 47. Fine A, Van Rooij B, Feldman Y, Shalvi S, Leib M, Scheper E, et al. Rule Orientation and behavior: Development and validation of a scale measuring individual acceptance of rule violation. Psychology, Public Policy, and Law 2016 22(3):314–29.
  48. 48. Fine A, Thomas A, van Rooij B, Cauffman E. Age-graded differences and parental influences on adolescents’ obligation to obey the law. Journal of Developmental and Life-Course Criminology. 2020:1–18.
  49. 49. Fine AD, van Rooij B. Legal socialization: Understanding the obligation to obey the law. Journal of Social Issues. 2021 77(2):367–91.
  50. 50. Kreps S, Kriner D. Model uncertainty, political contestation, and public trust in science: Evidence from the COVID-19 pandemic. Science Advances. 2020 6(43):eabd4563. pmid:32978142
  51. 51. Plohl N, Musil B. Modeling compliance with COVID-19 prevention guidelines: The critical role of trust in science. Psychology, Health & Medicine. 2021 26(1):1–12. pmid:32479113
  52. 52. De Coninck D, Frissen T, Matthijs K, d’Haenens L, Lits G, Champagne-Poirier O, et al. Beliefs in conspiracy theories and misinformation about COVID-19: Comparative perspectives on the role of anxiety, depression and exposure to and trust in information sources. Frontiers in Psychology. 2021 12:646394. pmid:33935904
  53. 53. Roozenbeek J, Schneider CR, Dryhurst S, Kerr J, Freeman AL, Recchia G, et al. Susceptibility to misinformation about COVID-19 around the world. Royal Society Open Science. 2020 7(10):201199. pmid:33204475
  54. 54. Gottfredson MR, Hirschi T. A general theory of crime: Stanford University Press; 1990. pmid:2278954
  55. 55. Pratt TC, Cullen FT. The empirical status of Gottfredson and Hirschi’s general theory of crime: A meta‐analysis. Criminology. 2000 38(3):931–64.
  56. 56. Pratt TC, Cullen FT. Assessing macro-level predictors and theories of crime: A meta-analysis. Crime and Justice. 2005 32:373–450.
  57. 57. Pratt TC, Lloyd K. Self-control and offending. In: Van Rooij B, Sokol DD, editors. The Cambridge Handbook on Compliance. Cambridge, UK: Cambridge University Press; 2021. p. 489–98.
  58. 58. Vazsonyi AT, Mikuška J, Kelley EL. It’s time: A meta-analysis on the self-control-deviance link. Journal of Criminal Justice. 2017 48:48–63.
  59. 59. Agnew R. Foundation for a general strain theory of crime and delinquency. Criminology. 1992 30(1):47–88.
  60. 60. Agnew R. Pressured into crime: An overview of general strain theory. Los Angeles: Roxbury; 2006.
  61. 61. Agnew R, White HR. An empirical test of general strain theory. Criminology. 1992 30(4):475–500.
  62. 62. Agnew R, Brezina T, Wright JP, Cullen FT. Strain, personality traits, and delinquency: Extending general strain theory. Criminology. 2002 40(1):43–72.
  63. 63. Baron SW. General strain, street youth and crime: A test of Agnew’s revised theory. Criminology. 2004 42(2):457–84.
  64. 64. Piquero NL, Sealock MD. Gender and general strain theory: A preliminary test of Broidy and Agnew’s gender/GST hypotheses. Justice Quarterly. 2004 21(1):125–58.
  65. 65. Botchkovar EV, Tittle CR, Antonaccio O. General strain theory: Additional evidence using cross‐cultural data. Criminology. 2009 47(1):131–76.
  66. 66. Brooks SK, Webster RK, Smith LE, Woodland L, Wessely S, Greenberg N, et al. The psychological impact of quarantine and how to reduce it: rapid review of the evidence. The Lancet. 2020 395(10229):912–20. pmid:32112714
  67. 67. Schultz PW, Nolan JM, Cialdini RB, Goldstein NJ, Griskevicius V. The constructive, destructive, and reconstructive power of social norms. Psychological Science. 2007 18(5):429–34. pmid:17576283
  68. 68. Cialdini RB, Demaine LJ, Sagarin BJ, Barrett DW, Rhoads K, Winter PL. Managing social norms for persuasive impact. Social Influence. 2006 1(1):3–15.
  69. 69. Goldstein NJ, Cialdini RB, Griskevicius V. A room with a viewpoint: using social norms to motivate environmental conservation in hotels. Journal of Consumer Research. 2008 35:472–82.
  70. 70. Cialdini RB, Goldstein NJ. Social influence: compliance and conformity. Annual Review of Psychology. 2004 55:591–621. pmid:14744228
  71. 71. Cohen LE, Felson M. Social change and crime rate trends: A routine activity approach. American Sociological Review. 1979 44(4):588–608.
  72. 72. Osgood DW, Wilson JK, O’malley PM, Bachman JG, Johnston LD. Routine activities and individual deviant behavior. American Sociological Review. 1996 61(4):635–55.
  73. 73. Spano R, Freilich JD. An assessment of the empirical validity and conceptualization of individual level multivariate studies of lifestyle/routine activities theory published from 1995 to 2005. Journal of Criminal Justice. 2009 37(3):305–14.
  74. 74. Clarke RV. Seven misconceptions of situational crime prevention. In: Tilley N, editor. Handbook of crime prevention and community safety. New York, NY: Routledge; 2005. p. 39–70.
  75. 75. Clarke RV. "Situational" Crime Prevention: Theory and practice. The British Journal of Criminology. 1980 20(2):136–47.
  76. 76. Age and Sex Composition in the United States: 2019 [Internet]. United States Census Bureau; 2019 [cited 2020 Dec 8]. Available from: https://www.census.gov/data/tables/2019/demo/age-and-sex/2019-age-sex-composition.html.
  77. 77. Kooistra EB, Reinders Folmer CP, Kuiper ME, Olthuis E, Brownlee M, Fine A, et al. Mitigating COVID-19 in a nationally representative uk sample: Personal abilities and obligation to obey the law shape compliance with mitigation measures. SSRN 3598221 [Preprint]. 2020 May 11 [cited 2021 Jul 7]. Available from: https://dx.doi.org/10.2139/ssrn.3598221.
  78. 78. Kuiper ME, De Bruijn AL, Reinders Folmer CP, Olthuis E, Brownlee M, Kooistra EB, et al. The intelligent lockdown: Compliance with COVID-19 mitigation measures in the Netherlands. SSRN 3598215 [Preprint]. 2020 May 13 [cited 2020 Sep 10]. Available from: https://doi.org/10.2139/ssrn.3598215.
  79. 79. De Bruijn AL, Feldman Y, Kuiper ME, Brownlee M, Reinders Folmer CP, Kooistra EB, et al. Why did Israelis comply with COVID-19 Mitigation Measures during the initial first wave lockdown? SSRN 3681964 [Preprint]. 2020 Aug 27 [cited 2021 Jul 7]. Available from: https://dx.doi.org/10.2139/ssrn.3681964.
  80. 80. Adler NE, Epel ES, Castellazzo G, Ickovics JR. Relationship of subjective and objective social status with psychological and physiological functioning: Preliminary data in healthy white women. Health Psychology. 2000 19(6):586–92. pmid:11129362
  81. 81. Fine AD, Rowan Z, Simmons C. Do politics trump race in determining America’s youths’ perceptions of law enforcement? Journal of Criminal Justice. 2019 61:48–57.
  82. 82. Hasson Y, Tamir M, Brahms KS, Cohrs JC, Halperin E. Are liberals and conservatives equally motivated to feel empathy toward others? Personality and Social Psychology Bulletin. 2018 44(10):1449–59. pmid:29739293
  83. 83. Wojcik SP, Hovasapian A, Graham J, Motyl M, Ditto PH. Conservatives report, but liberals display, greater happiness. Science. 2015 347(6227):1243–6. pmid:25766233
  84. 84. Tankebe J, Reisig MD, Wang X. A multidimensional model of police legitimacy: A cross-cultural assessment. Law and Human Behavior. 2016 40(1):11–22. pmid:26301708
  85. 85. Baker T, Gau JM. Female offenders’ perceptions of police procedural justice and their obligation to obey the law. Crime & Delinquency. 2018 64(6):758–81.
  86. 86. Gau JM. Procedural justice and police legitimacy: A test of measurement and structure. American Journal of Criminal Justice. 2014 39(2):187–205.
  87. 87. Wolfe SE, Nix J, Kaminski R, Rojek J. Is the effect of procedural justice on police legitimacy invariant? Testing the generality of procedural justice and competing antecedents of legitimacy. Journal of Quantitative Criminology. 2016 32(2):253–82.
  88. 88. McCright AM, Dentzman K, Charters M, Dietz T. The influence of political ideology on trust in science. Environmental Research Letters. 2013 8(4):044029.
  89. 89. Weinberger DA, Schwartz GE. Distress and restraint as superordinate dimensions of self‐reported adjustment: A typological perspective. Journal of Personality. 1990 58(2):381–417. pmid:2213473
  90. 90. Emsley R, Liu H. PARAMED: Stata module to perform causal mediation analysis using parametric regression models. 2013 Apr 26.
  91. 91. Harvey N. Behavioral fatigue: Real phenomenon, naïve construct, or policy contrivance? Frontiers in Psychology. 2020 11:589892. pmid:33224078
  92. 92. Mahase E. Covid-19: Was the decision to delay the UK’s lockdown over fears of “behavioural fatigue” based on evidence? BMJ. 2020 370:m3166. pmid:32769080
  93. 93. Feyman Y, Bor J, Raifman J, Griffith KN. Effectiveness of COVID-19 shelter-in-place orders varied by state. Plos One. 2020 Dec 31;15(12):e0245008. pmid:33382849
  94. 94. Kaste M. Police Back Off From Social Distancing Enforcement [Internet]. NPR; 2020 May 15 [cited 2020 Sep 10]. Available from: https://www.npr.org/2020/05/15/857144397/police-back-off-from-social-distancing-enforcement.
  95. 95. Van Rooij B. Weak enforcement, strong deterrence: Dialogues with Chinese lawyers about tax evasion and compliance. Law & Social Inquiry. 2016 41(2):288–310.
  96. 96. Apel R. Sanctions, perceptions, and crime: Implications for criminal deterrence. Journal of Quantitative Criminology. 2013 29(1):67–101.
  97. 97. Flanders C, Federico C, Harmon E, Klein L. "Terroristic threats" and COVID-19: A guide for the perplexed. University of Pennsylvania Law Review Online. 2020 169:63–89.
  98. 98. General OotDA. Memorandum For All Heads of Law Enforcement Components, Heads of Litigating Division, And United States Attorneys [Internet]. U.S. Department of Justice; 2020 March 24 [cited 2020 Dec 8]. Available from: https://www.politico.com/f/?id=00000171-128a-d911-aff1-becb9b530000.
  99. 99. Kushner Gadarian S, Wallace Goodman S, Pepinsky TB. Partisanship, health behavior, and policy attitudes in the early stages of the COVID-19 pandemic. Plos One. 2021 16(4):e0249596. pmid:33826646
  100. 100. Rothberger H, Wilson T, Rosenfeld D, Humphrey M, Moore A, Bihl A. Politicizing the covid-19 pandemic: Ideological differences in adherence to social distancing. PsyArXiv k23cv [Preprint]. 2020 Apr 22 [cited 2020 Sep 10]. Available from: https://psyarxiv.com/k23cv.
  101. 101. Mahmoud AB, Fuxman L, Mohr I, Reisel WD, Grigoriou N. “We aren’t your reincarnation!” workplace motivation across X, Y and Z generations. International Journal of Manpower. 2020 42(1):193–209.
  102. 102. Schmelz K. Enforcement may crowd out voluntary support for COVID-19 policies, especially where trust in government is weak and in a liberal society. Proceedings of the National Academy of Sciences. 2021 118(1):e2019378118. pmid:33443149
  103. 103. Lutz T, Pengelly M. Trump claims ’victory’ as US sees Covid-19 case records in multiple states [Internet]. The Guardian; 2020 Jul 4 [cited 2020 Sep 10]. Available from: https://www.theguardian.com/us-news/2020/jul/04/us-coronavirus-cases-fourth-of-july-holiday.
  104. 104. Reinders Folmer CP, Kuiper ME, Olthuis E, Kooistra EB, De Bruijn AL, Brownlee M, et al. Sustaining compliance with Covid-19 mitigation measures? Understanding distancing behavior in the Netherlands during June 2020. PsyArXiv xafwp [Preprint]. 2020 Sep 26 [cited 2020 Dec 8]. Available from: https://psyarxiv.com/xafwp/.
  105. 105. Reinders Folmer CP, Kuiper ME, Olthuis E, Kooistra EB, de Bruijn AL, Brownlee M, et al. Compliance in the 1.5 meter society: Longitudinal analysis of citizen’s adherence to COVID-19 mitigation measures in a representative sample in the Netherlands. SSRN 3624959 [Preprint]. 2020 Sep 3 [cited 2020 Dec 8]. Available from: https://doi.org/10.2139/ssrn.3624959.
  106. 106. van Rooij B, Sokol DD. Introduction: Compliance as the interaction between rules and behavior. In: van Rooij B, Sokol DD, editors. Cambridge Handbook of Compliance. Cambridge, UK: Cambridge University Press; 2021. p. 1–10.
  107. 107. Ansolabehere S, Schaffner BF. Does survey mode still matter? Findings from a 2010 multi-mode comparison. Political Analysis. 2014 22(3):285–303.
  108. 108. Hines DA, Douglas EM, Mahmood S. The effects of survey administration on disclosure rates to sensitive items among men: A comparison of an internet panel sample with a RDD telephone sample. Computers in Human Behavior. 2010 26(6):1327–35. pmid:20948972
  109. 109. Zhang X, Kuchinke L, Woud ML, Velten J, Margraf J. Survey method matters: Online/offline questionnaires and face-to-face or telephone interviews differ. Computers in Human Behavior. 2017 71:172–80.
  110. 110. Bauhoff S. Systematic self-report bias in health data: Impact on estimating cross-sectional and treatment effects. Health Services and Outcomes Research Methodology. 2011 11:44–53.
  111. 111. Van de Mortel TF. Faking it: Social desirability response bias in self-report research. The Australian Journal of Advanced Nursing. 2008 25(4):40–8.
  112. 112. Larsen MV, Nyrup J, Petersen MB. Do survey estimates of the public’s compliance with COVID-19 regulation suffer from social desirability bias? Journal of Behavioral Public Administration. 2020 3(2).
  113. 113. Google COVID19 Community Mobility Report [Internet]. 2020 [cited 2020 Sep 10]. Available from: https://www.gstatic.com/covid19/mobility/Global_Mobility_Report.csv?cachebust=5e35f7008c7c1554.
  114. 114. Hoeben EM, Bernasco W, Liebst LS, Van Baak C, Rosenkrantz Lindegaard M. Social distancing compliance: A video observational analysis. Plos One. 2021 16(3):e0248221. pmid:33720951