Introduction

The development of artificial intelligence (AI) tools is on a rapidly increasing trajectory. However, advanced AI brings with it also new risks and may exacerbate existing concerns, including surveillance and invasion of privacy1, misinformation and manipulation2, discrimination3, liability and accountability concerns4, social risks related to disruptions of labor markets5, cybersecurity threats6, terrorism and other forms of criminal misuse7, and autonomous warfare8.

Regulatory oversight over safety of AI research and deployment could help mitigate some of these risks. Regulatory oversight may include risk disclosure and audit requirements9, standards and restrictions10, and in some cases outright bans8. Regulatory oversight may be usefully complemented by other policy instruments, such as taxes11, and by industry norms12.

Data from two large surveys we conducted in Germany and Spain indicate there is considerable appetite for “much stricter” regulatory oversight of commercial AI research, which is supported or strongly supported by 62.2 and 63.5% of the German and Spanish participants, respectively. We explore whether socio-economic characteristics, beliefs about future economic impacts of AI, innovativeness, altruism, risk preferences, and time preferences predict support for AI regulation.

Method

Participants and procedure

We report exploratory analyses of data that are part of two larger studies. Study 1 (Spanish residents, n = 1434) and Study 2 (German residents, n = 1430) were conducted in September 2023 with quota samples recruited by a market research company (SurveyEngine GmbH) from their partner panels. The surveys were programmed by SurveyEngine and hosted on their online platform. Sample socio-demographics can be found in Table 1.

The studies received approval from SIKT (Norwegian Agency for Shared Services in Education and Research, reference no. 345930). Our institution did not require an additional approval for the studies. The studies were conducted according to the principles expressed in the Declaration of Helsinki, as well as with ethics guidelines by the American Psychological Association and with national and institutional regulations. Written informed consent was provided by all participants.

Table 1 Sample socio-demographics.

Measures

Support for AI safety oversight is our dependent variable, measured with the item “Do you support much stricter regulatory oversight over safety of commercial artificial intelligence research?”, with response options “strongly oppose” (coded as 1), “somewhat oppose” (2), “neither support, nor oppose” (3), “somewhat support” (4), and “strongly support” (coded as 5).

We explore several possible determinants of support for AI oversight. The first group of predictors are participants’ expectations regarding AI’s economic impacts in a 10-year horizon. Anticipated economic well-being was measured with the item “What will in your estimate be the most likely impact of artificial intelligence on your economic well-being in 10 years from now?”, with response options “large reduction” (coded as 1), “moderate reduction” (2), “small reduction” (3), “no change” (4), “small improvement” (5), “moderate improvement” (6), and “large improvement” (coded as 7). Anticipated job displacement was measured with the item “What percentage of workers in your occupation will in your estimate be replaced by artificial intelligence within the next 10 years?”, with the following response options: “0–10%” (coded as 1), “10–30%” (2), “30–50%” (3), and “more than 50%” (coded as 4).

We measured participants’ innovativeness using two items adopted from Hurt et al.13: “I have to see other people use new inventions before I consider using them myself” and “I am skeptical to new inventions”, with response options “strongly disagree” (coded as 1), “somewhat disagree” (2), “neither agree, nor disagree” (3), “somewhat agree” (4), and “strongly agree” (coded as 5). The two items were not sufficiently correlated to form a reliable scale (r = 0.35 in Study 1, r = 0.41 in Study 2) and were therefore used as separate measures in the analysis, labelled laggard (the first item) and skeptical (the second item). Note that responses are coded such that higher scores indicate being more of a laggard or more skeptical to innovations.

The third group of predictors are participants’ altruism, time preferences, and risk preferences. Altruism was measured with an item adapted from Falk et al.14: “How do you assess your willingness to share with others without expecting anything in return?”, with response options ranging from 1 = “completely unwilling to share” to 7 = “extremely willing to share” (only the two extreme response options were labelled). Time preferences were measured with an item adapted from Falk et al.14: “In comparison to others, are you a person who is generally willing to give up something today in order to benefit from that in the future or are you not willing to do so?”, with response options ranging from 1 = “completely unwilling to give up something today” to 7 = “extremely willing to give up something today” (only the two extreme response options were labelled). General risk preferences were measured with an item adapted from Falk et al.14: “Are you a person who is generally willing to take risks or do you try to avoid taking risks?”, with response options ranging from 1 = “completely unwilling to take risks” to 7 = “extremely willing to take risks” (only the two extreme response options were labelled). Economic risk preferences were measured with an item based on Eckel and Grossman15. Participants were asked to select one of eight hypothetical lotteries with different levels of risk (see Table 2). Selection of lotteries was coded such that higher-risk lotteries received a higher score.

Table 2 Response options for the economic risk preferences item.

The final group of predictors are participants’ socio-economic characteristics, including gender, age, education, income, and household size. Participants in addition stated their occupation by filling in a text field, but it was not possible to reliably categorize participants’ responses and this data was therefore excluded from analysis.

Results

Descriptive analysis

Figure 1 displays cumulative distributions of support for AI safety oversight in Studies 1 and 2. As can be seen from the figure, AI safety oversight enjoys similar levels of support in both countries. Only around 10% of participants oppose stricter AI safety oversight (11.9% of the Spanish participants in Study 1 and 10.4% of the German participants in Study 2). Around a quarter of participants do not have a preference either way (24.6% in Spain and 27.4% in Germany). The remaining nearly two thirds of participants somewhat support or strongly support much stricter AI safety oversight (63.5% in Spain and 62.2% in Germany).

Fig. 1
figure 1

Cumulative distributions of support for AI safety oversight.

Table 3 presents a breakdown of support for AI safety oversight by gender and age. Recall that a score of 4 indicates the participant “somewhat supports” strict AI oversight and a score of 3 indicates they “neither support, nor oppose” such policy. The majority of means reported in Table 3 fall within this range. While there are essentially no differences between men and women, participants reporting non-binary or other gender are descriptively speaking less supportive of AI safety oversight. Support for AI safety oversight tends to gradually increase with age, with about half a standard deviation change in support between the youngest and oldest age group. These patterns are formally tested in regression analyses reported below.

Table 3 Means and standard deviations of support for AI safety oversight by gender and age.

Regression analysis

Table 4 presents linear regression estimates, reporting standardized regression coefficients with 95% confidence intervals based on robust standard errors. We report p-values adjusted for multiple testing using the Benjamini and Hochberg method16. Collinearity is not an issue in either model (mean VIF = 1.17 in Study 1, mean VIF = 1.20 in Study 2).

Table 4 Determinants of support for AI safety oversight.

Respondents expecting AI to lead to more job displacement in their profession are less supportive of AI safety regulation (only marginally statistically significant in the Spanish sample). Bivariate correlations between anticipated job displacement and support for AI oversight are also negative, r = −0.16 in both studies. An explanation of this finding is not obvious, nevertheless it seems to speak against a motivated reasoning account which would be more consistent with a positive correlation between the two variables (wherein workers fearing job displacement could attempt to slow down AI progress through demanding strict regulation).

As expected, laggards and, at least in the German sample, those skeptical to new inventions are more supportive of AI safety regulation. Surprisingly, skepticism appears to be associated with less support for regulation in the Spanish sample – this is the least intuitive finding in our two studies and should be interpreted with caution.

The analysis also suggests that social and economic preferences may play an important role in shaping preferences for AI safety regulation. Altruistic, risk averse, and – in the German case – patient participants are more supportive of strict AI safety oversight.

Finally, older participants are more supportive and participants who identify as non-binary are less supportive of strict AI safety regulation (the latter finding is based on a small number of cases, five in Spain and four in Germany, and as potentially unreliable is therefore not considered in the following discussion). There do not appear to be other socio-demographic drivers of regulation support.

Looking at statistically significant predictors’ standardized beta coefficients, we see from Table 4 that a one standard deviation increase in a predictor is associated with between 8 and 15% of a standard deviation change in support for AI safety oversight. Unreported eta-squared statistics show that these predictors individually explain between 0.6% (age in Study 1) and 2.0% (laggard in Study 1) of variance in the dependent variable. We consider predictors uniquely explaining less than 1.0% of variance in the dependent variable to be of negligible importance and predictors explaining between 1.0 and 2.0% of variance to be of modest importance. From this perspective, of the statistically significant predictors, the importance of age would be considered negligible in the Spanish sample (uniquely explaining 0.6% of variance in the dependent variable) but modestly important in the German sample (uniquely explaining 1.8% of variance in the dependent variable). The importance of general risk preferences would be considered negligible in the Spanish sample (uniquely explaining 0.9% of variance in the dependent variable) but again modestly important in the German sample (uniquely explaining 1.5% of variance in the dependent variable). In summary, all statistically significant predictors in Table 4 show some degree of relevance.

Importantly, also because of the relatively weak correlations between predictors (|0.01| < r < |0.37| in Study 1 and |0.01| < r < |0.46| in Study 2, excluding non-binary participants), the models as a whole explain considerable amounts of variance in support for AI safety oversight – 10.4% in the Spanish sample and 13.0% in the German sample.

Concluding remarks

We report result from, to our knowledge, the first two studies on the general population’s preferences for AI safety oversight. The majority of respondents were in favor of strict safety oversight. Interestingly, socio-demographic factors only played a minor role (similarly see König et al.17). Less innovative participants were more supportive of safety oversight, largely as expected (although the results were somewhat mixed).

Participants anticipating greater job displacement due to AI over the next ten years were less supportive of AI safety oversight. This finding goes against the intuition that people concerned about job displacement may want to politically leverage demands for stricter AI oversight to slow down AI progress and its presumed labor market impacts with it. We should caution that the effect was small and only statistically significant in Germany (marginally so in Spain). Nevertheless, it may warrant further exploration in subsequent research. Jeffrey18 similarly found that making economic vulnerability to AI and automation more salient decreased support for certain welfare policies. One possible explanation for our finding is that participants might view strict AI safety regulation as a facilitator, not of AI research directly, but of commercial adoption of the now better regulated and more trusted AI tools. Another possibility is that participants concerned about job displacement could be primarily focused on legislation addressing this problem directly and may even oppose legislation that they feel could de-emphasize their policy priorities (see Gallego et al.19). These interpretations involve different assumptions about the electorate’s political and technological foresight. We therefore suggest that an analysis of the role played by foresight in shaping preferences for policies responding to large technological shifts could be promising.

Our results also reveal a connection between risk, time and altruistic preferences and support for AI safety oversight. An emerging literature similarly suggests associations between economic and social preferences and technology acceptance20. Subsequent research on AI safety regulation and AI regulation more broadly can build on the present findings by further examining the role of people’s economic and social preferences in this domain for example by manipulating perceptions of AI risk and of who, when and with what probability may be impacted by it.

The theoretical approach taken in this paper is purposefully eclectic given the highly exploratory nature of the present studies and lack of previous research on the topic. We focus on correlations with basic socio-economic variables and beliefs about the technology’s economic impact, with key economic preferences which have been shown to play a role in the broader technology adoption literature20 , and with consumer innovativeness. Subsequent research can build on this work by replicating our findings concerning risk, time and altruistic preferences, innovativeness, and socio-economic factors, as well as by incorporating other theoretical perspectives, such as influences postulated in the Unified Theory of Acceptance and Use of Technology21. Additional economic and social preferences, such as preferences for fairness and economic efficiency, should also be considered. For example, a recent study22 found that acceptance of AI-powered facial recognition tools, an AI-application associated with potential privacy risks, is predicted by perceptions of fairness. Given the novel nature of AI technologies and their potential broad social impacts, it could be beneficial to incorporate a wide range of influences when attempting to understand people’s preferences for whether and how the technology should be regulated.

Limitations

Measures collected for the purposes of this paper were part of longer questionnaires, and so simplicity and ease of comprehension were a priority. A downside of this approach was that single-item measures were used to operationalize most constructs, potentially leading to effect size attenuation due to measurement error. Our estimates may therefore be interpreted as lower bounds of the true associations. Where possible we relied on widely used measures, and where no widely used measures were available we constructed our own. Development of psychometrically validated scales to measure vulnerability to AI, in particular, could be a useful direction for subsequent research (besides items introduced in this paper see also other recent work18,19).

Subsequent research may not only consider inclusion of multi-item measures of relevant constructs, but also consider measuring different facets of the broader constructs. We took a first step in this direction by including a measure of general risk preferences14 and a measure of economic risk preferences15, anticipating that these could play different roles in predicting support for AI safety oversight. Indeed, only general risk preferences predicted support for AI safety oversight.

Finally, even though some effects replicated across studies (associations with innovativeness, altruism, risk preferences, and age), subsequent confirmatory research is needed to further substantiate the present findings.