I. Introduction
Insurers underwrite risks by analysing data. Insurers increasingly use Artificial Intelligence (AI): based on correlations found in data, the insurer estimates the expected claims cost of the consumer and assigns the consumer to a risk class. Two modern trends in insurance are (i) data-intensive underwriting and (ii) behaviour-based insurance. (i) Data-intensive underwriting means that insurers use and analyse more data for risk underwriting. Insurers could extend their analyses with machine learning. For example, insurers could collect more data about consumers and use machine learning to find new correlations in data to underwrite risks with. Some of those correlations could look odd to the consumer, such as the correlation between a consumer’s house number and the probability that they file an insurance claim (ii) A second trend is behaviour-based insurance: insurers can analyse the behaviour of an individual consumer to offer that consumer a discount. For example, some insurers offer a discount to the consumer if that consumer allows the insurer to monitor their driving behaviour (behaviour-based car insurance). Other insurers offer discounts to health or life insurance when consumers show, with a health tracker, that they have an active lifestyle.Footnote 1
Both modern insurance trends can have advantages. For instance, careful consumers or consumers who run less risk may be able to insure themselves for a lower premium.Footnote 2 But both trends may also have negative effects. For example, the insurer could, unintentionally, illegally discriminate certain groups. Other effects could be controversial or unfair, even if the insurer does not discriminate illegally. This paper focuses on such possibly unfair – but not necessarily illegal – effects.
While there is much research discussing these trends, less is known about what the public thinks about data-intensive and behaviour-based decisions in the insurance sector. We performed a survey of the Dutch population (N = 999) in which we asked respondents for their opinions about examples of (i) data-intensive underwriting and (ii) behaviour-based insurance. In general, we wanted to know whether, and under which circumstances, the general public finds the practices fair and acceptable. The paper explores the following questions.
-
(1) To what extent do people find data-intensive underwriting and behaviour-based insurance fair and acceptable?
-
(2) If people feel they have more influence on the premium, do they find an insurance practice fairer and more acceptable?
-
(3) If people find an insurance practice more logical, do they find the practice fairer and more acceptable?
-
(4) If an insurer only accepts certain parts of the population, do people find such practices less fair and acceptable?
-
(5) If an insurance practice leads to higher prices for poorer consumers, do people find the practice less fair and less acceptable?
The paper makes the following contributions to the literature. First, this is one of the first academic papers to research what the public finds fair in the context of modern applications of AI in insurance.Footnote 3 This is an important addition to the literature, as fairness is always context-specific.Footnote 4 Second, we are the first to ask respondents about specific types of insurance fairness. For example, we explore what respondents think about insurers basing the premium on consumer characteristics that may seem illogical, or irrelevant, to the respondent.
Our results can be useful for several groups of readers. For example, the results could inspire scholars or others interested in perceived fairness of AI and data use, as the survey gives an impression of the attitudes of the average citizen for real-life examples. Such surveys are scarce: the perspectives of insurance consumers are “generally an under-researched area.”Footnote 5 Second, insurers who use AI in their business, or plan to do so, may want to know what respondents think of modern insurance trends.Footnote 6 Third, policymakers who consider regulating AI in insurance could use the survey as a first impression of what the public thinks.Footnote 7 We do not claim that policymakers should always prohibit practices that seem unfair according to surveys.Footnote 8 The results could be of use in other sectors outside of insurance too, as insurance is an example of a sector that has relied on statistics and AI for a relatively long period of time.Footnote 9
We aim to avoid unnecessary jargon, so that the paper is understandable for readers without a background in insurance or actuarial science. We conducted the survey in Europe, and we discuss some EU law in the discussion section. The paper’s results can be useful outside the EU too, for example, by providing ideas for further research in other countries, allowing transnational applications of AI, or by enabling more (public) discussion about the topic.Footnote 10
In Section II, building on earlier work, we provide background to our survey questions. Section III discusses the method for the survey. Section IV presents the results, with each subsection discussing a different research question. In Section V we reflect on the results and place them in broader context, such as whether the law and fairness metrics can help against the perceived unfairness. Section VI discusses limitations of the research and provides suggestions for further research. Section VII concludes.
II. Background: AI in insurance and possibly unfair differentiation
The core business of insurers is underwriting risks: estimating the expected claims cost of a consumer via a risk assessment.Footnote 11 Insurers always discriminate, in the neutral sense, between groups of people. Many insurers use a form of AI to underwrite risks or to analyse past data. AI can be described as “the science and engineering of making intelligent machines, especially intelligent computer programs.”Footnote 12 For more than a century, insurers have gathered data directly from the consumers, such as data about a consumer’s age, smoking or drinking habits. Insurers used questionnaires to gather consumer characteristics such as the type of car they drive.Footnote 13
Now, insurers could analyse more data gathered from many different sources. Insurers seem captivated by two trends (i) First, insurers could analyse more and new types of data to assess risks more precisely: data-intensive underwriting. (ii) Second, insurers could the behaviour of individual consumers in real-time: behaviour-based insurance. For example, some car insurers offer a discount if the consumer agrees to being tracked by the insurer and drives safely. AI makes both types of practices easier to implement for insurers.Footnote 14
1. Data-intensive underwriting
The first trend is data-intensive underwriting: insurers use and analyse more data for underwriting. With machine learning, insurers can find new correlations in data with which the insurer could more accurately predict the expected claim costs of certain consumer groups. Insurers could analyse new datasets to find many new correlations in data.Footnote 15
Data-intensive underwriting comes with potential discrimination-related risks. A legal risk of data-intensive underwriting is that insurers could, accidentally, discriminate indirectly in an illegal way (disparate impact). For example, if an insurer uses a newly found correlation to set premiums, the insurer could inadvertently cause harm to certain ethnic groups, or other groups with legally protected characteristics.Footnote 16 There are other possibly unfair, or at least controversial, effects of data-intensive underwriting. Our paper focuses on those effects.
First, consumers may be confronted with premiums based on characteristics that the consumer has little influence on. For example, in the Netherlands, some insurers charged different car insurance premiums based on the consumer’s house number. A consumer paid more if their house number included a letter: for instance, 10A or 40B.Footnote 17 Consumers may feel that they cannot reasonably influence their house number. Moving to another house to lower one’s car insurance premium is hardly a serious option. Consumers can influence other characteristics more easily, such as the type of car they buy.
Second, insurers could use certain consumer characteristics to set premiums, while the consumer does not see the logic of using the characteristic.Footnote 18 In other words, the insurer could use seemingly illogical or irrelevant characteristics to set the premium. AI systems are good at finding correlations in large datasets. When insurers underwrite using more data, they might find many correlations in datasets that look strange to consumers. In theory an insurer could find a correlation between on the one hand the chance that a consumer files a claim, and on the other hand whether a consumer has an address with an even number, is born in a certain month, or “people who spend more than 50% of their days on streets starting with the letter J.”Footnote 19
Third, insurers could introduce insurance products that are only intended for specific groups. Suppose, for example, that according to an insurer’s analysis, people with a higher education submit fewer claims to their car insurer. The insurer could decide to offer car insurance only to a highly educated people. To illustrate, Promovendum, a Dutch insurer, markets its insurances as being only for the high-educated.Footnote 20 The insurer explains its reasoning: “Based on our years of experience and our claims statistics, we know that this target group has fewer claims. This enables us to offer our customers insurance products with low premiums and excellent terms and conditions.”Footnote 21 Another insurer offers a car insurance only to medical doctors.Footnote 22 Such insurance practices can be controversial because the insurer targets specific groups of the population and thereby excludes the rest.
Fourth, insurance practices could reinforce inequalities that already exist in society, such as the difference between rich and poor. Such an effect could occur if an insurer (on purpose or by accident) charges higher prices to poorer people. In sum, data-intensive underwriting could have several negative effects.
2. Behaviour-based insurance
A second trend in insurance is behaviour-based insurance: insurers can adapt the premium to individual consumers in real-time, based on how the consumer behaves. For example, some car insurers offer a discount if the consumer agrees to being tracked by the insurer (with a device in the car) and drives safely.Footnote 23 A life insurer could give discounts to people who show, with a health tracker or smart watch, that they lead an active life.Footnote 24 Insurers can improve their behaviour-based insurance with AI. For example, with a linear modelling approach, an insurer can consider characteristics such as weather and light conditions, mileage, time and driving habits such as speeding to assess the consumer’s driving behaviour.Footnote 25
The most important difference between data-intensive underwriting and behaviour-based insurance is that with behaviour-based insurance, the insurer bases the premium on the behaviour of an individual consumer.Footnote 26 The insurer monitors, for instance, how the consumer drives, or how active a consumer is every day, measured by a health tracker or step counter. Another difference is that with behaviour-based insurance, the insurer focuses more on preventing risks. For example, insurers could (at least in theory) prevent accidents by nudging risky drivers to drive safer.
Behaviour-based insurance comes with partly similar risks as data-intensive underwriting. For example, some people may be excluded from behaviour-based insurance. For bad drivers, behaviour-based car insurance might be so expensive that, practically speaking, they are excluded from car insurance. In a hypothetical market where all car insurers offer behaviour-based insurance, bad drivers may not have the possibility to insure themselves for an affordable price.Footnote 27
Second, behaviour-based insurance could reinforce financial inequality. Suppose for example, that on average, people with higher salaries have more time for sports or an active lifestyle. If richer people receive more discounts (based on their health trackers), they pay less. If richer people pay less, financial inequality is reinforced.
In conclusion, both data-intensive underwriting and behaviour-based insurance could have several possible negative effects. We do not claim that all these effects will occur; we merely highlight possible effects. We also note that insurers could combine both trends: insurers can aggregate individual behavioural data, and use the data for more data-intensive underwriting.Footnote 28 Insurers could then use that profit for more behaviour-based insurance.Footnote 29 Nevertheless because of their partly different effects, we present the two trends separately.
II. Method
1. Survey design and sample
We conducted an online survey about data-intensive underwriting and behaviour-based insurance among 1,000 Dutch citizens of 18 years and older, using a sample provided by panel service Panelclix in the Netherlands. In total, 1,102 respondents completed the questionnaire. We excluded respondents who completed the questionnaire in under 4 minutes (n = 96), and respondents failing both attention checks (n = 6).Footnote 30 One respondent requested to have their data removed after data collection; therefore, our final sample was N = 999.
The final sample consisted of 51.6% female, 46.6% male and 1.7% other/do not want to say. Mean age was 49.82 (SD = 17.13). Of the sample 30.5% was highly educated, 51.7% received mid-level education and 17.7% received primary or lower-level secondary education (Dutch population: 36.4%, 37.0%, and 26% respectively; CBS, 2024). Net monthly income was as follows: 6.9% earned less than €1000, 35.6% between €1000 and €2500, 35.4% between €2500 and €5000, and 4.8% over €5000 (net modal monthly income in the Netherlands in 2024 was €2.692; CPB, 2024). Hence, the sample consisted of a diverse demographic in terms of age, gender, education and income.
Respondents filled out a questionnaire on “practices in the insurance sector.” The median completion time was around 8 minutes. Respondents were compensated for their participation. The study complied with the criteria of Ethics Committee of the authors’ institution (reference no. ECSW-LT-2025-1-10-82717). All data and materials are available at OSF.Footnote 31
2. Procedure
Before respondents started the questionnaire, we asked them to give their informed consent. Subsequently, we asked respondents for their age. Next, respondents indicated their attitudes towards several short scenarios regarding data-intensive underwriting and behaviour-based insurance, related to several insurance domains (e.g., car, life, or burglary insurance).
We included two attention checks in the questionnaire by means of instructed response items. The first check was an embedded item, instructing the respondents to “click fully agree here.” A second check instructed respondents to click “green” in a list of colours.Footnote 32 Finally, respondents answered questions on socio-demographics. After completion, respondents were thanked and redirected to Panelclix collect their compensation. Respondents were paid 0.75 Euro for their participation.
3. Questionnaire
We presented respondents with various scenarios about different data-intensive insurance practices. Respondents read a short introduction, after which they were presented with several one-sentence examples of practices. For example, in one section (“block”) of the survey we told respondents:
This block is about how insurance companies determine the insurance premium (how much you pay each month). We want to know your opinion on this.
An insurance company (insurer) offers car insurance. The insurer calculates the monthly premium based on predictions of the number of claims for damages a consumer will make. The insurer makes these predictions based on information it has collected previously about the consumer, such as age, address, and type and year of the car. Below we present several examples of how an insurer can determine the premium. We ask for your opinion on each example.
Example 1: The insurer determines the premium based on the postal code of the consumer.
We wanted to know whether respondents found the practice that we described fair and acceptable. In other words, the two main outcome variables were Fairness and Acceptance. We measured Fairness by the statement: “I think this practice is fair.” Respondents indicated their agreement on a 5-point Likert scale: (1) completely disagree; (2) disagree; (3) neutral; (4) agree; (5) completely agree. Likewise, for Acceptance, respondents indicated their agreement with the statement: “I think this practice is acceptable.”
We also wanted to know whether respondents felt that a consumer could influence the premium, and whether they saw the logic behind the insurer using that characteristic to calculate the premium. Hence, additional variables were Perceived Influence and Perceived Logic. For Influence respondents indicated on a 5-point scale to what extent they agreed with the statement: “This way, the policyholder can influence the insurance premium”. For Logic, respondents indicated to what extent they agreed with the statement “I find it logical that the insurer determines the premium like this.” We asked the respondents the questions in Dutch. If we quote a question in this paper, we translate the question to English.
IV. Results
In this section, we present the results from the survey. Each subsection discusses a different research question of the paper. We present the results in text and figures. We include the full data in the tables in the Appendices.
1. Do people find (i) data-intensive underwriting and (ii) behaviour-based insurance fair and acceptable?
The first question is: do respondents find data-intensive underwriting and behaviour-based insurance fair and acceptable? A simple t-test compared the mean difference in fairness and acceptance between behaviour-based insurance practices vs. practices based on data-intensive data types. Here, two data types classified as sources for behaviour-based insurance, namely car tracker data (“car tracker”) and data based on a step counter (“steps”), whereas the other data types classified as sources for data-intensive underwriting. The t-test indicated that behaviour-based insurance practices overall score higher than data-intensive data types for insurance in terms of fairness (t (998) = 21.58, p < .001 (Mean Difference = .58)), and acceptance (t (998) = 18.26, p < .001 (Mean Difference = .51)).
As Figure 1 shows, respondents found it only (relatively) fair (i.e., scoring above the scale midpoint of 3) if an insurer adapts the premium based on the consumer’s driving behaviour, which is an example of behaviour-based insurance, and based on the maximum speed of the car, which is an example of data-intensive underwriting.Footnote 33 Respondents found it unfair and unacceptable if an insurer bases the premium on all the other types of data, including behaviour-based insurance based on a step counter (i.e., scoring below the scale midpoint of 3). Hence, the significant difference between behaviour-based insurance and data-intensive underwriting seems to be driven entirely by the higher fairness and acceptance values for the car tracker, and not by behaviour-based insurance with a step counter. In conclusion, respondents find almost all modern insurance practices that we described unfair. See Appendix A for an overview of significance levels of deviations from the scale’s neutral mid-point for each data type.

Figure 1. Public opinion of fairness and acceptance in insurance per data type.
2. If people feel they have more influence on the premium, do they find an insurance practice fairer and more acceptable?
As noted, insurers could calculate premiums based on characteristics that consumers cannot reasonably influence (see Section II). We asked respondents whether they felt that they could influence the premium if the insurer calculated the premium based on the following characteristics: the consumer’s gender, age or postal code, the car’s price, colour or maximum speed, the size of the city where the consumer lives, whether the consumer lived at an odd or even house number, and whether the consumer lived at a house number with a letter extension.
We also asked respondents whether they find it fair and acceptable if an insurer calculates the premium based on such characteristics.
We asked two questions about life insurance (we include the introductions in Appendix D). In the first example, the insurer charges higher prices to poorer people. We asked respondents whether they felt that the consumer can influence the premium in such a case. The question is based on a real-life example of an insurer charging higher premiums in poor neighbourhoods on purpose.Footnote 34
We asked a second question about this life insurance example, but with an extra explanation: “In the Netherlands, people with a migration background are on average poorer than those without a migration background. This insurer’s practice results in people with a migration background paying a higher premium on average than those without a migration background.” (The example is based on a real-life insurance practice.Footnote 35 ) Again, we asked respondents whether they thought that the consumer can influence the insurance premium.
We asked a similar question about burglary insurance.
We also asked respondents whether they felt that they could influence the insurance premium for two examples of behaviour-based insurance: behaviour-based car insurance (with a type of tracker in the car), and life insurance that gives a discount if the consumer walks more than 10,000 steps a day, as measured by a step counter.
Figure 2 shows the influence perceptions across all data types. For most data types, respondents feel that they can hardly influence the insurance premium. A one-sample t-test showed all values significantly deviated from the scale’s midpoint (3; See Appendix B for significance levels for each data type). Most practices deviate negatively from this neutral point (where respondents find the practices neither highly nor little influenceable; Figure 2; See also Appendix B). This means that for most practices, respondents feel that the consumer cannot really influence the premium. However, respondents feel that consumers have some influence on the premium in the case of behaviour-based insurance based on a car tracker and in the case of the price of the car they buy. Respondents do not feel that consumers have much influence on the premium in the case of behaviour-based insurance with a step counter.

Figure 2. Influence perceptions by data type.
We found a positive relationship between perceived influence and fairness and acceptance. Figures 3 and 4 give a graphic representation of the relationship between, on the one hand, perceived influence, and on the other hand perceived fairness (Figure 3) and acceptance (Figure 4). Regression analysis confirms the pattern that is shown in Figure 3, indicating that there is a positive relationship between perceived influence and the perceived fairness of the insurance practice across all cases (R 2 = .54, F(1, 997) = 1147.52, β = .732, p < .001). A similar conclusion can be drawn with regards to acceptance. Namely, as Figure 4 suggests, there is a positive relationship between perceived influence and perceived acceptance of the insurance practice (R 2 = .53, F(1, 997) = 1126.68, β = .728, p < .001).

Figure 3. The relationship between perceived influence and fairness across all cases.

Figure 4. The relationship between perceived influence and acceptance across all cases.
In addition, Figures 3 and 4 suggest that if respondents feel that they have little influence on the insurance premium, then almost all respondents find the practice unfair (Figure 3) and unacceptable (Figure 4). However, if respondents find that they have more influence on the premium, then some find the practice fairer than others; hence, respondents are more divided. We can see this in the error bar in the figure, which is more volatile, or varied, for data types for which respondents feel that they have more influence on the premium.Footnote 36
3. If people find an insurance practice more logical, do they find the practice fairer and more acceptable?
We asked respondents whether they saw the logic of certain insurance practices (see Section II about data-intensive underwriting). For various insurance practices, we asked respondents whether they found it logical that the insurer uses certain consumer characteristics to set the insurance premium. We also asked respondents whether they find it fair and acceptable if an insurer calculates the premium based on such practices.
In Figure 5, we show the perceived logic for all insurance practices. Respondents considered a situation where the insurer uses the price of the insured car as the most logical. Respondents considered a situation where an insurer uses the type of house number to calculate the premium for car insurance the least logical (i.e., an uneven house number and, to a slightly lesser extent, a house number with a letter).Footnote 37 Compared to the scales’ neutral midpoint of 3, respondents only considered situations in which car tracker data, or the price of a car, determined insurance fees as relatively logical (p’s <.001). Respondents considered all other situations as relatively illogical (p’s <.001), except for age, which the respondents considered neutral (p = .293; see Appendix B2 for all individual effects per insurance practice).

Figure 5. Perceived logic over all insurance practices.
We found a strong positive relationship between perceived logic and fairness. The more logical respondents found a certain insurance practice, the fairer they thought the practice was (R 2 = .84, F(1, 997) = 5142.06, β = .915, p < .001). Similarly, we found a positive relationship between perceived logic and acceptance (R 2 = .83, F(1, 997) = 4948.89, β = .912, p < .001, see also Figures 6 and 7). The error bars in the figures show that while respondents generally agree on the unfairness and unacceptability for examples that they find less logical, respondents disagree more often for more logical examples. In Section IV.3, we described a similar pattern for influence and fairness/acceptability. In sum, if respondents find an insurance practice more logical, they find the practice fairer and more acceptable.

Figure 6. The relationship between perceived logic and fairness across all insurance practices.

Figure 7. The relationship perceived logic and acceptance across all insurance practices.
4. If an insurer only accepts certain parts of the population, do people find such practices less fair and acceptable?
We asked whether respondents find it fair and acceptable if insurers only accept specific groups (see Section II). On average, respondents find all target group insurance completely unfair and unacceptable (all p’s < .001; see Figure 8; see also Appendix C for the individual significance levels of deviations from the scale’s midpoint). It hardly matters for respondents who the target group is. And, when they assess whether they find a specific target group insurance fair and acceptable, it seems that respondents find it irrelevant whether the target group has a low income (caregiver, unemployed) or a high income (lawyer, family doctor).

Figure 8. Respondents find target group insurance unfair and unacceptable.
5. If an insurance practice leads to higher prices for poorer consumers, do people find the practice less fair and less acceptable?
Respondents find it unfair when insurance practices result in higher insurance premiums for poor people (Figure 9). We asked a question about an insurer who charges higher premiums in neighbourhoods where more burglaries occur. We added: “This leads to poorer people paying higher premiums on average.” (The full question is quoted in Section IV.2). Hence, in this question, the fact that poorer people pay higher prices is an accidental effect of an insurance practice. The insurer did not charge higher premiums to poorer people on purpose, according to the question.

Figure 9. Public opinion of fairness and acceptance when poor population is indirect target group on the basis of risky neighbourhoods or migrant status.
Regardless, respondents find the burglary insurance practice unfair and unacceptable (t (998) = –41.880, Mean Difference = –1.22, and unacceptable, t (998) = –38.27, Mean Difference = –1.17). See Figure 9. So, at least in this example, respondents find it unfair and unacceptable if an insurance practice leads to higher prices for poorer consumers.
We also asked a question in which a life insurer charged a higher premium to poorer consumers on purpose. The reasoning of the insurer is that poorer people live shorter lives, and thus pay a premium for fewer years (see Section IV.2 for the full question).
We asked a second question about life insurance that was largely the same, but which included some extra information: “In the Netherlands, people with a migration background are on average poorer than people without a migration background. This insurer’s practice results in people with a migration background paying a higher premium on average than those without a migration background.” In a similar vein, respondents find it unfair if people with an immigrant background pay higher prices (t (998) = –29.19, p < .001, Mean Difference = –1.01), and unacceptable (t (998) = –27.78, Mean Difference = –.97). Figure 9 shows the results. All in all, respondents find it unfair and unacceptable if poorer consumers pay more for insurance.
V. Discussion
In this section, we reflect on the main findings from the survey. First, broadly speaking, if respondents find a practice unfair, they also find the practice unacceptable. For ease of reading, we only speak about “fair” and “unfair” below, rather than about “fair and acceptable” and “unfair and unacceptable”. Second, the survey respondents find almost all modern insurance practices that we described very unfair. Insurers who want to apply data-intensive underwriting or behaviour-based insurance should be aware that many people dislike those practices. Behaviour-based car insurance seems the least unfair of the practices: respondents found behaviour-based car insurance slightly fairer than neutral. This gives some preliminary evidence for a standard theory in insurance fairness: some say that behaviour-based insurance is more “actuarially fair” than traditional insurance.Footnote 38 A more expansive survey or experiment could confirm whether the same results hold.
Third, respondents find insurance practices fairer if they feel that they can influence the premium. This view of respondents is understandable. We can imagine that it feels unfair to be punished with a higher premium for your car insurance, just because you live in an apartment with a letter in the house number. Meanwhile, insurers sometimes set premiums based on such data (see Section II). In the future, insurers may use more data on which a consumer has little influence on to set premiums. By carrying out underwriting more data intensively, insurers could find more correlations that help them to predict claim costs. After all, machine learning is good at finding correlations in large datasets. It may be bad news for insurers that many consumers dislike it if insurers calculate premiums based on characteristics that the consumer cannot influence.
We asked about one other type of behaviour-based insurance: life insurance with discounts based on a step counter. Respondents find such step counter insurance unfair. What could explain this difference between respondents’ fairness perceptions about behaviour-based car insurance and behaviour-based insurance based on a step counter? We provide some preliminary thoughts. First, perhaps people feel that they cannot help it if they make few steps per day. Perhaps their legs hurt, or perhaps physical activity does not fit in with their busy schedules. People might think that, by comparison, they have more influence on their driving style. Second, perhaps some people do physical activity (e.g., strength training) that a step counter does not register well. Such people might see a step counter as a bad way to register physical activity. Third, we asked about a discount if you register more than 10,000 steps. Perhaps people’s reaction of respondents would have been different if we asked them about a different number of steps. More research on such questions would be welcome.
Fourth, respondents dislike it if insurers use seemingly illogical consumer characteristics to set premiums. A controversial characteristic that respondents found logical is Age. Perhaps in the scenario we presented, age was deemed fairer. Other surveys could ask about other situations. Respondents consider behaviour-based the most logical, but this does not seem true for the step counter. This could be because of the scenario we presented (10,000 steps). Respondents find it unfair if an insurer calculates the premium based on a consumer characteristic when the respondent does not see the logic, or relevance, of using that characteristic. The results raise questions about whether insurers can underwrite data-intensively in a way that consumers find fair. Insurers may find it difficult to convince consumers of the logic or relevance of the data that they use for underwriting, even if in the eyes of the insurer, the data are relevant. Suppose that an insurer only wants to use consumer characteristics to set premiums if it can explain the logic or relevance of using that characteristic. Such a fairness policy might mean that the insurer cannot use certain consumer characteristics, even though an AI system has found a correlation between those characteristics and filing claims.
Fifth, respondents dislike target group insurance. Respondents find all types of insurance that are only available for certain groups, such as car insurance specifically for family doctors, unfair. Meanwhile, some insurers do offer such target group insurance. Because in the Netherlands, target group insurance for doctors is long established, it is remarkable that respondents seem to consider all target group insurances equally unfair. Perhaps people feel discriminated if they are not allowed to buy certain insurance products. Perhaps people’s opinions change if an insurer gives arguments for a certain type of target group insurance. Sixth, respondents find it unfair if poor people pay higher insurance prices. Maybe respondents feel that such practices discriminate against poor people.
An insurer could try to make practices fairer by using a technical “fairness metric” that helps to prevent some unfair practices. A fairness metric is, roughly speaking, a score that expresses how fair a system is in terms of, for example, the number of false positives and false negatives).Footnote 39 However, for the examples we used in our survey, many fairness metrics do not seem suitable. Many metrics, such as Equality of Odds, focus on the outcome rather than whether a characteristic seems logical or is influenceable for the consumer.Footnote 40 Such metrics are useful, for example, if an insurer wishes to prevent the poor paying more. Perhaps metrics based on causality or control could provide some insight into the logic behind certain characteristics and whether consumers can influence the characteristic. But even then, it may be difficult for an insurer to foresee what characteristics and practices consumers see as logical.Footnote 41 Similar to recent fairness research, we think that bringing down fairness to a single metric is difficult.Footnote 42 Explainability may help insurers understand the correlations they come across while applying machine learning to new datasets. But even if insurers can explain a decision to the consumer, that explanation does not necessarily make the decision itself fairer.Footnote 43 For example, even if an insurer explains why it finds a house number relevant for car insurance, the consumer may disagree with the explanation. Most consumers are not statisticians.Footnote 44
It seems that the law in Europe can hardly protect people against the unfairness types highlighted in this paper. For instance, most insurance practices described in this paper do not violate non-discrimination law. Most non-discrimination statutes only protect people against discrimination based on certain protected characteristics, such as age, disability, gender, religion, ethnicity or sexual orientation. Through the concept of indirect discrimination (disparate impact) some insurance practices might be illegal, if they disproportionally harm people with a certain ethnicity or other protected characteristic.Footnote 45 Most rules in the AI Act only apply to health and life insurance, not to other types of insurance. And even for health and life insurance, the AI Act does not ban or clearly regulate the practices described in this paper.Footnote 46 EU law specific for insurance does not focus on the unfairness types in this paper either.Footnote 47 The best source for some, albeit limited, legal protection is probably the General Data Protection Regulation.Footnote 48 The GDPR may limit the amount of data that insurers can gather and analyse, and the GDPR’s transparency requirements could help to make some insurance practices less of a black box.Footnote 49 The GDPR’s rules on certain fully automated decisions with legal or similar effects may also offer some protection.Footnote 50 But the GDPR does not outright ban data-intensive underwriting or behaviour-based insurance. All in all, it appears that current law can offer, at most, some limited protection against the perceived unfairness discussed in this paper. An in-depth analysis of possible legal protection falls outside the scope of this paper.
We do not think that policymakers should ban every practice that people dislike or find unfair. To illustrate: many people dislike paying taxes. But the state needs tax income to run the country. Policymakers should consider many aspects of a problem. Similarly, we do not think that a practice is always morally fine, if a majority agrees with the practice. In our opinion, policymakers should give some attention to what people find fair and acceptable according to surveys, as long as policymakers scrutinise the survey in question, e.g., for data quality.Footnote 51
Finally, this paper did not show that, according to the respondents, certain insurance practices must be prohibited by a lawmaker or withdrawn by an insurer. What people consider unacceptable may not necessarily overlap with their buying decisions – a “fairness paradox.” People’s stated preferences (what people say in surveys) may differ from their observed behaviour (how people act). For example, somebody might say that it is unacceptable if a life insurer offers discounts to people who show an active lifestyle with a step counter, but that same person might buy such an insurance product because the option is more affordable. Related discussions exist about a “privacy paradox”: stated privacy preferences versus privacy behaviour.Footnote 52 Even if a fairness paradox exists, our paper can still inform the debate about the fairness and acceptance of data-intensive underwriting and behaviour-based insurance.
VI. Limitations and suggestions for further research
In this section we highlight limitations of this research and give suggestions for further research. This survey research was a first attempt to learn more about people’s opinions about two modern trends in insurance. The sample of 999 respondents gives a reasonable cross-section of Dutch society, but the sample is not fully representative.
There are exciting opportunities for further research, in many disciplines. Social scientists could do more empirical research. Survey research in more countries could highlight the differences between countries. And surveys could contain more, and more in-depth, questions about issues such as behaviour-based insurance, health insurance and insurance practices that reinforce financial inequality. Focus groups or interviews could provide more insight into why people find certain practices unfair. Further research could investigate whether there is there a “fairness paradox” in relation to data-intensive underwriting and behaviour-based insurance, i.e., the relationship between people’s stated fairness preferences and their buying decisions.
Computer scientists could analyse whether fairness metrics, as proposed by many papers of the ACM FAccT Conference for example, could help to protect people against the types of (perceived) unfairness as highlighted in this paper.Footnote 53
There are many open normative and ethical questions. For instance, from a normative perspective, when is it acceptable for insurance practices to increase financial inequality? Is it fair for people to bear the cost of characteristics or behaviours that they have no or little control over or find illogical?Footnote 54 Could a comprehensive normative framework be created for insurance?Footnote 55 Is behaviour-based insurance fairer than traditional insurance or data-intensive underwriting? Should insurers communicate the shift to more behaviour-based insurance with language implying solidarity or with language implying rationality?Footnote 56
Many legal questions deserve more research too. Perhaps an in-depth analysis of current law could show that the GDPR or other law offer some protection against the perceived unfairness discussed in this paper. And maybe additional regulation should be considered. Should policymakers do more to mitigate the risk that insurance practices reinforce social inequality? There is some precedent: in many situations, the law aims to protect the weaker party, such as in the field of labour law and consumer protection law. Are sector-specific rules needed for the use of AI in the insurance sector? Should the law do more to ensure that consumers are not excluded from certain insurance products, or should contract freedom stay the same? Normative and ethical research could discuss whether, and if so when, it is appropriate for the law to intervene.
VII. Concluding remarks
With a survey in the Netherlands, we explored people’s opinions about two modern trends in insurance: (i) data-intensive underwriting and (ii) behaviour-based insurance. The main results include the following. First, if survey respondents find an insurance practice unfair, they also find the practice unacceptable. Second, respondents find almost all modern insurance practices that we described unfair. Third, respondents find practices fairer if they can influence the premium. For example, respondents find behaviour-based car insurance based on a car tracker relatively fair. Fourth, respondents find it unfair if an insurer calculates the premium based on a consumer characteristic, while the respondent does not see the logic of using that characteristic. Fifth, respondents find all types of insurance that are only available for certain groups, such as a car insurance specifically for family doctors, unfair. Sixth, respondents find it unfair if an insurance practice leads to higher prices for poorer people. Overall, respondents worry about these practices, finding many of them unfair. Many open questions remain, for many disciplines: we suggest further research for social scientists, legal scholars, ethicists and computer scientists.
Acknowledgments
The authors thank Jeremias Adams-Prassl and the participants of the iHub Automated Decision-making Club (ADM), the Radboud Data Science Seminar (DSS) and the Second ENS Fellows and Friends Paper Workshop for their useful comments. The research was partially funded by Radboud AI, for the project “How do people feel about AI-driven decision-making?” https://www.ru.nl/en/departments/interdisciplinair/radboud-ai.
Funding Statement
Open access funding provided by Radboud University Nijmegen.
Appendix A
Table A1. Fairness by data type

Note. Negative t-values represent a negative deviation from the scale midpoint, whereas positive t-values represent positive deviations from the scale midpoint. Any mean differences are significant at the p < .001 level.
Table A2. Acceptance by data-type

Note. Negative t-values represent a negative deviation from the scale midpoint, whereas positive t-values represent positive deviations from the scale midpoint. Any mean differences are significant at the p < .001 level.
Appendix B
Table B1. Influence by data-type

Note. Negative t-values represent a negative deviation from the scale midpoint, whereas positive t-values represent positive deviations from the scale midpoint. Any mean differences are significant at the p < .001 level.
Table B2. Perceived logic by insurance practice

Note. Negative t-values represent a negative deviation from the scale midpoint (except Age, which is not significant from the scale’s midpoint), whereas positive t-values represent positive deviations from the scale midpoint. Any mean differences are significant at the p < .001 level.
Appendix C
Table C1. Fairness by target group

Note. Negative t-values represent a negative deviation from the scale midpoint, whereas positive t-values represent positive deviations from the scale midpoint. Any mean differences are significant at the p < .001 level.
Table C2. Acceptance by target group

Note. Negative t-values represent a negative deviation from the scale midpoint, whereas positive t-values represent positive deviations from the scale midpoint. Any mean differences are significant at the p < .001 level.
Appendix D. Survey questions on life and burglary insurance
“With life insurance, you pay a higher premium if the insurer expects you to die earlier. Because if you die earlier, you will pay a premium for a shorter period. The insurer wants to charge higher premiums to people who are likely to die earlier. In the Netherlands, poorer people live on average seven years shorter than richer people. The insurer therefore charges higher premiums to poorer people. People with a certain income often live in certain neighbourhoods.
The insurer charges higher prices in poorer neighbourhoods.”
“With burglary insurance, you pay a higher premium if the risk of burglary in your home is higher. Suppose poorer neighbourhoods have more burglaries. An insurer charges higher premiums in neighbourhoods with more burglaries. This leads to poorer people paying higher premiums on average.”






