1. Introduction
Philosophers and computer scientists debate whether there exist statistical criteria of fairness that can be used to assess the fairness of a machine learning (ML) model’s predictions. In this article, I bring attention to the additional values that are involved when these or other notions of fairness are integrated into an optimization procedure used to train an ML model. I focus on a major design choice in multiobjective optimization problems (MOOPs) between biased and unbiased search strategies, which are distinguished by whether they use priorities for the objectives to direct an optimization procedure. I argue that both reliably generate Pareto optimal (PO) solutions, but there are various additional, not purely epistemic, reasons to prefer one or the other.
The article will proceed as follows. In section 2, I explain how my analysis diverges from existing philosophical work on algorithms and values, and I specify the sense in which I consider values to be related to model design choices. In section 3, I introduce MOOPs and strategies for solving them, and I motivate why it is important to understand fairness optimization as a MOOP. In section 4, I argue that the methodological choice between biased and unbiased search strategies is unforced by the epistemic aim of finding PO trade-offs. In section 5, I discuss what additional values are relevant to this choice.
2. Values act as justificatory reasons for choices in optimization problems
Philosophers and social scientists have brought critical attention to the multitude of ways that algorithms are biased, often leading to unfairness when algorithmic predictions inform decisions. In this work, I offer a novel philosophical reflection on how algorithms implement values by focusing on the choice of what optimization method to use when incorporating fairness notions into ML design. This goes beyond existing philosophical work that highlights the relationship between values and statistical performance metric(s), including trade-offs between statistical criteria of fairness, and the values pertinent to data-driven practices as a whole (e.g., Fazelpour and Danks Reference Fazelpour and Danks2021). While fairness criteria can be incorporated at various stages of ML model design, including before, during, or after an optimization algorithm is run, my purpose here is to compare approaches that incorporate fairness notions into optimization design.
Because my goal is to analyze what justifies particular design choices in optimization, I will focus on a specific type of relation between values and choices where values act as justificatory reasons for choices. This type of relation is distinguished by Ward (Reference Ward2021), who characterizes it as involving an appeal to values in rational arguments used to support a certain course of action. Using Ward’s taxonomy, I will not attempt to answer what values motivate ML designers to choose various optimization strategies, which would require a separate psychological and sociological analysis. I will also not consider how values act as causes or effects of design choices—for instance, what values make designers and decision makers (DMs) more likely to formulate a decision as an ML problem or as a MOOP, or how a deployed ML model promotes certain social values and practices. Instead, I will focus on how values are (and should be) used as justificatory reasons in choices regarding the use of a particular optimization strategy. Notably, following Ward and others, my assessment focuses on pro tanto reasons for choices rather than on what reasons fully justify a choice.
3. Optimizing for fairness in ML is a multiobjective problem
Trade-offs arise in predictive modeling that aims to be fair in several ways, indicating that the problem of implementing almost any fairness notion in ML optimization is multiobjective. First, predictive modeling involves a cost-versus-benefit analysis. For example, solving a classification problem involves a search for a model that makes an optimal trade-off between the costs and benefits of each type of predictive result (e.g., positive or negative classifications, whether true or false). In supervised ML, these predictive preferences are typically chosen by the DM and encoded as a single performance metric that may be implemented as the objective function to optimize or used to evaluate the model’s final performance. Either way, during the training phase, the optimization algorithm varies the model’s parameters to minimize an objective function (also called the “loss” function) that expresses the loss of the costs relative to the benefits when predicting the target variable for the population represented in the training data set. The final trained model generates individual predictions based on an optimal risk score that represents the frequency of the target outcome given a particular set of features (in a Bayesian sense, it represents the posterior probability of the outcome given the observed features; Barocas et al. Reference Barocas, Hardt and Narayanan2023, 48).
Nonetheless, risk scores that are optimal for the whole population may not correspond to what is optimal for various subgroups, such that striving for fairness to groups involves trade-offs with overall predictive performance. Barocas et al. (Reference Barocas, Hardt and Narayanan2023) bring attention to how, in general, we should not expect the costs and benefits of predictions to be equally shared across groups if the predicted target variable is not statistically independent of group membership. Here, attempts to equalize various statistical criteria of fairness across groups generally decrease overall predictive performance. The only exceptions are fairness criteria based on the notion of sufficiency, which require that the risk score is well calibrated to the predicted outcomes within groups (Barocas et al. Reference Barocas, Hardt and Narayanan2023, 62). Yet even sufficiency trades off with population accuracy if models are “blinded” to group membership (e.g., for legal reasons) and the relationship between predictive features and outcomes differs across groups (Corbett-Davies et al. Reference Corbett-Davies, Gaebler, Nilforoshan, Shroff and Goel2023, 16–17).
Meanwhile, different statistical notions of what constitutes fairness to groups also involve inherent trade-offs such that employing more than one criterion incorporates conflicting goals into a predictive modeling problem. Barocas et al. (Reference Barocas, Hardt and Narayanan2023) categorize statistical nondiscrimination criteria into three types based on the notions of sufficiency, independence, and separation. Independence requires that a risk score is independent of a sensitive attribute such that the acceptance rate of a classifier is equal across groups. On the other hand, separation requires that a risk score be independent of a sensitive attribute within each stratum of equal claim, implying error rate parity. Barocas et al. prove that formal trade-offs exist between each family of criteria when the target variable is not independent of group membership (except for degenerate solutions). Thus, there is a three-way trade-off between the notions of sufficiency, independence, and separation such that if one is satisfied, the others cannot be (see also Chouldechova Reference Chouldechova2017; Kleinberg et al. Reference Kleinberg, Mullainathan and Raghavan2016; Miconi Reference Miconi2017).
Last, implementing even one statistical fairness criterion in a predictive modeling problem can benefit one demographic group while imposing significant costs for other subgroups of the population. Kearns et al. (Reference Kearns, Neel, Roth and Steven Wu2018) call this “fairness gerrymandering,” where an appearance of equity with respect to one subgroup comes at the expense of unfairness toward another. Thus, the challenge of achieving fairness with respect to multiple subgroups increases the number of conflicting goals in a predictive modeling problem.Footnote 1
Philosophers and computer scientists respond to these inherent trade-offs with the following four main claims. (1) Only one statistical criterion is normatively relevant (e.g., separation [Hellman Reference Hellman2020; Grant Reference Grant2023], calibration [Hedden Reference Hedden2021; Corbett-Davies et al. Reference Corbett-Davies, Gaebler, Nilforoshan, Shroff and Goel2023]). (2) Necessary statistical criteria of fairness should be assessed on a case-by-case basis, especially because it is possible to satisfy some measures from more than one family at the expense of others in those same categories (Miconi Reference Miconi2017). (3) The criteria could be relaxed, substantially altered to be compatible (Beigang Reference Beigang2023), or implemented as group-specific aims rather than as a summarized statistic (e.g., group distributionally robust optimization [DRO; Sagawa et al. Reference Sagawa, Koh, Hashimoto and Liang2020], minimax Pareto fairness [Martinez et al. Reference Martinez, Bertran and Sapiro2020]). (4) The criteria should be abandoned in favor of alternative debiasing or fairness-optimization strategies, such as striving toward a perfect predictor (Miconi [Reference Miconi2017] points out this option), implicit unfairness mitigation methods (see Wan et al. Reference Wan, Zha, Liu and Zou2023), or direct cost–benefit analysis using real-world quantities (Corbett-Davies et al. Reference Corbett-Davies, Gaebler, Nilforoshan, Shroff and Goel2023).Footnote 2
The diversity of these responses raises the question: What might a focus on multiobjective optimization add to this debate? I propose that analyzing the values that act as justificatory reasons for choices regarding how to optimize for multiple aims is part of the normative work required for establishing the legitimacy of almost any of the proposed approaches because they involve design choices that may or may not lead to Pareto optimality (defined later), and that might imply more or less precision in how well the solutions correspond to a DM’s preferences. (The main exception is for a perfect predictor [Miconi Reference Miconi2017].) For example, if only one statistical criterion is required for fairness in general or in a particular case, then it likely involves a trade-off with population performance (even for sufficiency, if a “blind” model is used). The advantage of recognizing the multiobjective nature of these optimization problems is that design choices may be evaluated for whether they generate a PO solution: one that offers the best possible population performance for the required fairness (e.g., Zafar et al. [Reference Zafar, Valera, Rodriguez and Gummadi2017] consider these and contrasting cases of “business necessity”). Alternatively, one or more statistical criteria of fairness might be relaxed in order to find an intermediate compromise with population accuracy or other fairness objectives. In these cases, treating the problem as a multiobjective one enables designers to plan for a PO trade-off solution: one without any unnecessary relaxation. Even Beigang’s (Reference Beigang2023) approach, which offers alternative, compatible criteria of algorithmic fairness, still does not reconcile the multiple goals of distributive justice that might arise from algorithmic decision making. Meanwhile, optimizing group-specific aims (without summary statistics) is also an inherently multiobjective problem, where Pareto optimality often matters for doing no unnecessary harm (Martinez et al. [Reference Martinez, Bertran and Sapiro2020] raise this point) but is not guaranteed by some approaches (e.g., group DRO).
In a MOOP, the main aim is to find one or more PO solutions that represent the best possible trade-offs between the objectives. A MOOP can be formulated as a minimization problem with m objectives (
$m\rm2$
) that map decision variables
$x \in X$
in decision space X into the objective space, defined in
${\mathbb{R}^m}$
:
A solution is PO if and only if it is not possible to improve in any one objective without degrading in at least one other. The notion of dominance distinguishes PO solutions from non-PO solutions: A solution x (1) dominates x (2) (equivalently, x (1) is nondominated by x (2)) if and only if two conditions hold (following the formalism of Deb et al. [Reference Deb, Sindhya and Hakanen2016, 151]):
1. x
(1) is no worse than x
(2) in all objectives (
${f_j}\;\left( {{{\bf{x}}^{\left( 1 \right)}}} \right)\; {\triangleright \hskip-0.5pc/} \;{f_j}\;\left( {{{\bf{x}}^{\left( 2 \right)}}} \right)$
for all j = 1, …, m).
2. x
(1) is strictly better than x
(2) in at least one objective (
${f_{\bar j}}\;\left( {{{\bf{x}}^{\left( 1 \right)}}} \right)\; \triangleleft \;{f_{\bar j}}\;\left( {{{\bf{x}}^{\left( 2 \right)}}} \right)$
) for at least
one
$\bar j$
$ \in $
{1, …, m}.
The PO set comprises the nondominated solutions attained when the entire feasible region of the decision space is searched (Deb et al. Reference Deb, Sindhya and Hakanen2016, 153). In most MOOPs, feasible solutions are defined by various constraints, including limits on the range of each parameter value that is searched. This means that in the objective function space, PO solutions mark the boundary between the feasible and infeasible regions (see fig. 1).

Figure 1. Trade-off solutions in objective function space when minimizing some notion of unfairness (F2) and population loss (F1).
Optimization strategies that implement multiple objective functions can be characterized according to two major design choices. The first choice depends on when the DM specifies preferences for how to trade off the multiple objectives. In a priori approaches, the DM decides before the stage of optimization, and these preferences are used to obtain a single PO solution. In contrast, a posteriori approaches present the DM with multiple PO solutions to choose from, representing different trade-offs, after the stage of optimization.Footnote 3 Typically, a posteriori approaches aim to find a large portion of the entire PO set (Deb et al. Reference Deb, Sindhya and Hakanen2016, 148). I will focus this article on a second major design choice, which concerns whether to use a search strategy that is guided toward a particular trade-off solution. I will use the term biased search for approaches that aim to find the single PO solution that corresponds to a certain set of priorities for the multiple objectives. They can be used in either an a priori or an a posteriori way, depending on whether the priorities are iteratively varied to map the Pareto front (PF). On the other hand, I will call search strategies that do not assign priorities to the multiple objectives as an optimization proceeds unbiased (i.e., in this limited sense; see table 1).
Table 1. Multiobjective Fairness-Optimization Strategies That Generate Pareto Optimal Solutions
| A Priori | A Posteriori | |
|---|---|---|
| Biased Search | • Linear scalarization (Kamishima et al. Reference Kamishima, Akaho, Asoh, Sakuma, Peter, De Bie and Cristianini2012) • Constrained optimization (Zafar et al. Reference Zafar, Valera, Rodriguez and Gummadi2017) • Rawlsian minimax Pareto fairness (Martinez et al. Reference Martinez, Bertran and Sapiro2020) |
• Epsilon constraint (Liu and Vicente’s 2022 extension of Zafar et al. Reference Zafar, Valera, Rodriguez and Gummadi2017) • Chebyshev scalarization (Wei and Niethammer Reference Wei and Niethammer2022) |
| Unbiased Search | • Constrained multigradient descent followed by automated selection of the final solution (Padh et al. Reference Padh, Antognini, Lejal-Glaude, Faltings and Musat2021) | • Stochastic multigradient descent (Liu and Vicente Reference Liu and Nunes Vicente2022) |
4. Choosing between biased and unbiased search strategies is an unforced methodological choice
Here, I argue that biased and unbiased search strategies for multiobjective optimization both reliably generate PO solutions, even for ML problems that indirectly optimize the goals. This implies that choosing between these strategies is an “unforced” methodological choice (Winsberg Reference Winsberg2012).
Biased search strategies often make use of optimization engines that offer guarantees of convergence (for convex functions), reducing the multiple objectives into a single function while incorporating priority information (e.g., as constraints on a primary objective [Zafar et al. Reference Zafar, Valera, Rodriguez and Gummadi2017] or as weights for a scalarizing function [Kamishima et al. Reference Kamishima, Akaho, Asoh, Sakuma, Peter, De Bie and Cristianini2012]). For example, many deterministic algorithms are guaranteed to converge on the global optimal solution to a convex objective function (where every local minimum is a global minimum, e.g., quadratic programming, gradient descent, Newton’s method). Alternatively, stochastic optimization engines also reliably converge on global optima (of either convex or nonconvex functions). To illustrate, appropriately sized stochastic variation of parameter values can prevent premature convergence to local but not global optima.
While unbiased search strategies optimize multiple objectives simultaneously rather than reducing them, they also reliably generate PO solutions. Unbiased strategies first attempt to map the PF and then incorporate a DM’s preference information (specified either a priori or a posteriori) to select a PO solution. In ML, unbiased methods map the PF by maintaining a list of nondominated solutions in various ways; for example, by comparing the list to new solutions generated at each epoch (Padh et al. Reference Padh, Antognini, Lejal-Glaude, Faltings and Musat2021) or to solutions generated by multiple runs of the algorithm with randomized initial starting points (Liu and Vicente Reference Liu and Nunes Vicente2022). Multigradient descent algorithms follow a vector of common descent (that represents a convex combination of the objective function gradients) until it vanishes. The resulting “Pareto stationary” solutions can be compared in terms of dominance. In general, repeatedly vetting solutions by dominance improves the quality of the PF.
Nonetheless, I submit a few remarks on the reliability of ML-specific multiobjective optimization because ML uses an “indirect” form of optimization (Goodfellow et al. Reference Goodfellow, Bengio and Courville2016, 268). In supervised ML, the ultimate performance of a model is judged by evaluating the objective function(s) on a specially reserved subset of the data (called the test set) that is not used during optimization. Still, granting that the same performance measure(s) (or suitable approximations) are used for both subsets and that standard ML assumptions hold (the training and test sets are independent and identically distributed), biased or unbiased search strategies also reliably generate PO solutions in ML.Footnote 4
Empirical evidence also supports my claim. Liu and Vicente (Reference Liu and Nunes Vicente2022) develop an unbiased search method for multiobjective fairness optimization based on stochastic multigradient descent (with the goal of using it in an a posteriori way). They compare their method to a biased search approach (the constraint-based optimization of Zafar et al. [Reference Zafar, Valera, Rodriguez and Gummadi2017]) on two convex optimization problems. The goals are to make accurate predictions of salary and minimize disparate impact with respect to either gender or race (they use the Adult Income data set; Kohavi and Becker Reference Kohavi and Becker1996). For accuracy, they use a logistic regression function, and for disparate impact, they use a convex approximation of Calders and Verwer’s (Reference Calders and Verwer2010; CV) score based on the decision boundary covariance. The comparison requires iterating over the constraint parameter in Zafar et al.’s method and storing nondominated solutions in order to obtain a full PF with the biased search strategy (i.e., adapting it to an epsilon-constraint method; see Haimes [Reference Haimes1971]). Liu and Vicente’s (Reference Liu and Nunes Vicente2022) results show that both methods generate high-quality PFs: When optimizing for minimal disparate impact with respect to gender, the Pareto front–stochastic multigradient (PF-SMG) algorithm dominates the epsilon-constraint (EPS)-fair algorithm in some regions of the trade-off surface (their fig. 1a, 521). However, when optimizing for minimal disparate impact with respect to race, the EPS-fair algorithm dominates in some regions (their fig. 2a, 522). Notably, the DM’s preferences (specified either a priori or a posteriori) might require targeting a PO solution from any region of these trade-off surfaces, where either method might dominate. Thus, neither approach offers an overall advantage in terms of Pareto dominance for these tests.Footnote 5
This means that values are relevant to normatively assessing the choice between biased and unbiased search strategies for finding PO solutions. Particularly, in the sense articulated by Ward (Reference Ward2021), values might act as nonepistemic justificatory reasons. Winsberg (Reference Winsberg2012) brings attention to how nonepistemic values affect what he calls “unforced methodological choices” in climate modeling, where one option is not “objectively” better than another, but each presents a benefit for a different set of preferences, such as preferences for inductive risks (130). While he argues that it is not possible to isolate the values that have affected the history of climate science because they hide in all its “nooks and crannies,” and current climate models also rely on past modeling choices, methods for fairness optimization are not yet so “generatively entrenched” (132). Therefore, I now suggest several values that are relevant to the choice between biased and unbiased search strategies.
5. Values in multiobjective optimization
There are several differences between biased and unbiased search procedures beyond their ability to generate PO solutions that might act as justificatory reasons for choosing between them. Here, I discuss three: precision in selecting a PO solution that corresponds to a certain set of preferences, computational efficiency, and dynamic adaptability. While not all of these appear to be purely epistemic, I do not attempt to classify them as epistemic or nonepistemic or to regress to the problem of identifying values that act as reasons for choosing between them.
First, perhaps counterintuitively, unbiased methods appear to have an overall advantage in offering more precise control over solution preferences. One reason for this is that it is somewhat difficult to achieve a dense and well-spread PF with a biased method. For weight-based methods, an evenly distributed set of weights does not necessarily generate an even spread of PO solutions (Deb et al. Reference Deb, Sindhya and Hakanen2016, 160). Meanwhile, constraint-based methods require advanced knowledge of the PF to achieve a sufficient resolution, including what upper bounds to use for the constraints and what step size to vary them with. Liu and Vicente’s (Reference Liu and Nunes Vicente2022) results demonstrate this difficulty: Their unbiased method produces a more dense and well-spread PF compared with the epsilon-constraint method (522–23). Also, the overall advantage of unbiased methods in generating higher-resolution PFs applies even when preferences are specified a priori. For example, if constraints are set too narrowly, no feasible solution will be obtained, and if they are broad enough to admit more than one PO solution, a highly resolved PF affords better fine-grained solution control. In addition, while it is possible in principle with some biased methods to discover nonconvex regions of a PF (e.g., epsilon constraint, Chebyshev scalarization), the challenge of tuning the weights or step size of the constraints means that it is more difficult than with unbiased methods. For this reason, designers of biased methods often choose to use convex approximations of the objective functions (e.g., Zafar et al. Reference Zafar, Valera, Rodriguez and Gummadi2017).
Second, biased and unbiased methods differ in the efficiency with which they generate one or more PO solutions. Because biased methods execute single-objective optimization routines, they are most efficient for a priori problems that aim to generate a single PO solution. However, they require iterative optimizations for solving a posteriori problems, so they may be inefficient for problems with many objectives (e.g., fairness optimization; see sec. 3). Also, if the objective functions are nonconvex, biased methods require iterative exploration of the PF to discover any nonconvex regions, which also makes them less efficient overall. On the other hand, while unbiased methods might be slightly less efficient at generating single PO solutions, they outperform biased methods in efficiently mapping the PF (Liu and Vicente [Reference Liu and Nunes Vicente2022] also measure this explicitly). Thus, different measures of computational efficiency are relevant to choosing between biased and unbiased methods.
Third, unbiased methods have the advantage that the PF may be dynamically adapted as new data becomes available. Liu and Vicente (Reference Liu and Nunes Vicente2022) highlight that this is an important feature of their design, and they present results that simulate a streaming scenario: The PFs they compute on successive batches of the training data adaptively converge to the final PF computed for the whole training data set (527, 535). Optimizing the multiple objectives simultaneously and stochastically sampling the data makes this efficient; it does not require restarting a series of biased searches to compute a new PF. Instead, new solutions computed with fresh data can be directly compared with previous solutions in terms of dominance. The ability to handle streaming data is important because in realistic socioeconomic systems, PFs evolve dynamically.
In sum, while neither biased nor unbiased search methods offer an overall advantage in terms of Pareto dominance, biased searches might be preferred for efficiently computing a single PO solution, and unbiased methods might be preferred for precise control over trade-off preferences, efficient PF mapping, and dynamic PO prediction.
6. Conclusion
I emphasized that because implementing fairness notions in ML optimization is generally a multiobjective problem, it is important to normatively assess the design choices involved in various methods that incorporate more than one objective in optimization. I argued that choosing between biased and unbiased search strategies is not forced by the epistemic aim of finding nondominated, PO solutions. Instead, values such as computational efficiency and preferences regarding various types of predictive risks (precision in trade-off control, robustness to distribution shifts) help justify this choice.
6.1. Future Work
My analysis is relevant to future ethical reflection on the design of fair algorithms—for instance, work that assesses the normative significance and priority that should be given to various values, such as the ones I have highlighted here. In particular, the importance of Pareto optimality must be ethically grounded, and here, I briefly highlight a couple of possible approaches stemming from Grant et al.’s (Reference Grant, Behrends and Basl2025) proposal regarding duties of evidential consideration owed to decision subjects. First, the accuracy of a predictive model relative to other available predictive models is a morally salient feature for showing due consideration to the substantive claims of decision subjects (11). Thus, if an optimization method that does not reliably generate PO solutions is chosen (e.g., group DRO; Sagawa et al. Reference Sagawa, Koh, Hashimoto and Liang2020), a decision subject may not have been shown due consideration because better methods exist for safeguarding against unnecessary loss of accuracy (i.e., loss that is not necessary for gain in some other normative goal). Second, optimizing for population accuracy might also raise duties to decision subjects to safeguard against unnecessary differential accuracy between groups or unnecessary loss of other kinds of fairness. These duties would safeguard against insensitivity to the multidimensional nature of the salient moral claims rather than establishing general priorities for various normative goals.
Also, the choice between a priori and a posteriori approaches should be normatively assessed. If a DM has clear requirements for trade-offs, it seems to justify the use of an a priori approach. However, if a DM is unable to precisely specify preferences, a posteriori approaches provide useful information for making ethical judgments: They can be used to identify “knee” solutions, where the curve of the PF changes rapidly between two objectives such that a small improvement in one objective causes a large degradation in at least one other. Branke et al. (Reference Branke, Deb, Dierolf, Osswald, Xin Yao, Burke, Lozano, Merelo-Guervós, Bullinaria, Rowe, Tiňo, Kabán and Schwefel2004) argue that in the absence of preference information, knee solutions are most likely to be interesting to a DM because other nearby solutions involve a large loss in at least one objective (i.e., knee solutions are locally nonextreme).
Acknowledgments
Thanks to support from the Northeastern Artificial Intelligence and Data Ethics (AIDE) Summer Training Program 2023 and National Science Foundation Award 2147220. This article also draws on research supported by the Social Sciences and Humanities Research Council of Canada (Ce article s’appuie sur des recherches financées par le Conseil de recherches en sciences humaines du Canada). I would also like to thank Chris Smeenk and Kathleen A. Creel for helpful discussions.

