Abstract
Kearns and Star have previously recommended that we measure the degree to which a reason supports a conclusion, either about how to act or what to believe, as the conditional probability of the conclusion given the reason. I show how to properly formulate this recommendation to allow for dependencies and conditional dependencies among the considerations being aggregated. This formulation allows us to account for how considerations, which do not themselves favour a specific conclusion, can modify the strength of a reason for that conclusion, and thus to explain the intensifiers and attenuators described by Dancy. The formulation also accounts for the workings of partial undercutters in epistemology. I then show how my account avoids the counterexamples that Brunero levied against probability-based theories of the strengths of reasons. My account supports the theory, suggested by Kearns and Star, that the strengths of reasons are measured by conditional probabilities. If my account is successful, then it will count in favour of the idea that the strengths of reasons are measured on the same scale as are conditional probabilities.
Similar content being viewed by others
References
Anderson, S. (2007). Combining evidence using bayes' rule. <cs.wellesley.edu/~anderson/writing/naive-bayes.pdf>
Broome, J. (2008). Reply to Southwood, Kearns and Star, and Cullity. Ethics, 119, 96–108.
Broome, J. (2013). Rationality through Reasoning. Wiley-Blackwell.
Brunero, J. (2009). Reasons and evidence one ought. Ethics, 119, 538–545.
Brunero, J. (2018). Reasons, evidence, and explanations. In D. Star (Ed.), The Oxford handbook of reasons and normativity (pp. 321–341). Oxford University Press.
Dancy, J. (2000). The particularist’s progress. In B. Hooker & M. Little (Eds.), Moral particularism. Oxford University Press.
Dancy, J. (2004). Ethics without principles. Clarendon Press.
Hawthorne, J., & Lasonen-Aarnio, M. (2009). Knowledge and objective chance. In P. Greenough & D. Prichard (Eds.), Williamson on knowledge. Oxford University Press.
Horty, J. (2007). Reasons as defaults. Philosophers’ Imprint, 7, 1–28.
Horty, J. (2012). Reasons as defaults. Oxford University Press.
Kagan, S. (1988). The Additive Fallacy. Ethics, 99, 5–31.
Kearns, S., & Star, D. (2008). Reasons: Explanations or evidence? Ethics, 118, 31–56.
Kearns, S., & Star, D. (2009). Reasons as evidence. In R. Shafer-Landau (Ed.), Oxford studies in metaethics (4) (pp. 215–242). Oxford University Press.
Kearns, S., & Star, D. (2013). Weighing reasons. Journal of Moral Philosophy, 10, 70–86.
Lord, E., & Maguire, B. (2016). An opinionated guide to the weight of reasons. In E. Lord & B. Maguire (Eds.), Weighing reasons (pp. 3–24). Oxford University Press.
Moss, S. (2018). Probabilistic knowledge. Oxford University Press.
Nair, S. (2016). How do reasons accrue? In E. Lord & B. Maguire (Eds.), Weighing reasons. Oxford University Press.
Nissan-Rozen, I. (2017). Reasoning with comparative moral judgements: An argument for Moral Bayesianism. In R. Urbaniak & G. Payette (Eds.), Applications of formal philosophy: the road less travelled (pp. 113–136). Springer.
Pollock, J. (1984). Reliability and justified belief. Canadian Journal of Philosophy, 14, 103–114.
Pollock, J., & Cruz, J. (1999). Contemporary theories of knowledge (2nd ed.). Rowman & Littlefield.
Schroeder, M. (2011). Holism, weight, and undercutting. Nous, 45, 328–344.
Schroeder, M. (2018). Getting perspective on objective reasons. Ethics, 128, 289–319.
Star, D. (2018). Introduction. In D. Star (Ed.), The oxford handbook of reasons and normativity (pp. 1–22). Oxford University Press.
Way, J. (2017). Reasons as premises of good reasoning. Pacific Philosophical Quarterly, 98, 251–270.
Williamson, T. (2002). Knowledge and its limits. Oxford University Press.
Acknowledgements
I would like to thank David White for editorial assistance in the preparation of this paper, and thank Erik Nelson, Sheldon Wein, Bill Seager, and members of the audience at a Dalhousie Philosophy Colloquium for helpful comments. A referee for this journal gave me excellent, critical, but much appreciated advice.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix: How to derive formula (G)
Appendix: How to derive formula (G)
Suppose we have a piece of moral reasoning where two reasons, R and \(R^{*}\), each either supports or counts against a conclusion, \({\Phi }\). To calculate the conditional credence of \({\Phi }\) based on the conjunction of R and \(R^{*}\), we start with the ratio-formula definition of conditional credence, \(Pr(A{|}B{)} = Pr\left( {A\& B} \right)/Pr\left( B \right)\) and substitute \({\Phi }\) for A and \(R\& R^{*}\) for B. By commuting the conjunctions in both the numerator and the denominator, we get:
From the ratio formula for conditional probability, we also get the product rule for the probability of a conjunction, \(Pr\left( {A \& B} \right) = Pr(A{|}B{)} \times Pr\left( B \right).\) When we apply the product rule to the numerator of (1) by substituting \(R^{*}\) for A and \(\left( {R\& {\Phi }} \right)\) for B, we get:
When we again apply the product rule to \({\text{Pr}}\left( {R\& {\Phi }} \right)\) in the numerator of (2), we get:
When we apply the product rule to the denominator of (3), we get:
By rearranging the terms in both the numerator and denominator of (4), we get:
Looking back into the main text, we can see that the last two terms on the right-hand side are the right-hand side of Bayes Theorem (BT).
Substituting this into (5) we get a general mathematical theorem (G) for the case of two reasons.
Notice that because all the operations involved in the proof are commutative, the proof will be symmetric in R and R*. We could equally well have derived the inverted formula, (G*).
We can extend the above derivation to produce corollaries of Bayes’ Theorem for inductive arguments that involve as many premises as we may wish, though the corollaries will rapidly become very complicated. (G) is a useful formula for combining the support offered by two moral premises into an overall degree of support for a conclusion. It shows right away that the degrees of support given to a conclusion by each of two considerations combine multiplicatively, not additively (Kagan 1998).
Rights and permissions
About this article
Cite this article
Kernohan, A. How to modify the strength of a reason. Philos Stud 179, 1205–1220 (2022). https://doi.org/10.1007/s11098-021-01703-3
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11098-021-01703-3