Loren K. Fryxell | Infinite Ignorance
This presentation was given at the 15th Oxford Workshop on Global Priorities Research, December 2024.
Loren K. Fryxell | Infinite Ignorance Read More »
Appear under internal videos on Talks page
This presentation was given at the 15th Oxford Workshop on Global Priorities Research, December 2024.
Loren K. Fryxell | Infinite Ignorance Read More »
This presentation was given at the 14th Oxford Workshop on Global Priorities Research, June 2024.
Tomi Francis | Aggregating Small Risks of Serious Harms Read More »
This presentation was given at the St Andrews-GPI Joint Workshop on the Long-Term Future, April 2024.
Elliott Thornley | A Non-Identity Dilemma for Person-Affecting Views Read More »
The sort of one-sentence summary of my talk is, when we’re thinking about benevolence, many people are naturally drawn to frame everything in terms of making a difference. And if you’re extremely careful about how you do it, which some people are, that is fine.
Hilary Greaves | On the desire to make a difference Read More »
So, as you know, our mission at GPI is to do foundational academic research on how to do the most good. And I thought I’d start by explaining one interpretation of what doing good is about. This isn’t really crucial to understanding the main part of the talk, but I thought it might put you in the right kind of mindset. So on this interpretation I’m going to give doing the most good is about reasons of beneficence…
Teru Thomas | The Multiverse and the Veil: Population Ethics Under Uncertainty Read More »
I want to explain why short-term proxies are a promising tool for improving our ability to make very long-term forecasts, mostly that is, before short-term proxies can help us evaluate forecasters that can incentivize information acquisition and that can be used for providing feedback for forecasting.
Benjamin Tereick | Creating short term proxies for long term forecasts Read More »
Longtermism holds roughly that in many decision situations, the best thing we can do is what is best for the long-term future. The scope question for longtermism asks: how large is the class of decision situations for which longtermism holds?
David Thorstad | The scope of longtermism Read More »
This talk aims to open a dialogue between philosophers working in decision theory and operations researchers and engineers whose research addresses the topic of decision making under deep uncertainty. Specifically, we assess the recommendation to follow a norm of robust satisficing when making decisions under deep uncertainty in the context of decision analyses that rely on the tools of Robust Decision Making developed by Robert Lempert and colleagues at RAND. We discuss decision-theoretic and voting-theoretic motivations for robust satisficing, then use these motivations to select among candidate formulations of the robust satisficing norm. We also discuss two challenges for robust satisficing: whether the norm might in fact derive its plausibility from an implicit appeal to probabilistic representations of uncertainty of the kind that deep uncertainty is supposed to preclude; and whether there is adequate justification for adopting a satisficing norm, as opposed to an optimizing norm that is sensitive to considerations of robustness.
We show that every theory of the value of uncertain prospects must have one of three unpalatable properties. Reckless theories recommend risking arbitrarily great gains at arbitrarily long odds for the sake of enormous potential; timid theories permit passing up arbitrarily great gains to prevent a tiny increase in risk; non-transitive theories deny the principle that, if A is better than B and B is better than C, then A must be better than C. While non-transitivity has been much discussed, we draw out the costs and benefits of recklessness and timidity when it comes to axiology, decision theory, and moral uncertainty.
Teru Thomas | A Paradox for Tiny Probabilities and Enormous Values Read More »
Christian Tarsney | Non-additive axiologies in large worlds Global Priorities Seminar Read the paper Presented as part of the Global Priorities Seminar series 26 June 2020 View full transcript Expand CHRISTIAN TARSNEY: (00:09) All right. So as Andreas said, the paper is called Non-Additive Axiologies in Large Worlds, and it’s joint work with Teru Thomas,