I am an Assistant Professor in the Department of Public Administration and International Affairs in the Maxwell School at Syracuse University.
My academic training is in philosophy. Before my current position I was a postdoctoral fellow in the McCoy Family Center for Ethics in Society at Stanford University. Part of my time I spent at Apple working on the ethics of machine learning and autonomous systems.
I work in moral and political philosophy as well as in applied ethics.
My research investigates the nature of human and collective agency, the value of agency, and its relation to moral responsibility. I am equally interested in the ethics of emerging technologies and quantitaive methods, such as self-driving cars, autonomous weapons systems, artificial intelligence and data science.
Moreover, I write on applied political philosophy and investigate ethical dilemmas of public policy.
I recently spoke with the Communications of the ACM about my research on self-driving cars
Here is an overview of papers and book reviews that I have written.
The ongoing debate on the ethics of self-driving cars typically focuses on two approaches to answering such questions: moral philosophy and social science. I argue that these two approaches are both lacking. We should neither deduce answers from individual moral theories nor should we expect social science to give us complete answers. To supplement these approaches, we should turn to political philosophy. The issues we face are collective decisions that we make together rather than individual decisions we make in light of what we each have reason to value. Political philosophy adds three basic concerns to our conceptual toolkit: reasonable pluralism, human agency, and legitimacy. These three concerns have so far been largely overlooked in the debate on the ethics of self-driving cars.
A central dispute in social ontology concerns the existence of group minds and actions. I argue that some authors in this dispute rely on rival views of existence without sufficiently acknowledging this divergence. I proceed in three steps in arguing for this claim. First, I define the phenomenon as an implicit higher-order disagreement by drawing on an analysis of verbal disputes. Second, I distinguish two theories of existence – the theory-commitments view and the truthmaker view – in both their eliminativist and their constructivist variants. Third, I examine individual contributions to the dispute about the existence of group minds and actions to argue that these contributions have an implicit higher-order disagreement. This paper serves two purposes. First, it is a study to apply recent advances in meta-ontology. Second, it contributes to the debate on social ontology by illustrating how meta-ontology matters for social ontology.
Future weapons will make life-or-death decisions without a human in the loop. When such weapons inflict unwarranted harm, no one appears to be responsible. There seems to be a responsibility gap. I first reconstruct the argument for such responsibility gaps to then argue that this argument is not sound. The argument assumes that commanders have no control over whether autonomous weapons inflict harm. I argue against this assumption. Although this investigation concerns a specific case of autonomous weapons systems, I take steps towards vindicating the more general idea that superiors can be morally responsible in virtue of being in command.
The asylum system faces problems on two fronts. States undermine it with populist politics, and migrants use it to satisfy their migration preferences. To address these problems, asylum services should be commodified. States should be able to pay other states to provide determination and protection-elsewhere. In this article, I aim to identify a way of implementing this idea that is both feasible and desirable. First, I sketch a policy proposal for a commodification of asylum services. Then, I argue that this policy proposal is not only compatible with the right to asylum, but also supported by moral considerations. Despite some undesirable moral features, a market in asylum facilitates the provision of asylum to those who need it.
Related publications: This proposal also made it to this book Wenn ich mir etwas wünschen dürfte (Steidl 2017) on the occasion of German general elections, and to a discussion in the Change My View Subreddit here.
Punishing groups raises a difficult question, namely, how their punishment can be justified at all. Some have argued that punishing groups is morally problematic because of the effects that the punishment entails for their members. In this paper we argue against this view. We distinguish the question of internal justice – how punishment-effects are distributed – from the question of external justice – whether the punishment is justified. We argue that issues of internal justice do not in general undermine the permissibility of punishment. We also defend the permissibility of what some call “random punishment.” We argue that, for some kinds of collectives, there is no general obligation to internally distribute the punishment-effects equally or in proportion to individual contribution.
Trolley cases are widely considered central to the ethics of autonomous vehicles. I caution against this by identifying four problems. (1) Trolley cases, given technical limitations, rest on assumptions that are in tension with one another. Furthermore, (2) trolley cases illuminate only a limited range of ethical issues insofar as they cohere with a certain design framework. Furthermore, (3) trolley cases seem to demand a moral answer when a political answer is called for. Finally, (4) trolley cases might be epistemically problematic in several ways. To put forward a positive proposal, I illustrate how ethical challenges arise from mundane driving situations. I argue that mundane situations are relevant because of the specificity they require and the scale they exhibit. I then illustrate some of the ethical challenges arising from optimizing for safety, balancing safety with other values such as mobility, and adjusting to incentives of legal frameworks.
Related publications: “The everyday ethical challenges of self-driving cars,” The Conversation, syndicated in The Boston Globe, and others.
This paper develops a taxonomy of kinds of actions that can be seen in group agency, human–machine interactions, and virtual realities. These kinds of actions are special in that they are not embodied in the ordinary sense. I begin by analysing the notion of embodiment into three separate assumptions that together comprise what I call the Embodiment View. Although this view may find support in paradigmatic cases of agency, I suggest that each of its assumptions can be relaxed. With each assumption that is given up, a different kind of disembodied action becomes available. The taxonomy gives a systematic overview and suggests that disembodied actions have the same theoretical relevance as the actions of any ordinarily embodied human.
This paper is about the status of collective actions. According to one view, collective actions metaphysically reduce to individual actions because sentences about collective actions are merely a shorthand for sentences about individual actions. I reconstruct an argument for this view and show via counterexamples that it is not sound. The argument relies on a paraphrase procedure to unpack alleged shorthand sentences about collective actions into sentences about individual actions. I argue that the best paraphrase procedure that has been put forward so far fails to produce adequate results.
Related publications: The paper prompted a discussion note, which you can find here.
This paper examines two questions about scientists' search for knowledge. First, which search strategies generate discoveries effectively? Second, is it advantageous to diversify search strategies? We argue pace Weisberg and Muldoon (2009) that, on the first question, a search strategy that deliberately seeks novel research approaches need not be optimal. On the second question, we argue they have not shown epistemic reasons exist for the division of cognitive labor, identifying the errors that led to their conclusions. Furthermore, we generalize the epistemic landscape model, showing that one should be skeptical about the benefits of social learning in epistemically complex environments.
Additional material: The model used for this article is written using NetLogo. The source code of our model is available here. It involves a swarm strategy, which draws on the model by Couzin et al. (2005) and the Boids model. You can find a simple simulation that I wrote to study the behaviour of this model here.
We are responsible for some things but not for others. In this thesis, I investigate what it takes for an entity to be responsible for something. This question has two components: agents and actions. I argue for a permissive view about agents. Entities such as groups or artificially intelligent systems may be agents in the sense required for responsibility. With respect to actions, I argue for a causal view. The relation in virtue of which agents are responsible for actions is a causal one. I claim that responsibility requires causation and I develop a causal account of agency. This account is particularly apt for addressing the relationship between agency and moral responsibility and sheds light on the causal foundations of moral responsibility.
Courses at LSE were taught as a teaching assistant; all other courses were taught as primary instructor.