Research
My research develops a social and structural approach to normative and philosophical questions. I study how values, knowledge, and meaning emerge—and sometimes fail to align—within complex systems of interaction among individuals, institutions, and technologies. A central insight guiding this work is that ethical failures such as mistrust, exclusion, disinformation, and inequitable outcomes rarely arise from isolated decisions or technical flaws. Rather, they are emergent features of social structure: how research is organised, how evidence is produced and evaluated, and how authority and expertise are distributed.
Drawing on the philosophy of science, social epistemology, and formal modelling (including evolutionary and game-theoretic methods), I connect ethical analysis directly to questions of governance, trust, and implementation. My research spans three areas: (1) the philosophy and ethics of artificial intelligence and data-intensive technologies, (2) the philosophy of autism and neurodiversity, and (3) social dynamics, norms, and conventions. Together, these projects form a unified framework for understanding ethical challenges in high-stakes domains.
1. Philosophy and Ethics of AI and Emerging Technologies.
One of my key research programmes examines the philosophy and ethics of artificial intelligence, with particular attention to the value alignment problem. I argue that alignment cannot be understood solely as a technical or abstract normative challenge; it is fundamentally social and institutional. This perspective is especially relevant for high-stakes applications of artificial intelligence, where misaligned incentives, proxy measures, and data practices can undermine equity and public trust.
Artificial Intelligence and the Value Alignment Problem
Traditional accounts frame the value alignment problem as partly technical (how to encode values in AI systems) and partly normative (which values should be encoded). I argue that this framing obscures the deeper source of alignment failures. In my book, Artificial Intelligence and the Value Alignment Problem: A Philosophical Introduction, I reconceptualise misalignment as an emergent property of multi-agent systems involving researchers, institutions, data infrastructures, and affected communities.
On this view, problems such as bias, fairness, transparency, and accountability arise from the organisation of research and deployment itself, rather than from isolated algorithmic failures. This analysis has direct implications for data governance, responsible innovation, and ethical oversight.
Language, Communication, and Measurement
Related projects extend this analysis to communication and measurement. I argue that robust value alignment requires socially grounded communicative capacities that current AI systems lack. Using signalling games as a formal framework, I study coordination in both human and artificial systems. This work connects the evolution of communication to contemporary challenges in AI ethics.
A further strand of research critiques benchmarking and standardised metrics in AI and healthcare, showing how they embed normative assumptions while obscuring lived experience. Together, these projects support a contextual, equity-oriented approach to evaluating emerging technologies.
Normativity and Artificial Intelligence
Along with a multi-disciplinary research team I have been reviewing the literature on artificial moral agency. We seek to provide a systematic account of the most promising approaches to developing artificial moral agents, addressing normative pluralism, conflicts, and codification. A central research question outlines the key issues and most promising implementation approaches for high- and low-level agents. Additionally, we explore the main benchmarking approaches for assessing the level of functionality, accuracy, or moral competence of artificial moral agents. My recent single- and co-authored research in this area has sought to critically examine the very idea of benchmarking ethics for AI systems. I explore why current approaches in computer science fail in light of category mistakes involved in using moral dilemmas from philosophy as benchmarks for whether an AI system acts morally, in addition to discussing metaethical problems for benchmarking ethics for AI.
2. The Philosophy of Autism and Autistic Philosophy.
Since Fall 2022, beginning with a graduate seminar that I designed to teach at Dalhousie University (Philosophy on the Spectrum: The Philosophy of Autism and Autistic Philosophy), I have been exploring philosophical research on autism spectrum disorders. Stuart Murray suggests that the very idea of an autistic person is a philosophical one. However, Kenneth Richman points out that ‘the philosophy of autism’ is not (yet) a subfield of philosophy insofar as philosophical work on autism has fallen primarily under ethics, philosophy of mind, philosophy of psychology, or philosophy of medicine. Although the philosophy of autism is a fruitful research area within and without these narrower domains, my main research interest concerns (what I refer to as) autistic philosophy.
A central strand of my research examines autism and neurodiversity through the lens of philosophy of science, ethics, and social epistemology. This work investigates how diagnostic categories, research paradigms, and clinical practices structure whose knowledge counts, which forms of difference are pathologised, and how trust is distributed between researchers, clinicians, and neurodivergent communities.
Cognitive Explanations of Autism
A major focus of this programme is a sustained critique of cognitive explanations of autism, particularly theory-of-mind–deficit accounts. In forthcoming work, I argue that despite persistent empirical and methodological challenges, these frameworks continue to structure autism research and clinical interpretation. Their persistence is best explained not by explanatory success, but by institutional, historical, and epistemic dynamics—such as incentive structures, disciplinary silos, and asymmetries of authority—that insulate dominant paradigms from meaningful revision.
By shifting attention from the correctness of individual theories to the social dynamics that sustain them, this work connects philosophical critique to concrete concerns about stigma, clinical translation, and public trust in health research.
Autistic Philosophy and Neurodiversity
Beyond the philosophy of autism, my work contributes to what I refer to as autistic philosophy. This approach begins from the recognition that much philosophical theorising implicitly assumes neurotypical perspectives, often misunderstanding autism and reinforcing harmful stereotypes. Autistic philosophy treats autistic perspectives as epistemic resources that challenge dominant assumptions in epistemology, ontology, and philosophy of science—particularly practices of categorisation, diagnosis, and pathologisation.
2. Social Dynamics, Norms, and Conventions.
My third research programme develops the general theoretical foundations that unify my work on autism and AI. Using evolutionary, agent-based, and game-theoretic models, I study how social norms, conventions, and epistemic practices emerge, stabilise, and sometimes sustain exclusion, injustice, and misinformation.
Language, Reflexivity, and Cooperation
I investigate the evolutionary origins of language, focusing on reflexivity—the ability to use language to talk about language—as a key innovation enabling complex communication and cooperation. I show how reflexivity generates feedback cycles that allow communication systems to modify themselves over time, linking language evolution to contemporary problems of coordination and alignment in AI systems.
Social Epistemology, Power, and Justice
Drawing on work supported by the National Science Foundation, I use formal models to analyse how social structure shapes the distribution of knowledge, belief, and resources. This research shows how small power asymmetries can produce large-scale inequities, and how false beliefs can persist in networks even after explicit correction. These findings have implications for theories of distributive justice, public trust, and the governance of science and technology.
Across these interconnected programmes, my research advances a structural and interactional philosophy of social phenomena. Whether examining autism research, AI systems, or the evolution of communication, I aim to illuminate how values, norms, and knowledge co-evolve within complex social systems—and how philosophical analysis can inform more just, inclusive, and trustworthy forms of research and innovation.
Books
Artificial Intelligence and the Value Alignment Problem: A Philosophical Introduction
Available for purchase.
From the publisher: "Written for an interdisciplinary audience, this book provides strikingly clear explanations of the many difficult technical and moral concepts central to discussions of ethics and AI. In particular, it serves as an introduction to the value alignment problem: that of ensuring that AI systems are aligned with the values of humanity. LaCroix redefines the problem as a structural one, showing students how many topics in AI ethics, from bias and fairness to transparency and opacity, can be understood as instances of the key problem of value alignment. Numerous case studies are presented throughout the book to highlight the significance of the issues at stake and to clarify the central role of the value alignment problem in the many ethical challenges facing the development and implementation of AI."
Advanced Praise:
"Travis LaCroix’s book on value alignment is, without a doubt, the best I have read on AI ethics. I highly recommend it to anyone interested in the ethics of artificial intelligence. The text is intellectually rigorous, and many of its ideas are genuinely novel. I found his discussion of measuring value alignment particularly insightful, along with the appendix on superintelligence and the control problem, which provides valuable depth to the topic.” -- Martin Peterson, Texas A&M University"
"LaCroix's Artificial Intelligence and the Value Alignment Problem offers an insightful overview and evaluation of the predicament we find ourselves in with respect to machine learning. The book doesn't shy away from engaging with the mathematical background of these challenges, but it does so in a way that's intelligible to readers with limited mathematical experience. The structural characterization of the alignment problem(s) provides a great conceptual tool for exploring the ways that values are (or fail to be) incorporated in machine learning systems. The discussions of values are also inclusive, incorporating views from Western, Eastern, and Indigenous philosophy. This book offers an up-to-date introduction to the topic at a level suitable for undergraduates while also providing a novel analytic tool for anyone already working in the area of AI ethics." -- Gillman Payette, University of Calgary
Recommended citation:
LaCroix, Travis. 2025. Artificial Intelligence and the Value Alignment Problem: A Philosophical Introduction. Broadview Press.
Peer-Reviewed Articles
Autism, Theory of Mind, and the Dynamics of Value-Laden Research Programmes
Psychological Inquiry (Accepted)
This article responds to commentaries on my analysis of the theory-of-mind-deficit explanation of autism, using Lakatos's methodology of scientific research programmes (LaCroix, 2026). The commentaries largely agree that the theory-of-mind paradigm faces serious problems, but shift focus from whether the framework is scientifically defensible to why it persists despite mounting counterevidence. Engaging with these responses, I clarify that the theory endures less because of its evidential strength and more because of historical, institutional, and epistemic dynamics that insulate research programmes from falsification. Building on the commentators' themes, I extend the original argument and defend the relevance of the pseudoscience thesis.
[Official version forthcoming.] (Open Access)
Recommended citation:
LaCroix, Travis. Accepted. "Autism, Theory of Mind, and the Dynamics of Value-Laden Research Programmes" Psychological Inquiry. 1-32.
Autism and the Pseudoscience of Mind
Psychological Inquiry (Accepted)
The theory-of-mind-deficit explanation of autism proposes that autistics lack a theory of mind, that autism comprises a theory-of-mind deficit (strong version); or, that autistics often have difficulty with theory-of-mind abilities (weak version). A growing body of critical research demonstrates how these explanations of autistic behaviour fail—both empirically and theoretically. The strong version lacks explanatory adequacy, while the weak version is undermined by methodological and empirical flaws in theory-of-mind research. Together, these issues suggest that the "science" of theory of mind in the context of autism is, at best, bad science. Nonetheless, researchers continue to pursue this line of inquiry in autism studies—often moving the goalposts or offering ad hoc rationalisations to preserve the theoretical framework. This article critically examines the theory-of-mind-deficit explanation of autism, focusing particularly on the widely-held view that autistics exhibit difficulties with theory of mind—i.e., the weak version of the theory-of-mind-deficit explanation. Drawing from the philosophy of science, I argue that ongoing adherence to this view exhibits all the hallmarks of a degenerating research programme. Hence, the fact that scientists have not abandoned this hypothesis entails that the research programme is pseudoscientific.
[Official version forthcoming.] (Open Access)
[PhilSci-Archive preprint available here.] (Please cite published version, if available.)
Recommended citation:
LaCroix, Travis. Accepted. "Autism and the Pseudoscience of Mind" Psychological Inquiry. 1-66.
What Do Philosophers Talk About When They Talk About Autism? (with Alexis Amero and Benjamin Sidloski)
Synthese (Accepted)
Several anecdotal claims about the relationship between philosophical discourse and the subject of autism have been forwarded in recent years. This paper seeks to verify or debunk these descriptive claims by carefully examining the philosophical literature on autism. We conduct a comprehensive scoping review to answer the question, what do philosophers talk about when they talk about autism? This empirical work confirms that the philosophy of autism is underdeveloped as a subfield of philosophy. Moreover, the way that philosophers engage with autism is often unreflective and uncritical. As a result, much work in the discipline serves to perpetuate pathologising, dehumanising, and stigmatising misinformation about autistics and autistic behaviour. By highlighting the significant gaps in the philosophical literature on autism, this review aims to deepen our understanding of philosophical thought surrounding autism and contributes to ongoing dialogues pertaining to neurodiversity, madness, and disability rights more generally.
[Official version available here] (Open Access)
[PhilSci-Archive preprint available here.] (Please cite published version, if available.)
Recommended citation:
LaCroix, Travis, Alexis Amero, and Benjamin Sidloski. 2025. "What Do Philosophers Talk About When They Talk About Autism?" Synthese. 206(63): 1-47.
Information and Meaning in the Evolution of Compositional Signals
Journal of Logic, Language and Information (2025)
This paper provides a formal treatment of the argument that syntax alone cannot give rise to compositionality in a signalling game context. This conclusion follows from the standard information-theoretic machinery used in the signalling game literature to describe the informational content of signals.
[Official version available here.] (Open Access)
[PhilSci-Archive preprint available here.] (Please cite published version, if available.)
Recommended citation:
LaCroix, Travis. 2025. "Information and Meaning in the Evolution of Compositional Signals." Journal of Logic, Language and Information. 34: 219-239.
Metaethical Perspectives on 'Benchmarking' AI Ethics (with Alexandra Sasha Luccioni)
AI and Ethics (2025)
Benchmarks are seen as the cornerstone for measuring technical progress in artificial intelligence (AI) research and have been developed for a variety of tasks ranging from question answering to emotion recognition. An increasingly prominent research area in AI is ethics, which currently has no set of benchmarks nor commonly accepted way for measuring the 'ethicality' of an AI system. In this paper, drawing upon research in moral philosophy and metaethics, we argue that it is impossible to develop such a benchmark. As such, alternative mechanisms are necessary for evaluating whether an AI system is 'ethical'. This is especially pressing in light of the prevalence of applied, industrial AI research. We argue that it makes more sense to talk about 'values' (and 'value alignment') rather than 'ethics' when considering the possible actions of present and future AI systems. We further highlight that, because values are unambiguously relative, focusing on values forces us to consider explicitly what the values are and whose values they are. Shifting the emphasis from ethics to values therefore gives rise to several new ways of understanding how researchers might advance research programmes for robustly safe or beneficial AI.
[Official version available Here.] (Open Access)
Recommended citation:
LaCroix, Travis, and Alexandra Sasha Luccioni. 2025 "Metaethical Perspectives on 'Benchmarking' AI Ethics." AI and Ethics. 5: 4029-4047.
The Linguistic Dead Zone of Value-aligned Agency, Natural and Artificial
Philosophical Studies (2024)
The value alignment problem for artificial intelligence (AI) asks how we can ensure that the “values”—i.e., objective functions—of artificial systems are aligned with the values of humanity. In this paper, I argue that linguistic communication is a necessary condition for robust value alignment. I discuss the consequences that the truth of this claim would have for research programmes that attempt to ensure value alignment for AI systems—or, more loftily, those programmes that seek to design robustly beneficial or ethical artificial agents.
[Official version available Here.] (Open Access)
Recommended citation:
LaCroix, Travis. 2024. "The Linguistic Dead Zone of Value-Aligned Agency, Natural and Artificial." Philosophical Studies. 1-23.
Power by Association (with Cailin O'Connor)
Ergo, an Open Access Journal of Philosophy (2022)
We use tools from evolutionary game theory to examine how power might influence the cultural evolution of inequitable norms between discernible groups (such as gender or racial groups) in a population of otherwise identical individuals. Similar extant models always assume that power is homogeneous across a social group. As such, these models fail to capture situations where individuals who are not themselves disempowered nonetheless end up disadvantaged in bargaining scenarios by dint of their social group membership. Thus, we assume that there is heterogeneity in the groups in that some individuals are more powerful than others.
Our model shows that even when most individuals in two discernible sub-groups are relevantly identical, powerful individuals can affect the social outcomes for their entire group; this results in power by association for their in-group and a bargaining disadvantage for their out-group. In addition, we observe scenarios like those described where individuals who are more powerful will get less in a bargaining scenario because a convention has emerged disadvantaging their social group.
[Official version available here.] (Open Access)
[PhilSci-Archive preprint available here.] (Please cite published version, if available.)
Recommended citation:
LaCroix, Travis and Cailin O'Connor. 2022. "Power by Association." Ergo. 8(29): 163-189.
Est-ce que vous Compute? Code-Switching, Culutral Identity, and AI (with Arianna Falbo)
Feminist Philosophical Quarterly (2022)
Cultural code-switching concerns how we adjust our overall behaviours, manners of speaking, and appearance in response to a perceived change in our social environment. We defend the need to investigate cultural code-switching capacities in artificial intelligence systems. We explore a series of ethical and epistemic issues that arise when bringing cultural code-switching to bear on artificial intelligence. Building upon Dotson's (2014) analysis of testimonial smothering, we discuss how emerging technologies in AI can give rise to epistemic oppression, and specifically, a form of self-silencing that we call 'cultural smothering'. By leaving the socio-dynamic features of cultural code-switching unaddressed, AI systems risk negatively impacting already-marginalised social groups by widening opportunity gaps and further entrenching social inequalities.
[Official version available here.] (Open Access)
[arXiv preprint available here.] (Please cite published version, if available.)
Recommended citation:
Falbo, Arianna and Travis LaCroix. 2022. "Est-ce que vous compute? Code-switching, cultural identity and AI." Feminist Philosophical Quarterly. 8(3-4): 9:1-9:24.
The Tragedy of the AI Commons (with Aydin Mohseni)
Synthese (2022)
Policy and guideline proposals for ethical artificial-intelligence research have proliferated in recent years. These are supposed to guide the socially-responsible development of AI for the common good. However, there typically exist incentives for non-cooperation (i.e., non-adherence to such policies and guidelines); and, these proposals often lack effective mechanisms to enforce their own normative claims. The situation just described constitutes a social dilemma—namely, a situation where no one has an individual incentive to cooperate, though mutual cooperation would lead to the best outcome for all involved. In this paper, we use stochastic evolutionary game dynamics to model this social dilemma in the context of the ethical development of artificial intelligence. This formalism allows us to isolate variables that may be intervened upon, thus providing actionable suggestions for increased cooperation amongst numerous stakeholders in AI. Our results show how stochastic effects can help make cooperation viable in such a scenario. They suggest that coordination for a common good should be attempted in smaller groups in which the cost for cooperation is low, and the perceived risk of failure is high. This provides insight into the conditions under which we should expect such ethics proposals to be successful with regard to their scope, scale, and content.
[Official version available here.]
[arXiv preprint available here.] (Please cite official version, if available.)
[Recorded talk available here.]
[Poster available here.]
Recommended citation:
LaCroix, Travis and Aydin Mohseni. 2022. "The Tragedy of the AI Commons." Synthese 200: 289:1-289:33. https://doi.org/10.1007/s11229-022-03763-2
Moral Dilemmas for Moral Machines
AI and Ethics (2022)
Autonomous systems are being developed and deployed in situations that may require some degree of ethical decision-making ability. As a result, research in machine ethics has proliferated in recent years. This work has included using moral dilemmas as validation mechanisms for implementing decision-making algorithms in ethically-loaded situations. Using trolley-style problems in the context of autonomous vehicles as a case study, I argue (1) that this is a misapplication of philosophical thought experiments because (2) it fails to appreciate the purpose of moral dilemmas, and (3) this has potentially catastrophic consequences; however, (4) there are uses of moral dilemmas in machine ethics that are appropriate and the novel situations that arise in a machine-learning context can shed some light on philosophical work in ethics.
[Official version available here.]
[PhilSci-Archive preprint available here.] (Please cite published version, if available.)
[arXiv preprint available here.] (Please cite published version, if available.)
Recommended citation:
LaCroix, Travis. 2022. "Moral dilemmas for moral machines." AI and Ethics. 2: 737-746. [https://doi.org/10.1007/s43681-022-00134-y].
Using Logic to Evolve More Logic: Composing Logical Operators via Self-Assembly
British Journal for the Philosophy of Science (2022)
I consider how complex logical operations might self-assemble in a signalling-game context via composition of simpler underlying dispositions. On the one hand, agents may take advantage of pre-evolved dispositions; on the other hand, they may co-evolve dispositions as they simultaneously learn to combine them to display more complex behaviour. In either case, the evolution of complex logical operations can be more efficient that evolving such capacities from scratch. Showing how complex phenomena like these might evolve provides an additional path to the possibility of evolving more or less rich notions of compositionality. This helps provide another facet of the evolutionary story of how sufficiently rich, human-level cognitive or linguistic capacities may arise from simpler precursors.
[Official version available here.]
[PhilSci-Archive preprint available here.] (Please cite official version.)
Recommended citation:
LaCroix, Travis. 2019. "Using Logic to Evolve More Logic: Composing Logical Operators via Self-Assembly." British Journal for the Philosophy of Science (2022) 73(2): 407-437.
Epistemology and the Structure of Language (with Jeffrey A. Barrett)
Erkenntnis (2022)
We are concerned here with how structural properties of language may evolve to reflect features of the world in which it evolves. As a concrete example, we will consider how a simple term language might evolve to support the principle of indifference over state descriptions in that language. The point is not that one is justified in applying the principle of indifference to state descriptions in natural language. Rather, it is that one should expect a language that has evolved in the context of facilitating successful action to reflect probabilistic features of the world in which it evolved.
[Official version available here.]
[PhilSci-Archive preprint available here.] (Please cite official version.)
Recommended citation:
Barrett, Jeffrey A. and Travis LaCroix. 2020. "Epistemology and the Structure of Language." Erkenntnis (2022) 87: 953-967. https://doi.org/10.1007/s10670-020-00225-4
Reflexivity, Functional Reference, and Modularity: Alternative Targets for Language Origins
Philosophy of Science (2021)
Researchers in language origins typically try to explain how compositional communication might evolve to bridge the gap between animal communication and natural language. However, as an explanatory target, compositionality has been shown to be problematic for a gradualist approach to the evolution of language. In this paper, I suggest that reflexivity provides an apt and plausible alternative target which does not succumb to the problems that compositionality faces. I further explain how proto-reflexivity, which depends upon functional reference, gives rise to complex communication systems via modular composition.
[Official version available here.]
[PhilSci-Archive preprint available here.] (Please cite official version, if available.)
Recommended citation:
LaCroix, Travis. 2020. "Reflexivity, Functional Reference, and Modularity: Alternative Targets for Language Origins." Philosophy of Science (2021) 8(5): 1234-1245.
The Dynamics of Retraction in Epistemic Networks (with Cailin O'Connor and Anders Geil)
Philosophy of Science (2021)
Sometimes retracted scientific information is used and propagated long after it is understood to be misleading. Likewise, sometimes retracted news items spread and persist, even after it has been publicly established that they are false. In this paper, we use agent-based models of epistemic networks to explore the dynamics of retraction. In particular we focus on why false beliefs might persist, even in the face of retraction.
[Official version available here.]
[PhilSci-Archive preprint available here.] (Please cite published version, if available.)
Recommended citation:
LaCroix, Travis, Anders Geil, and Cailin O'Connor. 2020. "The Dynamics of Retraction in Epistemic Networks." Philosophy of Science (2021) 88(3): 415-438. https://doi.org/10.1086/712817
Communicative Bottlenecks Lead to Maximal Information Transfer
Journal of Experimental and Theoretical Artificial Intelligence (2020)
This paper presents new analytic and numerical analysis of signalling games that give rise to informational bottlenecks—that is to say, signalling games with more state/act pairs than available signals to communicate information about the world. I show via simulation that agents learning to coordinate tend to favour partitions of nature which provide maximal information transfer. This is true in spite of the fact that nothing from an initial analysis of the stability properties of the underlying signalling game suggests that this should be the case. As a first pass to explain this, I note that the underlying structure of our model favours maximal information transfer in regard to the simple combinatorial properties of the ways in which the agents might partition nature into kinds. However, I suggest that this does not perfectly capture the empirical results; thus, several open questions remain.
[Official version available here.]
[PhilSci-Archive preprint available here.] (Please cite official version.)
Recommended citation:
LaCroix, Travis. 2020. "Communicative Bottlenecks Lead to Maximal Information Transfer." Journal of Experimental and Theoretical Artificial Intelligence (2020) 32(6): 997-1014.
Evolutionary Explanations of Simple Communication: Signalling Games & Their Models
Journal for General Philosophy of Science / Zeitschrift für allgemeine Wissenschaftstheorie (2020)
This paper applies the theoretical criteria laid out by D'arms et al. (1998) to various aspects of evolutionary models of signalling. The question that D'Arms et al. seek to answer can be formulated as follows: Are the models that we use to explain the phenomena in question conceptually adequate? The conceptual adequacy question relates the formal aspects of the model to those aspects of the natural world that the model is supposed to capture. Moreover, this paper extends the analysis of D'Arms et al. by asking the following additional question: Are the models that we use sufficient to explain the phenomena in question? The sufficiency question ask what formal resources are minimally required in order for the model to get the right results most of the time.
[Official version available here.]
[PhilSci-Archive preprint available here.] (Please cite official version.)
Recommended citation:
LaCroix, Travis. 2019. "Evolutionary Explanations of Simple Communication: Signalling Games & Their Models." Journal for General Philosophy of Science / Zeitschrift für allgemeine Wissenschaftstheorie (2020) 51(1): 19-43.
On Salience and Signalling in Sender-Receiver Games: Partial-Pooling, Learning, and Focal Points
Synthese (2020)
I introduce an extension of the Lewis-Skyrms signaling game, analysed from a dynamical perspective via simple reinforcement learning. In Lewis' (Convention, Blackwell, Oxford, 1969) conception of a signaling game, salience is offered as an explanation for how individuals may come to agree upon a linguistic convention. Skyrms (Signals: evolution, learning & information, Oxford University Press, Oxford, 2010a) offers a dynamic explanation of how signaling conventions might arise presupposing no salience whatsoever. The extension of the atomic signaling game examined here—which I refer to as a salience game—introduces a variable parameter into the atomic signaling game which allows for degrees of salience, thus filling in the continuum between Skyrms' and Lewis' models. The model does not presuppose any salience at the outset, but illustrates a process by which accidentally evolved salience is amplified, to the benefit of the players. It is shown that increasing degrees of salience allow populations to avoid sub-optimal pooling equilibria and to coordinate upon conventions more quickly.
[Official version available here.]
[PhilSci-Archive preprint available here.] (Please cite official version.)
Recommended citation:
LaCroix, Travis. 2018. "On Salience and Signaling in Sender-Receiver Games: Partial Pooling, Learning, and Focal Points." Synthese (2020) 197(4): 1725-1747.
Book Chapters
Ethics and Deep Learning (with Simon J. D. Prince)
Chapter 21 in Simon J. D. Prince (2023) Understanding Deep Learning
An overview of some ethical issues arising from deep learning approaches to artificial intelligence research.
[Complete book draft available here.]
Recommended citation:
LaCroix, Travis, and Simon J. D. Prince. 2023. "Ethics and Deep Learning". Chapter 21 in Simon J. D. Prince, Understanding Deep Learning. Cambridge, MA: The MIT Press. 420-435.
Conference Proceedings
Emergent Communication under Competition (with Michael Noukhovitch, Angeliki Lazaridou, and Aaron Courville)
Autonomous Agents and Multiagent Systems (AAMAS 2021)
Current literature in machine learning has only negative results for learning to communicate between competitive agents using vanilla RL. We introduce a modified sender-receiver game to study the spectrum of partially-competitive scenarios and show communication can indeed emerge in this setting. We empirically demonstrate three key takeaways for future research. First, we show that communication is proportional to cooperation, and it can occur for partially competitive scenarios using standard learning algorithms. Second, we highlight the difference between communication and manipulation and extend previous metrics of communication to the competitive case. Third, we investigate the negotiation game where previous work failed to learn communication between independent agents. We show that, in this setting, both agents must benefit from communication for it to emerge. Finally, with a slight modification to the game, we successfully learn to communicate between competitive agents. We hope this work overturns misconceptions and inspires more research in competitive emergent communication.
[Official version available here.]
[arXiv preprint available here.] (Please cite published version, if available.)
[Recorded Talk (Noukhovitch) Available Here.]
Recommended citation:
Noukhovitch, Michael, Travis LaCroix, Angeliki Lazaridou, and Aaron Courville. 2021. Emergent Communication under Competition. In U. Endriss, A. Nowé, F. Dignum, and A. Lomuscio (eds.), Proc. of the 20th International Conference on Autonomous Agents and Multiagent Systems (AAMAS 2021), London, UK, 3-7 May 2021, International Foundation for Autonomous Agents and Multiagent Systems (IFAAMAS) 974-982.
Biology and Compositionality: Empirical Considerations for Emergent-Communication Protocols
NeurIPS 2019 workshop Emergent Communication: Towards Natural Language (2019)
Significant advances have been made in artificial systems by using biological systems as a guide. However, there is often little interaction between computational models for emergent communication and biological models of the emergence of language. Many researchers in language origins and emergent communication take compositionality as their primary target for explaining how simple communication systems can become more like natural language. However, there is reason to think that compositionality is the wrong target on the biological side, and so too the wrong target on the machine-learning side. As such, the purpose of this paper is to explore this claim. This has theoretical implications for language origins research more generally, but the focus here will be the implications for research on emergent communication in computer science and machine learning—specifically regarding the types of programmes that might be expected to work and those which will not. I further suggest an alternative approach for future research which focuses on reflexivity, rather than compositionality, as a target for explaining how simple communication systems may become more like natural language. I end by providing some reference to the language origins literature that may be of some use to researchers in machine learning.
[ArXiv preprint available here.]
[Poster available here.]
Recommended citation:
LaCroix, Travis. 2019. "Biology and Compositionality: Empirical Considerations for Emergent-Communication Protocols." arXiv preprint: 1911.11668. https://arxiv.org/abs/1911.11668.
Book Reviews
Review of Ronald J. Planer and Kim Sterelny’s From Signal to Symbol
Philosophy of Science (2023)
Review of Ronald J. Planer and Kim Sterelny’s From Signal to Symbol – Ronald J. Planer and Kim Sterelny, From Signal to Symbol: The Evolution of Language. Cambridge: The MIT Press (2021), 296 pp., $35.00 (hardcover)
[Official version available here.]
Recommended citation:
LaCroix, Travis. 2023. "Review of Ronald J. Planer and Kim Sterelny's From Signal to Symbol." Philosophy of Sceince (Forthcoming): 1-4. https://doi.org/10.1017/psa.2023.75 .
Dissertation
Complex Signals: Reflexivity, Hierarchical Structure, and Modular Composition
My dissertation argues that what drives the emergence of complex communication systems is a process of modular composition, whereby independent communicative dispositions combine to create more complex dispositions. This challenges the dominant view on the evolution of language, which attempts to resolve the explanatory gap between communication and language by demonstrating how complex syntax evolved. My research shows that these accounts fail to maintain sensitivity to empirical data: genuinely compositional syntax is extremely rare or non-existent in nature. In contrast, my research prioritises the reflexivity of natural language—the ability to use language to talk about language—as an alternative explanatory target.
The first part of my dissertation provides the philosophical foundation of this novel account using the theoretical framework of Lewis-Skyrms signalling games and drawing upon relevant work in evolutionary biology, linguistics, cognitive systems, and machine learning. Chapter 1 introduces the signalling game and contextualises it with respect to problems in the realm of traditional philosophy of language. Chapter 2 examines empirical data from biology and linguistics and argues that complex syntax is not the most apt explanatory target for how language might have evolved out of simple communication. Chapter 3 then argues that the reflexivity of language is a more fruitful property to consider, showing how reflexivity aids the evolution of complex communication via a process of modular composition. This connects parallel research in the evolution of language, cognitive systems, and machine learning paradigms. Once such complexity is exhibited, at a small scale, it may lead to a 'feedback loop' between communication and cognition that gives rise to the complexity we see in natural language.
The second part of my dissertation provides a set of models, along with analytic and simulation results, that show precisely how (and under what circumstances) this process of modular composition is supposed to work.A more detailed summary of this work can be read HERE.
[Official version available here.]
Recommended citation:
LaCroix, Travis. 2020. Complex Signals: Reflexivity, Hierarchical Structure, and Modular Composition. UC Irvine. ProQuest ID: LaCroix_uci_0030D_16213. Merritt ID: ark:/13030/m5ps345j. Retrieved from https://escholarship.org/uc/item/5328x080
Current Projects
Accounting for Polysemy and Role Asymmetry in the Evolution of Compositional Signals
Several formal models of signalling conventions have been proposed to explain how and under what circumstances compositional signalling might evolve. I suggest that these models fail to give a plausible account of the evolution of compositionality because (1) they apparently take linguistic compositionality as their target phenomenon, and (2) they are insensitive to role asymmetries inherent to the signalling game. I further suggest that, rather than asking how signals might come to be compositional, we must clarify what it would mean for signals to be compositional to begin with.
[Unpublished draft available here.] (Please cite official version, if available.)
Recommended citation:
LaCroix, Travis. 2019. "Accounting for Polysemy and Role-Asymmetry in the Evolution of Compositional Signals." Unpublished Manuscript. May 2019, PDF File.
The Autistics and the Machines
Recent work in artificial intelligence claims that large language models (LLMs) exhibit theory of mind (ToM) abilities, based on their performance on classic benchmarks such as false-belief tasks. I argue that, rather than revealing emergent machine cognition, these results expose the normative, exclusionary, and linguistically biased assumptions embedded in how ToM has been operationalised and measured. Historically, ToM tasks emerged as proxies for social understanding and were quickly recruited to pathologise autistic cognition, framing it as a deficit relative to a neurotypical norm. The contemporary celebration of LLMs for "passing" these tests—while autistic people have historically been judged to "fail" them—marks a striking inversion that calls for philosophical scrutiny.
Drawing on philosophy of measurement and critical work in the philosophy of autism, I argue that treating these benchmarks as measures of understanding commits a category mistake and rests on the illusion of a non-existent ground truth for mindreading abilities. More broadly, I situate ToM benchmarking within a ritualised culture of quantification in both psychology and AI, in which proxy measures acquire unwarranted authority and predictive success is mistaken for representational adequacy. This conflation obscures the purposes and normative commitments that shape what is taken to be measurable. In response, the paper motivates a relational and interactional reconception of understanding that resists both machine anthropomorphism and the marginalisation of cognitively diverse humans.
[Preprint forthcoming.] (Please cite official version, if available.)
The Correction Game or, How Pre-Evolved Communicative Dispositions Might Affect Communicative Dispositions
How might pre-evolved communicative dispositions affect how individuals learn to communicate in a novel context? I present a model of learning that varies the reward for coordination in the signalling game framework under simple reinforcement learning as a function of the agents' actions. The model takes advantage of a type of modular compositional communicative bootstrapping by which the sender and receiver use pre-evolved communicative dispositions—a "yes/no" command—to evolve new dispositions.
[Unpublished draft available here.] (Please cite official version, if available.)
Recommended citation:
LaCroix, Travis. 2019. "The Correction Game or, How Pre-Evolved Communicative Dispositions Might Affect Communicative Dispositions." Unpublished Manuscript. April 2019, PDF File.
Learning From Learning Machines: Optimisation, Rules, and Social Norms (with Yoshua Bengio)
There is an analogy between machine learning systems and economic entities in that they are both adaptive, and their behaviour is specified in a more or less explicit way. It appears that the area of AI that is most analogous to the behaviour of economic entities is that of morally good decision-making, but it is an open question as to how precisely moral behaviour can be achieved in an AI system. This paper explores the analogy between these two complex systems, and we suggest that a clearer understanding of this apparent analogy may help us forward in both the socio-economic domain and the AI domain: known results in economics may help inform feasible solutions in AI safety, but also known results in AI may inform economic policy. If this claim is correct, then the recent successes of deep learning for AI suggest that more implicit specifications work better than explicit ones for solving such problems.
[arXiv Preprint available here.] (Please cite published version, if available.)
Recommended citation:
LaCroix, Travis and Yoshua Bengio. 2019. "Learning from Learning Machines: Optimisation, Rules, and Social Norms." arxiv.org/abs/2001.00006.
Relative Principals, Pluralistic Alignment, and the Structural Value Alignment Problem
The value alignment problem for artificial intelligence (AI)—ensuring AI systems act in accordance with human values—is framed as a purely technical or normative challenge, often focused on hypothetical future AI systems. This paper presents a reconceptualisation of value alignment as a structural problem grounded in real-world human–AI interactions. Drawing on the principal–agent framework from economics, value misalignment is described as arising whenever (a) an AI system's objectives are mis-specified or (b) informational asymmetries exist between human principals and artificial agents. On this description, misalignment arises along three interacting axes: objectives (when reward functions fail to capture true goals), information (when opacity or distributional shifts distort performance), and principals (reflecting plural, sometimes conflicting, human interests among developers, users, and affected stakeholders). Viewing alignment through these axes situates it within broader multi-agent coordination problems, revealing that alignment is not a unitary problem with a final solution but an ongoing matter of mitigation. This structural account integrates socio-technical factors—such as power, labour, and environmental dynamics—thereby providing conceptual clarity and practical guidance for addressing misalignment in both current and future AI systems.
[Preprint forthcoming.] (Please cite official version, if available.)
Purpose before Policy: Academic Integrity, Generative AI, and Rhetorical Stance (with Tristan Taylor)
The release of ChatGPT in 2022 sparked polarised responses in higher education, from rapid curricular adoption to warnings about academic misconduct and threats to academic integrity. This article argues that whether generative AI (GenAI) use constitutes misconduct depends on a prior question: what is the university's purpose? We contend that (apparent) rising GenAI-related misconduct cases reflect structural incoherence in the neo-liberal university, where technological enthusiasm, corporate influence, and policy enforcement often conflict. Such misalignment produces moral and institutional ambiguity, leaving students accountable for behaviours implicitly shaped by the institution itself. Drawing on an historical and rhetorical account of the Anglo-American university model, we analyse mission statements from leading university networks to reveal gaps between how institutional ideals are communicated versus how they are implemented. We conclude that universities cannot credibly enforce integrity standards in the age of AI without first ensuring coherence between their stated missions, pedagogical practices, and approaches to emerging technologies.
[Preprint forthcoming.] (Please cite official version, if available.)
Reference by Proxy and Truth-in-a-Model
Simchen (2017) brings to light the notion of 'scrambled truth' to show how productivist metasemantics is able to deal with problems of singular reference in a way that an interpretationist metasemantics (such as Lewisian reference magnetism) cannot. This serves to show that productivism is a live alternative, and indeed a rival to interpretationist metasemantics, and so cannot be subsumed by interpretationist theories.
I examine Simchen's challenge to interpretationist metasemantics by extending his theoretical problem in light of actual communicative exchanges. I show that when the problem is couched in these terms, the ability to refer depends inherently upon coordination—the onus of which is on the receiver. Thus, I show how the interpretationist stance, in this case, can reasonably be understood to encompass the productivist stance.
Saltationist versus Gradualist Approaches to Language Origins: A Critical Discussion
In spite of their vast differences, theories of language origins can be, more or less, partitioned into two exhaustive and mutually exclusive camps: saltationist and gradualist. Saltationism—from the Latin saltus, meaning 'leap'—is the view that (the human-level capacity for) language sprang into existence suddenly and recently, and that there is a complete discontinuity between the linguistic capacities of humans and the communication systems of non-human animals; whereas, gradualism—from the Latin gradus, meaning 'step'—is the view that language evolved slowly over long periods of time. However, rather than arguing for the plausibility of a gradualist versus a saltationist scenario, most researchers appear to fall into one or the other camp based purely upon external or pre-theoretic commitments regarding what they believe evolved [emerged] and how.
The purpose of this paper is to critically survey the respective commitments and entailments of saltationist and gradualist theories of language origins in order to make an explicit argument that the saltationist view is theoretically untenable. Under scrutiny, holding one or the other theoretical stance toward language origins will require or entail certain commitments, which vary in plausibility. It appears that many researchers either ignore these facts or are willing to bite the bullet with respect to them. However, arguments for why one should be so willing are often few and far between.
Strategic Polysemy in AI Discourse: A Philosophical Analysis of Language, Hype, and Power (with Fintan Mallory and Sasha Luccioni)
This paper examines the strategic use of language in contemporary artificial intelligence (AI) discourse, where metaphorical or colloquial terms like "agent", "hallucination", and "chain-of-thought" are coined and rapidly entrenched. These terms are often anthropomorphic and polysemous, which influences their interpretation by the broader community. We argue that this ambiguity is not merely incidental but exploited in the field. We examine how AI researchers, corporations, and other shareholders engage in what we call glosslighting—i.e., the practice of using technically redefined terms to evoke intuitive, often misleading associations while preserving plausible deniability through narrow technical definitions. We illustrate how this linguistic strategy fuels hype cycles, garners funding, and shapes public and policy perceptions of AI, while deflecting epistemic and ethical scrutiny.
[Preprint forthcoming.] (Please cite official version, if available.)
What Russell Can Denote: Aboutness and Denotation Between Principles and 'On Denoting'
How ought we to analyse propositions that are about nonexistent entities? Russell (1903) details the concept of denoting in Principles of Mathematics, and this theory appears to answer the question posed. However, in the paper 'On Denoting' (Russell 1905), we see that his theory of denoting has changed greatly. Hylton (1990) argues that the move from the former theory to the latter was unnecessary. The purpose of this paper is to show that, contra Hylton, the move to the theory found in 'On Denoting' was indeed necessary.
I argue that Hylton is correct to the extent that an answer to our first question relies on a different question concerning the ontology of nonexistent entities. However, this fails to take into account is a more interesting question regarding the truth values of propositions containing such puzzling entities. This question relies on Russell's notion of aboutness, and in this sense is more sensitive to his theory as a complete picture of denotation. If we take the aboutness relation seriously, then we see that the move from the former theory to the latter was necessary after all.
[Unpublished Draft Available Here.] (Please cite official version, if available.)
Recommended citation:
LaCroix, Travis. 2019. "What Russell Can Denote: Aboutness and Denotation Between Principles and 'On Denoting'." Unpublished Manuscript. May 2019, PDF File.
