Written by Dr John Danaher.
Dr Danaher is a Lecturer in Law at NUI Galway. His research interests include neuroscience and law, human enhancement, and the ethics of artificial intelligence.
A version of this post was previously published here.
Somebody recently sent me a link to an article by Jed Radoff entitled “Why Innocent People Plead Guilty”. Radoff’s article is an indictment of the plea-bargaining system currently in operation in the US. Unsurprisingly given its title, it argues that the current system of plea bargaining encourages innocent people to plead guilty, and that something must be done to prevent this from happening.
I recently published a paper addressing the same problem. The gist of its argument is that I think that it may be possible to use a certain type of brain-based lie detection — the P300 Concealed Information Test (P300 CIT) — to rectify some of the problems inherent in systems of plea bargaining. The word “possible” is important here. I don’t believe that the technology is currently ready to be used in this way – I think further field testing needs to take place – but I don’t think the technology is as far away as some people might believe either.
What I find interesting is that, despite this, there is considerable resistance to the use of the P300 CIT in academic and legal circles. Some of that resistance stems from unwarranted fealty to the status quo, and some stems from legitimate concerns about potential abuses of the technology (miscarriages of justice etc.). I try to overcome some of this resistance by suggesting that the P300 CIT might be better than other proposed methods for resolving existing abuses of power within the system. Hence my focus on plea-bargaining and the innocence problem.
Practical ethicists have become increasingly interested in the potential applications of neurointerventions—interventions that exert a direct biological effect on the brain. One application of these interventions that has particularly stimulated moral discussion is the potential use of these interventions to prevent recidivism amongst criminal offenders. To a limited extent, we are already on the path to using what can be described as neuro-interventions in this way. For instance, in certain jurisdictions drug-addicted offenders are required to take medications that are intended to attenuate their addictive desires. Furthermore, sex-offenders in certain jurisdictions may receive testosterone-lowering drugs (sometimes referred to as ‘chemical castration’) as a part of their criminal sentence, or as required by their conditions of parole.
On 13-14th April, a workshop (funded by the Wellcome Trust) focussing on the moral questions raised by the potential use of neuro-interventions to prevent criminal recidivism took place at Kellogg College in Oxford. I lack the space here to adequately explore the nuances of all of the talks in this workshop. Rather, in this post, I shall briefly explain some of the main themes and issues that were raised in the fruitful discussions that took place over the course of the workshop, and attempt to give readers at least a flavour of each of the talks given; I apologise in advance for the fact that I must necessarily gloss over a number of interesting details and arguments. Continue reading
The paper, “The Medicalization of Love” by Brian D. Earp, Anders Sandberg, and Julian Savulescu, has been accepted for publication at the Cambridge Quarterly of Healthcare Ethics. Scholars interested in submitting a short reply paper or peer commentary are encouraged to contact the editor, Tomi Kushner, at email@example.com.
The final deadline for commentaries/ papers is September 1st. The abstract for the paper is below; the accepted manuscript is available at this link. Inquiries to the editor should be sent as soon as possible.
Pharmaceuticals or other emerging technologies could be used to enhance (or diminish) feelings of lust, attraction, and attachment in adult romantic partnerships. While such interventions could conceivably be used to promote individual (and couple) well-being, their widespread development and/or adoption might lead to “medicalization” of human love and heartache—for some, a source of serious concern. In this essay, we argue that the “medicalization of love” need not necessarily be problematic, on balance, but could plausibly be expected to have either good or bad consequences depending upon how it unfolds. By anticipating some of the specific ways in which these technologies could yield unwanted outcomes, bioethicists and others can help direct the course of love’s “medicalization”—should it happen to occur—more toward the “good” side than the “bad.”
* image from http://www.metalsucks.net/2014/02/16/sunday-lurve/.
Things I’ve learned (so far) about how to do practical ethics
I had the opportunity, a few months back, to look through some old poems I’d written in high school. Some, I thought, were pretty good. Others I remembered thinking were good when I wrote them, but now they seem embarrassingly bad: pseudo-profound, full of clichés, marked by empty rhetoric instead of meaningful content. I’ve had a similar experience today with my collection of articles here at the Practical Ethics blog. And Oh, the things I have learned!
Here are just a few of the lessons that have altered my thinking, or otherwise informed my views about “doing” practical ethics — particularly in a public-engagement context — since my very first blog post appeared in 2011:
Announcement: “Brave New Love” – peer commentaries due October 7
Dear Practical Ethics readers,
The paper, “Brave new love: the threat of high-tech ‘conversion’ therapy and the bio-oppression of sexual minorities” by Brian D. Earp, Anders Sandberg, and Julian Savulescu has been accepted for publication in the American Journal of Bioethics: Neuroscience. Proposals for open peer commentaries are due this Monday October 7th.
The article may be accessed here, or at the following link: http://editorial.bioethics.net. Be sure to select AJOB:Neuroscience from the drop-down menu of journals. Here is an abstract of the argument:
Abstract: Our understanding of the neurochemical bases of human love and attachment, as well as of the genetic, epigenetic, hormonal, and experiential factors that conspire to shape an individual’s sexual orientation, is increasing exponentially. This research raises the vexing possibility that we may one day be equipped to modify such variables directly, allowing for the creation of “high-tech” conversion therapies or other suspect interventions. In this paper, we discuss the ethics surrounding such a possibility, and call for the development of legal and procedural safeguards for protecting vulnerable children from the application of such technology. We also consider the more difficult case of voluntary, adult “conversion” and argue that in rare cases, such attempts might be permissible under strict conditions.
Open Peer Commentary articles are typically between 500-1500 words and contain no more than 10 references. A guide to writing an Open Peer Commentary is available under the Resources section “Instructions and Forms” at http://editorial.bioethics.net. AJOB:Neuroscience asks that by Monday, October 7, 2013 you submit a short summary of your proposed Open Peer Commentary (no more than 1-2 paragraphs). Please submit your proposal online via the AJOB:Neuroscience Editorial site, following the instructions provided there. They ask that you do not prepare a full commentary yet. Once they have evaluated your proposal, they will contact you via email to let you know whether or not they were able to include you on the final list of those to be asked to submit an Open Peer Commentary.
You will then have until Friday, October 25, 2013 to submit your full Open Peer Commentary.
Antonio De Salles, Professor of Neurosurgery – UCLA
Lincoln Frias, postdoct UFMG-Brazil, International Neuroethics Society
Jorge Moll, D’Or Institute-Brazil, International Neuroethics Society
Psychosurgery has a bad name. The destruction or disconnection of brain tissue to treat mental illness was brought into disrepute by controversial figures of the past, who performed lobotomies with poorly defined clinical indications and without respect to even the most basic surgical practices of asepsis and hemostasis. The procedures were irreversible, unsafe, and often done without adequate informed consent. In many cases the surgeries drastically reduced the patients’ well-being and autonomy. To avoid this, governments put in place stringent regulations on these procedures. Coupled with developments in psychopharmacology, this left psychosurgery only as a last resort for extreme cases. The moral problem is that the stereotypes and stigma evoked by this kind of treatment are largely inadequate given current technology.