Skip to content

neuroethics

Reversibility, Colds, and Neurosurgery

By Jonny Pugh

This blog was originally published on the Journal of Medical Ethics Blog

 

Happy new year to readers of the blog!

I always approach the new year with some trepidation. This is not just due to the terrible weather, or even my resolution to take more exercise (unfortunately in the aforementioned terrible weather). Instead, I approach January with a sense of dread because it is always when I seem to come down with the common cold.

In my recent research, I have been interested in the nature and moral significance of reversibility, and the common cold is an interesting case study of this concept. In this blog, I will use this example to very briefly preview a couple of points that I make in a forthcoming open access article about reversibility in the context of psychiatric neurosurgery. You can read the open access paper here.

Read More »Reversibility, Colds, and Neurosurgery

Moral Responsibility and Interventions

Written by Gabriel De Marco

Consider a story about Joe, Louie, and Dr. White. Joe is a gambling man and has been for much of his life. In his late twenties, Joe began to gamble occasionally and after a while, he decided that he would embrace this practice of gambling. Although Joe gambles fairly often, he has his limits, and can often resist the desire to gamble.

Louie, on the other hand, is a frugal family man. With his wife, he has been saving money over the last year so that they can take their kids to Disneyland. Dr. White, an evil neurosurgeon who detests the thought of children enjoying themselves at Disneyland, wants to stop this trip. So, Dr. White designs and executes a plan. One night, while Louie is sleeping, Dr. White uses his fancy neuroscientific methods to make Louie more like Joe. He implants in Louie a strong desire to gamble, as well as further attitudes that will help Louie embrace this desire, such that Louie, for example, now values the thrill of gambling, and he desires that his gambling desires are the ones that lead him to action. In order to increase chances of success, Dr. White also significantly weakens some of Louie’s competing attitudes, like some of his family values, or his attitudes towards frugality. When Joe wakes up the next morning, he feels this strong desire to gamble, and although he finds it strange that it has come out of the blue, he fully embraces it (as much as Joe embraces his own gambling desires), having recognized that it lines up with some of his other attitudes about his desires (which were also implanted). Later in the day, while he is “out running errands,” Louie swings by a casino, bets the money he has been saving for the trip, and loses it. “Great success” thinks Dr. White. Since his goal of preventing some children’s joy at Disneyland has been achieved, he turns Louie back into his old self after Louie goes to sleep.

This story is similar to stories sometimes found in the debate about freedom and moral responsibility, though I will focus on moral responsibility. Intuitively, Louie is not morally responsible for gambling away these savings; or, at the very least, he is significantly less responsible for doing so than someone like Joe would be for doing something similar. If we want to make sense of these different judgments about Louie and Joe’s responsibility, we are going to need to find some difference between them that can explain why Louie is, at least, less responsible than regular Joe.

Read More »Moral Responsibility and Interventions

Cross Post: Fresh Urgency in Mapping Out Ethics of Brain Organoid Research

  • by

File 20181120 161641 npf87x.jpg?ixlib=rb 1.1

Written by Julian Koplin, University of Melbourne and

Julian Savulescu, University of Oxford

This article is republished from The Conversation under a Creative Commons license. Read the original article.

 

Researchers have grown groups of brain cells in the lab –
known as ‘organoids’ – that produce brain waves resembling
those found in premature infants.
from www.shutterstock.com

 

Scientists have become increasingly adept at creating brain organoids – which are essentially miniature human brains grown in the laboratory from stem cells.

Although brain organoid research might seem outlandish, it serves an important moral purpose. Among other benefits, it promises to help us understand early brain development and neurodevelopmental disorders such as microcephaly, autism and schizophrenia.

Read More »Cross Post: Fresh Urgency in Mapping Out Ethics of Brain Organoid Research

UK Supreme Court Decision Means Patients No Longer Forced to Live

By Mackenzie Graham

On July 30, The UK’s Supreme Court ruled that there is no requirement to obtain court approval before withdrawing clinically assisted nutrition and hydration (CANH), when there is agreement between physicians and the family that this is in the best interests of the patient.

In the judgement, Lady Black writes:

“If the provisions of the MCA [Mental Capacity Act] 2005 are followed and the relevant guidance observed, and if there is agreement upon what is in the best interests of the patient, the patient may be treated in accordance with that agreement without application to the court.”

Until now, requests to withdraw CANH needed to be heard by the Court of Protection to determine if withdrawing treatment was in the patient’s best interest. In addition to being emotionally difficult for families, this is a time-consuming and expensive process, and often results in the patient dying before a judgement is rendered.

I think this decision has much to be said in its favour. First, it means that when there is agreement that continued treatment is no longer in the best interests of a patient with a prolonged disorder of consciousness, these patients are no longer being ‘forced to live’ until the Court affirms that being allowed to die is in their best interests. In many cases, court decisions take months, meaning that a patient is forced to be kept alive, against their best interests and the wishes of their family. Making the decision to withdraw care from a loved one is highly distressing, and this is likely further compounded by the burden and distraction of court proceedings.

Read More »UK Supreme Court Decision Means Patients No Longer Forced to Live

Pain for Ethicists: What is the Affective Dimension of Pain?

This is my first post in a series highlighting current pain science that is relevant to philosophers writing about well-being and ethics.  My work on this topic has been supported by the W. Maurice Young Centre for Applied Ethics, the Oxford Uehiro Centre for Practical Ethics, and the Wellcome Centre for Ethics and Humanities, as well as a generous grant from Effective Altruism Grants

There have been numerous published cases in the scientific literature of patients who, for various reasons, report feeling pain but not finding the pain unpleasant. As Daniel Dennett noted in his seminal paper “Why You Can’t Make A Computer That Feels Pain,” these reports seem to be at odds with some of our most basic intuitions about pain, in particular the conjunction of our intuitions that ‘‘a pain is something we mind’’ and ‘‘we know when we are having a pain.’’ Dennett was discussing the effects of morphine, but similar dissociations have been reported in patients who undergo cingulotomies to treat terminal cancer pain and in extremely rare cases called “pain asymbolia” involving damage to the insula cortex.Read More »Pain for Ethicists: What is the Affective Dimension of Pain?

Addiction, Desire, and The Polluted Environment – Richard Holton’s 2nd Uehiro Lecture

By Jonathan Pugh

 

In the second of his three Uehiro lectures on the theme of ‘illness and the social self’, Richard Holton turned to the moral questions raised by addiction. In the first half of the lecture, he outlined an account of addictive behaviour according to which addictive substances disrupt the link between wanting and liking. In the second half of the lecture, he discusses the implications of this account for the moral significance of preferences, and for how we might structure environments to avoid triggering addictive desires.

 

You can find a recording of the lecture here

 

Read More »Addiction, Desire, and The Polluted Environment – Richard Holton’s 2nd Uehiro Lecture

Video Series: Tom Douglas Defends the Chemical Castration of Sex Offenders

The Minister of Justice in the UK wants to dramatically increase the use of chemical castration in sex offenders to reduce their risk of reoffending.Dr Tom Douglas (University of Oxford) argues that offering chemical castration to sex offenders might be a better option than current practices to prevent sex offenders from reoffending (e.g. incarceration), and… Read More »Video Series: Tom Douglas Defends the Chemical Castration of Sex Offenders

Neuroblame?

Written by Stephen Rainey

Brain-machine interfaces (BMIs), or brain-computer interfaces (BCIs), are technologies controlled directly by the brain. They are increasingly well known in terms of therapeutic contexts. We have probably all seen the remarkable advances in prosthetic limbs that can be controlled directly by the brain. Brain-controlled legs, arms, and hands allow natural-like mobility to be restored where limbs had been lost. Neuroprosthetic devices connected directly to the brain allow communication to be restored in cases where linguistic ability is impaired or missing.

It is often said that such devices are controlled ‘by thoughts’. This isn’t strictly true, as it is the brain that the devices read, not the mind. In a sense, unnatural patterns of neural activity must be realised to trigger and control devices. Producing the patterns is a learned behaviour – the brain is put to use by the device owner in order to operate it. This distinction between thought-reading and brain-reading might have important consequences for some conceivable scenarios. To think these through, we’ll indulge in a little bit of ‘science fiction prototyping’.

Read More »Neuroblame?

Guest Post: Mind the accountability gap: On the ethics of shared autonomy between humans and intelligent medical devices

Guest Post by Philipp Kellmeyer

Imagine you had epilepsy and, despite taking a daily cocktail of several anti-epileptic drugs, still suffered several seizures per week, some minor, some resulting in bruises and other injuries. The source of your epileptic seizures lies in a brain region that is important for language. Therefore, your neurologist told you, epilepsy surgery – removing brain tissue that has been identified as the source of seizures in continuous monitoring with intracranial electroencephalography (iEEG) – is not viable in your case because it would lead to permanent damage to your language ability.

There is however, says your neurologist, an innovative clinical trial under way that might reduce the frequency and severity of your seizures. In this trial, a new device is implanted in your head that contains an electrode array for recording your brain activity directly from the brain surface and for applying small electric shocks to interrupt an impending seizure.

The electrode array connects wirelessly to a small computer that analyses the information from the electrodes to assess your seizure risk at any given moment in order to decide when to administer an electric shock. The neurologist informs you that trials with similar devices have achieved a reduction in the frequency of severe seizures in 50% of patients so that there would be a good chance that you benefit from taking part in the trial.

Now, imagine you decided to participate in the trial and it turns out that the device comes with two options: In one setting, you get no feedback on your current seizure risk by the device and the decision when to administer an electric shock to prevent an impending seizure is taken solely by the device.

This keeps you completely out of the loop in terms of being able to modify your behaviour according to your seizure risk and – in a sense – relegates some autonomy of decision-making to the intelligent medical device inside your head.

In the other setting, the system comes with a “traffic light” that signals your current risk level for a seizure, with green indicating a low, yellow a medium, and red a high probability of a seizure. In case of an evolving seizure, the device may additionally warn you with an alarm tone. In this scenario, you are kept in the loop and you retain your capacity to modify your behavior accordingly, for example to step from a ladder or stop riding a bike when you are “in the red.”

Read More »Guest Post: Mind the accountability gap: On the ethics of shared autonomy between humans and intelligent medical devices

A Second Response to Professor Neil Levy’s Leverhulme Lectures.

  • by

Written by Richard Ngo , an undergraduate student in Computer Science and Philosophy at the University of Oxford.

Neil Levy’s Leverhulme Lectures start from the admirable position of integrating psychological results and philosophical arguments, with the goal of answering two questions:

(1) are we (those of us with egalitarian explicit beliefs but conflicting implicit attitudes) racist?

(2) when those implicit attitudes cause actions which seem appropriately to be characterised as racist (sexist, homophobic…), are we morally responsible for these actions?Read More »A Second Response to Professor Neil Levy’s Leverhulme Lectures.