Tuesday 27 October 2015

Emerging AI and Existential Threats


AI is much in the news recently. Google’s chairman Eric Schmidt believes AI is starting to make real progress whilst others such as Nick Bostrom believe AI might pose an existential danger to humanity (1). In this posting I want first to question whether any real progress is in fact being made and secondly examine the potential dangers involved. Before proceeding I must make it clear I don’t deny real AI is feasible for after all human beings have evolved intelligence. If intelligence can evolve due to natural selection then it seems feasible that it can be created by artificial means however I believe this will be harder to achieve than many people seem to believe.

At present computing power is rising fast and algorithms are increasing in complexity leading to optimism about the emergence of real AI. However it seems to me that larger faster computers and more complex algorithms alone are unlikely to lead to real AI. I will argue genuine intelligence requires a will and as yet no progress has been made to creating for or endowing AI with a will. Famously Hume argued that reason are the slave of the passions. Reason according to Hume is purely instrumental. It might be thought that better computers and better algorithms ought to be better at reasoning. I would question whether they can reason at all because I would suggest that reason cannot be separated from the will. Unlike Hume I would suggest that reason is not the slave of the passions. Reason and the will, the passions, are of necessity bound together. In the present situation seems to me that better computers and better algorithms only mean they are better instruments to serve our will, they don’t reason at all. The output of some computer program may indeed have some form but this form doesn’t have any meaning which is independent of us. The form of its output alone has no more meaning than that of a sand dune sculpted by the wind. However sophisticated computers or algorithms become if the interpretation of their output depends on human beings then they don’t have any genuine intelligence and as a result I believe it is misleading to attribute AI to such computers or algorithms. Real AI in this posting will mean computers, algorithms or robots which have genuine intelligence. Genuine intelligence requires reasoning independently of human beings and this reasoning involves having a will.

Let us accept that if some supposed AI doesn’t have a will that it doesn’t have any genuine intelligence. What then does it mean to have a will? According to Harry Frankfurt,

“The formation of a person’s will is most fundamentally a matter of his coming to care about certain things, and of his coming to care about some of them more than others.” (2)

For something to have a will it must be capable of ‘caring about’ or loving something. If computers, algorithms or robots are mere instruments or tools, in much the same way as a hammer is, then they don’t have any will and real AI is no more than a dream. How might we give a potential AI a will or create the conditions from which a potential AI will acquire an emergent will? Before trying to answer this question I want to consider one further question. If something has a will must we regard it as a person? Let us assume Frankfurt is correct in believing that for something to have a will it must be capable of ‘caring about’ something. Frankfurt argues that something

“to whom its own condition and activities do not matter in the slightest properly be regarded as a person at all. Perhaps nothing that is entirely indifferent to itself is really a person, regardless of how intelligent or emotional or in other respects similar to persons it may be. There could not be a person of no importance to himself.” (3)

Accepting the above means that to have a will is essential to being a person. It also suggests that if something has a will it might be regarded as a person. This suggestion has moral implications for AI. Clearly when we switch off our computers we are not committing murder however if we switched off a computer or terminated an algorithm which had acquired a will we would. I will not follow this implication further here.

Let us return to the question as to whether it is possible to seed a potential AI with a will or create the conditions in which it might acquire one. If we accept Frankfurt’s position then for something to have a will it must satisfy three conditions.

It must be able to ‘care about’ some things and care about some of them more than others.

It must ‘care about itself.

In order to ‘care about’ it must be aware of itself and other things.

Before being able to satisfy conditions 1 and 2 a potential AI must firstly satisfy condition 3. If we program a potential AI to be aware of itself and other things it seems possible we are only programming the AI to mimic awareness. For this reason it might be preferable to try and create the conditions from which a potential AI might acquire an emergent awareness of itself and other things. How might we set about achieving this? The first step must be to give a potential AI a map of the world it will operate in. Initially it need not understand this map and only be able to use it to react to the world. Secondly it must be able to use its reactions with the world to refine this map. If intelligence is to be real then the world it operates in must be our world and the map it creates by refinement must resemble our world. Robots react more meaningfully with our world than computers so perhaps real AI will emerge from robots or robot swarms connected to computers. However it seems to me that creating a map of the things in our world will not be enough for a potential AI to acquire emergent awareness. For any awareness to emerge it must learn to differentiate how different things in that world react to its actions. Firstly it must learn what it can and cannot change by physical action. Secondly and more importantly it must learn to pick from amongst those things it cannot change by physical action the things it can sometimes change by change by simply changing its own state. A potential AI must learn which things are aware of the potential AI’s states and perhaps by doing so become aware of itself satisfying the third of the conditions above. Meeting this condition might facilitate the meeting of the first two conditions.

For the sake of argument let us assume a potential AI can acquire a will and in the process become a real AI. This might be done by the rather speculative process I sketched above. Bostrom believes AI might be an existential threat to humanity. I am somewhat doubtful whether a real AI would pose such a threat. Any so called intelligent machine which doesn’t have a will is an instrument and does not in itself pose an existential threat to us. Of course the way we use it may threaten us but the cause of the threat lies in ourselves in much the same way as nuclear weapons do. However I do believe the change from a potential AI to a real AI by acquiring a will does pose such a theat. Hume argued it wasn’t “contrary to reason to prefer the destruction of the whole world to scratching of my finger.” It certainly seems possible that a potential AI with an emerging will might behave in this way. It might have the will equivalent to that of a very young child whilst at the same time possessing immense powers, possibly the power to destroy humanity. Any parent with a young child who throws a tantrum because he can’t get his own way will appreciate how an emerging AI with immense powers and an emergent will potentially might poses an existential threat.

How might we address such a threat? Alan Turing proposed his Turing test for intelligence. Perhaps we need a refinement of his test to test for good will, such a refinement might called the Humean test. Firstly such a test must test for a good will and secondly, but much more importantly, it must test whether any emergent AI might in any possible circumstances consider the destruction of humanity. Creating such a test will not be easy and it will be difficult to deal with the problem of deceit. Moreover it is worth noting some people, such as Hitler and Pol Pot, might have passed such a test. Nonetheless if an emerging AI is not to pose a threat to humanity the development of such is vital and any potential AI which is not purely an instrument and cannot pass the test should be destroyed even if this involves killing a proto person.


  1.  Nick Bostrom, 2004, Superintelligence, Oxford University Press
  2. Harry Frankfurt, 1988, The Importance of What We Care About. Cambridge University Press, page 91
  3. Frankfurt, 1999 Necessity, Volition, and Love. Cambridge University Press. Page 90.

Friday 18 September 2015

Do Same Sex couples have a greater right to Fertility Treatment?


Emily McTernan “argues that states have greater reason to provide fertility treatment for same sex couples than for heterosexual couples” (1). She bases her argument on the premise that greater access to fertility treatment for same sex couples will encourage a diversity in of ways of life and that this diversity is a social good. In this posting I will argue that she is mistaken and that same sex couples do not have a greater right to fertility treatment.

In what follows I will restrict my discussion of fertility treatment to IVF. McTernan argues that IVF cannot be justified as an element of general healthcare. Healthcare she assumes should be concerned with disease and infertility is not normally a disease. She defines disease as an adverse deviation from normal species functioning and this deviation is a deviation from what is statistically normal given someone’s age and sex. Of course there are exceptions.  A women with a specific problem such as blocked fallopian tubes has a disease and using the above definition has a right to fertility treatment, has a right to IVF. However for most couples, especially if the women is older, infertility is not a deviation from what is statistically normal, and as a result most couples do not have a right to IVF based on a right to healthcare. I agree with McTernan.

Accepting the above of course doesn’t automatically mean people don’t have some right to IVF or that the state shouldn’t provide IVF. After all the state provides such things as libraries, parks and sports fields. Accepting the above only means that the states provision of IVF should compete with the states’ provision of libraries, parks and other things which help its citizens flourish. Let us accept that the state should provide some funding for IVF commensurate with other requirements it has. McTernan argues that within this provision same sex couples should be prioritised in order to encourage a diversity in ways of life.

I now want to argue that McTernan’s argument is unsound and that in allocating IVF we shouldn’t prioritise same sex couples. Firstly I will argue McTernan’s reason for such prioritisation is unsound and secondly present an argument against any such prioritisation. McTernan believes that we should encourage diversity in ways of life. Offering priority in access to IVF to gay couples might increase diversity in child rearing. However if diversity in ways of life is an unqualified good then perhaps the state should reform the law on bigamy and even encourage polygamous marriage as by doing so it would encourage different sorts of relationships. Few people would support such a reform but even if such a reform could be justified other examples could be imagined to show that not all diversity in ways of life are good. Nonetheless let us accept that diversity is sometimes desirable, is a qualified good. It follows diversity in child rearing might be such a qualified good and hence should be encouraged. What exactly does McTernan want to increase diversity in? Does she want to increase diversity in child rearing or simply diversity in relationships? Child rearing involves loving, safe guarding, nurturing and guidance. I don’t believe McTernan wants to change these basics. It follows she wants to increase a diversity in relationships. However the state could also encourage a gay lifestyle, in order to increase a diversity in relationships, by tax incentives. Few would support such a proposal. Such a proposal seems to be mistaken for surely the amount of diversity in sexual orientation in a society should be determined by people’s natural inclinations rather than by government policy. Accepting the above means of course if sexual orientation in a society should be determined by people’s natural inclinations rather than by the state that the state has reason to permit gay marriage. Accepting the above also means the state has no reason to prioritise access to IVF for gay couples.

Gay couples cannot have children unaided by anyone else. It might be suggested that this fact means gay couples should be prioritised in accessing IVF. However even if gay couples cannot have children unaided by others IVF is not the only option open to them if they want to have children. Both male and female same sex couples might be able to adopt a child. Male couples might also use surrogacy and this need not involve IVF. Female couples can use AID. It seems to me the fact that gay couples cannot have children unaided does not mean they should be prioritised in accessing IVF.

I now want to argue there is a second reason as to why gay couples should not be given greater priority in accessing IVF. My argument is based on fairness. Let us assume that gay couples are given greater priority in accessing IVF. It might then be objected such prioritisation is unfair. Fairness requires that everybody’s needs are considered. It does not follow of course that everybody’s needs should be satisfied equally. However it does require that if some peoples’ needs aren’t satisfied equally that some reason can be given for this. Let us assume that people have a need to have children who are genetically related to them. Let us consider a gay and a heterosexual couple both of whom are unable to conceive children without IVF. Both couples have the same need. Fairness requires that if the needs of these couples are satisfied unequally that there that some reason can be given for this unequal satisfaction. The need of both couples are identical, to have children they are genetically related to. If the needs of both couples are the same then any reason given for unequal treatment must depend on the outcomes for any children so conceived or some benefit to society. The outcomes for any children depends on the parenting skills of the couples involved. Perhaps for instance either gay or heterosexual couples make better parents. However there seems no evidence to support such a reason. Perhaps then society might benefit from unequal satisfaction. It is difficult to see how society might benefit except for promotion of greater diversity, but I have argued above that whilst society must permit greater diversity it should not try to alter the natural diversity occurring within it. In conclusion it would seem that the encouragement of a diversity in ways of life does not give us a reason to prioritise IVF for gay couples over heterosexual couples. It further seems that fairness requires that all couples are given equal priority.


  1. Emily McTernan, 2015, Should Fertility Treatment be State Funded? Journal of Applied Philosophy, 32,3. Page 237



Historic wrongdoing, Slavery, Compensation and Apology

      Recently the Trevelyan family says it is apologising for its ancestor’s role in slavery in the Caribbean, see The Observer .King Ch...