I recently published an unusual article. At least, I think it is unusual. It imagines a future in which sophisticated sex robots are used to replicate acts of rape and child sexual abuse, and then asks whether such acts should be criminalised. In the article, I try to provide a framework for evaluating the issue, but I do so in what I think is a provocative fashion. I present an argument for thinking that such acts should be criminalised, even if they have no extrinsically harmful effects on others. I know the argument is going to be unpalatable to some, and I myself balk at its seemingly anti-liberal/anti-libertarian dimensions, but I thought it was sufficiently interesting to be worth spelling out in some detail. Hence why I wrote the article.
For the detail, you’ll have to read the original paper (available here, here, and here). But in an effort to entice you to do that, I thought I would use this post to provide a brief overview.
1. What is robotic rape and robotic child sexual abuse?
First things first, it is worth clarifying the phenomena of interest. I’m sure people have a general sense of what a sex robot is, and maybe some vaguer sense of what an act of robotic rape or child sexual abuse might be, but it’s worth being as clear as possible at the outset in order to head-off potential sources of confusion. So let’s start with the notion of a sex robot. In the article, I define a sex robot as any artifact that is used for the purposes of sexual stimulation and/or release with the following three properties: (i) a human-like form; (ii) the ability to move; and (iii) some degree of artificial intelligence (i.e. an ability to interpret, process and act upon information from its environment).
As you can see from this definition, my focus is on human-like robots not on robots with more exotic properties, although I briefly allude to those possibilities in the article. This is because my argument appeals to the social meaning that might attach to the performance of sexual acts with human-like representations. For me, the degree of human-likeness is a function of the three properties included in my definition, i.e. the more human-like in appearance, movement and intelligence, the more human-like the robot is deemed to be. For my argument to work (if it works at all) the robots in question must cross some minimum threshold of human-likeness, but I don’t know where that threshold lies.
So much for sex robots. What about acts of robotic rape and robotic child sexual abuse? Acts of robotic rape are tricky to define given that legal definitions of rape differ across jurisdictions. I follow the definition in England and Wales. Thus, I view rape as being non-consensual sexual intercourse performed in the absence of a reasonable belief in consent. I then define robotic rape as sexual intercourse performed with a robot that mimics signals of non-consent, where it would be unreasonable for the performer of those acts to deny that the robot was mimicking signals of non-consent. I know there is some debate as to what counts as a signal of non-consent. I try to sidestep this debate in the article by focusing on what I call “paradigmatic signals of non-consent”. I accept that the notion of a paradigmatic signal of non-consent might be controversial. Acts of robotic child sexual abuse are easier to define. They arise whenever sexual acts are performed with robots that look and act like children.
Throughout the article, I distinguish robotic acts from virtual acts. The former are performed by a human actor with a real, physical robot partner. The latter are performed in a virtual world via an avatar or virtual character. There are, however, borderline cases, e.g. virtual acts performed using immersive VR technology with haptic sensors (e.g. such as those created by the Dutch company Kiiroo). I am unsure about the criminalisation argument in such cases, for reasons that will become clearer in a moment.
2. What is the prima facie argument for criminalisation?
With that definitional work out of the way, I can develop the main argument. That argument proceeds in a particular order. It starts by focusing on the purely robotic case, i.e. the case in which the robotic acts have no extrinsic effects on others. It argues that even in such a case, there may be grounds for criminalisation. That gives me a prima facie argument for criminalisation. After that, I focus on extrinsic effects, and suggest that they are unlikely to defeat this prima facie argument. Let’s see how all this goes.
The prima facie argument works like this:
- (1) It can be a proper object of the criminal law to regulate conduct that is morally wrong, even if such conduct has no extrinsically harmful effects on others (the moralistic premise).
- (2) Purely robotic acts of rape and child sexual abuse fall within the class of morally wrong but extrinsically harmless conduct that it can be a proper object of the criminal law to regulate (the wrongness premise).
- (3) Therefore, it can be a proper object of the criminal law to regulate purely robotic acts of rape and child sexual abuse.
I don’t really defend the first premise of the argument in the article. Instead, I appeal to the work of others who have. For example, Steven Wall has defended a version of legal moralism that argues that actions involving harm to the performer’s moral character can, sometimes, be criminalised; likewise, Antony Duff has argued that certain public wrongs are apt for criminalisation even when they do not involve harm to others. I use both accounts in my article and suggest that if I can show that purely robotic acts of rape and child sexual abuse involve harm to moral character or fall within Duff’s class of public wrongs, then I can make the prima facie case for criminalisation.
This first premise is likely to be difficult for many, particularly those with a classic liberal or Millian approach to criminalisation. They will argue that only harm to others renders something apt for criminalisation. I sympathise with this view (which is why I am cagey about the argument as a whole) but, again, appeal to others who have tried to argue against it by showing that a more expansive form of legal moralism need not constitute a severe limitation of individual liberty and how it may be very difficult to consistently hold to the liberal view. I also try to soften the blow by highlighting different possible forms of criminalisation at the end of article (e.g. incarceration need not be the penalty). Still, even then I accept that my argument may simply lead some to question the moralistic principles of criminalisation upon which I rely.
Premise two is where I focus most of my attention in the article. I defend it in two ways, each way corresponding to a different version of legal moralism. First, I argue that purely robotic acts of rape and child sexual abuse may involve harm to moral character. This is either on the grounds that the performance of such acts encourages/requires the expression of a desire for the real-world equivalents, or on the grounds that the performance requires a troubling insensitivity to the social meaning of those acts. This is consistent with Wall’s version of moralism. Second, I build upon this by arguing that the insensitivity to social meaning involved in such acts (particularly acts of robotic rape) would allow for them to fall within Duff’s class of public wrongs. The idea being that in a culture that has condoned or belittled the problem of sexual assault, an insensitivity to the meaning of those acts demands some degree of public accountability.
In defending premise (2) I rely heavily on work that has been done on the ethics of virtual acts and fictional representations, particular the work of Stephanie Patridge. This reliance raises an obvious objection. There are those — like Gert Gooskens — who argue that our moral characters are not directly implicated in the performance of virtual acts because there is some distance between our true self and our virtual self. I respond to Gooskens by pointing out that the distance is lessened in the case of robotic acts. I rely on some work in moral psychology to support this view.
That is my defence of the prima facie argument.
3. Can the prima facie argument be defeated?
But it is important to realise how modest that argument really is. It only claims that robotic rape and robotic child sexual abuse are apt for criminalisation all else being equal. It does not claim that they are apt for criminalisation all things considered. The argument is vulnerable to defeaters. I consider two general classes of defeaters in the final sections of the paper.
The first class of defeaters is concerned with the possible effects of robotic rape and robotic child sexual abuse on the real-world equivalents of those acts. What if having sex with a child-bot greatly reduced the real-world incidence of child sexual abuse? Surely then we would be better off permitting or facilitating such acts, even if they do satisfy the requirements of Duff or Wall’s versions of moralism? This sounds right to me, but of course it is an empirical question and we have no real evidence as of yet. All we can do for now is speculate. In the article, I speculate about three possibilities. Robotic rape and robotic child sexual abuse may: (a) significantly increase the incidence of real-world equivalents; (b) significantly reduce the incidence of real-world equivalents; or (c) have an ambiguous effect. I argue that if (a) is true, the prima facie argument is strengthened (not defeated); if (b) is true, the prima facie argument is defeated; and if (c) is true then it is either unaffected or possibly strengthened (if we accept a recent argument from Leslie Green about how we should use the criminal law to improve social morality).
The second class of defeaters is concerned with the costs of an actual criminalisation policy. How would it be policed and enforced? Would this not involve wasteful expenditure and serious encroachments on individual liberty and privacy? Would it not be overkill to throw the perpetrators of such acts in jail or subject them to other forms of criminal punishment? I consider all these possibilities in the article and suggest various ways in which the costs may not be as significant as we first think.
So that’s it. That is my argument. There is much more detail and qualification in the full version. Just to be clear, once again, I am not advocating criminalisation. I am genuinely unsure about how we should approach this phenomenon. But I think it is an issue worth debating and I wanted to provide a (provocative) starting point for that debate.
You have never responded to Nick S.'s comment in response to this blog post of yours on another blog:
ReplyDelete"Haven’t read the paper, just this article, but there seems to be an important question entirely overlooked. At the moment, it is perfectly legal for two or more consenting adult partners to undertake sexual intercourse as an act of simulated rape or simulated child abuse. Why should it be legal for couples (or entire dungeons of consenting adults) to explore such fantasies, but not legal for a lone masturbator, playing with a sex toy? It doesn’t seem to make an awful lot of sense, unless you’re proposing that consenting adult partners should also be forbidden from engaging in role-playing of this kind."
So, what exactly is the meaningful difference between 2+ consenting adults engaging in simulated rape or simulated child abuse and a single adult doing it with a sex doll/robot? Why exactly should the former be considered legal and acceptable while the latter should be considered illegal and unacceptable?
Please keep in mind that very rarely an adult can look like an actual child. Shauna Rae, for instance, is 22 (or perhaps even slightly older now) but looks like she's 8. If there were millions of adults who looked like Shauna Rae, and who were willing to have consensual sex with pedophiles while also being willing to engage in ageplay and rape fantasy roleplaying together with pedophiles in the bedroom, many people would certainly find such behavior to be extremely creepy and bizarre but at the same time likely strongly balk at actually criminalizing this behavior or even firing someone from their job for engaging in this type of behavior. So, why exactly should people's attitude be any different if a single adult did all of these same things together with a child sex doll/robot instead?
There's also another factor here worth considering: Banning child sex dolls/robots and rapebots could significantly reduce some people's quality of life, similarly to how banning sodomy in the old days likely significantly reduced gay people's quality of life. If the net effect of making these things legal on children is neutral or ambiguous, then the case in favor of legalization would still be very strong due to the improved quality of life that this would provide for some adults. That's certainly nothing to scoff at, especially if some adults have no other harm-free way to achieve long-term sexual satisfaction.
I also think that there is another distinction to be made here: One can be sexually attracted to something (whether rape, child abuse, or something else) without actually "wanting it". For instance, me having sex with Ali Larter or Liz Hurley would be hot and sexually arousing, but I don't "want it" because I know that they would never actually give their consent to this. (Of course, you could say that I would want it if they were actually willing to give consent to this, but then one could likewise argue that the virtuous minor-attracted person would want to have sex with a person who looks like a child if this person was actually an adult of sound mind who simply looked like a child--as with my Shauna Rae example above--and who was actually willing to give consent to this, but also that virtuous minor-attracted persons would not be willing to have sex with anyone who cannot legally consent to sex, regardless of just how attractive they considered them to be.)
DeleteI suppose that one could argue that engaging in simulated rape and simulated child abuse exhibits a poor moral character to some degree, but surely this should be significantly compensated by one being unwilling--unlike many other people--to engage in the actual act itself (actual rape and actual child abuse), no? If someone rejects ever doing evil act A but is nevertheless willing to do B, which is a harmless simulation of evil act A, then such a person should still be viewed as being astronomically more praiseworthy than anyone who ever actually engages in evil act A itself. And as I mentioned above, what if engaging in symbolic harm is actually necessary for some people to have a good quality of life? This is true for child sex dolls/robots, rapebots, et cetera, but could also be true, say, in a purely hypothetical scenario where some white supremacist or black supremacist or whatever is clinically depressed and can only ever come out of this clinical depression and actually become a properly functional adult by placing an extremely violent and racist video game or whatever. In such a scenario, would it really be moral to arrest them for playing this game or even to merely legally prohibit them from playing this game?
Ultimately, I'm extremely wary of punishing people for thought crimes and for harmless simulations that result from these thought crimes. Should Kathy Griffin be arrested for her symbolically offensive Trump decapitation stunt, for instance?
One more point worth addressing here: There is a difference between engaging in an action which is intended to be offensive, such as, say, calling a black person the N-word in the middle of a fight or argument with them, and an action that is interpreted as offensive but isn't actually intended as such. Someone who orders and uses child sex dolls/robots and rapebots is primarily trying to satisfy their sex drive in the privacy of their own home (just like 2+ consenting adults who engage in rape fantasy roleplaying or realistic ageplay do the very same thing), not primarily trying to cause offense to children or rape victims. Just like, say, someone who draws a Muhammad cartoon might--at least sometimes--be trying to make (non-Muslim) people laugh rather than to offend Muslims; the extreme offense that many Muslims take to such cartoons is simply an unintended side effect of this.
Delete"This is true for child sex dolls/robots, rapebots, et cetera, but could also be true, say, in a purely hypothetical scenario where some white supremacist or black supremacist or whatever is clinically depressed and can only ever come out of this clinical depression and actually become a properly functional adult by *playing* an extremely violent and racist video game or whatever. In such a scenario, would it really be moral to arrest them for playing this game or even to merely legally prohibit them from playing this game?"
Delete(Corrected typo.)